6th European Conference - Academic Conferences
6th European Conference - Academic Conferences
6th European Conference - Academic Conferences
Create successful ePaper yourself
Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.
Proceedings<br />
of the<br />
2nd International<br />
<strong>Conference</strong> on Information<br />
Management and<br />
Evaluation<br />
Ryerson University<br />
Toronto<br />
Canada<br />
27-28 April 2011<br />
Edited by<br />
Professor Kenneth Grant<br />
Ryerson University<br />
Toronto<br />
Canada
Copyright The Authors, 2011. All Rights Reserved.<br />
No reproduction, copy or transmission may be made without written permission from the individual authors.<br />
Papers have been double-blind peer reviewed before final submission to the conference. Initially, paper<br />
abstracts were read and selected by the conference panel for submission as possible papers for the<br />
conference.<br />
Many thanks to the reviewers who helped ensure the quality of the full papers.<br />
These <strong>Conference</strong> Proceedings have been submitted to Thomson ISI for indexing.<br />
Further copies of this book and previous year’s proceedings can be purchased from http://academicconferences.org/2-proceedings.htm<br />
ISBN:97-1-906638-97-9 Book<br />
Published by <strong>Academic</strong> Publishing International Limited<br />
Reading<br />
UK<br />
44-118-972-4148<br />
www.academic-publishing.org
Contents<br />
Paper Title Author(s) Page<br />
No.<br />
Preface v<br />
Biographies of <strong>Conference</strong> Chairs,<br />
Programme Chair, Keynote Speaker and<br />
Mini-track Chairs<br />
Biographies of contributing authors vi<br />
Performance Management of the NHS’<br />
Mental Health Care Service Delivery in<br />
England: The Role of the Service’ Actors’<br />
(Clients) Collaboration Through Dialogism<br />
The Acceptance and Use of Educational<br />
Technology in Kingdom of Bahrain<br />
Comparison of Feature Selection Techniques<br />
Using Fully-Controlled Simulation-Based<br />
Datasets<br />
A Delphi-Multi-Criteria Decision Making<br />
Approach in the Selection of an Enterprise-<br />
Wide Integration Strategy<br />
Selection, Implementation and Post<br />
Production of an ERP System<br />
Evolution of eProcurement Model in the<br />
Italian Public Sector: From Government<br />
Centralisation to Regional Delocalisation<br />
ICTs as Weapons of Mass Interaction -<br />
Motivations and Implications of Mediated<br />
Control<br />
Investigating the Factors Inhibiting SMEs<br />
from Recognizing and Measuring Losses<br />
From Cyber Crime in South Africa<br />
Kwesi Korsa Aggrey 1<br />
Jaflah Al-Ammary 9<br />
Suzan Arslanturk, Mohammad Siadat,<br />
Theophilus Ogunyemi , Ishwar Sethi and<br />
Ananias Diokno<br />
Tridip Bardhan, James Ngeru, and Richard<br />
Pitts<br />
Imran Batada and Asmita Rahman 38<br />
Clara Benevolo and Renata Paola Dameri 45<br />
Kofi Agyenim Boateng 54<br />
Gino Bougaardt and Michael Kyobe 62<br />
Sharing Knowledge – The CoP way Sheryl Buckley and Apostolos Giannakopoulos 72<br />
An Empirical Framework of key Success<br />
Factors for Software Process Improvement<br />
Assessing Information Management<br />
Competencies in Organisations<br />
A Semiotic Approach to Analyse the<br />
Influencing Factors in Knowledge Transfer<br />
Exploring the Emotional Exhaustion of<br />
Healthcare Providers Using an In-Hospital<br />
Employee Communication Network<br />
The Implementation of RSS-Based Clinical<br />
Reminders<br />
Factors Determining the Adoption of ICTs in<br />
Healthcare Service Delivery: A Developing<br />
Country Context<br />
Evaluation of Application Embedded<br />
Knowledge Migration Issues<br />
Nuntarat Bunditwongrat, Mathupayas<br />
Thongmak and Atcharawan Ngarmyarn<br />
Andy Bytheway 91<br />
Supaporn Chai-Arayalert and Keiichi Nakata 103<br />
Cheng-Yi Chiang, Ya-Ying Kuo, Ying-Hui Hou<br />
and I-Chiu Chang<br />
Wen-Chou Ch, Chia-Hsien Wen, Sek-Kwong<br />
Poon and Shih-Che Huang 3<br />
Taurai Chikotie, Jonathan Oni and Vesper<br />
Owei<br />
vi<br />
18<br />
24<br />
82<br />
112<br />
120<br />
127<br />
Mitchell Cochran 134<br />
i
Paper Title Author(s) Page<br />
No.<br />
Motivations for IT Outsourcing in Public<br />
Sector Local Government<br />
Documenting Innovation: A Methodological<br />
Proposal and Application<br />
Legal, Privacy, Security, Access and<br />
Regulatory Issues in Cloud Computing<br />
Interoperability Monitoring for eGovernment<br />
Service Delivery Based on Enterprise<br />
Architecture<br />
Enterprise Resource Planning<br />
Implementation Differences Within the Same<br />
Methodology - Case Study From West<br />
Europe and Turkey<br />
A Meta-Analysis of how the Adoption of<br />
Innovative Web 2.0 Tools Like Digital<br />
Business Ecosystems can Lead to Improved<br />
SME Collaboration<br />
Ideas About Profitability in Research and<br />
Development and the Selective Pressure<br />
From Management Accounting<br />
Case Study on Information Evaluation by GIS<br />
for Aging Society Urban Planning: GIS<br />
Application on Urban Planning-<br />
Distortion Free Algorithm to Handle<br />
Secondary Watermark Attack in Relational<br />
Databases<br />
Engineering Change Through the Domains of<br />
Enterprise Architecture<br />
Evaluating Interoperability Artifacts for the<br />
Exchange of Public Service Information:<br />
Outline of a Conceptual Framework<br />
Free and Open Source Software for Public<br />
Sector Enterprise Applications in Sri Lanka<br />
Evaluating the Success of Customer<br />
Relationship Management (CRM) Systems<br />
Toward a Novel Methodology for IT Strategic<br />
Planning<br />
Applying an Innovative Information System to<br />
Improve the Effects of Community Health<br />
Promotion<br />
Information Security Practices in Zambian<br />
Copper Mines: An Investigation Into the<br />
State-of-Practice of Information Security<br />
Within Zambian Copper Mines Based on the<br />
ISO/IEC 27002 Standard<br />
Proposing an Optimized Change<br />
Management Process by Analyzing ITSM<br />
Frameworks<br />
Michael Cox, Martyn Roberts and John Walton 141<br />
Stefano De Falco 150<br />
Nomusa Dlodlo 161<br />
Badr Elmir, Nabil Alrajeh and Bouchaib<br />
Bounabat<br />
169<br />
Turan Erman Erkan 181<br />
Francesc Estanyol 187<br />
Albrecht Fritzsche 196<br />
Hiroatsu Fukuda, Yupeng Wang and Kiyoshi<br />
Shinriki<br />
Sajid Iqbal, Azhar Rauf, Huma Javed and<br />
Shabir Ahmad<br />
204<br />
214<br />
Tiko Iyamu 222<br />
Veit Jahns 231<br />
Srimal Jayawardena and Gihan Dias 239<br />
Farnoosh Khodakarami and Yolande Chan 253<br />
Mouhsine Lakhdissi and Bouchaib Bounabat 263<br />
Chih-Yu Lin, Wen-Yu Ho, Chia-Hsien Wen and<br />
Hui-Mei Hsu<br />
274<br />
Chishala Lukwesa and Christopher Upfold 281<br />
Negar Madani, Ali Suzangar, Mohamad Kajbaf,<br />
Shirin Nasher and Mehrdad Kalantarian<br />
ii<br />
291
Paper Title Author(s) Page<br />
No.<br />
The use of RFID and Web 2.0 Technologies<br />
to Improve Inventory Management in South<br />
African Enterprises<br />
Modeling the Genetic Schemes of Human<br />
Memory Information in the Process of<br />
Production and Reparation of Knowledge<br />
Copyright Protection for GIS Vector Map<br />
Based on Wavelet Transform<br />
Evaluation of IT Investment Methods and<br />
Proposing a Decision Making Model<br />
Critical Factors in the use of Mobile Learning<br />
by “Digital Natives” on Portuguese Teaching<br />
A Framework for Information Systems<br />
Innovation: A Case of Competitive<br />
Intelligence in Organisations<br />
RFID: A Review of its Relevance and<br />
Application in South African Retailing<br />
Systems<br />
Towards a Comprehensive Evaluation<br />
Framework for ICT for Development<br />
Evaluation – an Analysis of Evaluation<br />
Frameworks<br />
User Adoption of the New Health Information<br />
System - Individual and Organizational<br />
Determinants<br />
Relationship Between Organizational Factors<br />
and RFID Adoption in Iranian Gas Company<br />
Central Warehouses<br />
Business Intelligence as Decision Support in<br />
Business Processes: An Empirical<br />
Investigation<br />
Assessing Future Value of Investments in<br />
Security-Related IT Governance Control<br />
Objectives – Surveying IT Professionals<br />
Evaluation Information Extraction for Health<br />
Text Categories Using C4.5 and Naïve Bayes<br />
Optimizing Information Technology Value<br />
Governance Framework Based on Val IT<br />
Developing an Outpatient Electronic Medical<br />
Record System in Taiwan From a Physicians’<br />
Perspective<br />
MTN Foundation's Digital Library Project in<br />
Nigerian Universities: An Evaluative Study<br />
Cloud Computing-Based IT Solutions for<br />
Organizations With Multiregional Branch<br />
Offices<br />
Case Study on Information Evaluation by GIS<br />
for Aging Society Urban Planning:<br />
Information Evaluation of Shrinking Cities<br />
Sizakele Mathaba, Nomusa Dlodlo, Quentin<br />
Williams and Mathew Adigun<br />
Fatemeh Mohammadi, Tahereh Mohammadi<br />
and Bahram Rezaie<br />
300<br />
308<br />
Amer Sedeeq Mustafa 317<br />
Shirin Nasher, Mehrdad Kalantarian, Ahmad<br />
Akbari, Ali Suzangar, Mohammad Kajbaf and<br />
Negar Madani<br />
324<br />
Mário Carrilho Negas and Paulo Ramos 333<br />
Phathutshedzo Nemutanzhela and Tiko Iyamu 341<br />
Jonathan Oni, Edward Dakora and Vesper<br />
Owei<br />
351<br />
Caroline Pade-Khene and Dave Sewry 360<br />
Bahlol Rahimi 370<br />
Farnaz Rahimi and Gholamabas Arabshahi 378<br />
Ari Riabacke, Aron Larsson and Mats<br />
Danielson<br />
Waldo Rocha Flores, Teodor Sommestad,<br />
Hannes Holm and Mathias Ekstedt<br />
Klaokanlaya Silachan and Panjai<br />
Tantasanawong<br />
Ali Suzangar, Mehrdad Kalantarian , Shirin<br />
Nasher , Mohammad Kajbaf and Negar Madani<br />
Hsiao-Ting Tseng, Pei-Ting Chang and Ray<br />
Chang<br />
384<br />
393<br />
403<br />
412<br />
420<br />
Ngozi Blessing Ukachi 428<br />
Harris Wang 435<br />
Yupeng Wang, Hiroatsu Fukuda and Kiyoshi<br />
Shinriki<br />
iii<br />
441
Paper Title Author(s) Page<br />
No.<br />
The Impact of Software Test Constraints on<br />
Software Test Effectiveness<br />
Strategic Considerations for the Effective<br />
Mapping of Educational Technology to<br />
<strong>Academic</strong> Outcomes<br />
Grafton Whyte and Donovan Lindsay Mulder 450<br />
Hossein Zadeh, Arthur Shelley and Rod<br />
McCrohan<br />
PhD Research Papers 471<br />
Evaluating the Performance of ERP Systems<br />
in Saudi Arabian Higher Education: A<br />
Stakeholders’ Perspective<br />
The Benefits of ICT Adoption: An Empirical<br />
Study of Nigerian SMEs<br />
Establishing the Suitability of Dynamic Web<br />
Applications Development Platforms for<br />
Learning web Techniques in Tertiary<br />
Institutions<br />
Human Resources Transformation Beyond<br />
Boundaries in Outsourcing Business Model<br />
s- Expatriate Benchmarking<br />
The Diffusion of Mobile Phones for Business<br />
and Information Management in Kenya<br />
Querying DTI Analysis Results Within Deep<br />
Perisylvian Area<br />
Towards an Evaluation Method for<br />
Information Quality Management of Health<br />
Information Systems<br />
Developing a Change Management Model for<br />
Iranian <strong>Academic</strong> Libraries: A Delphi Study<br />
Mona Althonayan and Anastasia<br />
Papazafeiropoulou<br />
461<br />
473<br />
Idisemi Apulu and Ann Latham 483<br />
Johnson Dehinbo 492<br />
Swathi Duppada and Rama Chandra Aryasri 502<br />
Wakari Gikenye 511<br />
Sarmad Istephan and Mohammad Siadat 521<br />
Siti Asma Mohammed and Maryati Mohd Yusof 529<br />
Maryam Nakhoda and Sirous Alidousti 539<br />
Adding Action to the Information Audit Huan Vo-Tran 546<br />
Business Intelligence Best Practices for<br />
Success<br />
Joseph Woodside 556<br />
Work in Progress 563<br />
The Evolution of IT Governance in Multiple<br />
Industry Units and the Business Case for an<br />
Outside Partner<br />
The Relationship Between Quality<br />
Management and Knowledge Management in<br />
the Service Industries<br />
Construction and Validation of eSchool<br />
Success Model<br />
Ganeshprasad Chandrasekaran,<br />
Chandramohan Annavarapu and Balaji<br />
Rathinasamy<br />
565<br />
Amir Honarpour and Ahmad Jusoh 568<br />
Hesbon Nyagowa, Dennis Ocholla and<br />
Stephen Mutula<br />
iv<br />
572
Preface<br />
Following on from the continued success of the <strong>European</strong> <strong>Conference</strong> on Information Management and<br />
Evaluation, we are delighted at the Ted Rogers School of Management, Ryerson University to be able to<br />
host the 2nd International <strong>Conference</strong> on Information Management and Evaluation (ICIME 2011).<br />
ICIME aims to bring together individuals researching and working in the broad field of information<br />
management, including information technology evaluation. We hope that this year’s conference will provide<br />
you with plenty of opportunities to share your expertise with colleagues from around the world.<br />
This year’s opening keynote address will be delivered by Dr Catherine Middleton, Ted Rogers School of<br />
Information Technology Management, Ryerson University, Toronto, Canada.<br />
A second keynote address on day two will be delivered by Gary Davenport, Vice-President of Information<br />
Technology, AllStream Inc, Canada.<br />
ICIME 2011 received an initial submission of 180 abstracts. Of this 65 papers have been accepted for these<br />
<strong>Conference</strong> Proceedings as a result of the double blind, peer review process of full paper submissions.<br />
These papers come from all parts of the world including Australia, Canada, India, Iran, Italy, Japan,<br />
Malaysia, Morocco, Nigeria, Pakistan, Portugal, South Africa, Sri Lanka, Sweden, Thailand, Taiwan, the<br />
United Kingdom, the United States of America Zambia and Zimbabwe. So ICIME is set to be a truly<br />
international conference!<br />
We wish you a most interesting and enjoyable time in Canada<br />
Ken Grant<br />
<strong>Conference</strong> Chair<br />
April 2011<br />
v
Biographies of <strong>Conference</strong> Chairs, Programme Chairs<br />
and Keynote Speakers<br />
<strong>Conference</strong> Chair<br />
Ken Grant is an associate professor in the Ted Rogers School of Management at<br />
Ryerson University, His teaching and research interests are in the areas of Business<br />
& Technology Strategy, Innovation & Knowledge Management and Electronic<br />
Commerce. He holds the Ted Rogers School of Management Faculty Teaching Chair<br />
and is a former Director of the School of Information Technology Management at<br />
Ryerson. Prior to becoming part of the Ryerson community, Ken spent over 30 years<br />
in industry, mainly as a management consultant. He has held senior partner/vice<br />
president positions in major consulting firms, including A.T. Kearney, EDS and KPMG.<br />
He holds a BA from The Open University, UK., an MBA from York University, Toronto<br />
and is just completing his DBA at Henley Business School, University of Reading.<br />
Keynote Speakers<br />
Dr Catherine Middleton holds a Canada Research Chair in Communication<br />
Technologies in the Information Society. Her research focuses on the<br />
development and use of new communication technologies, with specific interests<br />
in mobile devices and fixed and wireless broadband networks. She is also<br />
interested in how Canadians use (or don't use) the internet in their daily lives. A<br />
member of the GRAND Networks of Centres of Excellence projece, she is the<br />
Leader of the New Media Challenges and Opportunities research theme and also<br />
leads the GRAND Digital Infrastructures project. Catherine teaches Innovation and<br />
Organization Theory in the MBA/MMSc program in the Ted Rogers School of<br />
Management, and teaches and supervises students in the Communication and<br />
Culture graduate program.<br />
Gary Davenport As Vice-President of Information Technology, Gary is responsible for<br />
the definition and implementation of transformational business initiatives using<br />
information technology for the Allstream business unit of MTS Allstream Inc. In this<br />
role, Gary leads a team of dedicated IT professionals focused on improving business<br />
capabilities in a very competitive telecommunications industry. He is also the<br />
Executive Sponsor for Allstream's Workplace 2.0 Program enabling much greater<br />
levels of employee mobility, productivity and engagement. Gary has participated on<br />
many North American and <strong>European</strong> CIO Councils and also served on the Board of<br />
Directors of the Voluntary Inter-industry Communications Standards (VICS)<br />
organization. He is currently a Board<br />
Member of the CIO Association of Canada. Gary holds a Bachelor of<br />
Arts<br />
degree from York University.<br />
Biographies of contributing authors (in alphabetical order)<br />
Kwesikorsa Aggrey is a PhD student of the University of Lincoln, UK. Kwesikorsa research area is in health<br />
service delivery, mental health service delivery in particular. His research topic is on collaboration between<br />
the NHS Trusts (Lincolnshire Partnership Foundation<br />
Trusts) and the Voluntary Sector in Performance of<br />
Mental<br />
Health Care Delivery in Lincolnshire.<br />
Jaflah Hassan AlAmmary is assistant professor in the Information System Department of the college of the<br />
IT at University of Bahrain. She holds PhD from University of Murdoch. Al-Ammary’s research interest<br />
focuses on Strategic alignment, Knowledge Management, and E-learning and Educational<br />
Technology. She<br />
has<br />
published many papers on these topics on conferences, journals and books.<br />
Mona Althonayan is a research student at the school of information<br />
systems, computing and mathematics,<br />
Brunel<br />
University, UK, working on ERP systems evaluation.<br />
Al-Mallah Amer is an Assistant Professor at the College of Science, he obtained a Ph.D. degree in<br />
Computer Science. Then he worked at the University of Technology. Subsequently, served as M.Sc.<br />
students supervisor at the postgraduate studies department. Dr. Amer<br />
has published eleven journal articles.<br />
His<br />
publications reflect his research interests in information security.<br />
vi
Idisemi Apulu is currently a PhD student at the School of Technology, University of Wolverhampton, United<br />
Kingdom where she also obtained her Masters degree in Strategic Information Technology Management.<br />
She also holds a BSc degree in Computer Science from the University of Abuja, Nigeria.<br />
Suzan Arslanturk is a masters student at the Department of Computer Science, Oakland University. She<br />
holds a B.S in Computer Engineering. Her research interests are image processing, pattern recognition and<br />
data mining. Suzan‘s master research focused on data mining and machine learning.<br />
Rahimi Bahlol, PhD: studied informatics at Linköping University. He is now assistant Professor at Urmia<br />
University of Medical Sciences. His research interests are; Individual, organizational and social issues and IT<br />
solution, collaborative work through ICT, IT solutions in health care setting, implementation and use and<br />
evaluation of integrated HIS, Adoption of IT solution.<br />
Tridip Kumar Bardhan is working as Chairperson and Director of Manufacturing Process Laboratory at the<br />
Department of Industrial and Systems Engineering (ISE) of Morgan State University in Baltimore, Maryland,<br />
USA. He received his BSc degree in Industrial Arts from Dhaka University and BSIE, MSIE and PhD in<br />
Industrial Engineering from Wichita State University, USA.<br />
Imran Batada is a senior IT Professional with over 10 years of experience in the IT industry. As an IT<br />
Professional he has worked in USA and Asia. He is currently working in Institute of Business Administration,<br />
Karachi Pakistan as Head of Information System Department. He is also responsible for overall ERP<br />
Implementation.<br />
Kofi Agyenim Boateng completed his PhD in Information Systems at the London School of Economics and<br />
Political Science, UK, in December 2009 and, currently, a lecturer at the British Institute of Technology & Ecommerce.<br />
Kofi’s research interests fundamentally lie in the innovative application of ICTs and their<br />
psychological consequences in organisational strategising.<br />
Sheryl Buckley is Deputy Head of Department in the Department of Business Information Technology (BIT)<br />
at the University of Johannesburg (UJ). Her passion lies in the Information Science discipline. She is a<br />
committee member of a number of international organisations as well as an active peer reviewer. She has<br />
presented and published papers locally and internationally.<br />
Nuntarat Bunditwongrat received her Bachelor’s Degree (2005) in Information Technology for Business<br />
from Chulalongkorn University, Thailand. She is now a Master’s degree student of Management Information<br />
Systems program at the Faculty of Commerce and Accountancy, Thammasat University. She also serves as<br />
a Senior Consultant at the Department of Global Business Services, IBM Thailand Company Limited.<br />
Supaporn Chai-Arayalert is a Ph.D. student in Informatics at Informatics Research Centre, Henley<br />
Business School, University of Reading, UK. Her research areas focus on knowledge management,<br />
semiotic approach and Green ICT. She graduated Master of Science in Management of Information<br />
Technology and she recently works as lecturer at Prince of Songkhla University, Thailand<br />
Yolande Chan is a Professor, MIS at Queen’s School of Business and Director, The Monieson Centre. She<br />
holds a Ph.D. from the Richard Ivey School of Business, an M.Phil. in Management Studies from Oxford, and<br />
S.M. and S.B. degrees in Electrical Engineering and Computer Science from M.I.T. She is a Rhodes Scholar.<br />
Dr. Chan conducts research on knowledge management and information technology strategy.<br />
Ganeshprasad Chandrasekaran B.E., M. Tech., M.B.A., is an associate of India’s largest Software<br />
Company TCS Limited, where he is focused with aligning the IT goals with Business priorities. He has been<br />
awarded with Chief Minister Award and University Gold medal for the best academic excellence. Ganesh<br />
has been a research scholar in SRM University, India.<br />
Taurai Chikotie is a student and researcher at the Cape Peninsula University of Technology in the fields of<br />
ICT and Healthcare for Development. He holds a BSc Honors degree in Information Science from the<br />
National University of Science and Technology in Zimbabwe and a Masters Degree in Information<br />
Technology cum laude from the Cape Peninsula University of Technology.<br />
Wen-Chou Chi is a Doctoral Student in the MIS program at the National Chung Cheng University in Taiwan.<br />
His current research interests include hospital information systems, knowledge management and decision<br />
support system.<br />
vii
Mitchell Cochran has been the Information Systems Manager for the City of Monrovia for 14 years. He has<br />
also worked in the Court System and for IBM. He is currently working on his Information Systems PhD from<br />
Claremont Graduate University and has completed Masters Degrees in Administration and Homeland<br />
Security. He has a CISM certification.<br />
Renata Paola Dameri is senior lecturer in Business Administration at the Faculty of Economics, University of<br />
Genova, Italy. She is professor in Accounting and Information Systems in Genova, visiting professor in IT<br />
Governance at Université de Paris Dauphine and Fellow of the SDA Bocconi School of Management, Milano,<br />
Italy. Her research interest covers IT investments evaluation, IT governance, IT security and compliance.<br />
Mats Danielson is a full Professor in Computer and Systems Sciences and Vice Dean of the Social Science<br />
Faculty at Stockholm University. He holds a Ph.D in Computer and Systems Sciences as well as degrees in<br />
Business Administration and Computer Engineering. He has been working with decision and risk analysis<br />
professionally and academically for almost 20 years.<br />
Stefano De Falco Holds a Degree and a Ph.D degree in electrical engineering, University of Naples<br />
Federico II. His research interests concern the modelling and testing of the measures for quality of firms. His<br />
publications are present in international congresses and journals. He is the inventor and the advisor of a<br />
national magazine “TT-Techology Transfer”.<br />
Johnson Dehinbo obtained B.Sc. degree in Computer Science & Statistics from Ogun State University in<br />
1989, and B.Sc. Honors in Information Systems from UNISA in 2000. He obtained M.Sc. degree in<br />
Information Systems specializing in software engineering from UNISA in 2006. He is currently busy his<br />
M.Phil/Ph.D studies in Informatics at the University of Pretoria.<br />
Nomusa Dlodlo holds a PhD in Computer Science. She is currently employed as a Senior Researcher at<br />
the Council for Scientific and Industrial Research (CSIR)'s Meraka Institute in Pretoria, South Africa. Dr.<br />
Dlodlo works in the Internet-of-Things (IoT) group. Her main research interests lie in applying IoT<br />
technologies to improve teaching and learning at primary and secondary school levels<br />
Swathi Duppada received her Master’s degree in Human Resource Management from Andhra University,<br />
Visakhapatnam, India, and is currently working with Satyam Computer Services Limited, Canada. She is<br />
currently a Ph.D. candidate at Jawaharlal Nehru Technological University, Hyderabad, India. Her research<br />
interests include Benchmarking of Human Resource Management, Modernization of HR using ERP and<br />
Software Testing.<br />
Badr Elmir is a Software Engineer graduated from ENSIAS (2002), holder of an Extended Higher Studies<br />
Diploma from ENSIAS (2006) and is a “Ph.D. candidate” at ENSIAS since 2009. His research focuses on<br />
interoperability monitoring within public administration. He is an integration architect on the Ministry of<br />
Economy and Finance of Morocco since 2002.<br />
Turan Erman Erkan is an Assistant Professor at the Department of Industrial Engineering in Atılım<br />
University. After a career in management consultancy, he worked as an international SAP consultant and<br />
then started a research career. His research and consultancy interests include ERP, BPR, SCM, CRM and<br />
performance measurement.<br />
Francesc Estanyol is research assistant at the University of Edinburgh Business School. He received his<br />
degree in Computer Engineering from the Universitat Pompeu Fabra, Spain, and also holds masters degrees<br />
in Bioinformatics for Health Sciences and in Management. Currently he is the Scientific Manager of<br />
EcoBusiness Marie Curie project. His research interests include how the adoption of innovative ICT tools can<br />
benefit SMEs and the translation of research into real market applications.<br />
Albrecht Fritzsche has a Master's degree in mathematics, and educational science, D.Phil. Vehicle<br />
scheduling expert in the IT Management department of Daimer, A.G., Stuttgart. Doctoral dissertation in<br />
industrial economics on heuristic search in complex environments, doctoral dissertation in philosophy on<br />
indeterminacy in technical systems. Interdisciplinary research on innovation at TU Darmstadt.<br />
Hiroatsu Fukuda is a professor at the University of Kitakyushu of Japan. A doctor of engineering, registered<br />
architect of Japan, member of Architectural Institute of Japan (AIJ). Research interests are environment<br />
design of architecture, recyclable buildings, and urban planning for compact cities, urban environmental<br />
evaluation and suggestion.<br />
viii
Waldo Rocha Flores received his B.Sc. in Business Administration and Economics from the Stockholm<br />
University in 2007 and his M.Sc. in Electrical Engineering from the Royal Institute of Technology (KTH) in<br />
Stockholm in 2008. He is now pursuing a Ph.D. in Industrial Information and Control Systems.<br />
Wakari Gikenye, B.A. M.A. University of Nairobi, PGD University of Wales, is a Senior Librarian at the<br />
University of Nairobi Library, Nairobi, Kenya, and a PhD Student at University of Zululand, South Africa. Has<br />
recently presented papers on the diffusion of Information and Communication Technologies in the informal<br />
sector in Kenya.<br />
Amir Honarpour is currently a PhD. Student at Faculty of Management and Human Resource Development<br />
(FPPMS) University Technology Malaysia. Amir graduated with a M.Sc. in system management in 2009.<br />
After graduation, he worked in a project to help Design an integrated research system among Iran's<br />
universities. His research interests include: Knowledge Management, Web-based Courseware Tools, Quality<br />
Management and Research information Systems.<br />
Sajid Iqbal received his Masters degree in computer science from University of Peshawar Pakistan.<br />
Currently he is a research student at Department of computer science, university of Peshawar Pakistan. His<br />
area of interest includes information security, relational database watermarking and data mining.<br />
Sarmad Istephan is a Computer Science PhD student at Oakland University. His research focuses on the<br />
storage and querying of Medical Images (e.g., MRI). In addition to being a PhD student, Istephan was a<br />
Java/C# Software Engineer and currently is a Senior SQL Server Database Engineer at Quicken Loans.<br />
Istephan eagerly awaits publishing more scientific papers in his field.<br />
Tiko Iyamu is a Professor of Information Systems at the Tshwane University of Technology, Pretoria. His<br />
research interests include Mobile Computing, Enterprise Architecture and Information Technology Strategy.<br />
Theoretically, he focuses on Actor Network Theory (ANT) and Structuration Theory (ST). Iyamu is author of<br />
numerous peer-reviewed journal and conference proceedings articles.<br />
Veit Jahns has a German Diploma in Computer Science and works as a software developer and consultant<br />
at the otris software AG in Dortmund, Germany. Additionally, he is finishing his Ph.D. thesis at the University<br />
of Duisburg-Essen. His research interests are the all aspects of information systems interoperability, in<br />
particular between information systems in public authorities.<br />
Srimal Jayawardena obtained his BSc Engineering from the University of Peradeniya and BIT from the<br />
University of Colombo School of Computing, both with first class honours. He has served in the Central Bank<br />
of Sri Lanka as an Assistant Director/IT d and at the Information and Communication Technology Agency of<br />
Sri Lanka as a Technology Specialist. He is currently a PhD candidate at the Australian National University.<br />
Mehrdad Kalantarian is ITSM specialist of Infoamn CO., a consulting firm that provides services and<br />
solutions for security, compliance & IT Management. He works with a professional team studying on IT fields<br />
such as ITIL, ISMS, COBIT and Val IT. His experimental field is COBIT. He has M.S. degree in ICT<br />
engineering and lives in Tehran.<br />
Farnoosh Khodakarami has an MSc in Management from Queen’s School of Business, Queen's University,<br />
Canada. Currently, she is working as a researcher at The Monieson Centre, Queen’s University. Her<br />
research interests include customer relationship management, e-commerce, information systems and<br />
knowledge management.<br />
Ya-Ying Kuo is a master of Healthcare Information Management at National Chung Cheng University,<br />
Chiayi, Taiwan. Her current research interests include hospital information systems and patient privacy.<br />
Michael Kyobe is an associate Professor in the department of Information Systems, University of Cape<br />
Town. He holds a PhD and an MBA. Prior to joining academia, Michael worked for over 15 years in the<br />
public and private sectors in the IS and Computer forensics fields. His research include information security,<br />
business alignment and KM<br />
Mouhsine Lakhdissi is a Partner in a consulting firm specialized in IT Architecture. He received a Master<br />
degree in Software Engineering in a leading engineering school in Morocco. He also served as Chief IT<br />
Architecture in many companies with international exposure. His research interests include Enterprise<br />
Architecture, IT Governance and processes and Software industrialization.<br />
ix
Aron Larsson, Ph.D. in Computer and Systems Science and MBA. Senior lecturer and researcher at<br />
Stockholm University and Mid Sweden University. Research interests include methods, procedures and<br />
applications of computer based decision support, as well as risk and decision analysis. Research projects<br />
include process models and methods for public decision making, landmine clearance activities, procurement<br />
processes, and distributed artificial intelligence in wireless networks.<br />
Chih-Yu Lin is a Doctoral Student in the MIS program at the National Chung Cheng University in Taiwan.<br />
His current research interests include hospital information systems, knowledge management and decision<br />
support system<br />
Sizakele Untonette Mathaba has competed her degree in Bsc Information Systems (2007) honours degree<br />
in Computer Science (2008) at the University of Zululand. In 2009, she joined CSIR (Council for Scientific<br />
and Industrial Research) as an internship student. Currently, she is registered for her Masters degree in<br />
Computer Science. She is working with the Internet of Things Engineering Group at CSIR (Meraka).<br />
Fatemeh Mohammadi has a Ph.D in Educational management. Faculty member of Islamic Azad University-<br />
Shiraz Branch, Iran.He is an inventor with 5 registered inventories. The author of 6 books and 24 essays<br />
articles and papers with 12 presentations in National <strong>Conference</strong>s. Instructor of 38 educational terms for<br />
faculty members and Theory builder of Human Information Life.<br />
Siti Asma Mohammed is currently a PhD student at National University of Malaysia. She finished Masters of<br />
IT specializing in Information Systems at University of Sydney, Australia. She worked as a Test Engineer for<br />
one year and Assistant Lecturer in Information Systems for four years before pursuing PhD. Her research<br />
interests are IS Evaluation and information quality.<br />
Maryam Nakhoda is a PhD candidate in Library and Information Science (LIS) at University of Tehran,<br />
faculty of Psychology and Education. She is the author of papers in Persian and English. Her research<br />
interests include Information Technology (IT) application in academic libraries, library management, and<br />
managing change in academic libraries.<br />
Shirin Nasher is ITSM specialist of Infoamn CO., a consulting firm that provides services and solutions for<br />
security, compliance & IT Management. She works with a professional team studying on IT fields such as<br />
ITIL, ISMS and COBIT. Her experimental field is Val IT. She has M.S. degree in IT management and lives in<br />
Tehran.<br />
Mário Carrilho Negas is an assistant professor of management at the Open University (Portugal). He<br />
received his Ph.D. in Management. His main research interests include the adoption of systems and<br />
information technology in SMEs, strategic planning of information systems and management of Innovation.<br />
Phathutshedzo Nemutanzhela is a Masters student at the Tshwane University of Technology. She has a<br />
Baccalaureus Technologies (BTech): Information Technology (Informatics). Her principle research interest is<br />
Competitive Intelligence and Information Systems.<br />
Hesbon Nyagowa is a Ph.D student at the University of Zululand, South Africa specializing in Information<br />
Studies. He obtained Master of Business and Administration specializing in Management Information<br />
Systems at University of Nairobi Kenya in 2002. His Bachelor’s degree was in Education (Science) obtained<br />
at Kenyatta University, Kenya in 1988. Is currently the <strong>Academic</strong> Registrar, Kenya Polytechnic University<br />
College.<br />
Jonathan Oni is a Post Graduate student and researcher at the Cape Peninsula University of Technology,<br />
Cape Town, South Africa, with an interest in e-business. He holds a BSc Honors in Computer Science.<br />
Jonathan consults for various Information Technology companies and involved in managing IT projects. He<br />
is also a part-time lecturer at a higher educational institution.<br />
Farnaz Rahimi is studying IT management in Alzahra University(MS degree) and work in contract<br />
department in Mashhad Gas Company. Farnaz is interested in the field of "knowledge management" and<br />
"implementing new IT technologies in organization"<br />
Azhar Rauf received his doctorate degree in computer science from Colorado Technical University,<br />
Colorado Springs CO, USA in 2007. Currently he is teaching as Assistant Professor at the Department of<br />
Computer Science, University of Peshawar, Peshawar, Khyber Pakhtunkhwa, Pakistan. His areas of interest<br />
include Relational Database Watermarking, Fine-Grained Security techniques in Relational Databases,<br />
Encryption, Anonymity and Information Security.<br />
x
Ari Riabacke, Ph.D. in Risk- and Decision Analysis (Computer Science), Head of Business Intelligence at<br />
the largest Swedish Management and IT Consultant Company, also has a M.Sc. in Organizational Decision<br />
Making, and is a member of the DECIDE Research Group at Stockholm University.<br />
Martyn Roberts spent the first few years of his career in industry working in information systems, but<br />
transferred to academia over 20 years ago. He is now Principal Lecturer at the University of Portsmouth. He<br />
has taught various aspects of IS on a wide of programmes at both undergraduate and post graduate levels.<br />
He has published mainly in the areas of strategic information systems and eCommerce.<br />
Klaokanlaya Silachan works at the Computer technology department, Facultly of Science Nakorn pathom<br />
Rajabhat University, Thailand. She is currently pursuing in computer information technology, Silpakorn<br />
University,Thailand major in data mining, Health medical data, ontology. She received her master’s degree in<br />
Information Management Technology from Mahidol University, Thailand, in 1998. She has publication in<br />
national conference and international conference proceedings.<br />
Ali Suzangar is a Chief System Officer of Infoamn CO., a consulting firm that provides services and<br />
solutions for security, compliance & IT Management. He works with a professional team studying on IT fields<br />
such as ITIL, ISMS and COBIT. His experimental field is Val IT. He has M.S. degree in IT management and<br />
lives in Tehran.<br />
Panjai Tantassanawong is currently pursuing his doctoral degree in Computer Science majoring in<br />
networking and software engineering at AIT . He received his master’s degree in Computer science from<br />
Chulalongkorn University, Thailand in 1992. He is Assistant Professional the Computing Program, Facultly<br />
of Science, Silpakorn University, Thailand. He has publications in national and international conference<br />
proceedings<br />
Hsiao-Ting Tseng is a graduate student of Healthcare Information Management at National Chung Cheng<br />
University, Chiayi, Taiwan. Her current research interests include patient privacy, electronic medical records<br />
and exchange of medical images.<br />
Ngozi Ukachi is a librarian at University of Lagos, and presently doing her PhD at University of Nigeria,<br />
Nsukka. A member of International Federation of Library Associations (IFLA), American Library Association<br />
(ALA) and, Nigerian Library Association (NLA). She was IFLA 2010 Essay Competition Award Winner<br />
(organized by IFLA <strong>Academic</strong> and Research Section).<br />
Chris Upfold is currently a lecturer in the Department of Information Systems at Rhodes University, South<br />
Africa. He also teaches in the Rhodes Business School. His areas of interest and research are Information<br />
Security, Radio Frequency Identification (RFID), Project Management, Virtual Teams and Corporate<br />
Communications.<br />
Huan Vo-Tran is a lecturer and the program director of the Bachelor of Business (Information and<br />
Knowledge Management) within the School of Business IT & Logistics at RMIT University. He is currently<br />
completing a PhD in business computing. Prior to becoming an academic he worked in various fields, which<br />
included project management, systems analysis and high school teaching. His areas of interest include<br />
information management and Web 2.0.<br />
Harris Wang is an associate professor in the School of Computing and Information Systems at Athabasca<br />
University, Canada. He received a PhD in computer science from the Australian National University,<br />
Australia. His research interests include advanced technology for education, information systems and<br />
information security.<br />
Joseph Woodside is a Doctoral candidate in Information Systems at Cleveland State University, with<br />
publications and research interests in topics of business intelligence, informatics, healthcare systems<br />
integration, geo-spatial-temporal modeling, HIT adoption, machine learning, and object-oriented database<br />
technology. Joseph is employed with KePRO, a national care management company, as the Director of<br />
Healthcare Informatics and Business Intelligence.<br />
Hossein Zadeh has taught undergraduate and postgraduate courses in Australia, Hong Kong, Vietnam,<br />
Singapore, and Sweden. Hossein is the recipient of 2008 University Team Teaching Award and 2009<br />
University Certificate of Achievement in innovative teaching. In 2004, Hossein was a visiting scholar at<br />
Linkoping University, Sweden, and in 2009/2010, was a Distinguished Visiting Scholar at IBM Almaden<br />
Research Labs, Silicon Valley (San Jose), USA. Hossein is the recipient of the prestigious 2010 IBM Faculty<br />
Award.<br />
xi
Performance Management of the NHS’ Mental Health Care<br />
Service Delivery in England: The Role of the Service’ Actors’<br />
(Clients) Collaboration Through Dialogism<br />
Kwesi Korsa Aggrey<br />
University of Lincoln, UK<br />
kaggrey@lincoln.ac.uk<br />
Abstract: A welfare state has a responsibility to provide health and social services to the citizenry. The state<br />
delegates that responsibility to a number of actors, each of which has its own performance management criteria. To<br />
ensure coherence, it is important to managing their performances collectively taking into consideration all the actors<br />
inputs rather than only that of government or its agent, the NHS This turns out to be difficult task especially that of the<br />
mental health care services. The author will be looking at a possible solution to this problem based on the<br />
assumption that it is possible to change the way the actors present their stories by bringing them together to share<br />
their purposes and common action through the notion of dialogism as way of maintaining plurality of logics: different<br />
voices (polyphonic), styles (stylistic), space-time conception (chronotopic), interanimating discourse (architectonics)<br />
and the dynamic interplay of different dialogisms (polypi). It is this process of plurality of logic that the researcher<br />
termed, the Third Cybernetics Evolution, as a way of sequential processes from the First Cybernetics through the<br />
Second Cybernetics to the Third Cybernetics. It is argued that implementing these, allows for improved<br />
communication among actors, as a way of achieving high quality service. It is suggested that implementation of these<br />
concepts and processes implies the use of storytelling as a facilitatory system. It is shown that all actors in the mental<br />
healthcare services delivery make use of this system, albeit often inefficiently. Therefore may lead to dissipation of<br />
the system in the future. To prevent such dissipation, the existing structure needs to be improved through spiral<br />
relationships via communication (dialogism)<br />
Keywords: third cybernetics evolution, mental health care system, NHS, collaboration and dialogism<br />
1. Introduction<br />
In qualitative research, one of the issues is how to explore the meaning that actors bring and generate<br />
within social interactions. However, in most researches on mental health care services, the concentration<br />
has been on the inputs of the strong actors namely the government, the NHS and the voluntary<br />
organisations with less attention to the weaker actors like service users, the family, or relatives, carers<br />
and advocates. This paper will focus not only on moving the ‘gear’ to higher level of service users’<br />
controlled and ‘emancipatory’ approach to research but further to the level of dialogism where all<br />
stakeholders are involved in a collaborative effort.<br />
The state through its agent, the NHS, is responsible for ensuring that the health needs of the citizenry are<br />
adequately met. The NHS sometimes sub-contracts some of the services to other organisations (actors,<br />
in this case, they are also clients) like the voluntary sector organisations, private organisations etc. The<br />
government provides the resources in terms of equipment, funds, facilities and personnel needed to<br />
deliver the services to the patients. The government in the process of resourcing the NHS sets targets<br />
through the Department of Health for health and social service providers. But for the health care system<br />
to be effective in meeting the local needs and expectations, it has to include the inputs of the clients and<br />
patients. According to some clients (patients) interviewed in Lincolnshire in England, the system is not<br />
performing because some of their colleagues had to go back to the hospital after their medical treatment<br />
because the ‘system’ has no adequate provision for social integration and social needs of the patients. It<br />
is argued that, where such provisions exist, the government decide what is needed and the criteria for<br />
measuring and managing the performance without considering the local needs of the citizens which may<br />
be different from one locality to another. The clients (patients) claimed that where the system seems to<br />
give attention to their needs, the government thinks the system is not performing because it is not<br />
meeting the targets set and therefore closes down such hospitals or departments in the hospitals even<br />
when the citizens protest.<br />
It is being argued that mainstream mental health care researches tend to be clinically dominated,<br />
reflecting the linear and hierarchical approaches of medical professionals who based their analysis on<br />
quantitative and experimental testing with less attention to qualitative research. One of the focuses of the<br />
paper therefore is to give attention to qualitative research in mental health services by analysing how to<br />
give audience to and manage service users, carers, the family and the community’s inputs through the<br />
actors-dialogism-system.<br />
1
2. Mental health care services in UK<br />
Kwesi Korsa Aggrey<br />
Mental health problems account for significant proportions of diseases in UK. However, access to<br />
treatment though available is not being fully utilised because of a number of issues in which<br />
reorganisation of the service is urgently needed. The issues of mental health care have become<br />
important because of its significant burden on the government and the community. It is argued that, it<br />
accounts for nearly 12% of the global burden of disease and will hit the target of 15% of disability<br />
adjusted life years lost to illness by 2020 (WHO, 2003).<br />
The economic and social costs to the state and the community have increased with the development of<br />
science and medicine. Before the 17 th century, issues of mental disorders were attributed to spiritual<br />
explanation and therefore its treatment. However, with the development in science and medicine, secular<br />
explanations were given to the mental disorder as a physical state of the mind leading to the confinement<br />
of such people in secured places.<br />
During the first part of the 18 th century however, the dominant view of mental disorders being incurable<br />
and the justification of being sub-humans and therefore their status, living conditions and physical<br />
restraints to places of confinement changed.<br />
The pressure from society and humanitarian groups about the human right of people with mental<br />
disorders led to a call for reforms in the treatment of mental patients and therefore the introduction of<br />
moral treatment programmes. This led to deinstitutionalisation or de-hospitalisation of the service delivery<br />
aimed at creating the opportunity to implement some of the ideas developed on social network and<br />
community support system as alternatives outside psychiatric institutions.<br />
However, the way mental health care service delivery is organised determines the effectiveness of the<br />
interventions and the ultimate fulfilment of the objective of the service. The quality of the service<br />
determines the likelihood of one achieving the desired outcome of becoming employable, independent,<br />
self organised and part of the community. The quality of the service delivery is important because it<br />
ensures that service users receive the care and support they require; it also ensures that there is<br />
improvement in their daily life activities and living; provide acceptable and relevant clinical and non<br />
clinical care to reduce the impact of mental health problems; and to make efficient and effective use of<br />
resources available to them (Murray and Lopez, 1996).<br />
In UK, the government’s policy on community care services was strongly developed with the third sector<br />
in mind. This is evidenced in the Community Care Act (Department of Health, 1990) which embraces the<br />
full involvement of other actors in mental health care services in particularly, the Third Sector. Despite<br />
this policy however, the collaboration between the statutory and the actors particularly the voluntary<br />
sector organisations to improve the quality of the service is weak (Simpson, 1996) and problematic,<br />
(Adams, 1990). The weakness of this collaboration has been attributed to lack of understanding of each<br />
other’s objectives, missions or vision and organisational practices, methods and significantly, differences<br />
in respect of their structures, resources (Wilson, 1994) and their value systems (Mosher and Burti, 1994).<br />
There is also an issue of confidence and trust by the professionals of the formal sector about the<br />
capabilities of semi skilled personnel of the Third Sector and the private sector. Gratham (1995) surveyed<br />
55 general practitioners in his area and found that only 18 per cent had made some referral to the<br />
Alzheimer’s disease Society. Some professional even see any sharing of skills, knowledge and<br />
experiences with the non-professional as a threat to their profession (Gussow and Tracy, 1976). The<br />
paper aims at proposing a method of resolving the problem of collaboration among the actors through<br />
participatory communication (dialogism). We will now look at the methodology used in this research to<br />
identify the problem of limited communication among actors of the system.<br />
3. Methodology and design<br />
The study is intended to expand on our understanding of how to improve performance or improve the<br />
quality service through increase of interaction among the stakeholders in health service delivery,<br />
particularly in mental health care delivery system. A phenomenological method was used in this research<br />
because it is one of the methods which help to identify phenomena through the perspectives of the actors<br />
in the situation. In human interactive situations like the mental health care service delivery, it enables the<br />
researcher to gather in depth information and perceptions through qualitative methods of interviewing and<br />
focal group discussions (Gerber, 2001).<br />
2
Kwesi Korsa Aggrey<br />
Phenomenology could be looked at as a philosophy, but also as a research method for gathering and<br />
analysing data (Goulding, 1999). As a method, phenomenology allows for the discovering of people’s<br />
‘world views’ thus capturing their subjective experiences. In all, eighteen (18) managers from the<br />
Lincolnshire Partnership Foundation Trust (LPFT), Lincoln County Council (LCC) Lincoln County Hospital<br />
(LCH) and some voluntary sector organisations in Lincolnshire were interviewed. In addition two focal<br />
groups’ discussions were also conducted.<br />
Using the phenomenological methods of interviews and focal groups discussions, the experiences of the<br />
participants were organised by means of storytelling and interwove with living story method. The living<br />
story theory, which is the bedrock of living story method is defined as the emergence, trajectory, and<br />
morphing of living story from ante-narrative-conception to the end of decomposition of a story (Boje,<br />
2007b; Boje, 2008). A living story is a story which lies in-between dead and alive narrative and between<br />
forgotten and revitalised story. It is therefore a critical ante-narrative, where a living story traces and predeconstructs<br />
an ongoing interweaving living story narrative and ante-narrative to become selfdeconstruct.<br />
In the process of deconstructing stories, some stories die and others rejuvenate or emerge<br />
and better still some revolve. The living story method is therefore interplay of dead (ended) narratives<br />
with living emergent stories. It tells the emergent stories in the context of fragmented dead narratives.<br />
The data collected were transcribed and analysed by reading and analysing the data in depth to search<br />
for differences and patterns in relation to how to improve service delivery. Each interview was used to<br />
probe into issues from the previous interviews which were not clear, in order to create a ‘rich picture’ of<br />
the situation (lack of adequate communication among the actors). The data were interpreted,<br />
continuously revised and the context broadened using the notion of ‘Ideal interpreter’.<br />
Interpretivism is assumed to be mental content judgement-dependent; that is, the facts about<br />
propositional attitude of people is exactly what is being captured by the judgement of the Ideal Interpreter<br />
(Johnston, 1993a). However, for Dennett and Davidson (sited in Wright, 1989), the Ideal Interpreter is a<br />
third-person who interpreters someone else. The interpreter can also be a first- personal account where<br />
the interpreter is taken to be the subject of interpretation. Dennett’s and Davidson’s version of interpreter<br />
is however, formulated on the thesis of biconditional: that X belief Q, therefore if there is an informed<br />
Ideal Interpreter, the Ideal Interpreter would be disposed to attribute to X the belief that is Q. There is<br />
however, a problem with this line of argument of interpretivism. There is no a priori guarantee that the<br />
Ideal Interpreter will find answers to all the meaning of the subjects’ beliefs particularly those that are<br />
linguistically expressed. It is therefore necessary that some sufficient conditions are held as a constraint<br />
to strengthen this identified weakness of interpretivism, that is, there is an appropriate informed Ideal<br />
Interpreter who would be disposed to attribute X the belief that Q. Interpretivism presupposes therefore<br />
that facts or believes given by the Ideal Interpreter will be the facts participants believe in an ideal<br />
situation, that is the assumption.<br />
For the purposes of this research, the Ideal Interpreter is assumed to have a sufficient database of nonintentional<br />
and intentional sources of information, knowledge and experience to interpret the subjects’<br />
meanings, believes, experiences and desires. The Ideal Interpreter uses daily life application of intentions<br />
in addition to scientific minded observations and experiences of non-intentional in order to arrive at the<br />
best possible meanings of the participants’ experiences and believes in order to detach the meanings of<br />
the participants’ experiences at the individual level. The next section looks at how interaction among the<br />
actors through participatory communication will help to achieve the ultimate level of collaboration that will<br />
encourage improvement in the quality of the service delivery system.<br />
4. Actors’ collaboration: Way forward to improve the quality of service<br />
It is proposed that to improve the quality of service, collaboration among all interested parties of the<br />
mental health care delivery system is necessary in order to meet the needs and expectations of clients.<br />
One can say therefore that, the collaboration among the stakeholders is not happening because of<br />
communication problems among the actors of the service delivery system. This was identified and<br />
confirmed from the data collected and organised as per the emerged sub themes (namely: state control<br />
of the mental healthcare system; organisational culture; service delivery as continuous process; service<br />
users needs and expectations; and participatory communication) from interviews conducted among<br />
managers and directors from the Lincolnshire Partnership Foundation Trust (LPFT), Lincoln County<br />
Hospital (LCH), Lincoln County Council (LCC),some voluntary sector organisations and some service<br />
users from the county, (these are some of the actors of the mental health care system).<br />
There are a number of possibilities to resolve the problem identified:<br />
3
Kwesi Korsa Aggrey<br />
One possibility is to focus on information overload in the organisation of the NHS. It is proposed, by<br />
increasing the level of interaction among all the interested parties via the use of participatory<br />
communication after a suitable destabilisation of the existing communication system, the quality of<br />
service will be improved.<br />
Another possibility is the proposal to use the method of ‘information overload’, i.e. an intentional increase<br />
in information through the communication channels of the NHS, as a way to support a suitable<br />
destabilisation of the interaction among all the interested parties to stimulate self-organised improvement.<br />
The third possibility, the one that is the paper’s focus is, the possible improvement in the organisation of<br />
the mental health care system within the NHS. It is proposed therefore that the use of ‘information<br />
overload’, a process to achieve participatory communication is necessary in order to increase the level of<br />
interaction among all the interested parties and also to stimulate self-organisation of the system using the<br />
dialogism approach. It is when the level of interaction in the decision making process is increased and<br />
more attention given to the beneficiaries of the system through the supporting groups’ representatives<br />
that the system may attain optimum performance and meet clients’ needs, expectations and in fact satisfy<br />
the needs of all stakeholders, that is achieve ‘complete collective satisfaction’.<br />
One theme that emerged from the interviews conducted was the organisational culture of the NHS, which<br />
mainly is about professionalism of work, staff and the way task is undertaken or service is rendered. The<br />
NHS: PCT and LPFT staffs have the idea that work should be based on specialisation of tasks leading to<br />
the departmentalisation and training of staff accordingly.<br />
To deal with the culture of professionalism and also to improve the services delivery, there is the need for<br />
collaboration, because after medication and discharge from hospital, service users may need the social<br />
support mainly provided by other actors including the Third Sector. It is a fact that the majority of mental<br />
health patients preferred to work with and receive help from the non-professionals than from the mental<br />
health professionals (Barker et al, 1990). Research has proven that, for mental health problems in<br />
particular, the benefit of social support cannot be swept under the carpet especially at the time of<br />
discharge from hospitals and that social networks has been very influential in the improvement of quality<br />
of life of service users. (Green et al, 2002).<br />
It is therefore important that collaboration is encouraged among all stakeholders to achieve complete<br />
collective satisfaction. This is necessary because no single organisation can meet the needs and<br />
expectations of mental health service users on its own, therefore a multi-agency model (actors-dialogismsystem)<br />
should be explored via collaboration; training of self help groups, sharing information, collective<br />
service evaluation and feedback to improve services. It is also important to incorporate other methods<br />
used by the Third Sector staff to strengthen the collaboration between the third sector and the formal<br />
sector (Simpson, 1996).<br />
It is important to note that in order to evaluate or manage the performance of the service delivery in<br />
mental health care and to improve the quality, one needs to go beyond the medical model, diagnosis of<br />
symptoms, treatment and side effects of disease. This may involve input/output system performance<br />
evaluation. This system of input/output evaluation is what the researcher calls the First Cybernetics (Boje,<br />
2008). However, for total treatment of mental problems, the inclusion of social well being of patients<br />
(service users) is necessary. This in a collective term is referred to as Health-Related Quality Of Life<br />
(HRQOL). The HRQOL generally includes the domains of physical functioning, psychological, well being<br />
(e.g. level of anxiety, depression, fear of recurrence, etc); and social functioning of the patients. The<br />
process of evaluating the performance of the mental service delivery system goes beyond the medical<br />
model, input/output approach, to include the system’s environment (open system approach).<br />
Cybernetics at the level of input/output approach is about negative feedback within self-stabilising loops<br />
that lead to a regulation of a system, in order to generate a state of equilibrium (Ashby 1957). When<br />
Cybernetics is related to external and centralised control of black boxes and goal-oriented behaviours, in<br />
order to sustain complex interactions with the environment over time (Beer, 1966 p.257), then, the<br />
system turns to become ‘open system’ a Second Cybernetics system or approach (Boje, 2008).<br />
Cybernetics might also manifest itself in connected networks of components in interaction. As a result of<br />
such interconnectedness, multiple feedback loops may be formed within the networks, which allow for<br />
self-regulation and the production of ‘global’ order. This implies that a system can also have distributed<br />
4
Kwesi Korsa Aggrey<br />
control, which does not depend on an external centralised control unit (Ashby, 1957), this is termed the<br />
First Cybernetics.<br />
The concept of negative feedback is crucial for achieving stability band adaptation. Negative feedback is<br />
a necessary corrective action, which when implemented produces a specific outcome by modifying a<br />
course of action. The consequence of this arrangement is that, the input of the loop is affected by its own<br />
output. Hence, a satisfactory outcome can be achieved following a negative feedback process, indicating<br />
whether the system has achieved its goal or fallen short. It does this by providing backward information<br />
on the resulting outputs, in order to be used for the manipulation of the inputs of the system (Ashby 1957<br />
p.53-54). This is what is expected from a participatory communication process via dialogism of actors’<br />
interaction if fully implemented.<br />
The First Cybernetic approach as described above is not very suitable for a complex system like the<br />
mental health care system. The reason being that as adaptation and stability is achieved; the system is<br />
bound to dissipate (dysfunction) with time because of the continuous changes and constant increase of<br />
the variety in its environment, which may exceed the capacity of the input/output cybernetics system. This<br />
is in line with the reverse interpretation of the Ashby’s Law of requisite variety. Therefore one needed to<br />
go beyond this level of abstraction to the Second Cybernetics system which is the open system thinking.<br />
It is an open system because it receives its feedback not only from within the system but also from the<br />
system’s environment. Again being an open system, it means the system does not only exchange energy<br />
with its environment but also matter from the environment. However, in both the First and Second<br />
Cybernetics approaches of improving performance or quality of service, there are elements of linearity<br />
and hierarchical, which in themselves create problems<br />
The mental health care system, as a human institution with different people with different needs and<br />
preferences, it is expected that the capacity of its meta-system transition (co-ordinated process) will not<br />
be able to cope with the variety of needs and preferences from both inside and outside the system and<br />
therefore dissipation is bound still to happen. This is what has resulted in various forms of reforms all<br />
aimed at resolving the problems associated with increase variety of needs and preferences by the<br />
various actors. One cannot avoid the dissipation because all the actors have their own different goaloriented<br />
values which are not static but grow together with a diversifying environment, which may be<br />
different from government’s target. The only way to resolve the problem in order to achieve the desired<br />
high quality of service delivery (complete collective satisfaction) is to move to a higher level of cooperation<br />
among the actors where the relationship between them or the functional imperative is neither a<br />
linear nor hierarchical but spiral relationship of polypi dialogism. This is an inter-co-ordinated system<br />
function of various dialogisms, namely the polyphonic, stylistic, chronotonic and architectonic.<br />
5. Dialogism<br />
A polyphonic is a written, visualised or orally told stories by all the stakeholders of a system as opposed<br />
to mono vocal narrative or written strategy by an expert or a dominant actor (Boje, 2008). It is a<br />
construction by many embodied voices, logics and perceptions. It could be visual arts, photos, décors,<br />
drama or oral storytelling that communicates through interaction. It involves getting stakeholders to<br />
engage in storytelling through the sharing of ideas, experiences, knowledge, skills such that each and<br />
every one’s voice and logic get enunciated and subjected to questioning until a form of understanding is<br />
reached on the particular issue and how to deal with it.<br />
However, the interaction or communication in the process of the sharing is not necessarily a function of<br />
producing a consensus but usually results in open situation where the issue is being accepted or<br />
rejected. The actors in the process therefore reach a point, which bifurcates into further alternatives. The<br />
process of bifurcation itself is a process of reducing complexity of the issues that come to the ‘floor’ of the<br />
dialogue into simple terms and meanings.<br />
The Stylistic is a dialogue; it is an orchestration of image or dialogism among various means of dialogue<br />
or communication namely oral telling of stories, print and video media, internet, gesture-theatrics, décor<br />
and other architecture modes of image expressions. It is a juxtaposition of varied styles for interaction. It<br />
can also be described as the interactivity of various modes of expressions of organisational image.<br />
For Weick (1995), sense making could be a way of looking at stylistic on the basis of public sense making<br />
control. Sense making is about the act of or the process of placement of items into frameworks,<br />
comprehending, redressing and constructing meanings from the frame such that interacting with it<br />
5
Kwesi Korsa Aggrey<br />
generate mutual understanding and patterns. This form of framing and reframing (producing and<br />
reproducing) is a stylistic feature. It also determines the style of delivery of the services and where<br />
necessary the revision of the style to meet certain moral obligations. The power of stylistics is in the way<br />
that stylistics is able to connect or depend on the working interplay of the roles of all stakeholders in the<br />
service delivery to achieve a common purpose.<br />
Chronotonic is a relativity of time and space, in terms of functions; it is like Einstein’s theory of relativity,<br />
with time being the fourth dimension of space. Chronotopic dialogism can also be described as a<br />
holographic relation of centring (centripetal of chronotopes) and amplifying (centrifugal of chronotopes).<br />
Holographic is storytelling that runs from one dimension (monogon) to multiple complexities. Storytelling<br />
is holographic in the sense that it can interrelate to more than one complexity. Storytelling is infectious<br />
and that can be tracked across space-time.<br />
Architectonic is the orchestration of ethics in relation to aesthetic and cognitive storytelling. Architectonic<br />
dialogism is mainly focus on the interaction of several societal discourses that affects organisation’s<br />
performance. The three basic discourses are ethics, aesthetic and cognitive which are answerable to one<br />
another. Kant (1993) and Bakhtin (1981) described architectonic as societal discourses. However, each<br />
author has his own divergent views on this notion. Kant invented the ‘cognitive architectonic’. For him<br />
architectonic is the art of constructing a system. He argues that reason cannot permit our knowledge to<br />
remain in an unconnected and rhapsodistic state, but requires that the sum of one’s cognitions constitute<br />
a system. Kant sees architectonic as cognitive notion deeply implicated in the construction of systemicity.<br />
Bakhtin preferred the term ‘consummation’ to construction and was careful not to assume a monophonic<br />
or mono logic or mono language system. He looked at a system as ‘systemicity’, that is unmergedness<br />
and unfinalisability of a system. This is the way forward for the proposed mental health care system,<br />
where there is no room for homeostasis state or dissipation because the process of ‘polypi’ is<br />
‘unmergedness’ ‘and unfinalisation’. Bakhtin went further from Kant’s cognitive architectonic discourse to<br />
add ethical and aesthetic discourses. He defined Ethics discourse as the notion of ‘answerability’ which is<br />
a description of how one domain of discourse is answerable to another. The Aesthetic is about how and<br />
for whom a given systemicity is constructed or designed.<br />
Polypi is the inter-dialogism of polyphonic, stylistic, chronotopic and architectonic. It is a multi-dialogised<br />
complexity where the four (4) notions collide. The polypi is therefore a coordination system of the<br />
dialogism<br />
6. Conclusion<br />
The complexity of dialogism, the polypi, which explores multiple of dialogisms, is in line with Letiches’s<br />
(2000) phenomenal complexity and also Stacey’s (2006) emergent complexity. The properties of the<br />
complexity are cumulative rather than successive. In cumulative, all the lower orders of phenomenal<br />
complexity (polyphonoic dialogism) intermingle with higher orders of dialogism (of fragmented antenarrative<br />
and petrified narratives that follow Aristotelian structures of wholeness and coherence of<br />
beginning middle and end in storytelling.<br />
Weick (1995) looks at the complexity in retrospective sense making of experiences and narrative of<br />
coherence and control by people whose current experiences fit into past meaning structures. This I think<br />
offers a revolutionary breakthrough in complexity thinking or approach .This approach is not only<br />
compatible with phenomenal complexity theory, but also provides a way to overcome a major crack or<br />
gap in knowledge creation. The same argument applies to synchronous and diachronous approaches to<br />
storytelling organisation. The synchronous approach (being the current experienced story) looks at<br />
storytelling organisation at a particular point in time, rather than over time while the diachronous<br />
approach (being narrative) looks at storytelling organisations in historical development. Therefore in order<br />
to achieve the wholeness, one has to bring both approaches together through the phenomena, the<br />
‘polypi’ which is the dialogism of history or background of current issue(s) plus current experiences of<br />
users of the mental health care delivery system.<br />
A typical and recent familiar example of narrative and emergent stories of dialogism is the death of<br />
Michael Jackson, the pop star. A network of organisations constructed narratives along side stories of<br />
Michael Jackson’s death; even now stories about his death continue to emerge. In the narrative and<br />
emergent stories of his death, the polypi of dialogism is contentious: polyphonic logics struggle to<br />
converge or agree; the multiple stylistics of verbal, written and posters or pictures contrast; the multiple<br />
6
Kwesi Korsa Aggrey<br />
chronotopes of varying temporalities and partialities diverge, and the ethical discourse of architectonic<br />
questions reverberates into many other discourses. It indicates therefore that the complexity of the<br />
narratives and emergent stories of Michael Jackson’s death is unmerged and ‘unfinalised’ if one want to<br />
get to the ‘bottom’of the story.<br />
Dialogism is a complex notion, it is a weave of many storytellers and listeners who together co-construct<br />
the meanings of a dynamic that reduce living story into complete collective meaning as a necessary<br />
outcome as opposed to ante-narrative amplification (Boje, 2008).<br />
The contribution of this paper to knowledge is the proposal that although linearity or hierarchical<br />
approaches to knowledge creation is making impact in society, the impact will be great and more<br />
beneficially however, if the interaction among the actors is on a spiral relationship, that is on a collective<br />
level by way of collaboration to resolve the problems identified collectively in order to achieve ‘complete<br />
collective satisfaction’ particularly in the service delivery system.<br />
In the case of mental health care delivery system ‘Complete collective satisfaction’ is achievable only if<br />
the mental health care system moves from the level of abstract categorisation such as planning or<br />
reforms to the level of collaboration. This means that the way forward is therefore, the re-organisation of<br />
the system which is exclusive to the NHS or government (target setting) to a level of multifunctional and<br />
interdependent interaction where consideration is given the environment, other service providers, clients,<br />
financiers, the public and in fact all stakeholders.<br />
References<br />
Adams, R. (1990), Self-help, Social work and Employment. Macmillan, London.<br />
Ashby, W. R. (1957) An Introduction to Cybernetics, Chapman & Hall, London<br />
Ashby, W. R. (1958) “Requisite variety and its implications for the control of complex systems” Cybernetics, Vol. 1<br />
No.2 pp83-99<br />
Ashby, W.R (1980) An Introduction to Cybernetics. London: Chapman and Hall L<br />
Bakhtin, M. M. (1981) Art and Answerability (Michael Holquist and Vadim Liapunor, eds; translation and notes by<br />
Vadim Liapimov; supplement translated by Kenneth Brostrom) Austin TX: University of Texas Press<br />
Barker, et al (1990) “Coping and help seeking in the UK adult population” British Journal of Clinical Psychology, Vol.<br />
29, pp.271-85<br />
Beer, S. (1966) Decision and Control: The Meanings of Operational Research and Management Cybernetics,<br />
London: Wiley.<br />
Beer, S. (1979). The Heart of Enterprise Chichester: Wiley<br />
Beer, S, (1985) Diagnosing the System for Organisations. John Wiley, Chichester<br />
Boje, D. M. (2007b) From Wilda to Disney: Living Stories in Family and Organisation Research. In Jean Clanddinin<br />
(ed.) Handbook of Narrative Inquiry, London: Sage<br />
Boje, M. D. (2008) Storytelling Organisations Sage Publications Limited London<br />
Department of Health (1990), National Health Service and Community Care Act, HMSO, London<br />
Donaldson, M. S. (2004) “Taking Stock of Health-Related Quality-of- Life Measurement in Oncology Practice in the<br />
United States” JNCI Monographs<br />
Vol. 2004 No. 33 pp. 155-167 Oxford University Press London<br />
Foucault, M. (1977b) Language, Counter-Memory, Practice: Selected Essays and Interviews by Michel Foucault, D.<br />
F. Bouchard (ed.). Trans. and intro Ithaca, NY: Cornell University Press.<br />
Gerber, R. (2001), “The concept of commonsense in workplace learning and experience”<br />
Goulding, C. (1999) “Consumer research, interpretive paradigms and methodological ambiguities”, <strong>European</strong> Journal<br />
of Marketing, Vol. 33 No. 9/10, pp. 859-73<br />
Graham, n. (1995) “GPs and voluntary organisations” British Journal of General Practice, Vol. 45, p. 273<br />
Green, G. et al (2002) “The role and impact of social relationships upon well being reported by mental health serviceusers:<br />
a qualitative study”, Journal of Mental Health, Vol. 11, pp.565-79.<br />
Gussow, Z.and Tracy, G.S. (1976), “The role of self-help clubs in adaptation to chronic illness and disability”, Social<br />
Science and Medicine, Vol. 10, pp. 407-14<br />
Johnston, M. (1993a) “Objectivity Refigure: Pragmatism without Verification” in J. Haldone and C. Wright, eds.<br />
Reality, Representation and Projection Oxford University Press.<br />
Kendall, J. and Knapp, M. (2000)” Measuring the Performance of Voluntary Organisations” Public Management 2(1):<br />
105-132<br />
Letiche, H. (2000) Phenomenal complexity theory as informed by Bergson. Journal of Organisational Changes<br />
Management, 13 (6): 545-58.<br />
Mosher, L.R. and Burti, L. (1994) Community Mental Health: Principles and Practice, New York, NY<br />
Murray, A. and Shepherd, G. (1996) “Suspicious minds”, Health Services Journal, Vol. 106, p. 27.<br />
Simpson, R. G. (1996) “Relationship between self-help organisations and professional health-care providers”, Health<br />
and Social Care in the Community, vol. 4, pp. 359-70<br />
7
Kwesi Korsa Aggrey<br />
Stacey, R. (2006) Complex responsive processes as a theory of organisational improvisation, pp128-41 London:<br />
Routledge<br />
Weick, K. E. (1995), Sense making in Organisations, Sage, Thousands Oaks, CA<br />
Wilson, J. (1994) “Aspirations and realities: community care at the crossroads”, Health and Social Care in the<br />
Community Vol.3, PP. 227-40<br />
World Health Organisation (WHO), Mental Health in the WHO <strong>European</strong> Region, Fact sheet EURO/03/03,<br />
Copenhagen, Vienna, 2003.<br />
Wright, C. (1989) “Wittgestein’s RuleFollowing Considerations and the Central Project of Theoretical Linguistics” in A.<br />
George, ed. Reflections on Chomsky, Basil Blackwell<br />
8
The Acceptance and use of Educational Technology in<br />
Kingdom of Bahrain<br />
Jaflah Al-Ammary<br />
University of Bahrain, Bahrain<br />
Jaflah@itc.uob.bh<br />
Abstract: In 2004, King Hamad Schools of Future Project (KHSFP) has been launched by the ministry of education<br />
at Kingdom of Bahrain with a goal of establishing a fundamental change in the way teaching and learning take place.<br />
The main objective of the project is to change the way in which students and teachers communicate, work and learn<br />
by transfer the traditional classroom into an open, interactive environment based on adopting new and more advance<br />
technologies. In fact at Kingdom of Bahrain, adopting technologies such as computers, smart board, or interactive<br />
board will increase the pressure on the schools to accommodate to these new conditions. This is because there are<br />
some schools in particular the elementary schools were only getting access to the technology and many teachers<br />
lack the required IT skills and capabilities. As a result, many efforts have been initiated to enhance the necessary<br />
skills and facilitate the functional development of the teachers. However, there is still a need to understand the<br />
teachers and student’s acceptance of the new technology to cope with KHSFP. Thus, to achieve a technological<br />
enhancement of the education system, students and teachers should accept the IT as an effective tool. The current<br />
research aims at investigating the acceptance and use of the new educational technology such as computer,<br />
interactive board and smart board, by teachers and students in the schools of future at Kingdom of Bahrain. Through<br />
an extension of the Technology Acceptance Model (TAM), five factors that influence the behavioral intention to<br />
accept and use education technology were examined. These factors include: subjective norms, motivation, computer<br />
self efficacy, previous experiences, and computer anxiety. Moreover, previous experiences was investigated to has<br />
an indirect effect on perceive ease of use through the computer anxiety. The results show that the computer self<br />
efficacy, previous experience, and motivation are the most critical factor for accepting and use educational<br />
technology at Kingdom of Bahrain. By considering these factors and investigating other factors, Ministry of Education<br />
at Kingdom of Bahrain can enhance the acceptance and use of the new educational technology by both teachers and<br />
students among the Kingdom schools.<br />
Keywords: acceptance technology model, computer anxiety, educational technology and previous experience<br />
1. Introduction<br />
In order to get benefit from the opportunities offered by the digital revolution and the big technological<br />
innovations in the field of education; the ministry of education at Kingdom of Bahrain decided to draw a<br />
strategy to employ ICT in education (Shaker, 2008). This strategy aims at developing a new environment<br />
for e-learning, not just providing computer laboratories with equipments and teaching computer as<br />
subject matter (ICE47, 2001). Having such action, future generations will be prepared to establish<br />
information society and be empowered with the basic skills necessary to transfer the kingdom into a<br />
knowledge based economy (Shaker, 2008, ICE47, 2001). The ministry of education is making its efforts<br />
to achieve its objective through the implementation of big projects that deal mainly with enhancing and<br />
improving the quality of education at Kingdom of Bahrain (ICE47, 2001). One of these projects is KHSFP.<br />
In 2004, King Hamad Schools of Future Project (KHSFP) has been launched by the ministry of education<br />
at Kingdom of Bahrain with a goal of establishing a fundamental change in the way teaching and learning<br />
take place. The main objective of the project is to change the way in which students and teachers<br />
communicate, work and learn by transfer the traditional classroom into an open, interactive environment<br />
based on adopting new and more advance technologies. (Al-Ammary et al. 2010a,b). Moreover, the<br />
project will reform the education system in Kingdom of Bahrain to prepare the citizen to invest in the ICT’s<br />
potentials to achieve the education quality and attain the curricula capabilities at all educational levels<br />
(ICE47, 2001). KHSFP includes an integrated systems comprise of an educational portal to achieve the<br />
qualitative turning point in the educational performance under a contemporary educational environment,<br />
supported by a modern educational and information technology (ICE47, 2001). This project has been<br />
implemented by first, establishing a website for the project, second, disseminating information about the<br />
project in the target schools and third, establishing a scientific documentary library for the project<br />
(ESCWA, 2005). The project will be executed by connecting the targeted schools (primary, elementary,<br />
and secondary level), with a speedy communication network via the central educational portal (ESCWA,<br />
2005).<br />
In fact at Kingdom of Bahrain, adopting technologies such as computers, smart board, or interactive<br />
board will increase the pressure on the schools to accommodate to these new conditions. This is<br />
because there are some schools in particular the elementary schools were only getting access to the<br />
9
Jaflah Al-Ammary<br />
technology and many teachers lack the required IT skills and capabilities. As a result, many efforts have<br />
been initiated to enhance the necessary skills and facilitate the functional development of the teachers.<br />
However, there is still a need to understand the teachers and student’s acceptance of the new technology<br />
to cope with KHSFP. Thus, to achieve a technological enhancement of the education system, students<br />
and teachers should accept the IT as an effective tool. Therefore, the current research has conducted to<br />
investigate the acceptance and use of the new educational technology such as computer, interactive<br />
board and smart board, by teachers and students in the schools of future at Kingdom of Bahrain. Through<br />
an extension of the Technology Acceptance Model (TAM), five factors which influence the behavioral<br />
intention to accept and use education technology were examined. These factors include: subjective<br />
norms, computer self efficacy, previous experiences, motivation and computer anxiety. Moreover,<br />
previous experiences was investigated to has an indirect effect on perceive ease of use through the<br />
computer anxiety.<br />
This paper consists of nine sections; each section will touch a significant component of this research. The<br />
following section spots the light on the literature review on the Technology Acceptance Model (TAM).<br />
Section three explains the research model and hypotheses. The survey instruments of this study are<br />
discussed in section four. Section five illustrates the research methodology for this study. Description of<br />
data analysis is presented in section six. In section seven, profile of the study participants are discussed.<br />
Finally, in the last two sections of this paper; 8 and 9, the findings of this study are discussed and a<br />
conclusion is presented and made ready for the educational institutes that intend to adopt IT.<br />
2. Research background<br />
2.1 Technology Acceptance Model<br />
There are many models have been developed to investigate and understand the factors affecting the<br />
acceptance of computer technology in the organization. Among those are the Theory of Reasoned Action<br />
(TRA) (Ajzen et al. 1980), Theory of Planned Behavior (TPB) (Ajzen, 1991), Technology Acceptance<br />
Model (TAM) (Davis et al., 1989), and Unified Theory of Acceptance and Use of Technology (UTAUT)<br />
(Venkatesh et al., 2003). However, it have been noticed that the most used model by Information System<br />
academics and practitioners is the TAM designed by (Davis, 1998). TAM is an adaptation of the Theory<br />
of Reasonal Action (TRA) to the field of IS. The objective of TAM is to provide an explanation of the<br />
determinants of the adoption and use of information technology (Davis et al., 1989). Perceived<br />
usefulness is the degree to which a person believes that using a particular system enhances his or her<br />
job performance. Perceived ease of use is the degree to which a person believes that using a particular<br />
system will be free of effort (Davis et al., 1989). Perceived ease of use and perceived usefulness are<br />
common in technology-usage setting (Tylor et al., 1995 a, b) and have been applied widely to solve the<br />
acceptance problem using different acceptance models (Wu et al., 2005).<br />
TAM originally suggested that perceived usefulness and perceived ease of use are involved in explaining<br />
the variance in user’s intention or behavior of intention (Davis et al., 1989). Perceived usefulness is also<br />
seen as being directly impacted by perceived ease of use. However, behavior intention to use functioning<br />
has hypothesized by TAM as a mediator of actual technology in use (Davis, 1998).<br />
3. Research model and hypotheses<br />
This research study has a main objective of investigating factors that influence the acceptance of<br />
educational technology in the schools at Kingdom of Bahrain. The literature revealed for many factors<br />
that may affect the acceptance and use of educational technology in any school. For the purpose of the<br />
current research, five factors have been selected which include: subjective norms, previous experience,<br />
computer anxiety, computer self efficacy, and motivation. The research model is developed as depicted<br />
in Figure (1).<br />
3.1 Theory factors: Perceived usefulness and perceive ease of use<br />
Information system researchers have investigated TAM and agreed on its validity in predicting the<br />
individual's acceptance of various systems (Chin and Todd, 1995; Doll et al, 1998; Segars and Grover,<br />
1993; Venkatesh and Davis, 2000). Most of them revealed that perceived usefulness and perceived ease<br />
of use have a positive effect on the behavioral intention to use system (Chin and Todd, 1995; Doll et al,<br />
1998). Moreover, perceived ease of use was found to have a positive effect on the perceived usefulness<br />
(Lee, 2006). The hypotheses relating to perceived ease of use and perceived usefulness have been<br />
studied by a number of TAM researchers (Srite, 2006). Although some researchers reveled that attitude<br />
10
Jaflah Al-Ammary<br />
toward systems use is postulated to mediate the effect of perceived ease of use and perceived<br />
usefulness on behavioral intention (Kim, el al, 2009), Davis, et al, (1989) suggest that perceived<br />
usefulness may impact on behavioral intention to use technology irrespective of attitude toward this<br />
system.<br />
Computer<br />
Self efficacy<br />
Motivation<br />
Previous<br />
experiences<br />
H12<br />
Computer<br />
Anxiety<br />
H<br />
H7<br />
H9<br />
H8<br />
H6<br />
H10<br />
H1<br />
H13<br />
H14<br />
H5<br />
Subjective<br />
Norms<br />
Perceived<br />
usefulness<br />
experiences<br />
H<br />
Perceived<br />
ease of use<br />
H<br />
H<br />
H<br />
Behavior<br />
Intention<br />
Figure 1: Research model<br />
Hypothesis 1: Perceived usefulness has a positive effect on students and teacher’s behavioral intention<br />
to use educational technology in school at Kingdom of Bahrain.<br />
Hypothesis 2: Perceived ease of use has a positive effect on students and teacher’s behavioral intention<br />
to use educational technology in school at Kingdom of Bahrain.<br />
Hypothesis 3: Perceived ease of use has a positive effect on students and teacher’s perceived<br />
usefulness of educational technology in school at Kingdom of Bahrain.<br />
3.2 Hypotheses on factors affecting the perceived ease of use and perceived usefulness<br />
3.3 Subjective norms<br />
According to Fishbein and Ajzen (1975), subjective norms refer to “the person’s perception that the most<br />
people who are important to him think he should or should not perform the behavior in question”. The<br />
literature reported on much theoretical and empirical evidence regarding the importance of the role of<br />
subjective norms on technology use directly, or indirectly through perceived usefulness in the workplace<br />
(Lee, 2006; Tylor et al., 1995a). Subjective norms have been found to be more important prior to, or in<br />
the early stages of innovation implementation when users have limited direct experience from which to<br />
develop attitude (Havelka, 2003). Venkatesh and Davis (2000) found in their study that there is a direct<br />
effect of subjective norms on the behavioral intention to use information systems. Lee [24] reported that<br />
the effect of subjective norms has significantly influenced perceived usefulness. Accordingly, the<br />
following hypothesis was developed accordingly:<br />
Hypothesis 4: Subjective norms have a positive effect on students and teacher’s behavioral intention to<br />
use educational technology in school at Kingdom of Bahrain.<br />
Hypothesis 5: Subjective norms have a positive effect on students and teacher’s perceived usefulness of<br />
educational technology in school at Kingdom of Bahrain.<br />
11
3.3.1 Computer self efficacy<br />
Jaflah Al-Ammary<br />
Computer self efficacy is defined as individual’s self confidence in his or her ability to perform behavior<br />
and use the computer in the context of information technology usage (Compeau et al., 1995; Compeau et<br />
al., 1999). According to Hayashe, et al. (2003), the computer self-efficacy is not concerned with what one<br />
has done in the past, but rather with judgments of what could be done in the future. The effect of<br />
computer self-efficacy on the performance of the learning has been studied extensively in teachinglearning<br />
settings (Hill et al. 1986; Compeau et al., 1999; Lee, 2006). Vijayasarathy (2004) found that the<br />
computer self-efficacy have a positive effect on the behavioral intention to use systems as the more<br />
experience one gain online; the more important are concerns of control over personal information.<br />
Moreover, the individuals’ confidence in their computer-related knowledge and skills can influence<br />
perception on the ease or difficulty of carrying out a specific task using a new technology, and how useful<br />
that new technology will be (Vijayasarathy, 2004). Therefore, computer self-efficacy has been revealed<br />
by the literature to have a critical role in terms of its effect on perceive usefulness and perceived ease of<br />
use (Venkatesh et al., 1996; Hayashe, et al., 2003; Lee, 2006). Therefore the following hypothesis was<br />
developed:<br />
Hypothesis 6: Computer self efficacy has a positive effect on students and teacher’s perceived<br />
usefulness of educational technology in school at Kingdom of Bahrain.<br />
Hypothesis 7: Computer self efficacy has a positive effect on students and teacher’s perceived ease of<br />
use of educational technology in school at Kingdom of Bahrain.<br />
3.3.2 Motivation<br />
Motivation to perform a behavior can be divided into two main types: intrinsic and extrinsic motivation<br />
(Hennessy et al., 2005). Extrinsic motivation refers to drive of behaviors to achieve valued outcomes that<br />
are distinct from the activity itself (Deci and Ryan, 1985; Hennessy et al., 2005). Intrinsic motivation is the<br />
satisfaction gained from performing the behavior (Hennessy et al., 2005). According to Deci and Ryan<br />
(1985) perceived ease of use is a form of intrinsic motivation. Within the TAM, extrinsic motivation is<br />
capture by perceived usefulness, while intrinsic motivation related to the perceived ease of use.<br />
Motivation in the current study is measured by both intrinsic and extrinsic motivation, and then it is<br />
proposed to affect both perceived usefulness and perceived ease of use.<br />
Hypothesis 8: Motivation has a negative effect on students and teacher’s perceived usefulness of<br />
educational technology in school at Kingdom of Bahrain.<br />
Hypothesis 9: Motivation has a negative effect on students and teacher’s perceived ease of use of<br />
educational technology in school at Kingdom of Bahrain.<br />
3.3.3 Previous experience<br />
If a user has already using computer in prior job or education institutions, then it is believed that this<br />
experience will affect his/her perception toward using computer or any technology (Chang, 2008). The<br />
repetition of task improves task performance by reducing the effort required to perform the task (Alba,<br />
1987) and provides the user with a greater opportunity to consider different aspects of performing the<br />
behavior in a relatively objective manner (Kim, et al, 2009). Moreover, Dyck and Smither, (1996)<br />
revealed that as gaining more experience in using computer, user will become more computer<br />
confidence and will show a less negative attitude toward using computers (Trocchia and Janda, 2000;<br />
Dyck and Smither,1996). Thus, the previous experiences have a relationship with the intentional behavior<br />
toward using computer (Kim, et al, 2009) through the perceived usefulness and perceived ease of use.<br />
Moreover, a previous experience is proposed to be as a factor influence the computer anxiety.<br />
Hypothesis 10: Previous experience has a positive effect on students and teacher’s perceived usefulness<br />
of educational technology in school at Kingdom of Bahrain.<br />
Hypothesis 11: Previous experience has a positive effect on students and teacher’s perceived ease of<br />
use of educational technology in school at Kingdom of Bahrain.<br />
Hypothesis 12: Previous experience has a positive effect on student and teacher’s computer anxiety on<br />
using educational technology in school at Kingdom of Bahrain.<br />
12
3.3.4 Computer anxiety<br />
Jaflah Al-Ammary<br />
Computer anxiety is a specific anxiety because it is a feeling that associated with a person’s interaction<br />
with computers (Saade and Kra, 2006). Computer anxiety is defined as a state of mind of being fearful or<br />
nervous when using a computer. It includes factors such as who may have initiated the person to the<br />
computer technology (Brosnan 1998), success or failure experiences with hardware or software (Saade<br />
and Kira, 2006), and detract cognitive resources from task performance (Sam et al., 2005). Computer<br />
anxiety has been revealed by many studies to be affected by previous computer experience and affect<br />
ease of use of information technologies using the technology acceptance model (Bertrand and Bouchard,<br />
2008; Saade and Kira, 2006).<br />
Hypothesis 13: Computer anxiety has a negative effect on students and teacher’s perceived usefulness<br />
of educational technology in school at Kingdom of Bahrain.<br />
Hypothesis 14: Computer anxiety has a negative effect on students and teacher’s perceived ease of use<br />
of educational technology in school at Kingdom of Bahrain.<br />
4. Data collection and research variables<br />
The study sample is comprised of teachers and students from twenty six public schools (male and female<br />
schools) at Kingdom of Bahrain. These schools consist of primary, elementary, and secondary levels.<br />
Two hundred and four surveys were distributed to the twenty six schools. The survey instruments for this<br />
study were developed using validated items from the prior research as a means of assessing the<br />
theoretical constructs of the extended TAM model for educational technology, using the TAM scales of<br />
perceived usefulness, perceived ease of use, and behavioral intention from Davis et al. (1989) and Davis<br />
(1998). A scale for measuring subjective norms was developed using the measurement of Venkatesh et<br />
al. (2003), scales used for measuring computer anxiety and motivation were adopted from the<br />
measurement proposed by Saade and Kira (2006), and scale for measuring computer self-efficacy was<br />
developed using the measures of both Lee (2006) and Karsten et al. (1998). However, measurements of<br />
previous experiences were developed by the author for the purpose of the current research.<br />
5. Data analysis and results<br />
A Partial Least Squares (PLS) was applied to test the current model using Smart-PLS 2.0. PLS is a<br />
structured equation modeling method that analyzes how the items load on their constructs simultaneously<br />
with estimating all the paths in the model (Chiu et al., 2005). Data analysis in the current research<br />
possessed in two stages. In the first stage, the measurement model was evaluated to validate the<br />
reliability and validity of the constructs. In the second stage, the structure model was estimated using<br />
hypotheses testing to test the significance of the path coefficients.<br />
6. Demographics<br />
The survey instrument provides a response rate of 86% which can be considered as high rate bearing in<br />
mind the difficulty in getting the permission from ministry of education at Kingdom of Bahrain to be able to<br />
conduct the survey in the public schools. Demographic characteristics of the overall participants both<br />
teachers and students are presented in Table (1). The results in the table revealed that most of the<br />
participants either teachers or students are male (64.9%, 81% respectively) and are from secondary<br />
schools (57% and 56.5% respectively). Moreover, the results show that most of the teachers have B.Sc.<br />
and master degree (79.4% and 12.6%) and at least six years of experience in teaching (79.6%). Most<br />
importantly, it has been revealed from the results that most of the participated teachers have ICDL or<br />
other computer certificates (57.9%). In addition, the results show that most of the students have taken<br />
class using educational technologies (62.4%) which include the computer (50%), interactive board (30%)<br />
and smart board (20%).<br />
Table 1: Selected characteristics of the sample<br />
Teacher Percentage Students Percentage<br />
Gender Gender<br />
Male 64.9% Male 64.9%<br />
Female 35.1% Female 35.1%<br />
Level of School Level of School<br />
Primary 71.42% Elementary 71.43%<br />
Elementary 28.57% Secondary 28.57%<br />
Secondary Age<br />
13
Jaflah Al-Ammary<br />
Teacher Percentage Students Percentage<br />
Qualification 10-13 33.7%<br />
B.Sc. 79.4% 14-17 45.5%<br />
Post-diploma 9.0% >17 20.8%<br />
Master 12.6%<br />
PhD<br />
ICDL or any certificate in<br />
0.0% Take class using educational<br />
technologies<br />
Computer<br />
Yes 62.4%<br />
Yes 57.9% No 37.6%<br />
No 42.1%<br />
Take class using educational<br />
technologies<br />
Year of experience Computer 50%<br />
1-5 20.4% Smart board 20%<br />
6-10 43% Interactive board 30%<br />
>10 36.6% others 5%<br />
7. Assessing the measurement model<br />
The strength of the measurement model is determined by its reliability and validity. Cronbach’ alpha was<br />
used to assess the reliability value of each dimension as demonstrated in Table (2). All the reliability<br />
values are higher than 0.7, except that of subjective norms (0.636). However, these values are accepted<br />
because they are closed to 0.7. Convergent validity was assessed by the examination of composite<br />
reliability and Average Variance Extracted (AVE) (Chin et al., 1995). The data indicates that the<br />
measures are robust in term of their internal consistency reliability. The composite reliabilities of the<br />
different measures ranged from 0.744 to 0.886 which exceed the recommended threshold value of 0.7 for<br />
each construct. Moreover, results in Table (3) show that AVE values were all above the recommended<br />
range (0.50) (Chin et al., 1995), with values ranging from 0.524 to 0.668, thereby establishing convergent<br />
validity for each construct. Moreover, to assess the convergent validity confirmatory, factor analysis with<br />
Varimax rotation was conducted to assess the underlying structure for the items of each research<br />
construct. The loading of each factor should be greater than or equal to 0.5 which has been achieved<br />
(results not shown in the current paper).<br />
Table 2: Reliability, and convergent validity<br />
Items Cronbachs Alpha Composite Reliability AVE<br />
Behavioral Intention 0.819 0.882 0.653<br />
Subjective Norms 0.636 0.744 0.524<br />
Perceived Usefulness 0.775 0.859 0.671<br />
Perceived Ease of use 0.761 0.857 0.668<br />
Computer Anxiety 0.779 0.853 0.596<br />
Previous Experience 0.774 0.842 0.579<br />
Computer self efficacy 0.762 0.832 0.553<br />
Motivation 0.838 0.882 0.653<br />
8. Structural model<br />
The causal relationships in the proposal research model were tested. Consistent with Chin et al., (1995)<br />
bootstrapping was applied to produce standard error and t-statistics. This permits us to measure the<br />
statistical significance of the path coefficients. The statistical objective of PLS is to show high R and<br />
significant t-values, thus rejecting the null hypothesis of no effect. The t-values need to be significant to<br />
support the hypothesized paths. R indicates the explanatory power of the latent endogenous variables.<br />
Explanation of variance for each equation in the hypothesized model is presented in Table (3), while<br />
properties of the causal paths, including standardized path coefficients and t-values are presented in<br />
Table (4).<br />
14
Table 3: Explanation of variance<br />
Jaflah Al-Ammary<br />
Factor R<br />
BI 0.511<br />
PEOU 0.447<br />
PU 0.485<br />
CA 0.210<br />
As expected by the TAM literature, perceived usefulness, perceived ease of use and subjective norms<br />
have a great impact on the behavioral intention to accept and use educational technology. Three of them<br />
together accounted for a high variance of behavior intention to use educational technology (R=0.511).<br />
Moreover, perceived ease of use affected perceived usefulness. Hence H1, H2, H3, and H4 which are<br />
based on TAM (except hypothesis related to subjective norms) were all supported (0.200 (2.071), 0.455<br />
(2.90), 0.18 (1.47) and 0.246 (2.903), respectively). On the other hand, perceived usefulness was directly<br />
influenced by the computer self efficacy, motivation, and computer anxiety. Thus, H6, H8, and H13 are<br />
supported. However, subjective norms and previous experience has no effect on the perceive usefulness.<br />
Thus H5 and H10 are not supported. Furthermore, the results show that perceived ease of use was<br />
directly influenced by computer self efficacy and previous experience, while motivation has no effect of<br />
this factor. Therefore, H7, H11, and H13 were supported while H9 was rejected. Finally, the results show<br />
that there is a negative impact for previous experience on computer anxiety.<br />
Table 4: Model testing results<br />
Hypothesis Path coefficient T-value<br />
H1 Perceived usefulness - Behavior Intention 0.200 2.071<br />
H2 Perceived ease of use - Perceived usefulness 0.455 5.05<br />
H3 Perceived usefulness - Behavior Intention 0.184 1.472<br />
H4 Subjective Norms - Behavior Intention 0.246 2.903<br />
H5 Subjective Norms - Perceived usefulness 0.064 0.623<br />
H6 Computer self efficacy Perceived usefulness 0.203 1.499<br />
H7 Computer self efficacy Perceived ease of use 0.239 1.570<br />
H8 Motivation Perceived usefulness 0.389 2.95<br />
H9 Motivation Perceived ease of use 0.069 0.525<br />
H10 Previous experience Perceived usefulness 0.091 0.681<br />
H11 Previous experience Perceived ease of use 0.222 1.881<br />
H12 Previous experience Computer anxiety -0.459 5.691<br />
H13 Computer anxiety Perceived usefulness -0.203 1.499<br />
H14 Computer anxiety Perceived ease of use -0.239 1.570<br />
9. Discussion and conclusion<br />
The main conclusion of the current research is that factors which were hypothesized to have an impact<br />
on the acceptance and use of educational technology in schools are fallen within two categories:<br />
behavioral and Information Technology factors. As such behavioral factors include: motivation, subjective<br />
norms and computer anxiety while Information Technology factors include computer self efficacy, and<br />
previous experience. In the first place, the results revealed that teachers and students behavioral<br />
intention to accept and use educational technology is influenced indirectly by the proposed behavioral<br />
and information technology factors via perceived usefulness and perceived ease of use. Moreover, the<br />
results show that perceived usefulness and perceived ease of use have a great positive effect on both<br />
teachers and student’s behavioral intention to accept and use the educational technology. In addition, it<br />
has been revealed by the results that computer self efficacy and computer anxiety are the only factors<br />
that have a positive indirect effect on the behavioral intention to accept educational technology through<br />
15
Jaflah Al-Ammary<br />
both perceived usefulness and perceived ease of use. Conversely, motivation has a positive indirect<br />
effect on the behavioral intention through perceived usefulness only, while previous experience has<br />
indirect effect on behavioral intention through perceived ease of use. Subjective norms, on the other<br />
hand, have been identified by the research to be an important predictor for the behavioral intention to use<br />
educational technology as it shown a great direct effect on the behavioral intention. However, most of the<br />
literature and especially those that have been conducted in the context of Kingdom of Bahrain (Al-<br />
Ammary 2008, 2010a, b) revealed that the subjective norms have an impact on the perceived usefulness<br />
of IT, the findings of the current research does not support that. It has been revealed from the results that<br />
teachers and students will not feel that the educational technology can simplify their work and make it<br />
more effective if their peers or important people are using or think to use it. Finally, the results show that<br />
even by having a good experience in using computer or any educational technology, teachers and<br />
students will still express a fear of being under surveillance and believe that their wok conditions will<br />
worsen with the introduction of such technologies.<br />
Schools of Future project which aim at establishing a fundamental change in the way teaching and<br />
learning takes place, and transferring the traditional classroom into an open, interactive environment<br />
based on a wide-ranging use of technology, is the first step toward the educational reform at Kingdom of<br />
Bahrain. Hence, the research results can present a good insight of the wider range of issues that<br />
academic institutions would need to address when adopting educational technology in schools. Schools<br />
and ministry of education leaders at Kingdom of Bahrain should invest in a strategic plan for the adoption<br />
of educational technology which should be focused more on developing the human resources and<br />
building IT skills and capabilities within the school. Actually the ministry of education has provided a great<br />
effort to enhance the technical skills of the teachers in all schools and especially those that involved in<br />
the KHSFP. For instant, ministry of education has integrated ICT in education program to be accessible<br />
to the teachers and students to gain computer skills. Moreover, ICT training programs have been<br />
developed with the aim to empower number of schools to move in pioneering on new ways of learning.<br />
The programs include number of teacher training courses on digital literacy including the use of key tools<br />
such as Microsoft office, visual basic and web-based applications. These programs will help teachers to<br />
get a certified qualification and to use their ICT ability in their daily classroom life.<br />
Acknowledgement<br />
I would like to thank my senior students including: Bayan Abdulla, Mona Ali, and Nawal Isa for the effort<br />
that they have provided in collecting the data used for the current research.<br />
References<br />
Ajzen I. and Fishbein, M., (1980), Understanding Attitudes and Predicting Social Behavior, Prentice Hall, Englewood<br />
Cliffs, New Jersey, 1980<br />
Ajzen, I., (1991), “The Theory of Planned Behavior”, Organizational Behavior and Human Decision Processes, Vol.<br />
50, pp. 179-211<br />
Al-Ammary, Jaflah (2010a),”Factors affecting accepting and using of computer and internet by elderly people at<br />
Kingdom of Bahrain”, in proceeding of the International <strong>Conference</strong> of Management and Evaluation (ICIME<br />
2010), Cape Town, South Africa, 25-26 March<br />
Al-Ammary, Jaflah and Hamad, Sherifa (2008), “Factors influencing adoption of E-learning at University of Bahrain”,<br />
in proceeding of the International <strong>Conference</strong> on Information Technology (ACIT 2008), University of Sfax,<br />
Tunisia, 16-18 December<br />
Al-Ammary, Jaflah, Al Rowaili, Fatima, Al Noaimi, Latifa, Dodhy, Mona, and Zain, Maasoma (2010b),”Building<br />
Professional Learning Community in the secondary schools: a step toward the educational reform at Kingdom<br />
of Bahrain”, in proceeding of the 7 th International <strong>Conference</strong> on Knowledge Management (ICKM 2010),<br />
Pittsburg, Pennsylvania, 22-23 October<br />
Alba, J.W. and Hutchinson, J.W. (1987), “Dimensions of consumer expertise”, Journal of consumer research, Vol. 13,<br />
Issue 3, pp. 411-454<br />
Bertrand, Manon and Bouchard, Stephane (2008), “Applying the technology acceptance model to VR with people<br />
who are favorable to ITS use”, Journal of Cyber Therapy and Rehabilitation, Vol. 1, Issue 2<br />
Brosnan, Mark (1998), ‘The impact of computer anxiety and self efficacy upon computer performance’, The Journal<br />
of Computer Aided learning, Vol.14, pp. 223-234<br />
Chang, Su-Chao and Tung, Feng-Chang (2008), “An empirical investigation of students’ behavioral intentions to use<br />
the online learning course website”, British journal of education technology, Vol. 39, Issue 1<br />
Chin, C. and Todd, P. (1995), “On the use, usefulness and ease of use of structural equation modeling in MIS<br />
research: a note of caution”, MIS quarterly, Vol. 19, Issue 2<br />
Chiu, C., Hsu, M., Sun, S., Lin, T., and Sun, P., (2005), “Usability, quality, value, and E-learning continuance<br />
decisions.”, Computers and Education, Vol. 45, Issue 4, pp. 399-416<br />
Compeau, D.R. and Higgins, C.A.,(1995), “Computer self-efficacy: development of a measure and initial test”, MIS<br />
Quarterly, Vol. 19, Issue 2, pp.189-211<br />
16
Jaflah Al-Ammary<br />
Compeau, D.R. and Huff, S., (1999), “Social cognitive theory and individual reactions to computing technology: a<br />
longitudinal study”, MIS Quarterly, Vol. 23, Issue 2, pp.145-158<br />
Davis, F.,(1998), “Perceived Usefulness, Perceived Ease of Use, and User Acceptance of Information Technology”,<br />
MIS Quarterly, Vol. 13, Issue 3, pp. 319-339<br />
Davis, F.D., Bagozzi, R.P, and Warshaw, P.R ,(1989),”User acceptance of computer technology: a comparison of<br />
two theoretical models”, Management science, Vol. 35, pages 982-1002, 1989.<br />
Doll, W.J, A. Hendrickson and X. Deng (1998), “Using Davis’s Perceived Usefulness and Ease-of-use Instruments for<br />
Decision Making: A Confirmatory and Multi-group Invariance Analysis”, Decision Sciences, Vol. 29, Issue 4, pp.<br />
839-869<br />
Dyck, J. L. and Smither, J. A (1996), “Older adults' acquisition of word processing: the contribution of cognitive<br />
abilities and computer anxiety”, Computer in Human Behavior, Vol. 12, Issue 1<br />
ESCWA (2005), National Profile for the information society in Bahrain, Economic and Social commission for Western<br />
Asia, United Nation, report, available online on http://www.escwa.un.org/wsis/docs/bahrain-2005-e.pdf<br />
Fishbein, M. and Ajzen, I.,(1975), Belief, attitude, intentions and behavior: an introduction to theory and research,<br />
Addison-Wesley, Reading, MA<br />
Havelka, D., (2003), “Students Beliefs and Attitudes Toward Information Technology”, In The Proceedings of<br />
ISECON 2003, Vol. 20 (San Diego), Revised in Information Systems Education Journal, Vol. 1, Issue 40, 2003.<br />
Hayashe, A.; Chen, C.; Ryan, T. and Wu, J., (2004),”Role of Social Presence and Moderating Role of Computer Self<br />
Efficacy in Predicting the Continuance Usage of E-Learning Systems”, The Journal of Information System<br />
Education, 2004<br />
Hennessey, B.A. and Amabile, T.M. (2005), “Extrinsic and intrinsic motivation”, in Nicholson, O.Audia, and M. Pillutia<br />
(Eds), Blackwell Encyclopedia of Mnagement, second edition Malden, MA., Blackwell publishing<br />
Hill, T.; Smith, N.D. and Mann, M.F., (1986),“Communicating Innovations: convincing computer phobic’s to adopt<br />
innovative technologies”, Advances in Consumer Research, Vol. 13<br />
ICE47 (2001), Main challenges facing the educational system at the beginning of the 21st century, UNESCO report,<br />
available online at: www.ibe.unisco.org/international/ICE47/English/natreps/Bahrain-en-part-2.pdf<br />
Karsten, R. and Roth, rm. (1998), “Computer self efficacy: a practical indicator of student computer competency in<br />
introducing IS courses’, Information Science, Vol. 1 Issue 3. Pp. 61-68<br />
Khan, Ted and Taber Ullah, Linda (1997),” learning by design: integrating technology into the curriculum through<br />
student multimedia design project”, New Horizons for Learning, Quarterly journal, available online at<br />
HTTP:www.newhorizons.org<br />
Kim, Young; Chun, Jae and Song, Jaeki (2009), “Investigating the role of attitude in technology acceptance from an<br />
attitude strength perspective”, International Journal of Information management, Vol. 29<br />
Lee, Y. C., (2006),”An empirical investigation into factors influencing the adoption of e-learning system”, Online<br />
Information review, Vol.30, No. 5<br />
Saade, R. and Kira, D. (2006), “The emotional state of technology acceptance”, Issues in Information Science and<br />
Information Technology, Vol. 3, pp. 529-540<br />
Sam, H.K.; Othman, A.E. and Nordin, Z.S. (2005), “Computer self-efficacy, computer anxiety, and attitudes toward<br />
the internet: a study among undergraduates in Unimas”, Educational Technology and Society, Vol. 8, pp. 205-<br />
209<br />
Segars, A., and Grover, V. (1993), “Re-examining perceived ease of use and usefulness: A confirmatory factor<br />
analysis”, MIS Quarterly, Vol. 17, Issue 4, pp. 517-525.<br />
Shaker, A. (2008), “Empowering Bahrain’s teachers and students to succeed in the digital future”, Partners in<br />
learning Middle East and Africa, available online<br />
Srite, Mark (2006), “Culture as an explanation of technology acceptance differences: an empirical investigation of<br />
Chinese an US users”, Australasian Journal of Information Systems, Vol. 14, Issue 1<br />
Taylor, S. and Todd, P.A., (1995 a),“Understanding information technology usage: a test of competing models”,<br />
Information Systems Research, Vol. 6 Issue 2, pp. 144-76<br />
Taylor, S. and Todd, P.A., (1995 b),“Assessing IT usage: the role of prior experience”, MIS Quarterly, Vol. 19, Issue<br />
4, pp. 561-70<br />
Trocchia, P. and Janda, S. (2000), “A phenomenological Investigations of Internet usage among older individuals”,<br />
Journal of Consumer marketing, Vol. 17, Issue 7, pp. 605-616<br />
Venkastesh, V. and Morris, M.G. (2000), “Why don’t men ever to stop to ask for directions? Gender, social influence,<br />
and their role in technology acceptance and usage behavior”, MIS Quarterly, Vol. 24, Issue 1<br />
Venkatesh, V. and Davis, F.,(2000),”A Theoretical Extension of the Technology Acceptance Model: Four Longitudinal<br />
Field Studies”, Management Science, Vol. 46, Issue 2, pp. 186-204<br />
Venkatesh, V. and Davis, F.D., (1996), “A model of the antecedents of perceived ease of use: development and test”,<br />
Decision Sciences, Vol. 27, pp. 451-481<br />
Venkatesh, V., Morris, M. G., Davis G. B., and Davis, F. D.,(2003), “User Acceptance of Information Technology:<br />
Toward a Unified View”, MIS Quarterly, Vol. 27, Issue 3, pp. 425-478<br />
Vijayasarathy, L. R., (1996), “Predicting consumer intentions to use on-line shopping: The case for an augmented<br />
technology acceptance model”, Information and Management, Vol. 41, Issue 6, 747-762, 2004.<br />
Wu, I. L. and Chen, J. L.,(2005), “An extension of Trust and TAM model with TPB in the initial adoption of on-line tax:<br />
An empirical study”, International Journal Human-Computer Studies, Vol. 62, 2005.<br />
17
Comparison of Feature Selection Techniques Using Fully-<br />
Controlled Simulation-Based Datasets<br />
Suzan Arslanturk 1 , Mohammad Siadat 1 , Theophilus Ogunyemi 1 , Ishwar Sethi 1 and<br />
Ananias Diokno 2<br />
1<br />
Oakland University, USA<br />
2<br />
Beaumont Hospital, USA<br />
sarslant@oakland.edu<br />
siadat@oakland.edu<br />
ogunyemi@oakland.edu<br />
isethi@oakland.edu<br />
adiokno@beaumont.edu<br />
Abstract: Data mining is the discipline of systematically reviewing datasets to determine what patterns,<br />
concurrences and/or rule sets can be discovered. In an effort to understand the effectiveness of a proposed mining<br />
methodology a well defined dataset is required for model verification. To this end, a fully-controlled simulation was<br />
created by which several different feature selection algorithms were evaluated. In this paper, we present a<br />
comprehensive comparison between attribute selection methods when noise, missing values and multicollinearity are<br />
in question. Our results show that "Relief" and "information gain" have outperformed other feature selection methods<br />
available in Weka when considering both sensitivity and specificity measures. We have evaluated the following<br />
features selection methods: J48, Relief, information gain, consistency based feature selection and correlation based<br />
feature selection to see which one handles additive noise better. The sensitivity of consistency based feature<br />
selection was 11% higher than the average sensitivity of other methods. However, it's specificity was 37% lower than<br />
that of the average. It is important to note that sensitivity or specificity alone does not give enough support to a<br />
method to say that it is the best way to handle the data. The best method, when both sensitivity and specificity are<br />
considered, was information gain. This method outperformed the average of other methods by 1% and 20.5% when<br />
we considered its sensitivity and specificity, respectively. Also in this regard, a goal of our study was to see which<br />
feature selection method outperforms within the missing set of values. In this case, when looking again at sensitivity<br />
and specificities, Relief and information gain proved to outperform the other methods of our study by 7.2% and<br />
12.4%, respectively. Our studies also show that when multicollinearity is embedded into the fully controlled dataset<br />
without any noise and missing values, the correlation based feature selection outperforms other methods. In<br />
summary Relief and information gain performed the best in all three situations in terms of its sensitivity and<br />
specificities.<br />
Keywords: attribute selection, rule extraction, classification, comparison techniques, simulation<br />
1. Introduction<br />
Feature selection is a machine learning method that selects an optimal subset of features by eliminating<br />
the ones which contain less predictive information. Reducing the dimensionality of a feature space<br />
improves the performance by diminishing the curse of dimensionality effect. Feature selection also gains<br />
advantage from efficiency in terms of storage and computational costs. Also, the execution time spent for<br />
both training and testing phases will decrease.<br />
The feature selection methods are categorized into three different forms: filter, wrapper or embedded<br />
(Molina 2002). It is important to evaluate the existing methods and figure out which one performs better in<br />
certain situations. Some algorithms perform well on correlated data while others can handle noise or<br />
missing values depending on the nature of the data. In filtering methods, feature selection is a<br />
preprocessing step independent of the induction algorithm. Information gain algorithm determines the<br />
importance of each feature by evaluating the uncertainty reduction, while ReliefF is sampling an instance<br />
and evaluating the difference between its nearest neighbors from both the same and the opposite class.<br />
Relevance scores are assigned to each attribute. Correlation based attribute selection methods eliminate<br />
one of the less important attribute that correlates with another attribute (Hall 1998). In embedded<br />
methods, inducer has its own feature selection algorithm embedded such as J48, a widely used decision<br />
tree algorithm. The occurrence of an attribute in a tree provides information about the importance of that<br />
particular attribute. Information gain and entropy reduction methods can be applied to each candidate<br />
feature of the decision tree node to evaluate the importance of each attribute (Sugumaran 2006). There<br />
have been several studies on feature selection methods. Hall and Holmes (Hall 1998) use the UCI<br />
dataset, a real world data set, that contains different data types such as categorical, continuous and<br />
multivariate. Six attribute selection methods are compared in terms of classification accuracy, reduction<br />
rate and speed. Molina et al., (Molina 2002) use a simulation based randomly created binary and nominal<br />
18
Suzan Arslanturk et al.<br />
valued dataset. Different types of syntactic functions are applied to the dataset to generate the class<br />
labels. Since it is a fully controlled scenario varying number of relevant, irrelevant and redundant features<br />
are placed in the dataset. However, the simulated dataset doesn’t lend itself to real longitudinal dataset of<br />
clinical trials and the functions used in this paper does not care about any particular attribute when<br />
deciding about the resultant, which does not fit to what we usually see in real world biomedical datasets.<br />
In this research, on the other hand, several feature selection algorithms are applied to a simulation based<br />
dataset with longitudinal trials in order to compare the performances of the algorithms in terms of different<br />
noise levels and different missing value levels and combination of both. Multicollinearity is added to the<br />
dataset to evaluate the most robust algorithms when dependencies between attributes are in question.<br />
Same attribute selection techniques are also applied to the combination of the longitudinal datasets. A<br />
simulation dataset is chosen since full control over the dataset is achieved.<br />
This paper is organized as follows. In Section 2 we explain different attribute selection techniques.<br />
Section 3 outlines how the database is created and how the rules are embedded into the dataset. In<br />
Section 4 the programming language used to embed the rules and the data mining tool to perform the<br />
feature selection algorithms are explained. Section 6 and Section 7 present the results.<br />
2. Attribute selection techniques<br />
Attribute selection techniques can be categorized into three different categories: filter, wrapper and<br />
embedded methods. In filtering methods, the feature selection method takes place before any learning<br />
algorithm. The undesirable attributes are filtered out before the classification step. All the training data is<br />
used in filtering methods (Hall 2003). In embedded methods, the learning algorithm has its own feature<br />
selection algorithm embedded in it (Molina 2002). J48 decision tree classification algorithm is a common<br />
example of an embedded method. In wrapper mode, on the other hand, the feature selection algorithm<br />
uses the learning algorithm as a sub-routine (John 1994).<br />
Five different attribute selection methods are applied to the dataset. Wrapper methods which are<br />
correlation based feature selection and Relief, a filtering method, information gain and an embedded<br />
method J48 decision tree based feature selection are applied to the dataset and the results are compared<br />
in terms of sensitivity and specificities.<br />
2.1 Correlation based feature selection<br />
Correlation based feature selection evaluates the dependencies between attributes and eliminate the<br />
ones which are correlated to each other. The irrelevant and redundant data has to be removed as much<br />
as possible. After the feature selection, the remaining data has to be highly correlated with the class and<br />
uncorrelated with each other.<br />
As equation 1 (Ghiselli, 1964) formalizes :<br />
Merits = (1)<br />
The feature subset S contains k different features where rfc is the feature to class correlation and rff is the<br />
feature to feature correlation. In order to have a good feature selection algorithm the merit has to be<br />
maximized.<br />
Symmetrical uncertainty can be evaluated as follows where H(X) and H(Y) are marginal entropies.<br />
Symmetrical uncertainty = 2.0 x (2)<br />
2.2 Consistency based feature selection<br />
The consistency of the class is evaluated by first figuring out all different combinations of the attributes.<br />
For each different combination the consistency is calculated by differentiating the number of occurrences<br />
of a particular attribute from the cardinality of the majority class (Hall 1998).<br />
19
2.3 Information gain<br />
Suzan Arslanturk et al.<br />
Information gain feature selection method the uncertainty of the class is evaluated with and without the<br />
attribute observation (Hall 1998).<br />
H(C) = - (3)<br />
H(C|A) = - (4)<br />
2.4 Relief<br />
Relief algorithm assigns a relevancy score to all the attributes in descending order. The algorithm selects<br />
an instance in each iteration and finds the nearest neighbor from the same and opposite class. The<br />
results should differentiate if the neighbor and the instance selected are in opposite classes (Hall 1998).<br />
The algorithm can handle noise if the neighbor number, k, is increased.<br />
2.5 J48 decision tree based feature selection<br />
J48 decision tree is a classification method that can also be used for feature selection. Each node of a<br />
tree involves an attribute and the occurrence of each attribute provide information about the importance<br />
of that particular attribute (Sugumaran 2006).<br />
The importance of each attribute can be evaluated by applying the information gain formula to each node<br />
of the tree.<br />
3. Dataset<br />
10 different simulation datasets are created each containing 20 attributes and 1500 subjects. In order to<br />
gain full control over the dataset the values in the dataset are formed to create rules (Agraval 1994). The<br />
last column of the datasets specify the resultants i.e. the class labels. Resultant is defined as the<br />
continence, expressed as 0, or incontinence, expressed as 1, of the subject. Resultant values are stored<br />
in the column space of the matrix.<br />
Subjects are defined as an individual and are represented in the row space of the matrix. The number of<br />
subjects is a function of the number of rules and the partition size for each rule. For example with ten<br />
rules and two hundred subject partition size there would be two thousand subjects represented in the<br />
matrix, with subjects one to two hundred being in the first partition, subjects two hundred and one to four<br />
hundred in the second, etc.<br />
Having defined the base dataset, it is now possible to manipulate the dataset for further methodology<br />
testing. There are two types of modifications that can be performed to the data set for this testing. First is<br />
the inclusion of noise to the data set. The second type of modification is to incorporate missing values,<br />
which shall be defined as an attribute missing a value.<br />
The expected results of the feature selection algorithms are the attributes that are the entities of the rules<br />
embedded to the datasets.<br />
A simulation dataset is used to allow maximum flexibility in creating and manipulating the data sets. The<br />
simulation allows for one base data set to be processed with multiple noise levels, multiple missing value<br />
levels and a mixture of both noise and missing values. The advantage of this methodology allows the<br />
researcher to understand the impact of varying levels of these factors on feature selection or any other<br />
metric of interest. Further, we can determine up to what missing value and noise levels a given data<br />
mining tool would be reliable. This is extremely important when one needs to use the tool on a real<br />
application. The current version of the matrix creation algorithm was intentionally restricted to binary data<br />
only. This was chosen primarily to validate the feature selection methods that were the driving force<br />
behind creating this simulated data. With the methodology described above modification to alternative<br />
data types, such as categorical and/or continuous, will be possible.<br />
4. Experimental methodology<br />
Our experiment applied the five different feature selection algorithms to the simulation datasets that were<br />
created using MATLAB. To apply the attribute selection methods Weka (Waikato Environment for<br />
20
Suzan Arslanturk et al.<br />
Knowledge Analysis) is used. Weka is an open source Java based data mining tool that performs pre<br />
processing, classification, clustering and many other machine learning tools by using a graphical user<br />
interface. First the dataset is uploaded to the Weka Explorer mode and the data is discretized. Then all<br />
the attribute selection algorithms are performed. Sensitivity and specificity are used to evaluate the<br />
performances in a statistical manner. Sensitivity i.e. true positive rate; measures the proportion of actual<br />
positives which are correctly identified. While specificity (1 – false positive probability) measures the<br />
proportion of negatives which are correctly identified. The algorithms with higher sensitivity and higher<br />
specificity are the ones that performs significantly better than the others.<br />
5. Results and discussions<br />
Our experiment applied five different feature selection algorithms to the simulation dataset and the results<br />
are compared when noise, missing values and multicollinearity are added to the data. The first set of<br />
experiments was designed to evaluate the following feature selection methods to see which one handles<br />
additive noise better: J48, Relief, information gain, consistency based feature selection and correlation<br />
based feature selection.<br />
In order to refer to a feature selection algorithm as robust, it has to both has a high sensitivity and a high<br />
specificity. Fig. 1 shows that J48 algorithm performs well in terms of sensitivity but there is a huge<br />
decrease in the specificity curve (Fig. 2) which makes the algorithm less desirable than the others.<br />
Without any noise Information gain and ReliefF algorithms both perform well. Consistency based feature<br />
evaluation has a low specificity and J48 decision tree classification algorithm has a low sensitivity. The<br />
results do not change in Cfs, Information Gain and Relief algorithms when the noise level is 2%, 5%,<br />
10%, 15%, ,respectively. In spite, in J48 decision tree the specificity decreases and the sensitivity<br />
increases rapidly when the noise level increase. The sensitivity of consistency based feature selection<br />
was 11% higher than the average sensitivity of other methods. However, its specificity was 37% lower<br />
than that of the average. It is important to note that data method that maximizes both the sensitivity and<br />
specificity is of interest. The best method, when both sensitivity and specificity are considered, was<br />
information gain. This method outperformed the average performance of the other methods by 1% and<br />
20.5% when we considered its sensitivity and specificity, respectively. Consistency based subset<br />
evaluation and J48 algorithms can not handle noise. The next set of experiments was designed to<br />
evaluate the feature selection methods to see which one handles missing values better. Fig. 3 and Fig. 4<br />
show the results of different missing value levels. In<br />
this case, when considering sensitivity and specificity, Relief and information gain were proved to perform<br />
better compared to the other methods of our study by 7.2% and 12.4%, respectively. Despite, the<br />
Consistency based feature selection and J48 algorithms can not handle missing values effectively. Our<br />
studies also showed that when multicollinearity was embedded into the dataset without any noise and<br />
missing values, the correlation based feature selection method outperformed other methods. In summary,<br />
Relief and information gain were the best in all three situations (noise, missing value, multicollinearity)<br />
when both sensitivity and specificity were considered.<br />
Figure 1: Noise-sensitivity<br />
21
Figure 2: Noise - specificity<br />
Figure 3: Missing values-sensitivity<br />
Figure 4: Missing values-specificity<br />
Suzan Arslanturk et al.<br />
22
6. Conclusions and future work<br />
Suzan Arslanturk et al.<br />
This paper presented a comparison between different attribute selection techniques by using a data<br />
mining tool, Weka. The results show that using different attribute selection techniques help the system by<br />
reducing the curse of dimensionality, time and storage. There is no single attribute selection technique<br />
that gives the best results. The advantages of each technique differs in different situations. One might<br />
give the most accurate results in a certain situation but the storage or time to run the algorithm might take<br />
too long which will effect the efficiency of the algorithm. However, the results show that the Relief and<br />
information gain algorithms give the most accurate results when missing values and noise are added.<br />
This study shows that if in a given dataset noise level is beyond 15% or missing value is beyond 12% the<br />
selected attributes are not reliable even when one uses Relief or information gain algorithms.<br />
The simulation was developed to allow for a particular randomly created matrix to be manipulated with<br />
differing noise, missing value, multicollinearity and combinations of both. Further development of this<br />
simulation model building tool will need to include categorical and continuous data, to allow for more<br />
realistic simulation models.<br />
References<br />
Agrawal, R., Srikant, R. (1994) Fast Algorithms for Mining Association Rules, Proceedings of the 20 th VLDB<br />
<strong>Conference</strong> Santiago, Chile.<br />
Ghiselli, E. E. (1964) Theory of Psychological Measurement.<br />
Hall, M., Holmes, G. (1998) Benchmarking Attribute Selection Techniques For Discrete Class Data Mining,<br />
Transaction on Knowledge and Data Engineering, Vol. 15, No. 3.<br />
Hall, M., Smith, L.A. (2003) Feature Selection for Machine Learning: Comparing a Correlation Based Filter Approach<br />
to the Wrapper, American Association of Artificial Intelligence.<br />
John, G., Kohavi, R. and Pfleger, K. (1994) Irrelevant features and the subset selection problem, in Proceedings of<br />
the International <strong>Conference</strong> on Machine Learning, San Francisco, CA.<br />
Molina, L., Belanche L., Nebot A.(2002) Feature Selection Algorithms: A survey and Experimental Evaluation.<br />
Second IEEE International <strong>Conference</strong> on Data Mining<br />
Sugumaran, V., Muralidharan, V., Ramachandran, K. I.(2006) Feature Selection Using Decision Tree and<br />
Classification through Proximal Support Vector Machine for fault Diagnostics of Roller Bearing, Mechanical<br />
Systems and Signal Processing.<br />
23
A Delphi-Multi-Criteria Decision Making Approach in the<br />
Selection of an Enterprise-Wide Integration Strategy<br />
Tridip Bardhan, James Ngeru, and Richard Pitts Jr<br />
Morgan State University, Baltimore, USA<br />
tridip.bardhan@morgan.edu<br />
jmngeru@gmail.com<br />
richard.pitts@morgan.edu<br />
Abstract: Driven by the need to be competitively effective and agile, the last few decades have witnessed explosive<br />
interest in the field of enterprise integration (EI). In spite of a great deal of research and advances in the EI discipline,<br />
there still seems to lack a comprehensive decision making framework which can assist decision makers during the<br />
crucial task of selecting an appropriate integration strategy that would optimally address most of the organization’s<br />
business and technical integration needs. Consequently, this paper attempts to fill the research gap by formulating<br />
and illustrating an application of such a framework through a case study. Since selection of an EI strategy can be<br />
regarded as a complex multi-criteria decision making problem that involves multiple participants who must weigh in a<br />
set of business and technical factors to evaluate different EI strategies, an integrated multi-criteria decision approach<br />
which combines Delphi techniques and the Analytic Network Process (ANP) is adopted. The use of the ANP is<br />
preferred over other multi-criteria decision analysis methods due to its ease of implementation and it provides a<br />
structure that allows for an intensive analysis of the criteria which shows the effects of various criterions to be taken<br />
into consideration. The Delphi technique on the other hand helps to ensure a consensus in the decision making<br />
process by allowing active participation of various decisions makers and assimilation of their different schools of<br />
thought. The formulated framework is then illustrated as a case study of a state government agency that is<br />
embarking on enterprise-wide integration. The choice of the case study in the public sector setting was preferred<br />
since these organizations are generally divided into several organizational functions; hence, such strategic decisions<br />
usually involve different decision makers who represent their functional units. As a result, EI decisions which are<br />
generally made at an enterprise-wide level are more challenging in public organizations. The results from the case<br />
study indicate that there exists a conflict of interest among decision makers who participate in the EI strategy<br />
evaluation and selection process. However, with the application of the proposed model, an effective decision making<br />
process can be realized with significant consensus among decision makers. The use of the proposed model in this<br />
context is a ground-breaking concept and represents a scientific and systematic approach to decision making that<br />
could minimize risks and improve the benefits generally associated with the selection of an enterprise-wide<br />
integration strategy.<br />
Keywords: enterprise integration, multi-criteria analysis, delphi technique, analytic network process<br />
1. Introduction<br />
Today’s business environment is characterized with globalization, customization, digitalization,<br />
virtualization and agility (Huang and Fan, 2007). Organizations are competing and creating new<br />
opportunities in global markets with global partners and consumers. The competitive and dynamic nature<br />
of the new business playground means new technologies methods and strategies are needed. To be<br />
competitive, businesses in both private and public sector must have enough agility to respond to<br />
business needs and with effective service delivery. This on the other hand is only possible with rapid<br />
sharing of information among business processes both within and across organization. For the last few<br />
decades, Enterprise Integration (EI) has emerged as a revolutionary concept and solution that promises<br />
to assist organization with achieving this goal. EI aims to link separate business processes giving them<br />
increased leverage (Lam and Shankararaman, 2004). Thus it is not surprising that for the past few years,<br />
survey after survey of top CIO have ranked EI initiatives, such as Service Oriented Architecture (SOA)<br />
adoption, Enterprise Resource Planning (ERP) adoption/upgrades, Data integration, in their list of top ten<br />
priorities (Gartner-inc, 2009; Nascio, 2009). To many top executives, EI is now a business necessity.<br />
However, implementation of an EI strategy is not an easy task. Past researchers have called for thorough<br />
evaluation of organization’s requirements before undertaking EI implementation (Themistocleous, 2004).<br />
Thus, this paper aims at presenting formulation and application of a systematic methodology for analysis<br />
in the selection of an EI strategy. Selection of an EI strategy is a multifaceted problem involving more<br />
than the choosing a vendor or integration software system. It is important to consider the organization’s<br />
vision, requirements and constraints; the technical merits and limitations of the chosen strategy.<br />
In spite of remarkable interest in overall EI concept, little has been done in development of a framework<br />
for EI implementation, especially in EI strategy selection. Most of the EI research has been centered in<br />
advancing the technology to enable integration. Janssen argues that organizational issues and<br />
stakeholders interests should be addressed in the selection process (Janssen and Cresswell,<br />
24
Tridip Bardhan et al.<br />
2005).Thus, the need to develop a comprehensive framework incorporating a compendium of<br />
organizational factors during the selection of an EI strategy is indispensable (Giachetiti, 2004).<br />
The methodology discussed in this paper utilizes multi-criteria analysis principles, specifically integrating<br />
ANP method with Delphi technique. Analytic Network Process (ANP) developed by Saaty (1980) was<br />
chosen for its ability to incorporate multiple factors with different relative weights and with dependencies<br />
characteristics. This gives the decision makers the ability to prioritize the importance of one factor over<br />
another. ANP has been used in complex problems involving technology selection, and thus it is not a far<br />
to see that it may be applicable in this situation. To further improve the results on using ANP within this<br />
context, Delphi technique is applied. The Delphi technique improves the process by facilitating active<br />
participation and consensus among evaluation process partakers representing different organization’s<br />
functions.<br />
The paper proceeds with an overview on EI, followed by methodology and a case study.<br />
2. An overview of Enterprise Integration (EI)<br />
Frye and Gulledge (2007) termed EI as “alignment of strategies, business processes, information<br />
systems, technologies, and data across organizational boundaries to provide competitive advantage”.<br />
Lack of this alignment is source of many organizational problems. The process of achieving integration<br />
involves all managerial and technological factors that enable cross-functional process integration (Frye<br />
and Gulledge, 2007). Integration is approached in various ways and at various levels, such as, physical,<br />
application, business and can also be achieved through enterprise modeling approach or any<br />
methodological approach facilitating consistent enterprise-wide decision making (Chen et al., 2008). EI is<br />
also classified on whether it involves integrating processes and applications within or outside<br />
organizational boundaries. Inter-organizational enterprise integration, attempts to integrate business<br />
processes between enterprises (B2B), such as Supply Chain Management systems (SCMs), or electronic<br />
purchasing processes (e-procurement). Intra-organizational integration is the integration of applications<br />
within the organization, attempting to integrate custom applications and packaged systems (Silveira and<br />
Pastor, 2006)<br />
Lammer et al (2008) acknowledged existence of a number of integration approaches and concepts,<br />
differentiated by integration’s level and architecture. However, Spackman et al (2004) and Chandra et al<br />
(2009) argued that beyond the marketing hype; “there is a great deal of overlap between the integration<br />
approaches”. Thus, at highest level of integration, while devising integration strategy for the enterprise,<br />
there are limited numbers of overall approaches available (Spackman and Speaker, 2004). EI<br />
approaches can therefore be broadly narrowed down into four basic categories, including (Spackman and<br />
Speaker, 2004; Chandra and Juarez, 2009): (i) Point-to-point based integration; (ii) ERP integration; (iii)<br />
Enterprise Application Integration (EAI) approach; (iv) and Service oriented integration approach.<br />
In early stages of integration project, determination of strategic integration approach is among the most<br />
important and challenging planning task. Decisions made during this process impacts negatively or<br />
positively most of the subsequent development activities, and consequently overall success and quality of<br />
the resulting system. This decision activity is usually complicated by the fact that it involves different<br />
participants, contributing in decision making process and who have different goals, criteria and perception<br />
about certain alternatives. The activity is further challenging as it comes in early stages of project<br />
lifecycle, hence hard to reason about the consequences of the decision made (Al-Naeem et al., 2005).<br />
2.1 Factors Influencing adoption of particular integration approach<br />
EI projects are strategic investments, and should be closely linked to organization vision, goals and<br />
strategies. These projects involve high expenditure, risk and with great impact in almost every<br />
organization’s aspects. Various organizations develop and implement integration projects motivated by<br />
completely different integration drivers. The initial rationale for EI adoption influences problem definition,<br />
methods of achieving goals and other subsequent activities (Wei et al., 2005).<br />
Lam et al (2007) categorized integration drivers into two categories namely; organizational drivers and<br />
project drivers. Puschmann et al (2001) identified five integration drivers including; software, financial,<br />
internal and external drivers. Themistocleous (2001) argued that many organizations are driven to<br />
integration initiatives based on organizational, operational, technical and strategic perceived benefits.<br />
These drivers and benefits should be taken into consideration when an enterprise is devising and<br />
25
Tridip Bardhan et al.<br />
evaluating different integration strategies (Themistocleous and Irani, 2001; Janssen and Cresswell, 2005;<br />
Themistocleous, 2004).<br />
As EI involves an attempt to connect a number of different types of software products, the selection of<br />
integration approach should not only be based on functional and business requirements, but should also<br />
take into consideration non-functional and other technical requirements. Non-functional requirements<br />
describe properties that a system should possess. They are defined by stakeholders and greatly<br />
influence adoption of a particular integration approach. Non-functional requirements should not be<br />
overlooked, since doing so could lead to a poor system quality, unsatisfied stakeholders and eventually<br />
unsuccessful investment (Al-Naeem et al., 2005; Al-Naeem et al., 2004). Spackman et al (2004)<br />
proposed a Software Quality Attribute Trading (SQUAT) technique, which takes into consideration known<br />
software qualities attributes to evaluate integration approaches and technologies. Silveira et al (2006)<br />
developed EAI evaluation tool by further characterizing and describing international software quality<br />
evaluation standard (ISO-9126) for integration tools. Some of these quality attributes includes;<br />
adaptability, platform Neutrality, scalability, security, reliability, modifiability, performance, interoperability,<br />
maintainability, flexibility and testability<br />
In general, selection and evaluation of integration approaches should take into consideration multiple<br />
factors, which include business, organizational, project and technical related factors (Bardhan and Ngeru,<br />
2009), as shown in Figure 1.<br />
Figure 1: EI approach factors (Bardhan and Ngeru, 2009)<br />
3. Methodology<br />
Evaluation and selection of integration approach is a complex and challenging process. It involves<br />
consideration of several organizational factors, from both technical and business perspective. Moreover,<br />
the process consists of several contributors representing different levels of organizational functions and<br />
with varying goals and preferences. Thus, the problem requires an elaborate and systematic approach<br />
that would allow participants involved in the process to evaluate different alternatives by considering and<br />
expressing their preferences on multiple factors. That been the case then, application of multi-criteria<br />
decision analysis (MCDA) techniques is warranted.<br />
A plethora of MCDA methods exists, where most common includes; Analytic Hierarchy Process (AHP),<br />
Analytic Network Process (ANP), Simple Additive Weight (SAW), Simple Multi-Attribute Rating Technique<br />
(SMART), Elimination and Choice Translating Reality (ELECTRE) and Technique for Order Preference<br />
by Similarity to Ideal Solution (TOPSIS) (Triantaphyllou, 2000). Guitouni (2004) argued that the purpose<br />
of every MCDA method is to assist in making a good recommendation. However, not all MCDA methods<br />
produce a good recommendation for all situations; since every method’s theoretical and axiomatic<br />
development is based on some assumptions and hypothesis. Given the nature of EI approach selection<br />
problem, ANP is deemed as the most appropriate method to utilize during the evaluation process. ANP is<br />
preferred for this case due to its ability to provide an elaborate structure that allows for an intensive<br />
analysis on problem elements and the resolution of inter-dependency characteristics among the<br />
elements. However, ANP framework does not fully support group decision making process. Since the EI<br />
approach selection process involves a group of contributors with varying views and preferences, a sense<br />
of consensus among the contributors is an important element in ensuring the process overall success. It<br />
is therefore posited that incorporation of Delphi technique in ANP framework, illustrated in Figure 2, will<br />
facilitate vigorous involvement of process contributors and bring forth a sense of unanimity among them.<br />
26
Tridip Bardhan et al.<br />
More discussion about ANP can be found in (Saaty, 1980; Saaty, 1996) and on Delphi technique in<br />
(Brown, 1968; Hartman et al., 2007; Okoli and Pawlowski, 2004).<br />
Figure 2: Delphi-ANP evaluation methodology<br />
The Delphi-ANP methodology presented in this article consists of four Delphi rounds, which are blended<br />
in with ANP procedure to yield the following eight steps; (i) Problem formulation and factor identification;<br />
(ii) Forming evaluation panel; (iii) Defining evaluation Criteria; (iv) Structuring the problem network model;<br />
(v) problem elements appraisal; (vi) judgment weights analysis; (vii) elements ranking approval; (viii) and<br />
finally the best alternative selection (See Figure 2)<br />
To illustrate the ANP-Delphi methodology, this paper will consider an actual case study in a state<br />
government agency, referred hereinafter as TGA (real name withheld).<br />
4. Application of Delphi-ANP methodology for EI approach selection<br />
The Delphi-ANP methodology was applied and validated at TGA, a state government agency in eastcoast<br />
of USA. TGA is composed of several organizational functions and manages several engineering<br />
and transportation facilities across the state. However, many of these functions are supported by different<br />
systems for their day-to-day operations. Unfortunately, most of these systems exist within technical and<br />
functional walls of separation that creates barriers to integration and sharing of data; and thus inefficient<br />
business processing. Accordingly, the agency has taken several measures to address its integration<br />
challenges as an effort to achieve its strategic goals of effectiveness, efficiency and service improvement.<br />
At the time of this study, TGA was embarked in efforts to integrate some of its backbone information<br />
systems, such as those used in agency’s financial management, supply chain management and<br />
maintenance operations. In other similar integration projects, TGA has successfully developed in-house<br />
integration solutions by either establishing direct interfaces between applications, replicating data<br />
between databases or simply databases-consolidation. However, for the particular initiative considered in<br />
this study, any of these traditional approaches may not have been viable option for several reasons. First,<br />
the systems intended to be integrated are the backbone of the agency’s operations; and must therefore<br />
integrate with other systems in a different platform and located within and outside the agency’s boundary.<br />
Secondly, these systems uses different data format, and thus the issue of data uniformity presented a<br />
27
Tridip Bardhan et al.<br />
major integration challenge. Finally, these systems are in constant upgrade to the latest versions, which<br />
also requires any direct interface existing between them to be upgraded or redeveloped, which is rather<br />
laborious and expensive.<br />
Consequently, the management was seeking for a forward-looking approach that will allow future<br />
expansion of the integrated network system, and resonate with organization’s strategic goals. The<br />
approach must also provide sound, cost effective and long-term solution that is efficient and meets all of<br />
the stipulated needs of the agency. As a start, some high level integration approaches were suggested.<br />
These suggested approaches consisted; (1) Implement a Service oriented based integration, (2) Adopt a<br />
vendor-based EAI adapter, (3) Implement a single enterprise system such as ERP that would either<br />
replace these systems or integrate them and (4) built in-house integration solution that establishes direct<br />
interface links between the systems.<br />
Deciding on the approach to commit the SGA presented a great challenge to the management, as these<br />
approaches were perceived with varying strengths and weakness. The formulated Delphi-ANP<br />
methodology was therefore used in assisting the TGA management to decide the appropriate EI<br />
approach for their agency.<br />
The Delphi-ANP methodology process involved the following steps;<br />
Step 1: Problem formulation: The step comprised of preparatory and necessary tasks to help<br />
understanding the problem domain and set forth the evaluation process. The initial task is to set the<br />
objectives and goals of the evaluation process. Then, a review of literature, as well as interviews with<br />
domain experts was conducted in order to broadly identify feasible integration approaches and a list of<br />
criteria, which will possibly be used in gauging the identified alternatives.<br />
Step 2: Evaluation panel establishment - This step involved setting up a team of participants to<br />
contribute in the evaluation process. In a Delphi evaluation process, a group of five to twenty willing<br />
experts (Rowe and Wright, 2001; Armstrong, 2001) is generally considered adequate. Thus, a panel of<br />
six members consisting of two IT planning manager, two project managers and two IT analysts was<br />
established. The members selected were believed to be knowledgeable about the integration subject and<br />
organizational requirements.<br />
Step 3: Selection of organization’s specific criteria factors: - In this step, also marked as Delphi first<br />
round, the evaluation panel, formed in the previous step, was presented with an overview explaining the<br />
problem at hand, the purpose of the evaluation task, the expectations and procedures of the entire<br />
evaluation task process, and a list containing description of feasible alternatives and possible evaluation<br />
criteria. The panel was requested to select and justify factors that they would consider important in<br />
evaluating the presented alternatives. They were also requested for additional factors that may have<br />
been excluded from the list. From this round, the group identified a list of eighteen factors, categorized as<br />
strategic related, cost related, technical related and other organizational factors. The final list of the<br />
approved criteria is presented in Table 1.<br />
Step 4: Structuring problem into an ANP network model: In this step, the problem was configured to<br />
form initial ANP network structure, which consists of all the problem elements (i.e. the goal, criteria and<br />
alternatives) and also indicating any relationship existing between the elements.<br />
This step also entailed Delphi second round, where the evaluation panel was presented with proposed<br />
ANP network structure as well as Delphi first round results, which contained other members’ opinions<br />
(anonymous feedback) and any of their discrepancies. The members were given an opportunity to<br />
reconsider their judgments’. The results from this round were similar to the result from round one, and<br />
thus, consensus on evaluation criteria, presented in<br />
Table 1 and final problem structure formed, shown in Figure 3.<br />
Step 5: Appraisal of problem’s model elements: - First, a questionnaire was developed to assess the<br />
relationships between the elements of the network structure, through pair-wise comparison. The<br />
questionnaire was instituted in MS-excel application so as to facilitate analysis, like checking consistency<br />
on the fly (see Figure 4).<br />
28
Table 1: Evaluation criteria for EI approaches<br />
Tridip Bardhan et al.<br />
Step five was also marked as Delphi round three, where the participants were presented with instructions<br />
on how to use excel-ANP template and the final problem network structure developed in the previous<br />
step. Using Saaty’s 1-9 scale (Saaty, 1980), the participants were requested to gauge (i) relative<br />
importance of criterions with respect to goal; (ii) relative preference of alternatives with respect to each<br />
criterion and; (iii) any relative influences existing between criterions within and outside the cluster.<br />
29
Tridip Bardhan et al.<br />
Figure 3: Integration approach selection problem network system<br />
Figure 4: Pair-wise questionnaire built in MS Excel<br />
30
Tridip Bardhan et al.<br />
Step 6: Gathered judgment weights analysis: - The pairwise comparison was done using MS-excel<br />
template, with in-built capability to calculate elements local priorities and pairwise-matrix consistency.<br />
Thus, the first task of the analysis step involved obtaining and organizing each panel member’s pairwise<br />
comparison matrices, and their respective element priority vectors. The pairwise matrices were also<br />
assessed for the completeness and consistency (see Figure 5).<br />
Figure 5: Pair-wise comparison matrix as built in Microsoft Excel ®<br />
The element priorities were calculated using Saaty’s (1980) additive normalization algorithm, which<br />
consists of the following three steps:<br />
For each pairwise comparison matrix :<br />
Where is pairwise matrix from panel member , for cluster and containing relative measures,<br />
assigned to element when compared to element (Note that and )<br />
Sum the values of each column in comparison matrix i.e.<br />
Divide each element in matrix with respective column sum i.e.<br />
pairwise comparison matrix.<br />
to obtain normalized<br />
Sum each row’s element and divide row’s sum by elements in the row, to<br />
obtain vector (Eigen vector) containing estimates of relative priorities of the elements in the<br />
cluster i.e.<br />
Where: is the local priority for element of cluster , for panel member<br />
The consistency of each pairwise comparison matrix was checked using Saaty’s (1980) consistency<br />
index (CI) calculated as:<br />
31
Where<br />
Tridip Bardhan et al.<br />
RI = consistency random index of randomly generated reciprocal matrix form scale 1-9, (see table 2), and<br />
Where,<br />
And<br />
Aggregating members’ priorities: - Since the local element priorities varied from one panel member to<br />
another, there was a need to aggregate the results for all the participants. Saaty (1980) argued that<br />
geometric mean can be used to correctly represent a group of experts’ consensus. Thus, geometric<br />
mean, calculated using below equation, was used to aggregate participants’ element priority into a group<br />
element priority.<br />
Where: = the group priority of element in cluster<br />
Building super-matrix – To resolve interdependencies characteristics existing between the elements of<br />
the problem network system and eventual derivation of final elements’ priorities, ‘supermatrix’ concept<br />
proposed by Saaty (1996) is applied. A supermatrix is partitioned matrix, where each supermatrix’s block<br />
is composed of a set of relationships between elements as represented by the problem model. The<br />
elements’ relative weights obtained in the preceding steps, which indicates relative importance and<br />
influence of elements with each other, are entered in their respective block of the supermatrix, to form unweighted<br />
supermatrix, shown in Figure 7. The un-weighted supermatrix’s columns are not stochastic i.e.<br />
each column of supermatrix does not add to one. Thus, to obtain stochastic supermatrix (shown in Figure<br />
8), each block of un-weighted supermatrix is multiplied by corresponding cluster priority (for this study<br />
case, all the clusters had relative-equal weights). To obtain long-term stable set of weights, the stochastic<br />
supermatrix is successively raised with large powers, usually 2k+1 (where k is a large arbitrarily number),<br />
until the process converges to a limit supermatrix. The final priority weights of the elements are obtained<br />
by normalizing the limit supermatrix, shown in Figure 9.<br />
Table 2: Average random consistency (Saaty, 1980)<br />
32
Tridip Bardhan et al.<br />
Figure 6: Combining of individual element priority using geometric mean<br />
Figure 7: Un-weighted super-matrix<br />
Step 7: Element Ranking Approval – In this step, marked as fourth and last Delphi round, a report<br />
describing the analysis results was prepared and sent to the evaluation participants for their review. The<br />
report contained all the elements ranking based on the final priority weights obtained from normalized<br />
limit supermatrix. The report underscored the elements’ ranking, by indicating each alternative<br />
performance with respect to the criterions and each criterion relative importance with respect to the goal.<br />
Any major discrepancies between members’ weights was highlighted, and members were given a last<br />
opportunity to revise and justify their pairwise relative judgments’, especially where notable discrepancy<br />
exists. In this last Delphi round, two members revised some of their pairwise comparison, arguing that,<br />
while undergoing through the evaluation process, they have become more acquainted on EI related<br />
issues, and therefore felt the need to make adjustment to reflect their true judgments. The final element<br />
rankings are presented in Table 3, Figure 10, Figure 11 and Figure 12.<br />
33
Figure 8: Stochastic (weighted) super-matrix<br />
Tridip Bardhan et al.<br />
Figure 9: Normalized limit super-matrix<br />
Step 8: Selection of best alternative – This step marked the end of the evaluation process. The final<br />
report, which included members’ comments and changes requested from the previous step, was<br />
produced. The final report emphasized on the alternative ranking, see Figure 10, and more in particular to<br />
the first ranked alternative, which is actually the selected alternative.<br />
5. Results and discussion<br />
The use of Delphi-AHP evaluation approach provided a systematic framework in selection and evaluation<br />
of integration approaches. From the case study, EAI was ranked as the most appropriate approach to<br />
commit TGA (See Figure 10). The Delphi-ANP analysis presented members with comprehensive<br />
breakdown on how they reached to the final decision, by prioritizing factors with respect to goal, in Figure<br />
34
Tridip Bardhan et al.<br />
11, and by how each alternative relatively support different factors, as shown by the radar chart in Figure<br />
12.<br />
Table 3: Alternative performance with respect to criteria elements<br />
Figure 10: Alternative final score and ranking<br />
The case study results also showed that implementation cost was the major factor influencing the final<br />
decision, scoring 12.8%, followed by implementation time with 12%. Scalability and security were<br />
considered as the major technical factors with each scoring 7.8% and 7.6% respectively. In strategic<br />
perspective, business agility, EI approach overall alignment with organization strategy and cooperation<br />
with other business partners are leading factors, each scoring 6.8%, 5.3%and 5.3%, respectively.<br />
Overall, factors deemed to enhance organization interoperability were well favored; although ranked<br />
behind the factors related to approach implementation and security. These results are presented in Table<br />
3, Figure 10 and Figure 11.<br />
EAI approach scored average in almost every criterion considered. SOA approach was viewed as the<br />
most outstanding approach as far as organization’s strategy is concerned, but performed relatively poor<br />
in terms of cost related factors. ERP scored relatively well in technical related factors, especially on<br />
issues related to security, maintainability and complexity. However, it was viewed as an expensive<br />
approach. Finally, in-house application integration was favored in terms of cost related factors, although<br />
viewed as an expensive approach to maintain. In addition, in-house application integration ranked quite<br />
high in terms of organizational related factors such as acceptability and technical support/expertise<br />
availability.<br />
35
Tridip Bardhan et al.<br />
Figure 11: Criteria ranked by their percentage contribution to overall decision<br />
Figure 12: Radar chart showing alternatives performance with respect to criteria<br />
6. Conclusion and future work<br />
Decisions related to EI implementation and technologies are unstructured and complex, due to<br />
consideration of wide range factors, and involvement of different stakeholders. This paper demonstrated<br />
a Delphi-ANP evaluation approach, which systematically considers multiple factors and improves the<br />
ANP decision process by incorporating Delphi technique. The Delphi technique improved the process by<br />
first, allowing selection of requirements considered by the participants. Many factors included in the<br />
problem network mean more pair-wise comparison, and this may affect the process ability. The Delphi<br />
feedback also helped in improving consensus in the decision making process.<br />
The use of the Delphi-ANP approach allows EI selection to be logically and systematically structured<br />
within the context of the organization’s strategic vision. It enables decision makers to comprehensively<br />
consider the strengths and weakness of each EI strategy against various criteria and sub criteria.<br />
36
Tridip Bardhan et al.<br />
Although the process is quite demanding, its adoption allows productive and informative participation of<br />
different stakeholders with varying perspective to arrive at a consensus with proper documentation for<br />
such a complex decision.<br />
Additional testing of this model is encouraged in different environments. Additionally, sensitivity analysis<br />
to determine robustness of the decision should be performed.<br />
References<br />
Al-Naeem, T., Gorton I., Babar, M. A., Rabhi, F. and Benatallah, B. (2005) A Quality-Driven Systematic Approach for<br />
Architecting Distributed Software Applications.<br />
Al-Naeem, T., Rabhi, F. A., Benatallah, B. and Ray, P. K.. (2004) Systematic Approaches for Designing B2B<br />
Applications. International Journal of Electronic Commerce, 9, 41 - 70.<br />
Armstrong, J. S. (2001) Principles of forecasting: a handbook for researchers and practitioners, Springer.<br />
Bardhan, T. K. and Ngeru, J. (2009) “Application of Fuzzy Analytical Hierarchy Process in Enterprise Integration<br />
Approach Selection,” Proceedings (in CD-ROM) of the 14 th Annual International <strong>Conference</strong> on Industrial<br />
Engineering – Theory, Applications, and Practice, Anaheim, CA, October 18-21, 2009.<br />
Brown, B. B. (1968) Delphi Process: A Methodology Used for the Elicitation of opinions of Experts.<br />
Chandra, S. and Juarez, R. (2009) A Practical Approach to Enterprise Integration. Defense AT&L. Defence<br />
Acquisition University.<br />
Chen, D., Doumeingts, G. and Vernadat, F. (2008) Architectures for enterprise integration and interoperability: Past,<br />
present and future. Computers in Industry, 59, 647-659.<br />
Frye, D. W. and Gulledge, T. R. (2007) End-to-End Business Process Scenarios. Industrial Management \& Data<br />
Systems, 107, 749-761.<br />
Gartner-Inc (2009) Gartner EXP Worldwide Survey of More than 1,500 CIOs Shows IT Spending to Be Flat in 2009.<br />
Gartner Newsroom.<br />
Giachetiti, R. E. (2004) A Framework to Review the Information integration of Enterpise. International Journal of<br />
Production Research, 42, 1147-1166.<br />
Guitouni, A. (2004) Decision-aid to improve organisational performance. <strong>European</strong> Journal of Operational Research,<br />
162, 581-583<br />
Hartman, F. T., Krahn, J. and Skulmoski, G. J. (2007) The Delphi Method for Graduate Research. Journal of<br />
Information Technology Education, 6, 1-21.<br />
Huang, S. and Fan, Y. (2007). Model Driven and Service Oriented Enterprise Integration---The Method, Framework<br />
and Platform. In: ALPIT '07: Proceedings of the Sixth International <strong>Conference</strong> on Advanced Language<br />
Processing and Web Information Technology (ALPIT 2007), Washington, DC, USA. IEEE Computer Society,<br />
504--509.<br />
Janssen, M. and Cresswell, A. M. (2005) Enterprise Architecture Integration in E-government. In: 38th Hawaii<br />
International <strong>Conference</strong> on System Sciences.<br />
Lam, W. (2007) Enterprise Architecture and Integration: Methods, Implementation, and Technologies, Idea Group Inc<br />
(IGI).<br />
Lam, W. and Shankararaman, V. (2004) An enterprise integration methodology. IT Professional, 6, 40-48.<br />
Lammer, A., Eggert, S. and Gronau, N. (2008) A Procedure Model for a SOA - Based Integration of Enterprise<br />
Systems. International Journal of Enterprise Information Systems, 4, 1-12.<br />
NASCIO (2009) State CIO Priorities 2009. National Association of State Chief Information Officers Website.<br />
Okoli, C. and Pawlowski, S. D. (2004) The Delphi method as a research tool: an example, design considerations and<br />
applications. Information & Management, 42, 15-29.<br />
Puschmann, T. and Alt, R. (2001) Enterprise Application Integration - The Case of the Robert Bosch Group.<br />
Proceedings of the 34th Hawaii International <strong>Conference</strong> on System Sciences, 34.<br />
Rowe, G. & Wright, G. (2001) Expert Opinions in Forecasting. Role of the Delphi Technique. In: ARMSTRONG, J. S.<br />
(ed.) Principles of forecasting: a handbook for researchers and practitioners. Boston: Kluwer <strong>Academic</strong><br />
Publishers.<br />
Saaty, T. L. (1980) The Analytic Hierarchy Process, New York, NY, McGraw-Hill.<br />
Saaty, T. L. (1996) Decision Making with Dependence and Feedback: The Analytic Network Process, Pittsburgh,<br />
RWS Publications.<br />
Silveira, R. and Pastor, J. A. (2006) A Model For Enterprise Application Integration Tools Evaluation. In: <strong>European</strong><br />
and Mediterranean <strong>Conference</strong> on Information Systems (EMCIS), Costa Blanca, Alicante, Spain.<br />
Spackmanm, D. and Speaker, M. (2004) Enterprise Integration Solutions, Microsoft Press.<br />
Themistocleous, M. (2004) Justifying The Decisions For EAI Implementations: A Validated Proposition Of Influential<br />
Factors. Journal of Enterprise Information Management, 17, 85.<br />
Themistocleous, M. and Irani, Z. (2001) Benchmarking the Benefits and Barriers of Application Integration.<br />
Benchmarking: An International Journa, 8, 317 - 331.<br />
Triantaphyllou, E. (2000) Multi-Criteria Decision Making Methods: A comparative Study, Springer.<br />
Wei, C. C., Chien, C. F. and Wang, M. J. (2005) An AHP-Based Approach to ERP System Selection. Int. J.<br />
Production Economics, 96, 47-62.<br />
37
Selection, Implementation and Post Production of an ERP<br />
System<br />
Imran Batada 1 and Asmita Rahman 2<br />
1<br />
Institute of Business Administration, Karachi, Pakistan<br />
2<br />
University of Georgia, Georgia, USA<br />
Imran.batada@khi.iba.edu.pk<br />
Asmitab@uga.edu<br />
Abstract: Enterprise Resource Planning (ERP) is an enterprise level system that integrates all the business<br />
processes of an organization into a single database. ERP System enables all the departments of an organization to<br />
share and manipulate the data. ERP system has a centralized structure consisting of one single central database<br />
which can be accessed by all the department(s). This centralized structure is one of the major reasons of current<br />
ERP system‘s popularity. Selection of an ERP system plays a vital role in the implementation of ERP system. We<br />
used AHP method for the selection. Once proper selection has been made, the implementation process begins. This<br />
process entails; accurate requirement gathering, user training, technical training, testing, data migration etc. Our<br />
research will give an in-depth analysis and recommended approaches for the preparation of RFP document,<br />
selection of ERP system, selection of vendor, requirement gathering, system design, testing and final roll-out. The<br />
objective of our research is to ensure successful implementation of ERP system. Our research also proposes a<br />
potential model for implementing an ERP System. Our model proposes the organization to have two (2) major<br />
committees: (1) ERP Implementation Committee (2) ERP Steering Committee. ERP Implementation Committee will<br />
drive the complete implementation process. It must include members of all the department of the organization. Since,<br />
the system is for the functional users, it is essential that all the stakeholders are part of this committee as well. In<br />
addition, the ICT (Information & Communication Technology) department should also be a part of this committee.<br />
The other committee is the Steering Committee; this committee should consist of members of top management of the<br />
organization. All business process reengineering should be directed to the ERP Steering Committee. Steering<br />
committee is also responsible for supervising the overall progress of the implementation. They will also decide the<br />
level of customization necessary for the ERP System.<br />
Keywords: ERP Implementation process, selection of an ERP System, RFP development of an ERP system,<br />
proposed implementation model of an ERP system<br />
1. Introduction<br />
Enterprise Resource Planning System (ERP) is a centrally integrated system within all the departments of<br />
an organization. With the rapid evolution of web 2.0 and web 3.0, ERP has now become a necessity of<br />
today’s world. Due to its centralized nature and numerous benefits, more and more companies have<br />
adapted the ERP Systems. Implementation of an ERP system is a very significant and crucial process. It<br />
requires involvement of not only the implementation team, but also all the stake holders (including users<br />
etc). In this paper we will discuss the implementation process of an ERP System in detail and review the<br />
process with the help of a case study. We will also look at some of the common issues faced during the<br />
implementation with proposed solutions.<br />
2. Development of strategic and implementation committee<br />
Creation of a strategic and implementation committee plays a critical role in successful implementation of<br />
an ERP system. It is recommended to involve all the stake holders in the implementation process, as it<br />
not only gives them a better understanding of the system, but will also help modify the system throughout<br />
the process to meet the needs of the organization. It will also play a vital role during the post-production<br />
implementation process.<br />
3. Buy vs build<br />
While deciding on the type of the ERP system, there are two major options. An organization can either<br />
buy an existing ERP solution or they can build their own custom ERP system based on their<br />
requirements. This is one of the major decisions to be made by the organization based on their needs<br />
and available resources. (Kim J., 2009: 10) If an organization chooses to use an existing available<br />
solution, they must be a little flexible on the desired functionality. Even though, customization is an option<br />
while using an existing solution, too much customization is not recommended. (Zhao Y, and Fan S.,<br />
2007: 13). Surplus amount of customization may result in making the system unreliable. In addition,<br />
applying new patch and release may become a laborious and expensive operation.<br />
38
Imran Batada and Asmita Rahman<br />
On the hand if the organization chooses to build their own custom solution, they can be more obstinate<br />
about their requirements; however, building a new solution will take more time and such a system will<br />
also need time to mature. In addition, it usually costs more than using an existing available solution.<br />
Regardless, of whichever choice an organization makes, the decision should be made by the entire<br />
steering committee of the organization<br />
4. Development of Request for Proposal (RFP) document<br />
Development of the RFP document is important as this document states the basic requirements of the<br />
system. If the requirements are stated clearly and precisely, the implementation process will become<br />
easier. Therefore, it is important to spend considerable amount of time and effort for the development of a<br />
clear RFP document. A sample RFP document may look like the Figure 1.<br />
Figure 1: Popular RFP templates, 2010: 2<br />
The first column indicates the hierarchy of the desired modules. The second column indicates the<br />
criterion of the desired module. The third column indicates the priority of the desired module in the<br />
system (may range from 1-10, with 10 as the most important). The fourth column indicates if the given<br />
module is mandatory or optional. Mandatory implies that the organization must have the functionality in<br />
order to be able to proceed. These four (4) columns should be filled by the organization. The remaining<br />
six (6) columns should be filled by the vendor (in case of using an existing solution) or the developmental<br />
committee (in case of building a custom solution). Out of the box means that the functionality is already<br />
built-in into the System. Modification via supported means the functionality can be made available with<br />
minor configuration changes. Supported via third party solution means the functionality is not available in<br />
the system but it can be integrated with 3 rd party solution. Minor Customization means it can be achieved<br />
with minor changes in the coding. Major customization means it can be achieved with the major changes<br />
in the source code. FUT means the functionality is not available right now in the system but it will be<br />
available in the future release or future patch. NS means not supported in the system and it is also not<br />
recommended to do the changes in the source code for such functionality. It is basically core business<br />
processing changes.<br />
5. Selection of an ERP system<br />
The RFP document developed in the previous stage gives a clear picture of the desired features with<br />
their respective priorities. Once the requirements have been clearly stated, the implementation committee<br />
will need to select the ERP System. We can use methods like AHP method for selection of an ERP<br />
System (Kunz J., 2010: 7). Implementation committee must review each and every requirement stated in<br />
the RFP document and carefully propose to choose the vendor that satisfies most or all the desired<br />
39
Imran Batada and Asmita Rahman<br />
requirements (Lall and Teyarachakul, 2006: 1). While planning the level of customization, one must not<br />
plan the desired customization more than 10% (percent). Final decision about the selection should be<br />
made by steering committee based on the facts and the recommendations of the ERP implementation<br />
committee.<br />
5.1 Practical example<br />
The proposed method was successfully implemented in a higher education institute in Karachi, Pakistan.<br />
The university core business was Campus Solution System which they used to connect faculty and<br />
students. However, the existing system was not integrated and had multiple databases in it. In addition,<br />
proper reporting system was also lacking and duplicate efforts had to be made due to multiple databases.<br />
In order to resolve this issue, university decided to go for an ERP System. Management was also flexible<br />
to re-engineer their process and was willing to adopt best practices in the organization. (Quiscenti M,<br />
Bruccoleri M , La Commare U, Noto La Diega S, and Perrone G., 2006: 14).<br />
5.2 Selection Method for ERP system<br />
Steering committee created a team that included business users, functional owner and the IS team for<br />
selecting the system. Series of ERP System workshops were arranged for evaluation process. We<br />
selected core business functionality to evaluate the mandatory needs of the system for the organization.<br />
We selected AHP Method for selection an ERP System and incorporated the evaluation data in it. We<br />
identified nine major criteria and under each criterion relevant sub criteria that covered the complete<br />
expectation of the users (Shapira A, and Simcha M., 2009: 6). Figure 2 shows the results of three major<br />
systems that were considered.<br />
Figure 2: Results of three major systems<br />
40
Imran Batada and Asmita Rahman<br />
Figure 3 shows the priority values of each aspect. For example, functionality was assigned maximum<br />
weight-age as compared to other aspects. Questionnaires were prepared with a list of mandatory<br />
functionalities (Wei C, Chein C, and Wang M., 2005: 5).<br />
Figure 3: Priority values of each aspect<br />
The following figure (Figure 4) shows the final results based on the individual values from Figure 2 and<br />
Figure 3.<br />
Figure 4: Final results<br />
6. Selection of vendors<br />
While selecting the vendors based on the requirements, we may end up with more than one vendor who<br />
is providing a potentially suitable solution (Grossman T, and Walsh J. 2004:3). The following are the four<br />
major criteria to be considered while making the final decision:<br />
1. Must have similar business experience<br />
2. Must have recommendations by clients in the same geographical region<br />
3. Must have success stories of implementing same or similar solutions.<br />
4. Must have proven experience in providing ERP support service<br />
7. Recommended organogram for the organization<br />
Figure 5 shows the recommended organogram for the organization. Project director would be the core<br />
responsible person of the project; however ERP project manager will play a vital role both technically and<br />
functionally. He would also be responsible to conduct functional training session(s) for all the users.<br />
Figure 5: Recommended organogram<br />
41
8. ERP implementation process<br />
Imran Batada and Asmita Rahman<br />
A proposed ERP implementation process comprises of the following phases:<br />
8.1. Requirements Gathering Phase<br />
8.1.1 Requirements Workshop(s)<br />
8.1.2 Requirements Documentation<br />
8.1.3 Requirements Review and Signoff<br />
8.2. Technical Training Phase<br />
8.3. System Design Phase<br />
8.3.1 Solution Design Workshop<br />
8.3.2 Solution Design Documentation<br />
8.3.3 Solution Design Review and Signoff<br />
8.4. Testing Phase<br />
8.4.1 Training for the users<br />
8.4.2 Functional Training & Testing Phases<br />
8.4.3 Signoff of Testing Phase<br />
8.5. Data Migration<br />
Signoff from all the stake holders must be obtained before proceeding from one phase to the other.<br />
8.1 Requirement gathering phase<br />
8.1.1 Requirement workshops<br />
This phase involves requirement gathering. (Parr A, and Shanks G., 2000: 9) This is one the very<br />
important phases of ERP implementation process. All the functional head(s) should be a part of this<br />
phase. Inputs from all the stake holders must be taken. GAP Analysis will be identified during this phase<br />
as well. Organization should be flexible enough to change the business process in order to avoid too<br />
much customization in the system as excessive customization makes the system unreliable. Also, adding<br />
of new patches or updates may result in creating problem(s) for the organization. Roles and rights of the<br />
user(s) should also be decided during this phase.<br />
8.1.2 Requirement documentation<br />
In this phase all the requirements based on requirement workshop should be documented so that it can<br />
ensure that all the stake holders are on the same page. Since the GAP Analysis is also performed based<br />
on the requirement, stakeholders will clearly understand what they will get in the new system.<br />
8.1.3 Requirement review and signoff<br />
All the stake holders will review the document and signoff. It is very important to involve all the functional<br />
users to be part of this process, so that they can have a feeling of ownership of the system.<br />
8.2 Technical training<br />
This phase is very critical and it has to be performed for the ICT (Information and Communication<br />
Technology) Department. This phase will provide technical training to them which will enable them to give<br />
42
Imran Batada and Asmita Rahman<br />
their maximum input during the design and testing phases. The lower the dependency on the vendor, the<br />
more successful the system would be.<br />
8.3 System design phase<br />
8.3.1 Solution design workshop<br />
Screen shots or prototypes of the system should be created and showed to the functional users based on<br />
requirement documentation. The prototype would help the non-technical functional users to understand<br />
the functioning of the system as the snapshots would give a preview of the actual system to be<br />
implemented. It will also help them understand their input parameters for all the forms and reports. They<br />
will also see the validation of the forms. During this phase all the changes as per the design can be<br />
accommodated.<br />
8.3.2 Solution design documentation<br />
Based on the design workshop, a detailed documentation should be made for the system design<br />
workshop. All the changes should be accommodated as per the requirement of the functional owners.<br />
8.3.3 Solution design review and signoff<br />
Once all the changes in the document have been incorporated, all the functional heads should sign the<br />
document.<br />
8.4 Testing phase<br />
8.4.1 Training for the users<br />
Proper training sessions should be arranged in this phase so that the users can obtain a better<br />
understanding of the system. This will aid the testing of the system.<br />
8.4.2 Functional training and testing phase<br />
Testing phase is entirely dependent on the training of the system. If proper training sessions have been<br />
conducted not only for functional heads but also for functional user(s), they will be able to properly test<br />
the system based on the requirements and design phase. Users can test the system with test data during<br />
this phase. Proper test cases should be developed to thoroughly check the system during this phase. ICT<br />
Department should be capable to thoroughly check the system along with all the users.<br />
8.4.3 Signoff of testing phase<br />
Once the system has been thoroughly tested, all the functional users need to signoff.<br />
8.5 Data migration<br />
After successfully completing the testing phase, it is time to migrate data from the legacy system to the<br />
new System. Deadlines (Cut-off times) should be established. All the data should be validated before it is<br />
made live. It is recommended to run the legacy system in parallel for one financial year. This helps verify<br />
proper functioning of the new system.<br />
9. Final rollout<br />
In this phase the newly developed system is made live and all the users in the organization are directed<br />
to start using the new system. Again, it is recommended to keep the previous traditional system running<br />
in parallel for some time to ensure proper functioning. It can also serve as a back-up.<br />
10. Training<br />
Training phase is a continuous phase for all the organization’s users. (Scott J., 2006:11). The more<br />
involved users are in this phase, the more comfortable they will feel in using the new system. Resistance<br />
of the staff will gradually decrease as they learn to understand and use the new system.<br />
43
11. Post production of an ERP system<br />
Imran Batada and Asmita Rahman<br />
This is the crucial phase. It is important to have vendor’s technical support during this phase. A vendor<br />
resource available at the premises is a good, however, costly solution. Ideally, the in-house team should<br />
be able to handle the issues, in case vendor’s support is limited or not available. One must identify the<br />
post production cost before making the final decision as it may get very expensive. (Wei C, 2008:12). It is<br />
recommended to have at least one (1) year post production support in order to fully use the system to its<br />
fullest capabilities.<br />
12. Conclusion<br />
Selection of ERP is the major decision for the organization. It is a proven fact that a wrong decision in<br />
selection will adversely effect in the overall performance of the organization. Involvement of the<br />
stakeholders for selecting the product and using the right method for evaluating is the correct path for<br />
successful implementation of the ERP System. AHP Methodology is a recommended technique for<br />
complex decision making. AHP Methodology permits evaluator to evaluate priorities of alternative on the<br />
basis of performance of each condition (Teltumbdey A., 2000: 4).<br />
Selection of the right ERP System using the right methodology makes the whole implementation process<br />
easier. Implementation of ERP System in the university or in any organization who are in the business for<br />
more than 50 years is itself a tough task. It is recommended to re-engineer the process instead of doing<br />
major customization in the system.<br />
In this study nine major criteria and the sub criteria have been identified for evaluating the ERP System.<br />
University used AHP Model. First they developed the weight for the criteria, then they developed the<br />
rating for each decision alternative for each criteria and finally they have calculated the weighted average<br />
rating for each decision alternative (Laia V, Truebloodb R, and Wong B. , 1999: 8).<br />
Implementation of ERP not only involves money, but also requires time, effort and dedication from all the<br />
stake holders. If the stakeholders are not part of the decision making process then they will not feel the<br />
ownership of the new system. Therefore, it is highly recommended to involve them from the beginning.<br />
Training of ICT department is also very necessary, they should be technically sound as eventually they<br />
will be responsible to train the non-technical user(s).<br />
References<br />
Grossman T, and Walsh J. (2004) ‘Avoiding the pitgalls of ERP system implementation’, Information Systems<br />
Management<br />
Kim J. (2009) ‘Activity-based framework for cost savings through the implementation of an ERP system’, International<br />
Journal of Production Research<br />
Kunz J. (2010) 'The Analytic Hierarchy Process (AHP)', Eagle City Hall Location Options Task<br />
Laia V, Truebloodb R, and Wong B. (1999) 'Case Study: Software selection: a case study of the application of the<br />
analytical hierarchical process to the selection of a multimedia authoring system', Information & Management<br />
Lall V, and Teyarachakul, S. (2006) ‘Enterprise Resource Planning (ERP) System selection: A data envelopment<br />
analysis (DEA) Approach’, Journal of Computer Information Systems.<br />
Parr A, and Shanks G. (2000) ‘A model of ERP project implementation’, Journal of Information Technology<br />
Popular RFP Templates (2010), Discrete Enterprise Resource Planning (Discrete ERP) RFP Template, [Online]<br />
Available: http://rfp.technologyevaluation.com/store.asp [November 10, 2010]<br />
Quiscenti M, Bruccoleri M , La Commare U, Noto La Diega S, and Perrone G. (2006) ‘Business process-oriented<br />
design of Enterprise Resource Planning (ERP) systems for small and medium enterprises’, International Journal<br />
of Production Research<br />
Shapira A, and Simcha M. (2009) 'AHP-Based Weighting of Factors Affecting Safety on Construction Sites with<br />
Tower Cranes', Journal of Construction Engineering and Management<br />
Teltumbdey A.(2000) 'A framework for evaluating ERP projects', Int. J. Prod. Res.<br />
Wei C, Chein C, and Wang M. (2005) 'An AHP-based approach to ERP system selection', International Journal of<br />
Production Economics<br />
Scott J. (2006) ‘Post- Implementation Usability of ERP Training manuals: The user’s perspective’, Information<br />
Systems Management<br />
Wei C, (2008) ‘Evaluating the performance of an ERP system based on the knowledge of ERP implementation<br />
objectives’, Int J Adv Manuf Technology<br />
Zhao Y, and Fan S. (2007) ‘Implementation approach of ERP with mass customization’, International Journal of<br />
Computer Integrated Manufacturing<br />
44
Evolution of eProcurement Model in the Italian Public Sector:<br />
From Government Centralisation to Regional Delocalisation<br />
Clara Benevolo and Renata Paola Dameri<br />
University of Genoa, Italy<br />
benevolo@economia.unige.it<br />
dameri@economia.unige.it<br />
Abstract: In this work the authors present the results of a survey regarding public eProcurement in Italy. The Italian<br />
Government has been starting to drive the introduction of eProcurement practices in the public sector since the end<br />
of the twentieth century. Its aim was especially to rationalise the public spending by the benefits offered by<br />
eProcurement, that is: more competition between the suppliers, more transparency in carrying on the tenders and<br />
auctions, lower prices thanks to higher volumes, speeding up of purchasing process. To quickly reach these goals,<br />
the Government opted for a centralised solutions. At that time, it was the best one to support the implementation of a<br />
new technology and to reach synergies and scale economies. In a second and more mature phase, the Government<br />
decided to encourage a delocalisation of public eProcurement, especially involving the twenty Italian Regions to<br />
implement their own eProcurement platform. The aim was to pursue not only quantitative benefits, but also<br />
qualitative ones, especially regarding a better alignment between the amount and nature of public spending and the<br />
real needs of citizens. The second phase is in progress yet, but surveying the actual implementation of<br />
eProcurement all over the twenty Italian Regions some results are already evident. The authors use their research on<br />
the Italian case to explore potentials and benefits, barriers and difficulties of public eProcurement implementation in<br />
complex administrative organisations, such as the Italian State, with very high level of public spending. Empirical<br />
evidences and lessons learned are explored, to define a comprehensive interpretative model of eProcurement in the<br />
public sector.<br />
Keywords: eProcurement, public administration, decentralisation, business case study<br />
1. Introduction<br />
During the latest ten years, the diffusion of eProcurement has been increasing both in private and in<br />
public sector. Governments and local public agencies consider eProcurement as a powerful instrument,<br />
especially to control the public spending, to gain efficiency and save money. However, even if<br />
eProcurement is based on a quite simple technology, it is not so easy to implement, because it affects<br />
several aspects, such as: IT infrastructure, software implementation, business transactions design,<br />
process reengineering, people behaviours and the relationship between costs and benefits. Moreover,<br />
several public agencies implemented eProcurement without any methodological approach, without<br />
applying nor theoretical models nor best practices. For these reasons, both successes and failures can<br />
be obtained, but it is not clear which critical factors influence the reached results.<br />
In Italy, public eProcurement implementation can be divided in two phases.<br />
In the first one, the Government implemented eProcurement solutions, especially aiming to reduce public<br />
expenses and looking at granting affordable and secure transactions. EProcurement was a centralised<br />
practice, supported by the strong commitment by the State able to develop this innovative instrument. But<br />
the centralised practice forced local public agencies to apply the same model, not ever suitable with their<br />
specificity.<br />
In the second phase, eProcurement was shifted from central to local level, involving Regional and other<br />
local public agencies to implement their own eProcurement applications, choosing both technical<br />
instruments and organisational solutions. The main aim was to gain efficiency and in the meantime to<br />
create a better fit between eProcurement applications, organizational solutions and local needs.<br />
The Italian case is interesting for several aspects: first, because in Italy the public expense is very high<br />
and it is crucial to improve the control on public spending and the quality of purchasing; second, because<br />
the territorial organisation of public bodies in Italy is complex and structured on different levels; third,<br />
owing to the evolution of the general framework of eProcurement implementation.<br />
In this paper, we examine the evolution of the Italian public eProcurement model, from the central to the<br />
regional one, and the key lessons learned in both the implemented experiences. We identify both<br />
reached goals and implementation problems, both critical success factors and winning choices, in order<br />
to design an implementation framework able to be applied to similar public agencies in other Countries.<br />
45
2. Literature review<br />
Clara Benevolo and Renata Paola Dameri<br />
EProcurement is an important topic, emerging in the latest ten years, both in companies and in the public<br />
sector. It regards the use of Internet potential to redesign business process and to improve operations<br />
quality. The focus of eProcurement has ever been double:<br />
On one side, eProcurement is seen like an instrument to control purchases, to reduce cost, to save<br />
money;<br />
On the other side, eProcurement is seen like a driver for business change, process quality<br />
improvement and more effective management (Ash and Burn 2003; Dooley and Purchase 2006).<br />
Also in the public sector, eProcurement is considered an important weapon to reach several goals<br />
(Croom and Brandon-Jones 2005):<br />
A better control on public purchases and spending, cost reduction and more effective use of public<br />
resources;<br />
The improvement of public administrative processes, to reduce bureaucracy and to increase<br />
transparency in the management of public agencies;<br />
The increasing of the public programs effectiveness, better focusing choices and expenses on the<br />
more important needs for citizens.<br />
However, the eProcurement implementation in the public sector should face several obstacles:<br />
organizational and cultural barriers, such as difficulties in evaluating the eProcurement benefits.<br />
On the organizational aspect, to implement eProcurement practices requires not only to create<br />
effective and appropriate web sites (Bruno et al. 2005) nor to implement a technological platform to<br />
manage tenders, agreements and eMarketplace, but it requires to rethink the administrative activity in<br />
public administration, that is to redesign all the activities included in public procurement and<br />
introducing eProcurement solutions in a redefined process. Without the organizational redesign,<br />
eProcurement perhaps can produce economic benefits, but it is not possible to reach the other (and<br />
perhaps more important) goals: administrative process quality, transparency, suppliers competition,<br />
public choices effectiveness (Reddick 2004).<br />
On the cultural aspect, eProcurement needs not only ICT competences, both regarding the use of<br />
web sites and in data processing, but also to change the administrative practices and to accept to<br />
modify their own habits. On the contrary, suspicion and resistance on implementation are the<br />
stronger barriers to successful and profitable eProcurement implementation in the public sector<br />
(Vaidya, Sajeev and Calender 2006; Wirtz, Sebastian and Schierz 2010).<br />
On the evaluation aspect, eProcurement benefits are relatively easy to evaluate, if limited to cost<br />
savings, but more difficult to measure if regarding process quality, transparency, effectiveness of<br />
public actions (Singer et al. 2009). The lack of a well conceived framework to evaluate also the<br />
intangible and qualitative benefits of eProcurement not only prevents to understand the real impact of<br />
this business practice on the efficiency and effectiveness of public purchases and spending, but also<br />
delays the improvement of administrative processes to integrate eProcurement (Rendon 2008).<br />
Despite barriers and difficulties, eProcurement is spread in several public administrations, all over the<br />
world (Moon 2005). Also in Italy, eProcurement practices and solutions have been adopted since the end<br />
of XX century. During the latest ten years, eProcurement was driven along with two paths:<br />
To enlarge its adoption, increasing both the number of public agencies using eProcurement and the<br />
amount of transactions using this channel;<br />
To improve the quality of eProcurement, by better ICT solutions, people training, implementation<br />
frameworks able to integrate eProcurement in more effective administrative processes and<br />
operations (Polimi 2009).<br />
Also in Italy, eProcurement faced the same difficulties and barriers noted in the international environment<br />
(Carpineti, Piga and Zanza 2006; Somasundaram 2004). Better implementation frameworks and<br />
knowledge sharing on best practices in eProcurement implementation in the public sector could help to<br />
assess the present and potential benefits of this technology. Governments and local administrative<br />
agencies could improve gains deriving from their eProcurement solutions, realising better returns on<br />
these technological and organisational investments.<br />
46
Clara Benevolo and Renata Paola Dameri<br />
3. eProcurement in the Italian public sector: The Government centralisation<br />
Starting from the end of Nineties, the Italian Government has been starting an innovation process, aiming<br />
to introduce eProcurement in the public sector. The main goals of this innovation were to reduce<br />
expenses, to increase transparency of purchasing and to improve efficiency of administrative processes<br />
in the Public Administration. In this first phase of eProcurement implementation in the public sector, the<br />
role of the central Government was crucial: indeed, only the State was able to support this innovative<br />
practice, helping all the public agencies to introduce this new technology into their administrative<br />
activities.<br />
Moreover, the Italian Government grounded its own implementation strategy on centralisation, to reach<br />
both the best efficiency and the best effectiveness of the innovation process, minimising in the meantime<br />
the cost of the organizational and technological solutions. The Central Public Administration defined all<br />
the phases of purchasing process and the operations to be automated, and also the software applications<br />
to be used for this aim.<br />
During the first phase of its implementation, three aspects are important to understand the evolution of<br />
eProcurement in the Italian public sector:<br />
The need of laws and rules to govern this practice;<br />
The public body that drove the implementation activity;<br />
The methods and instruments used to implement eProcurement.<br />
EProcurement implementation in the pubic sector needed for the first to define laws and rules to be<br />
applied to public contracts, especially regarding:<br />
The award of contracts concluded on behalf of the State, regional or local authorities and other<br />
bodies governed by public law entities;<br />
The opening-up of public procurement to competition;<br />
The verification of the suitability of tenderers, in open procedures, and of candidates, in restricted and<br />
negotiated procedures;<br />
The electronic signature.<br />
The successfully implementation of eProcurement in public sector depends not only on electronic<br />
solutions, but also on the affordability and compliance of contracting procedures. In the first phase of<br />
public eProcurement implementation, the central Government played the crucial role of regulator, both<br />
issuing its own laws and implementing <strong>European</strong> Directives. At present, all the rules regarding electronic<br />
activities of Italian Public Administration are collected in the “Codice dell’Amministrazione Digitale (CAD)”,<br />
Digital Administration Code. It is a set of rules governing the use of Information Technology in all the<br />
relations between the Italian Public Administration and the citizens (D.Lgs. 07.03.2005, n° 82). In Europe,<br />
eProcurement is ruled by Directive 2004/18/EC of the <strong>European</strong> Parliament and of the Council of 31<br />
March 2004 on the coordination of procedures for the award of public works contracts, public supply<br />
contracts and public service contracts.<br />
By these means, the Italian State defined its own eProcurement reference model (Alsac 2007), including<br />
rules regarding: the administrative process to be applied to award a public contract by electronic<br />
technologies, the affordability of economic operators, the public contracts transparency and the digital<br />
signature (Figure 1).<br />
The reference model includes all the steps and activities needed to drive to the final signed contract. That<br />
is:<br />
1. The assessment of economic operators to be admitted to electronic tenders, auctions and markets,<br />
by checking their affordability and compliance to the law regarding commercial relations with public<br />
agencies;<br />
2. The general framework of laws and rules to be applied to carry out all the procurement phases,<br />
complying with all the qualifications of a public tender such as transparency, accountability and so on;<br />
3. The procedure to be followed for a regular award of a valid public contract;<br />
4. The digital signature, for validating the contract.<br />
47
Clara Benevolo and Renata Paola Dameri<br />
Figure 1: The Italian public eProcurement reference model<br />
To better support the launch of eProcurement and its first implementation, the Italian Government choose<br />
a centralised model and set up a public agency, called Consip, to concentrate all the activities regarding<br />
public eProcurement. Consip SpA is a company belonging to the Italian Finance Office; its mission is to<br />
support e-Government initiatives with particular reference to eProcurement, with the aim to rationalise<br />
public purchasing and improve contracts transparency. Consip has in charge the design and<br />
development of new e-government processes and the related ICT solutions. Its main work consists in<br />
planning the best ideas to implement the strategic directives of Italian Finance Office, acquiring on the<br />
market the best technical solutions and competences to traduce these ideas in ICT applications. This<br />
innovative model permits to obtain two best results: to plan organisational processes for the public sector<br />
well aligned with the Government strategic aims, and to translate projects into practice applying the best<br />
technical solutions on the market.<br />
To implement public eProcurement, Consip uses three main instruments:<br />
Framework agreements;<br />
Electronic auctions and tenders;<br />
eMarketplace.<br />
By framework agreements, Consip awards tenders regarding the purchase of large amount of a well<br />
defined good or service by all the public bodies, with the same prices and conditions drew up in the<br />
contract. Consip is therefore responsible for awarding public contracts or framework agreements for all<br />
the public bodies. In view of the large volumes purchased, this technique helps to increase competition<br />
and streamlines public purchasing. Benefits are important especially for small public agencies, that<br />
purchase small volumes and could access only to local markets with a lower degree of competition.<br />
Moreover, all the public bodies using framework agreements could save the cost to carry out the<br />
agreement by themselves.<br />
By electronic auctions and tenders, a public agency could award a public contract using the technical<br />
platform provided by Consip. This platform complies with the rules drawn up by the Italian laws and<br />
<strong>European</strong> directives about equal treatment and transparency in public purchasing. Therefore the use of<br />
this platform grants the affordability of the technological instruments and prevents public bodies to<br />
develop their own ICT solutions, saving time and money.<br />
The eMarketplace realised by Consip is called MEPA (Mercato Elettronico della Pubblica<br />
Amministrazione/Public Administration Electronic Market). MEPA is a comprehensive and integrate<br />
platform able to manage all the stages of the public procurement: the tenderers selection and ranking, the<br />
48
Clara Benevolo and Renata Paola Dameri<br />
management of tenderers electronic catalogues, the request of quotation or request for order, the final<br />
contract award. On this eMarketplace, public bodies could meet several suppliers and tenderers<br />
previously qualified and therefore compliant with the rules about public purchasing; they can:<br />
Access at a large market and save money thanks to the higher competition between suppliers and<br />
tenderers,<br />
Streamline their procurement activities<br />
Save time thanks to the electronic platform.<br />
By means of the centralisation of public eProcurement by Consip, the Italian Government reached good<br />
goals in a short time. The technological platform for eProcurement is available for all the Italian public<br />
bodies, both small and big ones. This platform permitted to spread the eProcurement practices among all<br />
the public bodies unifying methods and procedures, sparing them the efforts and the expenses related to<br />
the implementation of their own eProcurement application and granting them the affordability and<br />
compliance of the Consip solution. During the period 2004-2010, eProcurement in the Italian public sector<br />
have been increasing from 76 million € to 6000 million €. Among these, about 2286 million € derive from<br />
transactions by Consip (Figure 2).<br />
Figure 2: The Italian public eProcurement volumes (source: Polimi 2010)<br />
Also qualitative benefits have been obtained. The most important benefits regard the process efficiency,<br />
that is, process streamlining, price savings and procurement time reduction. Also the quality of<br />
purchasing has been increasing, especially thanks to the larger number of tenderers and therefore the<br />
higher competition in the market of public procurement (Polimi 2010). Other obtained benefits regard<br />
documents dematerialisation and savings in human resources.<br />
However, the centralised model could not be the best solution for the Italian public eProcurement<br />
practice. Indeed, Italian public bodies and agencies are very heterogeneous and spread throughout the<br />
country. The centralised solution showed several difficulties in involving the smaller bodies and local<br />
administrations. The centralised platform is too generic to respond to the needs of a so large set of<br />
entities, and it is not suitable to meet the expectations and schedule of all the different public bodies, with<br />
their own characteristics and specific needs. Moreover, a centralised solution tends to incorporate best<br />
practices and to maintain them along the time, slowing down the innovation and the research of a better<br />
quality of eProcurement to positively impact on the effectiveness of public services to the citizens. Consip<br />
eProcurement centralised solution was certainly the best answer for the first time adoption, but it is too<br />
much focused on efficiency and it doesn’t take into consideration other aspects.<br />
For all these reasons, a second phase was started, to delocalise eProcurement at regional level and to<br />
drive public agencies to develop their own eProcurement solutions, in order to better reach specific goals<br />
related to their own needs.<br />
4. eProcurement in the Italian public sector: The regional delocalisation<br />
In 2007, the Italian Government issued a law (L. 27 December 2006, n. 296) to rule the adoption of<br />
electronic procurement solutions at the regional level. The Italian State is organised in twenty Regions;<br />
they have administrative autonomy and manage the Regional health system. The aim of this second<br />
49
Clara Benevolo and Renata Paola Dameri<br />
phase of public eProcurement was mainly to overcome the scarce involvement of small and local public<br />
agencies registered in the first phase.<br />
All the Regions, like Consip, have the role of contracting authorities, that is, they can select and authorise<br />
suppliers and award public contracts and framework agreements for all the public bodies and agencies in<br />
their own territory, included the health institutions, compelled to join the agreements.<br />
During the period 2007-2008, the regional contracting authorities carried out 2852 tenders for a total<br />
amount of about 9,7 billions €. 87% of tenders regard goods and services purchasing in the health sector,<br />
mainly in the north and center Italian Regions.<br />
This new phase is today in progress yet. However, it is possible to start to assess this delocalised model<br />
to understand if it is really better than the centralised one. Several aspects should be taken into<br />
consideration:<br />
The aims of this second phase of public eProcurement;<br />
The role of the Regions;<br />
The role played by the State or by another central agency, to govern and coordinate the<br />
delocalisation process.<br />
Unfortunately, the aims and goals of this second phase of eProcurement are not so different respect to<br />
the first phase, and it prevents the qualitative elevation of eProcurement projects. More efficiency of<br />
purchasing, more effectiveness of procurement process, streamlining of administrative activities and<br />
transparency are the main goals, also in this second step. Only two aspects are especially linked with the<br />
delocalisation: the larger participation of both small public bodies and local suppliers. That is, the<br />
enlargement of eProcurement volumes and of the relating benefits. The main goal to reach is ever to<br />
reduce the public spending, and eProcurement is certainly one winning instrument to be applied.<br />
However, the delocalisation has not be used to improve also the quality of public purchasing and the<br />
appropriateness of procurement practices. Also regional eProcurement is too much focused on cost<br />
reduction and administrative streamlining and the importance of the quality of purchasing for the quality of<br />
public services to the citizens is not taken into account.<br />
The role of Regions is the fulcrum of this second phase of public eProcurement. They could play a central<br />
role, especially in:<br />
Spreading the culture of eProcurement on their territory, involving also the smaller public bodies;<br />
Supporting small bodies in implementing eProcurement practices;<br />
Developing specific solutions to support particular needs in their environment;<br />
Studying the impact of eProcurement on the quality of public services delivered to the citizens,<br />
considering also the specific needs of their own area.<br />
At this moment, no surveys have been made, to assess the commitment of Regions and to verify if some<br />
of the roles listed above have been really performed. However, Regions are very different each others<br />
and the implementation of regional eProcurement platforms has been very heterogeneous. In Figure 3<br />
we can see that:<br />
Some Regions implemented eProcurement solutions able to support procurement for all the public<br />
bodies in their own area (the black ones);<br />
Some Regions implemented eProcurement solutions only to support purchasing in the health sector<br />
(the grey ones);<br />
Some Regions are developing during this period eProcurement solutions or have no projects in mind<br />
(the white ones).<br />
It means that each Region started its own eProcurement process, generally without working in<br />
cooperation with other Regions and choosing different ways to implement eProcurement (Avcp 2010).<br />
The further analysis of each regional solution shows that they are very different on behalf on: type of<br />
agency playing the role of contracting agency, type of implemented instruments, degree of participation of<br />
both public bodies and suppliers or tenderers, savings or other benefits obtained by regional<br />
eProcurement. The lack of an evaluation framework to be applied to public eProcurement, based on both<br />
50
Clara Benevolo and Renata Paola Dameri<br />
quantitative and qualitative parameters, prevent to better understand the effectiveness of different<br />
solutions.<br />
realised<br />
work in progress<br />
only for health<br />
Figure 3: The implementation of regional eProcurement<br />
Certainly, the delocalisation phase has partially reached its own goals and it helps to control and to<br />
reduce the public spending, especially in the health sector. More bodies, suppliers and tenderers have<br />
been involved and the competitiveness has increased. Also the administrative processes have been<br />
simplified and sped up.<br />
However, the second phase could be better based on the successes and failures of the first one, to<br />
overcome some problems and especially to aim at more important and large goals, considering also the<br />
quality of public purchasing and not only on the quantity of saved money. The delocalisation is the best<br />
weapon to reach a better alignment between local needs of the citizens, adequate public services and<br />
related purchasing. But to reach these goals two thinks are necessary: the clear definition of objectives to<br />
reach and an evaluation framework to assess if and in which measure the goals have really be reached.<br />
Moreover, in the second phase a central direction is lacked, to coordinate regional initiatives but<br />
especially to share knowledge, competences, best practices and lessons learned. For this reason the<br />
Italian State created a network between Consip and the regional purchasing authorities, aiming to<br />
rationalise the public spending and to create positive synergies in the implementation and use of ICT for<br />
public eProcurement.<br />
Perhaps a knowledge base on public eProcurement could help to better implement effective solutions,<br />
avoiding to repeat errors made by others. The comprehensive analysis of all the regional solutions, of<br />
their characteristics and of their specific goals is useful also to evaluate the trade off between<br />
centralisation and delocalisation. Indeed, to organise eProcurement at the regional level on one side<br />
improves the appropriateness of solutions and the involvement of local bodies; on the other side, it<br />
reduces the volumes of purchasing and therefore the economy of scale deriving from the centralised<br />
procurement. Only having well clear the desired results and behaviours it is possible to choice between<br />
the higher efficiency of the centralised model and the better specificity and adequacy of the delocalised<br />
one.<br />
5. Conclusions<br />
In Italy, public bodies are about 11.000. Among them, about 50% use eProcurement, but only a few<br />
hundreds are mature and intensive users (Polimi 2010). The more used instruments are eMarketplaces<br />
and online stores, auctions and framework agreements are less used. The savings are estimated<br />
between 10 and 50% of the electronic spending.<br />
Therefore, eProcurement in Italy could be considered a successful initiative, especially regarding the<br />
amount and volume of public purchasing, but it is not enough spread. The main role is played by Consip,<br />
that carries out 50% of purchasing; the prevalent model is hence the centralised one. The delocalised<br />
model have been starting, but with very heterogeneous results. Several regional purchasing authorities<br />
51
Clara Benevolo and Renata Paola Dameri<br />
are developing and implementing eProcurement solutions, but only nine regional eProcurement platforms<br />
are really working and only four Regions appear mature in the use of eProcurement. Moreover,<br />
procurement processes have not been redesigned and benefits are gained only on efficiency and<br />
spending; no evidences are available, regarding the actual impact of eProcurement on the quality of<br />
public services for the citizens.<br />
Examining the Regional implementation of eProcurement, some evidences are useful, to plan<br />
improvements to be spread all over the State.<br />
Only 4% of purchasing is carried out on the electronic channel; this number could be increased,<br />
especially regarding electronic auctions and framework agreements.<br />
The commitment of local authorities is crucial for the success of eProcurement initiatives, such as the<br />
role of central and regional purchasing authorities in supporting and promoting the use of<br />
eProcurement by smaller public bodies.<br />
To improve the efficiency of e-purchasing, a more effective requirements planning is necessary; it<br />
could permit to optimise quantity and quality of purchasing. No evidences are available, of<br />
requirements planning practices and instruments to support eProcurement.<br />
No evaluation frameworks are available, to quantify savings in money and time and to award the best<br />
public bodies and managers in eProcurement implementation. Moreover, it is necessary to involve<br />
the public opinion in the assessment of eProcurement and its actual or potential impact on public<br />
services. EProcurement results should be measured, evaluated respect to politic and administrative<br />
goals of each public body and disclosed to citizens and communities using the most adequate media<br />
and instruments.<br />
The topic of public eProcurement is therefore actual and it promises future interesting returns, if<br />
adequately studied in depth. Further researches could be useful, especially regarding the following<br />
aspects.<br />
To analyse some best cases, regarding public eProcurement in the Italian Regions. We discovered<br />
two or three successful platforms supported by best practices in process redesign, requirements<br />
planning and involvement of smaller public bodies. The case studies could be the basis for defining a<br />
general framework for public eProcurement implementation in local purchasing authorities.<br />
To better analyse the actual impact of public eProcurement on the efficiency of purchasing activities<br />
and practices, and on savings deriving from lower prices and higher competition among suppliers and<br />
tenderers. Till now, such surveys have been conducted only using panels and interviews and<br />
qualitative indicators. No affordable quantitative metrics are available. Moreover, the survey could<br />
better detailed the research, distinguishing between different techniques used for eProcurement<br />
(auctions, tenders, marketplaces, e-stores) to find relations between techniques and benefits, to<br />
answer to different purchasing requirements.<br />
To design a comprehensive evaluation method, able to assess not only the savings in time, money<br />
and human resources deriving from the process streamlining, but also to evaluate the impact of best<br />
eProcurement practices on the quality of public services for citizens; such an evaluation framework<br />
should include also a disclosure framework to involve communities in the assessment of<br />
eProcurement initiatives and their impact on the amount and quality of public spending.<br />
References<br />
Alsac, U. (2007), “Use of eProcurement in Turkey’s public health sector”, Journal of Public Procurement, Vol. 7, No.<br />
3, pp. 333-361.<br />
Ash, C. and Burn, J. (2003) “Assessing the benefits from e-business transformation through effective enterprise<br />
management”, Journal of Information Systems, 12, pp. 297-308.<br />
Avcp (2010) (Autorità per la Vigilanza sui Contratti Pubblici di Lavori, Servizi e Forniture) Censimento ed analisi<br />
dell’attività contrattuale svolta nel biennio 2007-08 dalle Centrali di Committenza Regionali e verifica dello stato<br />
di attuazione del sistema a rete, [online], www.avcp.it.<br />
Bruno, G., Esposito, E., Mastroianni, M. and Vellutino, D. (2005) “Analysis of public eProcurement web site<br />
accessibility”, Journal of Public Procurement, Vol. 5, No. 3, pp. 344-366.<br />
Cabras, I. (2010) “Use of EProcurement in Local Authorities' Purchasing and Its Effects on Local Economies:<br />
Evidence from Cumbria, UK”, <strong>European</strong> Planning Studies, Vol. 18, No. 7, p. 1133.<br />
Carpineti, L., Piga, G. and Zanza, M. (2006), The Variety of Procurement Practice: Evidence from Public<br />
Procurement, in Dimitri, N., Piga, G. and Spagnolo, G., eds., Handbook of Procurement, Cambridge University<br />
Press.<br />
52
Clara Benevolo and Renata Paola Dameri<br />
Croom, S. and Brandon-Jones, A. (2005) “Key issues in eProcurement: procurement implementation and operation<br />
in the public sector”, Journal of Public Procurement, Vol. 5, No. 3, pp. 367-387.<br />
Dooley, K. and Purchase, S. (2006) “Factors influencing eProcurement usage”, Journal of Public Procurement, Vol.<br />
6, No. 1/2, pp. 28-45.<br />
Hommen, L. and Rolfstam, M. (2009) “Public procurement and innovation: towards a taxonomy”, Journal of Public<br />
Procurement, Vol. 9, No. 1, pp. 17-56.<br />
MacManus, S. (2002) “Understanding the incremental nature of eProcurement implementation at the state and local<br />
level”, Journal of Public Procurement, Vol. 2, No. 1, pp. 5-28.<br />
Mishra, A., Konana, P. and Barua A. (2007) “Antecedents and Consequences of Internet Use in Procurement: An<br />
Empirical Investigation of U.S. Manufacturing Firms”, Information Systems Research, Vol. 18, No. 1, pp. 103-<br />
122.<br />
Moon, J. (2005) “EProcurement management in state governments: diffusion of eProcurement practices and its<br />
determinants”, Journal of Public Procurement, Vol. 5, No. 1, pp. 54-72.<br />
Polimi (2009) Innovazione ed efficienza nella Pubblica Amministrazione: il ruolo dell’eProcurement, Osservatorio<br />
eProcurement nella Pubblica Amministrazione, Politecnico di Milano – Dipartimento di Ingegneria Gestionale.<br />
[online], www.osservatori.net.<br />
Polimi (2010) eProcurement: una concreta via verso innovazione, efficienza e trasparenza per la PA italiana,<br />
Osservatorio eProcurement nella Pubblica Amministrazione, Politecnico di Milano – Dipartimento di Ingegneria<br />
Gestionale. [online], www.osservatori.net.<br />
Reddick, C. (2004) “The growth of eProcurement in American state governments: a model and empirical evidence”,<br />
Journal of Public Procurement, Vol. 4, No. 2, pp. 151-176.<br />
Rendon, R. (2008) “Procurement process maturity: key to performance measurement”, Journal of Public<br />
Procurement, Vol. 8, No. 2, pp. 200-214.<br />
Singer, M., Konstantinidis, G., Roubik, E. and Beffermann, E. (2009) ”Does eProcurement save the state money?“,<br />
Journal of Public Procurement, Vol. 9, No. 1, pp. 58-78.<br />
Somasundaram, R. (2004) "Diffusion of eProcurement in the Public Sector: Revisiting Centralized vs. Decentralized<br />
Debates As a Twist in the Tale", Proceedings of the 13 th <strong>European</strong> <strong>Conference</strong> on Information Systems, Turku,<br />
Finland, June.<br />
Tavi, J. (2008) “Learning From Global World-Class eProcurement Practices”, Strategic Finance, Apr., Vol. 89, No.<br />
10, pp. 24-30.<br />
Vaidya, K, Sajeev, A. and Callender, G. (2006) “Critical factors that influence eProcurement implementation success<br />
in the public sector”, Journal of Public Procurement, Vol. 6, No. 1/2, pp. 70-99.<br />
Walker, H. and Harland, C. (2008) “EProcurement in the United Nations: influences, issues and impact”, International<br />
Journal of Operations & Production Management, Vol. 28, No. 9, pp. 831-857.<br />
Wirtz, B., Sebastian, L. and Schierz, P. (2010) ”An Empirical Analysis of the Acceptance of EProcurement in the<br />
German Public Sector“, International Journal of Public Administration, Vol. 33, No. 1, pp. 26-42.<br />
53
ICTs as Weapons of Mass Interaction - Motivations and<br />
Implications of Mediated Control<br />
Kofi Agyenim Boateng<br />
British Institue of Technology & E-commerce, London, UK<br />
boateng@bite.ac.uk<br />
fiboat@yahoo.com<br />
Abstract: Interaction has long been an integral aspect of organisation’s life. Hence, in recent times interactions<br />
driven by Information Communications Technologies (ICT) have gained significant acceptance and momentum in<br />
contemporary organisational settings. This is demonstrated by their enormous adoption and wide-ranging<br />
employment across the various levels of an organisation’s hierarchy. Consequently, businesses have started to<br />
expand their investment in, and broaden the frontiers of, technology-mediated interaction. This has meant the<br />
involvement of such communication devices as mobile phones, personal digital assistants (PDA), BlackBerries,<br />
laptops, as well as other forms of portable and immovable computing technologies to provide enduring support for<br />
both enterprise-wide and external interactions. This study adds to the existing conceptualisations of technology<br />
mediated communication by delving into the fundamental rationalities, manifestations and implications of mediated<br />
control. To this end, an exploratory study is conducted in a moderately large UK food company that distributes semiprepared<br />
food to small restaurants, and examines the applications of both mobile and stationary technology tools for<br />
undertaking different enterprise-wide communication endeavours under dynamic work strategising commitments.<br />
The study reveals that first, the application of multiplicity of ICT-driven interaction tools generate diverse social<br />
implications, both pleasant and unpleasant to the business organisation. Second, multiple actors have different<br />
motivations for the application of varied technology tools in the accomplishment of a given organisational tasks,<br />
however similar such tasks may be. Finally, study as a result provides significant managerial and theoretical insights<br />
and implications for organisational interaction and the distribution and service delivery businesses in the briskly<br />
growing digital market.<br />
Keywords: ICT-driven interaction, social presence theory, media richness theory, emotion<br />
1. Prolegomenon<br />
The application of Information Communication Technologies (ICTs) in the many facets of organisational<br />
interaction is quite a recent phenomenon in spite of the fact interaction has always been an instrumental<br />
element of the life organisations. Business enterprises continue to expand their outlay in ICTs to reap the<br />
benefits afforded by technology mediated interaction in its varied breeds, such as web-driven media,<br />
mobile, fixed telephones, and short message service (SMS), among others (Bernett et al., 2002). An<br />
appealing aspect of technology mediated interaction is the flexibility and the vitality particularly associated<br />
with the many modes of its expression (Olson and Olson, 2000 p.141). A significant aspect of technology<br />
mediated interaction is the dynamism chiefly connected with the many avenues of its expression. Part of<br />
their flexible nature is demonstrated by their capacity of establishing interaction and exchange relations<br />
with others not collocated; what can conveniently be described as ‘virtual collaboration’(Olson and Olson,<br />
2000 p.141).<br />
Previous studies on technology mediated interactions have attended to the material features of the<br />
applied media (Daft and Lengel, 1986a), social considerations (Yoo and Alavi, 2001) reconstruction of<br />
interaction (Ngwenyama and Lee, 1997) and aspects of textual configurations (Orlikowski and Schultze,<br />
2004). While admitting that these studies made monumental impact on the mainstream discussion on<br />
ICT-driven interaction, they often overlook the emotional repercussions and the underlying psychological<br />
rationalities that come to the fore in such interactions. Still, the studies that intimately address the social<br />
elements of ICT interaction (Suchman, 1987) to tease out the emotional consequences that the<br />
interlocutors of ICT interaction experience. This paper wishes to broaden these investigations by having<br />
recourse to the theories of social presence and media richness and thereby advance the idea of how<br />
making sense of these ICT tools can diverge even across the people doing similar job within the same<br />
organisation.<br />
In particular, the paper advances the idea that ICT interaction can equally generate ‘e-ritual’ chains that<br />
are fundamental in getting agents undertake most social activities just as face to face interaction. We<br />
explore this concept by bringing into close scrutiny the limited perspective of Collins’ (2004) interaction<br />
ritual chains. This objective is achieved by embarking on a longitudinal study in a medium-size private<br />
company that specializes in supplying semi-prepared food to fast-food shops in and around London. The<br />
study seeks to address the following research queries:<br />
54
Kofi Agyenim Boateng<br />
To what extent can we better understand the rationalities and motivations that drive technology<br />
mediated interaction in collaborative environment?<br />
How does the use of ICT interaction tools affect the emotional make up of people?<br />
This paper resumes with a cursory look on the mainstream views on interaction with a touch on both<br />
media richness theory and social presence theory. The methodological approach to the study follows and<br />
the results from the field of investigation are immediately provided in that order. Eventually, the paper<br />
analyses the implications for both theoretical and practical standpoints and conclude by suggesting<br />
prospects for future areas of research.<br />
1.1 Interaction –contemporary considerations<br />
Interaction has witnessed many path-breaking studies and the masterpiece by Gofmann (1967) provides<br />
the foundation for many branches on the subject matter. Interaction has been conceptualised<br />
fundamentally in its mutual sense and therefore seeks to establish ‘reciprocal events that require at least<br />
two objects and two actions. Interaction occurs when these objects and events mutually influence one<br />
another’(Wagner, 1994 p. 8). The mutual aspect of interaction can take many possible forms such as<br />
collaboration, co-operation and coordination, all within the realm of communication. Interaction then,<br />
enables the establishment of networks for reaching out to contacts without necessarily their body<br />
presence (Peters, 1999) by means of the mediating capabilities of ICTs.<br />
ICT-driven interaction generates social relationships that form the basis of community integration<br />
(Calhoun, 1992) as various degrees of social relationships are likely to emerge from the ongoing<br />
exchanges. As such, human behaviour has impacted and informed the design, architecture and<br />
functionalities of these technologies. This mutual shaping between technology and human agency is<br />
observed elsewhere with the claim that ‘whatever effect technology may have on social forms is heavily<br />
shaped by the social, cultural and institutional relations in which the technology is embedded’ (Kallinikos,<br />
2007 pp. 278 - 279).<br />
The application and use of mediation tools in getting in touch with one’s interaction partner broadens the<br />
meaning and scope of interaction as the mediating capabilities of the technology. A capability that allows<br />
for ‘mutual tuning-in relationship’ (Schutz, 1971) based on mediated interaction, because it is predicated<br />
on shared dependency, common identity, mutual experience and joint context that is captured in a<br />
collective purpose directed toward common objectives (Schmidt and Bannon, 1992). This view suggests<br />
that interaction leads people to initiate strategic moves, develop routines, cultivate practices and<br />
procedures that give birth to organisations and provide the means for their vital sustenance (Hughes,<br />
1971). Within the spirit of symbolic interactionist (Blumer, 1969), the enduring nature of interaction does<br />
not only subsist in a context but it also constructs the context.<br />
It therefore comes as with a little surprise that Kakihara and Sørensen (2002) mainly link the building<br />
process of the phenomenon of knowledge (including an appreciation of the social backgrounds of the<br />
interactants) with the idea of interaction. The reason for this, the pair claim, is that the emergence of<br />
knowledge traces its sources from interaction. And this interaction forms the very rationale of complex<br />
exchanges inherently meshed in human society.<br />
Mediated or ICT-driven interaction recognises the use of technical elements to function as the platform on<br />
which interaction can be consummated. Thompson (1995 p. 83) recollects that technical medium like<br />
“paper, electrical wires, electromagnetic waves, etc.” facilitates the transmission of symbolic and<br />
information contents to parties who are either disconnected by time or space or occasionally both. The<br />
application of technical objects capable of setting two parties to engage in meaningful interaction<br />
underlies the concept of both media richness and social presence theories, the two of which produce<br />
different interaction effects for the interacting people. Nonetheless, mediated interaction affords the<br />
interactants very limited symbolic tokens for the reduction of uncertainty.<br />
1.2 Media richness and social presence theories – a case of mediated interaction<br />
Social presence holds the notion that mediated interaction has the potential to stimulate the features and<br />
benefits of face-to-face interaction Short, Williams et al (1976) and emphasises the extent to which one<br />
senses the immediate presence and significance of his or her interaction partner (Kayany et al., 1996).<br />
Social presence is crucial for building our knowledge and perception of other people’s behaviour and<br />
attitudes. Social presence is usually judged by the differences in meaning that borders on emotions or the<br />
55
Kofi Agyenim Boateng<br />
suitability of different media for a variety of similar organisational communication tasks (Argyle, 1969,<br />
Birdwhistle, 1970).<br />
The source of the media richness is directly connected to Daft and associates (1986b, 1987) when they<br />
made investigations into managers’ inclination to apply oral type of communication even when its<br />
mediated form is reasonable. The media richness theory suggests the ability of communicated<br />
information via a medium to limit the extent of equivocality on four counts, namely, the ‘speed of<br />
feedback, cue multiplicity, language variety, and personal focus’ (George and Carlson, 1999 p. 3). Key to<br />
media richness is the hypothesis that there is a disparity between jobs and specific characteristics;<br />
various media differ in the degree to which they can dispatch the kinds of information needed and the<br />
heterogeneous task characteristics.<br />
Consistent with the research questions outlined above, we conducted a field study to investigate how<br />
emotions and different psychological rationalities become evident in individuals making use of ICT<br />
interaction tools.<br />
2. Method<br />
The section addresses the ways and means of the whole research effort. The procedural method is<br />
fundamental for outlining the scheme for grounding the various parts and procedures of the entire study<br />
(Grunow, 1995). In short, the methodology goes to a greater length at laying down a sturdy and robust<br />
validity of the outcome of the research undertaking (Yin, 2003).<br />
The path to this methodological examination is followed with the aid of the three primary mechanics of<br />
human perception, namely, rationalism, empiricism and faith (Thieme, 2003). Rationalism identifies reality<br />
by means of reason. Empiricism does so through a support on the senses of sight, touch, taste, smell<br />
and sound. And faith defines reality via the confidence in the authority of another entity. Using all three<br />
systems of perception in the data-collection process would be beneficial in the sense that ‘rational proof<br />
and instruction do not fully exhaust the sphere of knowledge’ (Gadamer, 2004 p. 21). Therefore<br />
appropriating all three systems in a methodological process would ensure a demonstration of consistent<br />
and logical thinking thereby effectively limiting the researcher from going arbitrary with his judgements; so<br />
that the conclusions would not be treated ‘as a figment of my imagination’(Thieme, 2003 p. 2).<br />
To this end, the following data collection techniques were adopted for this study. The aim is to better<br />
appreciate phenomena-in-context inquiries against the background of certain professionals’ in their<br />
natural, operationally distinct but electronically connected environment and their cross-subjective<br />
predilections, this work mainly used qualitative data.<br />
2.1 Direct observation and interview<br />
The use of direct observation method allows the researcher to focus on the aspects of the research that<br />
are of particular interest to him without the possibility of being biased or influenced by ‘second-hand’<br />
information from the participants of the study. This necessarily occasioned long hours of sitting beside<br />
telesales reps, customer service reps, delivery drivers and team leaders as they handle both incoming<br />
and outgoing calls. There other situations where I was privileged enough to have been given the<br />
opportunity by a telesales rep to listen in to incoming telephone conversations from customers. I should<br />
admit that this ‘privilege’ was not officially permitted by top management; it was nonetheless made<br />
possible as a result of tactful arrangement I struck with a particular telesales rep. I also had the chance to<br />
travel to most delivery areas in the South-East of London with some of the delivery drivers.<br />
Most of these interviews lasted about 50minutes and some crossed the hour mark due to telephone<br />
interruptions from customers and work colleagues. Details from these interviews were electronically<br />
recorded and some notes also taken to fill data gaps during the transcription. Notes were read to the<br />
interviewees after every interview to allow confirmation and corrections to guarantee validity.<br />
2.2 Miscellaneous evidence<br />
Other forms of documentation such as quarterly progress meeting reports, company brochures,<br />
promotion leaflets, and many others were used to understand and make sense of the data that emerged<br />
from the interviews. The pieces of information gleaned from these materials proved quite helpful in later<br />
analysis and discussion of the data.<br />
56
Kofi Agyenim Boateng<br />
3. Findings –interaction activities at Adom Food Services Ltd<br />
Adom Food Services Ltd., henceforth designated as AFSL, trades in semi-prepared food and other food<br />
related products to small and medium-size restaurants in and around London. It is appropriate to point<br />
out at this stage that Adom is a fictitious nomenclature intended to protect the real identity of organisation<br />
where this study was carried out. AFSL has lately, through the use of technology, automated greater<br />
parts of the processes involved in commercial exchanges with its customers from order processing to<br />
delivery. Two unique services are provided: order processing by telesales representatives and order<br />
delivery by delivery drivers. The telesales reps are made up of about 12 different teams of roughly 5<br />
members, headed by a team leader each, while delivery is undertaken by a fleet of approximately 200<br />
vehicles.<br />
The study is contextualised around order processing and delivery activities which conspicuously feature<br />
telesales reps and delivery drivers in their use of portable, mobile and stationary ICTs in performing their<br />
assigned duties. The decision to use telesales reps and delivery drivers as the unit of analysis of this<br />
thesis is due to their particular use, and heavy reliance on ICTs to drive interaction in all its stripes in the<br />
enterprise-wide endeavours at AFSL. Secondly, the resolve to concentrate on ICT-driven interaction<br />
takes into due consideration the issues emotional repercussions that are significant in demarcating the<br />
context of this study.<br />
AFSL serves an extensive mix of customers including wholesalers and retailers. These include cafés,<br />
restaurants, burger bars, football clubs, ‘chicken and chips’ shops among others. The commercial<br />
activities at AFSL go almost seamlessly on a 24-hour basis from Monday to Friday and, up to noon on<br />
Saturdays and, sometimes Sundays. Supplying semi-packaged foods and drinks to small restaurants and<br />
shops requires registered customers and potential customers to telephone an all-purpose telephone<br />
number anytime between 9am and 5pm from Monday to Friday and, occasionally on Saturdays to make<br />
their orders. The details of the orders are then stored on a central database of the company until after<br />
5:30 p.m., when they have been matched against their respective contact details of the customers who<br />
ordered them.<br />
Building shipment resumes after all route planning activities have been done. Night loading is performed<br />
soon afterwards. Loading is done all night until about 6 a.m., by which time the individual drivers of the<br />
various trucks have arrived at the company’s premises.<br />
Having prefaced the diverse operational manifestations at AFSL, it is appropriate to bring out some of the<br />
details of the telephone conversations that emerged from the study. The thoughts from the interviewees<br />
(telesales reps, delivery drivers, telesales team leaders) reflect different philosophical orientations on the<br />
use of ICTs in mediated interactions. These orientations inform two broad thematic pathways, namely,<br />
emotional and psychological consequences on one hand and the motivational rationalities underlying the<br />
application of ICT-driven interactions. It must be said that all the names mentioned do not reflect the true<br />
identity of the people interviewed.<br />
One telesales rep expressed her frustration in promoting and selling on the phone to customers thus:<br />
‘When you’ve got like 20 calls holding and you know you gonna take the next call the<br />
customer is gonna be angry because they’ve been waiting, you know, and they don’t wanna<br />
hear you trying to sell them something; it’s hard’ – Theresa<br />
Another made the following remark.<br />
‘They [team leaders] are always on your back, I can’t force 11 people to get new lines’ –<br />
Georgina.<br />
‘Personally I don’t think I really like my job [a reference to the telesales work] anymore, I<br />
have got commitments so I have to come and get paid’ – Florence.<br />
‘I spoke to a customer who asked me to marry him over the phone … he went like ‘alright<br />
darling’, ‘okay sweetheart’ all this sort of talk’ – Louisa.<br />
‘You can sell a new line (in other words a new product) easier to a man than to sell to a<br />
woman… women, you know, just get on a bit but men oh!...they hear a woman’s voice and<br />
they know she is young and they go like: What’s your name, darling? Where’re you from? ...<br />
and stuff like this’ – Frances.<br />
57
Kofi Agyenim Boateng<br />
From the viewpoint of the delivery drivers, the following thoughts came to the fore during the interviews.<br />
A delivery driver’s reason for maintaining close relationship with a customer of the company is captured<br />
in the following way.<br />
‘My friend, the office people don’t know that this customer has my mobile number. I talk to<br />
her everyday and she knows when I am going to make deliveries. Sometimes she calls to<br />
find out when I am going to drop her order. And I say 10 minutes, half hour or even one<br />
hour, you understand? Me and this customer we are very good friends for a long time now’ –<br />
Gordon.<br />
In connection with the rationalities and motivational factors underpinning the application and preference<br />
of one technology tool to another the following matters surfaced during the interviews.<br />
‘Personally the email is a second option. I prefer the use of the telephone because that gives<br />
me a more appropriate description to the problems the staff may have. I don’t get immediate<br />
response to emails and I can’t put them any questions for instant response. That means I<br />
can’t really fish out what the problem is when I use the email’ – Thomas.<br />
‘Using emails wastes a lot of time, as talking on the phone gives me the opportunity to get<br />
instant feedback for my queries’ – Rosina.<br />
‘Talking over the telephone, you never know if the other person is listening to you so when<br />
you have the eye contact it makes a huge difference’ – Daniels.<br />
‘I send an email to the field sales reps and all of them get the same message<br />
simultaneously, it is not misinterpreted, and it is accurate, but if it is a telephone (mobile or<br />
landline), I have to send it one after the other. However, if you speak to someone, two or<br />
three different people asking the same question, you’d get two or three different<br />
interpretations…so that [an allusion to email] works very well for me’ – Robinson.<br />
One team leader expressed an innovative application of the email in enriching mediated interaction.<br />
According to him, he could use colour codes to highlight certain aspects of communicated messages.<br />
The research had the opportunity to witness an instance of colour-enriched mails.<br />
‘The beauty of email is you can prioritise your work, set reminders, colour code in terms of<br />
priority, arrange and delegate to other people’ – Katlin.<br />
The next section presents an analysis of the paper to flesh out the two principle themes of emotional<br />
consequences of mediated interaction and motivational factors triggering certain aspects of technology<br />
mediated interaction.<br />
3.1 ICT interaction and emotional dynamics<br />
ICT-driven interaction can result in certain psychological and emotional challenges that may impact<br />
negatively or positively on the performance of the individual concerned, especially if such emotional<br />
surges are not well handled by the people involved. It is obvious from the study that interactions<br />
empowered by ICTs have assumed nearly the same mechanical dimension as face-to-face interactions.<br />
The affordance by these technology objects makes it possible for what Collins (2004) refers to as<br />
‘emotional energy’ to be charged, the effect of which can match up to many close, proximal<br />
communication. Thus ICTs permit the essential logistical support for the breeding of networks, described<br />
as ‘networked individualism’ (Castells, 2001 p 130) as a central mechanism for flourished communication<br />
relationships. It can be stressed – without equivocation – the point that ICTs appear to have assumed a<br />
dominant role in configuring complex network of social organisation. These networks offer a gateway for<br />
individuals to share communication principles both ‘on-line and off-line, on the basis of their interests,<br />
values, affinities, and projects’ (Castells, 2001 p. 131).<br />
Proponents of the doctrine of social presence claim that face-to-face interaction can hardly be argued to<br />
have a perfect substitute (Culnan and Markus, 1987, Walther, 1992). However, it is evident from the<br />
study how repeated mediated interactions breed familiarity. This is found in the remarkable revelation of<br />
how ICT-driven interaction inspires, strengthens and sometimes affects the configuration of social<br />
relationships. The interview discovered breaches of strict organisational codes of practice as a result of<br />
familiarity between telesales reps, delivery drivers and customers occasioned by repeated and intensive<br />
mediated patterns of interaction.<br />
58
Kofi Agyenim Boateng<br />
The study for instance learned informally, if confidential, how a telesales rep developed amorous<br />
relationship with a customer of the organisation. He was promptly sacked when this ‘unofficial rapport’<br />
became known to the organisation. Therefore the material properties afforded by the technology enables<br />
what can take place face-to-face to equally prevail in a communication at a distance. This could be as a<br />
result of the observation elsewhere that in mediated interaction<br />
“there have to be signs, objects of some sort that are about some thing, objects whose<br />
function is reference rather than presence” (Borgmann, 1999 p.17).<br />
Hence mediated interaction makes it possible for initial encounters at the point of service delivery to<br />
develop into relationships paved by a bond of trust that follows dynamic interactions between the social<br />
agents. Some of the experiences of the telesales reps adequately capture this viewpoint. The discussion<br />
on this segment cannot be complete without recollecting certain responses from two telesales reps about<br />
product promotion on the phone which was laden with worry, frustration and intimidation. Indeed, one<br />
telesales rep could hardly suppressed her feelings when she declared in an interview that she feels quite<br />
apprehensive going to work at the thought of the challenge in meeting the daily sales target via the<br />
telephone. This worryingly frustrating experience connects well with an observation elsewhere that ‘the<br />
increasing competiveness of the modern economy has rendered people slaves to their workplaces’<br />
(Bolchover, 2005 p. 2). Also, Bunting’s (2005) ‘Willing Slaves…’ similarly has a remarkable resemblance<br />
with a workplace characterised by a stressed and overworked staff.<br />
3.2 Manifest rationalities in ICT-driven interaction<br />
Varying degrees of answers are understood to have motivated the application of various forms of<br />
communication tools for interaction predominantly, among the telesales reps. These declarations imply<br />
that user appropriation of ICTs depends on different situational factors connected with multiple rational<br />
explanations (Franz and Robey, 1994).The attribution of personal factors by the category of people<br />
identified in the study reflects sense-making posture in technology use that dwells largely on personal<br />
experience rather than any other social consideration. Different telesales reps assign different reasons for<br />
the use of the same technology device for almost the same job.<br />
It can also be conveniently argued that with respect to the manifest rationalities in ICT interaction, the<br />
technology opens up opportunities for manoeuvrability aimed at the achievement of other personal goals.<br />
For instance, colour coding email message to make it meaningfully rich and real, setting priorities of<br />
communicative trends with colleagues and the ability to transmit across spatial context makes email an<br />
appealing choice for engaging in certain aspects of mediated interaction.<br />
4. Concluding remarks<br />
Specific portions of the findings of this paper are unsympathetic to, and occasionally undermine, the<br />
fundamental postulates of the social presence theory. For instance, some interviewees showed they pay<br />
least regard to matters of social presence when instantiating interaction with their working colleagues.<br />
They have a firm indifference to the use of ICTs with specific reference to telephone and email for<br />
mediating their interaction. Their position seems to depend not essentially on the device being adopted<br />
but on the content of the message being broadcast via the device. This reflection was echoed by a<br />
couple of telesales reps’ opinions for which the following statement is illustrative.<br />
‘Just the same thing that I would say to the other person on the line is what I put down on the<br />
email; if I want to say something I would say the same on the email as I would via the phone’<br />
Collins (2004) largely sees emotions as a two-edged piece for erecting and changing social structures on<br />
the one hand, and a potential source of societal disagreements on the other, through the use of what he<br />
refers to as ‘emotional energy’. Mediated ICTs can build emotional energy through what I call ‘electronic<br />
ritual chains’ from the concept of Collins understanding of the ‘ritual’ in face-to-face encounters. This<br />
electronic means of building emotions by mediated technology needs to be taken further in all forms of<br />
communication at a distance. The emotional conditions necessitated by those communication activities<br />
are worthy of reflective consideration. For instance, evidence of emotional spillover in ICT-driven<br />
interaction arose from two divergent sources; one within the company (internal) and the other from<br />
outside the organisation (external).<br />
As emotions assumes the better part of a person, they are likely to become subjective to the degree that<br />
they tend, even if briefly, overlook their immediate allegiance to the cause of the organisation as a group<br />
committed to a common cause. A near quarreling remark by a telesales rep summarises the collective<br />
59
Kofi Agyenim Boateng<br />
effect of employee disenchantment when emotions surge to an intolerable level as a result of mediated<br />
interaction. This is because when pleasure, excitement and even enjoyment gain prominence in everyday<br />
interactions rising from intense communication, the possibility of getting involved personally, and with<br />
one’s emotions, can be unexpectedly sharp.<br />
Therefore our study of ICT-driven interactions should not limit our horizon to the artefacts and possibly<br />
the social implications of their implication into everyday life. Studies are encouraged to look intimately at<br />
how the behaviour of people changes overtime and the consequences and dynamics of such<br />
transformation. Again, this study opens the floodgate for future research to go beyond logistics to say<br />
emergency services, e.g. ambulance, the police or the fire service as a means for informing both theory<br />
and practice in ICT-driven interactions.<br />
References<br />
Argyle, M. (1969) Social Interaction, Methuen, London.<br />
Bernett, H. G., Masi, D. M. and Fischer, M. J. (2002), Web-enabled call centres: a progress report. Business<br />
Communications Review, 32, 7, 38-43<br />
Birdwhistle, R. L. (1970) Kinesics and Context, University of Philadelphia Press, Philadelphia.<br />
Blumer, H. (1969) Symbolic Interactionism: Perspective and method, Prentice Hall, Englewood Cliffs, NJ.<br />
Bolchover, D. (2005) The Living Dead: Switched Off, Zoned Out - The Shocking Truth About Office Lice, Capstone,<br />
Chichester, West Sussex.<br />
Borgmann, A. (1999) Holding On to Reality: The Nature of Information at the Turn of the Millennium, The University<br />
of Chicago Press, Chicago.<br />
Bunting, M. (2005) Willing Slaves: How the Overwork Culture is Ruling Our Lives, HarperPerennial, London.<br />
Calhoun, C. (1992), The infrastructure of modernity: Indirect social relationships, information technology and social<br />
integration In Social change and modernity (Eds, Haferkamp, H. and Smelser, N.) University of California Press,<br />
Berkeley, pp. 205-236.<br />
Castells, M. (2001) The Internet Galaxy: Reflections on the Internet, Business, and Society, Oxford University Press,<br />
Oxford.<br />
Collins, R. (2004) Interaction Ritual Chains, Princeton University Press, Princeton, NJ.<br />
Culnan, M. J. and Markus, M. L. (1987), Information Technologies In Handbook of Organizational Communication:<br />
An Interdisciplinary Perspective<br />
(Eds, Jablin, F. M., Putnam, L. L., Roberts, K. and Porter, L.) Sage, Newbury Park.<br />
Daft, R. L. and Lengel, R. H. (1986a), Organisational information requirements, media richness and structural design.<br />
Management Science, 32, 5, 554-571<br />
--- (1986b), Organizational Information Requirements, Media Richness, and Structural Design. Management<br />
Science, 32, 5, 554-571<br />
Daft, R. L., Lengel, R. H. and Trevino, L. K. (1987), Message Equivocality, Media Selection, and Manager<br />
Performance: Implications for Information Systems. MIS Quarterly, 11, 3, 355-366<br />
Franz, C. R. and Robey, D. (1994), An Investigation of User-Led System Design: Rational and Political Perspectives.<br />
Communications of the ACM, 27, 12, 1202-1209<br />
Gadamer, H.-G. (2004) Truth and Method, Continuum, London.<br />
George, J. F. and Carlson, J. R. (1999) In Proceedings of the 32nd Hawaii International <strong>Conference</strong> on System<br />
Sciences, Vol. 1 IEEE Computer Society Washington, DC, USA, Hawaii<br />
Grunow, D. (1995), The Research Design in Organization Studies: Problems and Prospects. Organization Science,<br />
6, 1, 93-103<br />
Hughes, E. C. (1971) The Sociological Eye: Selected Papers On Institution and Race (Part I) and Self and the Study<br />
of Society (Part II), Aldine Atherton, Chicago.<br />
Kakihara, M. and Sorensen, C. (2002), Exploring Knowledge Emergence: From Chaos to Organizational Knowledge<br />
Journal for Global Information Technology Management, 5, 1-14<br />
Kallinikos, Y. (2007), ICT, organizations, and networks In The Oxford Handbook of Information and Communication<br />
Technologies (Eds, Mansell, R., Avgerou, C., Quah, D. and Silverstone, R.) Oxford University Press, Oxford,<br />
pp. 273-292.<br />
Kayany, J. M., Wotring, E. C. and Forrest, E. J. (1996), Relational Control and Interactive Media Choice in<br />
Technology-Mediated Communication Situations. Human Communication Research, 22, 3, 399-421<br />
Ngwenyama, O. K. and Lee, A. (1997), Communication Richness in Electronic Mail: Critical Social Theory and the<br />
Contextuality of Meaning. MIS Quarterly, 21, 2, 145-167<br />
Olson, G. M. and Olson, J. S. (2000), Distance Matters. Human-Computer Interaction, 15, 139-178<br />
Orlikowski, W. J. and Schultze, U. (2004), A practice perspective on technology-mediated network relations: the use<br />
of Internet-based self-serve technologies. Information Systems Research, 15, 1, 87-106<br />
Peters, J. D. (1999) Speaking into the air: A history of the idea of communication, University of Chicago Press,<br />
Chicago.<br />
Schmidt, K. and Bannon, L. (1992), Taking CSCW Seriously: Supporting Articulation Work. Computer Supported<br />
Cooperative Work (CSCW), 1, 7-40<br />
Schutz, A. (1971), Making Music Together In Alfred Schutz, Collected Papers II: Studies in Social Theory (Ed,<br />
Brodes, A.) Martinus Nijhoff, The Hague, pp. 159-178.<br />
60
Kofi Agyenim Boateng<br />
Short, J., Williams, E. and Christie, B. (1976) The Social Psychology of Telecommunications, John Wiley & Sons,<br />
London.<br />
Suchman, L. (1987) Plans and Situated Actions: The Problem of Human-Machine Communication, Cambridge<br />
University Press, Cambridge.<br />
Thieme, R. B. (2003) The Plan of God, R B Thieme, Jr., Bible Ministries, Houston, Texas.<br />
Thompson, J. B. (1995) The Media and Modernity: A Social Theory of the Media, Stanford University Press,<br />
Stanford, California.<br />
Wagner, E. D. (1994), In support of a functional definition of interaction. The American Journal of Distance<br />
Education, 8, 2, 6-29<br />
Walther, J. B. (1992), Interpersonal Effects in Computer-Mediated Interaction: A Relational Perspective.<br />
Communication Research, 19, 1, 52-90<br />
Yin, R. K. (2003) Case Study Research: Design and Methods, SAGE Publications, Thousand Oaks.<br />
Yoo, Y. and Alavi, M. (2001), Media and group cohesion: Relative influences on social presence, task participation,<br />
and group consensus. MIS Quarterly, 25, 3, 371-390<br />
61
Investigating the Factors Inhibiting SMEs from Recognizing<br />
and Measuring Losses From Cyber Crime in South Africa<br />
Gino Bougaardt and Michael Kyobe<br />
University of Cape Town, Cape Town, South Africa<br />
michael.kyobe@uct.ac.za<br />
Abstract: The level of cyber attacks on organisations has increased tremendously in recent years. When such<br />
attacks occur, organisations need to assess the damage and loss from this crime. While large organisations have the<br />
mechanisms to determine such losses, SMEs lack such capability and often ignore the need to implement effective<br />
information security measures (Kyobe, 2008; Altbeker, 2000; Upfold and Sewry, 2005). Consequently, their risk<br />
exposure to cyber threats and the losses they incur from these attacks are often high (Ngo, Zhou, Chonka and Singh,<br />
2009). However, the current legislative requirements, costly legal liabilities for non-compliance, and increasing<br />
pressure by stakeholders (e.g., lenders, business partners) on SMEs to comply with good practices suggest that<br />
SMEs cannot ignore security any longer. In order to ensure accountability and compliance with security<br />
requirements, it is imperative for SMEs to identify, account and report cyber incidents and losses resulting from cyber<br />
attacks. This study investigated the factors that inhibit SMEs from recognizing and measuring losses from cyber<br />
attacks in South Africa. A survey involving twenty organisations from different business sectors was conducted and<br />
the results indicate that victimisation, resulting from a lack of awareness of cyber-crime has the greatest influence on<br />
SMEs’ ability to recognise and prepare losses from cyber attacks.<br />
Keywords: cybercrime, recognition and measuring losses, SMEs, victimisation<br />
1. Introduction<br />
Despite the availability of numerous methods and publications on how organisations can manage<br />
information security risks, SMEs still face serious challenges in managing cybercrime and resulting<br />
losses. Cybercrime are criminal activities involving the use of electronic devices and may lead to<br />
incidents such as theft of information; Sabotage of data of networks; loss of information due to<br />
eavesdropping ; financial fraud; denied access to information; and damage due to virus attacks (Kshetri,<br />
2009). These incidents result in financial and other losses to the organisations.<br />
Many SMEs fail to identify or recognise and account for these losses. They fail to manage these risks and<br />
continue to ignore the implementation of effective information security measures (Kyobe, 2008; Altbeker,<br />
2000; Upfold and Sewry, 2005). Consequently, their risk exposure to cyber attacks and the resulting<br />
losses continue to rise (Ngo, Zhou, Chonka and Singh, 2009). The current legislative environment,<br />
pressures from stakeholders (e.g., lenders, business partners) and liabilities in the event of violation of<br />
legal requirements make it imperative for SMEs to assess or estimate security risks and account for<br />
potential losses resulting from cyber attacks. Risk assessment is the process by which systems risks are<br />
identified and assessed in order to justify safeguards and protect systems from attacks (Carroll, 1996).<br />
Accounting for losses involves the evaluation and reporting of the damages when they occur. While<br />
studies on information security in SMEs are increasing, limited attention has been paid to SME inability to<br />
recognise and account for losses from cyber-crime.<br />
The objective of this study is to identify the factors that inhibit SMEs from identifying/recognizing and<br />
measuring losses from cyber attacks. In addition, we also determine the degree of influence each of the<br />
identified factors has on SME ability to recognise and prepare losses. In the following sections, the<br />
authors review literature on factors inhibiting recognition and measurement of losses. A conceptual model<br />
representing the relationship between these factors is presented. The results of an empirical test of this<br />
model are then presented. This is followed by conclusions and recommendations for future research.<br />
2. Literature review<br />
The process of identifying, recognising and reporting losses from cyber crime has not been easy for<br />
many organisations. Numerous challenges have been identified. For instance, difficulties in<br />
understanding what cyber crime represents, difficulties relating to risk identification and analysis,<br />
weaknesses in data recording and interpretation, poor design of security systems, risk management and<br />
human behaviour (Canhoto, 2010). Some of these challenges are examined in more detail below:<br />
62
Gino Bougaardt and Michael Kyobe<br />
2.1 Lack of awareness of cybercrime and lack of knowledge of being a cybercrime<br />
victim<br />
In his position paper to the Oxford Internet Institute, Baker (2010) argues that the process of collection<br />
and analysis of cybercrime data is often affected by a lack of understanding of what cybercrime means or<br />
represents. This lack of understanding of cybercrime, which is also evidenced by various ambiguous and<br />
conflicting interpretations of the term, impedes its recognition and measurement. Many organisations<br />
continue to be victimised because they are not aware of the nature or characteristics of this crime.<br />
Theories of victimisation explain the persistent victimisation of individuals and organisations by cyber<br />
criminals. Gottfredson and Hirsch (1990) in their General theory of crime, and Shreck et al (2003), in their<br />
work on victimisation show that victimisation results from low levels of self-control. A lack of self-control<br />
(defined here as the inability to control oneself or one’s emotions), is reflected in behaviours like shortsightedness,<br />
being insensitive or being impatient. SME managers are reported to be insensitive to<br />
cybercrime (Jacque, 2003; Zorz, 2003). Individuals with low self-control are believed to make decisions<br />
exclusive of those situations that increase their vulnerability and fail to change or mitigate their risk<br />
factors after the first victimisation (Forde and Kennedy (1997). This also renders several victimisation<br />
surveys inaccurate as they often underestimate incidence and prevalence rates (Fafinski, 2010).<br />
Solms and Solms (2004) assert that the lack of awareness sin is still committed by many companies.<br />
There are no proper awareness training programs and consequently, users are unaware of the risks of<br />
using their IT infrastructure and the potential damage they can cause to it.<br />
2.2 Lack of risk management skills<br />
The goal of risk management is to identify, measure, control and minimize losses associated with<br />
uncertain events (Patel and Zaveri, 2010). SMEs fail to recognise and measure losses because they do<br />
not engage in record keeping and IT risk assessment and management (Dimopoulos, Furnell and Barlow,<br />
2003). Dimopoulous et al. (2003) attributes this to the lack of funds, expertise and awareness of security<br />
risks. Risk analysis is perceived by SMEs as being complex, requiring specialist expertise and therefore<br />
something to be outsourced (Dimopouloulous et al (2003). It is also thought to disrupt management and<br />
employee activities throughout its duration. Furthermore, existing models (e.g. CRAMM) for evaluating<br />
the benefits of reducing the risks versus the investment in security technology are difficult to understand<br />
or use by SMEs (Dimopoulous et al., 2003). There are also challenges involved in comprehending the<br />
results and reports generated by these tools. The identification of risks is made harder for smaller<br />
organisations due to changes in technology and modes of vulnerabilities (Srinivasan and Abi-raad, 2003).<br />
2.3 Information system security design/infrastructure<br />
The design of the security system may also impact on the recognition and estimation of losses. Modern<br />
business environment comprises of many different applications and systems and each of these has its<br />
own threat profile (Conklin and Dietrich,2008). Such environment creates challenges for security<br />
practitioners responsible for developing security solutions. Consequently, these developers are forced to<br />
come up with piecemeal security designs, often disjointed, patched and can not monitor and<br />
comprehensively report on the security environment in the organisations (Conklin and Dietrich, 2008).<br />
Acccording to Canhoto (2010:1), the technical characteristics of the environment also influence how data<br />
on losses is derived and expressed e.g., data format, content and threshold, and which alerts and reports<br />
are produced. He states further that “formal aspects of the environment such as policies or regulations<br />
provide general definition of cybercrime, and may specify signs of alarm and the expected behaviour from<br />
analysts that detect such signs”. SMEs usually do not possess security and compliance policies (Kyobe,<br />
2008), and as indicated above, many do not engage in formal planning. Therefore such signs of alarm<br />
are often not identified before hand.<br />
2.4 Management attitude to security<br />
Individual cognitive processes, e.g., expectations, stereotypes and prior experiences, may influence<br />
attitude to security and the nature of data identified (Canhoto, 2010). Researchers argue that the failure<br />
by entrepreneurs and small business managers to proactively implement measures to handle risks has<br />
more to do with their personal characteristics (Ndubusi et al., 2005; Sjöberg et al., 2004; Nattaradol,<br />
2002; Orford et al., 2004). Proactive-risk handling is defined as the process in which potential risks to a<br />
business are identified in advance, analyzed, mitigated and prevented, and the cost of protection<br />
63
Gino Bougaardt and Michael Kyobe<br />
balanced with the cost of exposure to the risk. This does not appear to be done by small business<br />
managers or entrepreneurs. For instance, many entrepreneurs in South Africa are reported to have<br />
started businesses without giving proper consideration to economic, environmental, and cognitive<br />
limitations (Orford et al., 2004; Ladzani and Netswera, 2003). In his report on ICT adoption by rural SMEs<br />
in Thailand, Nattaradol (2002) also shows that lack of proactiveness and proper evaluation of ICT<br />
projects resulted in misjudgment or under-estimation of potential business and security risks. He<br />
identified several practices in ICT adoption which are indicative of irrational planning or behaviors (e.g.,<br />
failure to estimate project costs, use of unskilled or untrained staff to manage ICT installations, use of<br />
obsolete hardware and software and ignoring potential impact of hackers and sneakers).<br />
In addition, small business managers tend to have a strong desire for autonomy and control which could<br />
easily hamper the success of the organization (Kyobe, 2006). They tend to be rigid, traditional and<br />
usually do not draw up definitive duties or responsibilities for their subordinates. The high degree of<br />
generality resulting from undertaking heterogeneous work prevent employees from developing expertise<br />
in dealing with IT security issues (Kyobe, 2006).<br />
It is also widely reported that because of their false sense of security, small business managers are<br />
complacent about cyber-attacks and often shun good security systems and practices ( Zorz,<br />
2003;Jacque, 2003). Jensen (2004) found that security only became of much greater concern for the<br />
entrepreneurs once they had adopted e-commerce and experienced the reality of risks. Freeman (1999)<br />
explains this behaviour using Kübler-Ross’ (1969) loss model. He argues that organisations response to<br />
major environmental change is similar to individual response to loss. According to Kübler-Ross’ (1969)<br />
stage theory of loss, many who suffer loss proceed through these five stages – denial, anger, bargaining,<br />
depression, and acceptance. At the first stage, the person denies that the loss is inevitable. Freeman<br />
argues that this happens to organisations as well. Citing the work of Starbuck, Greve & Hedburg (1978),<br />
Freeman shows that managers would deny the extent of organisational crises to avoid blame.<br />
Another problem related to this is the unwillingness by managers to report cyber attack incidents. In many<br />
cases the victims withhold reporting (Lee, 1997). The importance of reporting incidents has been<br />
emphasised in many studies on information security and safety. Today, organisations are required to<br />
implement information security reporting schemes based on standards like BS-7799/ISO-17799 (ISO-<br />
IEC 27002 (ISO, 2007) (Calder and Watkins, 2005). Gonzalez (2005) maintains that information security<br />
reporting is a quality improvement process that is essential to reduce incidents and Sveen et al (2007)<br />
state that to learn from an incident and avoid it in the future the incident’s causes must be investigated by<br />
competent people. They caution however that the quality of an investigation is a function of the resources<br />
available and the workload. If for instance, resources are fewer than the allocated workload, this quality<br />
may be compromised. Phimister et al (2003) add that sporadic emphasis and management fear of liability<br />
may also hinder success in an incident reporting system. Eurim (2003) outlines several barriers to<br />
reporting including concern about confidentiality, disruption to business and loss of reputation. A culture<br />
of reporting does not exist in many SMEs due similar reasons outlined above (Kyobe, 2004;2006).<br />
2.5 Lack of knowledge of, and compliance with security regulations<br />
Cybercrime regulations require that organisations and individual recognise and account for damages<br />
resulting from cybercrime. Therefore a lack of knowledge of these security regulations and compliance<br />
with their requirements suggest inability to recognise and report on cyber-crime risks and losses. In South<br />
Africa cybercrime is regulated by the cyber crime section in Chapter XIII of the ECT Act, 2002<br />
(Michalson, 2005). This chapter introduces statutory criminal offenses relating to unauthorized access to<br />
data (e.g., through hacking), interception of data (e.g., tapping into data flows or denial of service<br />
attacks), interference with data (e.g., viruses) and computer related extortions, fraud and forgery. They<br />
also state that a person aiding those involved in these crimes will be guilty as an accessory. A person<br />
convicted of an offence related to the above is liable to a fine or imprisonment for a period not exceeding<br />
five years.<br />
Kyobe (2008) found that majority of the SMEs surveyed were not proactive in their approaches to<br />
compliance with such regulations. Few firms planned their security systems, allocated sufficient<br />
resources for compliance, trained their staff regularly and developed policies. Economical factors such as<br />
indirect compliance costs were found to be the main barriers to SME compliance in this study. Kyobe<br />
(2008) found that of the 30 South African Websites investigated, 40 percent had not implemented<br />
necessary legal requirements. 54% of all respondents never reported their effort towards compliance to<br />
64
Gino Bougaardt and Michael Kyobe<br />
auditors and 32 percent of the respondents were not aware of the requirements and liabilities of the ECT<br />
Act.<br />
2.6 Summary<br />
The key relationships as described in the previous sections are presented in the conceptual model in<br />
Figure 1 below. This model was then validated empirically in a study involving 20 SMEs as indicated in<br />
the following sections.<br />
Quality of Information system<br />
security design/ infrastructure<br />
Expertise in Info. Security & risk<br />
management<br />
Management attitude to security<br />
Awareness of cyber-crime and of<br />
being a victim of such crime<br />
Awareness of (and compliance with)<br />
security Regulations<br />
Figure 1: Conceptual model<br />
3. Methodology<br />
Recognition of Cyber crime<br />
and preparation of losses from<br />
this crime<br />
The sample consisted of SMEs in South Africa engaged in e-commerce activity since these are expected<br />
to experience greater risk exposed to e-crimes. SMEs across South Africa from various provinces and<br />
industries were targeted in order to generalise the findings (Saunders, et al., 2009, p. 217). The<br />
participants included senior management and employees responsible for IT infrastructure. These were<br />
specifically selected because they would be involved in making decisions relating to IT and its<br />
implementation. It would also enable us to collect and compare the views of different categories of<br />
respondents. An up-to-date database from the South African SME Toolkit web site<br />
(http://southafrica.smetoolkit.org) was used to identify the SMEs. The content administrator of this web<br />
site confirmed to us that the majority of the businesses registered on the web site are SMEs and that their<br />
details are updated annually. In order to increase the sample size, more SMEs were obtained from other<br />
sources, e.g., Waverly Business Park, Business Broadcaster, Facebook and postgraduate students.<br />
Twenty two responses were unfortunately received in total. This is however a typical problem in many<br />
SME surveys where responses rates are known to be very low (Kyobe, 2008).<br />
4. Data collection<br />
A questionnaire was used to collect the data since many of the organisations involved are located in<br />
different areas. The questionnaire was piloted with academics in the department. After making necessary<br />
65
Gino Bougaardt and Michael Kyobe<br />
adjustments, this was then emailed to the respondents together with a cover letter explaining the purpose<br />
of the study. The constructs and their measures were mainly adapted from previous research papers<br />
(Kyobe,2008; Upfold and Sewry, 2005; and Herath and Rao, 2009).<br />
Table 1: Constructs and measures<br />
Construct Measure Questions<br />
Adapted from<br />
Quality of Information Security<br />
Design/Infrastructure<br />
Lack of expertise in (Information<br />
security & risk management)<br />
-Our information security solution protects the entire<br />
business system<br />
-Use up-to-date software and Hardware<br />
-Use anti-virus software<br />
-Use a firewall and data encryption<br />
-Availability of IS/IT expertise in security<br />
-Seek expert assistance from external sources<br />
-Techniques employed are useful in determining<br />
cyber losses<br />
-experience difficulties using estimation techniques<br />
Management attitude to security -Mitigate threats or risks to business/IT systems<br />
-Discipline employees who violate security<br />
regulations<br />
- Train staff on IS risks<br />
- false sense of security<br />
Lack of awareness of cyber-crime<br />
and of being a victim of such<br />
attacks<br />
Awareness and compliance with<br />
regulations<br />
Recognise computer crime and<br />
prepare losses<br />
- readness/willingness to report incidents<br />
-Our system is often unavailable due to cyber attacks<br />
-Have suffered financial loss due to cyber attacks<br />
-<br />
- have documented a Security policy<br />
-Comply with the ECT Act<br />
-employees are aware of the ECT Act requirements<br />
and penalties<br />
- Perform record keeping<br />
-availability of expertise in preparing loss estimates<br />
- prepare financial loss estimates due to computer<br />
attacks<br />
-possess expertise in preparing financial losses<br />
we document information security activities<br />
- we conduct information security audits<br />
Kyobe (2008)<br />
Upfold and Sewry<br />
(2005)<br />
Kyobe(2006,<br />
2008)<br />
Kyobe(2008)<br />
Kyobe(2008)<br />
Upfold and Sewry,<br />
2005; and Herath<br />
and Rao, 2009<br />
The first part of the questionnaire obtained general information about the respondents. Subsequent<br />
sections measured the different aspects of the research model (i.e., quality of the security infrastructure,<br />
expertise in information security management, the attitudes of management towards information security,<br />
awareness of cyber-crime and acknowledgement of being a victim of such crime, knowledge of the<br />
security regulation and the extent to which SMEs recognise cyber crime and prepare losses..The<br />
demographic information of the SMEs is presented in Table 2 below.<br />
Table 2 indicates that 9 of the organisations are from the ICT business sectors. There are 11<br />
organisations based in the Western Cape and 6 organisations based in Gauteng. Further analysis<br />
revealed that 8 of the respondents were IT/IS staff while the rest were business owners/managers).<br />
Furthermore, business managers had on average internet trading experience of 5 years and computer<br />
experience of 15 years while IS/IT staff had 4 years and 11 years respectively.<br />
5. Reliability assessment<br />
The reliability test was conducted using the Cronbach’s alpha (Saunders, et al., 2009, p. 374). The<br />
Cronbach alpha coefficient above the recommended 0.70 would confirm the reliability of the questions<br />
66
Gino Bougaardt and Michael Kyobe<br />
used (Nunnally, 1978). Table 3 shows the reliability test results for the constructs in the conceptual<br />
model.<br />
Table 2: SME demographics<br />
1 4<br />
No. of<br />
Employee<br />
s Business Sector Location<br />
2 30<br />
Yrs<br />
trading<br />
on<br />
Internet<br />
Respondent<br />
Position<br />
Respondent<br />
yrs of<br />
computer<br />
experience<br />
Engineering<br />
Consulting<br />
Information<br />
Western Cape Owner<br />
Technology Western Cape Consultant (IT) 15<br />
3 12 NGO<br />
Broadband<br />
National 0 Software Developer 10<br />
4 150 provider-ICT<br />
Procurement to<br />
Western Cape 11 BP Manager- 9<br />
5 9 Mining & Industry<br />
Information<br />
Gauteng 4 Manager 10<br />
6 1 Technology Western Cape 1 Owner 20<br />
7 6 Telecoms North West 10 Sales Director 18<br />
8 2 Accounting Western Cape<br />
Limpopo/KZN/M<br />
4 Director 11<br />
9 24 Construction pumalanga Director 10<br />
10 2 IT GT 5 Director 15<br />
11 15 IT Consulting Gauteng 3 CEO<br />
CEO/Project<br />
17<br />
12 1 ICT Gauteng 1<br />
Manager 4<br />
13 5<br />
Management<br />
Consulting Gauteng 2 Managing Director 10<br />
14 1 Consultancy<br />
Information<br />
Western Cape 2 Owner 27<br />
15 7 Technology KZN 2 MD 20<br />
16 8<br />
Financial<br />
Services Western Cape 11 Director 20<br />
17 46 Building Western Cape 6 CEO 12<br />
18 (1)<br />
Financial<br />
Services Western Cape<br />
Western<br />
10<br />
19 80 Market Research Province 0<br />
Senior Architect/<br />
Strategist 16<br />
IT Development<br />
Team Leader 10<br />
20 80 Online Gaming Western Cape 5 Architect 5<br />
21 47 Construction Western Cape 9 Marketing Director 11<br />
22 20 Insurance Gauteng 3 Director 25<br />
The Cronbach alpha coefficients (in Table 3) are above the recommended 0.70 (Numally, 1978), thereby<br />
confirming reliability or internal consistency of the variables used in the present study. Most variables as<br />
stated above were adapted from prior studies. This also assisted in ensuring construct and content<br />
validity.<br />
67
Table 3: Assessment of reliability<br />
6. Analysis<br />
Gino Bougaardt and Michael Kyobe<br />
Constructs No.of<br />
Items<br />
Cronbach<br />
Alpha<br />
Quality of IS Security Design 3 0.70<br />
Lack of Expertise: Info. Security 4 0.82<br />
Management attitude to security 4 0.73<br />
Awareness of cybercrime (victimisation) 4 0.88<br />
Awareness of (and compliance with)security regulations 3 0.75<br />
Recognise computer crime & prepare losses from cyber crime 5 0.82<br />
Table 4: Mean responses<br />
No. of All respondents Business IT/IS staff<br />
Items<br />
Managers<br />
Constructs Mean Std Mean Std Mean Std<br />
Quality of IS Security Design 3 3.41 1.13 2.99 0.65 3.77 0.77<br />
Lack of Expertise: Info. Security 4 2.99 0.99 3.50 1.00 3.50 0.99<br />
Management attitude to security 4 2.49 0.91 2.19 0.65 3.90 0.97<br />
Awareness of cybercrime (victimisation) 4 3.01 1.16 2.27 0.88 3.98 0.88<br />
Awareness/compliance with security<br />
regulations<br />
3 2.45 0.73 2.59. 0.66 3.31 1.12<br />
Recognise & prepare losses from cyber crime 5 2.47 0.37 2.84 0.60 3.67 0.76<br />
6.1 Correlation analysis<br />
Correlation analysis was conducted to determine the association between variables.<br />
Table 5: Correlation analysis<br />
1 2 3 4 5 6<br />
1. Quality of IS security design 1<br />
2. Lack of expertise in IS 0.21 1<br />
3. Management attitude to security 0.39 0.44 1<br />
4. Awareness of cyber crime attacks (victimisation) 0.11 0.22 0.35 1<br />
5. Awareness/compliance with regulations 0.31 0.33 0.07 0.48 1<br />
6. Recognise & prepare losses from cyber crime 0.18 0.28 0.21 0.37 0.19 1<br />
* (correlations significant at .05 or less are shown in bold)<br />
6.2 Regression analysis<br />
Results of the regression analysis (Table 6) suggest that in the South African SMEs, lack of awareness or<br />
victimisation has the most influence on SMEs’ ability to recognise and measure losses from cyber crime<br />
than lack of expertise in information security management and management attitude to security. It is<br />
68
Gino Bougaardt and Michael Kyobe<br />
surprising to find in this study that quality of IS design and awareness and compliance with regulations<br />
did not have significant influence on recognition and measuring losses.<br />
Table 6: Regression summary for dependent variable: Recognising & measuring losses from cyber crime<br />
R = 0.8209 ; R Sq = 0.6739; Adjusted R Sq = 0.5720; F(5,16) =6.6132; p < 0.00162; Std.Error of estimate:<br />
0.68841<br />
Beta Std Err of<br />
Beta<br />
B Std Err of B T(16) P=level<br />
Intercept 2.5642 1.8677 1.8436 0.01887<br />
Quality of IS design -0.1238 0.1611 -0.2151 0.2798 -0.7678 0.45232<br />
Lack of expertise 0.3394 0.1895 0.35110 0.1959 1.7914 0.04932<br />
Management attitude<br />
to security<br />
0.4621 0.2177 0.1299 0.0612 1.1234 0.04365<br />
Lack of awareness<br />
(victimisation)<br />
0.8119 0.2034 1.11080 0.2783 3.990 0.00105<br />
Awareness &<br />
compliance with<br />
regulations<br />
0.5262 0.2431 1.0391 0.4800 2.1648 0.4589<br />
7. Discussion of findings<br />
Table 4 shows that most respondents did not consider the quality of IS security design to be good. The<br />
mean score on this item was 3.41 (uncertain). Respondents were not certain whether the security of the<br />
system (e.g., Firewall, software) really protected their businesses. Almost similar responses were<br />
obtained for the level of security expertise (2.99). Table 2 indicates that ½ of the respondents had more<br />
than 10 years of IS/IT experience. This suggests that possession of IT skills does not necessarily<br />
translate into proper management of IS security risks. There was a positive and significant association<br />
between information security design/infrastructure and the dependent variable (see Table 5, item 6 -<br />
recognise and prepare losses from cyber crime).Table 5 shows a positive and significant association<br />
between quality of information security design and the dependent variable. This suggests therefore that<br />
the former does influence the latter. Insufficient awareness of IT risks and computing limitations are major<br />
factors inhibiting small organizations from engaging in effective planning and monitoring of business<br />
operations (Kyobe, 2004).<br />
Table 4 also shows that overall most respondents seem to ignore information security requirements. The<br />
mean score was 2.49 indicating rather a negative attitude or lack of attention to security requirements.<br />
Management attitude to security was measured by asking respondents to indicate whether they mitigated<br />
risks, disciplined employees who violated security requirements, trained staff on IS risks and reported<br />
security violations. Further analysis of the responses of business managers and those in IT/IS staff<br />
confirms that information security is not a major concern for business managers (mean score for<br />
business managers was 2.19 compared to 3.90 by IT/IS staff). It is not surprising therefore that Table 5<br />
reveals a significant correlation between Management attitude to security and (quality of IS security<br />
design and lack of expertise in IS security). These influence SME ability to recognise losses as confirmed<br />
by the positive and significant relationship revealed in Table 5.<br />
In the case of awareness of cybercrime (and victimisation), the overall mean response was 3.01 (see<br />
Table 4), which suggests the respondents were uncertain. However, a breakdown of the responses<br />
indicates that only those with technical skills appear to be aware that their systems were unavailable due<br />
to cyber attacks and that they do suffer losses due to cyber attacks (mean score was 3.98). Business<br />
managers were mainly uncertain of the potential for victimisation.<br />
Compliance with security regulations remains a major problem in these organisations. The mean score<br />
was 2.45 (disagree). This may not be surprising given the fact that most business managers were also<br />
found to be least concerned about information security risks and also did not know they were potential<br />
victims of cyber attacks. A further analysis show that even IT/IS staff were not certain whether they<br />
complied with security regulations (mean score was 3.31). This confirms Jacque’s (2003) observation<br />
that SME managers were insensitive to cybercrime. Such level of insensitivity translates into lack of self<br />
control which results into re-victimisation (Shreck, 2003). The fact that this construct was found to be<br />
associated with recognition and reporting of losses from cyber-attacks confirms that lack of compliance<br />
influence SME ability to prepare losses.<br />
69
Gino Bougaardt and Michael Kyobe<br />
Most respondents did not agree that they do recognise and prepare losses from cyber attacks (Mean<br />
score 2.47). However, some IT/IS staff seemed to agree that they do to some extent (3.67). Recognition<br />
and preparation of losses from cyber attacks was measured by asking respondents to indicate if they<br />
kept records, possessed expertise in preparing losses, document information security incidents and<br />
conduct security audits (see Table 1). Table 5 shows that all other constructs were significantly correlated<br />
with this construct.<br />
However, when these relationships are analysed in a regression analysis, it is surprising to find that the<br />
quality of IS design, and awareness and compliance with regulations have no significant influence on<br />
recognition and measuring losses. This could perhaps be explained by the small sample size used in the<br />
present study<br />
Results of the regression analysis (Table 6) suggest that in these South African SMEs, lack of awareness<br />
of cyber crime (or victimisation) has the most influence on SMEs’ ability to recognise and measure losses<br />
from cyber crime than lack of expertise in information security management and management attitude to<br />
security. It is surprising to find in this study that quality of IS design and awareness and compliance with<br />
regulations did not have significant influence on recognition and measuring losses. This could also be<br />
attributed to the small sample size and the number of items used to measure these constructs. It is<br />
however interesting to find that our suspicion regarding three of the constructs (i.e., lack of awareness of<br />
cyber-crime (victimisation); management attitude to security; and lack of expertise in risk management)<br />
were found to be correct in this analysis.<br />
8. Conclusion<br />
This study shed more light on those factors influencing preparation of losses from cyber attacks in SMEs.<br />
Lack of awareness of cyber-crime appears to be a serious challenge for many SMEs and it is not<br />
surprising that these organisations continue to be victimised. Because SMEs managers do not<br />
understand what cyber-crime is, they simply may not take it seriously and may not find it necessary to<br />
engage in risk management initiatives. Risk management systems are vital to the evolution of SMEs<br />
(Sanchez, Ruiz, Fernandez-Medina, and Piattini, 2010). Management need to engage in risk<br />
management practices but this requires a change in attitude or in the way information security is<br />
perceived. Upfold and Sewry (2005) reported that SME management view information security policies as<br />
hard and expensive. SMEs managers can not continue rejecting good security practices due to their false<br />
sense of security (Kyobe, 2008).<br />
The findings presented in the present paper are however based on evidence gathered from only twenty<br />
two SMEs. Therefore precautions need to be taken when generalizing these findings. This study should<br />
be repeated with a much larger sample and the relationships between the constructs tested again in a<br />
regression analysis. Given the limited sample size, we could not compare the responses of business<br />
managers and IT/IS staff. This could reveal more details about the potential causes of different behaviors<br />
of managers towards security. Future studies should investigate these relationships and their impact on<br />
the dependent variable.<br />
References<br />
Altbeker, A. (2000). E-security: The race against computer crime. Retrieved March 06, 2010, from<br />
http://www.iss.org.za/PUBS/CRIMEINDEX/00VOL4NO6/ESecurity.html<br />
Baker, W.H. (2010). Thoughts on Mapping and Measuring Cybercrime.<br />
Oxford Internet Institute Forum Mapping and Measuring Cybercrime. http://www.sfu.ca/~icrc/content/oxford.foru<br />
m.cybercrime.pdf. Page 26.<br />
Calder, A., and Watkins, S., (2005), IT Governance, 3rd edition, Kogan Page, London<br />
Canhoto, A. (2010). What’ before ‘How’. Oxford Internet Institute Forum .<br />
http://www.sfu.ca/~icrc/content/oxford.forum.cybercrime.pdf.<br />
Carroll, J. M. (1996). Computer Security. Third Edition, Newton: Butterworth- Heinemann<br />
Conklin, W. A., and Dietrich, G. (2008, January 7). Systems Theory Model for Information Security. Proceedings of<br />
the 41st Hawaii International <strong>Conference</strong> on System Sciences - 2008, (pp. 1-9). Big Island, Hawaii.<br />
Dimopoulos, V., Furnell, s.m., and Barlow, I.M. (2003).Considering IT Risk Analysis in Small and Medium<br />
Enterprises. Proceedings of the 1st Australian Information Security Management <strong>Conference</strong> 2003 (InfoSec03),<br />
Perth, Australia, 24 November 2003<br />
Available: scissec.scis.ecu.edu.au/proceedings/2003/infosec/pdf/02_final.pdf<br />
EURIM (2003) IPPR E-Crime Study Partnership Policing for the Information Society. Working Paper 1: Reporting<br />
Methods and Structures. http://www.eurim.org/consult/e-crime/dec03/ECS_WP1_web_031209.pdf<br />
Fafinski, S. (2010) Mapping and Measuring Cybercrime, Position Paper. p75.<br />
Mapping and Measuring Cybercrime. http://www.sfu.ca/~icrc/content/oxford.forum.cybercrime.pdf. Page 36<br />
70
Gino Bougaardt and Michael Kyobe<br />
Forde, D., and Kennedy, L. (1997). Risky lifestyles, routine activities, and the general<br />
theory of crime. Justice Quarterly, 14, 265-289.<br />
Freeman, S. (1999). Identity Maintenance and Adaptation: A Multilevel Analysis of Response to Loss. Research in<br />
Organizational Behavior, 21, 247-294.<br />
Gonzalez, J.J. (2005), Towards a Cyber Security Reporting System – A Quality Improvement Process, Springer,<br />
Berlin Heidelberg<br />
Gottfredson, Michael R., and Travis Hirschi. 1990. A General Theory of Crime. Stanford, CA: Stanford<br />
University Press.<br />
Herath, T., and Rao, H. R. (2009). Encouraging information security behaviors in organizations: Role of penalties,<br />
pressures and perceived effectiveness. Decision Support Systems , 47 (2), 154-165.<br />
Jaques, R. (2003). Survey finds firewall and antivirus software considered unimportant by SMEs, URL (Consulted<br />
September, 2004): http://www.frame4.com/php/printout689.html<br />
Kyobe, M. (2006). Entrepreneur behaviors on e-commerce security, In M. Khosrow-Pour (ed.) Encyclopedia of ecommerce,<br />
e-government, and mobile commerce, pp. 437-444).<br />
Kyobe, M. (2008). Evaluating Information Security within SMEs engaged in E-commerce in South Africa. Retrieved<br />
March 08, 2010, from http://www.isbe.org.uk/Kyobe<br />
Ladzani, M.W., and Netswera, F.G. (2003). Lessons from successful South African rural entrepreneurs, URL<br />
(Consulted October, 2007): http://www.sbaer.uca.edu/research/icsb/2003/papers/48.doc<br />
Lee, W. (1997). A Deterrent Measure Against Computer Crime:<br />
Knowledge-Based Risk-Analytic Audit, Singapore Management Review,<br />
January, pp 19-45.<br />
Michalson, (2009). “ Protection of Personal Information Bill - the implications for you”. Michalson.com. [Online]<br />
Available: http://www.michalsons.com/protection-of-personal-information-bill-the-implications-for-you/3041.<br />
Access date: 10 January 2010<br />
Nattaradol, P. (2002), Harnessing ICT potential for the benefits of farmers and the rural poor: Experience and vision<br />
of Bank for agricultural cooperatives (BAAC), Thailand. BAAC. URL (Consulted October, 2007):<br />
http://www.adb.org/annualmeeting/2002/seminars/presentation/nattaradol_presentation.pdf.<br />
Ndubisi, N.O., Gupta, O.K., and Ndubisi G.C. (2005). The Moguls’ Model of Computing: Integrating the Moderating<br />
Impact of Users’ Persona into The Technology Acceptance Model, Journal of Global Information Technology<br />
Management, 8(1): 27-47.<br />
Ngo, L., Zhou, W., Chonka, A., and Singh, J. (2009). Assessing the Level of I.T. Security Culture Improvement:<br />
Results from Three Australian SMEs. IECON '09 35th Annual <strong>Conference</strong> of IEEE 2009, (pp. 3189-3195).<br />
Porto, Portugal.<br />
Nunnally, J. C. (1978). Psychometric Theory. New York: McGraw-Hill.<br />
Orford, J., Herrington, M., and Wood, E. (2004). South African Report Global Entrepreneurship Monitor. Retrieved<br />
January 22, 2005, from www.gsb.uct.ac.za/gsbwebb/userfiles/GEM_ 2004.pdf .<br />
Patel, S. and Zaveri, J. (2010). Assessment model of cyber-attacks on information systems. Journal of computers,<br />
15(3): 352-359.<br />
Phimister,J., Oktem,U., Kleindorfer,P.R, and Kunreuther, H. (2003). Near-Miss Incident Management in the<br />
Chemical Process Industry. Risk Analysis, Volume 23, Issue 3, pages 445–459, June 2003<br />
Sanchez, C.R, Fernández-Medina,E., and Piattini, M. (2010): Managing the Asset Risk of SMEs. ARES 2010: 422-<br />
429<br />
Saunders, M., Lewis, P., and Thornhill, A. (2009). Research methods for business students (5th ed.). Harlow:<br />
Pearson Education Limited.<br />
Schreck, C.J., and Miller, J.M. (2003). Sources of fear of crime at school: What is the relative contribution of<br />
disorder, individual characteristics, and school security?, Journal of School Violence 2 (2003), pp. 57–77.<br />
Solms, B, and Solmns R. (2004). The 10 deadly sins of information security management, Computer & Security<br />
23(5): 371-376.<br />
Sjöberg, L., Moen,B., and Rundmo, T. (2004). Rotunde Publications 84. Explaining risk perception. An evaluation of<br />
the psychometric paradigm in risk perception research, URL (Consulted February, 2005):<br />
http://www.svt.ntnu.no/psy/Torbjorn.Rundmo/psychometric_paradigm.pdf<br />
Srinivasan, G., and Abi-raad, M. (2003). Risk factors associated with e-buziness infrastructure of SMEs. Paper 21. 1 st<br />
Australian Information Security Management <strong>Conference</strong>, Perth, W.Australia.<br />
Sveen, F., Rich, E., and Jager, M. (2007), "Overcoming organizational challenges to secure knowledge<br />
management", Information Systems Frontiers, Vol. 9 No.5, pp.481.<br />
Upfold, C. T., and Sewry, D. A. (2005). An Investigation Of Information Security In Small And Medium Enterprises<br />
(SME's) In The Eastern Cape. Retrieved March 06, 2010, from<br />
http://icsa.cs.up.ac.za/issa/2005/Proceedings/Research/082_Article.pdf<br />
Zorz, J.(2003, May 15). Small firms ‘shun’ PC security. BBC NEWS. Retrieved October 15, 2004, from<br />
http://www.net_security.org/news:php?id=2650.<br />
71
Sharing knowledge – The CoP way<br />
Sheryl Buckley and Apostolos Giannakopoulos<br />
University of Johannesburg, South Africa<br />
sbuckley@uj.ac.za<br />
paulg@uj.ac.za<br />
Abstract: Managing academics’ knowledge to the benefit of all is a challenge to any institution. One way to share<br />
knowledge is through a community of practice. However, knowledge sharing within communities is complex.<br />
Challenges such as a lack of trust, a lack of incentives and a poor culture of learning in an institution have an effect<br />
on effective knowledge sharing. A questionnaire consisting of open- and closed-ended questions was administered<br />
to academics. This article reports on academics’ awareness and views of a community of practice at a higher<br />
education institution in South Africa. Empirical evidence shows that academics would be willing to share knowledge.<br />
However, time constraints, an unwillingness among academics to share knowledge, and a lack of support or<br />
participation from management could become obstacles to knowledge sharing.<br />
Keywords: knowledge sharing, communities of practice, higher education, trust, learning culture, knowledge<br />
1. Introduction<br />
Knowledge is the driving force of a knowledge economy. Therefore, the way knowledge is shared and<br />
created and the way these actions are managed could lead to either a competitive advantage for an<br />
organisation or it could lead to its demise. The same rationale applies to an academic institution. The<br />
knowledge that academics possess, forms the largest percentage of such an institution’s market value.<br />
Since knowledge (both tacit and explicit) resides in the minds of the people and some of it can be<br />
subsequently codified and become ‘common to all knowledge’ (Smith and McKeen 2004, 398), managing<br />
academics’ knowledge becomes a challenge to the institution.<br />
One way to share knowledge is through a community of practice (CoP) (Wenger 2004). A CoP consists<br />
of groups of people who come together voluntarily, both face-to-face and virtually, to share and to learn<br />
from one another. For a CoP to exist there must be people, shared interest, shared competencies and<br />
shared activities performed on a regular basis that advance learning and knowledge (Wenger, 2006).<br />
Knowledge sharing within communities is complex due to the lack of shared consensual knowledge or<br />
shared sense of identity (Barab and Duffy 1998; Thorpe 2003). Firstly, the development of trust is seen to<br />
be a key foundation which requires to be developed before effective knowledge sharing can occur. To<br />
facilitate effective community knowledge sharing, effort needs to be invested in developing the social<br />
relationship (and hence trust) (Muller 2006) between members of the communities. Secondly, appraisal<br />
systems need to be coupled to a proper system of rewards for those staff that enhance the institution’s<br />
stock of knowledge and use it to creative ends (Sallis and Jones 2002, 43). Lastly, a learning culture is a<br />
key component for ensuring that an institution can benefit from the knowledge at its disposal (Sallis and<br />
Jones 2002, 78). It can be argued that academics looking for knowledge connect with those who possess<br />
it and in this way they are able to improve their practice and create a competitive sustainable advantage<br />
for themselves and the institution. If the above issues are not addressed, knowledge sharing will be<br />
compromised.<br />
This article reports on participants’ awareness and views of communities of practice at a HE institution<br />
and proposes a framework for a CoP in an academic environment.<br />
2. Impact of knowledge and information sharing in higher education<br />
A university creates, investigates, evaluates and conveys culture through its scientific research and<br />
teaching activity (Brzezinski 1997, 202). Research universities are crucial factors in the development of<br />
new knowledge. Jarvis (2007, 90) writes: the university as whole, not just individual faculty members<br />
within the institution needs to be measured on the extent to which it serves the knowledge needs of<br />
society as a whole, as well as how it serves basic teaching and research.<br />
University expansion at the end of the 20 th century has led to a burgeoning of new programmes, modules<br />
and courses (Lea 2005, 180-181). The need for creating learning or knowledge communities rather than<br />
teaching factories is greater than ever before, if universities are to transform to modern universities (Msila<br />
2006, 90). As a university’s core business is teaching and learning, knowledge plays the central role in<br />
both (Jacob 2000, 141-142). The impact of failure to transmit knowledge successfully has a detrimental<br />
effect on the institution as well as the learner. Wick (2000, 519) contends that knowledge itself has limited<br />
72
Sheryl Buckley and Apostolos Giannakopoulos<br />
value because it becomes obsolete very quickly; and unless the institution ensures that its academics<br />
keep updating their knowledge, stagnation will prevail. However, it is not knowledge itself that is so<br />
valuable; rather it is the ability of an organisation’s members to generate knowledge and innovate, using<br />
that knowledge (Wick 2000, 520).<br />
For Duderstadt (2005, 81-82) “Knowledge can be created, absorbed and applied only by the educated<br />
mind, and … universities in particular, will play increasingly important roles as our societies enter this new<br />
age.” In a sense, knowledge is the medium of the university.<br />
There is widespread belief that information sharing is beneficial because it fosters collaboration, enables<br />
coordination, enhances capabilities and facilitates growth. According to (Lamberton, 2005, 155)<br />
information and communication technologies (ICT) can be used effectively as part of a toolbox for<br />
addressing global problems. Technology is a critical enabler and can be separated into two categories:<br />
infrastructure and collaborative applications (2004, 71). Infrastructure is the hardware and software that<br />
allow anyone to communicate with anyone else from any place at any time. Once the infrastructure is in<br />
place, collaborative technology (or groupware) applications can be effectively implemented to address<br />
communication across time and space.<br />
In discussing excellence in university teaching, Sherman, Armistead, Fowler, Barksdale and Reif (1987,<br />
72) consider excellence as developmental and such development as ‘a function of an interaction between<br />
individuals and their environments in an appropriate and facilitative learning atmosphere and not merely a<br />
function of the passage of time’. Such individuals are known to operate collaboratively with other<br />
colleagues as they see themselves as lifelong apprentices, always willing to improve their teaching<br />
practices. By implication, if a university aims for excellence, it is not sufficient to have highly qualified and<br />
knowledgeable academics, but such academics should interact with others to improve their practices. In<br />
an indirect way, they form unofficial CoPs, with excellence as one of their main objectives.<br />
Innovation is vital for any university wishing to be proactive in the marketplace and have the competitive<br />
edge (Du Toit 2000, 188; Skyrme 2003). Du Toit (2000, 188) argues that nowadays competitiveness in<br />
the marketplace is essential for the survival of South African universities and that the ability to sustain<br />
innovation is recognised as a strategic advantage; it has become evident that knowledge must be<br />
generated and integrated within the university at an accelerated pace. Such knowledge is essential for<br />
innovation or knowledge creation within a university. As academics are surrounded by other academics<br />
with the same purpose, it stands to reason that there should be close cooperation between them. Wenger<br />
(2004) argues that the social aspect then plays a central role in achieving their purpose.<br />
3. Challenges for institutions in implementing a community of practice<br />
Witt, McDermott, Peters and Stone (2007) believe that the education sector is now exploring the<br />
usefulness of building CoPs and connecting individuals with a common interest where knowledge is<br />
created, sustained and transformed. Already familiar to many academics is the idea of working in groups<br />
such as project teams, programme teams, special interest groups and forums to discuss, share and<br />
develop practice. According to Wenger (2004) the salient features of a CoP are the domain, the<br />
community and the practice. The domain is the area of interest, the community is formed by the<br />
relationships (conversations and discussions) between members and the practice is what community<br />
members do with learning derived from their interaction.<br />
3.1 Building trust<br />
Without trust and norms of openness, employees are likely to be concerned about the repercussions<br />
caused by expressing their views openly. When there are few norms of sharing and reciprocity,<br />
employees see little need and value in sharing knowledge with others. Employees who do not identify<br />
with the institution are unlikely to be willing to put effort into participating in knowledge-sharing activities<br />
(Pee and Kankanhalli 2007, 134). Trust is a crucial knowledge issue and it is a pivotal and essential<br />
element in long-term social relationships. However, building it is difficult and challenging at HE<br />
institutions. One of the best ways for HE leaders to engender trust is to be trustworthy themselves, and to<br />
act with integrity. Trust grows out of acts of trustworthiness (Sallis and Jones 2002, 35; Hislop 2004, 41).<br />
An HEI’s trust-building process has a number of linked stages. The first stage of building collaboration<br />
has its basis in the need to develop rapport. Groups, meetings and processes need to be established so<br />
that each employee can learn from the skills and expertise of others. It is a process of bottom-down<br />
73
Sheryl Buckley and Apostolos Giannakopoulos<br />
sharing, with management encouraging and establishing the space for teamworking to develop (Sallis<br />
and Jones 2002, 36).<br />
During the second stage the members of groups become more focused and active. This is when the<br />
prevailing corporate culture can encourage group activities by rewarding innovative thoughts and action<br />
(Chetley and Vincent 2003). At this time management needs to encourage people to look outwards<br />
towards their colleagues, rather than inwards. Network leaders need to break down old habits and ensure<br />
that the trust required for collaboration can thrive (Sallis and Jones 2002, 36). Techniques such as<br />
storytelling and learning conversations play a vital role in this process and should encourage employers<br />
to shift from hoarding to sharing knowledge. Technological infrastructures can also be useful for<br />
encouraging active collaboration (Hayes and Walsham 2000, 61; Sallis and Jones 2002, 36).<br />
To reach the third stage in the trust-building exercise, academics need to have easy access to<br />
information, ideas and solutions and need to know where old ideas can be unearthed and used again.<br />
Chat rooms, online communities and other inventive ways can be instrumental in supporting the sharing<br />
activities that build trust, and move the culture to one of knowledge sharing. As more and more people go<br />
online, informal communities spring up. Such communities allow important learning to take place; they<br />
know that learning works best where there is trust (Sallis and Jones 2002, 37).<br />
The final stage is rewarding the contribution of great teamwork. People are inclined to want some sort of<br />
scorecard that acknowledges their knowledge-sharing activity. The rewards should be personal (Sallis<br />
and Jones 2002, 37).<br />
3.2 Rewarding knowledge creation<br />
HE institutions need to recognise the value of their intellectual capital and the importance of the<br />
knowledge that their staff have. Appraisal and other feedback systems need to provide clarity about what<br />
is expected of staff, and to concentrate on staff contributions to knowledge creation. Davidson and Voss<br />
(2002, 99) emphasise that ‘[a] KM initiative will not get very far if – however actively you champion<br />
knowledge sharing – the existing remuneration system rewards knowledge hoarding’. Galunic and Weeks<br />
(1999, 189) suggest a variety of recognition and financial awards for knowledge sharing to the practice as<br />
a whole, including promoting developing intellectual capital as one of the four criteria used when<br />
determining promotion and bonuses. Liebowitz and Chen (2003, 431) feel that knowledge sharing is its<br />
own reward and that introducing formal incentive schemes might have the opposite of their intended<br />
effect. Among the incentives named by Liebowitz and Chen (2003, 432) are recognition; duty or need; a<br />
good frame of reference; a sense of give and take (quid pro quo); feedback mechanisms for letting<br />
knowledge sharers know their knowledge is being put to use; and the pleasure of helping someone attain<br />
their goals.<br />
HE institutions need to show their appreciation of staff contributions to the institution’s intellectual capital.<br />
Measurement could include skills acquisition as well as a willingness to undertake additional training and<br />
research output (Sallis and Jones 2002, 43).<br />
3.3 Culture of learning<br />
HE institutions need to build a culture of learning that allows for continuous knowledge creation and<br />
transformation (Sallis and Jones 2002, 77). Argyris (1994, 82) argues that institutions fail when their<br />
culture inhibits learning. A learning organisation encourages its staff to learn by giving them the<br />
mechanisms to share experience and best practice, and to improve their skills and capabilities. Leonard<br />
(1995, 63) identified a number of important implications for educational institutions. First, the importance<br />
of what she calls ‘firm specific skills’ or tacit knowledge indicates why it is difficult to replicate the<br />
performance of a successful institution. She argues that innovations cannot simply be imported from one<br />
institution to another. Educational institutions need to focus on developing the organisational-specific<br />
knowledge of its members; talent needs to be home-grown.<br />
Learning organisations have a knowledge-sharing culture, with knowledge as the ‘cement’ holding them<br />
together. The development of informal knowledge communities alongside more formal structures<br />
empowers staff to learn as they work, and builds a common language and context in which tacit<br />
knowledge can be shared effectively. This allows the institution to build its knowledge and share best<br />
practices (Sallis and Jones 2002, 78). CoPs can also be established by management (Thorpe 2003, 9)<br />
74
Sheryl Buckley and Apostolos Giannakopoulos<br />
rather than form naturally. While management cannot command CoPs to form, they can facilitate their<br />
growth (Thorpe 2003, 9).<br />
The perpetuation of a learning culture depends on the institution’s ability to develop practices, procedures<br />
and customs that draw new people into a CoP and compensate for membership losses (Buysse,<br />
Sparkman and Wesley 2003, 267).<br />
White and Weathersby’s (2005, 294) view is that the traditional culture of educational institutions often<br />
runs counter to the values of the learning organisation. <strong>Academic</strong>s value academic freedom, intellectual<br />
development, exploration and examination, peer review, acquisition of knowledge for its intrinsic worth,<br />
and collegiality. <strong>Academic</strong>s are also members of a loosely-coupled community, working as experts in<br />
their self-defined domain of a larger field, collaborating when necessary, and governed increasingly by<br />
professional academic administrators.<br />
White and Weathersby (2005, 295) contend that educators have the opportunity of choosing to become<br />
members of multiple CoPs within their own organisation and larger profession. Creating learning-oriented<br />
CoPs among faculty in even small, incremental ways is a challenge that may involve risk taking.<br />
4. Research methodology<br />
The research on which this article is based aimed at investigating CoP awareness and views of<br />
academics at a higher education institution in SA. A mixed method approach was used. Qualitative data<br />
was aimed at gaining a better understanding of the awareness and views of academics using a CoP. The<br />
sample consisted of 784 academics. A questionnaire consisting of four parts (section A, B, C and D)<br />
including open- and closed-ended questions was administered. Section A contained background<br />
information and for the purpose of this article, this data will not be discussed. Section B consisted of five<br />
closed-ended questions and aimed to determine whether academics were prepared to share their<br />
knowledge and experience with others in the same field; whether the character of the other members<br />
would play a role; whether they preferred to work alone or were team players, and which form of<br />
communication they preferred. Section C consisted of 16 closed-ended questions informing participants<br />
about CoPs and establishing their willingness to take part in a CoP. Section D (open-ended) was an<br />
opportunity for participants to highlight any other factors not mentioned about CoPs in section B and C.<br />
The questionnaire was administered (by an independent agency) online with a weekly email reminder to<br />
the participants for the duration of three weeks. Only 180 participants completed the questionnaire and<br />
this could be a limitation. A reason could be that academics were busy with assessments and preparation<br />
of mid-year examinations during that period. The sample may therefore be considered too small to<br />
generalise but trends can be noted.<br />
Content analysis and open coding was used to analyse the qualitative data. The reliability of the<br />
quantitative section of the questionnaire was determined by using Cronbach’s Alpha test to determine<br />
whether the measuring instrument was homogeneous.<br />
5. Results and discussion<br />
The questionnaire was completed by 22.9% (180) of the respondents.<br />
Section B produced the following results: All respondents indicated that they would be prepared to share<br />
knowledge and experience with others in the same field. This is a positive sign when formulating a<br />
strategy for the implementation of CoPs. The character of the other members would play a role for 77.8%<br />
(140) while sharing knowledge. This could be indicative of trust or social issues experienced by some<br />
respondents. Muller (2006) argues that effective knowledge sharing requires investing in social<br />
relationships. While 88.3% (159) of the respondents were of the opinion that they were team players,<br />
43.3% (78) of the respondents preferred to work alone. This was surprising, as 88.3% (159) of the<br />
respondents had indicated previously that they felt they were team players. Face-to-face communication<br />
was preferred by 85.5% (154) of respondents and 72.7% (131) preferred online communication. This is<br />
yet another positive sign for implementing a CoP, because as stated earlier, a CoP is a group of people<br />
who come together to learn from one another, both face-to-face and virtually (Wenger 2004).<br />
Correspondence was also indicated as an option.<br />
Section C aimed at informing participants about a CoP and at determining whether they would be willing<br />
to participate (see Table 1).<br />
75
Table 1: Informing participants about a CoP<br />
*Legend for<br />
theme<br />
categories<br />
Sheryl Buckley and Apostolos Giannakopoulos<br />
Question Disagree Agree Strongly<br />
agree<br />
KS I benefit from knowledge sharing 3 107 70<br />
PD I will learn more from a colleague 57 93 24<br />
KS We all possess tacit knowledge 6 107 64<br />
KS I would like to discuss my teaching with others 90 57 9<br />
KS I would like to share my teaching methods 37 121 17<br />
BR Sharing my personal knowledge builds trust 11 108 61<br />
ATT Sharing my knowledge increases my status 21 108 50<br />
PD By sharing knowledge, knowledge base will increase 1 70 109<br />
MP Sharing knowledge will influence decisions 34 113 27<br />
ATT I have knowledge for my peers 16 135 28<br />
KS Sharing knowledge is not detrimental to performance 17 124 37<br />
PR CoPs are created out of passion and ‘die’ from lack of<br />
it<br />
23 116 35<br />
PI Same identity creates a strong bond 22 124 29<br />
VP Since voluntary, I can opt out anytime 26 116 33<br />
VP, BR Friendships can develop based on trust 6 118 56<br />
MP My employer benefits from knowledge sharing 1 110 65<br />
* See Table 2 for explanation of legend.<br />
Comments given by respondents in Section D were the following: ‘Employers that are senior should<br />
encourage knowledge sharing and reward us for sharing good ideas, they should not use it as their own<br />
platform of success’ and ‘I was not sure what you meant with CoP but having considered your definition<br />
and survey, I conclude that CoP principles were instrumental in advancing my research abilities and<br />
industry reputation.’ Other factors noted were excessive administrative requirements, fear of the<br />
unknown, time limitations, ignorance, lack of knowledge about CoPs, heavy workload, no promotion, nonexistence<br />
of a CoP, not being established in the institution, organisational culture and personal reasons.<br />
Section B, C and D were grouped and categorised into themes using content analysis and open coding<br />
(see Table 1 and Table 2) and are discussed below.<br />
Knowledge sharing: The desire to share knowledge was supported by the majority of respondents i.e.<br />
98.3% (177), who felt that they would benefit. Sharing knowledge was not seen as being detrimental to<br />
performance 91% (163) and 95% (171) agreed that we all possess tacit knowledge. This supports<br />
Wenger’s (2004) and Wick’s (2000, 520) findings. There were 76.6% (138) of the respondents who were<br />
willing to share their unique and effective teaching methods with others.<br />
Build relationships: Here again the majority of the respondents felt that they were team players. Muller<br />
(2006) contends that it takes time to invest in social relations. A small minority, namely 3.3% (6),<br />
disagreed that by associating voluntarily with others to share knowledge, friendships can develop, based<br />
on trust. Most respondents, namely 93.9% (169), agreed that sharing helps build trust among peers.<br />
Communication: Most respondents, namely 85.5% (154), also indicated that they would use face-to-face<br />
communication but 72.7% (131) indicated that they would prefer to communicate online. Communication<br />
was not a detrimental factor for the respondents.<br />
Personal development: The majority, namely 99.5% (179), agreed that by sharing knowledge, their<br />
knowledge base would increase. This view is supported by Leonard (1995, 63),Sallis and Jones (2002,<br />
78) and Wick (2000, 519). The results show that respondents are willing to socialise and in the end their<br />
knowledge base increases.<br />
Attitudes: Only a small percentage of the respondents, namely 11.7%, (21), disagreed that sharing<br />
personal knowledge and experiences would maintain or increase their status amongst their peers.<br />
Findings by Smith and McKeen (2004, 399) concur with these findings. There was a desire by 90.5%<br />
(163) to share knowledge with their peers. The results show that the respondents were willing to share.<br />
76
Sheryl Buckley and Apostolos Giannakopoulos<br />
Management participation: The majority of respondents, namely 97.2% (175), agreed that the<br />
organisation also benefits from the sharing of knowledge among the employees. This is supported by<br />
Sallis and Jones (2002, 36-37, 43) and Du Toit (2000, 188). Furthermore, 77.7% (140) agreed that<br />
sharing personal knowledge and experiences would increase their power to influence decisions. As is the<br />
case in any institution, management approval in an HEI needs to be granted before a CoP may be<br />
started especially if it will be institutionalised. Therefore management involvement and encouragement<br />
may be crucial.<br />
Voluntary participation: Almost full consensus was obtained among the respondents that by associating<br />
voluntarily with others to share knowledge, friendships can develop, based on trust. White and<br />
Weathersby (2005, 295) support this standpoint. However, 14.4% (26) felt that they could not opt out at<br />
any time. Perhaps they felt they would be obligated to stay in such a community once they were<br />
members. Since a CoP primarily implies voluntary participation, members may leave whenever they wish<br />
to do so.<br />
Practice: Most of the respondents, namely 83.8% (151), agreed that CoPs are created out of passion for<br />
one’s work. This is one of the characteristics of a successful CoP (Wenger 2004; Buysse et al. 2003, 267;<br />
Sherman et al. 1987, 72). The practice is the end goal of what has been derived from the learning<br />
between the interactions with others. We all possess tacit knowledge and through our interactions, our<br />
knowledge base is increased.<br />
Professional identity: Most of the respondents, namely 85% (153), agreed that sharing the same identity<br />
creates a strong bond among the members of a CoP. This is supported by previous research (Pee and<br />
Kankanhalli 2007, 134; Thorpe 2003; Barab and Duffy 1998). Wenger (2004) claims that the strong<br />
feeling of identity of members belonging to a CoP is mostly regarded as a positive aspect. Perhaps<br />
respondents in the same field feel comfortable to share with others in a similar field.<br />
Advertise: A respondent commented that participants could be informed about CoPs by advertising the<br />
existence of the CoP. It is important to promote CoPs continually to engage new members, re-engage<br />
existing members and sustain the community over its lifespan.<br />
These themes (see Table 2) were further grouped into three main themes (see Table 3): domain,<br />
community and practice, which form the elements of a CoP.<br />
Table 2: Initial categories of themes<br />
1 Management participation MP 8 Knowledge sharing KS<br />
2 Practice PR 9 Voluntary participation VP<br />
3 Communication COMM 10 Professional identity PI<br />
4 Building relationships BR 11 Advertise AD<br />
5 Attitudes ATT 12 Trust TR<br />
6 Personal development PD 13 Time TI<br />
7 Infrastructure INF<br />
Table 3: Final themes<br />
Domain 8<br />
Practice 1, 2, 6, 9, 10, 11, 13<br />
Community 3, 4, 5, 12<br />
77
Sheryl Buckley and Apostolos Giannakopoulos<br />
Factor analysis was carried out on the data of the survey to confirm the validity and reliability of the<br />
findings. The purpose of the factor analysis was to determine whether the 16 items could be organised or<br />
grouped into a smaller set of underlying factors. Four factors were generated from the 16 items and are<br />
presented in Table 4. For this article, 0.700 was used as the benchmark against which to measure the<br />
Cronbach’s Alpha values. The four factors that were generated were personal development (factor 1),<br />
mutual beneficiation (factor 2), self-esteem (factor 3) and self-recognition (factor 4).<br />
There was no noteworthy difference between the gender of the respondents for the four factors used. It is<br />
also possible that the respondents who participated in the survey met some of the criteria of a CoP.<br />
There was also no noteworthy difference between the ages, current positions or number of years<br />
employed of the respondents for the same factors used.<br />
Table 4: Factor analysis<br />
Factors generated Cronbach’s Alpha<br />
Personal development .740<br />
Mutual beneficiation .729<br />
Self-esteem .704<br />
Self-recognition .652<br />
It was concluded that the prerequisite conditions for CoP were present, so a framework for CoP was<br />
developed to help start and maintain CoPs. CoPs are formed by people who engage in a process of<br />
collective learning in a shared domain. The findings of this research confirm most of the findings of the<br />
literature that the three key elements that are critical for a CoP should be taken into consideration when<br />
creating a CoP, namely: domain, community and practice.<br />
Figure 1 presents a framework based on the themes specified in Table 2. These themes were grouped<br />
into the three key elements of a CoP: domain, community and practice. This framework could be used to<br />
assist in the creation of new CoPs and to improve existing CoPs where a society of CoPs can ultimately<br />
develop. The findings positively suggest that a CoP might increase knowledge sharing in a university.<br />
Domain has been linked to management, as management is perceived to have the resources to create<br />
awareness among the academic community, to educate, encourage use, provide support (financial and<br />
technical) and reward sharing. Without the support and approval of management, no CoP can be<br />
created. Community and practice have been linked to academics. <strong>Academic</strong>s could develop, use,<br />
support, mentor and evaluate CoPs. Finally, CoPs are integrated to form global societies of CoPs in<br />
which academics participate.<br />
6. Conclusion<br />
The main elements of a CoP either featured, or were accepted by most of the respondents. One factor<br />
that appears to inhibit either the flourishing of an existing CoP or the formation of a new one is too little<br />
time. Most of the respondents accepted the other elements. This is very encouraging, since if the<br />
necessary elements to establish a CoP are available, then half of the objective has already been<br />
achieved. The two factors that appear from the findings to play an important role in the effectiveness of<br />
knowledge sharing are the attitudes of the academics and participation by management. Management<br />
has the ability to create an atmosphere that can be conducive to the sharing and creating of knowledge<br />
among academics by focusing on the domain, community and practice.<br />
It is the combination of a domain, community and practice that constitute a community of practice. And it<br />
is by developing these three elements in parallel that such a community is cultivated. In a higher<br />
education institution, where knowledge is created, disseminated and transmitted, management perhaps<br />
needs to play an even more active role in developing communities of practice if such communities are to<br />
be successful.<br />
78
If existing<br />
CoP<br />
Management<br />
<strong>Academic</strong>s<br />
Create<br />
awareness<br />
(advertise<br />
existing and<br />
new CoPs)<br />
Use CoP<br />
Sheryl Buckley and Apostolos Giannakopoulos<br />
Support CoP<br />
Educate &<br />
encourage<br />
(workshops)<br />
Mentor CoP<br />
<strong>Academic</strong>s<br />
develop or<br />
build new<br />
CoPs<br />
Willingness to<br />
participate in a<br />
CoP<br />
Figure 1: Framework for the creation of a CoP (adapted from Buckley 2009, 178)<br />
Evaluate CoP<br />
Use CoPs<br />
Themes identified<br />
from questionnaire<br />
survey<br />
79<br />
Support CoP<br />
<strong>Academic</strong>s<br />
Integrate CoPs<br />
Knowledge sharing<br />
Mentor CoP<br />
Domain<br />
Attitudes<br />
Build relationships<br />
Community<br />
Evaluate CoP<br />
Create global and<br />
societies of CoPs<br />
Practice<br />
Integrate CoPs<br />
Advertise<br />
Communication<br />
Management participation<br />
Personal development<br />
Practice<br />
Professional identity<br />
Voluntary participation<br />
Time
References<br />
Sheryl Buckley and Apostolos Giannakopoulos<br />
Argyris, C. (1994). Good communication that blocks learning. Harvard Business Review 72(4):77-85.<br />
Barab, S.A. and Duffy, T.M. 1998. From practice fields to communities of practice.<br />
http://crlt.indiana.edu/publications/complete.pdf. (Accessed 7 December 2009).<br />
Brzezinski, L. (1997). Reflections on the university. In J. Brezezinski and L. Nowak (eds). The idea of the university.<br />
Amsterdam: Rodopi. Pp.201-221.<br />
Buckley, S.B. (2009). Knowledge sharing through communities of practice at institutions of higher education.<br />
University of Johannesburg.<br />
Buysse, V., Sparkman, K.L. and Wesley, P.W. (2003). Communities of practice: Connecting what we know with what<br />
we do. Exceptional Children 69(3):263-277.<br />
Chetley, A. and Vincent, R. (2003). Learning to share learning: An exploration of methods to improve and share<br />
learning. www.healthcomms.org. (Accessed 14 April 2010).<br />
Davidson, C. and Voss, P. (2002). Knowledge management: An introduction to creating competitive advantage from<br />
intellectual capital. Auckland: Tandem Press.<br />
Du Toit, A.S.A. (2000). Knowledge management: An indispensable component of the strategic plan of South African<br />
universities. South African Journal of Education 20(3):187-191.<br />
Galunic, C. and Weeks, J. (1999). Managing knowledge at Booz-Allen and Hamilton: Knowledge on-line and off.<br />
Fontainebleau, France: Instead.<br />
Gorelick, C., Milton, N. & April, K. (2004). Performance through learning: knowledge management in practice.<br />
Elsevier Butterworth-Heinemann.<br />
Hayes, N. and Walsham, G. (2000). Competing interpretations of computer supported co-operative work.<br />
Organization 7(1):49-67.<br />
Hislop, D. (2004). The paradox of communities of practice: Knowledge sharing between communities. In P.M.<br />
Hildreth & C. Kimble (Eds). Knowledge networks: Innovation through communities of practice. Hershey: Idea<br />
Group.<br />
Jacob, M. (2000). Imaging the future university. In M. Jacob and T. Hellström (Eds). The future of knowledge<br />
production in the academy. Buckingham: The Society for Research into Higher Education, Open University<br />
Press.<br />
Jarvis, P. (2007). Globalisation, lifelong learning and the learning society: Sociological perspectives. Lifelong learning<br />
and the learning society, Volume 2. Routledge.<br />
Lamberton, D.M. (2005). Information sharing. In D. Rooney, G. Hearn and A. Ninan. Handbook of the knowledge<br />
economy. Edward Elgar Publishing Limited.<br />
Lea, M.R. (2005). Communities of practice in higher education. Useful heuristic or educational model? In D. Barton<br />
and K. Tusting (Eds). Beyond communities of practice. Language, power and social context. Cambridge:<br />
Cambridge University Press.<br />
Leonard, D. (1995). Wellspring of Knowledge: Developing and Sustaining the source of Innovation. Boston: Harvard<br />
Business School Press.<br />
Liebowitz, J. and Chen, Y. (2003). Knowledge sharing proficiencies: The key to knowledge management. In C. W.<br />
Holsapple (Ed.), Handbook on knowledge management (pp. 408-438). New York: Springer-Verlag.<br />
Msila, F. (2006). Massification: Preparing faculty for open learning systems. Progressio 28(1, 2):82-96.<br />
Muller, P. (2006). Reputation, trust and the dynamics of leadership in communities of practice. Journal of<br />
Management and Governance 10(4):381-400.<br />
Pee, L.G. and Kankanhalli, A. (2007). Development of knowledge management capability in an institution of higher<br />
learning in Singapore. In K.S. Ning (Ed.). Knowledge management: Singapore perspectives 2007. Singapore:<br />
Information and Knowledge Management Society.<br />
Sallis, E. and Jones, G. (2002). Knowledge Management in Education. Enhancing Learning and Education. London:<br />
Kogan page Limited.<br />
Sherman, T.M., Armistead, L.P., Fowler, F., Barksdale, M.A. and Reif, G. (1987). The quest for excellence in<br />
university teaching. The Journal of Higher Education 58(1):66-84.<br />
Skyrme, D. (2003). Knowledge management: Making sense of an oxymoron. www.skyrme.com/insights/22km.htm.<br />
(Accessed 1 September 2009).<br />
Smith, H.A. and McKeen, J.D. (2004). Knowledge management in organisations: The state of current practice. In<br />
C.W. Holsapple (Ed.). Handbook on knowledge management 2. Knowledge directions (p. 395-410). Berlin:<br />
Springer-Verlag.<br />
Thorpe, M. (2003). Communities of practice and other frameworks for conceptualising, developing and evaluating<br />
NCSL’s initiatives in linking staff and school communities. Unpublished report for the National College of School<br />
Leadership.<br />
Wenger, E. (2004). Knowledge management as a doughnut: Shaping your knowledge strategy through communities<br />
of practice. Ivey Business Journal, January/February.<br />
http://www.iveybusinessjournal.com/view_article.asp?intArticle_ID=465. (Accessed 7 December 2009).<br />
Wenger, E. (2006). Communities of practice: A brief introduction.<br />
http://www.ewenger.com/theory/communities_of_practice_intro.htm. (Accessed 7 December 2009).<br />
White, J. and Weathersby, R. (2005). Can universities become true learning organizations? The Learning<br />
Organization 12(3):292-298.<br />
Wick, C. (2000). Knowledge management and leadership opportunities for technical communicators. Technical<br />
Communication 47(4):515-529.<br />
80
Sheryl Buckley and Apostolos Giannakopoulos<br />
Witt, N., McDermott, A., Peters, M. and Stone, M. (2007). A knowledge management approach to developing<br />
communities of practice amongst university and college staff.<br />
http://www.ascilite.org.au/conferences/singapore07/procs/witt.pdf. (Accessed 30 September 2009).<br />
81
An Empirical Framework of key Success Factors for<br />
Software Process Improvement<br />
Nuntarat Bunditwongrat, Mathupayas Thongmak and Atcharawan Ngarmyarn<br />
Thammasat University, Bangkok, Thailand<br />
nuntarat.b@gmail.com<br />
mathupayas@gmail.com<br />
a_ngarmyarn@yahoo.com<br />
Abstract: During the last decade, software industry has concerned more about the quality of software process, as<br />
indicated by an increasing number of proposed software process improvement (SPI) standards and models.<br />
However, the adoption rate of these standards and models in software development is still low and the failure rate of<br />
SPI implementations is very high. For these reasons, this paper is aimed at investigating the key success factors for<br />
SPI implementation to answer two research questions. The first question is, what factors do respondents think that<br />
they have an impact on SPI implementation success?; and the second question is, does the organizational size have<br />
moderating effect on the relationship between human factors / organizational factors and SPI implementation<br />
success? The conceptual model and hypotheses of this study are proposed by extending and integrating conceptual<br />
models from prior research, collecting the human, organizational, and implementation factors associated with SPI<br />
implementation in industry. Furthermore, the contextual factor that is organizational size was added to the model to<br />
test the sensitivity of the variation in software organizations. Then, a quantitative questionnaire survey was gathered<br />
from 23 software organizations in Thailand which apply CMMI (Capability Maturity Model Integration) as their SPI<br />
guidelines in order to test the conceptual model and hypotheses of the study. The results indicate that SPI<br />
implementation success depends on six factors, i.e. management commitment, staff involvement, allocated<br />
resources, pilot projects, setting SPI goals, and defined process model, while organizational size has no moderating<br />
effect on SPI implementation success. Finally, the main contribution of this paper is the guidance for SPI practitioners<br />
on key factors that practitioners need to focus during an SPI implementation. Moreover, this guidance will be useful<br />
in the processes of allocating resources and prioritizing tasks to effectively implement SPI in the organization.<br />
Keywords: software process improvement, SPI, CMMI, key success factors<br />
1. Introduction<br />
Software plays an increasing role in today’s society. Modern software organizations operate in a highly<br />
dynamic market, under time and cost constraints (Abrahamsson 2001). Software quality has received<br />
attention from both academic and software industry (Niazi, Wilson, and Zowghi 2003a, 2003b).<br />
Organizations have started to apply SPI initiatives to increase the maturity and quality of their software<br />
processes (Abrahamsson 2001).<br />
During the last decade, the software industry has concerned more about software process improvement<br />
as indicated by an increasing number of proposed standards and models. All of these standards are<br />
claimed to increase the likelihood of SPI success (Dyba ํ 2001, 2005). Investment in process improvement<br />
has significant benefits such as improving productivity and product quality, reducing cost and time to<br />
market, increasing organizational flexibility and customer satisfaction, and enabling opportunities to enter<br />
the international market (Abrahamsson 2001; Brietzke and Rabelo 2006; Jalote 2002; Niazi, Wilson, and<br />
Zowghi 2003b).<br />
The number of proposed SPI standards and models has been increased; however, there has been<br />
limited success for many SPI efforts. Several companies have been carrying out SPI projects but some of<br />
them give up before realizing its benefits and others take much longer than expected to get it<br />
accomplished. The current problem with SPI is a lack of an effective strategy to successfully implement<br />
those standards or models (Abrahamsson and Iivari 2002; Babar and Niazi 2008; Brietzke and Rabelo<br />
2006; Hardgrave and Armstrong 2005; Niazi, Wilson, and Zowghi 2003b, 2005).<br />
CMMI is currently the most widely used framework for SPI. The recent report from SEI indicated that<br />
there was an increasing in CMMI appraisal from software companies around the world up to 15% during<br />
the last six months of 2008 (ThailandSPIN 2009). Thailand is the second highest number of appraised<br />
companies in Southeast Asia; however, the number is still low compared with all software companies in<br />
Thailand (Poopaka 2009; Siamturakij Media 2009). So, it is interesting to investigate the key success<br />
factors for SPI to assist software companies in effectively planning SPI implementation strategies. In this<br />
paper we focus on two research questions:<br />
82
Nuntarat Bunditwongrat et al.<br />
RQ1: What factors do respondents think that they have an impact on SPI implementation success?<br />
RQ2: Does the organizational size have moderating effect on the relationship between human<br />
factors/organizational factors and SPI implementation success?<br />
2. Background<br />
2.1 SPI concept<br />
During the last decade, SPI has become a popular approach to deliver improvements in software<br />
products. Many companies have either a formal or informal SPI program based on SPI model, such as<br />
CMM, CMMI, and ISO/IEC 15504 (SPICE). These models are SPI frameworks for defining and<br />
measuring processes and practices that can be used by software companies (Babar and Niazi, 2008;<br />
Hall, Rainer, and Baddoo 2002; Niazi, Wilson, and Zowghi 2003b). The report from Data & Analysis<br />
Center for Software (DACS) support that successful SPI programs reduce number of defects by 95%,<br />
reduce software development schedules by 71%, and increase productivity in terms of lines-of-code or<br />
function points per day by 222%. Additionally, Software Engineering Institute (SEI) reported the ratio of<br />
average return of 5:1 for investments in successful SPI programs (Abrahamsson 2001).<br />
2.2 CMMI<br />
CMMI is a model used to develop and refine software development processes which measures both<br />
capabilities and maturity levels. The model was developed by SEI, University of Carnegie Mellon, USA.<br />
CMMI is the successor of the CMM which integrate and standardize the separate models of CMM, and to<br />
eliminate other drawbacks of CMM. It is accepted and widely used in software development over the<br />
world (Malaiwong 2004, 2008a).<br />
CMMI identified processes to efficiently improve software quality starting from design, delivery, and<br />
maintenance. The CMMI version 1.2 have five maturity levels consisting of 25 process areas which is the<br />
group of best practices needed to comply to achieve those task objectives. Higher maturity levels require<br />
more set of process areas to be complemented (Phillips 2005; National Electronics and Computer<br />
Technology Center 2009).<br />
The result from survey indicates that software organizations which apply CMMI as their SPI guideline can<br />
work more systematic and easier to track with the flexible standard practices. By using these standard<br />
practices, managers need to seek an adaptive strategy to deal with any changes that may occur all the<br />
times. As a result, managers can see working situation more clearly in order to solve problems efficiently<br />
and well prepare to solve unexpected problems that may occur (Malaiwong 2008b).<br />
3. Conceptual model and hypotheses<br />
The conceptual model and hypotheses of the study are proposed by extending and integrating<br />
conceptual models from prior research. We focus on the human, organizational, and implementation<br />
aspects. Also, the contextual factor was added to the model to test the sensitivity of the variation in<br />
organizational size. Figure 1 shows the research model.<br />
3.1 Human factors<br />
3.1.1 Management commitment<br />
Commitment concerns with the extent to which resources are made available for SPI and management's<br />
interest in SPI (EI-Emam, Goldenson, McCurley, and Herbsleb 2001). Without commitment from all<br />
organizational levels, the initiative will likely fail or the results are not reached. The concept of<br />
commitment has been brought up as one of the most important human factors determining whether a<br />
well-planned process improvement program will succeed or not (Abrahamsson 2001).<br />
SPI initiatives involve a large number of people. Overall, SPI initiative does not get support unless<br />
management shows strong commitment to provide both people and financial resources for the initiative,<br />
and to spend their own time in participating, monitoring, and resolving issues. Active involvement of<br />
management provides strong motivation to the people participating in the initiative (Jalote 2002). In<br />
addition, Steltzer and Mellis (1999) ranked management commitment as number one of the ten factors<br />
83
Nuntarat Bunditwongrat et al.<br />
for SPI success. Management should also show that they believe in the initiative. Thus, we hypothesize<br />
that<br />
H1: SPI success is positively associated with management commitment<br />
3.1.2 Staff involvement<br />
Staff involvement is essential because if staff do not buy into the proposed changes, the improvement<br />
initiative is useless. Furthermore, they have detailed knowledge of the current processes. Steltzer and<br />
Mellis (1999) also ranked staff involvement as number two of important success factors with two reasons.<br />
First, their first hand experience of the processes is valuable. Second, participation from staff reduces the<br />
resistance to the developed processes.<br />
While Dyba ํ (2003, 2005) defines staff involvement as the extent to which employees use their<br />
knowledge and experience to decide, act, and take responsibility for SPI. Staff participation should be<br />
offered and managed to allow them improving their work and feeling a sense of contribution. Thus, we<br />
hypothesize that<br />
H2: SPI success is positively associated with staff involvement<br />
Human Factors:<br />
Management Commitment<br />
Staff Involvement<br />
Organizational Factors:<br />
Allocated Resources<br />
Pilot Projects<br />
Implementation Factors:<br />
Setting SPI goals<br />
Defined Process Model<br />
Figure 1: Conceptual research model<br />
3.2 Organizational factors<br />
3.2.1 Allocated resources<br />
Moderating Factor:<br />
- Organizational Size<br />
H7A H7B H7C H7D<br />
H5<br />
H6<br />
H1<br />
H2<br />
H3<br />
H4<br />
SPI Success<br />
Allocated resources are defined as the amount of staff time and resources dedicated to SPI (Goldenson,<br />
and Herbsleb 1995). SPI requires planning, dedicated people, management time, tools, and capital<br />
84
Nuntarat Bunditwongrat et al.<br />
investment. Organizations must commit their resources and constantly motivate people to contribute<br />
(Borjesson and Mathiassen 2004). SPI can’t be sustained without dedicated resources. Shen and Ruan<br />
(2008) recommend that at least one full time person for the SPI project.<br />
Training may be considered as a tool for implementing a successful SPI program and it requires allocated<br />
resources as well (Niazi, Wilson, and Zowghi 2003a, 2005). Training involves transferring important<br />
information and communicating process innovations to all relevant members. Providing enhanced<br />
understanding is mentioned in 72% of the ISO cases and in 87% of the CMMI cases (Stelzer and Mellis<br />
1999). It is important for capturing and disseminating experiences to less experienced staff (Rainer and<br />
Hall 2002). Thus, we hypothesize that<br />
H3: SPI success is positively associated with allocated resources<br />
3.2.2 Pilot projects<br />
After the process model was approved, the newly revised processes went through a testing phase (Shen<br />
and Ruan 2008). A careful planning to pilot new processes is critical because only some projects will be<br />
selected for pilot implementation. Good pilot planning is essential to maintain on-time SPI initiatives<br />
(Serrano, Montes, and Cedillo 2003).<br />
Practitioners are advised to implement SPI program within a particular department first. After proper<br />
planning and using pilot implementation experience, results of pilot implementation should be illustrated<br />
to other departments to get support and confidence (Niazi, Wilson, and Zowghi 2003b, 2005). Thus, we<br />
hypothesize that<br />
H4: SPI success is positively associated with pilot projects<br />
3.3 Implementation factors<br />
3.3.1 Setting SPI goals<br />
Stelzer and Mellis (1999) defined SPI goals as the degree to which everyone can make the translation<br />
from top management goals to the goals that each person is being asked to achieve. The goals must be<br />
decomposed to specific measures that everyone can see how their efforts relate to organization's<br />
success. Setting relevant goals means that the improvement efforts attempt to contribute to the success.<br />
While setting realistic goals means that the goals may be achieved in the foreseeable future with a<br />
reasonable amount of resources. This helps to direct the efforts towards common goals and to motivate<br />
people.<br />
The role of process improvement has been recognized as essential factor to achieve business objectives.<br />
So, clearly defined SPI program driven by business needs, have been suggested as a key success factor<br />
for SPI (Dyba ํ 2005). Thus, we hypothesize that<br />
H5: SPI success is positively associated with setting SPI goals<br />
3.3.2 Defined process model<br />
The major activities of defining process model involve in some definition of new processes and refine of<br />
existing processes. Frequently, process definition is the main challenge in an SPI initiative. Besides, it is<br />
imperative to have deployment mechanisms in place for deploying the processes to make the task of<br />
changing processes easier (Jalote 2002; Niazi, Wilson, and Zowghi 2003a; Shen and Ruan 2008).<br />
Creating process action team is also considered as a mechanism to effectively change or deploy new<br />
processes in the organizations. It is recommended to have a core group of full-time people that may<br />
include external agents (Hall, Rainer, and Baddoo 2002; Jalote 2002). The rest of the team should be<br />
users in processes, e.g. project managers, developers, testers, etc, to minimize the risk of resistance<br />
(Shen and Ruan 2008). Thus, we hypothesize that<br />
H6: SPI success is positively associated with defined process model<br />
85
3.4 Organizational size<br />
Nuntarat Bunditwongrat et al.<br />
Existing software engineering and organization development literature acknowledges that there are<br />
fundamental operational differences between small and large organizations. Small organizations seem<br />
more concerned about practice, while large organizations seem more concerned about formal process<br />
(Dyba ํ 2003). Despite this recognition, there has been no systematic attempt to verify whether small and<br />
large software organizations implement SPI programs differently. So, it is interesting to verify the<br />
importance of organizational size on SPI success in this study. Thus, we hypothesize that<br />
H7A: Organizational size moderates the effect of management commitment on SPI success.<br />
Management commitment influences the SPI success for large organization to a greater extent than that<br />
of small organization.<br />
H7B: Organizational size moderates the effect of staff involvement on SPI success. Staff involvement<br />
influences the SPI success for large organization to a greater extent than that of small organization.<br />
H7C: Organizational size moderates the effect of allocated resources on SPI success. Allocated<br />
resources influence the SPI success for large organization to a greater extent than that of small<br />
organization.<br />
H7D: Organizational size moderates the effect of pilot projects on SPI success. Pilot projects influence<br />
the SPI success for large organization to a greater extent than that of small organization.<br />
3.5 SPI success<br />
Dyba ํ (2003, 2005) measured SPI success based on the performance of the organization for the past<br />
three years with respect to cost reduction, cycle time reduction, and customer satisfaction. While<br />
Goldenson, Gibson, & Ferguson (2004) identified seven categories of performance measures: process<br />
adherence, cost, schedule, productivity, quality, customer satisfaction, and return on investment.<br />
Goldenson and Herbsleb (1995) asked the survey respondents how they would describe their<br />
organizations with respect to six performance characteristics: ability to meet schedule and budget<br />
commitments, address process predictability, product quality, staff productivity, staff morale / job<br />
satisfaction, and customer satisfaction. The overall patterns are quite clear that higher process maturity<br />
does appear to pay off in better organizational performance.<br />
4. Research design<br />
The data was collected between March and April 2010 from target sample of 293 employees who involve<br />
in SPI program of 23 software companies listed by Software Park, an agency supporting software<br />
companies in Thailand, using purposive sampling method. The questionnaires were distributed through<br />
both online and offline channels.<br />
The appraised companies comprise with the group of companies in maturity level 2, level 3, and level 5.<br />
There is no company in Thailand which achieved at level 4. According to statistics during the last two<br />
years, the appraised companies in Thailand has been increased up to 4 times, total of 30 companies<br />
divided into 9, 20, and 1 companies which achieve CMMI maturity level 2, level 3 and level 5 respectively.<br />
Notwithstanding, after contacting to the company which is mentioned to level 5 according to the list, we<br />
found that this company has not achieved at level 5, they are still at level 3. Hence, the sample of this<br />
research will cover only those companies achieved at level 2 and 3.<br />
5. Data analysis and result<br />
Our measurement model includes 40 items describing six latent constructs. The reliability of the multipleitem<br />
measurement scales were evaluated using Cronbach’s alpha coefficient. All constructs have a<br />
higher composite reliability than the benchmark of 0.7 recommended by Nunally (1978) as shown in<br />
Table 1.<br />
The unidimensionality of the six measurement scales was evaluated by performing an exploratory factor<br />
analysis using principal components analysis with varimax rotation. The six components were extracted<br />
86
Nuntarat Bunditwongrat et al.<br />
with eigenvalues greater than 1. The rotate matrix component is shown in Table 2. All item loadings for<br />
each component were greater than 0.5, which achieved an acceptable degree of unidimensionality.<br />
Table 1: Reliability analysis<br />
Independent Variables Number of Items α<br />
Management Commitment 6 0.847<br />
Staff Involvement 5 0.866<br />
Allocated Resources 10 0.939<br />
Pilot Projects 5 0.867<br />
Setting SPI Goals 6 0.893<br />
Defined Process Model 8 0.933<br />
Table 2: Rotated component matrixa<br />
Measurement Items<br />
1<br />
Component<br />
* 2 * 3 * 4 * 5 * 6 *<br />
A process has been established to distribute lessons<br />
learned to all relevant staff<br />
.734<br />
SPI implementation plan has been established .729<br />
The processes for improvement are clearly defined .721<br />
Mechanism has been established to manage changes .719<br />
Mechanism has been established to collect and analyze<br />
feedback data and to extract lessons learned<br />
.716<br />
Processes are adequately defined at an appropriate level<br />
of detail<br />
.714<br />
Continuously improve a SPI implementation methodology .664<br />
Mechanism has been established to monitor the progress<br />
of each process action team<br />
.576<br />
Sufficient budget for staff to participate in SPI training .722<br />
SPI training is provided for all related staff .698<br />
Sufficient time for staff to participate in SPI training .681<br />
Required tools have been provided for SPI implementation<br />
properly<br />
.673<br />
Required people have been provided for SPI<br />
implementation properly<br />
.663<br />
Sufficient staff time is allocated to SPI efforts .649<br />
SPI training helps developing skills and knowledge needed<br />
to perform SPI implementation<br />
.608<br />
Sufficient budget has been allocated to SPI implementation<br />
properly<br />
.571<br />
Training programs are reviewed on a periodic basis .537<br />
All future group or individual trainings of SPI are planned .520<br />
Staff members actively involve in defining their operational<br />
processes<br />
.780<br />
Staff members actively involve in setting goals for SPI<br />
activities<br />
.780<br />
Staff members have responsibility to involve in SPI<br />
activities<br />
.683<br />
Staff members interest and commit in SPI effort .675<br />
Opportunity for staff to exchange their ideas related to SPI<br />
implementation<br />
.549<br />
Management provides strong commitment for SPI .760<br />
Management actively participates in SPI activities .742<br />
Management actively supports SPI activities .679<br />
Management commits to provide resources for SPI .675<br />
Management establishes standard practices for SPI .638<br />
Management actively monitors the progress of SPI .578<br />
The results of pilot implementation are used for planning<br />
SPI implementation across the organization<br />
.664<br />
Careful planning of the piloting new processes .649<br />
Staff members satisfy with the performance of<br />
methodology in the pilot projects<br />
.646<br />
SPI implementation methodology has been tried and tested<br />
in pilot projects<br />
.567<br />
87
Nuntarat Bunditwongrat et al.<br />
Measurement Items<br />
1<br />
Component<br />
* 2 * 3 * 4 * 5 * 6 *<br />
The results of pilot implementation are communicated .540<br />
SPI goals are measurable .660<br />
SPI goals are aligned with organization’s business goals .607<br />
SPI goals are realistic .589<br />
SPI goals are well understood within the organization .589<br />
Fine balance between short-term and long-term SPI goals .544<br />
SPI goals are clearly stated .524<br />
Cronbach’s Alpha .847 .866 .939 .867 .893 .933<br />
a. Rotation converged in 7 iterations.<br />
*1 - Defined Process Model *2 - Allocated Resources *3 - Staff Involvement<br />
*4 - Management Commitment *5 - Pilot Projects *6 - Setting SPI Goals<br />
For respondent demographics, the ratio of female (51.54%) is slightly more than male (48.46%). The<br />
majority of the respondents’ age ranged between 25-35 years. Most of them have been working in the<br />
jobs related to software development process for 1-2 years. Most of the respondents’ company have staff<br />
dedicated to SPI in the range of 50-200 people with CMMI maturity level 3 and has started SPI<br />
implementation for 1-2 years.<br />
Multiple regression analysis was used to test the hypotheses. The regression model was run without<br />
constant because we used factor score to run in the model. Hence, the unstandardized and standardized<br />
coefficients were equal. Hypotheses 1 through 6 consider the individual relationships between SPI<br />
success and each of the six independent variables to answer our first research question. Table 3<br />
summarizes the results of our analysis for these hypotheses. The results show that hypotheses 1 through<br />
6 were significant. So, all six independent variables investigated have an impact on SPI implementation<br />
success.<br />
The contextual factor was added to the model to test the sensitivity of the variation in organizational size<br />
which is in accordance to our second research question. However, the estimated coefficients of the<br />
independent variables of the original model were not significantly changed by the introduction of the<br />
organization size. Thus, our analysis rejects Hypotheses 7A through 7D.<br />
Table 3: Regression analysis for hypotheses H1 – H6<br />
Model Standardized Coefficients (Beta) t<br />
Management Commitment .141 2.233*<br />
Staff Involvement .265 .757***<br />
Allocated Resources .218 2.514*<br />
Pilot Projects .229 .715**<br />
Setting SPI Goals .177 1.977*<br />
Defined Process Model .197 2.281*<br />
Dependent Variable: SPI Success<br />
R = .514 R2 = .265 SE = 0.865 (F = 17.202***)<br />
*p < .05 **p < .01 ***p < 0.005<br />
6. Discussion<br />
The results show that SPI success depends on six factors. Staff involvement seems to be associated with<br />
the highest explanatory power since it achieved the highest standardized regression coefficient ( =<br />
0.265). Next come to pilot projects ( = 0.229), followed by allocated resources ( = 0.218), defined process<br />
model ( = 0.197), setting SPI goals ( = 0.177), and finally, management commitment ( = 0.141). A<br />
88
Nuntarat Bunditwongrat et al.<br />
surprising result was the lowest explanatory power of management commitment in predicting SPI<br />
success. All the main studies on which this investigation is based share a strong belief in the importance<br />
of management commitment for the successful implementation of any type of change initiatives.<br />
However, the result is in agreement with Abrahamsson (2000)’s findings that many SPI initiatives do not<br />
require management commitment beyond obtaining the resources needed. It can be stated that if a<br />
committed manager should bring energy, passion, and excitement to improvement initiative, then<br />
management commitment is not a necessity. On the other hand, if management commitment is viewed<br />
as someone who provides resources that enables the process improvement to grow, then the<br />
management commitment is a necessity. Furthermore, the success in some cases as presented in<br />
Abrahamsson’s study could be the concept of champion, who makes a decisive contribution to the<br />
innovation by actively and enthusiastically making the improvement happen with having little support from<br />
management. There could be several explanations for this seemingly surprising result.<br />
With respect to the contextual factor, we found that organizational size did not have moderating effect on<br />
the relationship between human factors / organizational factors and SPI implementation success.<br />
However, the result is in agreement with Dyba ํ (2005)’s findings that the results hold for small, as well as<br />
large, software organizations, and for organizations operating in stable as well as in turbulent<br />
environments. Similarly, Goldenson and Herbsleb (1995) found that organizational size alone is unrelated<br />
to overall SPI success. Respondents who represent larger organizations are no more, or less, likely to<br />
claim such success than are those who come from smaller organizations. This means that the size of the<br />
organization does not limit its potential for SPI success. We believe that the success factors discussed<br />
here give many interesting insights that will be relevant and applicable to organizational change in most<br />
software organizations. If the success factors are implemented, they facilitate the success of<br />
improvement initiatives. If they are not implemented or not implemented correctly, this makes process<br />
improvement difficult to achieve, or may even cause failure of the initiative.<br />
7. Limitations<br />
Our participants belonged to only one standard, CMMI, and only one country, Thailand, which is a<br />
limitation of this study as the findings cannot be widely generalized to practitioners using other standards<br />
or models, and for other countries that have different environment and culture. As mentioned earlier, the<br />
sample of this research covers only those companies achieved at level 2 and 3. Nevertheless, software<br />
companies in Thailand are very productive and it would be another interesting case to investigate since<br />
they are well supported by many government agencies as witnessed in the statistics that the appraised<br />
companies in Thailand have been increased up to 4 times during the last two years (Malaiwong 2004).<br />
Another issue is that the questionnaire-based studies are usually based on self-reported data that may<br />
reflect what people think they should say, not necessarily what they actually observe or perceive. Our<br />
results are limited to the respondents’ knowledge and beliefs about the reasons, potential benefits and<br />
barriers that can undermine SPI initiatives. This situation can cause problems when practitioners’<br />
perceptions may be inaccurate. However, like many other studies based on opinion data, we have<br />
collected data from practitioners working in quite diverse roles and directly involved in SPI activities within<br />
their organizations to ensure that they could be representatives for our target population.<br />
8. Conclusion and future research<br />
The study focused on identifying the key factors for success in SPI by a quantitative survey. The results<br />
indicate that SPI implementation success depends on six factors, i.e. management commitment, staff<br />
involvement, allocated resources, pilot projects, setting SPI goals, and defined process model. While<br />
organizational size has no moderating effect on SPI implementation success. From a theoretical<br />
perspective, these findings add an important new dimension to empirical software engineering research<br />
to verify the importance of organizational size for SPI success. From a practical perspective, the main<br />
contribution of this paper is an actionable guidance for SPI practitioners on key factors that practitioners<br />
need to manage during an SPI effort. This guidance will be useful in processes of planning and<br />
prioritizing the implications, so that software organizations can evaluate their readiness and current<br />
practices to effectively implement SPI. Future research is needed to investigate such variable in other<br />
countries against the results of this study to see whether the different cultures and/or environments have<br />
any effect to the results. Furthermore, in case that the sample can be separated into more variety group<br />
of maturity levels, it is interesting to investigate the relationship between the success factors and maturity<br />
levels. Consequently, we would know that do the respondents from lower and higher maturity companies<br />
think that the same factors impact the implementation of SPI. Likewise, we can compare the factors<br />
recognized by companies with successful and non-successful SPI programs, as well as comparing CMMI<br />
to other standards/models, e.g. ISO/IEC 15504 (SPICE) or ITIL. In addition, qualitative research by in-<br />
89
Nuntarat Bunditwongrat et al.<br />
depth interview or case study should be done to gather insight into the SPI implementation success on<br />
how to efficiently establish SPI implementation strategies.<br />
References<br />
Abrahamsson, P. (2000) “Is Management Commitment a Necessity After All in Software Process Improvement?”,<br />
Proceedings of the 2<strong>6th</strong> Euromicro <strong>Conference</strong>, Vol. 2, pp 246-253.<br />
Abrahamsson, P. (2001) “Commitment Development in Software Process Improvement: Critical Misconceptions”,<br />
23rd International <strong>Conference</strong> on Software Engineering, pp 71-80.<br />
Abrahamsson, P. and Iivari, N. (2002) “Commitment in Software Process Improvement In Search of the Process”,<br />
Proceedings of the 35th Annual Hawaii International <strong>Conference</strong> on System Sciences, Vol. 8, pp 1-10.<br />
Babar, M.A. and Niazi, M. (2008) “Implementing Software Process Improvement Initiatives: An Analysis of<br />
Vietnamese Practitioner’s Views”, 2008 IEEE International <strong>Conference</strong> on Global Software Engineering, pp 67-<br />
76.<br />
Borjesson, A. and Mathiassen, L. (2004) “Successful Process Implementation”, IEEE Software, Vol. 21, No. 4, pp 36-<br />
44.<br />
Brietzke, J. and Rabelo, A. (2006) “Resistance Factors in Software Process Improvement”, CLEI ELECTRONIC<br />
JOURNAL, Vol. 9, No. 1.<br />
Dyba ํ, T. (2001) “An Instrument for Measuring the Key Factors of Success in Software Process Improvement”,<br />
Empirical Software Engineering, Vol. 5, No. 4, pp 357-390.<br />
Dyba ํ, T. (2003) “Factors of Software Process Improvement Success in Small and Large Organizations: An Empirical<br />
Study in the Scandinavian Context”, Foundations of Software Engineering, pp 148-157.<br />
Dyba ํ, T. (2005) “An Empirical Investigation of the Key Factors for Success in Software Process Improvement”, IEEE<br />
TRANSACTIONS ON SOFTWARE ENGINEERING, Vol. 31, No. 5, pp 410-424.<br />
EI-Emam, K., Goldenson, D., McCurley, J. and Herbsleb, J. (2001) “Modelling the Likelihood of Software Process<br />
Improvement: An Exploratory Study”, Empirical Software Engineering, Vol. 6, No. 3, pp 207-229.<br />
Goldenson, D.R. and Herbsleb, J.D. (1995) “After the Appraisal: A Systematic Survey of Process Improvement, its<br />
Benefits, and Factors that Influence Success”, Software Engineering Institute, pp 1-50.<br />
Goldenson, D.R., Gibson, D.L. and Ferguson R.W. (2004) “Why Make the Switch? Evidence about the benefits of<br />
CMMI”, Software Engineering Institute, pp 1-42.<br />
Hall, T., Rainer, A. and Baddoo, N. (2002) “Implementing Software Process Improvement: An Empirical Study”,<br />
Software Process Improvement and Practice, Vol. 7, pp 3-15.<br />
Hardgrave, B.C. and Armstrong, D.J. (2005) “Software Process Improvement: It’s a Journey, Not a Destination”,<br />
Communications of the ACM, Vol. 48, No. 11, pp 93-96.<br />
Jalote, P. (2002) “Lessons Learned in Framework-Based Software Process Improvement”, Proceedings of Ninth<br />
Asia-Pacific Software Engineering <strong>Conference</strong>, pp 1-5.<br />
Malaiwong, K. (2004) “Transition to CMMI”, [online], www.drkanchit.com/cmm/cmmitransition.pdf.<br />
Malaiwong, K. (2008a) “CMMI Implementation”, [online], www.drkanchit.com/cmm/CMMI_ ImplementationNew.pdf.<br />
Malaiwong, K. (2008b) “Benefits of CMM”, [online], www.drkanchit.com/cmm/cmm04.html.<br />
National Electronics and Computer Technology Center (2009) “Introduction to CMMI”, [online],<br />
cmmi.wikidot.com/knowledge.<br />
Niazi, M., Wilson, D. and Zowghi, D. (2003a) “A maturity model for the implementation of software process<br />
improvement: an empirical study”, The Journal of Systems and Software, Vol. 74, No. 2, pp 155-172.<br />
Niazi, M., Wilson, D. and Zowghi, D. (2003b) “A model for the implementation of software process improvement: a<br />
pilot study”, Proceedings of the Third International <strong>Conference</strong> on Quality Software, pp 196-203.<br />
Niazi, M., Wilson, D. and Zowghi, D. (2005) “A framework for assisting the design of effective software process<br />
improvement implementation strategies”, The Journal of Systems and Software, Vol. 78, No. 2, pp 204-222.<br />
Nunnally, J. C. (1978) Psychometric theory (2nd ed.), McGraw-Hill, New York.<br />
Phillips, M. (2005) “CMMI® Version 1.2 and Beyond …a Tutorial”. [online], Carnegie Mellon University,<br />
www.dtic.mil/ndia/2005cmmi/monday/phillips.pdf.<br />
Poopaka, S. (2009) “SPI@ease: CMMI Achievement Series #6 “CMMI vs Agile: Why Not Embrace Both !!!”,<br />
[seminar], Software Park Thailand.<br />
Rainer, A. and Hall, T. (2002) “Key success factors for implementing software process improvement: a maturitybased<br />
analysis”, The Journal of Systems and Software, Vol. 62, pp 71-84.<br />
Serrano, M.A., Montes, C. and Cedillo, K. (2003) “An Experience on using the Team Software Process for<br />
Implementing the Capability Maturity Model for Software in a Small Organization”, Proceedings of the Third<br />
International <strong>Conference</strong> On Quality Software, pp 327-334.<br />
Shen, B. and Ruan, T. (2008) “A Case Study of Software Process Improvement in a Chinese Small Company”, 2008<br />
International <strong>Conference</strong> on Computer Science and Software Engineering, Vol. 2, pp 609-612.<br />
Siamturakij Media (2009) “TMC supports software development of Thai private organizations”, [online],<br />
www.siamturakij.com/home/news/display_news.php?news_id=%20413339 303.<br />
Stelzer, D. and Mellis, W. (1999) “Success Factors of Organizational Change in Software Process Improvement”,<br />
Software Process: Improvement and Practice, Vol. 4, No. 4, pp 227-250.<br />
ThailandSPIN. (2009) “CMMI-appraised software companies in Thailand has been increased 3 times”, [online],<br />
Software Park Thailand, www.thailandspin.com/%20News/tabid/71/articleType/ ArticleView/articleId/42/-3-<br />
.aspx.<br />
90
Assessing Information Management Competencies in<br />
Organisations<br />
Andy Bytheway<br />
Cape Peninsula University of Technology, Cape Town, South Africa<br />
bythewaya@cput.ac.za<br />
Abstract: The history of the management of information systems includes many ideas that were intended to simplify<br />
the complexities of the management task, but there is still a great deal of wasted investment that produces no<br />
significant benefits. Much of the thinking has been rational and structured, but it can be argued that structured<br />
thinking will not solve the problems presented by the ever-increasing scope and depth of information systems, the<br />
need for improved responsiveness and agility, and the need to deal with a range of requirements that are sometimes<br />
behavioural and sometimes legislative. Three of the more frequently cited frameworks for information management<br />
(Zachman, Henderson & Venkatraman, Ward), are briefly reviewed and found to have common characteristics. They<br />
are combined into a new, simple arrangement of the central (and critically important) ideas. This new framework has<br />
been used as the basis of a survey instrument that is introduced and explained; it works at two levels - the "micro"<br />
and "macro" levels. It assesses perceptions of organisational capability to manage information well, as seen by<br />
respondents who are normally employees working in different roles with varying responsibilities. The survey<br />
instrument comes with an analysis and reporting package that is found to be suitable for the needs of busy<br />
managers, and the way in which micro and macro data is presently analysed and presented is demonstrated using<br />
data from a reference dataset, a CIO workshop, an investigation within a real estate agency and a large financial<br />
services organisation. The contribution of this work to the research programme from which it emanated is<br />
summarised and future directions briefly explained.<br />
Keywords: Information management; perceptions; IS/IT strategy; alignment; assessment<br />
1. Background<br />
There is a history of difficulty in delivering benefits from information technology investments, and for the<br />
last 40 years experts have worked to ease the problems.<br />
During the 1990s business managers and academics strove to find answers to critical questions. An<br />
early, extensive, review of literature concerning information systems "success" (DeLone & McLean, 1992)<br />
was well received and has since been updated; specific research has looked at process-based<br />
collaboration across corporate boundaries (Bytheway & Braganza, 1997); academic attention focused on<br />
strategic alignment (Kearns & Lederer, 2000; Chan & Reich, 2007); concerns about "agility", ethics and<br />
alignment have emerged (Tallon et al., 2000; Sambamurthy et al., 2003; Symons, 2005).<br />
More recently, there has been increasing attention to the management of benefits (Chatterji, 2007; Ward<br />
& Daniel, 2005) but business newspapers still report problems (Anon, 2008) and there is still difficulty<br />
with enterprise-wide systems (Seddon et al., 2010). It is reported that information systems strategy is still<br />
not properly understood (Chen et al., 2010), and ethics (Mingers & Walsham, 2010) and behavioural<br />
issues (Beaudry & Pinsonneault, 2010) are of concern. There are staggering losses involved in<br />
information technology investments in the public sector (Anon, 2010).<br />
In the face of all this, how can the complexities of managing information technology and systems be dealt<br />
with? When costs of technology are spiralling, how can they be justified to senior business managers in<br />
the board room? Concepts of information technology strategy might mirror business strategy, but exactly<br />
what do the "alignment" of these IT and business strategies really mean?<br />
This paper briefly reviews some ideas from this history and derives a new framework that proves to be<br />
comprehensible and workable. The framework leads to a survey instrument that assesses organisational<br />
capability to manage information well.<br />
2. Some important ideas<br />
2.1 Zachman reveals complexity<br />
The complexities of information technology management were first revealed in the 1980s in a framework<br />
for information systems architecture (Zachman, 1987); Zachman took a broad view of the issues but his<br />
ideas were necessarily detailed, and his frequently cited six-by-six matrix, with layers of technology<br />
management down one side and different perspectives of the business across the top, was beyond many<br />
91
Andy Bytheway<br />
managers' willingness to work with. 36 different points of concern (the intersection of the six rows and six<br />
columns of the matrix) were just too many to handle. Although there has been subsequent reference to<br />
the Zachman framework (Frankel et al., 2003) it has failed to gain currency in the general context.<br />
2.2 Henderson and Venkatraman seek simplicity<br />
Others have sought simpler views of the problem, and one frequently cited example is the Henderson<br />
and Venkatraman framework. It relates business and IT issues at the internal and external levels using a<br />
two-by-two matrix. It shows that there is a need for "functional integration" between the business and IT<br />
domains, and that there is a need for "strategic fit" between the internal and external worlds. Each of the<br />
four quadrants embodied in the framework are detailed in turn: scope, competencies and governance in<br />
the external portion and processes, skills and infrastructure or architecture in the internal portion.<br />
Internal External<br />
Strategic fit<br />
Distinctive<br />
competencies<br />
Business strategy IT strategy<br />
Business<br />
scope<br />
Administrative<br />
infrastructure<br />
Business<br />
Governance<br />
Processes<br />
Skills<br />
Organisational infrastructure and processes<br />
Automation Linkage<br />
Functional integration<br />
Systemic<br />
competencies<br />
Technology<br />
Scope<br />
Achitectures<br />
IT Governance<br />
Processes<br />
Skills<br />
IS Infrastructure and processes<br />
Business Information technology<br />
Figure 1: The Henderson and Venkatraman framework, linking business and IT from the internal and<br />
external viewpoints (Henderson & Venkatraman, 1993)<br />
Questions arise from a close examination of this model. It is interesting to see that competencies and<br />
skills are included, but why are they in different parts of the framework? A skill can be seen as a low level<br />
thing ("I can work this computer") but a competency is something else ("I can use this computer to<br />
produce a useful econometric model"). The implication of "processes" being in both the internal business<br />
and internal IT quadrants is that business processes and IT processes must be functionally integrated,<br />
but how is that possible? A single IT process might contribute to a wide range of business processes;<br />
conversely a typical business process might depend on many information systems. Why do we have<br />
"administrative infrastructure" on the left, and "architectures" on the right?<br />
The Henderson and Venkatraman model is simple at first sight, but it leads to a range of questions and<br />
lacks the sort of elegance and symmetry that makes these things memorable. It implies dependencies<br />
and relationships between its conceptual components, but these are not immediately evident on a first<br />
reading.<br />
92
2.3 The issue of alignment<br />
Andy Bytheway<br />
Efforts to improve the return on information technology investments have focused on the ways in which<br />
business and technology strategies can be more effectively aligned. However, this high-level approach to<br />
the management of IT and IS related investments has found only limited success.<br />
It has been argued (Yayla & Hu, 2009) that alignment is achieved by means of simple steps: strategising<br />
processes, increasing the level of communication, formalising policies, and so on. Easy to say, but is this<br />
taking us forwards? Where is the detail that tells us how to realise these worthy and somewhat obvious<br />
steps to success?<br />
"Alignment" is an appealing word but it has to be judged by the benefits that are delivered. The<br />
management of benefits has been an issue from the very early days (Baets, 1992) right through to recent<br />
times; it is now dealt with in standard texts (Ward & Peppard, 2002; Ward & Daniel, 2005).<br />
2.4 The concept of value from IT investments<br />
"Alignment" is appealing, but "value" is more so, especially when talking to senior management.<br />
Venkatraman introduced the idea of "eras" (Venkatraman, 1994), his observations concerned the<br />
increasing scope and reach of information systems, the increasing degree of business change that was<br />
required to benefit from them, and the increasing value to be gained thereby.<br />
Inevitably, practice takes time to adopt and adapt to new ideas and practical concerns continued to be<br />
reported at about that time (Uchitelle, 1996). But what Venkatraman had given us was a clear message<br />
that we are concerned with the management of systems and the information that comprises the essence<br />
of those systems, not just the technology.<br />
Progressive organisations worked along these lines, and references to "IT management" were<br />
supplanted by references to "IS management", and then to "Information management", as in the case of<br />
BP Chemicals (Cross, 1995). There was academic attention to these ideas based on workshops with<br />
working managers (Bytheway, 1996), and one feature of the work at BP chemicals led ultimately to the<br />
framework that is presented here. John Cross dunned his idea "Jacob's ladder" - a management stairway<br />
to a strategic heaven?<br />
2.5 Jacobs' ladder<br />
The arrangement of the four steps in BP Chemical's "Jacob's ladder" is shown in the figure below: At the<br />
bottom is the technology that comprise infrastructure for systems and business activity, and at the top the<br />
business processes that serve the business strategy (Cross & Earl, 1997).<br />
Value<br />
creation<br />
Value<br />
realisation<br />
Business processes<br />
Information<br />
Applications<br />
Infrastructure<br />
In-house<br />
expertise<br />
Outsourced<br />
expertise<br />
Figure 2: Jacob's ladder as promoted in BP Chemicals<br />
It is interesting that the creation of value is seen at the top, with in-house expertise; the realisation of<br />
value is seen at the bottom, based on the use of outsourced expertise (outsourcing was one of the<br />
principal outcomes for BP Chemicals at this time).<br />
93
Andy Bytheway<br />
However, this notion of value generation is seen in other work. Zachman had already articulated six<br />
"levels", from the representation of technology infrastructure (program code and data definitions), through<br />
technology, systems, enterprise and context of the enterprise (Zachman, 1987); Venkatraman was<br />
promoting the idea that data, information, knowledge, action and result were all related in a similar<br />
progressive way (Venkatraman, 1996), and Ward has promoted the idea that benefits from IT<br />
investments are achieved at different levels (Ward & Daniel, 2005).<br />
These ideas are summarised in the table below, and related to the five levels of the Information<br />
Management Body of Knowledge (IMBOK) that are discussed in the next section.<br />
Table 1: Linking information technology to business strategy<br />
Zachman Venkatraman Ward IMBOK<br />
Highest Context Result Strategic<br />
objective<br />
Business strategy<br />
Enterprise model Action Functional benefit Business benefit<br />
Lowest Representation of<br />
system<br />
Knowledge Business change Business process<br />
System model Information Enabling change Information<br />
system<br />
Technology Data Technology Information<br />
technology<br />
Zachman's "representation of system" is really only of interest to technology management, but his other<br />
five levels show strong empathy with the other viewpoints. One can argue that there are actually five<br />
levels at which management must operate (from the lowest to the highest):<br />
Information technology must be acquired, configured, and used to provide the requisite infrastructure<br />
so that a business can store its data and operate its …<br />
Information systems, some of which will necessitate changes to the way that the business operates<br />
its …<br />
Business processes, wherein knowledge is deployed in order to initiate the actions that are expected<br />
to deliver the desired …<br />
Business benefits, that will in turn realise the organisation's …<br />
Business strategy.<br />
This arrangement of ideas strongly reflects the historical thinking that is available, and it has been<br />
adopted as the foundation of the Information Management Body of Knowledge, which currently<br />
comprises a knowledge base, a handbook, a developing survey instrument, and a supporting community<br />
web site (http://www.imbok.org).<br />
3. The Information Management Body of Knowledge<br />
As can be seen from the table above, at the heart of the IMBOK is the idea of the business process,<br />
incorporated by Zachman (with other ideas) in his enterprise model, seen by Venkatraman as the layer<br />
within which knowledge is deployed (in the taking of informed decisions), and by Ward as the level where<br />
business change is to be found. Zachmans system model is echoed strongly in the IMBOK, where<br />
Venkatraman and Ward choose slightly different ideas concerning information (Venkatraman) and<br />
enabling change (Ward). Whilst a detailed evaluation of the ideas would take more space than is<br />
available here, perusal of the table is revealing in that all perspectives align well, even though they see<br />
things from different perspectives. The merit of the IMBOK is that it is strongly aligned to the generation<br />
of value from an investment in information technology, and it is entirely aimed at fulfillment of the<br />
business strategy as the ultimate objective.<br />
Other threads of research have informed the development of the IMBOK. In its embryonic stages there<br />
was a very extensive review of the literature that established the validity of many of the ideas that have<br />
been used (Lambert & Peppard, 1993); the process value adding viewpoint was developed and published<br />
94
Andy Bytheway<br />
(Edwards & Peppard, 1997) and ideas about skills, competencies and capabilities were developed in<br />
conjunction with working business managers (Bytheway & Lambert, 1998). The concept of value came<br />
strongly into focus (Peppard et al., 2001) and is evident in standard works dealing with IS/IT strategy<br />
(Ward & Peppard, 2002).<br />
In South Africa, these ideas were incorporated into a research project that provided new learning material<br />
and new learning opportunities appropriate to the South African context (HICTE, 2003), and the IMBOK<br />
handbook was the result (Bytheway, 2004). It has since been adopted as a standard course text in South<br />
Africa, Europe and North America.<br />
3.1 The IMBOK framework<br />
Pictorially, the IMBOK identifies five domains of management, and four two-way interfaces between<br />
them.<br />
Information<br />
technology<br />
System<br />
projects<br />
System<br />
requirements<br />
Information<br />
system<br />
Business<br />
change<br />
Business<br />
needs<br />
Business<br />
process<br />
Business<br />
operations<br />
Business<br />
capability<br />
Business<br />
benefit<br />
Business<br />
results<br />
Goals and<br />
objectives<br />
Business<br />
strategy<br />
Figure 3: The Information Management Body of Knowledge<br />
It is clear that one of the principal difficulties faced by management is the preservation of the quality and<br />
detail of thinking that passes between the five management domains, and that there is not just one point<br />
of alignment (as is so often argued) but four. If information systems projects do not deliver systems that<br />
are well supported by the technology, if systems do not support business processes, if business<br />
processes do not deliver the performance improvements that are expected, and if those improvements<br />
are not what strategy demanded, then all will be in vain. Hence, the interfaces between the management<br />
domains take on special significance.<br />
The IMBOK framework is useful, because it allows assessment and analysis of the competencies that are<br />
needed to manage the successful delivery of benefits from information technology investments. From the<br />
preceding work, and in particular the review of a very wide range of literature (Lambert & Peppard, 1993),<br />
a set of 144 competencies have been identified that can be organized into nine groups that correspond to<br />
the five management domains in the IMBOK and the four gaps between them. These competencies are<br />
the substance of the survey instruments deployed in this work, and taken together they represent the<br />
capability of the organisation to achieve effective information management.<br />
3.2 The survey instruments<br />
The two survey instruments work at different levels:<br />
The first, with 144 statements each representing one of the 144 competencies; this is referred to as<br />
the "micro" level of working,<br />
The second, with a simplified form using just nine statements to represent each of the nine groups of<br />
competencies; this is referred to as the "macro" level of working.<br />
They are based on Likert-scale worksheets with statements, each representing one competency;<br />
respondents are asked to agree/disagree with the statements in order to indicate their perceptions of the<br />
competency of the organization to do things well. All statements are phrased in a positive sense, so that<br />
"agreement" is good news and "disagreement" is bad news. The focus on perceptions, as opposed to<br />
any absolute measure (such as might be found in using COBIT or ITIL) helps to take more careful<br />
account of people's feelings and emotions. Most previous thinking has revolved around structured,<br />
95
Andy Bytheway<br />
rational, right-brained thinking, but management is as much concerned with perceptions as with<br />
absolutes.<br />
3.3 The micro level<br />
Working at the micro level is of course potentially difficult: no respondent can be expected to respond to<br />
144 individual statements in what is hoped to be a quick but effective survey. Because a large number of<br />
respondents were expected to be involved, worksheets were prepared using randomly selected<br />
statements from the total set of 144, 20 statements at a time - enough worksheets would ensure that<br />
adequate coverage of the nine macro domains would be achieved.<br />
A fragment of such a worksheet is shown below:<br />
Figure 4: A sample worksheet from the survey instrument<br />
The statements are in no particular order and the significance of each statement (in terms of its position<br />
in the framework) is not revealed, so that they appear completely random to the respondent.<br />
Experience at the micro level soon indicated that at the start of an assessment there was a need for a<br />
short, simple but compatible approach. This was undertaken at the macro level.<br />
3.4 The macro level<br />
The simplified macro-level worksheet has just nine statements, one for each of the five domains and one<br />
for each of the four interfaces. Two additions were made: an additional statement concerning culture for<br />
change was included, and a rating of importance of each of the statements (high to low) was included so<br />
as to gather data about the relative importance of the different domains.<br />
Figure 5: The short survey instrument used at the macro level<br />
This short questionnaire proved to be an effective way of dealing with the hundreds of respondents<br />
involved with the financial services company.<br />
3.5 Four cases assessed<br />
The survey has been deployed in four cases, resulting (at the time of writing) in a total of almost 600<br />
responses and some thousands of individual opinions:<br />
A random sample of different businesses (and other organizations) in Cape Town, undertaken at the<br />
micro level.<br />
A purposeful sample of South African Chief Information Officers, also at the micro level.<br />
A purposeful sample of staff in all departments of a real estate agent, also at the micro level.<br />
96
Andy Bytheway<br />
A representative sample of staff working in a large financial services company, this time at the macro<br />
level.<br />
The results provide an interesting insight into the perceptions of different stakeholder groups: IT<br />
specialists, business specialists, administrative staff, management and others. For example, it is found<br />
that IT specialists often under-estimate their capability and fail to realise that their work is actually highly<br />
valued by others. In other cases senior management have a poor view of organisational ability to manage<br />
IT-specific issues such as technology acquisition, IT projects, and business change management.<br />
The paragraphs that follow present selected findings to illustrate these observations, and to show how<br />
the data is collated and presented back to the respondents using an analysis package that has itself<br />
developed over the period of the work.<br />
4. Results<br />
The opinions were analysed using simple, descriptive statistical methods, directed at the production of<br />
"radar charts". It has been found that radar charts are an effective way of presenting complex data to<br />
managers, and because the arrangement of summary measures around the circle is fixed, the audience<br />
for the results becomes used to the shapes that arise, indicating problems and opportunities.<br />
An analysis package has been developed, first to work with data at the micro level and later eanced to<br />
work at the macro level. It allows the selection of different datasets, sectors (of business), organizations<br />
within sector, and so on right down to the gender and education of the individual respondent, as will<br />
become evident. It automatically produces the radar charts (see below) that have ten spokes, or axes, as<br />
follows:<br />
IT The information technology management domain<br />
IT-IS The implementation of information systems<br />
IS The information systems management domain<br />
IS-BP The implementation of business change arising from new systems<br />
BP The business process management domain<br />
BP-BB The delivery of business benefits from improved business processes<br />
BB The business benefit management domain<br />
BB-BS The fulfillment of business strategy through performance<br />
BS The business strategy management domain<br />
(The five principal domains of the IMBOK are written in bold text, the four interfaces in italics).<br />
The tenth vertical "null" axis is reserved for "culture" measures - there was no data for that measure at<br />
this stage in the work.<br />
4.1 The reference sample<br />
The results for all received opinions (1391 in total) is shown in red (and bold); the results for the reference<br />
dataset (1076) are shown in blue (and feint); the standard deviation is shown in the centre in pale green.<br />
The nine measures (from IT through to BS, clockwise) all result in an average assessment of a little less<br />
than "3". This indicates a tendency on the part of respondents to respond to the statements using the<br />
third box from the right (they were coded "7" to "1", left to right):<br />
97
Figure 6: The results for the reference dataset<br />
Andy Bytheway<br />
Because the overall result is based on 1391 opinions, it might be expected that the results would "level<br />
out" (if perceptions of capability are indeed measured in an equitable way by the survey) - it is therefore<br />
pleasing that there is no excessive variation around the nine axes. Also, it can be seen that the standard<br />
deviation of data (on all axes) is in the range 1.3 to 1.5, which suggests that there are no areas of strong<br />
disagreement. It follows that deviations from these broadly-based figures will probably be significant.<br />
The figure below shows two selections that illustrate what is found in the detail:<br />
Figure 7: Results for "education" and "government" sectors selected from the reference dataset<br />
There are significant differences between these results: education is strong on strategy, but government<br />
is very weak; government considers its IT management to be strong but education does not. The<br />
perceived benefits to education are extremely poor, perhaps because of the strength of the strategic<br />
vision and the weakness of the management of the technology that should assist in its realization; the<br />
98
Andy Bytheway<br />
ability of government to see benefits coming out of routine business activity (in the business processes) is<br />
extremely poor.<br />
4.2 Chief Information Officers<br />
At a workshop in Cape Town, about 20 CIOs were asked to assist by completing worksheets at the micro<br />
level.<br />
Figure 8: Results for Chief Information Officers<br />
The results show a distinct optimism, especially when compared with some of the data from senior<br />
managers in the financial sector (see below). Perhaps this is not surprising, a CIO should feel optimistic<br />
about the merits of his work (but it would be a shock if this optimism was in distinct disagreement with the<br />
opinions of others in their organizations).<br />
4.3 Real estate agent<br />
The first in-company assessment concerned about 80 employees in a real estate agency, where the<br />
different views of the different groups proved to be interesting.<br />
Figure 9: Results from a real estate business<br />
Perhaps this is a typical result for a typical organisation, where overall there is a positive and optimistic<br />
view of information management (compared with the reference dataset). After all, this is an information<br />
intensive industry and it is critical to manage information well. This data was gathered at a time when<br />
there was a major upgrade taking place on the company web site, and therefore it is interesting to see<br />
more optimism in the IT department than in the Marketing department. When this analysis was made<br />
99
Andy Bytheway<br />
available to the management, it was possible to work to restore the balance. This is a marketing-led<br />
business, where it is essential that marketing staff share the general confidence in getting the technology<br />
to work well and deliver the benefits.<br />
4.4 Financial services company<br />
Because of the large number of potential respondents in this large financial services company, and<br />
because of the need to undertake the analysis quickly, this was the first application of the "simple"<br />
version of the survey instrument (the macro level). The analysis package being used was further<br />
developed, and so this data was more exhaustively analysed than before. The enhanced analyser allows<br />
for the selection of different departments and job levels, and it presents averages and standard<br />
deviations in detail for both capability and importance.<br />
Just two analyses are presented below: in the main image are the results of the whole set of 224<br />
responses (1906 opinions); overlaid at left-centre are the results for one senior manager who has quite<br />
different views than the overall average.<br />
The overall results show that there is broad concern about the level of capability to deal with IT, systems,<br />
and systems development. The results are less levelled than in the reference dataset. Given the large<br />
number of responses, this is a result that needs further examination. The senior manager has deep<br />
concerns about technology and systems capabilities (this is typical of other senior managers in this<br />
survey) and it is probably time to make sure that something is done to redress this concern.<br />
Having said all that, it is interesting that when compared with the reference dataset, the general results<br />
for this financial services business are good, averaging something between 4 and 5 (compared with less<br />
than 3 in the reference dataset). When presented with this data, the CIO of the financial services<br />
business was pleased, and indicated that more detailed results would be needed in order to guide<br />
management actions intended to improve information management capability. Work on this case<br />
continues.<br />
Figure 10: Analysis of macro data from a financial services business (overall result, with senior manager<br />
for comparison)<br />
100
5. Conclusions<br />
Andy Bytheway<br />
This paper has presented a review of the history of management thinking about information management<br />
issues, it has presented a framework that synthesizes the more interesting ideas, and has shown how<br />
this can be formalized as a new body of knowledge, a framework for thinking, and a capability<br />
assessment instrument.<br />
At the start of the paper, questions were asked that provide a basis for our conclusions:<br />
How can the unavoidable complexities be dealt with?<br />
How can costs be justified in the board room?<br />
Exactly what does the "alignment" of IT and business strategies mean?<br />
First, it is contended that the IMBOK deals with the complexities of information management in an<br />
effective and comprehensible way. It shows that there is no such thing as "alignment of IT and business<br />
strategies", it is necessary to align at four critical interfaces between five domains of management.<br />
Second, the IMBOK provides a means to communicate effectively with senior management. There are no<br />
technical terms, just a clear indication of the dependencies that justify the cost drivers (principally the<br />
acquisition of information technology and services) in light of the systems needed, the improvements to<br />
business processes, and the fulfillment of organizational strategy through appropriate benefits.<br />
Third, it is found that alignment needs a capability to effectively undertake all that is needed within the<br />
five domains of management, and at their interfaces. Space has precluded a detailed discussion of the<br />
144 competencies, but herein lies the secret to success: if any one of these many things is not done well,<br />
then the whole investment - value chain is at risk. No wonder that managing information well is difficult,<br />
and risky.<br />
It is still early days, but this research will continue. There must be more validation of the ideas and careful<br />
consideration of the merits of this kind of perceptions-based survey work. The analysis that is undertaken<br />
here is merely descriptive, and the potential for full inferential analysis of the data (statistically) remains to<br />
be examined.<br />
Acknowledgements<br />
This work would not have been possible without the enthusiastic involvement of staff and students at the<br />
University of the Western Cape and the Cape Technikon (now "CPUT"). Particular acknowledgement is<br />
due to the "ICT in Higher Education" research team led by Derek Keats, and to the Carnegie Corporation<br />
of New York that provided funding on that occasion.<br />
References<br />
Anon (2010) Labour's computer blunders cost £26,326 bn. The Independent. Available from:<br />
[Accessed 24 August 2010].<br />
Anon (2008) Making Sure Tech Is in Step. Wall Street Journal - Eastern Edition, 252 (59), p.B9.<br />
Baets, W. (1992) Aligning information systems with business strategy. The Journal of Strategic Information Systems,<br />
1 (4), pp.205-213.<br />
Beaudry, A. & Pinsonneault, A. (2010) The other side of acceptance: Studying the direct and indirect effects of<br />
emotions on information technology use. MIS Quarterly, 34 (4), pp.689-A3.<br />
Bytheway, A. (1996) Strategic information modelling: some experiences with a new framework. Journal of Applied<br />
Management Studies, 5 (1), pp.17-45.<br />
Bytheway, A. (2004) The Information Management Body of Knowledge. 1st ed. World Wide Web. Available from:<br />
.<br />
Bytheway, A. & Braganza, A. (1997) Process based collaboration: strategies, tactics and benefits. Cranfield UK,<br />
Cranfield School of Management.<br />
Bytheway, A. & Lambert, R. (1998) Organisational competencies for harnessing IS/IT. Available from:<br />
[Accessed 7 May 2010].<br />
Chan, Y.E. & Reich, B.H. (2007) IT alignment: what have we learned? Journal of Information Technology, 22 (4),<br />
pp.297-315.<br />
Chatterji, S. (2007) Bridging business and IT strategies with Enterprise Architecture: Realising the real value of<br />
business-IT alignment. Information Systems Control Journal|, 3, pp.1-2.<br />
Chen, D.Q., Mocker, M., Preston, D.S. & Teubner, A. (2010) Information Systems strategy: Reconceptualization,<br />
measurement and implications. MIS Quarterly, 34 (2), pp.233-A8.<br />
Cross, J. (1995) IT outsourcing : British Petroleum's competitive approach. Harvard Business Review, 73 (3), p.94.<br />
Cross, J. & Earl, M. (1997) Transformation of the IT function at British Petroleum. MIS Quarterly, 21 (4), pp.401-424.<br />
101
Andy Bytheway<br />
DeLone, W. & McLean, E. (1992) Information Systems Success: The Quest for the Dependent Variable. Information<br />
Systems Research, 3 (1), pp.60-95.<br />
Edwards, C. & Peppard, J. (1997) Operationalizing strategy through process. Long Range Planning, 30 (5), pp.753-<br />
767.<br />
Frankel, D.S., Harmon, P., Mukerji, J., Odell, J., Owen, M., Rivett, P., Rosen, M. & Soley, R.M. (2003) The Zachman<br />
Framework and the OMG's model driven architecture. Business Process Trends (Whitepaper). Available from:<br />
[Accessed 22<br />
November 2010].<br />
Henderson, J.C. & Venkatraman, N. (1993) Strategic alignment: leveraging information technology for transforming<br />
organizations. IBM Systems Journal, 32 (1), pp.4-16.<br />
HICTE (2003) ICT in Higher Education. Cape Town, University of the Western Cape.<br />
Kearns, G.S. & Lederer, A.L. (2000) The effect of strategic alignment on the use of IS-based resources for<br />
competitive advantage. The Journal of Strategic Information Systems, 9 (4), pp.265-293.<br />
Lambert, R. & Peppard, J. (1993) IT and new organizational forms: destination but no road map? Journal of Strategic<br />
Information Systems, 2 (3), pp.180-206.<br />
Mingers, J. & Walsham, G. (2010) Toward ethical information systems: The contribution of discourse ethics. MIS<br />
Quarterly, 34 (4), pp.833-854.<br />
Peppard, J., Lambert, R. & Edwards, C. (2001) Whose job is it anyway? Information Systems Journal, 10 (4), pp.291-<br />
322.<br />
Sambamurthy, V., Bharadwaj, A. & Grover, V. (2003) Shaping agility through digital options: Reconceptualizing the<br />
role of IT in contemporary firms. MIS Quarterly, 27 (2), pp.237-264.<br />
Seddon, P.B., Calvert, C. & Yang, S. (2010) A mulit-project model of key factors affecting organizational benefits<br />
from Enterprise Systems. MIS Quarterly, 34 (2), pp.305-A11.<br />
Symons, C. (2005) IT Governance Framework. Forrester. Available from:<br />
[Accessed 22 November 2010].<br />
Tallon, P.P., Kraemer, K.L. & Gurbaxani, V. (2000) Executives' perceptions of the business value of information<br />
technology: a process-oriented approach. Journal of Management Information Systems, 16 (4), pp.145-173.<br />
Uchitelle, L. (1996) What has the computer done for us lately? New York Times, p.21.<br />
Venkatraman, N. (1994) IT-enabled business transformation: from automation to business scope redefinition. Sloan<br />
Management Review, 35 (2), pp.73-87.<br />
Venkatraman, N. (1996) Managing IT resources as a value center. IS Executive Seminar Series, Cranfield School of<br />
Management.<br />
Ward, J. & Daniel, E. (2005) Benefits Management: Delivering Value from IS and IT Investments. Chichester, Wiley.<br />
Ward, J. & Peppard, J. (2002) Strategic Planning for Information Systems (3rd Edition). Chichester, Wiley.<br />
Yayla, A. & Hu, Q. (2009) Antecedents and drivers of IT-Business Strategic Alignment: Empirical validation of a<br />
theoretical model. In: Proc 17th <strong>European</strong> <strong>Conference</strong> on Information Systems. Verona Italy.<br />
Zachman, J.A. (1987) A Framework for Information Systems Architecture. IBM Systems Journal, 26 (3), pp.590-616.<br />
102
A Semiotic Approach to Analyse the Influencing Factors in<br />
Knowledge Transfer<br />
Supaporn Chai-Arayalert 1 and Keiichi Nakata 2<br />
Henley Business School, University of Reading, UK<br />
s.chai-arayalert@pgr.reading.ac.uk<br />
k.nakata@henley.reading.ac.uk<br />
Abstract: With the rapid growth of information and technology, knowledge is a valuable asset in organisation which<br />
has become significant as a strategic resource. Many studies have focused on managing knowledge in<br />
organisations. In particular, knowledge transfer has become a significant issue concerned with the movement of<br />
knowledge across organisational boundaries. It enables the exploitation and application of existing knowledge for<br />
other organisations, reducing the time of creating knowledge, and minimising the cost of organisational learning. In<br />
order to achieve their goals and objectives, organisations need to combine knowledge and performance<br />
measurement tools such as Key Performance Indicators (KPIs) which are powerful agents of organisational change<br />
that help organisations to measure their progress towards achieving their goals. In this paper, we discuss how<br />
organisations can transfer knowledge through KPIs effectively by identifying the relationship between knowledge<br />
transfer and KPIs. In addition, we analyse localisation factors of knowledge transfer based on a semiotic approach:<br />
the organisational containment analysis to scope the influencing localisation factors, and the semiotic framework to<br />
analyse the socio-technical aspects of these factors.<br />
Keywords: knowledge management, knowledge transfer, key performance indicator, semiotics<br />
1. Introduction<br />
Knowledge is a valuable asset in organisations and it has become significant as a strategic resource.<br />
Many studies have focused on managing knowledge in organisations with various processes and<br />
methods (Alavi and Leidner 2001). Knowledge transfer has recently become a significant issue in<br />
knowledge management concerned with the movement of knowledge across the boundaries (Von Krogh<br />
and Roos 1996, Cummings 2003). In addition, it enables the exploitation and application of existing<br />
knowledge for an organisation’s purposes. In order to achieve its goals and objectives, the organisation<br />
needs to combine knowledge and performance measurement tool such as Key Performance Indicators<br />
(KPIs). These are powerful agents of organisational change that help organisations to measure their<br />
progress towards achieving their goals. One of the most important reasons for measuring performance is<br />
to assess the success rate of the business as well as its current performance. KPIs are often used to<br />
help an organisation to achieve its long term organisational goals. Therefore, it is useful to analyse how<br />
performance indicators represent knowledge which in turn could be used for knowledge transfer. In this<br />
paper, we discuss how organisations can transfer KPIs as knowledge effectively by identifying the<br />
relationship between knowledge transfer and KPIs. In addition, we analyse localisation factors in<br />
knowledge transfer based on the semiotic approach. This paper is organised as follows. First, the<br />
concept of knowledge management and knowledge transfer is discussed, followed by a review of the<br />
concepts of KPIs and organisational semiotics. Then, methods to analyse the localisation factors of<br />
knowledge transfer based on semiotics are proposed and followed by the conclusion.<br />
2. Background<br />
In this section, we review knowledge management and its activities and the notion of knowledge transfer,<br />
followed by an overview of key performance indicators and their relation to knowledge.<br />
2.1 Knowledge management<br />
In recent years, there has been a transformation from the resource based view of an organisation into the<br />
knowledge based view (Alavi and Leidner 2001). As a result, more organisations are treating knowledge<br />
as a vital organisational resource alongside other assets, which has led the organisations to recognise<br />
the importance of knowledge management.<br />
In the organisational context, there are many definitions of knowledge which reflects the multiple ways in<br />
which knowledge is considered to be used within organisations. Nonaka and Takeuchi (1995) define<br />
knowledge as a dynamic human process of justifying personal belief toward the truth. Prusak and<br />
Davenport (1998) define knowledge to include experiences, beliefs, values, how we feel, motivation and<br />
information. Additionally, they focuses on the function or purpose of knowledge which is a framework that<br />
people use for evaluating and incorporating new experiences and information through embedded<br />
103
Supaporn Chai-Arayalert and Keiichi Nakata<br />
routines, processes, practices, and norms. Arthur Anderson and APQC (1996) define knowledge as the<br />
relationship between knowledge and organisation as well as information that has value and the collective<br />
experience of the organisation. In the same vein, Demarest (1997) defines knowledge as the actionable<br />
information embodied in work practices, theories-in-action, skills, equipment, processes and heuristics of<br />
firm's employees.<br />
Knowledge management (KM) is to identify and leverage the collective knowledge in an organisation to<br />
help the organisation to be competitive (Von Krogh and Roos 1996). Activities in KM consist of four basic<br />
processes of creating, storing or retrieving, transferring, and applying knowledge (Von Krogh and Roos<br />
1996). Knowledge creation means developing new contents or replacing existing knowledge within<br />
organisational knowledge. Knowledge storage or retrieval refers to organisational memory involving<br />
documentation, database, and codified human knowledge stored in expert systems, etc. Moreover, an<br />
important process in KM is knowledge transfer. Knowledge transfer is the re-creation of a source’s<br />
knowledge-related elements in the recipient (Cummings 2003). Knowledge application is a vital aspect of<br />
knowledge-based theory of the firm, which is the source of competitive advantage. This process is<br />
considered as an activity to apply knowledge to different places and learning from it (Bouthillier and<br />
Shearer 2002).<br />
2.2 Knowledge transfer<br />
Knowledge transfer has become one of the significant KM processes concerned with the movement of<br />
knowledge across the boundaries created by specialised knowledge domains (Carlile 2002, Carlile and<br />
Rebentisch 2003). It is the movement of knowledge from one place, person or ownership to another.<br />
Furthermore, knowledge transfer enables the exploitation and application of existing knowledge for the<br />
organisation’s purposes.<br />
The concept of knowledge transfer based on the communication theory is defined as “a process of<br />
exchange of explicit or tacit knowledge between two agents, during which one agent purposefully<br />
receives and uses the knowledge perceived by another” (Kumar and Ganesh 2009 p.163). Moreover, the<br />
knowledge transfer is “identifying (accessible) knowledge that already exists, acquiring it and<br />
subsequently applying this knowledge to develop new ideas or enhance the existing ideas to make a<br />
process/action faster, better or safer than they would have otherwise been” (Liyanage et al. 2009, p.122).<br />
This suggests that in knowledge transfer, consideration should also be given to the acquisition of<br />
knowledge as well as how it can be exploited to improve organisational activities.<br />
Successful knowledge transfer can be explained as the results of transfer in the receiving end<br />
accumulating or assimilating new knowledge. In other words, success of knowledge transfer can be<br />
assessed by how the underlying knowledge elements have been re-created in the recipient to conform to<br />
those of the source (Cummings 2003). While the barriers of knowledge transfer from source,<br />
communication channel and recipient have been studied (Khamseh and Jolly 2008, McLaughlin 2007,<br />
Jensen and Szulanski 2004), previous studies have not analysed both human and technical aspects<br />
together.<br />
2.3 Key performance indicators and their relation to knowledge<br />
Key performance indicators (KPIs) measure organisation’s overall health and well being (AusIndustry<br />
1995). Importantly, it focuses on the areas of organisation’s performance which are critical for ongoing<br />
and future success. KPIs are driver tools to change an organisation and its culture; at the same time,<br />
organisations can apply KPIs as an improvement strategy where they will need to change (Swan and<br />
Kyng 2004). KPIs are commonly used to help an organisation define and evaluate its success, in terms of<br />
making progress in long term organisational goals and toward knowledge-based goals which are difficult<br />
to quantify. The differences in KPIs depend on the nature and characteristics of the organisation including<br />
its strategies and goals. This section examines the concept of KPIs and the relationship between KPIs<br />
and organisational knowledge.<br />
2.3.1 The concept of key performance indicators<br />
KPIs are often defined as the quantifiable metric which reflects the organisational performance in<br />
achieving its goals and objectives, which in turn it reflects strategic drivers and critical success factors of<br />
organisational activities and processes (Parmenter 2007, Thanyaphut 2006). Furthermore, KPIs are<br />
indicators at the operational and tactical level obtained by breaking down strategic objective decision<br />
104
Supaporn Chai-Arayalert and Keiichi Nakata<br />
making and act as pointers for monitoring the effect of implementation of macro strategic decision (Zhou<br />
et al. 2006). KPIs can align to units, departments and individuals of an organisation. An organisation can<br />
have one or more KPIs to measure its outcomes. KPIs have to be created based on the organisational<br />
vision, mission, and strategy. Therefore, KPIs usually differ in each organisation due to the different goals<br />
and strategies. Organisations can utilise KPIs to monitor progress, communicate performance results,<br />
and determine priorities (Promput et al. 2006). Organisations can be monitored by setting KPIs, and KPIs<br />
are also created to identify the organisational position compared to its competitors and to communicate<br />
performance results within the organisations. KPIs show goal achievement results based on facts, not on<br />
emotion or supposition so that employees will accept and understand the results. To determine work<br />
priorities, KPIs show the target employees need to achieve and helps them set the priorities of their work.<br />
2.3.2 Organisational knowledge and key performance indicators<br />
As described earlier, knowledge management (KM) is recognised as a part of strategy to improve<br />
business performance. KPIs complement performance measurement and they are a part of<br />
organisational knowledge which is most vital in improving performance. Therefore, it has been suggested<br />
that managing knowledge could be incorporated into KPIs, and other performance measurement<br />
approaches (Okkonen 2004). KM consists of activities and tools which are used to improve core<br />
competencies and to make the use of information more efficiently in organisational context (Okkonen<br />
2004). A relationship between KM and performance measurement has also been identified (Okkonen et<br />
al. 2002), i.e., performance measures provide the information needs in KM activities. In addition, the<br />
process of formulating and implementing measures or indicators may be through gathering and<br />
implementing organisational knowledge. As KPIs depend on objectives or goals of organisational<br />
overview and break down strategic knowledge into lower levels or operational knowledge in organisation,<br />
they can be considered as explicit knowledge due to the fact that they are captured as written<br />
documents, forms, tables etc. (Promput et al. 2006).<br />
There are a number of limitations of developing and using KPIs. Firstly, the development KPIs is specific<br />
to an organisation and uses information from different sources. The relevant KPIs can be found in internal<br />
or external documents within the organisation, e.g., policies, mission statements, business plan, job<br />
descriptions, and laws. Therefore, acquisition of knowledge of KPIs development may be incomplete and<br />
vague (Popova and Sharpanskykh, 2010). Secondly, KPIs can be developed by using organisational<br />
knowledge, time, and needs experts to facilitate the KPI development process. However, some<br />
organisations are not conveniently set up to capture knowledge, find it time consuming, and sometimes<br />
lack expertise (Popova and Sharpanskykh, 2010). Thirdly, although KPIs are useful in performance<br />
measurement due to the fact that the organisation’s objectives are broken down into KPIs that are<br />
applied to measure some aspects against targets in order to improve the problems areas, the KPI<br />
implementation can also suffer from the same reasons performance measurement initiatives fail, as<br />
argued by Neely and Bourne (2000) and Masayna et al (2007). They state that these initiatives fail<br />
because they do not reflect the strategies and help people to understand what the organisation’s priorities<br />
are. Another reason is the lack of infrastructure which is focused on the management of data in<br />
developing and using KPIs (Neely and Bourne, 2000). This problem of data management in the<br />
organisation is that they are often unrelated, unlinked, and inconsistent (Masayna et al., 2007).<br />
3. Analysing localisation factors of knowledge transfer<br />
This section discusses the relationship between KPIs and organisational knowledge, the concept of<br />
knowledge transfer through KPIs, and factors that influence this form of knowledge transfer. It begins by<br />
describing organisational semiotics which is applied for problem analysis, especially the semiosis model,<br />
the organisational containment analysis, and the semiotic framework. This is followed by an exploration<br />
of localisation factors based on the semiosis model. Finally, we analyse these localisation factors in<br />
knowledge transfer using organisational containment analysis and semiotic framework.<br />
3.1 Organisational semiotics<br />
Semiotics presents signs as things in a world of objects, actions and relationships. As a branch of<br />
semiotics, organisational semiotics (OS) is a discipline which aims to study the nature, functions,<br />
characteristics and effects of information and communication within organisational contexts (Liu 2000).<br />
OS helps us understand the cooperative workings and interactions among individuals, and between<br />
human beings and technology. It defines organisations as systems where signs are created and used for<br />
communication and business purposes (Liu et al. 1999). It deals with the use of signs and the<br />
construction of shared meanings within and among organisations (Liu 2000).<br />
105
3.2 Peirce’s triadic model of semiosis<br />
Supaporn Chai-Arayalert and Keiichi Nakata<br />
Semiosis, introduced by Charles Sanders Peirce (1839–1914), is the process of constructing meaning<br />
from represented signs. The process is shown by Peirce’s Triadic Model of semiosis. Semiosis contains<br />
sign, object and interpretant (Liu 2000). Sign is the signification without reference to anything other than<br />
itself. The object is the signification in relation to something else. The interpretant is shown as to meditate<br />
the relationship and helps establish the mapping between the sign and the object (presented by the<br />
dotted line in figure 1). The sign is related to its referent or the object with the assistance of the<br />
interpretant which is the interpretation process (from sign to object). The sign can be understood or<br />
misunderstood in different ways depending on the interpretant. The semiosis model can assist the<br />
analysis of knowledge transfer, as the interactions between the sign, object and interpretant.<br />
Figure 1: Peirce’s triadic model of semiosis (Liu 2000)<br />
Based on this view, we can analyse semiosis gaps in knowledge transfer processes. There are two<br />
possible gaps: representation gaps and interpretation gaps (Tan et al. 2003). ‘Representation’ is to<br />
describe something or illustration of a sign. A representation gap occurs when the two corresponding<br />
signs referring to the same object are not aligned. Secondly, an interpretation gap is the displacement of<br />
objects or description leading to a mismatch of inference. This gap happens when understanding of<br />
object differs and results in a distorted understanding of intended meaning. Employing the semiosis<br />
model can explain gaps that might occur in knowledge transfer.<br />
The figure 2 illustrates the notion of semiosis gaps.<br />
Figure 2: The semiosis gaps<br />
106
Supaporn Chai-Arayalert and Keiichi Nakata<br />
The semiosis model presents two significant gaps occurring in the sign (S) and the interpretant (I). The<br />
sign represents KPIs which an organisation intends to use as a vehicle to transfer knowledge to another<br />
organisation. In this model, S1 represents developed KPIs from source’s organisation and S2 occurs when<br />
S1 is adapted through adaptation procedures that are concerned with localisation barriers of transferring<br />
KPIs. The object (O) is knowledge. In the interpretant (I) is the process of relating KPIs to organisational<br />
knowledge. We adopt the view that knowledge transfer can be carried out through adapting KPIs of one<br />
organisation to another, the transfer of knowledge to external organisation and different countries can be<br />
influenced by factors which affect the success of transfer. These factors are represented by gaps<br />
between the interpretant (I1) and interpretant (I2). In the next section, we use the semiotic framework to<br />
identify the vital localisation factors.<br />
3.3 Analysis of influencing localisation factors based on semiotic approach<br />
This analysis of localisation factors that influence knowledge transfer begins with collecting data from<br />
literature, after which the data is scoped by organisation containment analysis and analysed by the<br />
semiotic framework.<br />
3.3.1 Scoping the influencing localisation factors based on organisational containment analysis<br />
The organisational containment analysis (figure 3) presents a model of organisation as a system in terms<br />
of informal, formal and technical layers (Liu 2000). The informal layer covers the whole business<br />
organisation and it interacts directly with the external environment and context. It focuses on subcultures,<br />
behaviours, beliefs and practices. The activities of this layer can be formalised in the formal<br />
layer or as business rules. The formal layer consists of forms, rules, and procedures. These formalisms<br />
are developed to clarify and replace meanings, intentions, commitments, and responsibilities of the<br />
informal layer. In the technical layer, technical systems are designed to automate parts of formalised<br />
aspects, which in turn rely on the informal aspect.<br />
Figure 3: An organisational containment analysis for scoping localisation factors<br />
Here we attempt to identify the localisation factors that influence knowledge transfer. The informal layer<br />
includes the national culture and norms (Liyanage et al. 2009, Kluge et al., 200, Barson et al. 2000); the<br />
experience of source and recipient organisation (Liyanage et al. 2009, Khamseh and Jolly 2008, Jensen<br />
and Szulanski 2004); the ability to identify, assimilate, transform and apply external knowledge or<br />
absorptive capacity (Khamseh and Jolly 2008); and language barriers (Cranefield and Yoong 2007,<br />
Jensen and Szulanski 2004). In the formal layer, the factors related to the characteristics of target<br />
organisations including strategy and objectives (AusIndustry 1995), organisational structure (Cranefield<br />
107
Supaporn Chai-Arayalert and Keiichi Nakata<br />
and Yoong 2007, McLaughlin 2007, Barson et al. 2000, AusIndustry 1995), technical terms (Cranefield<br />
and Yoong 2007, McLaughlin 2007), regulations and rules (McLaughlin 2007), standards (McLaughlin<br />
2007), policies (AusIndustry 1995). As described above, the technical layer refers to specific factors of<br />
information technology involving hardware, software and communication channels (Barson et al. 2000,<br />
Wathne et al. 1996, Khamseh and Jolly 2008, Cranefield and Yoong 2007).<br />
3.3.2 Analysing the influencing localisation factors based on the semiotic framework<br />
Based on the scoping of localisation factors above, we apply the semiotic framework to analyse the<br />
localisation factors which influence the transfer of knowledge through KPIs. The semiotic framework was<br />
developed in order to understand different dimensions of signs (Stamper 1996). This framework is<br />
divided as human and technical aspects. The technical aspects are the physical world, the empiric layer<br />
and the syntactic layer. The upper three layers address the human aspects in terms of the semantic<br />
layer, the pragmatic layer and the social world (table 1).<br />
Table 1: The semiotic framework for analysing localisation factors of knowledge transfer<br />
The physical world is concerned with the physical aspects of signs at the level of signals and marks (Liu<br />
2000). The localisation factors in this layer focus on channel infrastructures, technologies and resources.<br />
The channel infrastructures can facilitate the knowledge transfer process (Wathne et al. 1996). However,<br />
there are barriers to transfer in terms of digital divide in some countries. Therefore, some countries<br />
cannot conveniently transfer and adapt knowledge from another organisation. In addition, the barriers of<br />
inter-and-intra knowledge transfer are technologies in term of existing resource, available technology and<br />
legacy systems (Barson et al., 2000).<br />
The empiric layer is statistical properties of signs when different physical media and devices are used<br />
(Liu 2000). This layer is concerned with the statistical behaviour of signs and it can be matched to the<br />
statistical characteristic of the media. In empirics view, information is viewed as a stream of signals which<br />
is transmitted from sender to receiver. Therefore, the analysis of localisation factors is related to video-<br />
108
Supaporn Chai-Arayalert and Keiichi Nakata<br />
conference, telephone, internet, documents, and face-to-face meetings(Cranefield and Yoong 2007)<br />
which have an effect on transferring knowledge.<br />
The syntactic layer is concerned with the complex structure of a language (Liu 2000). Study of this layer<br />
will answer question as to how signs are structured and used in language. We identify the localisation<br />
factors by looking at the characteristics of knowledge, i.e., metrics, variables, and formula (McLaughlin<br />
2007, Khamseh and Jolly 2008, Zhihong 2008). In addition, the procedure and/or process, software and<br />
database technology are considered as localisation factors which affect both the source and the recipient<br />
(Khamseh and Jolly 2008).<br />
The semantic layer is considered with the relationship between a sign and what it refers to (Liu 2000).<br />
The transfer and adaptation of knowledge through KPIs between organisations of different backgrounds<br />
need to share the same understanding and meaning of knowledge. The knowledge transfer from the<br />
source to the recipient cannot take place, if they do not understand and share the meanings when<br />
adopting and adapting KPIs. The localisation factors in this layer are technical terms, specific definitions,<br />
standards and language (Cranefield and Yoong 2007).<br />
The pragmatic layer focuses on the relationships between signs and behaviours of agents (Liu 2000).<br />
This layer considers signs that can be used for certain purposes or objectives. In the case of transferring<br />
knowledge through KPIs to difference countries, although organisations may have a similar domain, they<br />
may differ in their context. Therefore, not all indicators can be transferred and applied as objectives and<br />
strategies in another organisation. It depends on the organisational context such as types of organisation,<br />
organisational structure, strategies, and objectives which directly influence the behaviours of adapting<br />
indicators (AusIndustry 1995, Cranefield and Yoong 2007, McLaughlin 2007, Jensen and Szulanski<br />
2004). The difference in organisational context can make it difficult to transfer knowledge between<br />
organisations.<br />
The social world is the effect of the use of signs in social and human affairs (Liu 2000). In particular, it<br />
implicates social norms, beliefs, expectations, commitments, functions, contracts, law, and culture (Filipe<br />
and Liu 2000). KPIs can be used to measure performance in organisations to achieve objectives and<br />
goals. They are applied to measure the performance against targets which are intended to solve or<br />
improve areas of weakness. Knowledge can be embedded in KPIs which can be considered as<br />
representation of knowledge which is generated from several sources. Formalising KPIs requires tangible<br />
and intangible resources. Tangible resource includes reports, manuals, stories, and records etc. and<br />
intangible resource are culture, experience, norms, and commitments. The knowledge transfer through<br />
KPIs occurs in the social world layer; however, this is not a trivial process for the limitation of KPIs<br />
development mentioned above (Alwaer and Clements-Croome 2010). The localisation factors can be<br />
captured as the interpretant gap in the semiosis model, and includes national culture, laws and<br />
regulations, experiences, and absorptive capacity. National culture constrains knowledge adaptation as<br />
well as governs the process of transfer (Davenport and Prusak 1998). As such, national culture is clearly<br />
a significant and complex issue in knowledge transfer. Others suggest that knowledge transfer process<br />
cannot succeed if the source and recipient are unwilling to share knowledge because of cultural<br />
difficulties (Liyanage et al. 2009). Experience is found as one of the important factors. A number of<br />
studies analysed experience as a key factor impacting on knowledge transfer. Cranefield and Yoong<br />
(2007) identified the key factors that impacted on knowledge transfer and the result showed the prior<br />
experience had an impact on knowledge transfer during the first stage of transfer. Absorptive capacity is<br />
the ability to identify, assimilate, transform, and apply valuable external knowledge (Cohen and Levinthal<br />
1990). Therefore, the recipient may be unable to utilise external sources of knowledge because they may<br />
lack absorptive capacity which affects their ability to value, assimilate and apply new knowledge<br />
(Szulanski 1996).<br />
4. Conclusion<br />
An effective acquisition and management of knowledge has become a competitive advantage in the use<br />
of organisational resources. The organisational knowledge can be captured through learning, which<br />
includes the transfer of knowledge among organisations. However, the transfer of knowledge is not<br />
straightforward as it depends not just on the nature of knowledge itself but also on the process of<br />
acquiring and assimilating it. This study explored methods of analysis and identification of localisation<br />
factors. It aimed to identify the relationships between knowledge transfer and its vehicles such as KPIs<br />
through the use of semiotic approach. The localisation factors are also identified using the semiotic<br />
approach. The outcomes of this paper are as follows. Firstly, we used organisational containment<br />
109
Supaporn Chai-Arayalert and Keiichi Nakata<br />
analysis, which analyses an organisation as a system made up of informal, formal and technical layers, to<br />
scope the influencing localisation factors when knowledge transfer is carried out through KPIs. As a<br />
result, the localisation factors with the informal layer have been identified as the culture and norms; the<br />
experiences of source and recipient organisation; the ability to identify, assimilate, transform and apply<br />
external knowledge or absorptive capacity; and language barriers. In the formal layer, the factors are<br />
related to the characteristics of target organisations including strategy, objectives, organisational<br />
structure, technical terms, definitions, language, rules, standards, policies, and laws. The technical layer<br />
refers to specific factors of information technology involving hardware, software and communication<br />
infrastructure. Secondly, the influencing localisation factors were divided into human and technical<br />
aspects based on the semiotic framework. In the future work, we plan extend to this research by<br />
developing a conceptual framework of knowledge transfer based on the result of this analysis, and<br />
validate it through case studies in the inter-organisational and inter-cultural knowledge transfer.<br />
References<br />
Alavi, M. and Leidner, D. E. (2001) 'Review: Knowledge management and knowledge management systems:<br />
Conceptual foundations and research issues', MIS quarterly, pp.107-136.<br />
AusIndustry (1995) Key performance indicators manual : a practical guide for the best practice development,<br />
implementation and use of KPIs., Business & Professional Publishing..<br />
APQC. (1996) The Knowledge Management Assement Tool, Arthur Anderson and The American Productivity &<br />
Quality Center.<br />
Alwaer, H. and Clements-Croome, D. J. (2010) 'Key performance indicators (KPIs) and priority setting in using the<br />
multi-attribute approach for assessing sustainable intelligent buildings', Building and Environment, 45(4), 799-<br />
807.<br />
Barson, R. J., Foster, G., Struck, T., Ratchev, S., Pawar, K., Weber, F. and Wunram, M. (2000) Inter-and intra-<br />
Organisational barriers to sharing knowledge in the extended supply-chain,IOS Press.<br />
Bouthillier, F. and Shearer, K. (2002) 'Understanding knowledge management and information management: the<br />
need for an empirical perspective', Information research, 8(1).<br />
Carlile, P. (2002) 'Transferring, Translating and Transforming: An Integrative Framework Managing Knowledge<br />
across Boundaries', Effective Enterprise Learning, pp. 555-568.<br />
Carlile, P. R. and Rebentisch, E. S. (2003) 'Into the black box: the knowledge transformation cycle', Management<br />
Science, 49(9), pp.1180-1195.<br />
Carrillo, P. M., Robinson, H. S., Anumba, C. J. and Al-Ghassani, A. M. (2003) 'IMPaKT: A framework for linking<br />
knowledge management to business performance', Electronic Journal of Knowledge Management, 1(1), pp.1-<br />
12.<br />
Cohen, W. M. and Levinthal, D. A. (1990) 'Absorptive capacity: a new perspective on learning and innovation',<br />
Administrative science quarterly, 35(1), pp 128-152.<br />
Cranefield, J. and Yoong, P. (2007) 'Organisational factors affecting inter-organisational knowledge transfer in the<br />
new zealand state sector: a case study', Journal of Organizational Virtualness, 1, pp. 9-33.<br />
Cummings, J. (2003) Knowledge Sharing: a review of the literature, The world Bank Operations Evaluation<br />
Department, The World Bank, Washington D.C.<br />
Davenport, D. L. and Hplapple, C. W. (2006) 'Knowledge organizations' in Schwartz, D., ed. Encyclopedia of<br />
knowledge management,Idea Group Reference, pp.451-458.<br />
Davenport, T. and Prusak, L. (1998) Working knowledge: How organizations manage what they know, Harvard<br />
Business School.<br />
Demarest, M. (1997) 'Understanding knowledge management', Long Range Planning, 30(3), pp. 374-384<br />
Filipe, J. and Liu, K. (2000) The EDA Model: An Organizational Semiotics Perspective To Norm-Based Agent Design,<br />
Workshop on Norms and Institutions in Multi-agent systems at Normative agent.<br />
Jensen, R. and Szulanski, G. (2004) 'Stickiness and the Adaptation of Organizational Practices in Cross-Border<br />
Knowledge Transfers', Journal of International Business Studies, 35(6), pp.508-524.<br />
Khamseh, H. M. and Jolly, D. R. (2008) 'Knowledge transfer in alliances: determinant factors', Journal of Knowledge<br />
Management, 12(1), pp.37-50.<br />
Kluge, J., W. Stein, and T. Licht. (2001) ‘Knowledge unplugged: The Mckinsey & Company global survey on<br />
knowledge management’, Palgrave Macmillan.<br />
Kuhn, O. and Abecker, A. (1997) 'Corporate memories for knowledge management in industrial practice: Prospects<br />
and challenges',Journal of Universal Computer Science, 3(8), pp.929-954.<br />
Kumar, J. A. and Ganesh, L. S. (2009) 'Research on knowledge transfer in organizations: a morphology', Journal of<br />
Knowledge Managemnt, 13(4), pp.161-174.<br />
Liu, K. (2000) Semiotics in information systems engineering, Cambridge:Cambridge University Press.<br />
Liu, K., Alderson, A., Shah, H., Sharp, B. and Dix, A. (1999) 'Applying semiotic methods to requirements recovery',<br />
Methodologies for Developing and Managing Emerging Technology-Based Information Systems, pp.142-152.<br />
Liyanage, C., Elhag, T., Ballal, T. and Li, Q. (2009) 'Knowledge communication and translation–a knowledge transfer<br />
model', Journal of Knowledge Management, 13(3), 118-131.<br />
Masayna, V., Koronios, A., Gao, J. and Gendron, M. (2007) 'Data quality and KPIs: a link to be established', in The<br />
2nd World Congress on Engineering Asset Management (EAM) and The 4th International <strong>Conference</strong> on<br />
Condition Monitoring,<br />
110
Supaporn Chai-Arayalert and Keiichi Nakata<br />
McLaughlin, S. (2007) Identifying knowledge transfer barriers within a complex supply chain organization.,<br />
unpublished Ph.D.thesis University of Glasgow<br />
Neely, A. and Bourne, M. (2000) 'Why measurement initiatives fail', Measuring Business Excellence, 4(4), 3-7.<br />
Nonaka, I. and Takeuchi, H. (1995) 'The knowledge-creating company: How Japanese companies create the<br />
dynamics of innovation'<br />
Okkonen, J. (2004) The Use of Performance Measurement in Knowledge Work Context, unpublished Ph.D.thesis<br />
University of Tampere.<br />
Okkonen, J., Pirttim ki, V., Hannula, M. and L nnqvist, A. (2002) 'Triangle of Business Intelligence, Performance<br />
Measurement and Knowledge Management', in 2nd Annual <strong>Conference</strong> on Innovative Research in<br />
Management, Stockhome, Sweden, 2002, pp.1-14<br />
Parmenter, D. (2007) Key performance indicators: developing, implementing, and using winning KPIs, Wiley.<br />
Popova, V. and Sharpanskykh, A. (2010) 'Modeling organizational performance indicators', Information Systems,<br />
35(4), 505-527.<br />
Promput, W., Pingsuttiwong, S. and Lerdmahakit, C. (2006) KPIs principle to practice, Bangkok:Siriwattana Interprint.<br />
Stamper, R. (1996) 'Signs, Norms and Information Systems, Signs at Work' in al, Holmqvist B. et al, ed.,Berlin:<br />
Walter de Gruyter.<br />
Swan, W. and Kyng, E. (2004) An Introduction to Key Performance Indicators, Manchester, UK:Centre for<br />
Construction Innovation.<br />
Szulanski, G. (1996) 'Exploring internal stickiness: Impediments to the transfer of best practice within the firm',<br />
Strategic management journal, 17(1), pp.27-43.<br />
Tan, S., Liu, K. and Xie, Z. (2003) A Semiotic Approach to Organisational Modelling Using Norm Analysis,<strong>6th</strong><br />
International <strong>Conference</strong> on Enterprise Information Systems,Porto, Portugal,.<br />
Thanyaphut, A. W., P. (2006) 'Key performance indicators: Strategic into action', Journal of Quality Guarantee, 7(1),<br />
pp.1-12.<br />
Von Krogh, G. and Roos, J. (1996) Managing knowledge: perspectives on cooperation and<br />
competition,Representationism: the Traditional Approach to Cooperative Strategies, London:Sage Publications<br />
Ltd.<br />
Wathne, K., Roos, J. and von Krogh, G. (1996) 'Towards a theory of knowledge transfer in a cooperative context',<br />
Managing Knowledge-Perspectives on cooperation and competition, pp.55-81.<br />
Watson, S. and Hewett, K. (2006) 'A Multi-Theoretical Model of Knowledge Transfer in Organizations: Determinants<br />
of Knowledge Contribution and Knowledge Reuse*', Journal of Management Studies, 43(2), pp.141-173.<br />
Zhihong, L. (2008) A Study on the Factors Influencing Knowledge Transfer among Organizations in Strategic<br />
Alliances,pp.822-825.<br />
Zhou, N., Chen, X.-y. and wang, J.-h. (2006) 'A Study on the design method of KPI system in Chinese<br />
communication enterprises', Journal of Communication and Computer, 6(12), pp.63-69.<br />
111
Exploring the Emotional Exhaustion of Healthcare Providers<br />
Using an In-Hospital Employee Communication Network<br />
Cheng-Yi Chiang 1 , Ya-Ying Kuo 1 , Ying-Hui Hou 2 , and I-Chiu Chang 1<br />
1<br />
Institute of Healthcare Information Management, National Chung Cheng<br />
University, Chiayi, Taiwan<br />
2<br />
Kai Nan University, Taoyuan, Taiwan<br />
g98556002@mis.ccu.edu.tw<br />
yaying.kuo@gmail.com<br />
yhhou@mail.knu.edu.tw<br />
misicc.pig@gmail.com<br />
Abstract: Competition between hospitals in Taiwan is often fierce. Information transparency has replaced the former<br />
information asymmetry situation. Healthcare service providers will have to attract patients in new ways. Healthcare<br />
service providers need to have a positive attitude. The problems of emotional labor and emotional exhaustion of the<br />
front line healthcare givers cannot be over emphasized. It’s important for them to have an emotional support system.<br />
Lifestyles have changed with the development of the Internet and Information Technology. Establishment of an<br />
employee communication network (ECN) in hospitals is essential. Due to the increasing emotional exhaustion of<br />
healthcare providers, this study focuses on one hospital with in-hospital ECN to investigate the extent of healthcare<br />
givers’ emotional exhaustion. This study will further discuss the usage and perceived usefulness of healthcare givers<br />
who have used the ECN. We also hope to ease healthcare givers’ occupational burnout.<br />
Keywords: healthcare providers, nurse, in-hospital employee communication network, emotional exhaustion,<br />
perceived usefulness<br />
1. Introduction<br />
In the 43 major events of Social Readjustment Rating Scale (SRRS) edited by Homels and Ruhe in 1967,<br />
three major events within the top five events are hospital-related ones. The first one is “Death of a<br />
Spouse” with a score of 86. The second one is “Death of Close Family Member” with a score of 77. The<br />
fifth one is “Personal Injury or Illness” with a score of 61. The three events are sources of daily life stress<br />
that will not only influence individuals and families, but also trigger more demands and direct emotional<br />
reactions. Front line healthcare givers who need to face patients or patients’ family members usually bear<br />
considerable blame or negative emotional burdens.<br />
Bolton (2000) mentioned that nurses' 'emotional labor” is hard and productive work and should be valued<br />
in the same way as physical or technical labor. The foundation for organization to benefit comes from<br />
healthcare givers’ emotional labor supply to customers. Organizations are trying to modify and control<br />
staff to show emotions that deliver the expectation of the organization to patients in order to improve<br />
patient satisfaction (Hochschild, 1983).<br />
While healthcare providers face high frequency IT contact and upsurging emotional labors to deal with<br />
daily works, they can employ the virtual in-hospital ECN to: (1) communicate or vent their emotions<br />
through ECN to express negative feelings at the moment and transform emotions. (2) Such<br />
communication methods can temporarily release healthcare givers’ highly emotional labors, so that<br />
negative emotions and effects of healthcare givers can be solved. The stress of healthcare givers can be<br />
reduced to avoid emotional exhaustion. The two points mentioned above are critical keys of this study. If<br />
these two points can be achieved, this can be viewed as a tremendous contribution. For medical<br />
institution executives, they should consider the necessity of building ECN to eliminate job-quitting<br />
problems caused by emotional exhaustion.<br />
2. Literature review<br />
2.1 Emotional labor<br />
Hochschild (1983) first defined “emotional labor” as “the management of feeling to create a publicly<br />
observable facial and bodily display; emotional labor is sold for a wage and therefore has exchange<br />
value.”<br />
112
Cheng-Yi Chiang et al.<br />
Emotional labor must correspond to the following three characteristics: First, workers must create a<br />
publicly observable face-to-face interaction. Second, workers must have an emotional status to effect<br />
others. Third, emotional labors shown by workers are restrained by employees (Wharton, 1993). For<br />
service workers, emotional work has become part of the service (Wichoroski, 1994). Emotional work can<br />
be mostly seen in front-line workers (Ashforth and Humphrey, 1993).<br />
Hochschild (1983) categorized American occupations into high emotion labor occupation and low<br />
emotional labor. Nurses are regarded as professional, technical workers.<br />
2.2 Emotional exhaustion<br />
The definition of Emotional Exhaustion is: a certain reaction which is pressure-related. It is an<br />
overextended emotional phenomenon while interacting with customers or clients. Emotional exhaustion is<br />
often considered as a key element of job stress (Saxton, Phillips and Blakeney, 1991). Jackson, Schwab<br />
and Schuler (1986) believed that workers are involved too much in their jobs. They cited Maslach’s study<br />
and pointed that those work roles involved too much on jobs would have a higher degree in emotional<br />
exhaustion. Such as: workers need to have face-to-face interaction with customers, high emotional<br />
intensity workers or long-term interaction with customers will have a higher degree in emotional<br />
exhaustion. Morris and Feldman (1996) proposed several issues: “frequency of emotional display has<br />
positive correlation on emotional exhaustion”, “attentiveness to emotional display rules has positive<br />
correlation on emotional exhaustion” and “variety of emotions has positive correlation on emotional<br />
exhaustion.”<br />
According to Wharton’s (1993) study focused on bank tellers and hospital workers, it was indicated that<br />
for those who engaged in emotional works have potentially negative consequences of emotional<br />
exhaustion. It depends on authority of job, job involvement and self-control. For high emotional labor<br />
workers with higher authority and self-control are less likely to have emotional exhaustion. Workers with<br />
higher working hours and seniority are likely to have emotional exhaustion. For low emotional labor<br />
workers with higher job involvement, they are less likely to have emotional exhaustion.<br />
The influence of emotional exhaustion is tremendous. Many scholars hope to find out the extraneous<br />
variables of emotional exhaustion to provide a better suggestion and to ease emotional exhaustion<br />
occurrence. Research shows that social supervisory support moderated emotional exhaustion and had a<br />
negative correlation to emotional exhaustion (Dignam et al, 1986; Kirmeyer and Dougherty, 1988).<br />
2.3 Technology Acceptance Model<br />
Technology Acceptance Model (TAM) is designed to understand new information system acceptance<br />
behaviors of IT users. Davis et al. (1989) utilized the “Theory of Reasoned Action” (TRA) as a foundation<br />
and applied information system usage to propose the “TAM.” TAM is mainly to explore the relationship of<br />
the “Perceived Usefulness”, “Perceived Ease of Use”, “Intention to Use” and “Actual System Use” of<br />
information users’ opinions toward certain information systems. TAM claims that “Actual System Use” is<br />
directly determined by “Behavior Intention to Use”; “Behavior Intention to Use” is determined by “Attitude<br />
toward Using” and “Perceived Usefulness”. External variables include: system usage, training, user<br />
intervention of system design process, nature of system design. These variables will effect users’<br />
behavior intentions to use and actual system use indirectly. TRA and TAM both consider that belief will<br />
influence the attitudes of individuals; attitude will influence their willingness and further influence the<br />
behavior performance of individuals.<br />
2.3.1 Hypotheses of TAM<br />
Hypotheses of TAM are as follows (Davis, 1989; Agarwal and Prasad, 1999):<br />
1. Technology usage behavior of humans can be predicted through user willingness of technology<br />
usage. Individual behavior intentions on new technology will be influenced by attitude toward using<br />
new technology.<br />
2. Two determinants of influencing the attitude toward using new technology are “Perceived<br />
Usefulness” and “Perceived Ease of Use.” “Perceived Usefulness” is the key factor for individuals to<br />
accept new technology. “Perceived Ease of Use” is the secondary influencing factor for individuals to<br />
accept new technology.<br />
113
Cheng-Yi Chiang et al.<br />
3. “Perceived Usefulness” influences not only “Attitude toward Using” but also “Intention to Use.”<br />
4. “Perceived Ease of Use” will effect individuals’ perception on “Perceived Usefulness.” The higher<br />
degree of “Perceived Ease of Use” is, the higher individuals’ perception on “Perceived Usefulness”<br />
toward new technology will be.<br />
5.”External Variables” means external factors that will potentially influence users “Perceived<br />
Usefulness” and “Perceived Ease of Use.” Such as individual differences, system character and<br />
environments. These external variables will influence user perceptions on “Perceived Usefulness”<br />
and “Perceived Ease of Use.” (Venkatesh and Davis, 1996; Szajna, 1996).<br />
2.3.2 Revised Technology Acceptance Model<br />
Not many concrete studies were done in understanding the relationship among “Perceived Usefulness”,<br />
“Attitude toward Using”, “Intention to Use”, and “System use”, which resulted in debates of many<br />
scholars. Scholars like Adams et al, (1992) and Straub et al, (1996) have proposed the revised model of<br />
TAM. Their findings suggest that the “Behavior Intention to Use” of TAM should be removed. Igbaria<br />
(1997) further applied the revised TAM on the issue of technology acceptance in small enterprises. The<br />
IT discussed in this study (in-hospital ECN) is a website which is built by the hospital itself. It is not<br />
complicated as other information systems and healthcare givers are capable of operating Internet.<br />
Hence, this study only takes “Perceived Usefulness” as its research variable.<br />
3. Research method<br />
3.1 Objectives and sampling method<br />
The research method of this study is a multi-case study. Questionnaires were dispatched to one juridical<br />
person medical center and healthcare givers of regional hospitals. Healthcare givers include all-levels of<br />
staffs and executives. The sampling method of this study is a stratified random sampling. Distribution of<br />
questionnaires is calculated by the ratio of healthcare givers in the three medical institutions. 80, 850, and<br />
70 questionnaires were given to the three medical institutions, respectively. Method of questionnaire<br />
distribution: A workshop for this questionnaire will be held in one week before the collection date.<br />
Questionnaires will be given to all divisions by this study’s researcher to ask for lower nursing managers’<br />
assistance in the questionnaire distribution. When questionnaires are completed, they will be sent back<br />
by lower nursing managers or be collected by researcher of this study.<br />
3.2 Research framework and hypotheses<br />
This study hopes to find out the relativity of emotional exhaustion through perceived usefulness, usage<br />
pattern and motivation of using the in-hospital ECN. The structure of this study is shown as Figure 1.<br />
H1 H2<br />
Figure 1: Structure of this study<br />
H1: In-hospital healthcare givers’ perceived usefulness toward ECN has significant correlation with using<br />
experience.<br />
H2: The experience of using this ECN has significant correlation with emotional exhaustion.<br />
3.3 Questionnaire design<br />
This questionnaire in this study is a structure questionnaire. Variables include: the degree of emotional<br />
exhaustion of healthcare givers, personal information, experiences of using in-hospital ECN, perceived<br />
114
Cheng-Yi Chiang et al.<br />
usefulness of using this website, perceived usefulness, character of in-hospital ECN and one<br />
open-ended question which is suggestions to this website. The extent of emotional exhaustion and<br />
perceived usefulness are continuous variables. Points are measured by the use of the Likert 7-point<br />
scale.<br />
3.3.1 Validity analysis<br />
A structured questionnaire is being used in this study. Some questions in Maslach and Johnson (1981)<br />
and Davis’ studies (1989) have been used in this questionnaire. Seven general experts with experience<br />
in the field (3 of them are Information Management experts and the others are psychiatric experts) will be<br />
invited to check and review the 21 continuous variables for its adequateness and necessity. A five-point<br />
scale is being used as its rating scale. Questions higher than 4 points will be adopted and validity<br />
measurement of experts have an average score of 4.5 points.<br />
3.3.2 Reliability analysis<br />
This study employs Cronbach’s α value to measure the consistency of every question,Non-research<br />
samples will be chosen for testing when questionnaires are examined for content validity. Modification will<br />
be done based on respondents’ opinions when questionnaires are being collected. The value of<br />
Cronbach’s α is more than 0.8 which stands for a high validity. The whole scale was 0.917 for emotional<br />
exhaustion and the Cronbach’s α of perceived usefulness of ECN was 0.971. Both of them reached<br />
significant validity.<br />
3.4 Data analysis method<br />
3.4.1 Descriptive statistics<br />
Descriptive statistics is used to analyze the perceived usefulness, using experience, motivation,<br />
emotional exhaustion awareness, and characteristics of the ECN. Normal distribution and usage<br />
distribution will also be included in the statistics process and multiple choice questions will be utilized to<br />
analyze the motivation and characteristics of in-hospital ECN.<br />
3.4.2 One-Way MANOVA<br />
One-Way MANOVA is mainly used to explore the correlation of the using experiences and emotional<br />
exhaustion of in-hospital ECN and the correlation of variables of using experience and perceived<br />
usefulness.<br />
3.4.3 Independent t-test<br />
Independent t-test is used to analyze the correlation of emotional exhaustion and perceived usefulness of<br />
in-hospital ECN.<br />
3.4.4 Pearson correlation<br />
Pearson correlation is employed to examine the correlation of healthcare givers’ experiences of<br />
in-hospital ECN with the perceived usefulness and emotional exhaustion awareness.<br />
4. Results and discussion<br />
1000 questionnaires were dispatched and were numbered from 1 to 1000. Questionnaires will be<br />
collected two weeks after distribution. Of the 1000 dispatched questionnaires, 918 were returned for a<br />
91.8% return rate. Effective respondent rate is 91.1%.<br />
4.1 Descriptive statistics<br />
4.1.1 Using experience<br />
The analysis of using experience: (1) Continuous users take the majority for 73.47%. 3/4 of the<br />
healthcare givers still keep using this website. (2) Users with five years or above using experiences have<br />
rate of 35.6% and users with 2-3 years of using experience have a rate of 13.7%. (3) Frequency of using<br />
ECN is mostly about 1 to 3 times. (4) Time of use is predominantly within 5 minutes (63.9%) and if those<br />
using time within 5 to 15 minutes are included in the calculation; the rate of time of use within 5 to 15<br />
115
Cheng-Yi Chiang et al.<br />
minutes will be 93.6%. (5) The main reason for never using this website is lack of interest (57%) and<br />
other reasons of never using this website has a rate of 21.9%.Non-continuous users have a rate of 64.5%<br />
and the main reason is lack of time. Void of content comes to the next which has a rate of 14.0%. Other<br />
reasons of using this website are ordered from: “lack of time”, “worried of being interviewed by<br />
supervisors and privacy of information exchanging is not secured.”<br />
4.1.2 Motivation<br />
Multiple choice percentages of motivation analysis is as follows: (1) Entertainment (19.6%) (2)<br />
Information sharing (18.5%) (3) Emotional relief (17.2%) (4) Learning (15.7%) (5) Emotional support<br />
(13.0%). The main reason for using this website us for entertainment and reason related to emotional<br />
relief and emotional support has a rate of 30.2%. 10.9 of the respondents use this website to monitor the<br />
condition of hospital.<br />
4.1.3 Characteristics of In-hospital ECN<br />
Characteristics of In-hospital ECN are ordered from: (1) Variety of topics (31.2%) (2) Enthusiastic<br />
interaction among members (23.6%) (3) Privacy of information (18.1%) Almost one-third of the<br />
respondents will like to have various topics for information exchanging and one-seventh (15.1) of the<br />
respondents will like to receive feedback for hospitals.<br />
”Other factors” in Analysis Chart of Characteristics of In-hospital ECN has a rate of 2.4%. Most opinions<br />
in the open-ended question are “hoping questions raised can be responded positively by hospitals”,<br />
“worried of being interviewed by supervisors, lack of privacy, hoping information could be secured.”<br />
4.2 Perceived usefulness and emotional exhaustion distribution<br />
4.2.1 Perceived usefulness of Healthcare Givers toward In-hospital ECN<br />
Perceived Usefulness and Emotional Exhaustion Distribution of Healthcare Givers toward In-hospital<br />
ECN are ordered from: (1) Beneficial to solve daily problems and with a score of 4.19. (2) Useful in daily<br />
jobs and with a score of 4.18. (3) Ease patients’ unpleasant emotion. (4) Less stressed in work emotion<br />
and with a score of 4.11. (5) Expressing ideas or suggestions to hospitals and with a score of 4.09.<br />
Average perceived usefulness score toward in-hospital ECN is 4.0 and only 25% of the respondents have<br />
an average score less than 4.0. Mode of the respondents is 5.0.<br />
4.2.2 Emotional exhaustion of Healthcare Givers<br />
The distribution of respondents’ emotional exhaustion is ordered from: (1) “Exhausted after duty” has the<br />
highest average score of 5.69. (2) “Emotional exhaustion of daily work” comes the next and with an<br />
average score of 5.20. (3) “Seems to have occupational burnout” ranked the third and with an average<br />
score of 5.02. (4) “Feel stressed when contact with customers” ranked the lowest score of 3.99.<br />
The average score of respondents’ emotional exhaustion is 4.80 and only 25% of the respondents have<br />
an average score less than 4.0. Mode of the respondents is 5.0. This phenomenon reveals that the e<br />
degree of emotional exhaustion of healthcare givers is relatively high.<br />
4.3 Related analysis<br />
4.3.1 Correlation between whether healthcare givers use the In-hospital ECN and emotional exhaustion<br />
awareness<br />
Whether healthcare givers use the in-hospital ECN or not has significant correlation with emotional<br />
exhaustion awareness. It conceals that there will be a significant difference in emotional exhaustion when<br />
healthcare givers do not use the website. Healthcare givers who employ the in-hospital ECN have lower<br />
emotional exhaustion awareness. Hence, H2 above of this study is supported; see Table 1.<br />
116
Cheng-Yi Chiang et al.<br />
Table 1: Independent sample t-test analysis of whether healthcare givers use in-hospital ECN and<br />
emotional exhaustion<br />
Item Use ECN Number Mean SD Standard error of mean<br />
Sum<br />
No 189 45.132 10.4402 .7594<br />
emotional<br />
exhaustion Yes 614 42.800 10.9403 .4415<br />
Levene equal<br />
variance test<br />
T test of equal mean<br />
F test Significant t df Significant Standard<br />
error<br />
Sum emotional exhaustion Assuming<br />
equal<br />
variance<br />
2.123 .146 2.590 801 .010 .9005<br />
Not<br />
assuming<br />
equal<br />
variance<br />
2.655 325.178 .008 .8784<br />
4.3.2 Correlation between Healthcare givers’ perceived usefulness and using experience toward<br />
in-hospital ECN<br />
This study is shown as Table 2:<br />
1. The perceived usefulness has a significant correlation with frequency of using the website. This<br />
means that the perceived usefulness of In-hospital ECN for healthcare givers is different in statistics,<br />
so is its frequency. Hence, the above H1 of this study is supported.<br />
2. The perceived usefulness has significant correlation with time of use. This shows that when the<br />
perceived usefulness is different in statistics and there will be great difference on time of use. The<br />
above H1 is supported.<br />
Table 2: Perceived usefulness and using experience and Pearson correlation analysis chart<br />
Usefulness<br />
Use frequency<br />
Use time<br />
5. Conclusions and suggestions<br />
Usefulness Use frequency Use time<br />
Pearson correlation 1 .331(**) .388(**)<br />
Significant .000 .000<br />
Number 804 804 801<br />
Pearson correlation .331(**) 1 .663(**)<br />
Significant .000 .000<br />
Number 804 835 832<br />
Pearson correlation .388(**) .663(**) 1<br />
Significant .000 .000<br />
Number 801 832 832<br />
1. The degree of emotional exhaustion and the mode of occupational burnout are greater than<br />
average value. It appears that healthcare givers are stressful at work and supervisors should take<br />
such problems seriously to avoid follow up questions.<br />
2. Reasons for those who do not use this website are: “lack of time “is the main reason; “worried of<br />
being interviewed and information is not secured “ranked the second reason. Healthcare givers are<br />
lack of time to use in-hospital ECN instead of lack of willingness. Some consider that the website is<br />
void of content and time-wasting. Moreover, it can be verified by the value of frequency and time of<br />
use. (Frequency and time of use are of low value)<br />
3. The main motivation of using this website is for entertainment; however, 30% of the respondents<br />
use this website to seek for emotion expression and emotional support. It is clear that healthcare<br />
givers still need a way to vent their feelings. 30% of the respondents use this website for<br />
information sharing and learning. Healthcare givers have the motivation of learning new knowledge. If<br />
this virtual space is being utilized properly by supervisors, healthcare givers can have their<br />
information exchanged and reached the goal of learning new knowledge and information sharing. For<br />
117
Cheng-Yi Chiang et al.<br />
the motivation of “monitoring the whole hospital”, healthcare givers expect to realize the policy and<br />
decision of hospitals and obtain related answers.<br />
4. Average score of perceived usefulness is 4.0. “Daily work assistance” scores higher among all the<br />
items within the questionnaire of in-hospital ECN. It reveals that healthcare givers can solve daily<br />
problems by the use of information sharing and correspond to the “information sharing” of motivation.<br />
Perceived usefulness of “less stressed at work”, ”releasing unpleasant emotion while taking care of<br />
patients” and “less frustration at work ”corresponds with “emotional expression and emotional<br />
support” of motivation in this study. Perceived usefulness of “feel stressed while interacting with<br />
patients”, “emotional exhaustion” and “exhausted” can still not be improved by the use of in-hospital<br />
ECN. Perceived usefulness of” same as solving working problems” can be traced back to the<br />
biggest demands of healthcare givers. Unpleasant emotion or energy of healthcare givers need to be<br />
released timely. It will be effective even if healthcare givers merely use this website to express their<br />
feeling after work. Hence, in-hospital ECN is relatively critical.<br />
5. Healthcare givers are expecting the in-hospital ECN to have characteristics as follows: “varied<br />
topics”(31.2%), “enthusiastic interaction”(23.6%), “information to be secured”(18.1%), “feedback from<br />
hospitals”(15.1%) and “stability of this website”(9.7%) The great concern of healthcare givers is the<br />
diversification of in-hospital ECN; not only merely emotion expression. This finding conforms to<br />
references and this study. Healthcare givers hope to build a multi-functional virtual space and this<br />
opinion has been mentioned in the open-ended question. In the aspect of “feedback from hospitals”,<br />
questions raised by healthcare givers did not receive adequate responses; it ended up with opinions<br />
of “worried of being interviewed by supervisors”, ”information is not being secured” and “lack of<br />
privacy.” Supervisory agencies only scold healthcare givers instead of understanding their thoughts<br />
and feelings. If questions raised by healthcare givers could be bulletined in public, healthcare givers<br />
and hospital executives can understand the other party’s needs and expectations. Hence, the<br />
establishment of this website will be meaningful. Executive should be more tolerating to avoid<br />
positive suggestions being neglected. Hospital should create a virtual space for healthcare givers to<br />
express their thoughts and ideas. In other words, healthcare givers will no longer use the in-hospital<br />
ECN to express their genuine feelings.<br />
The average score of emotional exhaustion is 5.0. However, objectives of this study are first line<br />
healthcare givers and executives. Their sources of pressure are different than those of high-level<br />
executives. The degrees of emotional exhaustion may also differ in the source of pressure. Emotional<br />
exhaustion of high-level executives should be explored further to serve as a reference of emotional<br />
expression.<br />
This study only discusses the using experience and perceived usefulness of in-hospital ECN. Content of<br />
information exchanging is not being explored in this study. Follow-up researchers can probe further the<br />
content of information exchanging of in-hospital ECN for further application to practical field and<br />
academia. (Ex: turnover prediction, work-related stress, work satisfaction, burnout prediction and<br />
e-counseling service etc.)<br />
References<br />
Adams, D.A., Nelson, R.R. and Todd P.A. (1992) “Perceived Usefulness, Ease of Use, and Usage of Information<br />
Technology:A Replication”, MIS Quarterly, Vol. 16, No.2, pp 227-247.<br />
Agarwal, R. and Prasad, J. (1999) “Are Individual Differences Germane to the Acceptance of Information<br />
Technology. Decision Sciences, Vol. 30, No. 2, pp 361-391.<br />
Ashforth, B.E. and Humphrey, R.H. (1993) “Emotional Labor in Service Roles:The Influence of Identity”, <strong>Academic</strong> of<br />
Management Review, Vol. 18, No. 1, pp 88-115.<br />
Bolton, S. (2000) “Who Cares Offering Emotional Work as a ‘gift’ in the Nursing Labor Process”, Journal of Advanced<br />
Nursing, Vol. 32, No.3, 580-586.<br />
Davis, F. (1989) “Perceived Usefulness, Perceived Ease of Use, and User Acceptance of Information Technology”,<br />
MIS Quarterly, Vol. 13, No. 3, pp 319-340.<br />
Davis, F.D., Bagozzi, R.P. and Warshaw, P.R. (1989) “User Acceptance of Computer Technology:A Comparison of<br />
Two Theoretical Models”, Management Science, Vol. 35, No. 8, pp 982-1003.<br />
Dignam, J.T., Brarrera, M.Jr. and West, S.G. (1986) “Occupational Stress, Social Support, and Burnout among<br />
Correctional Officers”, American Journal of Community Psychology, Vol. 14, No. 2, pp 177-193.<br />
Hochschild, A. R. (1983) The Managed Heart: Commercialization of Human Feeling. Berkeley: University of<br />
California Press.<br />
Holmes, T. and Rahe, R. H. (1967) “The Social Readjustment Rating Scale”, Journal of Psychosomatic Research,<br />
Vol. 11, No. 2, pp 213-218.<br />
118
Cheng-Yi Chiang et al.<br />
Igbaria, M., Zinatelli, N., Cragg, P. and Cavaye, A.L.M. (1997) “Personal Computing Acceptance Factors in Small<br />
Firms: A Structural Equation Model”, MIS Quarterly, Vol. 21, No. 3, pp 279-305.<br />
Jackson, S.E., Schwab, R.L. and Schuler, R.S. (1986) “Toward an Understanding of the Burnout Phenomenon”,<br />
Journal of Applied Psychology, Vol. 71, No. 4, pp 634-640.<br />
Kirmeyer, S. L. and Dougherty, T. W. (1988) “Work Load, Tension, and Copying: Moderating Effects of Supervisor<br />
Support”, Personnel Psychology, Vol. 41, No. 1, pp 125-139.<br />
Maslach, C. and Jackson, S. E. (1981) “The Measurement of Experienced Burnout”, Journal of Organizational<br />
Behavior, Vol. 2, No. 2, pp 99-113.<br />
Morris. J. A. and Feldman, D. C. (1996) “The Dimensions, Antecedents, and Consequences of Emotional Labor”,<br />
The Academy of Management Review, Vol. 21, No. 4, pp 986-1010.<br />
Saxton, M. J., Phillips J. S. and Blakeney, R. N. (1991) “Antecedents and Consequences of Emotional Exhaustion in<br />
the Airline Reservations Service Sector”, Human Relations, Vol. 44, No. 6, pp 583-595.<br />
Straub, D., Limayem, M. and Karahanna-Evaristo, E. (1995) “Measuring System Usage : Implications for IS Theory<br />
Testing”, Management Science, Vol. 41, No. 8, pp 1328-1342.<br />
Szajna, B. (1996) “Empirical Evaluation of the Revised Technology Acceptance Model”, Management Science, Vol.<br />
42, No. 1, pp 85-92.<br />
Venkatesh, V. and Davis F. (1996) “A Critical Assessment of Potential Measurement Biases in the Technology<br />
Acceptance Model: Three Experiments”, International Journal of Human-Computer Studies, Vol. 45, No. 1, pp<br />
19-45.<br />
Wharton, A.S. and Erickson, R.J. (1993) “Management Emotions on the Job and at Home:Understanding the<br />
Consequences of Multiple Emotional Roles”, The <strong>Academic</strong> of Management Review, Vol. 18, No. 3, pp<br />
457-486.<br />
Wichoroski, M.A. (1994) “The Secretary: Invisible Labor in the Workworld of Women. Human Organization, Vol. 53,<br />
No. 1, pp 33-41.<br />
119
The Implementation of RSS-Based Clinical Reminder<br />
Wen-Chou Chi 1 , Chia-Hsien Wen 2 , Sek-Kwong Poon 3 and Shih-Che Huang 3<br />
1 National Chung Cheng University, Chiayi, Taiwan<br />
2 Providence University, Taichung, Taiwan<br />
3 Taichung Veterans General Hospital, Taichung, Taiwan<br />
y6312002@gmail.com<br />
chwen@pu.edu.tw<br />
Abstract: Computer base reminder system can help physicians to get right information and make right decision in<br />
daily clinical work. This study presents a RSS-based Clinical Reminding System (RCRS) that was designed and<br />
implemented for reminding clinicians to deal with their varied unfinished clinical works. RCRS provides XML-formatted<br />
RSS clinical messages through automatically collecting clinical information for every clinician from the hospital<br />
information system (HIS). There are two types of system structure in our design, one is stand along module which is<br />
working in specific PC, and the other is web-based module, integrated with the CPOE system. According to this<br />
design, clinicians will receive and read these messages whenever he/she logs on the HIS. A hyperlink, which<br />
connects to appropriate applications and let the clinician easily make some response, was provided by each RSS<br />
message. The system can help clinicians focus on patient care without keeping track of clinical chores. Information<br />
“Content” of this system was suggested to modify by system reviewers, but information “Accuracy”, “Formats”, “Ease<br />
of use” and “Timeless” of this system is appropriate for system design purposes.<br />
Keywords: RSS, CPR, XML, clinical reminder<br />
1. Introduction<br />
Information system has been applied in medical service for a while, there are two important contributions<br />
of these computer-based service systems: reduce medical error and increase service efficiency (Aspden,<br />
et al., 2006). Because the huge database conduct by imputed clinical orders and Electronic Medical<br />
Record (EMR), there are a lot useful information in the system and these information will help us to do<br />
appropriate decision in clinical service through Clinical Decision Support System (CDSS). According the<br />
request proposed by clinical staff, CDSS use the data base form EMR and the algorithms that made by<br />
evident to provide some recommendation to these staff.<br />
The computer reminder is one of the CDSS used in hospital whose role is more and more important in<br />
medical service. Physicians can get important information from the computer reminder and use these<br />
information to do decision. But most reminder system are individual system, it can only provide one single<br />
event alert (Hunt, Haynes, Hanna and Smith, 1998). Recently, some reminder systems start to use in<br />
disease management or prevent disease, which means this system start to provide “a series of alert”, but<br />
still a single system (Garg, et al., 2005). In the clinical daily work, the physicians need a lot of information<br />
from many subsystem of Hospital Information System (HIS), every subsystem have its own alert, if we can<br />
collect all these information, not just one event or one disease, and provide physicians an “to do list”, it<br />
would be helpful in prevent error and increase service quality.<br />
The effeteness of computer reminders is debatable, some studies conclude that computer reminders<br />
produced much smaller improvements than those generally expected from the implementation of<br />
computerized order entry and electronic medical record systems (Shojania, et al., 2010); the others<br />
conclude in opposite way (Hunt, Haynes, Hanna and Smith, 1998; Garg, et al., 2005). Infect, the useful<br />
reminder has few characteristics (Lobach, 2005): automatically prompted to use the system; integration<br />
with charting or order entry system and response required. RSS, which stands for Really Simple<br />
Syndication, Rich Site Summary or Resource Description Framework (RDF) Site Summary, is a data<br />
exchange protocol for sharing web contents, may fit the requires of clinical reminder system. It may<br />
simplify and speed the users’ access to the latest updated information which is published by a web site.<br />
The publisher web site creates an RSS file, also known as RSS feed or RSS channel, in which the content<br />
to be published is transformed into an XML-based format. Any user can read the RSS feed at different<br />
sites at the same time by way of an RSS aggregator (RSS reader).<br />
An RSS aggregator is an application providing the means to read the content of RSS file before it can be<br />
displayed (Cadenhead, Curry and Zellers, 2005a; Bray, et al., 2004; Ola and Niclas, 2005). In general,<br />
there are three common types of RSS aggregators. The first type RSS aggregator, Desktop RSS<br />
aggregator, is the standalone program. Users download this program to their computers from Internet and<br />
use it to collect the feeds and refresh items when a feed is updated. Web-based aggregator, the second<br />
120
Wen-Chou Chi et al.<br />
type, is a kind of online web service. Users can use it to collect the feeds and browse messages in<br />
Internet. The final type is Plug-in aggregator which allows users to view RSS feeds with an existing<br />
program such as Microsoft Outlook (Ola and Niclas, 2005). RSS is generally used in real-time news<br />
reading and publication information management (Hu, Lim and Rizos, 2005; Klein, Mika and Serban;<br />
Çelikbaş). There are some RSS-based applications used in health care territory, but no one is applied in<br />
hospital based medical care (Phillippi and Buxton, 2010; Ku and Huang, 2010).<br />
The proposed RSS-based clinical information reminder is implemented under the support of a medical<br />
center in Taiwan. The reminder is coupled with the HIS. It will periodically collect clinical information<br />
required by clinicians from HIS and transform them into RSS messages automatically. A clinician will<br />
receive these RSS clinical messages sent to him/her whenever he/she logs on HIS. The reminder is also<br />
associated with the LDAP server of HIS for user authentication to protect patient privacy.<br />
RSS is an effective tool to spread internal information. One of the functional components of a CPR system<br />
is integrated communication support. A patient-oriented medical record will be the communication tool for<br />
team members who take care of a patient. Traditional computer-based patient-record (CPR) systems<br />
generally provide passive pull-type communication because each team member has to scan through the<br />
record to find other’s opinion. By including RSS mechanism in a CPR system, each team member’s<br />
opinion will automatically push to other members.<br />
2. System design<br />
We have built the RSS-based clinical information reminder. In order to help user operation be more<br />
effective and efficient, we designed and built two modes for user operation in our system: one is the<br />
standalone system which installs the PCs in the clinic and station; the other is the web-based module<br />
which integrates with the CPOE system. The system architecture has shown schematically in figure 1,<br />
uses the following components and sources of clinical information.<br />
Figure 1: The system architecture and its operation flow of RSS-based clinical information reminder<br />
2.1 Standalone mode<br />
In standalone mode, PCs and Workstations installed in the standalone system for front-end of reminder<br />
and used by healthcare workers to login in the reminder through the intranet. LDAP Server receives the<br />
user’s login request and identifies the user’s identity and authority. In RSS reminder, the patient’s<br />
information only allows the related healthcare workers to access. HIS which stores the clinical information<br />
and translated information to XML-formatted files. RSS Server includes the RSS aggregator and the<br />
121
Wen-Chou Chi et al.<br />
database server of the reminder. The reminder works by the way of coupling itself with the HIS of a<br />
hospital. The RSS server periodically collects XML-formatted clinical information from HIS and saves them<br />
into the RSS database as RSS clinical messages. The RSS feed is created concurrently.<br />
2.2 Web-based mode<br />
In our case, the web-based mode is similar to the standalone mode, but there are some differences in<br />
them. In this mode, we develop a RSS web service module and integrate with the CPOE system because<br />
the clinicians are familiar with CPOE in the hospital. When clinician uses the RSS service in the CPOE,<br />
the RSS web service will send the clinician id and patient’s id to the RSS server through the delivered web<br />
address. RSS server receives the address and interprets it, then returns the patient’s clinical information<br />
to the RSS web service. The example of web address has shown as<br />
http://xxx.xxx.xxx./?doctor_id=Y123009 &patient_id=A123008772.<br />
2.3 Data structure of the reminder<br />
Figure 2 shows an example of RSS clinical messages stored in the clinical information reminder. The data<br />
structure of the reminder consists of an index-file and many data-files.<br />
Figure 2: An example of clinical messages stored in the clinical information reminder<br />
The index-file is the RSS feed that lists some important attributes, such as patient id, data-file name, and<br />
physician id of all data-files. A data-file records detailed data of a clinical information. Both of the index-file<br />
and data-files are in XML format conformed to RSS 2.0. A element of the RSS feed represents<br />
a class of RSS clinical messages. There will be n elements in a channel, if there are n RSS clinical<br />
messages in the class. The patient id, data-file name, and physician id of a RSS clinical message are<br />
122
Wen-Chou Chi et al.<br />
correspondingly stated in element, element, and element of an item. Since a<br />
physician will not always use a specific client, the RSS aggregator of the reminder is designed as a<br />
web-based, the second type, aggregator. The RSS server performs event-driven RSS syndication,<br />
because only those messages with matched doctor’s id will be retrieved and displayed to the user (Ola<br />
and Niclas, 2005; Coile, 2002). Whenever a physician logs in HIS, the LDAP server will pass the address<br />
of the RSS aggregator and ciphered user data to the RSS server to initiate the reminder after<br />
authentication. The aggregator will get the RSS feed and pick up those items whose doctor’s id in<br />
element match with the one obtained from the LDAP server. A matched item represents a<br />
sound message. Then the aggregator passes the list of sound messages to the RSS server and requests<br />
it to retrieve sound messages from RSS database. All sound messages are displayed for advanced<br />
process by the clinician.<br />
3. System implementation<br />
The reminder is developed using the VB.net and ASP.net programming language and Microsoft SQL<br />
Server 2000. In our system, there are four classes of RSS clinical messages currently processed in the<br />
reminder: Abnormal laboratory results notification, abnormal examination results notification, discharge<br />
note notification, and consultation notice notification. Other kinds of clinical information will be joined to the<br />
reminder gradually.<br />
An abnormal laboratory results notification messages is sent to notify a physician that the result of a<br />
laboratory test he/she ordered is abnormal and critical. If a test result is verified to be abnormal and out of<br />
the acceptable range set in advance, relative information about the abnormal laboratory result, such as<br />
patient’s name, the laboratory result, and the order date-time will be immediately sent to the physician who<br />
ordered this test. Figure 3 shows an example screen about abnormal laboratory results notification in the<br />
standalone mode.<br />
Figure 3: An example screen about abnormal laboratory results notification<br />
The left part of the screen is a class list of RSS clinical messages. Archived messages will be shown<br />
Archived messages will be shown while “archived messages” item in the class list was clicked. The upper<br />
right part of the screen lists laboratory results notification messages that the clinician was received. If one<br />
of the messages in the list is clicked, detail information of the message will be shown on the lower right<br />
part of the screen. The physician may enter into laboratory information system by clicking the asterisk in<br />
the “link” field of the message.<br />
Figure 4 shows an example screen about the abnormal examination results notification. Similar to the<br />
appeal of abnormal laboratory results notification, it is used to notify a physician that the result of an<br />
imaging examination or a pathology examination he/she ordered is abnormal.<br />
123
Wen-Chou Chi et al.<br />
Figure 4: An example of transforming an abnormal examination results notification message into an<br />
archived one<br />
Discharge note notification is used to notify a clinician that one of his/her patient is discharged and he/she<br />
has to accomplish the discharge note before deadline. Figure 5 shows the RSS clinical messages about a<br />
clinician’s discharge note notification.<br />
Figure 5: An example of deleting a discharge note notification message<br />
A consultation notice notification message notifies a clinician that there is a consultation he/she has to<br />
hold. An example screen about consultation notice notification is shown in figure 6.<br />
4. Evaluation<br />
The case hospital is a medical center in central Taiwan. At present, the case hospital has more than 500<br />
physicians and 1,500 beds. The project team members come from the information department and clinical<br />
informatics research & development center (CIRD). Two project leaders are director from the CIRD.<br />
During the reminder design stage, they were responsible for discussing the reminder functions in the<br />
physicians group meeting. After reminder was implemented, they were responsible to train and collect<br />
feedback from the physicians. So, we chose the two project leaders as our interviewee to give a<br />
124
Wen-Chou Chi et al.<br />
preliminary evaluation of the system. The interview results were recorded using the taxonomy developed<br />
by Doll and TorKzadeh (1998) in The Measure of End-User Computing Satisfaction.<br />
Figure 6: An example screen about consultation notice notification<br />
The interviewees summarized that most users suggested that the “content” of the system needs to be<br />
screened in order to avoid information overload. Because the information provided by the system is not<br />
“created” from the system, but “collected” from other system, both evaluators believed the information<br />
provided by the system is “Accurate”. Regarding to the “Ease of use” and “Timeless” of the system, the<br />
evaluators scored “satisfied” with two recommendations. Firstly, they suggested that the end users need<br />
more authorizations to manipulate the information they want, because some events may be meaningful for<br />
some users but not others. Secondly, use the remote technology to deliver message to end user may<br />
increase the accessibility of the system.<br />
5. Discussion<br />
Providing too many information will cause the physician “information overlord” (Rebitzer, Rege and<br />
Shepard, 2008), so do this system. In the future, if this system can screen information and verify which<br />
information are important and which are not, it would be helpful. Unfortunately, it need time to build up this<br />
screen system.<br />
Because of the information which physicians need were come from different sub system in hospital<br />
information system, all the reminding system need is collect information from different subsystem. RSS<br />
has these features, the user don’t have to pick up information around hospital, the information we need will<br />
active “feed ” the RSS reminding system. There are two advantage of this system, more accuracy of these<br />
information and more easy to get information.<br />
The amount of control for the system is considered difficult to determine. Too much control by users<br />
makes it difficult to follow, and too little control limits the control by users (Sprague, 1980). In order to<br />
prevent the loss of essential information, RSS does not allow users to determine the type of information<br />
provided by the program. Multi-disciplinary care is usually involved in patient care. Before the consensus<br />
was made, RSS provided standardized information for all professionals involved in the patient care.<br />
In order to reduce the amount of errors in patient care, getting correct information on time is important.<br />
Although RSS is able to provide information on time, the medical professionals need to be using the<br />
specific computers to access the information. The medical professionals rarely stay in front of a computer<br />
during work. Therefore, it is important to develop a system that is capable of sending real-time information<br />
to the medical professionals via wireless devices such as mobile phones.<br />
125
6. Conclusion<br />
Wen-Chou Chi et al.<br />
RSS provides a mean to acquire up-to-date information quickly from Internet. According to the evaluation<br />
by project leaders, this RSS-base reminder system is useful and effective and suggested to the<br />
superintendent not only to expand the RCRS to include all kinds of messages needed by the physicians<br />
but also put this idea into the design of the next generation Hospital Information System. They also point<br />
out that physicians may be overburdened if provided too much information may not be relevant. If this<br />
system can screen information and verify important information would be helpful and increase the users’<br />
acceptance of this system. This paper presents the design and implementation of an RSS-based clinical<br />
information reminder. A clinician may use the reminder to receive all kinds of clinical messages and<br />
arrange them as a to-do list. Although the pioneer system only includes four classes of clinical messages,<br />
it proves that the reminder will help clinicians in managing clinical chores.<br />
Acknowledgments<br />
This paper is based upon work supported by National Science Council (NSC), Taiwan under grants<br />
no. NSC96-2221-E-126-006, NSC96-2221-E-126-004-MY3 and NSC95-2218-E-007-025. Any opinions,<br />
findings, and conclusions or recommendations expressed in this material are those of the authors and do<br />
not necessarily reflect the views of the NSC.<br />
References<br />
Aspden, P. et al., (2006) Committee on Identifying and Preventing Medication Errors. Preventing medication errors:<br />
quality chasm series. Washington(DC): The National Academies Press.<br />
Bray, T. et al., Extensible Markup Language (XML) 1.0 (Third Edition) W3C Recommendation 04 February 2004, The<br />
World Wide Web Consortium, 2004, last visited 2005-02-18, http://www.w3.org/TR/2004/REC-xml-20040204/<br />
Cadenhead, R. Curry, A. and Zellers, S. (2005a) “RSS at Harward Law”,[online], RSS Advisory<br />
Board, http://blogs.law.harvard.edu/tech/rss#whatIsRss<br />
Çelikbaş, Z. “What is RSS and how can it serve libraries?” Learning for the Future: e-Learning, pp. 277-292, İstanbul<br />
(Turkey).<br />
Coile, R.C. Jr. (2002) “The Paperless Hospital, Healthcare in a Digital Age,” Foundation of the American College of<br />
the Healthcare Executives.<br />
Doll, W.J., TorKzadeh G.(1998) "The Measure of End-User Computing Satisfaction. Management Information<br />
Systems Quarterly." Vol. 12, No.2, pp 259-274.<br />
Garg, A.X. et al., (2005) "Effects of computerized clinical decision support systems on practitioner performance and<br />
patient outcomes: a systematic review." JAMA, Vol 293, No. 10, 9 Mar, pp 1223-1238.<br />
Hu, Y.H. Lim, S. and Rizos, C. (2005) “Delivering GNSS Data Over the Internet Using RSS for Post-processing<br />
Applications” International Symposium on GPS/GNSS.<br />
Hunt, D.L. Haynes, R.B. Hanna, S.E. Smith, K. (1998)” Effects of computer-based clinical decision support systems<br />
on physician performance and patient outcomes: a systematic review.” JAMA, Vol 280, No. 15, 21 Oct, pp<br />
1339-1346.<br />
Klein, M. Mika, P. and Serban, R. “Semantics-based Publication Management using RSS and FOAF”<br />
SemanticDesktop.org.<br />
Ku, H.H. and Huang, C.M. (2010) "Web2OHS: A Web2.0-Based Omnibearing Homecare System." IEEE transactions<br />
on information technology in biomedicine, VOL. 14, NO. 2,<br />
Lobach, D.F. (2005) "Improving clinical practice using clinical decision support systems: a systematic review of trials<br />
to identify features critical to success." BMJ, Vol 330, No.7494, 2 Apr, pp 765.<br />
Ola, A. and Niclas, (2005) L. “RSS – The future of internal communication?” Reports from MSI, ISSN 1650-2647.<br />
Phillippi, J. C. and Buxton, M. (2010) "Web 2.0: Easy Tools for Busy Clinicians." Journal of Midwifery & Women’s<br />
Health. Vol 55, No. 5, pp 472-476.<br />
Rebitzer, J. B., Rege, M., and Shepard, C.(2008) "Influence, information overload, and information technology in<br />
health care." Adv Health Econ Health Serv Res, Vol. 19, pp. 43-69.<br />
Shojania, K.G. et al., (2010) "Effect of point-of-care computer reminders on physician behaviour: a systematic review."<br />
CMAJ, Vol 182, No.5, 23 Mar, pp 216-225. Kawamoto K, Houlihan CA, Balas EA,<br />
Sprague, R.H. (1980) "A Framework for the Development of Decision Support Systems." Management Information<br />
Systems Quarterly, Vol. 4, No.4, pp 1-26.<br />
126
Factors Determining the Adoption of ICTs in Healthcare<br />
Service Delivery: A Developing Country Context<br />
Taurai Chikotie, Jonathan Oni and Vesper Owei<br />
Cape Peninsula University of Technology, Cape Town, South Africa<br />
chikotiet@cput.ac.za<br />
jo45ng@yahoo.com<br />
OweiV@cput.ac.za<br />
Abstract: The proliferation of Information and Communication Technologies (ICTs) in healthcare service delivery has<br />
brought with it great advancements in this sector. This has created the concept of eHealth which is a relatively new<br />
concept in health care service delivery in most developing countries. According to previous research studies, ICT<br />
solutions are rapidly changing the way health organizations and stakeholders communicate with each other (Ouma<br />
and Herselman, 2008). Many people are using such communication technologies to communicate with others to<br />
gather information, with healthcare being one of the most important areas of such applications. It is therefore<br />
understandable that the value and importance of ICTs in health has increasingly been recognized world wide. The<br />
literature in innovation adoption notes that there are still inadequacies in identifying the factors that are crucial in<br />
promoting eHealth adoption in developing countries (Drury, 2005). Although, in Africa, the advent of eHealth has<br />
brought exciting opportunities to reduce or control the growing healthcare inequities, a lot still needs to be done in the<br />
adoption of these technologies (UNESCAP, 2007). In a developing country like South Africa, the disparity in the<br />
access to ICTs in healthcare service delivery is wide due to the dichotomous demography of first and second<br />
economies existing in this country. This has had a negative impact on the adoption eHealth technologies thus,<br />
suggesting that even though there has been an appreciation in ICTs in healthcare, very little has been done to<br />
ensure the adoption of such technologies. This paper, however attempts to ascertain such factors that determine<br />
adoption of ICTs in healthcare service delivery from both the management and patients’ perspectives. Drawing from<br />
the theories of innovation adoption in healthcare service delivery, the paper argues for the need to examine eHealth<br />
service adoption factors in developing countries and provides recommendations on how to tackle the challenges to<br />
adoption. The paper will conclude by recommending further research on issues in a number of key areas that need to<br />
be resolved to improve upon the efficient use and adoption of ICTs in healthcare service delivery amongst<br />
developing countries. South Africa is used as a case in this paper.<br />
Keywords: eHealth, challenges, theoretical models of adoption, ICTs in healthcare, developing countries<br />
1. Introduction<br />
The widespread use of new ICTs into almost every aspect of our professional and social lives is already<br />
gripping most developing countries (NTIA, 2007). Many people are using ICTs to communicate with<br />
others to gather information, with healthcare being one of the most important areas of such applications.<br />
It is therefore understandable that the value and importance of ICTs in health is increasingly being<br />
applauded world wide (Marconi, 2002). ICT adoption by various players in the health sector has been<br />
accelerating as well as spreading geographically in the recent years and this has led to the coinage of the<br />
new term ‘eHealth’ (UNESCAP, 2007).<br />
EHealth has several connotations, though all of them hover around the use of ICTs in healthcare service<br />
delivery. EHealth is defined as an emerging field in the intersection of medical informatics, public health,<br />
and business. It refers to health services and information delivered or enhanced through the Internet and<br />
related technologies (JMIR, 2008). Although in Africa the advent of eHealth has offered an exciting<br />
opportunity to reduce or control the growing healthcare inequity, a lot still needs to be done in the<br />
adoption of these technologies. In South Africa the gap in the use of ICTs to effective healthcare service<br />
delivery between the private and the public sector is still wide and this has had a negative impact on the<br />
adoption and use of ICTs in the healthcare sector (Bassett, 2000). This suggests that even though there<br />
has been an appreciation in ICTs in healthcare, very little has been done to ensure the diffusion and<br />
adoption of such technologies in this country.<br />
In developing countries, ICTs have been crucial to the development of effective health information<br />
sharing and dissemination systems, with users’ adoption of ICTs being one of the most important<br />
success factors of eHealth projects (Pannarunothai and Speedie, 2001). Many eHealth studies have<br />
been carried in Africa we however note that most of it has been focusing on particular technologies and<br />
only a few on the socio-technical and economic factors affecting the pace of adoption of such<br />
technologies in the healthcare sector. In modern trends of Information and Communication Technology, it<br />
is difficult for a project to be implemented without the consideration of personal, institutional and national<br />
127
Taurai Chikotie et al.<br />
factors especially in the adoption of new technologies (Ankem, 2004). This paper therefore outlines the<br />
research question as: What factors determine the adoption of ICTs in the healthcare service delivery<br />
sector in South Africa? The main purpose of this paper is to identify those factors that are crucial to the<br />
adoption of ICT in the healthcare delivery sector in South Africa.<br />
The rest of the paper is structured as follows: The next section is a literature review of healthcare and ICT<br />
development. This is followed by theoretical models of adoption and healthcare service delivery in section<br />
3. Section 4 discusses our research methods. An analysis of our findings is presented in section 5. Also<br />
in this section, the findings are discussed. The paper concludes in section 6 with a summary and a<br />
discussion of eHealth service delivery in developing countries.<br />
2. Healthcare and ICT development<br />
2.1 Global ICT developments in healthcare service delivery<br />
In recent years, the rates of adoption and diffusion of ICTs have become important indicators of<br />
development. Developing countries have put technological innovation on top of their national<br />
development agenda and invested heavily in ICTs (Detmer, 2000). For example, in the healthcare sector,<br />
the adoption of ICTs is perceived as the main driving force in the unfolding healthcare reforms in many<br />
developed and developing countries (Gladwin, Dixon, and Wilson, 2003). The influx of ICT in healthcare<br />
has increasingly been recognized by government agencies, health practitioners, private sector, academia<br />
and NGOs across the world as a milestone in quality health delivery. The United Nations Economic and<br />
Social Commission for Asia and the Pacific (UNESCAP) report on ICT for development (2007) stated<br />
that, judging from current news, research and reports, ICTs adoption by various players in the health<br />
sector has been accelerating as well as spreading geographically in the recent years and this is<br />
confirmed by the number of eHealth initiatives that have sprung up spreading from metropolitan areas to<br />
remote rural areas, increasingly, in both developed and developing countries. Wallace (1997:88) notes<br />
that in today’s information society, health professionals need to maximize the potential benefits offered by<br />
ICT as a means to improving public access to health care and information. Globally the convergence of<br />
ICTs has changed the way in which the health sector and society at large are organised and managed.<br />
From the foregoing it is seen that, it is important for every part of the health sector to adapt to these<br />
changes as ICTs can enhance the delivery of health services in various ways.<br />
2.2 ICT development in healthcare service delivery: South Africa<br />
Citizens in most of the emerging economies are beginning to taste the positives of communication<br />
technologies especially in the way they have improved the quality of their health lifestyle (United Nations,<br />
2008). In an era in which the state of health for developing countries, communities, and people in the<br />
world is at best stagnant and is probably in deterioration, the potential value of new information<br />
technologies for more effective health and development action can appear to be manna from heaven<br />
(Feek and Long, 2003:2). Emerging and existing health and ICTs initiatives have been put in place for the<br />
benefit of communities and also exposing key issues that need addressing in future projects.<br />
Most governments in developing countries are appreciating the critical role of ICTs to change their<br />
countries and have started to adopt these ICTs for national development. A lot of transformations in the<br />
way governments operate have been brought about by the advent of ICTs (Scott et al, 2005:97). Enakrire<br />
and Onyenania (2007:15) allude that most African countries have acknowledged the information<br />
revolution, although the adoption of communication technologies differs from country to country.<br />
The South African government has emphasized the development of an ICT sector through the formation<br />
of a national ICT strategy that addresses ICTs penetration and adoption particularly in underserved<br />
communities.<br />
According to the Department of Industry and Trade (2007), there has been the formulation of the South<br />
African Information Technology Industry Strategy (SAITIS) which is a bilateral project between the South<br />
African government and the Canadian government. This initiative seeks to contribute to sustainable<br />
economic growth, social upliftment and empowerment. Apart from SAITIS and Info 2025 Vision that have<br />
been assigned responsibility of building ICTs infrastructure, three taskforces were put in place to address<br />
to address ICTs deployment as a socio-economic development enabler and these are:<br />
128
Taurai Chikotie et al.<br />
The Presidential International Taskforce on information Society, which targets mainly the global ICTs<br />
market;<br />
The National Information Technology Taskforce, which deals with local ICTs initiatives and;<br />
The IT Council, responsible for local and provisional government information technology functions<br />
(Digital Opportunity Initiative, 2001: 13).<br />
There have also been a number of ICT initiatives in South Africa focusing on equitable information<br />
access and dissemination in the healthcare service delivery sector. Such initiatives include the Project<br />
Masiluleke (BBC, 2008); the telemedicine project done by the Medical Research Council and the<br />
Department of Health (DoH, 2007); the LoveLife Project (Chetty, 2007) and the Closed Health Broadcast<br />
Channel (Feek and Long, 2003). The main objective of such initiatives was to tap the power of ICTs as a<br />
high impact, low cost and efficient tools to deliver healthcare services at a distance to South African<br />
communities. They also seek to provide underserved communities with access to quality health<br />
information through the use of information communication technologies. However, research studies<br />
suggest that South Africa is poorly utilising Information and Communication Technologies (ICTs) in<br />
healthcare in disadvantaged communities although it is the very country that is leading in ICT<br />
development in the African Continent (Research ICT Africa, 2008; Gillward et al, 2005).<br />
3. Theoretical models of adoption and healthcare service delivery<br />
Several theoretical models were discussed to gather information on factors that foster the adoption of<br />
innovation from all perspectives, that is, social, technical and institutional or managerial. Such factors are<br />
deemed crucial in the adoption of ICTs within the healthcare sector from both the patient and the<br />
management or institution’s point of view. The most widely used of these theoretical models are the<br />
Diffusion of Innovation (DOI) which seeks to identify how an innovation is communicated through certain<br />
channels over time among the members of a social system (Rogers,1995); the Social Learning Theory<br />
which considers that people learn from one another, including such concepts as observational learning,<br />
imitation, and modeling (Bandura,1977); the technology acceptance model (TAM) which uses<br />
perceptions to predict attitude towards using or rejecting a technology or innovation (Davis, 1989); and<br />
the Fit between Individuals, Task and Technology (FITT) which seek to use the interaction of users, task<br />
and technology as the basis to better understanding IT adoptions in healthcare environments<br />
(Ammenwerth, Iller and Cornelia ,2006). All of the above mentioned theories seek to argue the adoption<br />
of innovations from the technology’s point of view and not much from the institutional or management’s<br />
perspective though they do address some of the social concerns. The identification of various factors on<br />
the management and patient view in South Africa is examined and proposed in the development of a<br />
pragmatic framework. Figure 1 depicts the proposed framework that will further be used in the findings<br />
from survey questions and interview.<br />
4. Experiment<br />
4.1 Research methods<br />
Document analysis: Extensive academic literatures on eHealth development and community informatics<br />
(Mbarika, 2007; Eysenbach, 2008; Samake and Mbarika, 2004; and Gurstein, 1999), ICT policies in<br />
healthcare (Rodrigues, 2008 and World Health Organisation, 2008), challenges to the adoption of ICT in<br />
healthcare (Kiplang’at & Ocholla, 2005; Migiro, 2006) were reviewed; including journals on previous<br />
related research innovations adoption in healthcare (Bates, Manuel and Oppenheim, 2007) and<br />
(Borgatti,2005), as well as white papers. These articles were analyzed and the gaps in ICTs adoption in<br />
healthcare were identified. The objective was to ascertain the factors critical in the assimilation of ICTs in<br />
the healthcare sector in a developing country scenario.<br />
For the research survey and interviews, patients and healthcare personnel from two healthcare centres<br />
(both public and private) in the City of Cape Town participated.<br />
Questionnaires: Survey questionnaires were issued to 20 patients in which they had to respond to<br />
questions on their use of ICTs of any type in health information communication, their frequency of use of<br />
such ICTs, the benefits they have gained and the challenges the are facing.<br />
129
Taurai Chikotie et al.<br />
Figure 1: Proposed framework for eHealth service delivery<br />
Interviews: Face to face interviews were conducted with 5 senior healthcare personnel with an<br />
understanding of ICTs within their organizations. These were used as the key informants to complement<br />
most of the data from questionnaires.<br />
The consultants were asked about their specific experiences working with ICTs and what they found<br />
when working with such technologies. Most of the responses focused on their brief, internal and external<br />
healthcare projects (i.e. benefits, challenges, strategies or policies etc).<br />
5. Analysis of findings<br />
Our findings identified several critical factors that are instrumental in the adoption of ICTs in healthcare<br />
service delivery most notably within a developing country context. The responses of the interviews and<br />
questionnaires are presented and discussed from both the management and patient’s perspectives. Most<br />
of these responses are closely linked to the understanding of ICTs usages and the challenges being<br />
faced in using these technologies. Important to note is the fact that while there has been several ICTs<br />
and healthcare related projects in developing countries, such projects have not been able to sustain<br />
national needs with some failing to meet their objectives as their models or frameworks are based on<br />
alien foundations different from what developing countries need.<br />
Several factors were identified from the interviews with senior healthcare personnel and these<br />
include:<br />
Knowledge: Most of the senior healthcare personnel concurred to having serious challenges on issue<br />
such as lack of end-user training on the systems they use, lack of online health resources awareness and<br />
poor research skills in finding authoritative information online. This was confirmed by another interviewee<br />
who noted that their academic background is to blame for the slackness in ICT skills.<br />
Organizational: some respondents brought issues such as the need to satisfy diverse stakeholders, the<br />
need for timely response to service requests, resistance to change, and the need to integrate services<br />
with the workflow of medical, technical and administrative staff.<br />
130
Taurai Chikotie et al.<br />
Regulatory and policy: these include the lack of concrete security standards for medical records, Issues<br />
of interstate licensure, credentialing of health care professionals and liability for malpractice, and lack of<br />
suitable payment mechanisms. The healthcare sector is a very sensitive field which needs stern<br />
measures in the case of information sharing and dissemination. Just like in developed countries, the<br />
issue of ethics and regulations play a crucial role in the use of ICTs in healthcare service delivery.<br />
Economic: includes the lack of sustainable funding models, and pace with the ever-changing technology<br />
while controlling costs. In South Africa the healthcare sector is always operating at shortfall due to low<br />
funds being allocated to the department. This leaves most of the management to weigh priorities for<br />
healthcare service delivery, that is, it is either drugs or technology of which the latter always play second<br />
fiddle.<br />
From the patient’s perspective the following factors were identified:<br />
Poor user awareness on new technologies: in spite of the support for the exciting benefits of eHealth, a<br />
number of obstacles continue to stand in the way of its widespread adoption by health organizations and<br />
consumers. Most consumers are still not aware that they may access specialist knowledge online (Kedar<br />
et al, 2003). Most patients link this to poor awareness on the abilities of ICTs especially in health<br />
information sharing and dissemination.<br />
ICT Infrastructures: South Africa, like any other developing country, still faces a lot of challenges in ICT<br />
infrastructural development. This contributes significantly to the poor adoption of eHealth in South Africa.<br />
Many African countries with South Africa included have witnessed considerable local and foreign<br />
investments in ICT infrastructures development, primarily in the telecommunications sector and it is up to<br />
the policy makers to take this as an opportunity to enhance the use of ICTs in the health sector.<br />
Education: most responses also implied that both well educated health staffs and patients have positive<br />
impacts on the development, implementation, adoption, and diffusion of various eHealth initiatives.<br />
Tornatzky and Klein (1982) write that, an effective education facilitates and increases the expression and<br />
identification of patients’ needs, the communication between providers and patients, and the patients’<br />
awareness and appreciation of the benefits of the eHealth project. Incorporating ICTs into medical<br />
educational interventions is a challenge. It is therefore vital that staff is trained on ICTs use to ensure<br />
maximum effectiveness and efficiency for support in the process of rendering health services to patients<br />
(Sargeant, 2007).<br />
Content divide: information provided on most ICTs is in English, which is also a hindrance to non-English<br />
speaking people who might need to understand the whole concept coming with these health<br />
technologies, and South Africa is not an exception (Mbarika and Kifle, 2006). This is a concern to most<br />
patients especially those who use mobile phones and have a poor English background.<br />
6. Discussion and conclusion<br />
6.1 Summary<br />
From the findings presented in this paper, it is evident that adoption of ICTs in healthcare service delivery<br />
from both the management and patients’ perspectives comes with various challenges. Drawing from the<br />
theories of innovation adoption in healthcare service delivery, the paper argues for the need to examine<br />
eHealth service adoption factors in developing countries. On the patience perspective several issues like<br />
poor user awareness on new technologies, education, content divide, etc. has been identified. These<br />
factors will go a long way in assisting policy makers to evaluate the need and effectiveness of healthcare<br />
service delivery especially in a developing country context. However, senior management in healthcare<br />
will especially play an important role in sharing and dissemination of eHealth policies to patients. Just like<br />
in developed countries, the issue of ethics and regulations play a crucial role in the use of ICTs in<br />
healthcare service delivery.<br />
6.2 Government policies<br />
The lack of clear government policies and strategies in the promotion of eHealth in most developing<br />
countries has led to the slow uptake of the technologies. Where initiatives have been implemented, there<br />
has not been adequate monitoring and evaluation as to how these initiatives are performing (Ouma, and<br />
131
Taurai Chikotie et al.<br />
Herselman, 2008). It is important that government policies support most of the initiatives to promote<br />
eHealth.<br />
Consequently, some of these factors may differ from one community to another and may not directly<br />
reflect a particular eHealth environment. But it is hoped that this study would be applied to various<br />
organizations, governments, and eHealth policy practitioners in planning eHealth adoption. This paper<br />
will also be of practical relevance in the decision making process of senior management and consultants<br />
of ICTs in service delivery. The various factors identified as well as findings from interviews in a<br />
developing world context will address and further highlight the need to promote eHealth especially in the<br />
developing countries.<br />
References<br />
Ammenwerth, E, Iller, C and Corneila, M. 2006. IT-adoption and the interaction of task, technology and individuals: a<br />
fit framework and a case study. BMC Medical Informatics and Decision Making. Vol 6(3)<br />
Ankem, K. 2004. Use of Internet resource-based value-added processes by faculty in LIS education. Library and<br />
Information Science Research, 28:482-500<br />
Bandura, A. 1977. Social Learning Theory. Englewood Cliffs, NJ: Prentice-Hall.<br />
Bassett, H. 2000. Healthcare in South Africa. MedHunters Articles. URL:<br />
http://medhunters.com/healthcareInSouthAfrica.html. (Online) Retrieved: 07 September 2010.<br />
Bates, M., Manuel, S. & Oppenheim, C. 2007. Models of early use of ICT innovations in higher education. ARIADNE,<br />
50:1-10.<br />
Borgatti, S. P. 2005. Thinking theoretically. URL: http://www.analytictech.com/mb870/handouts/theorizing.htm :<br />
[Accessed: 8 October 2010].<br />
Chetty, K.S.2007. EHealth: the South African Context. URL: www.doh.gov.za/healthini.htm . [Accessed: 14<br />
November 2010]<br />
Department of Health. 2007. Health Strategic Framework 1999-2004. URL: http://www.doh.gov.za/docs/policy-f.html.<br />
[Accessed: 23 March 2010]<br />
Department of Trade and Industry, 2004. Benchmarking of technology trends and technology developments report.<br />
Johannesaburg: Bluepeter/AMI.<br />
Detmer, D.E. 2000. Information Technology for Quality Health Care: A Summary of United Kingdom and United<br />
States Experiences. Quality in Health Care, 9: p. 181-189.<br />
Digital Opportunity Initiative, 2001. Creating a development dynamic: Final report of the Digital Opportunity Initiative.<br />
Okinawa: UNDP<br />
Drury, P. 2005. The eHealth agenda for developing countries. World Hosp Health Serv; 41: 38–40.<br />
Enakrire, T. R. and Onyenania, O. G. 2007. Factors affecting the development of information infrastructure in Africa.<br />
Library Hi Tech news, 2: 15-20.<br />
Eysenbach, G. 2001. What is eHealth?. Journal of Medical Internet Research 3(1):e20. URL:<br />
http://www.jmir.org/2001/2/e20/index.htm. (Online) Retrieved: 3 October 2010.<br />
Feek, W and Long, G. 2003. The Digital Pulse: The Current and Future Applications of Information and<br />
Communication technologies for developmental Health Priorities. A Report Published by the Communication<br />
Initiative. URL: www.comminit.com [Accessed: 7 January 2010].<br />
Gillwald, A. and Esselaar, S. 2005. South African ICT Sector Performance Review. Wits University.School of Public<br />
and Development Management, Johannesburg.<br />
Gladwin, J., Dixon R.A., and Wilson T.D. 2003. Implementing a new health management information system in<br />
Uganda. Health Policy and Planning. 18(2): p. 214-224.<br />
Gurstein, M and Dienes, B.1999. “A ‘Community Informatics’ Approach to Health Care for Rural Africa”. Presented to<br />
The Africa Telemedicine Project: <strong>Conference</strong> ’99<br />
Jai Ganesh, A.U. 2004. EHealth - drivers, applications, challenges ahead and strategies: a conceptual framework.<br />
International Journal of Medical Informatics. Vol (20)2<br />
JMIR. What is eHealth? Journal of Medical and Internet Research. 3(1): e20. URL:<br />
http://www.jmir.org/2001/2/index.htm. (Online) Retrieved : 8 November 2010<br />
Kedar, I et al. 2003. Internet based consultations to transfer knowledge for patients requiring specialised care. British<br />
Medical Journal. Vol (326). 696-699<br />
Kiplang’at, J. & Ocholla, D. N. 2005. Diffusion of information and communication technologies in communication of<br />
agricultural information among agricultural researchers and extension workers in Kenya. South African Journal<br />
of Libraries and Information Science, 71(3):234-246.<br />
Marconi, D. 2002.Health Connectivity for development in America. Social Science and Medicine. University of<br />
Michigan: University Press. 40-43<br />
Marcus, J.A. 1986. Diffusion of innovations and social learning theory: Adoption of the context text-processing<br />
system at Stanford University. Dissertation Abstracts International, A 46/12, 3553(University Microfilms No. AAT<br />
8602506)<br />
Mbarika, V and Kifle, M.. 2006. Telemedicine in sub-Saharan Africa: The case of tele-ophthalmology and eye care in<br />
Ethiopia. Journal of the American Society for Information Science and Technology. 57(10), 1383 -1393<br />
Migiro, S. O. 2006. Diffusion of ICTs and e-commerce in manufacturing SMEs in Kenya. South African Journal of<br />
Library and Information Science, 72(1):35-44.<br />
132
Taurai Chikotie et al.<br />
NTIA (National Telecommunications and Information Adminstration). 2007. Falling through the Net: Towards digital<br />
inclusion. Department of Commerce: Washington, DC. 47-53<br />
Ouma, S and Herselman, M.E. 2008. EHealth in rura areas: case study of developing countries. Proceedings of<br />
world academy of science, engineering and technology. Vol (30) 560-566<br />
Pannarunothai, S., Speedie S. 2001. Factors influencing health information technology adoption in Thailand's<br />
community health centers: Applying the UTAUT model. International Journal of Medical Informatics, Vol (78)6.<br />
404-416<br />
Research ICT Africa .2008. ICT Access and Usage in South Africa, LINK Centre. University of Wits, Johannesburg,<br />
SA.<br />
Rodrigues, R.J. 2008. EHealth Strategies in America. The Commonwealth Health Ministries Reference Book. 5(3):<br />
e4. URL: http://www.ehealthstrategies.com .[Accessed: 19 June 2010]<br />
Rogers, E.M. 1995. Diffusion of Innovations. 4 th Ed. New York, N.Y.: Free Press. 58-63<br />
Samake, K.D. and Mbarika, W.A. 2004. EHealth in Africa: A vision for a healthier Africa. American Journal of Public<br />
Health. 96(1). 73-78<br />
Sargeant, J.M. 2007. Medical education for rural areas: Opportunities and challenges for information and<br />
communication technologies. Journal of Postgraduate Medicine. 51(4): 301-307<br />
Scott, V., Chopra, M., Conrad, E and Ntuli, A. 2005. Low equitable is the scaling up of HIV service provision in South<br />
Afnca? SAfrMedf. 95:I09-n3.<br />
Tornatzky, L and Klein, K. 1982. Innovation characteristics and innovation adoption- implementation: a meta analysis<br />
of findings. IEEE transactions on Engineering Management. Vol. 29(1), 28-48<br />
[6] UNESCAP. 2007. ICT for development in South East Asia and the Pacific. Report on developing countries and<br />
ICT revolution for the Scientific and Technological Options Assessment Panel. URL:<br />
www.unescap.org/search/rpts/ictdcp_asia466878764.pdf. (Online) Retrieved: 24 September 2010<br />
United Nations Foundation.2009.Global ICTs Report 2009.<br />
World Health Organisation. 2008. ICT and healthcare in the African region. URL:<br />
http://www.who.org/search/africa_sa.htm. [Accessed: 27 October 2010]<br />
Websites: http://newsvote.bbc.co.uk/2/hi/technology/7688268.stm. (Online) Retrieved: 11 October 2010.<br />
133
Evaluation of Application Embedded Knowledge Migration<br />
Issues<br />
Mitchell Cochran<br />
Claremont Graduate University, Claremont, USA<br />
mcochran@ci.monrovia.ca.us<br />
Abstract: As computing has matured, more organizations are purchasing best of breed applications as opposed to<br />
developing them in-house. From a Knowledge Management point of view, the organizations are renting the use of<br />
knowledge that is embedded in the applications. The organizations may own the data but the host application<br />
company owns the intellectual capital that creates the knowledge. For any of a number of reasons organizations will<br />
have to move to new applications and in turn new knowledge. It is assumed that the organization will be able to<br />
migrate current data and print reports but it does not own the original base knowledge. The issue is to understand<br />
what knowledge is imbedded in the old application and how can it be integrated into the new system. As the<br />
knowledge is inventoried, the new vendor can then determine if the knowledge will be available in the system. After<br />
that determination, the user might have to decide if the information is obsolete or possibly lost data. The migration<br />
issue also can put the onus of development on the end user. Consider the conversation of the developer and the end<br />
user where the end user asks for a feature that the developer has not seen. The end user is looking for features in<br />
the old system and the developer is going to say that it is up to the end user to tell them what they want. The issue is<br />
that the end user may now know what they want. The knowledge embedded in the code of the old application<br />
provided the information. The information basis is the intellectual property of the outgoing vendor and they may not<br />
have any reason to work with the incoming vendor. The paper will evaluate migration issues based on a case study<br />
of the migration of a financial application for a small city. The paper will also discuss some of the assumptions of<br />
knowledge management and a knowledge inventory to help an organization prepare to move applications to a new<br />
vendor.<br />
Keywords: knowledge management, application, migration<br />
1. Introduction<br />
An organization is responsible to understand all of the knowledge which is necessary to run its business.<br />
The knowledge is the intellectual capital of the organization. Rastogi defines intellectual capital as the<br />
“holistic meta-level capability of an organization to generate creative and effective responses to extant<br />
and emerging, present and potential challenges facing it, in an ongoing manner”. (Rastogi, 2000)<br />
The firm may or may not have developed explicit practices to maintain the knowledge base. Instead, it<br />
may be dependent on informal practices. Nelson and Winter comment that much of an organization’s<br />
knowledge is tacit because the crucial know-how resides in the minds of the organization’s members.<br />
(Nelson and Winter, 1980) Badaracco comments that embedded knowledge tends to be non-migratory<br />
where it “resides primarily in specialized relationships among individuals and groups and in the particular<br />
norms, attitudes, information flows and ways of making decisions that shape their dealings with each<br />
other.” (Badaracco, 1991) The organization needs to develop a knowledge inventory to develop an<br />
understanding of the critical process or application knowledge. Without a formal procedure to identify the<br />
knowledge inventory the company may not be able to fully document, develop or migrate knowledge to<br />
new applications.<br />
Many authors have identified the importance of documenting knowledge. Bots and deBruin created a<br />
knowledge value chain which includes a ‘Make Inventory Determine Knowledge Needed’. (Bots and<br />
deBruin 2002). Sage and Rouse identified issues that also relate to knowledge identification and<br />
migration: (Sage and Rouse 1999)<br />
Modeling processes to identify knowledge needs and sources<br />
KMS strategy for the identification of knowledge to capture and use and who will use it<br />
An understood enterprise knowledge structure<br />
Buhler states that only developing learning organizations that use knowledge that they acquire can<br />
continue to adapt and respond to their changing environment (Buhler 2002). McManus et al summarize<br />
the challenge to managers: ‘Managers must adequately plan for and implement KM practices that will<br />
allow them to effectively operate despite the multifarious influences that threaten the knowledge bases of<br />
their business.’ (McManus et al, 2003) The KM practice that this paper focuses on is developing an<br />
inventory of knowledge.<br />
134
Mitchell Cochran<br />
The importance is not which tool is used to define the knowledge but that the organization defines and<br />
follows a plan to document the knowledge. There any number of tools that could be used. Grossman et al<br />
list four different KM models: the Balanced Scorecard, The Intangible Assets Monitor, the Skandia<br />
Navigator, and the Intellectual Capital Index. (Grossman et al, 2006). If the organization is not using a<br />
tool, they may not have a KM strategy in place for either the functional business area or the overall<br />
organization. In 2010, an informal survey of municipal information system directors for the State of<br />
California asked how many agencies had complete documentation of all of their systems. Approximately<br />
10% of the 300 cities or special districts responded that they had completely documented their systems.<br />
The survey was only intended to get an initial impression of the issue’s scope. The entities had detailed<br />
documentation on which applications were in use and some went so far as to have documented the<br />
system settings. Most have only created electronic backups without a way to recreate the application<br />
settings should it need to be reinstalled from the vendor supplied media. The backups can be applied to<br />
new equipment but the organizational knowledge within may not be understood. The systems have been<br />
documented but not the embedded knowledge.<br />
If there is a plan or procedure, an implied part of the plan is to be able to identify all of the knowledge<br />
used by the firm. The original sources of knowledge may be internally or externally developed such as<br />
outsourcing. In either case, the components of the organization’s functional knowledge may readily<br />
available. Externally provided knowledge may be obtained by purchasing applications, outsourcing<br />
services, or using consultants. However, the knowledge may be tacit in nature so that it can not be easily<br />
recreated or available in a raw form. The knowledge may be available from the consultants or have to be<br />
purchased for a second time. Internal knowledge may have been developed by sources that are no<br />
longer with the firm which can be the case for retiring employees. The foundations and assumptions that<br />
created the legacy knowledge may not be documented.<br />
One of the critical areas would be for the firm to understand what processes are known and what are<br />
unknown. The knowledge inventory would help to identify what knowledge points are known and which<br />
attributes need to be discovered. Without a process to indentify what the organization knows, it cannot<br />
identify what points are missing or unidentified.<br />
Without a detailed understanding of the knowledge base the firm may not be able to recreate it in other<br />
forms. The issue is that the firm may not understand how to move the company’s knowledge base to new<br />
applications or processes. From a disaster recovery context, many organizations may not understand the<br />
knowledge inventory which is required to operate an organization without any automated processes.<br />
Many organizations do not map all of the information that is used by the organization. They may map<br />
what they consider as the critical missions. One of the concerns is that not all of the knowledge is<br />
identified. The critical base of knowledge may be dependent on processes that are not documented. The<br />
organization needs to identify and understand all of the processes whether the processes are internal or<br />
external, and or known or unknown.<br />
As part of a knowledge inventory the organization needs to understand which knowledge it controls and<br />
knowledge that is controlled by an outside entity. Some of that knowledge is obtained from consultants<br />
and available to be inventoried. The consultant will typically deliver a product or report that documents<br />
the results of the services provided. Some of the knowledge base will not be available since it was<br />
purchased as an application. The application takes inputs, applies an embedded knowledge and then<br />
develops an output. The process knowledge may not be known. The software will probably include a use<br />
license as opposed to an ownership agreement which might infer a right to re-engineer the application.<br />
Many applications with embedded knowledge can not be re-engineered due to those legal or contractual<br />
agreements.<br />
Should a firm use outsourcing services, the firm may not own the knowledge of the application, just the<br />
data used to create it. The organization will need to understand the process if the organization plans on<br />
providing the capability internally in the future. This situation can be found in hosted applications or many<br />
cloud computing environments. A simple example would be a outsource firm that provides payroll<br />
processing. The firm provides the knowledge of current personnel benefit, tax and withholding issues.<br />
An organization needs to have a method to document the organization’s knowledge. A framework or<br />
process needs to identify what knowledge is required and or which knowledge is readily available. If the<br />
data or knowledge is not available then there will be a cost for its recreation. One example is to identify<br />
135
Mitchell Cochran<br />
the level of knowledge that the organization requires about human resource practices. The organization<br />
can easily outsource HR practices to industry leading firms. The knowledge that is easily found in the<br />
industry can be easily migrated. The more specialized the information is the more difficult it will be to find<br />
outside vendors to provide it. The organization needs to understand if that knowledge is strategic to the<br />
key practices of the organization. The organization’s knowledge base will include organizational memory<br />
theory to develop the existing structure of retained knowledge.<br />
Many forms of knowledge will have a shelf life. The organization will have to determine what the cost<br />
would be to recreate any data that is not maintained. One example is the data on an organization’s<br />
infrastructure. If the information updates are not applied to the knowledge base then the knowledge is<br />
lost. All that will be available will be the obsolete data. Consider what it would take to recreate the US<br />
space program so that a person could be landed on the Moon in the near future. The technology has<br />
advanced over 30 years but the operational knowledge of the engineers was lost due to personnel<br />
retirements. That operational knowledge would have to be recreated. (Leonard, Kiron, 2002)<br />
Another concern is to understand if a process is outsourced, what would be the cost to bring it back into<br />
the organization. Some operations may be outsourced due to cost savings. If the outsourcing is not as<br />
effective as desired or if the outsourced operation has become more integrated into the critical mission it<br />
may need to be brought back into the organization. The initial operational knowledge may have to be<br />
recreated or purchased.<br />
Unidentified or missing knowledge will also be an issue. Over time, the data origins or the knowledge<br />
itself may be unavailable. Tacit knowledge may leave as employees leave. Legacy applications may<br />
contain knowledge but not the foundations that created that knowledge. An organization may understand<br />
the processes or rules from legacy applications but they need to understand the assumptions and<br />
theories needed to create those processes. The organization needs to answer the question of why it did<br />
what it did. As the environment changes, the organization will have to adapt the rules. If it does not<br />
understand the knowledge behind the rules or logic then it will be costly to alter the logic. If it is not<br />
documented, the knowledge will only be retained by the memory of the staff. The issue of knowledge<br />
retention is especially important with the retirement of the experienced baby boomer workers. A simple<br />
example can be considered when deciding if an accounting operation is taxable. There are rules that<br />
govern the process but how the organization handles an exception will require documentation of the<br />
assumptions or circumstances. The rules try to create black and white situations but there will always be<br />
a grey area. The answers to how and why the organization achieved the result are just as important as<br />
the result. Without the assumptions or background information the operation results may not be<br />
repeatable.<br />
For disaster recovery purposes, company needs to be run an operation without its primary tools. Should<br />
the organization not have access to its applications, it would have to recreate its processes using other<br />
procedures or manual tools.<br />
Many knowledge concepts are embedded in applications. The knowledge is made tacit by the<br />
programmer in the application. The issue is that the knowledge is implicit to the application and may not<br />
be available to the user. The user can own the data and use the application to generate results but he or<br />
she may not understand the knowledge area. One simple example would a tax application. The rules are<br />
documented and available but the user does not want to spend the resources to understand the<br />
application area. The user would rather purchase the application where the domain knowledge is then<br />
made accessible. An interesting example would be how the user would respond to a tax audit based on<br />
the results of the application. The user can only attest to the data that was input in the application. The<br />
user cannot attest to the appropriateness of the results, only a review of the application logic can provide<br />
that support.<br />
The organization’s application base provides knowledge of the firm’s operations. The organization will<br />
optimize procedures based on the applications. Over time, those applications will change due to vendor<br />
requirements, hardware obsolescence, or new customer requirements. If the current application process<br />
and procedures are not understood then it will be difficult to migrate to a new application. The current<br />
vendor owns the application and has no incentive to help an organization move to a new application. The<br />
user has ownership of the data but not the knowledge. Consider the following conversation between a<br />
customer and vendor:<br />
136
Mitchell Cochran<br />
Customer: Please provide a similar function that we had in the old system.<br />
Vendor: How did the function work?<br />
Customer: We don’t know, the program provided the information.<br />
Vendor: As a vendor we can’t re-engineer the old code so we don’t know what it did or how it did<br />
it. You will have to tell us what you want and how to do it.<br />
Customer: But we don’t know how it works.<br />
The current vendor may not be able to provide technical assistance or a description of the knowledge that<br />
the application contains. The new vendor asks the customer what they would like. The customer<br />
responds that they want the application to perform like the last one. The new vendor may not have an<br />
understanding of the old application. The onus will be on the customer to develop a knowledge base of<br />
what the old application did and what the new application should do. If the customer does not understand<br />
the knowledge base they will have to incur the cost to recreate the knowledge.<br />
2. Case study<br />
The City of Monrovia, California was required to upgrade its Cayenta financial application during 2007.<br />
The vendor was no longer going to support the installed version. The City would have been content to<br />
continue with installed version but the change was required by the vendor’s decision to terminate support.<br />
The new version included significant new features and would cost $500,000 for both the software license<br />
fee and migration efforts. Due to the large dollar amount, the City was required to investigate alternatives.<br />
The Request for Proposal or RFP process had eight proposals which ranged in cost from $300,000 to<br />
$1,000,000. The City chose a new vendor, Munis, which bid approximately $350,000 for the software<br />
license fees and migration costs. The evaluation did not show enough value in the $500,000 Cayenta<br />
proposal to justify the additional $150,000 expense to stay with the original vendor. The new vendor only<br />
had one similar financial system installed in California. There are special reporting mechanisms for the<br />
California State Public Employee Retirement System or PERS. Each vendor needs to develop the<br />
custom reports. The new vendor promised to develop the required reporting code in a timely manner.<br />
MUNIS had similar reporting requirements for other states so they assumed that the level of effort would<br />
be comparable. The California cities of Lakewood, Palmdale, Redondo Beach and Chino also purchased<br />
the financial system approximately at the same time. The cities joined together into a local users group.<br />
All of the cities had the same retirement system reporting requirements.<br />
The previous vendor had a working retirement reporting system. The new vendor had yet to develop its<br />
system. The new vendor required the users group to determine the needs for the new system. The users<br />
group acted as the development team. The new system vendor would develop prototypes and the users<br />
group was required to troubleshoot each version. The MUNIS developers underestimated the level of<br />
effort required. The development project took over two years. Some of the Cities had working systems,<br />
others where not completed as of this point. One of the development difficulties is that each city had<br />
different requirements. Fire and police departments have unique retirement reporting requirements.<br />
Some cities did not have public safety components which simplified their reporting efforts. The cities<br />
without either fire or police departments where able to implement a correctly working reporting system.<br />
The issue is that the vendor would commit the resources to develop the system but the responsibility for<br />
a correct submission lies with the City. If the system is delayed, which might be expected on a new<br />
system, the City has to find a work around or a manual method. The City of Monrovia was forced to take<br />
output from the new financial system and enter the batch totals into the old system to get the correct<br />
submission information. In addition to any normal or expected migration efforts, the PERS system<br />
announced that it was going to change the reporting requirements and structure. The new system was<br />
then put on hold pending the State implementation of the new requirements. The project was further<br />
complicated by delays in the State implementation of the new standards.<br />
The ultimate responsibility for submission of correct data is that of the end customer which in this case is<br />
the City. The vendor committed to develop the system but it would need specific requirements. The<br />
contractual liability of the vendor is typically limited to the amount of the fees paid by the client. The City<br />
has been forced to maintain the old system to act as a migration tool for creating the correct reports.<br />
Even with known requirements from the retirement authority, the vendor was not able to develop the<br />
137
Mitchell Cochran<br />
reporting system in a timely manner. They could develop a system that works for some entities but not all<br />
of the California clients. The issue is the complexity of integrating the new reporting requirements into the<br />
existing system. There are two parts of knowledge that need to be considered. The first is the<br />
requirements that can be developed or presented by the customer. The second part is how to integrate<br />
the new requirements into the existing system. The structure of the existing program may not be capable<br />
of handling new information. The application may not have internal structures that can support the new<br />
requirements. Knowledge management theory which discusses lost or missing data can add some insight<br />
into the migration issue.<br />
The knowledge migration requirements have not been met in trying to implement the new system.<br />
3. Knowledge inventory<br />
Most organizations collect their knowledge in a written form. The policies and procedures can be clearly<br />
documented. One of the main issues would be how current the information is. The line staff would have<br />
the most current information. As an example, emergency management guidelines recommend that the<br />
facilities managers be the first people contacted in a facilities disaster. They will have the most up to date<br />
information about the contents of the buildings and staffing. The emergency operations documentation<br />
may not be kept current.<br />
From a KM point of view, training manuals are documentation of the organization’s knowledge. Many<br />
agency training manuals take a manufacturer’s and then customize them for the local environment. The<br />
customization is one piece of the knowledge inventory. Training videos can be a tool to capture<br />
knowledge via storytelling. NASA partnered with the NASA Engineering and Safety Center (NESC) to<br />
extract knowledge from NESC experts and incorporate it into targeted training courses. (KMEdge, 2010)<br />
It is also important for the organization to understand how to create the information. Meta-data or data<br />
about the creation or use of the data is also important. Data turns into knowledge with a context. The<br />
meta-data can provide the assumptions and background to describe the context. In a Geographical<br />
Systems (GIS) Environment, the meta-data describes the mathematical model or projection that is used<br />
to describe the spatial data. If an analyst does not know or validate the projection that information cannot<br />
be combined with other spatial data. Knowledge can also be context sensitive. The basis of that<br />
knowledge is based on information and assumptions. Should the knowledge’s base information change,<br />
the current knowledge would have to be verified against the current assumptions or environment.<br />
The organization may choose to outsource some of the procedures that it does not understand, is not a<br />
cost effective user, or decides that it is not part of the core operations. For the procedures that are<br />
purchased, the knowledge may not be owned or understood. From a programmer’s viewpoint, the<br />
knowledge is embedded in the application but it is explicit since the programmer can determine the<br />
relationships and procedures. From the user’s viewpoint, the knowledge is tacit within the application.<br />
Fuller in his book, Knowledge Management Foundations, points out that programmers are converting<br />
tacit, human knowledge into explicit, transferable intellectual property. (Fuller, 2002) The knowledge is<br />
the intellectual property of the developer. The user or owner of the data may or may not have the legal<br />
ability to see the code. Many license agreements have limitations on the deconstructing or reverse<br />
engineering the application code.<br />
To develop a knowledge inventory the organization needs to address some of the following points:<br />
The necessary steps that the organization uses to identify what kept knowledge is important.<br />
The organization needs to understand the key information points that make the organization run.<br />
The organization may be a straight forward processor of information such as an accounting<br />
department which has industry standards that dictate the majority of the knowledge processes.<br />
However, the organization needs to understand what makes up the culture of the organization.<br />
The organization needs to identify what are the items or knowledge points that make the organization<br />
unique. It needs to create a process or method to understand the information customizations or<br />
specialties.<br />
The organization should document the use of standard KM practices such as lists, rules, expert<br />
systems, or story-telling?<br />
138
Mitchell Cochran<br />
Many organizations cannot run without a central control system. The organization needs to develop<br />
procedures to be able to operate in a limited form if the technology is not available. The understanding or<br />
knowledge of the processes will be critical for maintaining a manual methodology. An organization should<br />
always be able to issue payments with paper checks for a short duration. A longer duration might create<br />
problems for tracking expenses, sales, deliveries or other business tracking methodologies.<br />
4. Strategic knowledge<br />
The organization needs to be able to identify the various units of knowledge as part of the inventory. The<br />
knowledge or business processes need to be broken down into logical groups. The groups can be as<br />
large as business units or as small as any functional concern that cannot be recovered if lost. The key<br />
issue is the organization needs to identify what knowledge can be recreated and the cost to recover it.<br />
The concept is to identify the knowledge that is global in nature or easily available outside the<br />
organization. From a black box type of operation which consists of an input, process and output, the<br />
process is not necessarily known. The issue is for the organization to be able to determine when the<br />
process is known or unknown and its importance.<br />
It is just as important to identify the knowledge that is unimportant to the organization. Many business<br />
functions could be outsourced such as human resources, payroll, billing, or accounting. Many small<br />
businesses use a PC based accounting package such as QuickBooks for those tasks. The knowledge<br />
needs to be identified but it can easily be recreated if necessary. It is important to realize what knowledge<br />
can be recreated or does not need to be maintained. A messenger process may require a vehicle but the<br />
type of vehicle may be immaterial.<br />
The knowledge inventory is going to require that the organization determine which units are proprietary<br />
and how they are maintained. The resources need to be identified that are necessary to maintain that<br />
knowledge. Some knowledge units will be obtained from outside of the organization. The sources,<br />
processes, assumptions, and necessary resources need to be identified.<br />
5. Migration issues<br />
As the organization documents and understands its knowledge base, it will be better prepared for<br />
maintaining or migrating that knowledge to new applications. The organization will have created<br />
documentation that describes the inputs, processes, and desired outputs of each knowledge unit. It will<br />
understand the data relationships or models that have been used in the applications. An organization<br />
should consider the following activities to aid in migration of knowledge:<br />
Develop and maintain a complete data dictionary. The data dictionary needs to include knowledge<br />
that is created or maintained both internal and external to the organization.<br />
Use application tools which allow for the methodology to be visible. Application toolsets or power<br />
tools can provide power users with the capability to modify the code. Since the application logic is<br />
visible, it can be understood and migrated.<br />
The organization needs to document and understand the meta-data or assumptions that validate the<br />
knowledge.<br />
Have the outside vendor contractually obligated to create reports which can be used to export data to<br />
new applications.<br />
Have a vendor document what information will be lost in moving to a new application. Reports can be<br />
generated prior to migration to allow for the capturing of information that may be lost.<br />
A common migration issue is moving information from one system into another where the fields do not<br />
have the same configuration. There is the potential for information to be lost if the description of the new<br />
data element is smaller than the existing one. Consider a case where the last name of an individual is<br />
moved from a description of 25 alpha numeric characters to a field that is limited to 20 characters. There<br />
is the potential that a last name will be incorrect due to truncation of the data element after the migration<br />
process.<br />
The migration effect needs to evaluate if there are any non-standard implementations. In the case of the<br />
City of Monrovia, certain unused data fields were ‘borrowed’ to define custom City information. The<br />
custom information helps to create a unique service or resource for the organization that creates it.<br />
139
Mitchell Cochran<br />
However, as the application is updated or migrated the vendor will not be aware of how non-standard<br />
implementations might impact new levels of code.<br />
6. Conclusion<br />
Many organizations follow standardized business processes. In general, they understand how their<br />
business works and what their industry expertise is. The issue is if the organization documents or<br />
understands what makes the business unique. Many organizations don’t have a detailed inventory of the<br />
knowledge that is used by the operations of the business. Many authors have discussed the issues of<br />
tacit knowledge within the workforce. The issue described in this paper is determining what knowledge is<br />
tacit within the applications. Much of that tacit knowledge may not be available during disaster recovery<br />
or during a migration to a new application. The new vendor may or may not have the same knowledge<br />
base. The onus is on the customer to determine what knowledge will be lost and what will need to be<br />
migrated or created; caveat emptor or buyer beware.<br />
The organization needs to develop a knowledge inventory of the processes it uses. The knowledge base<br />
needs to be broken down into units that can be treated as business functions. The knowledge is identified<br />
as coming from either internal or external sources. The importance of the knowledge units also needs to<br />
be identified.<br />
The detailed knowledge inventory can be used as a base to understand what migration efforts will be<br />
needed when moving to new applications.<br />
References<br />
Badaracco, J.(1991) The knowledge link, Harvard Business School Press, p. 79.<br />
Bots, P.W.G. and H. de Bruiin, (2002) “Effective Knowledge Management in Professional Organizations: Going by<br />
the rules,” 35th Hawaii International <strong>Conference</strong> on System Sciences, IEEE Computer Society Press.<br />
Buhler, Patricia, (2002) “Building the Learning Organization for the 21st Century: A Necessary Challenge.<br />
SuperVision Vol. 63. Iss.12 pp. 20.<br />
Fuller, S. (2002) Knowledge management Foundations, Butterworth and Heinemann, Boston.<br />
Grossman, Martin, (2006) “An Overview of Knowledge Management Assessment Approaches”, The Journal of<br />
American Academy of Business, Vol. 8, Num. 2, March.<br />
KMEdge.org, (2010) APQC, “Overcoming Knowledge Loss: Key Knowledge Retention and Transfer Approaches”,<br />
Kmedge.org/wp/overcomingknowledge3.html.<br />
Leonard, Dorothy and Kiron, David, (2002) “Managing Knowledge and Learning at NASA and the Jet Propulsion<br />
Laboratory (JPL)”, Harvard Business Review, Sept 27.<br />
McManus, Denise J., Synder, Charles A., and Wilson, Larry T. (2003) “The Knowledge Management Imperative,”<br />
DSS 2004 <strong>Conference</strong> Proceedings,<br />
http://www.knowledgeharvesting.com/documents/KnowledgeManagementImperative.pdf,.<br />
Nelson, Richard R., Winter, Sidney G. (1982),An Evolutionary theory of Economic Change, Cambridge, MA: Belknap.<br />
Rastogi, P.N., (2000)” Knowledge management and intellectual capital – the new virtuous reality of competitiveness”,<br />
Human Systems Management, Vol, 19 No. 1. pp39-48, January 01.<br />
Sage, A.P. and W.B. Rouse, (1999) “Information Systems Frontiers in Knowledge Management,” Information<br />
Systems Frontiers, Volume 1, Number 3, pp. 205-219.<br />
140
Motivations for IT Outsourcing in Public Sector Local<br />
Government<br />
Michael Cox, Martyn Roberts and John Walton<br />
University of Portsmouth Business School, UK<br />
Michael.Cox@port.ac.uk<br />
Martyn.Roberts@port.ac.uk<br />
John.Walton@port.ac.uk<br />
Abstract: This paper examines the approach taken to Information Technology (IT) outsourcing in four local<br />
government councils in the UK. This is important because, whilst outsourcing has become a significant issue in the<br />
restructuring of organisations and is increasingly used within both the private and public sectors, there has been a<br />
lack of research into IT outsourcing in the public sector and particularly within local government. This paper provides<br />
an in-depth study into how outsourcing is managed in local councils and how successful it has been; especially<br />
considering its sometimes controversial nature and the mixed press results it receives. To complete this study,<br />
interviews, containing both qualitative and quantitative questions, were conducted with key people at the four<br />
councils. These interviews examined the rationale for IT outsourcing. The findings from the interviews were then<br />
compared to the current literature on IT outsourcing to identify best practice. This research shows that, whilst cost<br />
savings remain important, councils focus on achieving best value when outsourcing IT rather than simply lowest cost.<br />
Indeed, it shows that whilst outsourcing can result in improved efficiency, councils that focus primarily on cost<br />
savings are often less successful. However, whilst the results revealed that IT outsourcing was more successful at<br />
councils who focused on long-term strategic goals, the interviewees considered the strategic benefits of outsourcing<br />
less important than improving the service. The structured selection process that is imposed by legislation allows<br />
council managers to gain a better understanding of the outsourcing requirements and make informed decisions to<br />
achieve best value, however the need for cost efficiency can result in a more short-term focus. The cost of the<br />
process and its inflexibility makes it more difficult for councils to focus on long-term goals. The study concludes that,<br />
whilst councils recognise that both the contract and trust are important to ensure that outsourcing is successful, the<br />
culture of risk aversion in the public sector tends to lead to a ‘play it safe’ mentality resulting in an overemphasis on<br />
the contract. This can lead to a short-term focus that could make it difficult for the council and the provider to work<br />
together to meet long-term goals. The councils were generally skeptical of developing partnerships; however, the<br />
research reveals that councils who focused predominantly on the contract were less successful than those who<br />
developed partnerships with their providers. The authors therefore recommend that, in order to achieve greater<br />
success, councils should develop partnerships and focus on best value and long-term strategic goals when<br />
outsourcing IT.<br />
Keywords: Information Technology (IT); Information Systems (IS); outsourcing; public sector; local government.<br />
1. Introduction<br />
The use of outsourcing, although not a new phenomenon, has increased in recent years as firms seek to<br />
lower costs and increase efficiency in response to higher levels of global competition. Child (2005, p.179)<br />
describes outsourcing as “the contracting out of activities that need to be undertaken on a regular basis,<br />
which otherwise would be conducted within an organisation”. Kern and Willcocks (2000, p.322) define IT<br />
outsourcing as “a decision taken by an organization to contract out or sell the organizations IT assets,<br />
people and/or activities to a 3 rd party supplier, who in exchange provides and manages assets and<br />
services for monetary returns over an agreed time period”. Outsourcing has become a significant issue in<br />
the restructuring of organisations and many commentators agree that it is currently “one of the fastestgrowing”<br />
and most important activities in business (Burnes & Anastasiadis, 2003, p.355; Weinert &<br />
Meyer, 2005, p.1).<br />
Outsourcing offers numerous advantages; however there is also evidence that outsourcing often fails.<br />
The potential benefits include cost savings, efficiency gains, improved flexibility, access to world-class<br />
expertise and focus on core competencies. However, outsourcing also poses numerous risks that must<br />
be managed in order for outsourcing to be successful (Kremic, 2006, p.467).<br />
Outsourcing is identified in two categories; outsourcing of core value-chain operations and outsourcing of<br />
support activities. The core value-chain operation is the outsourcing of supply chain activities such as<br />
distribution, whereas outsourcing of support activities includes HRM and facilities management (Child,<br />
2005, p.181). Outsourcing initially involved the outsourcing of non-core activities to reduce costs and<br />
improve efficiency and had a relatively short-term focus. However, more recently outsourcing has been<br />
used more strategically as firms increasingly seek to become more efficient by focusing on areas where<br />
they can achieve competitive advantage (DiRomualdo & Gurbaxani, 1998, p.1). However, some theorists<br />
141
Michael Cox et al.<br />
suggest that you should not outsource your core competency, whilst others argue that it can be difficult to<br />
define what is core and non-core. Alexander and Young (1996), cited by Que’lin and Duhamel (2003,<br />
p.648), distinguish between activities that “are critical to performance” but only support the core and<br />
those that actually provide competitive advantage.<br />
This is a notable problem when outsourcing IT because, whilst some consider it a non-core activity, IT is<br />
becoming critical to service delivery. The fact that IT is complex and affects all activities within an<br />
organisation make it one of the most important activities outsourced (Weinert & Meyer, 2005, p.1). Lacity<br />
and Willcocks (1995, p.226) note that IT changes rapidly and switching costs to new technologies or<br />
suppliers can be high because IT demands high investment costs. They suggest this makes IT<br />
outsourcing unique compared to the outsourcing of other activities. This will have implications for how IT<br />
outsourcing is conducted and managed.<br />
1.1 Outsourcing in the public sector<br />
The public sector has followed the trend set by the private sector and IT outsourcing is now one of the<br />
most important outsourced activities that by “1996 had led to contracts worth £2 billion” (Lacity &<br />
Willcocks, 1997, p.85).<br />
In the 1980’s the government introduced Compulsory Competitive Tendering (CCT), this “involves<br />
government or firms using a competitive bidding process to help decide who should have the right to<br />
produce or deliver goods or services” within the public sector (Rimmer, 1994, p.79). This process was<br />
driven by “marketing and political trends” with the aim of making the public sector more competitive<br />
(Marco-Simó, et al, 2007, p.51). It was assumed that market-based competition would be able to provide<br />
improved efficiency and lower cost whereas public monopolies provided no incentives to improve<br />
performance (Besley & Ghatak, 2003). However, traditional outsourcing contracts in the public sector<br />
were often granted to the lowest priced bidder, which can lead to lower quality provision. Thus, the<br />
government updated CCT in 1998 to focus on best value rather than simply lowest cost, whether<br />
internally or externally (Butler, 2003).<br />
However, whilst outsourcing has become one of the most important tools in public sector management,<br />
failures tend to attract greater publicity, due to political accountability, whereas successes often go<br />
unnoticed (Maughan, 2003). These failures have led to an increasing interest in how outsourcing is<br />
managed in local councils and to what extent it has been successful.<br />
2. Literature review<br />
Much of the literature available deals with IT outsourcing in general without specifically addressing cases<br />
within the public sector or local councils. Indeed, Vilvovsky (2008, p.337) notes that the available<br />
research is “limited and fragmented”. De Looff (1996) finds this interesting since he shows that in the<br />
Netherlands 30% of IT outsourcing is accounted for by the public sector, a figure not too dissimilar to the<br />
UK. However, despite this, Marco-Simó et al (2007, p.52) note that outsourcing in the public sector “has<br />
not produced a level of research interest proportional to its economic importance”. This has led to an<br />
increased interest in how outsourcing is managed within the public sector; especially considering the<br />
increasing use of IT outsourcing within the public sector the mixed press results it receives.<br />
The lack of research in this area can be problematic for making comparisons. However, it does provide<br />
an opportunity to explore a gap in the literature and to discover if anything can be learnt from how<br />
outsourcing is managed in local councils. This is important because whilst the private and public sectors<br />
share some similarities they also have some notable differences. Vilvovsky (2008, p.338) notes that the<br />
public and private sectors are divided both “ideologically and operationally”. Both have similar values in<br />
terms of “responsiveness, honesty and accountability” but the public sector is not motivated primarily by<br />
financial performance. Thus, public sector managers may adopt a different mindset to outsourcing, based<br />
on goals and values other than cost efficiency, whilst being further constrained by political accountability.<br />
This could result in a more cautious approach to outsourcing based on minimising risk. Burnes and<br />
Anastasiadis (2003, p.365), in their comparison of a police force and an insurance company, further<br />
highlight differences in supplier selection, contractual arrangements and in the management of the<br />
relationship between the public and private sectors. One major difference is that public agencies are<br />
legally required to outsource any activity that can be done cheaper by outside contractors, as long as<br />
they meet the minimum specifications (Burnes & Anastasiadis, 2003, p.359). The private sector, in<br />
142
Michael Cox et al.<br />
contrast, has more choice over whether they outsource and to whom and can therefore do so more<br />
strategically and approach the best suppliers directly (Lin, et al, 2007, p.164).<br />
Currie (1996, p.226), in a comparison of the experiences of IT outsourcing in the UK’s private and public<br />
sectors, shows that IT outsourcing poses significant challenges to both and should not simply be seen as<br />
a “quick-fix” solution. Lacity and Willcocks (1997, p.106) investigated two case studies in the US public<br />
service. They have previously suggested that the UK has had “long standing problems” with outsourcing<br />
and that the results have often fallen short of expectations. In this article, they suggest that politicians<br />
often believe that outsourcing will automatically save money whereas in reality outsourcing either has<br />
failed or has yet to show any real benefits.<br />
Frederick (1994), however, contends that CCT has been successful and has produced both cost savings<br />
and increased efficiency. Indeed, Maughan (2003) suggests that the public sector is better at managing<br />
outsourcing relationships than the private sector because they devote staff exclusively to managing the<br />
relationship whilst the private sector generally does not which can create a lack of focus and clarity. He<br />
also states that the public sector benefit from having clear guidelines and a lengthy procurement process,<br />
which allows for multiple tenders, imposed upon them by legislation. Vilvovsky (2008, p.342) contends<br />
that public sector managers often find legislation and guidelines “a legal constraint rather than an aid”<br />
which may create inflexibility and lead to contracts being awarded to the lowest cost bidder rather than<br />
the best quality. Despite this, Burnes and Anastasiadis (2003, p.365) argue that the private sector could<br />
learn from the public sector’s structured approach. They suggest that government guidelines can create a<br />
greater understanding of outsourcing requirements and may produce more stable relationships. This<br />
research explores these issues further.<br />
3. Methodology<br />
This research takes the form of a multiple case study. Creswell defines a case study as “a single,<br />
bounded entity, studied in detail, with a variety of methods, over an extended period” (Maylor &<br />
Blackmon, 2005, p.243). Marco-Simó et al (2007, p.59) have argued that we need more “real cases”<br />
before “we can generalise the findings”. A case study approach was chosen because this type of<br />
research is useful when ‘how’ or ‘why’ questions are being asked (Saunders et al, 2007, p.139).<br />
The case study allows an in-depth examination to develop an enhanced understanding of the subject<br />
area and develops areas for further research (Flyvbjerg, 2004, p.420). Commentators agree that this<br />
method allows for a greater understanding of complex issues and is more comprehensive than surveys,<br />
due to its ability to “come closer….to the complexity of real organisation setting” (Maylor & Blackmon,<br />
2005, p.242).<br />
Yin (1984), cited by Gable (1994, p.113-115), believes that a single case study is useful in areas that are<br />
under-researched and require exploration, whereas multiple case studies are used to test for patterns<br />
and draw comparisons.<br />
The purpose of this research was to explore a complex issue in-depth and to identify some common<br />
trends rather than to present a generalised industry view. Indeed, Flyvbjerg (2004, p.420) argues that the<br />
case study is a reliable method in itself due to its ability to incorporate multiple data collection methods.<br />
3.1 Case studies<br />
The case studies are based on four local government councils in the south of the UK.<br />
Table 1: Description of the cases<br />
Council Description<br />
A Council A is a district council that serves a population of around 110,000 people. IT supports<br />
the council’s everyday operations and is critical to business performance and service delivery.<br />
IT is one of the most important activities outsourced at this council. A budget report meeting in<br />
2008 revealed that the contract for IT services is the second most expensive contract at the<br />
council, at over £700,000 per year. The ICT strategy at the council is not only to improve the<br />
efficiency and effectiveness of the service but also to increase its accessibility to all.<br />
The Council signed a 5-year contract with their service provider in 2005, worth £2.5 million.<br />
This contract was to provide the council with a full IT managed service that included desktops,<br />
servers, infrastructure and applications support.<br />
143
Michael Cox et al.<br />
Council Description<br />
B Council B is a unitary authority that serves a population of around 200,000 people. Council B<br />
chose to outsource to one supplier in a 10-year contract. This contract included not only the<br />
provision of IT but also related services such as HR, payroll, Customer Service, local taxation<br />
and benefits, procurement and property services. The council sought one strategic partner to<br />
ensure greater integration between the councils services but also because greater cost<br />
reductions could be achieved by outsourcing many services to the same provider. The main<br />
reason that the council opted for a long-term agreement was because of the high investment<br />
needed at the start of the agreement. This meant that the council needed to allow sufficient<br />
time for the provider to recoup their investment costs.<br />
C Council C is a unitary authority that serves a population of 150,000. This council outsourced for<br />
seven years. The council outsourced its entire IT department. Services provided include<br />
procurement, technology refresh, network development and support, offsite server hosting<br />
support and offsite mainframe services, as well as application support and management, 24x7<br />
services monitoring and helpdesk.<br />
D Council D is a district council that serves a population of 110,000. Council D do not outsource<br />
IT.<br />
3.2 Data collection methods<br />
To answer the research questions semi-structured interviews were conducted at four local councils. The<br />
interviews were held with the heads of IT at the councils. The interviews were the primary data collection<br />
method and allowed the authors to explore the interviewees’ experience and attitudes towards IT<br />
outsourcing. This paper presents a summary of the data collected.<br />
4. Analysis of results<br />
4.1 Motivations<br />
Table 2 summarizes the main issues as described by the four interviewees at the four councils.<br />
Table 2: Summary of responses<br />
Council A In 1998 Recruitment issues were the main reason to outsource because the council had difficulty<br />
attracting and retaining staff. However, this became less important after the 2005 outsourcing<br />
agreement began. Cost savings and improved quality were rated as the second and third most<br />
important reasons to outsource. However, outsourcing was also designed to set the strategic<br />
direction and improve the council’s responsiveness to the public.<br />
Council B The council primarily outsourced to improve quality and to gain much needed investment. Cost<br />
savings, however, were considered the second most important reason because the service had<br />
become costly to maintain. The council also outsourced to focus on strategy but this was<br />
considered the sixth most important reason.<br />
Council C Recruitment issues were the most important motivation to outsource IT, with improving quality and<br />
gaining access to expertise second and third. Cost savings were unimportant because the council<br />
had to initially pay more for the outsourced service to improve quality. The council also wanted<br />
outsourcing to support strategy, which is one reason they opted for a partnership agreement.<br />
Council D This council does not outsource IT. However, cost savings are considered the most important<br />
reason that councils outsource, but, in order of importance, cost savings, access to expertise,<br />
focus on core and improving quality are other important reasons. Also that councils outsource to<br />
pass on the responsibility and risk of managing the IT service.<br />
The interviewees from the four councils were asked to asked to rank the seven motivational criteria (7<br />
being the most important, 1 the least). Table 3 and Figure 1 illustrate the responses.<br />
144
Michael Cox et al.<br />
Table 3: Interviewees ranking of main motivations<br />
Interviewee A Interviewee B Interviewee C Interviewee D<br />
Aggregate<br />
Score<br />
External expertise 2 4 5 6 27<br />
Improved Service Quality 5 7 6 4 22<br />
Cost savings 6 6 2 7 21<br />
Cost Restructure 1 5 1 3 10<br />
Focus on core 4 2 4 5 15<br />
Increased flexibility 3 3 3 1 10<br />
Recruitment issues 7 1 7 2 17<br />
25<br />
20<br />
15<br />
10<br />
5<br />
0<br />
External<br />
Expertise<br />
The motivations for IT outsourcing<br />
Improved<br />
service<br />
quality<br />
Aggregate ranking<br />
Cost<br />
savings<br />
cost<br />
restructur<br />
e<br />
Focus on<br />
Core<br />
Flexibility Recruitme<br />
nt issues<br />
Aggregate ranking 17 2 2 2 1 10 15 10 17<br />
Figure 1: Summary of individual rankings<br />
In Table 4 and Figure 2 the interviewees from the four councils were asked to state whether the<br />
motivations were: Very Important, Important, Unimportant or Irrelevant to their decision to outsource.<br />
Table 4: Interviewees categorization of motivations<br />
Interviewee A Interviewee B Interviewee C Interviewee D<br />
External expertise Unimportant Important Important Important<br />
Improved Service<br />
Quality<br />
Important Very important Important Important<br />
Cost savings Important Very important Unimportant Very important<br />
Cost Restructure Irrelevant Important Irrelevant Unimportant<br />
Focus on core<br />
competencies<br />
Important Important Important Important<br />
Increased flexibility Important Important Unimportant Unimportant<br />
Recruitment issues Very important Unimportant Very important Unimportant<br />
145
Recruitment Issues<br />
Flexibility<br />
Focus on Core<br />
Cost Restructure<br />
Cost Savings<br />
Improved Service Quality<br />
External Expertise<br />
Michael Cox et al.<br />
The importance of the main motivational factors<br />
0% 50% 100%<br />
Figure 2: Importance of motivational factors<br />
4.2 Discussion<br />
4.2.1 Improved quality and cost reduction<br />
Very Important<br />
Important<br />
Unimportant<br />
Irrelevant<br />
Don’t Know<br />
This research shows that quality improvement and cost savings were the most important reasons to<br />
outsource IT at local councils (see Figure 1). Although cost savings were a major reason for outsourcing<br />
IT, improvements in quality of service often took priority meaning that cost savings could not always be<br />
achieved immediately.<br />
Several of the councils believed that they could have improved quality and reduced costs in-house given<br />
time, but that results can be achieved far quicker through outsourcing.<br />
Councils that focused excessively on price can result in the provider cutting back in other areas such as<br />
innovation and supporting the council strategically and have a negative long-term impact.<br />
Several councils sought to achieve best value. However, these cases showed that a focus on long-term<br />
goals and building a working relationship can lead to better results.<br />
Thus, whilst cost savings are important, councils should adopt a balanced approach to outsourcing to<br />
achieve best value rather than simply lowest cost. Indeed, this research shows that the results of<br />
outsourcing are not instantaneous and supports Daly’s conclusion (2001, p.13) that those that primarily<br />
focus on cost savings are often less successful than those who target “more value adding activities”.<br />
4.2.2 Access to expertise<br />
Access to external expertise was considered an important reason to outsource IT by three Interviewees<br />
and was overall considered the third most important reason. This is because IT now requires a range of<br />
different skills, not all of which councils can provide internally, and is constantly changing. In contrast,<br />
service providers specialise in IT and can therefore call upon a larger pool of skilled workers (Vilvovsky,<br />
2008, p.338).<br />
A number of interviewees commented that whilst gaining access to expertise is important, councils need<br />
to retain key skills in-house.<br />
146
4.2.3 Recruitment issues<br />
Michael Cox et al.<br />
Recruitment issues were the most important motivation for outsourcing in local government 10 years ago.<br />
This is evident at Council A and Council C, who both outsourced initially around 2000, who found that the<br />
inability to attract skilled staff, due to higher pay in the private sector, was the main reason to outsource.<br />
This was partly because the millennium bug had created a shortage in the supply of skilled IT staff and<br />
pushed wages up.<br />
Recruitment issues now appear to be less important. This could be because councils are beginning to<br />
recognise the importance of IT and the need to offer higher wages to attract skilled staff and have<br />
become more commercially orientated.<br />
4.2.4 Focus on core competence<br />
The councils also outsourced IT to transfer the responsibility and risk of managing routine activities so<br />
that the in-house team could spend more time and energy focusing on strategy. This allows councils to<br />
transfer responsibility for service failures, ongoing costs, and staffing whilst the council only have to<br />
manage the contract. All Interviewees considered this an important reason to outsource, although as<br />
Figure 2 shows none considered it the most important reason. This illustrates that improving the service<br />
by reducing costs and improving quality takes priority over strategic considerations at local councils.<br />
Indeed, both Interviewee A and B support Lacity and Hirschheim’s (1993, p.260) conclusion that IT is<br />
more complicated than other support functions that are traditionally outsourced, and therefore it takes<br />
longer than anticipated managing the contract and dealing with problems, thus resulting in less time to<br />
focus on strategy.<br />
The councils, however, are becoming increasingly reliant on IT and see IT as the key to transforming<br />
service delivery. Indeed, whilst some private companies may consider IT a support function, IT at local<br />
councils is critical to service delivery. The councils, thus, expected outsourcing to support strategy.<br />
Council B, for instance, signed a long-term Strategic Services Partnership (SSP) with their provider,<br />
which was designed to create greater integration between departments and to transform service delivery<br />
(Whitfield, 2008, p.3). Council C, similarly, wanted their provider to support them strategically. The<br />
councils that opted for long-term contracts concentrated more on improving quality and transforming the<br />
service. However, even Council A, who outsourced primarily to reduce costs, expected outsourcing to<br />
support strategy. Indeed, Interviewee A noted that the lack of strategic support was one of the main<br />
reasons outsourcing did not meet expectations.<br />
The use of outsourcing to support strategy could explain the recent increase in SSPs, signed between<br />
councils and service providers (Whitfield, 2008, p.4), and the increasing focus on shared services. This is<br />
where councils come together to pool resources, such as combined data centres, to achieve higher levels<br />
of efficiency in-house. This, combined with the recruitment of private sector managers with commercial<br />
experience, can give councils greater leverage when outsourcing and allow them to focus more on<br />
strategy.<br />
4.2.5 Flexibility and restructured costs<br />
The interviewees found that it was difficult to write flexible contracts and changes were subsequently<br />
difficult and costly to negotiate unless they were in the provider’s interest. One council reported improved<br />
flexibility. However, they only achieved flexibility after putting the contract aside and developing a<br />
partnership. This suggests that partnerships based on mutual benefit will make the provider more open to<br />
changes.<br />
Furthermore, whilst Gartner (2008) suggest that “changing the cost structure of IT” will allow firms to<br />
respond more quickly to changing needs, the councils did not consider cost restructuring an important<br />
reason to outsource IT. The interviewees acknowledged the benefits of paying only a monthly service<br />
charge. However, only one council outsourced to gain investment because they had suffered from<br />
numerous years of under-investment and needed investment to transform the service. This allowed the<br />
council to gain immediate investment but pay it off over the length of the agreement. This gives the<br />
provider greater incentive to improve the service although the council had to sign a long-term contract to<br />
enable the provider to recoup their investment costs.<br />
147
5. Conclusion<br />
Michael Cox et al.<br />
The literature review suggested that private companies are focusing less on cost reductions when<br />
outsourcing (DiRomualdo & Gurbaxani, 1998, p.3), whereas with local government cost savings remain<br />
very important to councils due to budget constraints. However, whilst Child (2005, p.187) argues that<br />
outsourcing offers “significant, often immediate cost savings”, this was not the case at two of the three<br />
councils that outsourced. This is because, as Daly (2001, p.12) notes, councils are now focusing on<br />
achieving best value and recognise that improving the quality of the service has an impact on price. This<br />
research supports Daly (2001, p.13) and Vilvovsky’s (2008, p.342) conclusion that those that focus<br />
primarily on cost savings are less successful than those who target more value adding activities.<br />
Although Hart et al (1997, p.1128) argue that cost reductions often take priority because quality is difficult<br />
to specify within contracts, two of the three councils that outsourced achieved greater success by<br />
balancing cost savings with improving quality to achieve best value (Daly, 2001, p.12).<br />
Indeed, whilst Lacity and Willcocks (1997, p.106) argue that outsourcing has often fallen short of<br />
expectations in the public sector, the three councils that outsourced all noted that outsourcing had<br />
improved quality and produced cost savings. This supports Frederick’s (1994) conclusion that CCT has<br />
been successful and has produced both cost savings and increased efficiency. However, whilst<br />
DiRomualdo and Gurbaxani (1998, p.3) argue that private sector organisations are focusing more on the<br />
strategic benefits of outsourcing, this research shows that improving quality and reducing costs are the<br />
most important reasons to outsource IT at local councils.<br />
The councils, did however, outsource to gain access to external expertise because, as Vilvovsky (2008,<br />
p.338) observes, IT has become increasingly complicated and councils do not have all the necessary<br />
skills in-house. The providers, in contrast, specialise in IT and can call upon a larger pool of skilled<br />
workers. However, the councils retained small in-house IT teams with key skills to allow them to develop<br />
new ideas and support strategy to reduce their dependence on their provider. Furthermore, recruitment<br />
issues are no longer the most important issues at local councils. Although Vilvovsky (2008, p.338) argues<br />
that the public sector pays lower wages, this research shows that councils are beginning to recognise the<br />
importance of IT and are becoming more commercially orientated and therefore willing to offer higher<br />
wages to attract the best workers.<br />
Finally, the research shows that significant time and effort needs to be taken to manage the service after<br />
outsourcing, and that outsourcing is seen very much as a way of making organizations responsive to<br />
business change.<br />
References<br />
Besley, T., & Ghatak, M. (2003). Incentives, choice and Accountability in the Provision of Public Services (the<br />
Institute for Fiscal Studies WP03/08). Retrieved April 8, 2009, from the Institute of Fiscal Studies website:<br />
http://www.ifs.org.uk/wps/wp0308.pdf.<br />
Burnes, B., & Anastasiadis, A. (2003). Outsourcing: a public-private sector comparison [Electronic version]. Supply<br />
Chain Management: An International Journal, 8 (4), 355–366.<br />
Butler, P. (2003). Timeline: outsourcing and the public sector. Retrieved April 17, 2009, from Guardian website:<br />
http://society.guardian.co.uk/microsite/outsourcing_/story/0,13230,933819,00.html.<br />
Child, J. (2005). Organization: contemporary principles and practice. Oxford: Blackwell.<br />
Currie, W. L. (1996). Outsourcing in the private and public sectors: an unpredictable IT strategy [Electronic version].<br />
<strong>European</strong> Journal of Information Systems, 4, 226–236.<br />
Daly, G. (2001). Outsourcing in Government – pathways to value, Retrieved July 26, 2009, from the Accenture<br />
website: http://www.accenture.com/NR/rdonlyres/438B5454-BEED-499F-A94E-<br />
2CFF389F2890/0/outsourcing.pdf.<br />
De Looff, L. A. (1996) IS outsourcing by Public Sector organisations. Paper presented at the International Federation<br />
for Information Processing World <strong>Conference</strong> on Advanced IT Tools, Australia. Retrieved April 2, 2009, from<br />
http://www.acs.org.au/president/1996/ifip96/i96iso.htm.<br />
DiRomualdo, A., & Gurbaxani, V. (1998). Strategic Intent for IT Outsourcing [Electronic version]. Sloan Management<br />
Review, 39 (4), 1-16.<br />
Flyvbjerg, B. (2004). Five misunderstandings about case-study research. In C. Seale, G. Gobo, J. Gubrium, & D.<br />
Silverman (eds.), Qualitative research practice (pp.420-434). London: SAGE.<br />
Frederick, D. (1994). Why Compulsory Competitive Tendering for Local Government Services is not as good as<br />
Privatisation. Retrieved April 6, 2009, from http://www.libertarian.co.uk/lapubs/econn/econn052.pdf.<br />
Gable, G. (1994). Integrating case study and survey research methods: an example in information systems<br />
[Electronic version]. <strong>European</strong> Journal of Information Systems, 3(2), 112-126..<br />
Gartner. (2008). Gartner says changing the cost structure of IT will become a business imperative for most CIOs.<br />
Retrieved June 15, 2009, from Gartner website: http://www.gartner.com/it/page.jsp?id=776812.<br />
148
Michael Cox et al.<br />
Hart, M. (2008). Is short-termism holding back public sector outsourcing? Retrieved August 19, 2009, from the Silicon<br />
website: http://www.silicon.com/publicsector/0,3800010403,39273472,00.htm..<br />
Kern, T., & Willcocks, L. (2000). Exploring information technology outsourcing relationships: theory and practice<br />
[Electronic version]. Journal of Strategic Information Systems, 9, 321-350.<br />
Kremic, T. (2006). Outsourcing decision support: a survey of benefits, risks, and decision factors [Electronic version].<br />
Supply Chain Management: An International Journal, 11 (6), 467–482.<br />
Lacity, M. C., & Hirschheim, R. (1993). Information Systems Outsourcing: Myths, Metaphors and Realities.<br />
Chichester: John Wiley & Sons Ltd.<br />
Lacity, M. C., & Willcocks, L. P. (1995). Interpreting Information Technology Sourcing Decisions From A Transaction<br />
Cost Perspective: Findings and Critique [Electronic version]. Accounting, Management and Information<br />
Technology, 5 (3/4), 203-244.<br />
Lacity, M.C., & Willcocks, L. (1997). Information systems sourcing: examining the privatization option in USA public<br />
administration [Electronic version]. Information Systems Journal, 7, 85–108.<br />
Lin, C., Pervan, G., & McDermid, D. (2007). Issues and recommendations in evaluating and managing the benefits of<br />
public sector IS/IT Outsourcing [Electronic version]. Information Technology & People, 20 (2), 161-183.<br />
Marco-Simó, J M., Macau-Nadal, R., & Pastor-Collado, J. A. (2007). Information Systems Outsourcing in Public<br />
Administration: An Emergent Research Topic. Paper presented at the <strong>European</strong> and Mediterranean<br />
<strong>Conference</strong> on Information Systems. (pp.51-61).<br />
Maughan, A. (2003). Stop laughing and start learning from public sector outsourcing. Retrieved April 8, 2009, from<br />
the Silicon website: http://comment.silicon.com/0,39024711,10003991,00.htm.<br />
Maylor, H., & Blackmon, K. (2005). Researching business and management. Basingstoke: Palgrave Macmillan.<br />
Que’Lin, B., & Duhamel, F. (2003). Bringing Together Strategic Outsourcing and Corporate Strategy: Outsourcing<br />
Motives and Risks [Electronic version]. <strong>European</strong> Management Journal, 21 (5), 647–661.<br />
Rimmer, S. J. (1994). Competitive Tendering and Contracting: Theory and Research [Electronic version]. The<br />
Australian Economic Review, 27 (3), 79-85.<br />
Saunders, M., Lewis, P., & Thornhill, A. (2007). Research methods for business students (4 th ed.). Harlow: Financial<br />
Times Prentice Hall.<br />
Vilvovsky, S. (2008). Differences between public and private IT outsourcing: common themes in the literature. Paper<br />
presented at the Proceedings of the 2008 international conference on Digital government research,<br />
International <strong>Conference</strong> Proceeding Series (337-346). Montreal, Canada.<br />
Weinert, S., & Meyer, K. (2005). The evolution of IT outsourcing: From its origins to current and future trends.<br />
Whitfield, D. (2008). PPP DATABASE: Strategic Service-Delivery Partnerships for local authority ICT, corporate and<br />
technical services in Britain.<br />
149
Documenting Innovation: A Methodological Proposal and<br />
Application<br />
Stefano De Falco<br />
University of Naples, Federico II, Italy<br />
sdefalco@unina.it<br />
Abstract: Nowadays the business is driven by a constant and ever-changing demand for innovation. In order to<br />
remain competitive the companies must know how to present new products, processes and services to respond<br />
more quickly and effectively to the needs of consumers. Today as never before, companies have faced a similar<br />
challenge, to hone the best of his ability to innovate. In this highly competitive business environment, only those with<br />
managers and employees highly motivated and prepared, and can count on a creative process properly, it can be<br />
said to be ready to conquer success. However, if this is the scenario in which to move in order to achieve the<br />
objectives of efficaciousness, managers also need to consider efficiency aspects. Often the actions and the projects<br />
designed to introduce elements of innovation within companies require a lot of energy. We just refer to this term<br />
borrowed from physics that makes the concept of power’s absorption for a certain time corresponding to the different<br />
phases of the innovation process with regard to human resources, construction, infrastructure, equipment and<br />
financial requirements involved in the innovation process. But this power’s consumption can be optimized and thus<br />
reduced, with the same results (or to remain in physical jargon) of charge to be met, through a system of<br />
documentation of the actions of innovation introduced over time by the company. In this paper a methodology for<br />
documenting innovation, as a strategic key factor for firms, is developed and practical application is provided through<br />
the contribution of a young company working on information systems, the “Solvendo Company”(www.solvendo.net)<br />
for the construction of an information system finalized to innovation’s documentation. It is also offered an application<br />
based on RFID, radio frequency devices, for the process of data acquisition instrumentation, machinery, plant, in<br />
order to have a quick reference tool for operating on a single component that aggregates the “query- innovation<br />
processes “ activity over time.<br />
Keywords: innovation, tracking, tracciability, rintracciability, Rfid<br />
1. Documenting Innovation: A technology overview<br />
The analysis of the competitive landscape in which companies move, has highlighted the increasingly<br />
urgent need, in this global economy, to trace the development of innovative activities from within the<br />
company to the end customer, helping to control the production and quality , attesting the originality of the<br />
product and its features, certifying that the product / service has been built following the existing rules on<br />
employment protection and health, and enabling real-time communication between businesses and<br />
business customers in partnership. This need can be answered only using the technology infrastructure<br />
in terms of all modern tracking and communications devices. (Top Management Forum on Strategic<br />
Management of Technology and Innovation 2010).<br />
This approach allows managers to start always from a step higher each time there is a new innovation<br />
process, avoiding to start from scratch with no memory of previous interventions.<br />
To frame the issue in the context of this analysis, in the following a summary of the opportunities that<br />
ICTs offer for SMEs to innovate their processes, is reported.<br />
The evolution and integration among the various ICT (Web, Mobile & Wireless, RFID, etc.) always has an<br />
impact more significant and pervasive within all main processes of the enterprise value chain, including<br />
the most complex and cross (Supply Chain Management, Customer Relationship Management,<br />
Knowledge Management, etc..) and provides new channels of communication with customers of the<br />
company as the possibility of communication via the Web, online marketing tools, the Mobile CRM<br />
services, etc.. that are profoundly changing the relationships between the company and its customers.<br />
But these benefits are exploited only if you have a system to document over time the improvements<br />
offered by the technology that determines them.<br />
The development of applications mentioned above, however, is conditioned by the necessity of<br />
adjustment of basic infrastructure:<br />
Infrastructure hardware<br />
This infrastructure consists of:<br />
150
Stefano De Falco<br />
Systems used to collect and manage information, access to applications and personal productivity,<br />
such as servers, clients, mobile devices, etc.<br />
The basic software and environments found in server systems, such as operating systems, software<br />
for database management (DBMS) and application server environments<br />
(For example, the Web server enables you to publish web applications);<br />
Peripheral devices (printers, scanners, plotters, etc. RFID transponders and readers.)<br />
Systems designed to ensure reliability of services and data management, enable the automatic<br />
replication of critical data on both the server and / or client (backup) or storage and sharing through<br />
the use of large volumes of network data.<br />
Security systems<br />
Even the SMEs in their computerization process, can not neglect the security aspect both provided by the<br />
instruments "centralized" (for the protection of servers and shared computing resources) and both for the<br />
protection of the resources used by individual users.<br />
Communications infrastructure<br />
Systems Voice over IP (VoIP) have become increasingly important due to a number of the benefits these<br />
systems can offer, in terms of reducing the costs of new advanced services enabled. The types of VoIP<br />
systems used by the company are essentially three types:<br />
VoIP on PC is the use of communications software (eg Skype, MS Messenger, etc..) Installed on<br />
each PC connected to the Internet. The installation is free and simple to use;<br />
Systems-based adapters, adapters are used, connected to the line data and analog phones to the<br />
company, making the conversion from analog to digital. The approach turns out to be so cheap for<br />
the company, which can preserve their traditional phones, but sometimes with poor quality<br />
communication;<br />
PBX systems based on IP-PBX, PBXs same mode allowing connectivity to mobile VoIP over the<br />
traditional to the new digital phones. This solution enables you to preserve the investments already<br />
incurred and to migrate in a flexible and gradual to VoIP technology.<br />
The benefits from the introduction of VoIP systems consist in the reduction of costs:<br />
Communication costs, because calls between IP-connected sites are free, while calls to the outside<br />
world does not have significant additional costs compared to the use of data network and there is<br />
also the possibility of using GSM dual mode mobile / Wi -fi within the walls of company headquarters,<br />
with a further reduction of the cost of mobile telephony;<br />
Infrastructure-management costs, thanks to a general simplification of systems and services and the<br />
use of a single infrastructure for voice and data and the ability to use advanced services such as:<br />
Video conferencing costs (at significantly lower costs than traditional systems);<br />
Messaging costs (the ability to manage e-mail address in an integrated service with voice mail) and<br />
the ability to share files and documents while communicating.<br />
Business applications can be distinguished:<br />
Management systems;<br />
Business intelligence applications;<br />
Management systems<br />
a) simple software package.<br />
Functionality is relatively limited and generalized (administration, warehouse, manufacturing). These<br />
packages can be integrated with significant customization is needed to extend the functional coverage in<br />
some areas of the company or to accompany the growth in size and complexity of the undertaking.<br />
b) vertical systems and / or developed ad hoc.<br />
Are software applications developed for the specific needs of the enterprise.<br />
151
Stefano De Falco<br />
These applications are usually run on proprietary architectures. The development is generally performed<br />
by software houses side by side, in the case of more structured companies, by the internal IT function.<br />
These systems offer a good level of support built in, but having been developed over time, are often<br />
based on outdated technology. Some software companies linked to specific industrial and regional<br />
contexts have also developed packages "vertical" that, based on a solid knowledge sector, it reduces the<br />
risk of adoption, in terms of time and budget, from customers.<br />
c) Systems with a high level of functional integration.<br />
They are made by software houses national (or domestic), which deal directly with end customers install<br />
or use a limited number of partners. Distinguishing factor of these applications and adherence to the<br />
specific regulatory, operational and business of the Italian context.<br />
At the most complete enterprise integration are the ERP systems offered by IT companies that typically<br />
operate on international markets (eg SAP) ERP systems that are capable of supporting the whole key<br />
business processes in an integrated manner.<br />
The introduction of an integrated management system pushes the company to analyze their processes,<br />
and this is the first step to identify opportunities for improvement. Only by learning to structure the<br />
activities and information flows associated with them, and then introducing operating logic of explicit<br />
processes, the company may lay the basis for identifying indicators of efficiency and effectiveness on<br />
which then act to increase their level of competitiveness to improve the performance parameters<br />
(reduction of delivery time, increase productivity in back-office activities (administration, order<br />
management, purchasing management, etc..), reduction of errors in direct activities (production,<br />
deliveries, sales, etc..) and indirect.<br />
Generally, the management system plays an important role in supporting the change in the company.<br />
From this it appears evident the importance of flexibility in the management system adopted. As more the<br />
system is prepared for the expansion of its functional coverage, the more the management system<br />
becomes an asset for the company to guarantee flexibility (adaptation to the logical management) in<br />
terms of opening opportunities for change offered by evolving business scenarios.<br />
Applications of Business Intelligence (BI)<br />
BI applications are intended to exploit the massive amounts of data (commercial transactions, financial<br />
and administrative, navigation patterns on web email, text and hypertext, etc..) available to the company.<br />
These applications are based theTechnologies DWH / ETL for the collection and storage should be off<br />
line of dates collected, provide access to the dates from different funds to analyze, structure, calculation<br />
of appropriate indicators and generate reports and dashboards with the aim of supporting better<br />
decisions by management enterprise.<br />
Examples of applications are:<br />
Analysis of the overall performance of the company: they allow the employer and / or top<br />
management to monitor key performance (as measured through appropriate KPIs) of the enterprise<br />
as a whole and the level of individual business units or corporate functions;<br />
Business analysis: to analyze the dates for customers, sales and marketing;<br />
Financial analysis, to control the flow of cash input / output, trade receivables and payables and any<br />
critical issues identified and anticipated;<br />
Analysis of suppliers to analyze the dates relating to the supply chain (procurement and inbound<br />
logistics), measuring the basic performance (number of suppliers, costs, delivery times etc..).<br />
The benefits of BI applications, therefore, consist essentially offering the user to quickly and appropriately<br />
to their needs decision-making information and data in large quantities but easily accessible, allowing you<br />
to monitor over time the outcomes of decisions. These features are not offered by "traditional" software<br />
because:<br />
They are often given the use of the information residing on different databases (eg, customer<br />
database, it is used by the CRM system, it's products in the database management, etc.). BI<br />
applications can simultaneously access data instead of these dates in different bases, to analyze and<br />
integrate information structure of decision support;<br />
152
Stefano De Falco<br />
The information presented in the various applications are typically structured in<br />
Analytical and synthetic form, the business intelligence applications, allow instead of synthesizing<br />
and presenting such information through an effective graphic form;<br />
Requires analysis of the relationship between different variables, the analysis is not possible with<br />
software applications management "traditional".<br />
CAD to PLM<br />
The CAD and Product Lifecycle Management (PLM) are used to support the activities of product design<br />
and, generally, created to manage the entire lifecycle of the product itself, from its birth until after the<br />
sale.<br />
Alongside the traditional CAD in recent years has significantly developed the methodology / 3D display<br />
technology, enables the realistic simulation of a product from different points of view (style, design,<br />
production and even maintenance), providing considerable support to different decisions on product<br />
development.<br />
In addition, CAD systems have enabled more and more communication with other systems Computer<br />
Aided (CAX) systems such as Computer Aided Styling (CAS) used by product designers and CAM<br />
systems for the automatic generation of programs in machine control numerically.<br />
The PLM is the integration of various ICT applications generate and use the product data from<br />
engineering systems (CAD described above), the software supporting the organization of their work<br />
(PDM - Product Data Management), the document management systems. Within the PLM approach, you<br />
can define different "features" on managing the lifecycle of the product: these features ranging from<br />
collaborative design and distributed to the management of the bill of product in its different views (the<br />
view of the designer For example, different from that of those who make technical assistance), to portfolio<br />
management and product configurations.<br />
CAD systems can reduce design time and improve the quality of the project through the possibility of a<br />
better view of the project, the automatic production of drawings, the easier management of revisions of<br />
the project. PDM systems and "tending to PLM" is usually taken with the specific objective-or to "put<br />
order" in the product data and manage the communication between different actors. In some cases, the<br />
PLM system will increase communication between the technical department and other business<br />
functions, or even with external actors (co-designers and component suppliers), while in others it may<br />
support a more thorough cost analysis life cycle of a product.<br />
Web applications and online services<br />
Web applications and online services typically provide features not provided by traditional management<br />
systems. The company may use their own web applications, developed and customized based on your<br />
needs (Intranet), or use services provided by external providers, are all available remotely as an ASP<br />
(application service provider).<br />
The intranet can enhance such efficiency and the management of the business records of sales activities.<br />
Of particular interest to SMEs are:<br />
The use of shared platform in ASP mode on the standard of services requiring low levels of<br />
personalization (such as management of pay slips);<br />
Services Customer Relationship Management (CRM) for managing customer relationships, which<br />
relate to specific activities such as building a customer database, the creation of reports and<br />
statistics, newsletters, etc..<br />
The use of ASP services online it has the added advantage compared to the intranet, you use these<br />
applications at a cost linked to the real use, freeing the company from start-up costs of investment and<br />
maintenance of software.<br />
Web applications and services online B2B<br />
It intends to intercompany B2B set processes supported by customer-supplier Web applications. We can<br />
then distinguish the applications and services based on the type of process supported.<br />
153
1) Buying (e-Sourcing, e-Catalog)<br />
Stefano De Falco<br />
The e-Sourcing applications are applications and services to find new suppliers, their qualification and<br />
certification. The services and applications to enable e-Catalog:<br />
Increased efficiency of the purchasing process, as a result of both automation of some activities,<br />
such as the handling of request tobuying, workflow and approval process of the cycle of<br />
administration and accounting, both of "Reversal" in a logical self-service, certain activities directly<br />
external user;<br />
A reduction in delivery times, thanks to a streamlining of the overall management process and the<br />
request to purchase order (thanks to the delegation of purchase) and a possible reduction in the time<br />
of receipt of goods in the case of special delivery arrangements with suppliers;<br />
A reduction of maintenance costs in stock.<br />
Also in the field of e-Sourcing and e-Catalog is willing to operate the offer its technology solutions in ASP<br />
mode with value-added services related to the use of the platform. The goal of these operators and to<br />
allow SME users to benefit from advanced technology without having to develop in-house.<br />
2) sales<br />
These applications enable the company to market the products to its business customers through the use<br />
of Web catalogs.<br />
Also in this case, the company may decide to develop their own application or extranet use the services<br />
provided by external providers. Extranet applications sales support, in general, the publication of the<br />
catalog of products, with an option for business customers (eg other users of the product on the market<br />
or the distribution network operators) to search for items of interest, to place orders online, to monitor the<br />
progress of the order and receive a set of pre-and post-sales.<br />
The main benefits arising from the use of catalogs are:<br />
Possibility to reduce non-value added activities involving, for example, inclusion of net sales orders<br />
from within the management system, etc.<br />
Reducing the risk of errors in the compilation of trade documents (Eg, orders, etc.).<br />
An increase in the level of service to their clients, with the possibility of offering support services to<br />
the selection and customization of products, services, order tracking and after sales services to<br />
support your use of the product.<br />
3) information exchange and collaboration in the supply chain<br />
These applications extranet or B2B services provided by specific operators, which support the exchange<br />
of information and / or documents between the company and its supply chain partners (eg, contractors,<br />
suppliers, etc.).<br />
Extranet or use the services provided by external providers that, in recent years, played a particularly<br />
important role. These operators provide, in particular SMEs in industrial districts, services that support the<br />
exchange of documents (eg, work orders, notes, business documents) and informations through the use<br />
of a shared platform, based on technology Internet.<br />
The main benefits are:<br />
An increase in labor productivity of people involved in operational interface is the front office to back<br />
office;<br />
Improve the quality of processes, through a reduction of errors in documents;<br />
A more timely exchange of information;<br />
Reduced costs of communication with their supply chain partners.<br />
Web applications and services online B2C<br />
154
Stefano De Falco<br />
The main applications or services that the firm can use to support the relationship with consumers and,<br />
more generally, with all those (stakeholders) that have some interest in the enterprise and that can be<br />
classified into three main categories:<br />
i) Institutional sites: they serve to provide a range of information to stakeholders on the company:<br />
slag, products / services, financial data, job opportunities, etc.. These applications are simple and<br />
inexpensive that have no particular impact on the processes and organization of the company and<br />
therefore are particularly popular even among SMEs;<br />
ii) Places of service: used primarily for the provision of services on line of pre-sales and / or post-sale;<br />
iii) e-commerce sites used in a "eCommerce Service Provider".<br />
Again the company intends to market its products through the internet to end consumers have the choice<br />
of a site e-business owner, as his shop "virtual", or the services offered by "eCommerce Service<br />
Provider", bringing together in one place a set of virtual stores, offering a range of common services<br />
(marketing, payment, delivery, etc...)<br />
The advantage for an SME to use the eCommerce Service Provider is the ability to leverage the brand,<br />
visibility and the traffic generated from them, the were able to develop a broad community of users. In<br />
particular, through this new commercial channel an SME may be able to:<br />
Extend their reach the market from traditional channels;<br />
It can better serve its customer base in a logic CRM, through the online channel by providing<br />
increased and more timely information (about products, news and promotions) as well as an<br />
additional way, convenient and always available to carry out orders;<br />
The possibility of increasing the effectiveness of distance selling, and the case of operators<br />
specialized in selling via catalog, telephone, television, etc.., The present, already in their traditional<br />
business model and a value chain particularly suitable for sale on the Web.<br />
The internet offers the opportunity to promote products and brands on the market particularly effective<br />
way. The main marketing tools for Internet-based are:<br />
Purchase of keywords: provides for the promotion of a website on engines research through the<br />
purchase of sponsored links appear in the results of research generated by visitors through the use<br />
of specific keywords. Sponsored links consist of a title, a brief description of the product c / o service<br />
and a link to the Punta advertised site. Paid according to the number of visitors actually click on the<br />
link, the ads are sorted based on what the advertiser is willing to pay for each click received;<br />
Banner, net consists publish graphic panels of various sizes form (banner), which can be positioned<br />
in different parts of a web page, textual or visual content, a fixed or rotating. Banners can also be<br />
inserted in the newsletter, the are sent to a database of members, usually focused on a specific<br />
theme.<br />
Direct eMailing, this tool is based on sending emails to potential (or regular) customers using its CRM<br />
or buying from specialized society, lists of email addresses of the people have expressed their<br />
consent to receive advertisements.<br />
Mobile and Wireless Applications (M & W)<br />
Have applications, taking a growing interest in the support business processes for SMEs. M & W<br />
applications are based on:<br />
Cellular network, whatever the term used (cell phone, smartphone, palmate, finished industrial<br />
notebook with a connect card, box or trasponder3);<br />
Red Wi-Fi, no matter what it terms used (smartphone, palmate, the Hindu-dustry, laptop, box or<br />
transponder);<br />
RFID technology.<br />
These technologies can be of support to:<br />
1) Activities to be carried on cameo as: sales (purchase orders via mobile devices (cell phones,<br />
PDAs, laptops), a collection of dates for operational marketing activities (merchandising), technical<br />
support and maintenance, etc.;<br />
155
Stefano De Falco<br />
2) operational activities aimed at monitoring progress (tracking) workmanship. Through terminals<br />
equipped with an optical reader, or RFID, placed in appropriate positions along the production line<br />
inside or at the contractors / suppliers / clients are automatically read the information contained tag /<br />
transponder (put on products or pallets) as, for example, the progress of the production process, the<br />
outcome of the tests performed by the workers, the information on the subsequent stages of the<br />
manufacturing process.<br />
This overview is a preliminary useful step in the proposed methodology to characterize the technology<br />
level adopted by the firm to document innovation.<br />
2. Track and trace innovation<br />
The first step in the application of a methodology aimed at using technology as a tool for documenting<br />
innovation the enterprise knowledge, is the management capability comprehensively from the two<br />
aspects of supporting capability from enterprise knowledge organization and technology environment and<br />
three aspects of knowledge production, diffusion and application in knowledge activities system by dint of<br />
method of knowledge audit, designed evaluating index system of knowledge management capability and<br />
evaluated the knowledge management capability by practical case. (Suo; Wang; Zhang 2009).<br />
The general objectives of the technology infrastructure are to obtain information about the flow of<br />
activities related to interventions of innovation in the enterprise.<br />
This means keeping track of the various activities performed on a single business process in terms of<br />
traceability from upstream to downstream, and traceability in the opposite direction.<br />
A innovation traceability system allows to relate the human resources and structural and infrastructure<br />
resources of a company with innovation activities developed over time, in terms of searches, research<br />
partners outside the company, calls for funding of programs research used for specific research activities,<br />
results (if any prototypes, patents and research agreements, conventions, etc.) (Sixto 2006).<br />
Moreover, this innovation traceability system, as a corollary, allows to define the responsibilities of units<br />
involved in each activity, among other things to improve quality control, to combat counterfeiting, to<br />
increase collaboration among units, to control the increasing complexity of the units and so on.<br />
To implement the proposed innovation traceability system, we consider each business unit as being<br />
composed of a set of processing sites and their operators. At each processing run specific activities<br />
(functions) on the parts compose the whole product or a product. The sites of processing may be related<br />
to the same company or different companies or business units. (Zhonghua and Aihua 2008).<br />
The traceability of innovation must be directed to those units.<br />
In a manufacturing site and the next may be an intermediate storage (referred to as "Buffer"), or a<br />
transport system (in case of processing stations belonging to different sites or different organizational<br />
units).<br />
In general, units can also be regarded as a group or batch of units, if the members of that group have the<br />
same production history. The innovation supply chain traceability is achieved by recording all the<br />
information about the research activities carried out on the unit, and allowing that information to flow<br />
throughout the chain. In this way, we reach the goal of tracking is the tracking unit.<br />
Track a unit is to identify the source going back in the chain, while drawing a unit is to follow the path in<br />
the chain from producer to consumer with regard to the effects of innovative research activities. Tracking<br />
can be also related to each portion of the product concerned from the research for its innovation, and as<br />
such may be capable of tracking each company has played a role in the formation of this portion and thus<br />
was in the innovation process.<br />
The traceability of innovation is therefore also a powerful instrument of control processes. The<br />
management of material flows in critical areas provides insight into a given time what processes are<br />
being developed in order to innovate, to identify the causes greater rapidity and safety of noncompliance,<br />
to manage operations in time and materials-there is the improvement and standardization of<br />
quality, cost reduction, the rationalization of logistics flows and processes. Thus, the traceability must not<br />
only allow the registration of providers, products and locations of materials contribute to the formation of<br />
156
Stefano De Falco<br />
the product, but also the association of other important data for process control, such as time, quantity,<br />
indicators, etc..<br />
Description of the technological infrastructure<br />
In the proposed approach we suggest the use of RFID technology to fast detect the past and present<br />
innovation activities implemented in the firm, as a key factor of success (Bradford and Florin 2003).<br />
RFID (Radio Frequency Identification) is considered as the main technology for realizing the ubiquitous<br />
environment by providing the identity to the physical object. (Rosenberg 2006). The RFID tag, which is<br />
consisted of an antenna, transceiver, and integrated circuit with memory, has been widely used for supply<br />
chain and inventory management to increase the business efficiency. Also, it can be used for ID badges,<br />
access control, fleet maintenance, equipment and parts maintenance, parking log access control, toll<br />
gate system, car tracking, manufacturing line control (RFID Implementation in Retail and Manufacturing<br />
IEEE RFID 2008), e-government services (Carter and Bélanger 2005) etc. The tag is attached or<br />
integrated to an object such as a product can be read by RFID reader and then we can identify the object<br />
with the information in the tag.<br />
In this approach this technology is used to give a tool for managers to know fast the existing links<br />
between firm’s resources and research processes and research partners over time (Figure 1) (Brown,<br />
and Russell 2007), (Gentag 2006).<br />
Figure 1: RFID application for tracking research processes<br />
In terms of infrastructure, the traceability system as a model for the documentation of business innovation<br />
is made up of different terminal unit (TU) equipped with a reader / writer and RFID tags, the acquired data<br />
and transmit them to a pin or storage unit (SU). The SU record data sent by UT. Processing unit (PU) is<br />
responsible to collect and analyze data from the SM. Although the data are distributed in different SU, the<br />
PU may ask for a SU only, which will contact the other SU on behalf of the PU. The TU can be mobile<br />
(such as palmate, with screen and keypad) or "bench" (only the reader), and connected to a SU wireless<br />
technology or cable network, respectively. The TU enable operators to collect and send data on the<br />
production flow. The innovation documentation platform offers the opportunity to pose several queries to<br />
the system, including use of keywords, to search specific innovation processes under development or<br />
completed and archived.<br />
This platform was developed in research partnership with the “Solvendo Company”. Solvendo is an italian<br />
software house whose main mission is finding and developing smart solutions to real problems of SME.<br />
The main activities concern document management and dematerialisation, ontology-based knowledge<br />
157
Stefano De Falco<br />
management and e-learning, business software, web applications development. Solvendo is actually<br />
involved in technology transfer activities with Federico II University of Naples and is committed in<br />
research projects.<br />
The operation of the platform is as follows: When a request comes of an innovation, both process or<br />
product or system innovation, following a call for research funding, or an agreement with a research<br />
institution, or a customer need, then every element involved in the modification, extension, or integration<br />
process, is identified and associated with that specific request of innovation and it is also associated with<br />
keywords to the action implemented. This builds a database of innovation characterized by many fields<br />
useful for next query (Symbol_Reader, Handheld RFID Readers, (2007). Thus to new each request of<br />
innovation is first carried out a query on the database of innovation to see if there were other actions can<br />
serve as a starting point for the current action and especially if there are already skilled human resources<br />
in the company on that precise issue and physical resources more already evolved into the current target.<br />
Also another big plus of the platform is the possibility to classify, over time, all successful and not<br />
successful innovation activities. Therefore, every new innovation activities on the question of innovation<br />
database can be done by keyword, by date, research partners, by choice of process and by result<br />
(positive or negative). In figure 2 a resource registration form is reported and in figure 3 a palmate reader<br />
used in the implementation of the proposed system is reported (EPC 2007 and EPC GLOBAL 2005).<br />
Figure 2: Resource registration form in the proposed tracking innovation platform<br />
Figure 3: Reader in the proposed tracking innovation platform<br />
158
3. An application<br />
Stefano De Falco<br />
The methodology was actually applied at an Italian medium / large firm that produces and distributes<br />
dairy products in Italy and abroad, named ICCA spa. With the helpful cooperation of the processes chief,<br />
the processes of the company have all been surveyed, have been traced all human resources and<br />
equipment of the industrial plant and were made the report, in the time window 2000-2010, all research<br />
projects involving the Icca company and their involvement with these resources, as reported in table I.<br />
Table 1: Innovation grid for “Icca” company<br />
Structural Infrastructural Material Human Existing Innovation Research<br />
Processes Processes Resources Resource Technology Process ID Partner (s)<br />
Preparation -Power supply. -Oven. Unit 1 VoIP R72010 Dep.of<br />
products<br />
-Fuel. -Hydraulic<br />
PLM<br />
Informatics<br />
-Cleaning of the press.<br />
University<br />
workplace.<br />
of Naples<br />
-Calibration<br />
equipments.<br />
-Maintenance<br />
equipments.<br />
-Communications<br />
-Quality Control.<br />
Federico II<br />
Packaging Power supply. Packaging Unit 2 VoIP G32009<br />
-Fuel.<br />
-Cleaning of the<br />
workplace.<br />
-Calibration<br />
equipments.<br />
-Maintenance<br />
equipments.<br />
-Communications<br />
-Quality Control.<br />
-Communications<br />
-Quality Control.<br />
equipment.<br />
CAD<br />
Distribution Product Delivery Vehicles Unit 3 CRM<br />
ERP<br />
-<br />
Sale In outsourcing In<br />
In<br />
- -<br />
outsourcing outsourcing<br />
Promotion In outsourcing In<br />
In<br />
- -<br />
outsourcing outsourcing<br />
4. Results and discussions<br />
A result of applying the methodology of innovation relates to a request by the network of distributors of<br />
the products of “Icca” Company. The request was the possibility of implementing a new telephony<br />
software for order management. The application of the methodology has allowed a quick feasibility study<br />
of the needs required by the network of distributors who provided a positive solution to the case, as the<br />
preceding process of innovation (in the database identified innovation as a process of innovation<br />
R72010) introduced the transition to voice over IP telephony system that facilitates the ability to<br />
implement the new software to manage orders. In order to have committed human resources already<br />
formed in the previous process of innovation and also contacted the Dep. of Informatics of University of<br />
Naples Federico II, which had provided advice on the previous trial.<br />
5. Conclusions<br />
The proposed work concerns the very topical theme of innovation in operational terms, and introduces a<br />
methodology with a case of a real application through development of a platform that can really make an<br />
improvement both for the users of 'innovation, ie companies, but also for the suppliers of upstream<br />
research for innovation, universities and research centers.<br />
In particular, the proposed work addresses the problem from an innovation point of view circular, in the<br />
sense that often sees the technology as a result of the innovation process, but is itself an instrument to<br />
facilitate it. In the approach proposed has chosen to use technology to acquire data using RFID devices<br />
with the aim of documenting all activities relating to innovation developed over time by the company.<br />
159
Stefano De Falco<br />
The pursuit of this objective requires a series of preliminary steps, principally, according to Porter's<br />
approach, the classification of all processes which is characterized by the company, both structural and<br />
infrastructural. This arrangement allows the operator of innovation to have a functional association<br />
between processes and resources affected by human activity and structural innovation.<br />
The advantages of a system for documenting innovation in the company consist of the opportunity for<br />
managers of innovation in the enterprise to optimize the choice of the most appropriate resources to<br />
specific objectives related to project financing facility agreements with research centers on innovation<br />
needs of their customers, since they already know the "state of innovation" of each element and also<br />
what process it is inserted.<br />
Acknowledgements<br />
The author thanks to Eng. Fabio Di Marino of “Solvendo” society, for his valuable contribution to this<br />
research work, and thanks to Icca Company for having granted to work on their business processes and<br />
systems to test the proposed methodology.<br />
References<br />
Asian Productivity Organization, Top Management Forum on Strategic Management of Technology and Innovation,<br />
Strategic Management of Technology and Innovation ©APO 2007, ISBN: 92-833-7063-5 Report of the APO<br />
January 2010.<br />
Bradford, M., and Florin, J. Examining the Role of Innovation Diffusion Factors on the Implementation Success of<br />
Enterprise Resource Planning Systems” International Journal of Accounting Information Systems, 4, (2003),<br />
pp.205-225.<br />
Brown, I., and J. Russell, Radio Frequency Identification Technology: An Exploratory Study on Adoption in the South<br />
African Retail Sector, International Journal of Information Management, 27,(2007), pp. 250–265<br />
Carter, L. and Bélanger, F., The utilization of e-government services: citizen trust, innovation and acceptance factors,<br />
Information Systems Journal, 15(1), (2005), pp. 5-25.<br />
EPC2007, EPC Information Services (EPCIS) Version 1.0 Specification, April, (2007).<br />
EPCglobal, The EPCglobal Architecture Framework, July (2005).<br />
Gentag_NFC, Gentag to Piggyback on NFC Technology to Read RFID Sensors, September, (2006).<br />
O. J. Sixto, Is near-field communication close to success?, IEEE Computer, vol. 39, p. 3, March (2006).<br />
RFID Implementation in Retail and Manufacturing, IEEE International <strong>Conference</strong> on RFID (IEEE RFID 2008), Las<br />
Vegas, April 16-17, (2008).<br />
B. Rosenberg, RFID : applications, security, and privacy. Upper Saddle River, NJ: Addison-Wesley, (2006).<br />
Suo Bai-min; Wang Jia-bin; Zhang Han-bing, Primarily Research on Knowledge Audit for Evaluating Enterprise<br />
Knowledge Management Capability Management and Service Science, 2009. MASS '09. International<br />
<strong>Conference</strong> on Digital Object Identifier: 10.1109/ICMSS.2009.5302226 Publication Year: 2009 , Page(s): 1 - 5<br />
Symbol_Reader, Handheld RFID Readers, (2007).<br />
Zhonghua Liao and Aihua Yang, Knowledge Sharing Based on Project Management, China Human Resources<br />
Development, vol. 222, 2008, pp. 95–98.<br />
160
Legal, Privacy, Security, Access and Regulatory Issues in<br />
Cloud Computing<br />
Nomusa Dlodlo<br />
CSIR – Meraka Institute, Pretoria, South Africa<br />
ndlodlo@csir.co.za<br />
Abstract: Cloud computing is a sufficiently new research area. Since it is in its development stages, the information<br />
on the field is slowly being compiled by the researchers and practitioners from their experiences. Some of the areas<br />
in which there is still a gap on reporting on legal, privacy, security, access and regulatory issues. This paper raises<br />
an awareness of legal, privacy, security, access and regulatory issues that are associated with the advent of cloud<br />
computing. An in-depth literature survey is conducted on these and an analysis is drawn from the issues that are<br />
identified through the literature survey. Recommendations are then given on how the issues identified in the analysis<br />
can be mitigated. These recommendations centre around the issues of policy interventions, standards, privacy and<br />
data protection, traffic and congestion management, business continuity planning, security and regulation. This<br />
research is an advancement of knowledge in that field and is meant to initiate further debate on cloud computing<br />
Keywords: cloud computing, security, privacy, regulation, legal issues, interoperability, business continuity planning<br />
1. Introduction<br />
The emergence of very large specialised data centres that host thousands of servers has created a<br />
surplus of computing resources that has come to be called the cloud. The cloud is the term for networked<br />
computers that distribute processing power, applications and large systems among machines. This<br />
means that, computing is no longer on local computers but on centralised facilities operated by third party<br />
compute and storage facilities (Foster, 2010). Cloud computing transforms once-expensive resources like<br />
disk storage and processing cycles into a readily-available cheap commodity. By deploying Information<br />
Technology (IT) infrastructure and services over the network, any organisation can purchase these<br />
resources on an as-needed basis and avoid capital costs of software and hardware. By offering<br />
enterprises the opportunity to decouple their IT needs and their infrastructure, cloud computing has the<br />
likely ability to offer enterprises long-term IT savings, including reducing infrastructure costs and offering<br />
pay-for-service models.<br />
One of the definitions of cloud computing given in literature is as follows (Cloud computing, 2009):<br />
“Cloud computing refers to both the applications delivered as services over the Internet and<br />
the hardware systems and software in the datacentres that provide those services. The<br />
services themselves have long been referred to as Software as a Service (SaaS). The<br />
datacentre hardware and software is what we will call the cloud. When a cloud is made<br />
available in a pay-as-you-go manner to the public, we call it a public cloud; the service being<br />
sold is utility computing. We use the term private cloud to refer to internal datacentres of a<br />
business or other organisations that are not made available to the public. Thus cloud<br />
computing is the sum of SaaS and utility computing, but does not normally include private<br />
clouds”<br />
The National Institute of Standards and Testing (NIST) defines cloud computing under 5 identified<br />
characteristics as follows (How cloud computing, 2010):<br />
On-demand self service, which allows business units to get the computing resources they need<br />
without having to go through the IT department<br />
Broad network access, which allows applications to be built in ways that align with how businesses<br />
operate today – mobile, multi-device, etc.<br />
Resource pooling, which allows for pooling of computing resources to serve multiple consumers<br />
Rapid elasticity, which allows for quick scalability or downsizing of resources depending on demand<br />
Measured service, which means that business units only pay for the computational resources they<br />
use. IT costs match business success<br />
Kushida, et.al (Kushida, 2010 ) give the operating definition of cloud computing as:<br />
“Cloud computing provides on-demand network access to a computing environment and<br />
computing resources delivered as services. There is elasticity in the resource provision for<br />
161
Nomusa Dlodlo<br />
users, which is allocated dynamically within providers’ datacentres. Payment schemes are<br />
typically pay-as-you-go models”.<br />
With the advent of the cloud arises legal issues and those of privacy, security, access and its regulation.<br />
This paper gives a review of these issues and how they can be mitigated.<br />
2. Problem statement<br />
Since cloud computing is a field that is in its development stages, the information on the field is slowly<br />
being compiled by the researchers and practitioners from their experiences. Some of the areas in which<br />
there is still a gap on reporting are on are legal , privacy, security, access and regulatory issues. This<br />
paper raises an awareness of legal, privacy, security, access and regulatory issues that are associated<br />
with the advent of cloud computing.<br />
2.1 Process<br />
An in-depth literature survey is conducted on the legal, privacy, security, access and regulatory issues of<br />
cloud computing and an analysis is drawn from the issues that are identified through the literature survey.<br />
Literature that is relevant to this article was compiled through Internet searches, and searches of<br />
databases of online journals and conference proceedings. The search was done randomly, on the basis<br />
of the associated keywords identified. Recommendations as identified from current literature are then<br />
given on how the issues identified in the analysis can be mitigated. These recommendations centre<br />
around the policy interventions, standards, privacy and data protection, traffic and congestion<br />
management, business continuity planning, security and regulation.<br />
2.2 Question and objectives<br />
The main research question in this paper is, “ What is the current state of affairs on the soft issues of<br />
cloud computing and what does the literature say on the recommended way forward to further the issues<br />
of cloud computing.”<br />
The objectives are:<br />
Identify the legal, privacy, security, access and regulatory issues prevailing currently in the area of<br />
cloud computing<br />
Identify recommendations on the way forward on these issues<br />
3. Legal implications of cloud computing<br />
(Legal implications, 2010; Legal issues, 2010; Wisdom of clouds, 2008)<br />
There is the issue of “reasonable security” in the cloud computing context, and potential liability arising<br />
out of security breaches in the cloud. A company that provides a service to handle the personal<br />
information of another organisation has the responsibility to ensure that there is reasonable security to<br />
protect personal and confidential information.<br />
The data centres of cloud service providers are located in various locations all over the world. That<br />
means data on the cloud could be stored in any country. The ‘physical location’ raises the question of<br />
legal governance over the data. In case of a conflict between the cloud vendor and the customer the<br />
question of which country’s court system will settle the dispute comes to the fore. In cases where there is<br />
a litigation, an organisation will have to deal with a third party cloud provider to gain access to information<br />
relevant to the litigation. Considering the multiple copies of data that may be created, stored, recompiled,<br />
reused, dispersed and reassembled, what constitutes a “record” for evidence may be difficult to grapple<br />
with the cloud.<br />
The number of trademark filings covering cloud computing brands, goods and services is increasing as<br />
companies seek to better position themselves for cloud computing branding and marketing efforts.<br />
Therefore ensuring the uniqueness of a trademark with the advent of a cloud has been further<br />
complicated.<br />
Sharing and transferring data within the cloud is a problem. Organisations are legally prohibited from<br />
transferring personal information to countries that do not provide the same level of protection with respect<br />
to personal information. That means cloud providers will not be in a position to make any contractual<br />
162
Nomusa Dlodlo<br />
promises to their clients because in many cases they cannot say which countries data will be transferred<br />
to and from.<br />
Systems are vulnerable to damage or interruption from earthquakes, terrorist attacks, flood, fires, etc.<br />
Customers have to ensure therefore that they are insured against loss of business due to such potential<br />
losses. This is essential. If there is a breach of privacy due to the fault of the cloud vendor, the carrier<br />
should be liable for the compensation. Vendors on the other hand should do their level best to meet<br />
service level targets committed with any customer.<br />
Ideally, the data that is of the customer’s creation is protected under the intellectual property rights of a<br />
country. Therefore there should be compensation for infringement. Customers own the data. No vendor<br />
can claim ownership of any data that is uploaded or associated with intellectual property. Customer data<br />
includes all data maintained by the customer<br />
In one frequently cited scenario (Gurav, 2010), a government agency presents a subpoena or search<br />
warrant to the third party that has possession of a customer’s data. Had they retained physical custody,<br />
the customer might still have been compelled to surrender the information, but at least would have been<br />
able to decide for themselves whether or not to contest the order. The third party service is presumably<br />
less likely to go to court on behalf of a customer. In some circumstances the customer might not even be<br />
informed that their documents have been released.<br />
With the cloud there is lesser privacy protection under the law. To search a house or office (including<br />
documents stored on a computer), police need a warrant of arrest. To get the information that is stored<br />
on a third party’s web server they only need a subpoena, which is easier to obtain. This kind of search<br />
can happen without the cloud customer’s knowledge.<br />
4. Security issues in cloud computing (hidden risks, 2010)<br />
Due to its distributed nature, the cloud results in weak security systems that are easy to break into. The<br />
security of the system is only as strong as the weakest user’s set-up. Weak password recovery<br />
workflows, phishing attacks, and keyloggers present bigger security risks. In collaborative web<br />
applications that are built for groups, like Google Apps or any web-based project management software,<br />
any breach of security spreads across all participants.<br />
In cloud computing an organisation’s data is locked-in and the third party in control. When you participate<br />
in the cloud, you depend on a third party to make decisions about your data and platforms. Cloud<br />
computing also comes with chances of server unavailability and account lock-out. When the Internet goes<br />
down, access to one’s data is cut off. An important measure of security often overlooked by companies is<br />
how much downtime a cloud service provider experiences. The client should request the provider’s<br />
reliability reports to determine whether these meet the requirements of their business. Exception<br />
monitoring systems is another important area which companies should ask their service providers about<br />
(Binning, 2010).<br />
The biggest concern with cloud computing is that it puts all of a company’s data and applications in one<br />
place. Businesses should be wary of putting sensitive company information in public clouds. They should<br />
instead stick to low-risk, low volume applications and build internal and private clouds to enable<br />
collaboration within the organisation and externally with partners.<br />
Security is one of the most often-cited objections to cloud computing (Zhang, 2010). Cloud users face<br />
security threats from both outside and inside the cloud. This responsibility is divided among the cloud<br />
user, the cloud vendor and any third party vendors that users rely on for security-sensitive software or<br />
configurations. The cloud user is responsible for application-level security. The cloud provider is<br />
responsible for physical security, and likely for enforcing external firewall policies. Security for<br />
intermediate layers of the software stack is shared between the user and the operator. Cloud providers<br />
must guard against theft or denial-of-service attacks by users. One last security concern is to protect the<br />
cloud user against the provider. The provider will by definition control the ‘bottom layer’ of the software<br />
stack, which effectively circumvents most known security techniques. Users also need to be protected<br />
from one another. The primary security mechanism in today’s clouds is virtualisation. It is a powerful<br />
defence, and protects against most attempts by users to attack one another or the underlying cloud<br />
infrastructure.<br />
163
Nomusa Dlodlo<br />
Virtualisation is the enabler technology for the cloud and uses physical resources such as a server which<br />
is divided into virtual resources called virtual machines (VM). Customers cannot protect their VMs on their<br />
own. Cloud service providers are making substantial effort to secure their systems, in order to minimise<br />
the threat of insider attacks and reinforce the confidence of customers. For example, they restrict access<br />
to hardware facilities, adopt stringent accountability and auditing procedures, and minimise the number of<br />
staff who have access to critical components of the infrastructure. Security management of virtualisation<br />
technologies is required to reduce the risk of security exposures and enable security policy enforcement.<br />
A cloud-specific security issue is that of running arbitrary VM images. This is only one aspect of making<br />
sure the right data is available to the right user at the right time. The Cloud Security Alliance is directly<br />
pursuing in addition, issues of authentication, authorisation, privacy, integrity and non-repudiation and<br />
data reliability and availability.<br />
According to Chow et.al. (Chow, 2009), security concerns are categorised as:<br />
Traditional security<br />
Availability<br />
Third party data control<br />
Traditional security concerns involve computer and network intrusions or attacks that will be made easier<br />
or possible by moving to the cloud (Chow, 2009). VM-level attacks are a problem because of potential<br />
vulnerabilities in the VM technology. There are also cloud provider vulnerabilities. These could be<br />
platform level such as SQL-injection or cross-site scripting. Phishers and other social engineers have a<br />
new attack vector. The cloud user must protect the infrastructure used to connect and interact with the<br />
cloud, a task complicated by the cloud being outside the firewall in many cases. The enterprise<br />
authentication and authorisation framework does not naturally extend into the cloud. Traditional digital<br />
forensic methodologies permit investigators to seize equipment and perform detailed analysis on the<br />
media and data recovered. The likelihood therefore of the data being removed, overwritten, deleted or<br />
destroyed by the perpetrator in this case is low.<br />
Availability concerns centre on critical applications and data being available (Chow, 2009). As with<br />
traditional security concerns, cloud providers argue that their server uptime compares well with the<br />
availability of the cloud user’s own data centres. There are more single points of failure and attacks in the<br />
cloud. They may lack an assurance of computational integrity.<br />
The legal implications of data and applications being held by a third party are complex (Chow, 2009).<br />
Therefore there are many questions that remain unanswered. There is also a potential lack of control and<br />
transparency when a third party holds the data. If served a subpoena can a cloud user compel the cloud<br />
provider to respond in the required time frame? How can a cloud user be guaranteed that data has been<br />
deleted by the cloud provider? Audit difficulty is another side effect of the lack of control in the cloud. Is<br />
there sufficient transparency in the operations of the cloud provider for auditing purposes? There are<br />
contractual obligation issues also. One problem with using another company’s infrastructure besides the<br />
uncertain alignment of interests is that there may be surprising legal implications. Cloud provider<br />
espionage is the worry of the theft of company proprietary information by the cloud provider. How can a<br />
cloud user avoid data lock-in? The data itself might locked in proprietary format and there are also issues<br />
with training and processes. There is also a problem of the cloud user having no control over frequent<br />
changes in cloud-based services. Another possible concern is that the contracted cloud provider might<br />
use subcontractors, over whom the cloud user has even less control and who must also be trusted.<br />
5. Data access and interoperability<br />
The issue of data access and interoperability continues to be an outstanding matter for inherently<br />
distributed applications and federated organisations. Common best practices and standards are needed<br />
to achieve the fundamental properties of portability and interoperability for cloud applications and<br />
environments.<br />
A major challenge of moving applications to the cloud for most organisations is the need to master<br />
multiple languages and operating environments (Gurav, 2010). In many cloud applications a back-end<br />
process relies on a relational database, so part of the code is written in SQL, or other query language. On<br />
the client side, program logic is likely to be implemented in JavaScript embedded within HTML<br />
documents. Standing between the database and the client is a server application that might be written in<br />
a scripting language (such as PHP, Java and Python). Information exchanged between the various layers<br />
164
Nomusa Dlodlo<br />
is likely to be encoded in some variation of XML. Any web application needs to be available to legitimate<br />
visitors from all over the world. A true cloud spans the entire globe, with a server presence in multiple<br />
simultaneous locations.<br />
Besides technical issues, a cloud provider could suffer outages for non-technical reasons, including going<br />
out of business or being the target of regulatory action. Therefore organisations should be wary of this<br />
and put in place measures to ensure business continuity and service availability when outages occur<br />
(Zhang, 2010).<br />
Data lock-in by the service provider is a contentious issue for the customer. Software stacks have<br />
improved interoperability among platforms, but the storage Application programming Interfaces (APIs) for<br />
cloud computing are still proprietary. Thus, customers cannot extract their data and programs from one<br />
site to run on another. This has prevented some organisations from adopting cloud computing. Customer<br />
lock-in may be attractive to cloud computing providers, but their users are vulnerable to price increases,<br />
to reliability problems, or even to providers going out of business (Zhang, 2010).<br />
Applications continue to be more data intensive. Data bottlenecks are likely to occur as more users<br />
subscribe to a cloud service. Cloud users and cloud providers have to think about the implications of<br />
placement and traffic at every level of the system if they want to minimise costs (Zhang, 2010).<br />
6. Privacy issues in cloud computing<br />
Privacy is a fundamental right enshrined in the UN Universal Declaration of Human Rights. There are<br />
various forms of privacy, including ‘ the right to be left alone” and “control of information about ourselves”<br />
(Pearson, 2009). There are different types of information that need to be protected. These include any<br />
information that can be used to identify or locate an individual (e.g. name, address, credit card number<br />
and IP address). Sensitive information such as personal financial information and job performance<br />
information is considered private. Behavioural information such as viewing habits for digital content,<br />
user’s recently visited websites or product usage history need to be protected as well.<br />
Violation of privacy occurs as a result of a number of cloud dynamics. In the cloud the infrastructure is<br />
shared between organisations and is off-premise. Therefore there are threats associated with data being<br />
stored remotely and because of virtualisation. Virtualisation is a method of running multiple independent<br />
virtual systems on a less physical resource making one computer act as many, and sharing the resources<br />
of hosts across multiple environments. The cloud is also a dynamic environment. Services can be<br />
aggregated and changed dynamically by customers and service providers can change the provisioning of<br />
services anytime. Sensitive data may move around within an organisation and across organisational<br />
boundaries. Legal compliance and adequate protection has to be maintained therefore. The speed and<br />
flexibility of adjustment to vendor offerings that benefits business and provide a strong motivation for the<br />
use of cloud computing might come at the cost of compromise to the safety of data. Cloud computing<br />
enables new services to be made available in the cloud by combining other services, e.g. a ‘print on<br />
demand’ service can be provided by combining a printing service with a storage service. This procedure<br />
of service combination is typically under less control than previous service combinations carried out<br />
within traditional multi-party enterprise scenarios. There may be varied degrees of security and privacy in<br />
each of the components.<br />
Privacy risks for cloud computing may also lie in the following (Pearson, 2009):<br />
For the cloud service user: being forced or persuaded to be tracked or give personal information<br />
against their will<br />
For the organisation using the cloud: non-compliance to enterprise policies and legislation, loss of<br />
reputation and credibility<br />
For implementers of cloud platforms: exposure of sensitive information stored on the platforms<br />
(potentially for fraudulent purposes), legal liability, loss of reputation and credibility, lack of user trust<br />
and take up.<br />
For providers of applications on top of cloud platforms: legal non-compliance, loss of reputation,<br />
‘function creep’ using the personal information stored on the cloud, i.e. it might later be used for<br />
purposes other than the original cloud service intention<br />
For the data subject: exposure of personal information<br />
165
7. Regulatory issues<br />
Nomusa Dlodlo<br />
In the cloud there are a number of issues that need to be regulated. Potential physical location of data<br />
centres could be anywhere, with geography-blind distribution of applications and data. As a practical<br />
commercial matter, national regulations should be able to influence the actual deployment of cloud<br />
services in countries around the globe.<br />
Without concrete guarantees on the privacy of data held by cloud providers, the diffusion of cloud<br />
services may be hampered by the perceived risk in entrusting sensitive data to external cloud services. In<br />
the US and Europe the regulations require some cloud offerings to allow users to stipulate the country in<br />
which their data will be stored. Non-US firms whose servers are located in the US can have their<br />
information accessed by the US government under the US Patriot Act and Homeland Security Act<br />
(Carrigan, 2008). This impacts on information privacy policy.<br />
Strongly related to the notion of service level agreements and policy, is that of governance – how to<br />
manage sets of virtual resources. At the infrastructure level, applications may consist of many virtual<br />
machines, virtual storage and virtual networks. Managing these virtual missions, or virtual data centres,<br />
requires policy and enforcement from both the provider and consumer (Lee, 2010).<br />
In a private cloud, the infrastructure for implementing the cloud is controlled completely by the enterprise.<br />
Typically, private clouds are implemented in the enterprise’s data centre and managed by internal<br />
resources. A private cloud maintains all corporate data in resources under the control of the legal and<br />
contractual umbrella of the organisation. This eliminates the regulatory, legal and security concerns<br />
associated with information being processed on third party computing resources.<br />
In a public cloud however, external organisations provide the infrastructure and management required to<br />
implement the cloud. Public clouds have the disadvantage of hosting data in an offsite organisation<br />
outside the legal and regulatory umbrella of the organisation. In addition, as most public clouds leverage<br />
a worldwide network of data centres, it is difficult to document the physical location of data at any<br />
particular moment. These issues result in potential regulatory compliance issues which preclude the use<br />
of public clouds for certain organisations or business applications.<br />
According to Enki, et.al (Enki, 2010), the identified regulatory issues in the cloud are in the areas of<br />
service level agreements (SLA), service and support and performance. Cloud-computing services define<br />
an SLA as some guarantee of how much time the server, platform or application will be available. For<br />
example, a cloud provider will provide 99.99% uptime, or five minutes downtime a year, with a 10%<br />
discount on charges for any month in which that availability is not achieved. Since its infrastructure is not<br />
built to reach this uptime, it is effectively offering a 10% discount on services in exchange for the benefit<br />
of claiming that reliability. Another trick is to compute the SLA on an annualised basis. This means that<br />
customers are eligible for a service only after one year has passed. The end-user should pay close<br />
attention to the details of the SLA being provided and weigh that against what business impact it will have<br />
if the service provider misses the committed SLA and regulatory authorities should chip in to level the<br />
playing field.<br />
One of the greatest attractions of cloud computing is that it enables computing to be available to a large<br />
community. In addition, the elimination of the responsibility for physical hardware removes the need for<br />
data-centre administration staff. As a result, there is an increasing number of people responsible for<br />
production computing who do not have systems administration backgrounds, which creates demand for<br />
comprehensive cloud vendor support offerings and thereby greatly inconveniencing the consumer. Round<br />
the clock live support staff costs a great deal.<br />
The cloud is a pervasive federated network in which unregulated personal area networks and local area<br />
networks will interoperate with traditionally regulated electronic communication services. Regulators need<br />
to carefully monitor the challenges posed by these networks, taking action as necessary to regulate for<br />
technical interoperability, consumer protection, support for competition and the appearance of<br />
opportunities for the exploitation of market power.<br />
166
8. Discussion and way forward<br />
Nomusa Dlodlo<br />
This sections draws recommendations from literature on the way forward in cloud computing. These<br />
recommendations centre around the issues of policy interventions, standards, privacy and data<br />
protection, traffic and congestion management, business continuity planning, security and regulation.<br />
Any large scale deployment needs to adhere to certain standards. The cloud spans multiple industries<br />
and differs widely in application scenarios and user requirements. The standardisation of the cloud<br />
should cover common communication protocols, at, for example, the carrier level; terminal description<br />
and service discovery mechanisms and application data switching mechanisms such as technologies<br />
based on XML, SOAP and web services. The latter covers terminal, communication protocols and<br />
application specifications.<br />
Concerns over privacy and data protection are widespread. Protecting privacy must not be limited to<br />
technical solutions but encompass regulatory, market-based and socio-ethical considerations. There<br />
should be a concerted effort involving government, civil society and private sector players to protect these<br />
values. One of the hazards of shared infrastructure is that one customer’s usage pattern may affect other<br />
customers’ performance. The cloud must incorporate traffic and congestion management. This will sense<br />
and manage information flows, detect overflow conditions and implement resource reservation for timecritical<br />
and life-critical data flows.<br />
Policy interventions towards the adoption of cloud computing should include (Etro, 2010):<br />
International agreements in favour of unrestricted flow of data across borders (since data centres are<br />
located in different countries with different privacy laws, data portability remains a key issue for the<br />
diffusion of cloud computing)<br />
Agreements between government and industry leaders on a minimum set of technological standards<br />
and process standards to be respected in the provision of cloud computing services to guarantee<br />
data security and privacy and promote a healthy diffusion of new technology<br />
Expansion of broadband capacity<br />
Introduction of fiscal incentives for the adoption of cloud computing and a specific promotion in<br />
particular dynamic sectors ( for instance, governments could finance, up to a limit, the variable costs<br />
of computing for all domestic and foreign firms that decide to adopt a cloud computing solution)<br />
Some examples of cloud computing risks for the enterprise that need to be managed include (Cloud<br />
computing: business benefits, 2009):<br />
Reputation, history and sustainability of the provider to ensure reliability of service provision<br />
The cloud provider should take responsibility for information handling and be held liable for loss of<br />
confidentiality and privacy<br />
Business continuity and disaster-recovery plans must be well documented and tested<br />
Compliance to regulations and laws in different geographic regions can be a challenge for<br />
enterprises. It is critical to obtain proper legal advice to ensure that the contract specifies the areas<br />
where the cloud provider is responsible and liable for ramifications arising from potential issues.<br />
Business continuity planning is identifying core operational systems and work processes that an<br />
organisation requires in order to deliver services and products to their customers. This involves<br />
identifying key suppliers, business partners and staff. From an IT perspective this means three things<br />
(Cloud computing, 2009):<br />
Architecting your IT infrastructure and application systems to be distributed with no single point of<br />
failure<br />
Ensuring the systems have built-in data and application redundancy<br />
Having the ability for these systems to be accessed securely from any location at any time.<br />
The problem that has made business continuity and disaster recovery extremely expensive has always<br />
been the need for redundant hardware, both on-site and in remote sites. The advent of cloud computing<br />
has made the provision of dynamically scalable and virtualised resources widely and cheaply available.<br />
Security is one of the largest concerns for the adoption of cloud computing. Seven risks a cloud user<br />
should raise with vendors before committing are (Foster, 2010):<br />
167
Nomusa Dlodlo<br />
1. Privileged user access: sensitive data processed outside the enterprise needs the assurance that<br />
they only accessible and propagated to privileged users<br />
2. Regulatory compliance: A customer needs to verify if a cloud provider has external audits and<br />
security certifications and if their infrastructure complies with some regulatory security requirements<br />
3. Data location: since a customer will not know where her data will be stored, it is important that the<br />
cloud provider commit to storing and processing and processing data in specific jurisdictions and to<br />
obey local privacy requirements on behalf of the customer.<br />
4. Data segregation: one needs to ensure that one customer’s data is fully segregated from another<br />
customer’s data<br />
5. Recovery: it is important that the cloud provider has an efficient replication and recovery<br />
mechanism to restore data if a disaster occurs<br />
6. Investigative support: Cloud services are especially difficult to investigate, if this is important for a<br />
customer, then such support needs to be ensured with a contractual commitment<br />
7. Long-term viability: your data should be viable even if the cloud provider is acquired by another<br />
company.<br />
To avoid customer lock-in, customers would want to see open/standard APIs. Cloud users must not<br />
entrust mission-critical applications to cloud service providers so as to avoid outages when cloud service<br />
providers go out of business. They should also keep backups of the applications and data on premises.<br />
They should also secure favourable service-level agreements from the cloud service provider (Kim, 2009)<br />
9. Conclusion<br />
Cloud computing has transformed the knowledge society by offering enterprises the opportunities to<br />
decouple their IT needs and their infrastructure. It has therefore given rise to new business models and<br />
given the opportunity for those enterprises that would not have had the resources to compete in the<br />
knowledge society a renewed opportunity. As a result the area of cloud computing deserves to be given<br />
attention and further developed. This research does exactly that by contributing to the advancement of<br />
knowledge in that field.<br />
References<br />
Binning, D., Top five cloud computing security issues, [online],<br />
http://www.computerweekly.com/Articles/2010/01/12/235782/Top-five-cloud-computing-security-issues.htm<br />
Carrigan, M., Alex, T., Ward, C., The US Patriot Act deconstruction, Civil Liberties and Patriotism, Journal of<br />
Business and Economic Research, Vol. 6, No. 3., pp. 19-30, 2008<br />
Chow, R., Golle, P., Jakobsson, M., Masuoka, R., Molina, J., Controlling data in the cloud: outsourcing computation<br />
without outsourcing control, CCSW’09, Novembr 13, 2009, Chicago, Illinois, USA, pp. 85-90.<br />
Cloud computing as a business continuity plan, [online], http://www.tectonic.co.za/2009/05/cloud-computing-as-abusiness-continuity-plan/<br />
Enki, D., Why cloud computing will never be free, ACMQUEUE, Distributed Computing, pp. 1-10, 2010.<br />
Etro, F., The economics of cloud computing, [online], http://www.voxeu.org/index.php?q=node/4671<br />
Foster, I., Zhao, Y., Raicu, I., Lu, S., Cloud computing and grid computing 360-degree compared<br />
Gurav, U., Shaikh, R., Virtualisation – a key feature of cloud computing, International <strong>Conference</strong> and Workshop on<br />
Emerging Trends in Technology (ICWET 2010), pp. 227-229, TCET, Mumbai, India, 2010<br />
Hidden risks of cloud computing, [online], http://lifehacker.com/5325169/the-hidden-risks-of-cloud-computing<br />
How cloud computing can transform business, [online],<br />
http://blogs.hbr.org/cs/2010/06/business_agility_how_cloud_com.html<br />
Kim, W., Kim, S. D., Lee, e., Lee, S., Adoption issues for cloud computing, proceedings of MoMM2009, December<br />
14-16, 2009, Kuala Lumpur, Malaysia, pp. 2-5.<br />
Kushida K.E., Breznitz,D., Zysman, J., Cutting through the fog: understanding the competitive dynamics in cloud<br />
computing, The Berkeley Roundtable on the International Economy (BRIE) Working Paper 190 (Beta), May 1,<br />
2010<br />
Lee, C.A., A perspective on scientific cloud computing, HPDC 2010, June 20-25, 2010, Chicago, USA.<br />
Legal implications of cloud computing – Part One (the Basics and Framing the Issues), [online],<br />
http://www.llrx.com/features/cloudcomputing.htm<br />
Legal issues associated with cloud computing, [online], http://www.labnol.org/internet/cloud-computing-legalissues/14120/<br />
Pearson, S., Taking account of privacy when designing cloud computing services, CLOUD’09, ICSE’09 Workshop,<br />
pp. 44-52, Vancouver, Canada, May 23, 2009.<br />
The wisdom of clouds, [online], http://blog.jamesurquhart.com/2008/08/cloud-computing-bill-of-rights.html<br />
Zhang, Q., Cheng, L., Boutaba, R., Cloud computing: state-of-the-art and research challenges, Journal of Internet<br />
Serv Appl, Vol 1, pp. 7-18, 2010<br />
168
Interoperability Monitoring for eGovernment Service Delivery<br />
Based on Enterprise Architecture<br />
Badr Elmir 1 , Nabil Alrajeh 2 and Bouchaib Bounabat 1<br />
1 Université Mohammed V – Souissi, Morocco<br />
2 King Saud University, Saudi Arabia<br />
b.elmir@daag.finances.gov.ma<br />
nabil@ksu.edu.sa<br />
bounabat@ensias.ma<br />
Abstract: Public administration has to prepare itself to deliver fully integrated eGovernment services. This delivery<br />
often requires cooperation via business processes interoperability across two or more departments. In this context,<br />
public departments and agencies need to implement interoperability using enterprise architecture techniques to<br />
structure business processes, and service oriented models to achieve their integration. Thus, it’s quite interesting to<br />
adopt enterprise architecture paradigm and techniques to analyse, track and control the evolution degree of<br />
processes interoperability from the existing “as-is” state to the future “to-be” state. The present paper proposes a<br />
periodic monitoring approach based on an assessment method which considers three main aspects of interoperation:<br />
1. Potentiality, reflecting the preparation to interoperate. The objective is to foster interoperation readiness by<br />
eliminating barriers that may obstruct the interaction. 2. Compatibility, referring to interoperation implementation<br />
through adequate engineering process. It aims to study the relation between the external interfaces of processes and<br />
the surrounding environment in order to ensure effective interaction. 3. Performance efficiency, focusing on<br />
monitoring operational performance. It consists of the availability assessment of the communication infrastructure<br />
and the supporting system in general. It considers also end users satisfaction of interoperation in use. The proposed<br />
method supporting tool, (IMT) for interoperability monitoring tool, assesses interoperability degree periodically<br />
through five steps: (i) Delineating the scope of interoperation; (ii) Quantifying the interoperation potentiality; (iii)<br />
Calculating the compatibility degree; (iv) Evaluating the operating performance; (v) Aggregating the degree of<br />
interoperability. In addition to its capacity to track the evolution of interoperation degree in time, the IMT measures<br />
the required effort to reach a planned degree of interoperability. Finally, to better illustrate how to use the proposed<br />
interoperability monitoring approach, we present a practical example of integrated public eService. It’s a citizen<br />
oriented eService proposed by a public hospital that offers special fees for persons covered by social security<br />
insurances. It includes government to business collaboration and government to government one.<br />
Keywords: integrated public eService, enterprise architecture, interoperability assessment, periodic monitoring, and<br />
eHealth<br />
1. Introduction<br />
Public administration has to prepare itself to provide fully integrated online services for citizens and<br />
businesses. In this context, horizontal cooperation in the public domain is a key enabler for eGovernment.<br />
Indeed, the delivery of most useful online governmental services often requires cooperation between two<br />
or more public administrations or agencies. This cooperation starts from simple information exchange<br />
and can reach business processes interoperability among public departments (Klischewski 2004).<br />
The present work focuses on monitoring interoperability between automated business-processes<br />
involved in the provision of an integrated public eService. The studied processes may be located within a<br />
single organization or across a group of public partners. Therefore, the proposed approach is based on a<br />
five step measurement method (Elmir 2010b), and takes into account three main aspects:<br />
Interoperability maturity level of the environment surrounding the studied eService.<br />
Compatibility degree between the external interfaces of the involved business processes.<br />
Operational performance of the support systems used to provide the online service.<br />
The objective of this work is to: (1) Identify the most important characteristics of interoperability used to<br />
deliver integrated public services. This work proposes a set of criteria used to assess interoperability in<br />
this context considering all aspects of collaboration. (2) Describe a monitoring approach of<br />
interoperability. This allows to know what is needed to reach a desired level of interoperability.<br />
In this article, the second section is devoted to eGovernment system interoperability. The third section<br />
presents the assessment method based on a set of IT indicators for interoperability measurement. The<br />
fourth section proposes the monitoring approach model adopted in this study. This section presents also<br />
the platform developed to support interoperability monitoring in the context of integrated public eServices<br />
169
Badr Elmir et al.<br />
delivery. Finally, a sample case study is presented. It is about a citizen oriented eHealth service that<br />
encloses government-to-government and government-to-business collaboration.<br />
2. Interoperability in eGovernment system<br />
2.1 eGovernment service delivery<br />
Government-to-government collaboration is a key factor in the success of almost all eGovernment<br />
strategies. In this context, Public online eServices target multiple groups of potential customers that could<br />
be organized into four main categories (Vintar 2002):<br />
Government-to-government services (Chen 2009): generally known by e-administration, it refers to<br />
public interdepartmental coordination and collaboration in order to deliver electronic services.<br />
Integrating the intergovernmental systems is the first step in implementing electronic government<br />
solutions.<br />
Government-to-citizens services (Peristeras 2009): this field implements the relationship between the<br />
government and the citizen, and provides communication information and essential services that<br />
interest individuals or residents in order to ease their life. It represents the main aim of eServices<br />
projects.<br />
Government-to-business services (Shambour 2010): it refers to online interaction between<br />
government and the business sector. This category includes optimizing administrative procedures,<br />
public tenders, various permits, authorizations, electronic transaction services including<br />
<br />
procurements and bids; and payment services of various taxes and public charges.<br />
Government-to-employees services (Bercea 2010): it includes a set of employees and managers self<br />
services giving the ability to view and update personal, administrative, payroll and benefits<br />
information via human resources portals. These include also communication, knowledge<br />
management, and collaboration tools made available to employees to well perform their daily<br />
operations.<br />
In term of service nature, public online services have different types (Chen 2006):<br />
Informational use: in which public departments and agencies publish information to educate,<br />
entertain, influence, or reach their potential customers;<br />
Transactional use: where they support a coordinated sequence of user and system activities to<br />
provide a specific service;<br />
Operational use: when an agency provides a new mechanism for conducting business operations by<br />
integrating information systems, human intellect, and other resources into synergistic networks.<br />
In this context public administration can be viewed as a business collaboration network (BCN) (Elmir<br />
2010b). In fact, BCN enables organizations to communicate and collaborate with their customers,<br />
partners and suppliers in a productive way (Sterling 2010). This cooperation takes different forms that<br />
start from simple information exchange, and can reach business processes interoperability among<br />
independent administrations (Sun 2007) (Shishkov2009).<br />
In this sense, this paper studies more precisely the back office integration of public administration<br />
(government-to-government) in order to provide operational integrated eServices essentially those<br />
oriented to citizens and businesses. In order to preserve the autonomy of the actors, this integration<br />
takes usually the form of business process interoperability which represents an obvious prerequisite of<br />
integrated public eService delivery.<br />
2.2 Interoperability<br />
Interoperability characterises the ability, for any number of processing information systems, to interact<br />
and exchange information and services between them.<br />
To implement interoperability, public administration faces technical and semantic difficulties (Gupta 2007)<br />
but also organizational challenges (Goldkuhl 2008). Moreover, monitoring this quality is not easy on such<br />
a macroscopic level.<br />
In fact, interoperability is an information system quality that can be viewed from various perspectives.<br />
Several taxonomies have been proposed in this direction. In this sense, there are:<br />
170
Badr Elmir et al.<br />
Many levels of interoperability scope: business, process, service and data level (Guijaro 2007).<br />
Various approaches to implement this quality: integrated, federated, and unified approach (Missikoff<br />
2004).<br />
Multiple barriers could handicap interoperation: conceptual, organizational and technical barriers<br />
(Arms 2002).<br />
Different scopes of application: within the same organization, cross independent organizations<br />
(Guédria 2008),<br />
Different transactional aspects of cooperation: synchronous or asynchronous collaboration<br />
(Michelson 2006).<br />
Diverse measurement perspectives: potentiality, compatibility, performance efficiency (Chen 2008).<br />
Thus, in terms of scope, there are various levels where interoperability takes place within public<br />
administration (Guijaro 2007):<br />
Business level that refers to how to work within a business network in harmonized way in order to<br />
collaborate.<br />
Process level aims making various processes working together. In the case of a networked<br />
administration, internal processes of two departments are connected to create a common macro<br />
process.<br />
Service level is concerned with identifying, composing, and making function together with various<br />
applications.<br />
Data level refers to making synergy between different data models and heterogeneous conceptual<br />
schemas.<br />
This organization coincides with the four levels of enterprise architecture metamodel.<br />
Also, in terms of barriers, the interoperability implementation faces (Arms 2002):<br />
Conceptual barriers which are related to the syntactic and semantic problems of information to be<br />
exchanged.<br />
Organizational barriers which refer to the definition of responsibilities and authority so that<br />
interoperability can take place.<br />
Technical barriers which deal with the use of adequate protocols, languages and infrastructure in<br />
communication.<br />
In this case, interoperation compatibility check has to consider these barriers on each one of the four<br />
enterprise architecture layers cited before.<br />
2.3 Enterprise Architecture in government structure<br />
Public services users expect to perceive public administrations as a homogeneous and coherent unit in<br />
order to have a unified access to services they need. So, Public administration must be prepared to<br />
interact effectively with all the surrounding actors. This requires essentially openness and willingness to<br />
break functional, organizational and technological barriers.<br />
In this context, several governments have established Enterprise Architecture programmes (Liimatainen<br />
2007). These programmes are well known by Government Enterprise Architecture (GEA). They aim to<br />
eliminate overlapping projects, to support reuse, and to enhance interoperability between public<br />
departments. On another hand, some national strategies were limited to the single issue of interoperation<br />
and developed interoperability frameworks. They are mainly addressing technical problems by<br />
referencing the main specifications recommended to facilitate and promote cooperation between different<br />
government agencies (Guijaro 2007).<br />
At the same time, the concept of enterprise architecture (EA) attracted a lot of interest during the past<br />
decade. It aims to provide a structure for business processes and systems that supports them. It<br />
represents an information system using models in order to illustrate the interrelationship between its<br />
components and its relationship with the ecosystem.<br />
171
Badr Elmir et al.<br />
The Enterprise Architecture (EA) proposes to take an inventory of information system components by<br />
considering: (1) organization procedures, etc. (2) business process (3) IT applications, (4) technical<br />
infrastructure.<br />
Also, with the emergence of service delivery environments, independent businesses become able to<br />
collaborate in order to have beneficial results for all (Aisopos 2009). Among the main forms of<br />
cooperation, occurs integrated service providing to clients.<br />
The service-oriented interaction model implements less coupled connections between various distributed<br />
software components. The approach seeks to provide abstraction by encapsulating functionality and<br />
allowing reuse of existing services.<br />
To facilitate interoperation within a public collaboration network, usually public administration tends to<br />
adopt the enterprise architecture as a strategic choice of organization using "the service oriented"<br />
paradigm and techniques to implement and deploy services.<br />
3. Interoperability assessment method<br />
3.1 Interoperability measurement aspects<br />
Concerning measurement, Chen (2008) differentiates between the following complementary<br />
characteristics (see Figure 1):<br />
Interoperation potentiality: it is an «internal quality» of a system that reflects its preparation to<br />
interoperate. This involves identifying a set of characteristics that have an impact on communication<br />
with peer’s systems without necessarily having concrete information on them. The objective is to<br />
foster interoperability readiness by eliminating barriers that may obstruct the interaction.<br />
Interoperation compatibility: it represents an «external quality». In fact, the interaction between two<br />
support systems is ensured through an adequate engineering process.<br />
Interoperation performance: the third aspect characterizes the «quality in use». It focuses on<br />
monitoring operational performance. It consists of an assessment of the communication infrastructure<br />
availability, and the supporting system in general.<br />
Figure 1: Operational aspects of interoperability (Elmir 2010b)<br />
3.2 Interoperability indicators<br />
Many interoperability maturity models (IMM) were introduced to describe the interoperation potentiality.<br />
They are mostly inspired by the CMM/CMMI model (Chrissis 2003). Pardo (2009) Lists among others:<br />
ITIM (It Investment Management),<br />
LISI (Level of Information System Interoperability),<br />
OIMM (Organizational Interoperability Maturity Model),<br />
EIMM (Enterprise Interoperability Maturity Model),<br />
172
Badr Elmir et al.<br />
GIMM (Government Interoperability Maturity Matrix),<br />
SPICE (Software Process Improvement and Capability dEtermination).<br />
Each model adopts a specific vocabulary to express the levels of maturity. However, the models have in<br />
general five scales ranging from low to high:<br />
An organization with a low level of interoperability is characterized as working independently or in<br />
isolation from other organizations and in an ad hoc or inconsistent manner.<br />
An organization with a high level of interoperability is characterized as being able to work with other<br />
organizations in a unified or enterprise way to maximize the benefits of collaboration across<br />
organizations.<br />
In terms of compatibility and in order to dematerialise a business process and to interconnect it with its<br />
ecosystem, there is a need to study the external interfaces of its support systems. In this case, the<br />
degree of compatibility «DC» is calculated on the basis of a mapping between the underlying<br />
components and the adjacent processes.<br />
Several studies have focused on the characterization of the interoperation compatibility. Kasunic (2004)<br />
identifies for instance several indicators to describe this compatibility.<br />
The operational performance «PO» measurement is done on the basis of IT dashboards of involved<br />
public departments. It takes into account indicators as the availability score of the application servers, the<br />
quality of service of communication, and the end users degree of satisfaction about the interoperation in<br />
use. This information is collected based on surveying end users.<br />
3.3 Interoperability assessment method<br />
The present work proposes a five step method to assess interoperability needed to deliver a specific<br />
integrated public eService. These steps are as follows (see Figure 2):<br />
1. Delineating the scope of the study.<br />
2. Quantifying the interoperation potentiality.<br />
3. Calculating the compatibility degree.<br />
4. Evaluating the operating performance.<br />
5. Aggregating the degree of interoperability.<br />
Figure 2: Five steps of interoperability measurement (Elmir 2010a)<br />
173
3.4 Delineating the scope of the study<br />
Badr Elmir et al.<br />
Assessing interoperability, whether used or required, to deliver a specific integrated public eService<br />
requires the knowledge of its ecosystem.<br />
In practical terms, the study focuses on a macro business process consisting of a set of sub automated<br />
processes in independent departments. These sub processes are linked together via several interfaces<br />
identified in advance. In this case, the preliminary phase consists of identifying the context of the studied<br />
online service delivery and then lists its underlying automated processes.<br />
This step includes identifying:<br />
Public departments and agencies involved in the cooperation.<br />
Sub process within each department in order to study compatibility.<br />
Information systems that support automated business processes within each department.<br />
3.5 Quantifying the interoperation potentiality<br />
The calculation of the potential for interoperability within the k th department «PIk» requires the adoption of<br />
one of these maturity models mentioned above. The organization is classified then on one of these five<br />
levels noted IMML (for interoperation maturity model level). To identify the potential degree of<br />
interoperability, we propose then the following mapping (See Table 1):<br />
Table 1: Quantification of the maturity of the interoperability<br />
Maturity Level (IMML) Potentiality quantification<br />
1 0.2<br />
2 0.4<br />
3 0.6<br />
4 0.8<br />
5 1<br />
Within each department, the potential is calculated using the following equation:<br />
The final interoperation potentiality is given by:<br />
3.6 Calculating the degree of compatibility<br />
To assess the compatibility degree, the present work uses the compatibility matrix of Chen and Daclin<br />
(Chen 2008).<br />
The compatibility matrix (see Table 2) consists of a combination of the “interoperability levels perspective”<br />
and “interoperability barriers perspective” seen in section 2.2. In practical terms, we enumerate<br />
conceptual, technical and organisational barriers in the different layers of interoperability scope: business,<br />
process, service and data.<br />
Table 2: interoperability compatibility<br />
Conceptual Organizational Technology<br />
syntactic semantic<br />
authorities<br />
responsibilities<br />
Organization platform communication<br />
Business dc11 dc12 dc13 dc14 dc15 dc16<br />
Process dc21 dc22 dc23 dc24 dc25 dc26<br />
Service dc31 dc32 dc33 dc34 dc35 dc36<br />
Data dc41 dc42 dc43 dc44 dc45 dc46<br />
174
Badr Elmir et al.<br />
By noting the elementary degree of interoperation compatibility «dcij» (i takes values from 1..4, and j<br />
takes values from 1..6). Therefore, if the criterion in an area mark satisfaction the value 0 is assigned to<br />
dcij; otherwise if a lot of incompatibilities are met, the value 1 is assigned to dcij.<br />
The degree of compatibility «DC» is given as follows:<br />
3.7 Evaluating operating performance<br />
By Denoting:<br />
«DS» the overall availability rate of application servers.<br />
«QoS» service quality of different networks used between interacting components. QoS is<br />
represented mainly by the overall availability of networks.<br />
«TS» end users’ satisfaction level about interoperation.<br />
Given the cumulative nature of these three rates, the evaluation of operational performance is given by<br />
the geometric mean (DeFusco 2007) as the following equation:<br />
3.8 Aggregating the degree of interoperability<br />
The final calculation of the ratio characterising the interoperability process in question is by aggregating<br />
the three previous indicators using a function f defined in [0,1]3 [0,1]<br />
RatIop=f (PI, DC, PO)<br />
Given the independent nature of these three indicators, we opt for the arithmetic mean (DeFusco 2007)<br />
as follows:<br />
4. Periodic interoperability monitoring approach<br />
4.1 Interoperability monitoring approach<br />
This section shows the metamodel we propose for the interoperability monitoring approach. It includes:<br />
The exposed public integrated online service.<br />
The business processes supporting the delivery of the public online service.<br />
The connections and the compositions that exist between the involved business processes.<br />
The public departments that participate to the provision of the online service.<br />
The maturity model used in every department and their levels and the prerequisites to reach each<br />
level.<br />
The end users Satisfaction level.<br />
The enterprise architecture layers that coincide with the level of interoperation scopes.<br />
The elementary barriers that may obstruct interoperation situations.<br />
The periods within which interoperability is assessed.<br />
The metamodel serves as a basis for interoperability assessment periodically. It considers existing IT<br />
indicators within the public collaboration network like availability rate of application servers and the<br />
network. It includes end users satisfaction about used interoperation. This metamodel includes maturity<br />
score of each involved department. It references furthermore compatibility aspects on all levels of the<br />
enterprise architecture of collaboration in use.<br />
175
Badr Elmir et al.<br />
The metamodel is represented by using UML diagrams: class diagram (See Figure3).<br />
Figure 3: Interoperability monitoring Metamodel (class diagram)<br />
4.2 Periodic interoperability monitoring tool<br />
The Interoperability monitoring tool (IMT) includes three principal modules. The first one is dedicated to<br />
the interoperability assessment at a specific period. Figure 4 describes the interoperability assessment of<br />
an automated macro process within a single department which uses GIMM as maturity model. In this<br />
specific case, we notice that there are a lot of conceptual and organizational incompatibilities on a<br />
business layer.<br />
Figure 4: Screen from Interoperability monitoring tool<br />
176
Badr Elmir et al.<br />
In addition to its capacity to track the evolution of the interoperation degree periodically, the IMT gives the<br />
possibility to propose a scenario to reach a planned degree of interoperability. For instance, in the<br />
example shown on Figure 5, we plan to move the interoperability ratio calculated from 79% to 89%.<br />
IMT proposes to (i) improve the interoperability maturity to reach the fourth level, (ii) optimize the<br />
availability of involved application servers, (iii) better meet end users expectations and (iv) resolve<br />
conceptual incompatibilities.<br />
Figure 5: Planning of Interoperability optimization<br />
Interoperability monitoring tool is able to recognize prerequisites of going from a specific level of maturity<br />
to the next one.<br />
Figure 6: Sample of interoperability maturity optimization details<br />
5. Case study<br />
This section presents an illustration example on monitoring interoperability of an integrated public<br />
eService. The case in this paper consists of an online payment for health care services. It’s a citizen<br />
177
Badr Elmir et al.<br />
oriented eService proposed by a public hospital that offers special fees for persons covered by social<br />
security insurances.<br />
The studied online service is designed and implemented over an inter organizational information system<br />
within four public departments as described in figure 7. This eService includes:<br />
The identification and notification services within the public hospital system.<br />
The fee payment in national treasury department system.<br />
The administrative data validation in the national general mutual: it is oriented essentially to the civil<br />
servants.<br />
The administrative data validation in national social security fund: it is oriented to employees in the<br />
private sector.<br />
The data validation within private insurance: it is oriented to citizens who contracted a complementary<br />
insurance.<br />
Figure 7: Sample of online payment situation in eHealth system<br />
The interoperability is assessed every quarter in this collaboration network. During this period, the<br />
maturity level is brought up. The end user satisfaction is improved. The IT indicators are optimized (See<br />
Figure 8).<br />
Figure 8: Quarterly interoperability monitoring eHealth online payment<br />
178
5.1 Discussion<br />
Badr Elmir et al.<br />
The provided case study illustrates the results of interoperability monitoring within a specific collaboration<br />
context. This is provided by the interoperability monitoring tool (IMT) that collects existing IT indicators<br />
like availability score of Application servers and the end users satisfaction level. The IMT logs, in a<br />
convenient way, the interoperation incompatibilities in every architecture layer surrounding the delivery of<br />
the integrated eService.<br />
This automated method enables the establishment of an action plan that aims to improve the integrated<br />
eService delivery.<br />
The IMT is able to propose a scenario to reach a planned result for interoperability ratio. However the<br />
current version of the IMT is not yet able to propose the best scenario to achieve this result efficiently.<br />
This work is a prerequisite for several projects launched in parallel, and dealing with the applicability of<br />
control theory in the information system interoperability field. In this case, the future version of the IMT<br />
will use such techniques to enable the optimal control of interoperability.<br />
6. Conclusion<br />
The delivery of most useful public online services often requires cooperation between two or more public<br />
administrations. This cooperation takes, in general, the form of business processes interoperability<br />
among public departments. The present paper has presented a periodic monitoring approach for<br />
interoperability to deliver integrated public eServices. This approach is based on a five step assessment<br />
method that uses existing indicators within involved departments like quality maturity indicators,<br />
information technology dashboards, etc.<br />
The result of this automated method is a ratio metric enabling the measurement of this quality by taking<br />
into account three main operational aspects: interoperation potentiality, interoperation compatibility and<br />
operational performance.<br />
The proposed method supporting tool tracks the evolution of interoperation degree in time and is able to<br />
propose a suitable scenario to reach planned degrees of interoperability.<br />
References<br />
Aisopos, F., Tserpes, K., Kardara, M., Panousopoulos, G., Phillips, S. and Salamouras, S. (2009) “Information<br />
exchange in business collaboration using grid technologies, Identity in the Information Society”, Springer 2009,<br />
Volume 2, Number 2, pp. 189-204.<br />
Arms, W. Y., Hillmann, D., Lagoze, C., Krafft, D., Marisa, R., Saylor, J. et al. (2002) “Spectrum of Interoperability”:<br />
The Site for Science Prototype for the NSDL, D-Lib Magazine, 8 (1), January (2002)<br />
Bercea, L., Nemtoi, G. and Ungureanu, C. (2010), “The government of state’s power bodies by means of the<br />
Internet”, Journal of Computing, Vol. 2, Issue 2, February 2010.<br />
Chen, D. and Daclin, N. (2006) “Framework for Enterprise Interoperability”. IFAC TC5.3 workshop EI2N, Bordeaux,<br />
France.<br />
Chen, D., Vallespir, B. and Daclin, N. (2008) “An Approach for Enterprise Interoperability Measurement”. Paper read<br />
at MoDISE-EUS 2008. France.<br />
Chen, J. (2009) “Qualitative Evaluation of EGovernment Service Based on Participatory Index and Balanced<br />
Scorecard Approach”, iciii, International <strong>Conference</strong> on Information Management, Innovation Management and<br />
Industrial Engineering, Vol. 4, pp. 380-383.<br />
Chrissis, M.B., Mike, C., Sandy S. (2006) CMMI: Guidelines for Process Integration and Product Improvement.<br />
Addison-Wesley. USA.<br />
DeFusco, R., McLeavey, D.W. and Pinto J.E. (2007) Quantitative investment analysis. Vol. 2 of CFA Institute<br />
investment series, pp. 127, John Wiley and Sons, Hoboken, NJ, USA.<br />
Elmir, B. and Bounabat, B. (2010a) “Integrated Public EServices Interoperability Assessment”. International Journal<br />
of information Science and Management, Special issue, October 2010, pp 1-12.<br />
Elmir, B. and Bounabat, B. (2010b) (Elmir-b) “EService Interoperability Measurement within Business Collaboration<br />
Networks”. Paper read at MICS 2010. International <strong>Conference</strong> on Models of Information and Communication<br />
Systems. Morocco.<br />
Goldkuhl, G. (2008) “The challenges of interoperability in eGovernment: towards a conceptual refinement”. Pre-ICIS<br />
2008 SIG eGovernment Workshop, Paris.<br />
Guédria, W., Naudet, Y. and Chen, D. (2008) “Interoperability maturity models - survey and comparison-“. Lecture<br />
Notes in Computer Science, Springer Berlin / Heidelberg. Vol. 5872/2009, pp. 216-225,<br />
Guijaro, L. (2007) “Frameworks for fostering cross-agency interoperability in eGovernment”. Electronic government:<br />
information, and transformation, within the advances in management information system (AMIS) series (ME<br />
Sharp).<br />
179
Badr Elmir et al.<br />
Gupta, K.M., Zang, M., Gray, A., Aha, D.W., and Kriege, J. (2007) “Enabling the interoperability of large-scale legacy<br />
systems”. (Technical Note AIC-07-127). Washington, DC: Naval Research Laboratory, Navy Center for Applied<br />
Research in Artificial Intelligence.<br />
Kasunic, M. and Anderson W. (2004) “Measuring Systems Interoperability: Challenges and Opportunities”. Technical<br />
Note CMU/SEI-2004-TN-003. Carnegie Mellon University, Pittsburgh.<br />
Klischewski, R. (2004) “Information Integration or Process Integration? How to Achieve Interoperability in<br />
administration”. In Traunmüller, R., Electronic Government: Third International <strong>Conference</strong> EGOV 2004, Lecture<br />
Notes in Computer Science 3183, Berlin: Springer Verlag, pp.57-65.<br />
Liimatainen, K., Hoffmann, M. and Heikkilä, J. (2007) “Overview of Enterprise Architecture work in 15 countries”.<br />
Ministry of Finance, State IT Management Unit, Research reports 6b/2007.<br />
Michelson, B. (2006) “Event-Driven Architecture Overview”. Technical Report. Patricia Seybold Group. Boston, USA.<br />
Missikoff, M., Schiappelli, F. and D'Antonio F. (2004) “State of the Art and State of the Practice Including Initial<br />
Possible Research Orientations”. Deliverable D8.1. <strong>European</strong> Commission, INTEROP Network of Excellence.<br />
Pardo, A. and Burke B. (2009) “Improving government interoperability. A capability framework for government<br />
managers”. Technical Report. Center for technology in government. University at Albany, NY, USA.<br />
Peristeras, V., Mentzas, G., Tarabanis, K.A. and Abecker, A. (2009) “Transforming EGovernment and E-participation<br />
through IT”, IEEE Intelligent Systems, Vol. 24, no. 5, pp. 14-19, Sep./Oct. 2009.<br />
Shambour, Q. and Lu, J. (2010) “A Framework of Hybrid Recommendation System for Government-to-Business<br />
Personalized EServices”, Seventh International <strong>Conference</strong> on Information Technology, pp. 592-597.<br />
Shishkov, B., Sinderen, M. and Verbraeck, A. (2009) “Towards flexible inter-enterprise collaboration: a supply chain<br />
perspective”, Lecture Notes in Business Information Processing, 1, Volume 24, Enterprise Information Systems,<br />
III, pp. 513-527.<br />
Sterling (2010) Optimize and Transform Your Business Collaboration Network, Sterling Commerce white paper.<br />
Sun, H.H., Huang S. and Fan, Y. (2007) “SOA-Based Collaborative Modeling Method for Cross-Organizational<br />
Business Process Integration”, Lecture Notes in Computer Science, 2007, Volume 4537, Advances in Web and<br />
Network Technologies, and Information Management, pp 522-527.<br />
Vintar, M. and Leben, A. (2002) “The Concepts of an Active Life-event Public Portal”. In Proceedings of the First<br />
International <strong>Conference</strong> on Electronic Government. Springer-Verlag, Berlin, pp. 383-390.<br />
180
Enterprise Resource Planning Implementation Differences<br />
Within the Same Methodology - Case Study From West<br />
Europe and Turkey<br />
Turan Erman Erkan<br />
At l m University, Ankara, Turkey<br />
ermanerk@atilim.edu<br />
Abstract: Enterprise Resource Planning (ERP) systems are vital for competitive edge in today’s business world.<br />
ERP is an integrated information system that serves all departments within an enterprise. Therefore, business<br />
processes should be optimised before ERP implementation in order to catch a perfect implementation. Before year<br />
2000 all the giant firms realised their ERP projects and after that best practises took place. After those trial and error<br />
based experiences, ERP vendors need to form an ERP implementation methodology. Big ERP vendors developed<br />
implementation methodologies, they were almost the same; starting from project preparation to selection,<br />
implementation to maintenance and control. Those project methodologies improved the success rate of ERP<br />
implementations. This research is both empirical and qualitative. In fact it consists of two monographic researches.<br />
They are both about ERP implementations in the same sector. One of the firms is a national one and other one is a<br />
multinational one. Both implementations done by the same consultant team with the same methodology, so the<br />
evaluation of the project is more objective than any other questionnaire based ones, which are filled by different<br />
implementation teams. The findings of the comparison are surprisingly different from each other within the<br />
implementations, although they both used the same methodology. Basic steps of the ERP implementation such as:<br />
project preparation, business blueprint, realization, final preparation and go live & support differs from each other<br />
both in national and multinational firms. The major difference is seen in the first steps which are project preparation<br />
and business blueprint. Multinational firm seems to have more chance than the national one in achieving<br />
organisational efficiency through successful ERP implementation. The research outcome is useful for professionals<br />
running implementation projects and those making decisions on ERP implementation. The results can also be used<br />
by practitioners managing ERP projects in order to avoid from implementation methodology illusion.<br />
Keywords: enterprise resource planning, ERP project management, ERP implementation, cultural differences<br />
1. Introduction<br />
The business environment is dramatically changing. Companies today face the challenge of increasing<br />
competition, expanding markets, and rising customer expectations. This increases the pressure on<br />
companies to lower total costs in the entire supply chain, shorten throughput times, drastically reduce<br />
inventories, expand product choice, provide more reliable delivery dates and better customer service,<br />
improve quality, and efficiently coordinate global demand, supply, and production (Umble, 2003).<br />
Therefore, the changes in the business perspectives, goals objectives and strategies are pressuring on<br />
organizations and on their structure to be upgraded in spite of their cultures and values. Knowledge<br />
sharing is not a competitive advantage anymore the challenge is in knowledge management.<br />
As the business world evolved it was no longer adequate for companies to merely offer their goods for<br />
sale, in order to stay viable they had to keep their competitive advantage (Tersine 2003).<br />
in the ’60s industry concentrated on how to produce more (quantity),<br />
in the ’70s how to produce it cheaper (cost)<br />
in the ’80s how to produce it better (quality)<br />
in the ’90s how to produce it quicker (lead time)<br />
in the 21st century how to offer more (service<br />
ERP systems have been considered an important development in the corporate use of information<br />
technology in the 1990s, enhancing organizational cross-functional efficiency and effectiveness through<br />
the seamless integration of all the information flowing through a company (Davenport, 1998).<br />
ERP is the business backbone. It is a cross-functional enterprise system that integrates and automates<br />
many of the internal business processes of a company, particularly those within the manufacturing,<br />
logistics, distribution, accounting, finance, and human resource functions of the business. Thus, ERP<br />
serves as the vital backbone information system of the enterprise, helping a company achieve the<br />
efficiency, agility, and responsiveness required to succeed in a dynamic business environment<br />
(Davenport, 1998) and (Olson, 2004). ERP software typically consists of integrated modules that give a<br />
181
Turan Erman Erkan<br />
company a real-time cross-functional view of its core business processes, such as production, order<br />
processing, and sales, and its resources, such as cash, raw materials, production capacity, and people.<br />
However, properly implementing ERP systems is a difficult and costly process that has caused serious<br />
business losses for some companies, which underestimated the planning, development, and training that<br />
were necessary to reengineer their business processes to accommodate their new ERP systems. ERP<br />
literature suggest that an ERP system alone cannot improve the company performance unless an<br />
organization restructures its operational processes, and this is generally accomplished through business<br />
process reengineering (Bingi, Sharma, and Godla 2003, Olson 2004 and Davenport 1998).<br />
1.1 Enterprise resource planning project management<br />
In order to study ERP implementation projects, some researchers are using a critical success factors<br />
(CSFs) approach (Esteves and Pastor 2001, Nah et al 1999). However, little has been done in relation to<br />
the management and the operationalization of these CSFs. Project evaluation is critical to the<br />
understanding, control and monitoring of the CSFs of an ERP implementation project. ERP project<br />
success is influenced by a large number of factors, and most of the times it is difficult to measure them<br />
objectively. Adequate Business Process Redesign (BPR) is one of the most cited CSFs in ERP<br />
implementation projects (e.g. Bancroft et al. 1998, Bingi et al. 1999, Holland et al. 1999, Nah et al. 2001).<br />
Comprehensive BPR is related with the alignment between business processes and the ERP business<br />
model and associated best practices (Esteves and Pastor 2000). As Valiris and Glykas (1999) mention, in<br />
the literature there has been some confusion regarding the use of terms like reengineering, process<br />
improvement and redesign. They suggest that reengineering is synonymous to radical change and<br />
process improvement to incremental change and that both, reengineering and process improvement are<br />
included in the definition of redesign.<br />
This CSF stream focuses on the details associated with implementing ERP systems and their relative<br />
success and cost. Specifically, according to Motwani et al (2005); the articles in this stream include topics<br />
such as the implementation procedures (Abdinnour et al 2003, Al-Mudimigh et al 2001, Mandal and<br />
Gunasekaran 2003 and Umble 2003), critical success factors (Al-Mashari et al 2003, Nah et al 2001,<br />
Clemmons and Simon 2001, Davison 2002 and Hong and Kim 2002), pitfalls and complexities in ERP<br />
implementation (Ip et al 2002, McAlary 1999 and Ribbers and Schoo 2002), and successful strategies for<br />
effective ERP implementation (Aladwani 2001, Cliffe 1999, Markus et al 2000a and 2000b, Scheer and<br />
Habermann 2000, Willis and Brown 2002 and Mabert, Soni and Venkataramanan 2003).<br />
Based on the preceding review of the literature done by Motwani (2005) and also on the research by<br />
Akkermans and van Helden (2002), Grabski, Leech and Lu (2001), and Somers and Nelson (2001),<br />
Grabski, Stewart, and Leech (2007) developed a list of ERP implementation controls,<br />
Business process reengineering<br />
Consultants' involvement<br />
Top management support<br />
Active steering committee<br />
Knowledgeable project team<br />
Close working relationship between the project team and consultants<br />
Detailed requirements specification<br />
Detailed implementation plan<br />
Frequent communication with the users<br />
Managing people<br />
User involvement<br />
Training<br />
Involvement of internal audit<br />
System testing prior to implementation<br />
Close monitoring after implementation<br />
Change management and transition management<br />
182
Develop users' project ownership<br />
In-depth, up front project planning<br />
Project management skills<br />
Project sponsor from top management<br />
Clearly identified objectives<br />
Specified measures of success<br />
Ways to manage risk<br />
Turan Erman Erkan<br />
Detailed tracking of actionable items by internal audit<br />
Monthly internal audit reports on project risk items to steering committee.<br />
Therefore, both ERP vendors and clients start caring about those controls and improve their<br />
implementation methodologies.<br />
1.2 SAP ASAP methodology<br />
The ASAP solution was developed to ensure the successful, on-time delivery of a project. SAP delivers<br />
the Accelerated SAP (ASAP) methodology for project management and system implementation.<br />
Developed by SAP to optimize the success of implementing the SAP Business Suite, ASAP streamlines<br />
the implementation by providing templates, methods, tools, and accelerators that have been built on the<br />
success of thousands of previous SAP implementations (SAP, 2001)<br />
The ASAP Roadmap provides the methodology for implementing and continuously optimizing your SAP<br />
System. It divides the implementation process into five phases and offers detailed Project Plans to assist<br />
you (in Microsoft Project format). The documentation stored at each level of the Roadmap tree structure<br />
contains recommendations on implementing your SAP System and links to helpful tools and accelerators<br />
(Niefert 2009, Portougal and Sundaram 2006)<br />
The implementation of your SAP System covers the following phases:<br />
Project Preparation<br />
In this phase you plan your project and lay the foundations for successful implementation. It is at this<br />
stage that you make the strategic decisions crucial to your project:<br />
Define your project goals and objectives<br />
Clarify the scope of your implementation<br />
Define your project schedule, budget plan, and implementation sequence<br />
Establish the project organization and relevant committees and assign resources<br />
Business Blueprint<br />
In this phase you create a blueprint using the Question & Answer database (Q&Adb), which documents<br />
your enterprise’s requirements and establishes how your business processes and organizational<br />
structure are to be represented in the SAP System. You also refine the original project goals and<br />
objectives and revise the overall project schedule in this phase.<br />
Realization<br />
In this phase, you configure the requirements contained in the Business Blueprint. Baseline configuration<br />
(major scope) is followed by final configuration (remaining scope), which can consist of up to four cycles.<br />
Other key focal areas of this phase are conducting integration tests and drawing up end user<br />
documentation.<br />
Final Preparation<br />
183
Turan Erman Erkan<br />
In this phase you complete your preparations, including testing, end user training, system management,<br />
and cutover activities. You also need to resolve all open issues in this phase. At this stage you need to<br />
ensure that all the prerequisites for your system to go live have been fulfilled.<br />
Go Live & Support<br />
In this phase you move from a pre-production environment to the live system. The most important<br />
elements include setting up production support, monitoring system transactions, and optimizing overall<br />
system performance.<br />
2. Case study: ERP implementations<br />
A case study approach was employed to identify the projects steps that were held differently among the<br />
national and international firm projects. The criterion used to select the case study companies was that<br />
each of the case studies should use ERP software from the same vendor. Data was collected primarily<br />
through questionnaires and observations.<br />
This case study consists of two ERP implementations. In fact, this a monographic study of two distinct<br />
firms. Firms are from Turkey because of confidentiality the sector would not be shared. one is<br />
multinational the other one is national. The purpose of the case study is to examine the ERP<br />
implementation method difference between the national and multinational firms.<br />
Both enterprises implemented the same ERP with same modules as it can be seen from Table I. The<br />
author also participated to both of the projects as a BPR and ERP consultant. Therefore, data collected<br />
from the original source. In case of confidentiality the enterprises would take place as Firm A<br />
(international) and Firm B (national)<br />
Table 1: ERP modules used in firms<br />
ERP Module Firm A Firm B<br />
FI Financial Accounting X X<br />
CO Controlling X X<br />
TR Treasury X X<br />
IM Investment Management X X<br />
PP Production Planning X X<br />
MM Materials Management X X<br />
SD Sales and Distribution X X<br />
QM Quality Management X X<br />
WM Warehouse Management X X<br />
PM Plant Maintenance X X<br />
CS Customer Service X X<br />
PS Project System X X<br />
HR Human Resources X X<br />
In Firm A implementation lasted 14 months and in Firm B implementation lasted 8 months. In both firms<br />
ASAP methodology was used. As seen from Table 1 same modules of ERP had been implemented by<br />
the same project team. There is something like 50% difference in implementation time of projects. Thus 6<br />
months difference is mainly because of project preparation and business blueprint steps. Those initial<br />
and vitals steps considered seriously in the international firm rather than the national one.<br />
Table 2 indicates the grades of the project evaluation that was realized by consultant team out of 10 for<br />
each project step. Average and standard deviation of those grades calculated for each step of ASAP<br />
methodology. According to Table 2, in all steps of ASAP methodology there is an obvious performance<br />
evaluation difference between multinational and national firms.<br />
The biggest difference is in project preparation (9.90 to 5.80) and business blueprint steps (9.80 to 6.00)<br />
which are the essential steps of an ERP project. Starting from realization (9.00 to 7.10) and continuing<br />
with final preparation (9.80 to 6.80) and go live & support (9.90 to 8.10) the difference between national<br />
and international implementation reduce. Overall average of five steps (project preparation, business<br />
184
Turan Erman Erkan<br />
blueprint, realization, final preparation, go live & support) are 9,68 and 6.76 respectively in international<br />
and national firms. In other words simply one can easily say that implementation quality is 96.8% in<br />
international firm and 67.6% in national firm.<br />
Table 2: ERP project steps evaluation realized by consultant team<br />
Firm A Firm B<br />
ASAP Steps Average St. D.* Average St. D.*<br />
Project Preparation 9.90 0.32 5.80 1.03<br />
Business Blueprint 9.80 0.42 6.00 0.67<br />
Realization 9.00 0.82 7.10 0.88<br />
Final Preparation 9.80 0.42 6.80 0.79<br />
Go Live & Support 9.90 0.32 8.10 0.74<br />
Overall Average 9.68 6.76<br />
St.D.*: Standard Deviation<br />
3. Conclusion<br />
This research attempted to answer the question. ‘‘Is there a difference in implementing ERP via same<br />
project methodology between national and international firms?” The answer was quite surprising; primary<br />
steps of the ERP implementation such as: project preparation, business blueprint, realization, final<br />
preparation and go live & support differ from each other both in national and multinational firms. The<br />
major difference is seen in the first steps which are project preparation and business blueprint. Therefore,<br />
multinational firm seems to have more probability than the national one in achieving a successful ERP<br />
implementation. Since the questions were asked to the ERP consultants that they participated to the<br />
project, there was no problem about the perception of the questions and therefore answers were<br />
objective.<br />
Although this research is a monographic one, it indicates the project management difference between<br />
national and international firms in Turkey. As discussed in the previous sections both project period (14 to<br />
8 months) and evaluated performance (9.68 to 6.76) differ among international and national firms.<br />
National firms usually do not have budget and time for business process reengineering, organizational<br />
change management and so on.<br />
In conclusion, there is a recordable ERP project implementation difference between multinational and<br />
national firms although they use the same methodology. The difference is mainly on the way of doing<br />
business, the planning method and application. This might be because of cultural differences and<br />
business circumstances. National firms’ ERP projects are mostly about configuration of the software to<br />
the enterprise not a tailor made customization for them to achieve organisational efficiency through<br />
successful ERP implementation. National firms need to redesign their organizations, reengineer their<br />
processes, enforce employees in order to increase their chance of ERP well implementation.<br />
The research outcome is useful for professionals running implementation projects and those making<br />
decisions on ERP implementation. The results can also be used by practitioners managing ERP projects<br />
in order to avoid from implementation methodology illusion. This study was subject to further limitations<br />
due to the measurements of cultural effects and economically readiness of firms that could be a subject<br />
of an other study.<br />
References<br />
Abdinnour-Helm, S. Lengnick-Hall, M. and Lengnick-Hall, C. (2003) “Pre-implementation attitudes and Organizational<br />
Readiness for Implementing an Enterprise Resource Planning System”, <strong>European</strong> Journal of Operational<br />
Research 146 (2) pp 258-273.<br />
Akkermans, H. and K. Helden, K. (2002) “Vicious and virtuous cycles in ERP implementation: a case study of<br />
interrelations between success factors” <strong>European</strong> Journal Information Systems, 11 (1) pp 35–46.<br />
Aladwani, A. (2001) “Change management strategies for successful ERP implementation”, Business Process<br />
Management 7 (3) pp 266-275<br />
Al-Mashari, M. Al-Mudimigh, A. and Zairi, M. (2003) “Enterprise Resource Planning: A Taxonomy of Critical Factors”,<br />
<strong>European</strong> Journal of Operational Research 146 (2) pp 352-364.<br />
Al-Mudimigh, A. Zairi, M. and Al-Mashari, M. (2001) “ERP Software Implementation: an Integrative Framework”,<br />
<strong>European</strong> Journal of Information Systems 10 (4) pp 216-226.<br />
Bancroft, N. Seip H.and Sprengel A. (1998) "Implementing SAP R/3", 2nd ed., Manning Publications.<br />
185
Turan Erman Erkan<br />
Bingi, P. S. Sharma M. K., and Godla J. (1999) “Critical issues affecting an ERP implementation”. Information<br />
Systems Management,16(3) pp7–14.<br />
Clemmons, S. and Simon, S. (2001) “Control and Coordination in Global ERP Configuration”, Business Process<br />
Management Journal 7 (3) pp 205-215.<br />
Cliffe, S. (1999) “ERP implementation”, Harvard Business Review 77 (1) pp 16-17.<br />
Davenport, T.H. (1998) “Putting the Enterprise into the Enterprise System” Harvard Business Review 76 (4), pp.<br />
121–131.<br />
Davison, R. (2002) “Cultural complications of ERP”, Communications of the ACM 45 (7) pp109-111<br />
Esteves, J. and Pastor, J. (2000) "Towards the Unification of Critical Success Factors for ERP implementations",<br />
10th Annual BIT conference, Manchester, UK.<br />
Esteves, J. and Pastor, J. (2001) "Analysis of Critical Success Factors Relevance along SAP Implementation Phases<br />
", Americas <strong>Conference</strong> on Information Systems.<br />
Esteves, J. Pastor, J. and Casanovas, J. (2002) "Monitoring Business Process Redesign in ERP Implementation<br />
Projects", Americas <strong>Conference</strong> on Information Systems (AMCIS), Dallas (USA).<br />
Grabski, S. V. Leech, A. and Lu, B. (2001) “Risks and Controls in the Implementation of ERP Systems”. Int J Digit<br />
Account Res;1(1) pp 51–78.<br />
Grabski, S. V. Stewart, A. and Leech,A. (2007) “Complementary Controls and ERP Implementation Success”<br />
International Journal of Accounting Information Systems 8 pp 17–39.<br />
Holland, C. P. Light, B.and Gibson, N. (1999) "A Critical Success Factors Model for Enterprise Resource Planning<br />
Implementation", <strong>European</strong> <strong>Conference</strong> on Information Systems, pp. 273-279.<br />
Hong, K. and Kim, Y. (2002) “The critical success factors for ERP implementation: an organizational fit perspective”,<br />
Information and Management (1) pp 25-40<br />
Ip, W. Chau, K. and Chan, S. (2002) “Implementing ERP through continuous Improvement”, International Journal of<br />
Manufacturing 4 (6) pp 465-478<br />
Mabert, V. Soni, A and Venkataramanan, M. (2003) “Enterprise Resource Planning: Managing the Implementation<br />
Process”, <strong>European</strong> Journal of Operational Research 146 (2) pp 302-314<br />
Mandal, P. and Gunasekaran, A. (2003) “Issues in implementing ERP: a case study”, <strong>European</strong> Journal of<br />
Operational Research 146 (2) pp 274-283<br />
Markus, M. Axline, S. Petrie, D. and Tanis, C. (2000b) “Learning from adopters’ experiences with ERP: problems<br />
encountered and success achieved”, Journal of Information Technology 15 (4) pp 245–266.<br />
Markus, M. Tanis, C. and Fenema, P. (2000a) “Multisite ERP implementations”, Communications of the ACM 43 (4)<br />
pp 42–46.<br />
McAlary, S. (1999) “Three pitfalls in ERP implementation”, Strategy and Leadership 27 (6) pp 49-53.<br />
Motwani, J, Subramanian, R and Gopalakrishna, P (2005) “Critical factors for successful ERP implementation:<br />
Exploratory findings from four case studies”, Computers in Industry Volume 56, Issue 6, pp 529-544<br />
Nah, F. Lau, J. and Kuang, J. (2001) "Critical Factors for Successful Implementation of Enterprise Systems",<br />
Business Process Management Journal, 7(3), pp. 285 -296.<br />
Nah, F. Lee-Shang J. Lau, J. and Kuang, J. (2001) “Critical Factors for Successful Implementation of Enterprise<br />
Systems”, Business Process Management Journal 7 (3) pp 285-296.<br />
Niefert, W. (2009) SAP Business ONE Implementation, Packt Publishing, Birmingham.<br />
Olson, L. (2004) Managerial Issues of Enterprise Resource Planning Systems, McGraw Hill, NY,.<br />
Portougal, V. and D. Sundaram, (2006) Business processes: Operational Solutions for SAP Implementation, IRM<br />
Press London<br />
Ribbers, P. and Schoo, K. (2002) “Program management and complexity of ERP implementations”, Engineering<br />
Management Journal 14 (2) pp 45-54.<br />
SAP (2001), ASAP Methodology service.sap.com/asap, [Accessed 7.12.2001]<br />
Scheer, A. and Habermann, F. (2000) “Making ERP a Success”, Communications of the ACM 43 (4) pp 57–61.<br />
Somers, T. and Nelson, K. G. (2001) “The impact of critical success factors across stages of enterprise resource<br />
planning implementations” Proceedings of the 34th Hawaii International <strong>Conference</strong> on Systems Science.<br />
Tersine R., (2003) Open University, Available: www.ou.edu/class/tersine.<br />
Umble, E. J. Haft, R. R. Umble, M. M. (2003) “Enterprise resource planning: Implementation procedures and critical<br />
success factors” <strong>European</strong> Journal of Operational Research 146 pp. 241–257.<br />
Valiris, G. and Glykas, M. (1999) “Critical Review of Existing BPR Methodologies”, Business Process Management<br />
Journal, 5(1), pp. 65-86.<br />
Willis, T. and Brown, A. (2002) “Extending the value of ERP”, Industrial Management and Data Systems 102 (1/2) pp<br />
35-38<br />
186
A Meta-Analysis of how the Adoption of Innovative Web 2.0<br />
Tools Like Digital Business Ecosystems can Lead to Improved<br />
SME Collaboration<br />
Francesc Estanyol<br />
University of Edinburgh Business School, Edinburgh, UK<br />
francesc.casals@ed.ac.uk<br />
Abstract: This work studies the increasing role of Web 2.0 tools as enabler of SME collaboration. The author analyses<br />
how innovative Information and Communication Technologies (ICT) like Digital Business Ecosystems (DBE)<br />
could help small and medium enterprises (SMEs) to reduce the gap with large corporations. Nowadays Web 2.0 is<br />
permitting to perform some tasks unimaginable only few years ago. Individuals have evolved from a passive to an<br />
active role and currently are the main actors creating Internet content. However, compare to individuals, business still<br />
in the first stage of Internet revolution and are not taking profit of all the possibilities of Web 2.0. Most of them are<br />
using these tools only to improve existing procedures but not for creating new ways of doing business or interact with<br />
other SMEs. In this scenario, DBE has emerged as a new paradigm to solve some of these problems and to permit<br />
businesses behave in a similar way to what individual users are doing nowadays. In this work, after an analysis of the<br />
nature of co-operative behaviour and the importance of Web 2.0 facilitating its emergence by promoting selforganisation<br />
of their participants, the author evaluates this new paradigm in order to determine if it is a valid approach<br />
for SMEs. Finally, based on previous works of the author about the needs of SMEs, the study concludes with<br />
the definition of MADBE, an innovative multi-agent Digital Business Ecosystems inspired by natural ecosystems<br />
which aims to facilitate collaboration between SMEs.<br />
Keywords: SME, collaboration, multi-agents, digital, ecosystems, business<br />
1. Introduction<br />
The field of SME collaboration has become very popular from the 90s. Terms such as networking, cooperation<br />
or alliances have been widely used in literature in works like (Hoffmann and Schlosser 2001),<br />
(Varamaki and Vesalainen 2003) or (Narula 2004) but in spite of some studies like (Coombs et al. 1996)<br />
have demonstrated the advantages of adopting collaboration and co-operation approaches for improving<br />
business performance and competitiveness, SMEs, due their scarcity of time, knowledge and resources,<br />
have experienced problems implementing and maintaining successful alliances.<br />
Considering a SME as defined in (EU 2009), SMEs represent 99 percent of non-financial businesses in<br />
US and EU and, in addition to their presence, play a key role creating employment: about two-thirds of<br />
total employment in the private sector; economic growth: contributing to more than half of the total valueadded<br />
created by businesses, innovation and social integration (Audretsch et al. 2009). This influence is<br />
not exclusive of western countries and works like (Zeng et al. 2010) present similar figures for Asian<br />
countries like China, Hong Kong or Taiwan.<br />
However, the above-mentioned studies also show that small firms have lower profitability, employee<br />
compensation, and labour productivity than large enterprises. In this context collaboration emerges as a<br />
potential solution but most of firms are reluctant to collaborate even if there is evidence of mutual benefits.<br />
In (Estanyol 2010), the author identifies a set of problems and barriers SME face in order to adopt<br />
collaborative approaches. Among of them, the self-centred behaviour of participants giving priority to their<br />
individual interests or for fear that others will take advantage of their co-operative decision is one of the<br />
most important. In addition, the lack of resources, mainly time, money and personnel, force SMEs to focus<br />
on their daily activities, losing the opportunity to create new business, enter new markets or create<br />
new products in collaboration with other SMEs.<br />
On the other hand, as summarised in (Nooteboom 1994) traditionally large corporations have been more<br />
efficient using external networks. They take advantage of their higher capital reserves and negotiation<br />
power to obtain favourable alliances with other organizations while SMEs need to find imaginative ways<br />
to survive. In uncertain economic periods, this gap increases and the ability to collaborate with others can<br />
make the difference between surviving or not. Fortunately, nowadays this situation is changing and the<br />
emergence of Web 2.0 technologies like DBE are creating a new range of options for collaboration, permitting<br />
SMEs to combine the use of these tools with their major flexibility in order to overcome their<br />
weaknesses by developing smart alliances and to reduce the gap with large corporations (Narula 2004).<br />
187
Francesc Estanyol<br />
The objectives of this work are as follows: a) to study the current context of SME collaboration in order to<br />
b) analyse how Web 2.0 technologies facilitate the emergence of co-operative behaviour and c) they role<br />
as enabler of business collaboration. Knowing the current situation permits to understand the problems<br />
SME face in order to adopt innovative ICT tools and to d) introduce DBEs as an alternative to facilitate<br />
businesses co-operation. Finally, the main objective of this study is to e) define and present MADBE, the<br />
first ecologically inspired multi-agent to DBE and its main contribution to research.<br />
2. The importance of Web 2.0 for co-operation<br />
During the last decade, ICT and Internet have changed our habits in a wide range of activities. Nowadays,<br />
it is difficult to understand our lives without e-mails, on-line tickets, digital photography or mp3 music.<br />
Although important, is not until the advent of Web 2.0 that the use of the Internet can be defined as<br />
revolutionary. Before that, services and content were made by IT specialists and users had a passive role<br />
interacting with them, simply viewing or downloading it. Software companies and Internet businesses<br />
were the players, the owners and who decided what, where, how and when to publish information on the<br />
World Wide Web (WWW).<br />
The term Web 2.0, firstly coined by Dancy DiNucci in 1999 but popularised by Tim O’Reilly in (O’Reilly<br />
2005) refers to the evolution of the Web and it is associated with applications where users are the main<br />
actors creating the content of the WWW. Although most researchers agree on some common concepts<br />
and principles around Web 2.0, currently there is no accepted definition. In this work, the author uses the<br />
term Web 2.0 as defined in (Hoegg et al. 2006), where the authors clearly separate the concept from<br />
specific technologies and define it as a philosophy or a new way to create web content which implies a<br />
greater collaboration among Internet users and the possibility to create communities.<br />
Independently of the scientific definition, most Web 2.0 applications like Facebook, Twitter, Flickr, Blogger<br />
or Wikipedia share some principles such as information sharing, interoperability, collaboration and cooperation.<br />
But the most powerful consequence of Web 2.0 is that it permits the creation and sharing of<br />
content in real-time. As pointed out by Clay Shirky in (Shirky 2009), nowadays it is possible to be informed<br />
about news, local traffic, deals or just friend information using these types of applications and all<br />
this faster than the official information channels such as TV, radio or newspapers.<br />
2.1 Self-organisation and collaborative behaviour<br />
With this revolution, the role of users has changed radically during the last years. They evolved from passive<br />
to active and nowadays are able to organise themselves, build communities and share information<br />
independently of the external channels.<br />
Web 2.0 facilitates self-organisation of participants and the emergence of collaborative behaviour, two<br />
key aspects to ensure co-operation success. Nowadays, individuals are able to identify something they<br />
would like to do, find others who would enable that action to be a success and access the resources to<br />
move to action. When users are involved in these types of actions, collaborative behaviour emerges naturally<br />
and facilitates the sharing of resources and the creation of a feeling of belonging to a community and<br />
therefore, makes users less dependent on traditional channels.<br />
Web 2.0 applications like CouchSurfing.com, HomeExchange.com or Compartir.org enable users to<br />
specify their interests (e.g. location, days, traffic route, timetable, etc.) and meet others offering the same.<br />
Apart from obvious economic and environmental benefits, these initiatives help to establish new relationships<br />
between users and to promote the creation of social communities with similar preferences.<br />
These tools have created a new way to perform old activities and represent an alternative to traditional<br />
services and products. This is possible because with Web 2.0 users are able to comment inputs, make<br />
recommendations and evaluate, rate, vote or tag users’ information. These actions permit to rank the<br />
users depending on their reputation and help the creation of trust and collective intelligence within the<br />
community.<br />
3. Web 2.0 as enabler for business collaboration<br />
The examples presented in previous section are useful to illustrate how Web 2.0 is changing people’s<br />
habits and routines but businesses, and SMEs in particular, are far from this situation. Despite of abundant<br />
proofs that the Internet is effective for gaining a competitive advantage, most small firms are not<br />
making full use of it (Martin and Matlay 2003).<br />
188
Francesc Estanyol<br />
Favoured by the declining costs of technology, businesses currently have a wide range of options to organise<br />
themselves and improve the relationships with their providers/clients. Business networking applications<br />
like LinkedIn; Business Intelligence (BI) tools like Pentaho or MicroStrategy; or Customer relationship<br />
management (CRM) solutions permit to control organisational activity more effectively, support<br />
the decision making process, establish links with other partners or facilitate co-operation processes that<br />
require flexibility and trust (Caby-Guillet et al. 2006).<br />
But compare with individuals, most firms are still in the first stage of the Internet revolution and are using<br />
these tools mostly for upgrading existing mechanisms, for example, replacing postal mail by e-mail or<br />
communicating their corporate vision using websites instead of other traditional media but not for creating<br />
new ways of doing businesses or interact with other SMEs, missing the real possibilities that Web 2.0<br />
brings (Hoegg et al. 2006).<br />
In this context, recently DBEs have emerged as a new Web 2.0 paradigm to solve some of these problems.<br />
DBEs permit to access dynamically to valuable information about other participants (e.g. background,<br />
previous collaboration, results, etc.) and to introduce reputation, voting or tagging systems to<br />
ensure security, trust and commitment among the participants. These characteristics promote the creation<br />
of digital environments in which businesses co-evolve and collaborate, permitting enterprises to<br />
adapt quickly to a changeable environment and to save resources.<br />
The next section describes DBEs and how SMEs can benefit from the opportunities that this new paradigm<br />
offers.<br />
3.1 Digital business ecosystems<br />
Although the definition of a DBE may differ considerably depending on the author, in this work it refers to<br />
those digital environments that reproduce natural ecosystems and in which businesses can interact in a<br />
similar way to what people are doing in social networks.<br />
DBEs permit the interoperability of firms thorough the implementation of intelligent mechanisms of publishing<br />
characteristics, services and needs. In this way, SMEs are able to find other partners and to establish<br />
potential relationships with other entities of the ecosystem in a similar way individuals are doing in<br />
the applications presented in Section 2.1.<br />
Moreover, they are an innovative approach to supporting the adoption and development of ICT, which, as<br />
identified in (Nachira 2004), is a crucial element for improving productivity and for enabling business networking<br />
in a competitive knowledge-based economy.<br />
Like other Web 2.0 tools, DBEs enable the self-organisation of their participants that in the case of SMEs,<br />
can lead to improved business networking, greater competitiveness and a better exploitation of new opportunities.<br />
In these environments, businesses are represented by software components that exchange information,<br />
create coalitions and find new opportunities to compete and improve their individual position within the<br />
digital ecosystem but also for the benefit of the community they belong to. In this scenario, SMEs live,<br />
interact, evolve and become extinct as natural species do in natural ecosystems, and working together,<br />
create collaborative behaviour that leads to an improvement of the competitiveness of the whole set of<br />
cooperating companies.<br />
In order to be considered a digital ecosystem, the collection of autonomous digital entities must present<br />
the following characteristics:<br />
Complexity. According to (Lewin 2000), a complex system is one whose properties are not fully explained<br />
by an understanding of its parts.<br />
Self-organisation. Self-organisation is defined as the ability of complex systems to create new order<br />
and coherence. It relies on the spontaneous emergence of order by the interaction of individuals in an<br />
ecosystem.<br />
Emergence. Emergence is a concept closely linked to self-organisation and evolution and it is defined<br />
as the process which creates new order together with self-organisation.<br />
189
Francesc Estanyol<br />
Co-evolution. Defined in (Mitleton-Kelly 2003) as the evolution of one domain or entity is partially<br />
dependent on the evolution of other related domains or entities, it means that one domain or entity<br />
changes in the context of the others.<br />
Adaptation. Adaptation is an old concept from Darwin’s “Origin of Species” and it refers to the capacity<br />
of individuals to adapt to a changing environment.<br />
In addition to these principles, a DBE must be loosely coupled, meaning that each participant does not<br />
need to expose the details of its internal work-flows and requires only that it presents a service interface.<br />
Although DBE is a relatively new concept, recent projects like (OPAALS 2007) or (ONE 2006) have<br />
shown its potential to promote collaboration. These initiatives focused on providing organisations in the<br />
service industry with more sophisticated mechanisms to concentrate and publish services, improve negotiation<br />
and facilitate learning.<br />
Nonetheless, most of these tools use the Service Oriented Architecture (SOA) paradigm that, although it<br />
has proved to be successful in works like (Briscoe and De Wilde 2006), do not permit to implement all the<br />
characteristics of DBEs, limiting their use and potential. In that sense, Multi-Agent Systems (MAS)<br />
emerge as a perfect alternative because fulfil all DBEs principles. Moreover, MAS have been applied<br />
successfully in a large number of disciplines, from transport systems like (Balbo and Pinson 2005) to<br />
medical automatic diagnosis systems like HealthAgents (Lluch-Ariet et al. 2008).<br />
By definition, MAS facilitate the implementation of distributed architectures in which each entity is represented<br />
by an autonomous intelligent agent which pursues its goals and executes its tasks trying to improve<br />
its performance considering the conditions of the environment and interacting with other agents.<br />
Despite the parallelism between both paradigms seems clear, it is difficult to find MAS approaches to<br />
DBE and currently there is no application focused on SMEs. The work presented in (Muntaner-Perich and<br />
De la Rosa Esteva 2007) and the EveSim simulation tool resulting from the OPAALS project are the only<br />
initiatives to have succeeded in the creation of a multi-agent tool for DBEs but they failed on implementing<br />
the concepts of natural ecosystems.<br />
4. MADBE, a multi-agent digital business ecosystem for SME collaboration<br />
In moving beyond the current approaches and understanding digital ecosystems from an artificial intelligence<br />
perspective, MADBE represents the first ecologically inspired multi-agent digital business ecosystem<br />
for SME collaboration. In MADBE, the digital entities (SMEs) are represented by agents with their<br />
characteristics, services and needs. Therefore, the interactions with other agents are the representation<br />
of business-business relationships.<br />
Apart of the new technological approach, the step forward from similar initiatives is the introduction of<br />
mutualistic ecological network concepts to define the business interactions. Digital entities are defined by<br />
rules found in natural ecosystems in order to reproduce the characteristics of mutualistic networks, where<br />
species interact in a co-operative way and benefiting each other. Additionally, as summarised in (Bastolla<br />
et al. 2009), this type of network minimises competition and promotes species diversity and stability to the<br />
network itself, resulting in an environment with a larger number of SMEs collaborating with each other<br />
and establishing long-term mutual benefit relationships.<br />
Following these principles, the agents in MADBE are defined based on the set of ecological concepts<br />
presented in next section. This set of features are a starting point for designing the particular characteristics<br />
of each of the entities willing to join the digital ecosystem and create the necessity, covered by this<br />
study, to translate them into business concepts defining SME collaboration. In order to define the business<br />
concepts to include in MADBE regarding SME collaboration, the author used the findings of (Estanyol<br />
2010), a study which summarises why SMEs should co-operate, the problems they face adopting<br />
collaborative approaches and the factors influencing interfirm collaboration effectiveness. Therefore, the<br />
mapping process presented in next section is based on the results of that work, which permitted to understand<br />
SME collaboration and to select only the relevant business concepts, discarding ecological concepts<br />
with no applicability in the business domain and vice versa.<br />
190
Francesc Estanyol<br />
4.1 Mapping ecological concepts to the domain of SME collaboration<br />
The combination of natural ecosystems and the characteristics of real SME collaboration result in the<br />
definition of the MADBE agents by the following ecological concepts:<br />
Traits: a set of characteristics that define any entity within the system. In terms of SMEs, these traits are<br />
related to the characteristics of the firm and its expertise/interest in collaboration. The following items<br />
were selected in order to find complementarities and compatibilities between SMEs and facilitate partner<br />
search and selection. They cover most of the factors identified to do with collaboration success and to<br />
overcome the barriers to collaboration identified in (Estanyol 2010).<br />
Number of employees.<br />
Age of the company.<br />
Type of SME (individual enterprise, SL, SA, association, other).<br />
Goal. A short description of the firm’s global vision and main objectives.<br />
Average employee education level (no studies, undergraduate, graduate, postgraduate).<br />
Average age of employees (less than 25, between 25 and 35, between 35 and 50, more than 50).<br />
IT skilled personnel (yes/no).<br />
Technological capability (none, low, medium, high).<br />
Management style (owners, owners with management studies, professionals).<br />
Interest in co-operation (no interest, low, medium, high).<br />
Previous collaborations (yes, number of collaborations/no).<br />
Previous collaboration with the candidate (yes/no).<br />
Currently involved in co-operation projects (yes, number of projects/no).<br />
Level of satisfaction with previous collaborations (unsatisfied, low, medium, high, very high).<br />
List of services and resources (core competences). An array of a dynamic list of services and resources.<br />
The company has the option of adding non-existing items, making them available to other<br />
firms and increasing the list of system known services and resources.<br />
List of needs and resources. The same concept of the previous traits but about the needs of the<br />
firms.<br />
Degree of complementarity: the degree to which any trait of a given info-habitant is complementary to<br />
another trait possessed by another info-habitant. In MADBE, for each service, a degree of complementarity<br />
is given. It is normally a value between 1 and 5. For example, if a SME is looking for local transport, it<br />
could specify a maximum distance from its location. In this case, the degree of complementarity depends<br />
on the distance from the candidate to this location, the closest being the maximum. Although this ranking<br />
method has been adopted as a standard approach to define the degree of complementarity, the possibility<br />
of specifying different types of services and resources implies to treat each case individually and to<br />
create a system which permits the users to define how they want to specify the complementarity of their<br />
core competences.<br />
Habitat and Meta-communities: habitat refers to the place where an organism most commonly occurs.<br />
Basic market information like trend or the type of industry items is included in this section. Additionally,<br />
MADBE also needs to include those characteristics of the environment which influence the activity of the<br />
SME. In that sense, as it has been identified that SMEs look for external help and public funding when it<br />
comes to collaboration, MADBE includes a couple of lists indicating the presence of collaboration agents<br />
and funding organisations in the environment. In addition, SMEs tend to collaborate with firms from the<br />
same area due their preference for face-to-face contact, which required the inclusion of the firm location.<br />
On the other hand, meta-communities are aggregations of entities that belong to different habitats (or<br />
regions) and that occasionally interact when certain conditions are met. This concept is important for<br />
those situations when SMEs are looking to collaborate within the framework of a bigger association, not<br />
individually.<br />
Type of industry (monopoly, oligopoly, perfect competition).<br />
Market trend (unknown, downward, stable, upward).<br />
191
Francesc Estanyol<br />
Government regulations (low, medium, strict, very strict).<br />
List of collaboration agents. Information about agents which can help and provide support to SMEs<br />
during the collaboration process.<br />
List of public/private funding organisations. List of organisations promoting and funding collaborative<br />
initiatives.<br />
SME Location.<br />
A list containing the associations which the SME is a member (cluster, science park, etc).<br />
Niche: the niche is the “job” of the organism within the habitat. This concept groups those characteristics<br />
about the interactions and role of the agent in the ecosystem. The niche concept is also useful in explaining<br />
how seemingly similar species can coexist in the same community. In MADBE, the niche is the multidimensional<br />
space composed by the identified characteristics of the habitat and the way agents interact<br />
with their environment. Therefore, it is necessary to include the business context of the company and its<br />
sector. Additionally, SMEs within small areas tend to have similar world-views and to speak similar languages,<br />
which have been identified as key factors for collaboration success. It is for those reasons that,<br />
apart from location and culture, it is important to include the languages known by the company. Finally,<br />
the niche concept also explains how species from a specific habitat collaborate temporarily with species<br />
from a different environment, which is why it has been included a parameter to determine the willingness<br />
of the company to create new businesses in different business contexts.<br />
Business context (local, national, <strong>European</strong>, international).<br />
Business sector (research, innovation, product development, marketing, etc.)<br />
Interest in creating new businesses in a different business domain (low, moderate, high, very high).<br />
Languages. List of languages known by the company.<br />
Size of the SME in relation to the market size (very small, small, medium, big, very big).<br />
Employee mobility (yes/no). Parameter to detect the capacity of the company to send employees to<br />
other companies for learning and transfer of knowledge purposes.<br />
Fitness: a measure that is used to determine how well an entity is doing during its lifetime in the digital<br />
ecosystem. An intelligent tool like MADBE needs fitness measures in order to evaluate the different options<br />
during simulations and to provide the optimum results to the user. Unfortunately, there is no consensus<br />
about a specific set of measures to determine the fitness of SME collaboration. Some authors<br />
defend objective measures like the duration of the co-operation or the number of sales while others prefer<br />
to consider the subjective perception of their participants. MADBE has two set of parameters in order to<br />
implement both. A first group including basic accounting concepts and useful to evaluate the fitness of<br />
the SMEs within the ecosystem and a second group including those parameters to evaluate the collaboration<br />
process itself.<br />
Individual SME performance:<br />
Number of current collaborations. Useful to identify the role and the influence of the company in the<br />
ecosystem.<br />
Average of previous collaborations duration.<br />
5 year annual turnover (i.e. sales).<br />
5 year proportion of annual turnover of new products/services.<br />
5 profit margin = pre tax net profit/turnover.<br />
5 year RoA (return on assets) ratio.<br />
5 year Liquid Ratio = (current assets - inventories) / current liabilities.<br />
5 years percentage increase of number of employees.<br />
Performance of the collaboration:<br />
Reputation rate. An array of evaluations including the opinion about the SME from partners with<br />
whom it has collaborated. A reputation mechanism to overcome the trust, credibility and compromise<br />
problems and to evaluate the co-operation skills of the company. This measure permits to identify reliable<br />
partners and discard partners that break collaborations or fail to carry out their assigned tasks.<br />
Duration of the co-operation.<br />
192
Francesc Estanyol<br />
Number of new products/services. The proportion of current sales from products newly introduced<br />
during the co-operation.<br />
Subjective perception (bad, neutral, positive, very positive).<br />
Number of new jobs created.<br />
Access to new markets (yes/no).<br />
New relationships (yes/no).<br />
The concepts of traits, niche and fitness have been employed for the design of the agents interacting and<br />
living together within MADBE, and are therefore responsible for much of the behaviour displayed by them<br />
in any given interaction. More details about the ecological concepts and their role specifying rules of interactions<br />
and specific details of MADBE architecture can be found in (Lurgi and Estanyol 2010).<br />
5. Conclusions<br />
On the basis of the findings from the literature review, the author explored the current context of SME<br />
collaboration and how Web 2.0 applications are changing the way individuals interact. By using current<br />
examples, it has been demonstrated that nowadays users are able to self-organise themselves creating<br />
communities in order to satisfy their needs independently of the traditional and external channels. These<br />
communities facilitate the emergence of collaborative behaviour among the participants but businesses,<br />
and in particular SMEs, are not taking profit of it.<br />
In order to understand why, the author analysed the type of existing ICT tools, concluding that most applications<br />
are not specifically designed considering SMEs needs and they focus on how improve their<br />
current business practices instead of looking for mechanisms to create new ways of doing businesses.<br />
In that context, the new paradigm of DBEs has been introduced as a possible solution to overcome these<br />
problems and, after studying the existing alternatives, the work concluded with the presentation of<br />
MADBE, the main contribution to research of this study and the first ecologically inspired multi-agent<br />
DBE.<br />
The author believes that DBEs can play an important role enabling and improving SMEs co-operation. In<br />
line with Nachira, who in (Nachira 2002) proclaims that “DBEs produce an extraordinary competitive advantage<br />
for SMEs”, the author thinks that DBEs designed and developed considering SMEs needs could<br />
benefit them in the long-term and enhance their chances of competing with larger organisations.<br />
In addition, the adoption of multi-agent technology in MADBE and the mapping between business and<br />
ecological concepts found in natural mutualistic networks permits to define an application fulfilling SME<br />
needs and truly inspired in ecology. Moreover, by treating SMEs as autonomous intelligent agents, the<br />
system requires a minimum level of dedication and permits to analyse the individual behaviour of each<br />
company in a collaborative environment.<br />
At this point, it is important to remark that this work only covered the definition of the ecological concepts<br />
from a business perspective, which is a fundamental step to demonstrate that the concepts explaining<br />
natural mutualistic networks of interaction could be translated into the business domain and report benefits<br />
for SMEs but the system is still under development and therefore, it has not been deployed to a real<br />
market scenario and there are no results yet. It will take time to develop the system and convince SMEs<br />
to use it in order to collect real data. In that sense, the current participation of the author in (EcoBusiness<br />
2010) ensures the continuity of this research.<br />
Another aspect to consider when evaluating DBEs as enablers for collaboration is that the utilisation of<br />
any Web 2.0 tool is not enough to ensure collaborative success. In order to become part of Web 2.0 it is<br />
necessary a change of mentality. E-business is more than e-commerce and it is not enough to use Web<br />
2.0 tools for collaboration without a planned strategy. It is obvious that companies cannot take the same<br />
risks as individuals and that their bureaucracy and risk aversion make sudden changes difficult, but it is<br />
also true that Web 2.0 opens up a wide range of business opportunities for collaboration and cooperation<br />
that companies should consider. In order to succeed, it is necessary that firms interested in<br />
collaboration have a clear understanding of why they want to co-operate and which results they expect<br />
before the process starts. Collaboration requires a favourable attitude and an analysis of the individual<br />
characteristics of the participants in order to find complementarities and synergies while avoiding the<br />
emergence of conflict.<br />
193
Francesc Estanyol<br />
Finally, a positive consequence of DBEs is that business environments are represented as complex systems,<br />
permitting to analyse and understand the principles of their formation, evolution and interrelationships<br />
and exploit the research carried out in other sciences like biology, mathematics, computer science<br />
or social science in the business domain. For example, by applying network analysis algorithms in a system<br />
like MADBE would be possible to identify the key firms of the ecosystem, the “leadership companies”<br />
or the ones who have a strong influence on the others. In that sense, MADBE will not benefit only the<br />
involved SMEs but also third parties like public governments, who, for example, could use it to identify<br />
which companies should receive more funding or those reaching undesired monopolistic positions.<br />
Acknowledgments<br />
The author wants to thank Miguel Lurgi for the specification of MADBE from an ecological perspective<br />
and for the identification of the key ecological concepts to be included in the system. This work is part of<br />
the EcoBusiness project, funded by the <strong>European</strong> Commission through the Seventh Frame work programme<br />
Marie Curie Actions – Industry-Academia Partnerships and Pathways (IAPP). Grant agreement<br />
no.: 230618.<br />
References<br />
Audretsch, D, van der Horst, R, Kwaak, T, and Thurik, R.. (2009) “First section of the annual report on EU small and<br />
medium-sized enterprises”, January.<br />
Axelrod, R. (1985) “The importance of being nice, retaliatory, forgiving and clear”, The Economist, pp 99–102, November,<br />
9.<br />
Balbo, F and Pinson, S. (2005) “Dynamic modelling of a disturbance in a multi-agent system for traffic regulation”,<br />
Decision Support Systems, Elsevier, Vol 41, No. 1, pp131–146.<br />
Bastolla, U, Fortuna, M, Pascual-García, A, Ferrera, A, Luque, B, and Bascompte, J. (2009) “The architecture of<br />
mutualistic networks minimizes competition and increases biodiversity”, Nature, Nature Publishing Group, Vol<br />
458, No. 7241, pp 1018–1020.<br />
Briscoe, G and De Wilde, P. (2006) “Digital ecosystems: Evolving service-orientated architectures”, Bio-Inspired<br />
Models of Network Information and Computing Systems. Vol 1, pp 1–6.<br />
Caby-Guillet, L, Crave, S, and Ladame, S. (2006) “How ICT can facilitate trust inside networks of SME: the role of<br />
professional virtual communities”, International conference on Concurrent Engineering, Milano<br />
Coombs, R, Richards, A, Saviotti, P, and Walsh, V. (1996) Technological Collaboration: The dynamics of cooperation<br />
in industrial innovation, Edward Elgar Pub, ISBN: 1858982359.<br />
Dickson, K. (1996) “How informal can you be? trust and reciprocity within co-operative and collaborative relationships”,<br />
International Journal of Technology Management, 11, Inderscience Publishers, Vol 1, No. 2, pp 129–<br />
139.<br />
EcoBusiness Website (2010) URL http://www.ecobusiness.cat, Accessed date 19/11/2010.<br />
Estanyol, F. (2010) “The SME co-operation framework: A multi-method secondary research approach to SME collaboration”,<br />
Proceedings of ICEME 2010, Hong Kong.<br />
EU. (2009) “On the implementation of commission recommendation of 6 may 2003 concerning the definition of micro,<br />
small and medium-sized enterprises”, Commission Staff Document, October.<br />
Hoegg, R, Meckel, M, Stanoevska-Slabeva, K, and Martignoni, R. (2006) “Overview of business models for web 2.0<br />
communities”, proceedings of GeNeMe, pp 23–37.<br />
Hoffmann, W, H and Schlosser, R. (2001) “Success factors of strategic alliances in small and medium-sized enterprises<br />
- An empirical survey”, Long Range Planning, Elsevier, Vol 34, No. 3, pp 357–381.<br />
Lewin, R. (2000) “Complexity: Life at the edge of chaos”. University of Chicago Press.<br />
Lluch-Ariet, M, Estanyol, F, Mier, M, Delgado, C, González-Vélez, H, Dalmas, T, Robles, M, Sáez, C, Vicente, J and<br />
Huffe, Sl. (2008) “On the implementation of healthagents: Agent-based brain tumour diagnosis”, Agent Technology<br />
and e-Health, pp 5–24.<br />
Martin L. M., and Matlay, H. (2003) “Innovative use of the internet in established small firms: the impact of knowledge<br />
management and organisational learning in accessing new opportunities”, Qualitative Market Research: An International<br />
Journal, MCB UP Ltd, Vol. 6, No. 1, pp 18 – 26.<br />
Mitleton-Kelly, E. (2003) “Ten principles of complexity and enabling infrastructures”, Complex systems and evolutionary<br />
perspectives on organisations: The application of complexity theory to organisations, Pergamon, pp 23–50.<br />
Muntaner-Perich, E and de la Rosa Esteva, J. (2007) “Using dynamic electronic institutions to enable digital business<br />
ecosystems”, Coordination, Organizations, Institutions, and Norms in Agent Systems II, Springer, pp 259–273.<br />
Nachira, F. (2002) “Towards a network of digital business ecosystems fostering the local development”. URI<br />
http://hdl.handle.net/2038/529<br />
Narula, R. (2004) “R&D collaboration by SMEs: New opportunities and limitations in the face of globalisation”, Technovation,<br />
Elsevier, Vol 24, No. 2, pp 153-161.<br />
Nooteboom, B. (1994) “Innovation and diffusion in small firms: theory and evidence”, Small Business Economics,<br />
Springer, Vol 6, No. 5, pp 327–347.<br />
O’Reily, T. (2005) “Web 2.0: Compact definition?” Blog, October, URL http://radar.<br />
oreilly.com/archives/2005/10/web-20-compact-definition.html, accessed date 16/06/2010.<br />
ONE. (2006) One negotiation environment project. Website, URL http://one-project.eu/<br />
site/modules/content/index.php?id=1, accessed date 08/08/2010.<br />
194
Francesc Estanyol<br />
OPALS. (2007) Open philosophies for associative autopoietic digital ecosystems. Website, URL http:<br />
//www.opaals.eu/, accessed date 15/08/2010.<br />
Shirky., C (2009) “How social media can make history”, <strong>Conference</strong> Talk, TED@State, June, URL<br />
http://www.ted.com/talks/clay_shirky_how_cellphones_twitter_facebook_can_make_history.html, Accessed<br />
date 03/07/2010.<br />
Varamaki, Eand Vesalainen, J. (2003) “Modelling different types of multilateral co-operation between SMEs”, Entrepreneurship<br />
& Regional Development, 15:2747.<br />
Zeng, S, Xie, X, and Tam, C. (2010) “Relationship between cooperation networks and innovation performance of<br />
SMEs”, Technovation, Elsevier, Vol 30, No. 3, pp 181 – 194.<br />
195
Ideas About Profitability in Research and Development and<br />
the Selective Pressure From Management Accounting<br />
Albrecht Fritzsche<br />
Technical University, Darmstadt, Germany<br />
a.m.fritzsche@gmx.net<br />
Abstract: The selective pressure on the development of products and production technologies is not exerted directly<br />
by customer demand, but by its interpretation in a company. The decisions on technical development usually take place<br />
a long time before a product reaches the market on the basis of calculations about profitability in management accounting.<br />
The profitability can depend on various factors, including production facilities, supply networks, intellectual<br />
property, distribution channels and competition. The calculations that lead to decisions on an investment tend to be<br />
very complex. Nevertheless, they do not represent fully realistic market conditions, but only a simplification for the sake<br />
of reliable estimates. In a conventional simple evolutionary model of innovation, technical development does not express<br />
expert knowledge. In addition, it adapts as much to the shortcomings of the financial calculations as to the<br />
prognostic information they are based on. If the selective pressure from management accounting is high, technical<br />
development is likely to focus on the exploitation of the missing cost factors in the calculations and neglect the reality of<br />
market demand. In practice, innovations in research and development are not propelled by arbitrary change. Expanding<br />
the conventional model, the expertise in research and development can be represented by introducing intentional<br />
improvement operators. These operators show a potential to avoid unwanted directions of adaptation, which<br />
gives reason to believe that common sense in research and development play an important role in avoiding detrimental<br />
effects of incomplete cost calculations.<br />
Keywords: computational models of innovation, investment decisions on R&D, technology assessment and management<br />
accounting, simulation of evolutionary adaptation, profitability calculations and human expertise<br />
1. Background<br />
During the last decades, it has become very popular to use Genetic Algorithms for the solution of complex<br />
computational problems. In contrast to conventional optimization techniques, such as gradient-based or<br />
simplex-based methods, Genetic Algorithms require only a minimum of analytic representation of the<br />
problem structure. They can therefore also be applied to problems with a very high level of complexity.<br />
Genetic Algorithms represent a heuristic approach to problem solving. There is no guarantee that they<br />
produce optimal results. Nevertheless, they have proven to be very useful in a large variety of different<br />
scenarios, from factory scheduling over jet engine design to financial portfolio management (see Goldberg<br />
2002).<br />
Thanks to their success as a search method, Genetic Algorithms have become the object of extensive<br />
scientific research. The concept of Genetic Algorithms is based on the principle of evolutionary adaptation.<br />
Evolutionary approaches are quite common in many different disciplines, for example in economics, sociology<br />
and philosophy (Dennett 1996, Broesel et al. 2007). In economy, evolutionary models are used, for<br />
example, to describe the dynamics of technological change. While analytic approaches focus on the<br />
identification of optimal economic transactions and the way how they can be achieved, evolutionary approaches<br />
put the emphasis on the explanation of innovative processes in complex environments where<br />
optimal, equilibristic states are unknown (Dopfer 2008).<br />
2. Evolutionary approaches to technical development<br />
2.1 Technology, economics and evolution<br />
Evolutionary approaches in economy are strongly related to the idea of a dynamic relationship between<br />
supply and demand on a free market. Consumption and production are interdependent: Consumers react<br />
to price changes and the introduction of new products; producers offer goods up to the extent to which<br />
selling them is profitable. Technical development, in terms of new products as well as new production<br />
methods, is an adaptive process that can be compared to natural evolution, which, according to Darwin,<br />
advances through repetitive arbitrary changes under selective pressure from outside (see e.g. Hodgson<br />
1993). Finding exact economic correspondences for the reproductive processes in nature, the genetic<br />
carriers of information and the genotype of adaptation is rather difficult. The general principle of evolution<br />
as an adaptive activity without intentional direction, however, applies very well to the dynamics business on<br />
an unrestricted market (see e.g. Broesel t al. 2007, Beck-Sickinger/ Petzold 2009).<br />
196
Albrecht Fritzsche<br />
The idea to describe the development of trade and industry similar to the biological models of natural<br />
adaptation and selection dates back to the late nineteenth century (Marshall 1890, Veblen 1898). In the<br />
twentieth century, Joseph Schumpeter has been very influential with his evolutionary approach to innovation<br />
(Schumpeter 1912). Schumpeter describes innovation as a source of continuous change, driven by the<br />
search for competitive advantage of the companies on a market. After several decades in which economic<br />
research showed little interest in evolutionary models, the approach regained increasing popularity during<br />
the 1980ies (Hodgson 2005), mainly due to the seminal paper by Richard Nelson and Sidney Winter<br />
(Nelson/ Winter 1982).<br />
Evolutionary economics is a relatively open field of research, which has been influenced by various scientific<br />
concepts, such as bounded rationality or complexity studies (Simon 2005, Frenken 2006). It is concerned<br />
with both microscopic and macroscopic phenomena of economic adaptation and their dependency<br />
on socio-cultural or administrative structures. It is characteristic for evolutionary economics to assume that<br />
the economic development cannot be guided and controlled (comp. Dopfer 2008, Winter 2005). Nevertheless,<br />
it is possible to influence the process of their adoption on a market, which depends on their representation,<br />
distribution and the selective pressure exerted on them.<br />
2.2 Simulated evolution in Genetic Algorithms<br />
An important source of influence on the studies of evolutionary processes in general has been the development<br />
of electronic data processing and the rise of modern information technology. Computers can<br />
execute a large number of logical operations in short time; they are therefore a very useful tool to simulate<br />
adaptive phenomena that are caused by the combination of many simple change routines. It also turned<br />
out very quickly that it was possible to apply these simulations of evolutionary developments and similar<br />
procedures in solving optimization problems in complex search spaces. For this purpose, the core elements<br />
of the evolutionary process were formalised in the concept of what is now known as Genetic Algorithms.<br />
Figure 1 shows a rough sketch of the design of Genetic Algorithms (see e.g. Goldberg 1998). The search<br />
starts with a set of possible alternatives to solve the problem, called a population. In course of the search,<br />
the alternatives in the population repetitively go through some changes, usually caused by mutations or<br />
re-combinations of the alternatives.<br />
Figure 1: Concept of evolutionary search in Genetic Algorithms<br />
197
Albrecht Fritzsche<br />
As the genetic metaphor indicates, these operations are inspired by processes in biological reproduction.<br />
With a mutation, certain characteristics of the alternatives are modified at random; a re-combination puts<br />
together arbitrary characteristics from two (parent-) alternatives in order to create a new one. The operations<br />
have no knowledge about the quality of the alternatives. The evaluation of the changes in the population<br />
happens separately in a subsequent step of the procedure, where the best alternatives are selected<br />
to form a new population. Since these changes are performed on alternatives which already possess<br />
positive characteristics, it can be expected that the population will step by step develop into set of optimal<br />
alternatives, similar to the survival of the fittest in nature.<br />
2.3 Imitation and innovation<br />
Genetic Algorithms are just one heuristic search method among many others, but they play an outstanding<br />
role as a representation of evolutionary processes. In fact, Genetic Algorithms have widely replaced the<br />
biological model of natural development as the point of reference for the concept of evolution, thanks to the<br />
simplifications they provide and their implementation on computers. It is much easier to understand the<br />
principle of evolution on the example of a Genetic Algorithm than by looking at the development of species<br />
in nature, which retains importance as the origin and the symbol for the general importance of evolutionary<br />
processes.<br />
In particular, Genetic Algorithms have been used to research how technical change becomes possible.<br />
According to David Goldberg, the basic reproduction methods of Genetic Algorithms, the mutation and the<br />
re-combination, in combination with the selective pressure from outside, represent the two fundamental<br />
activities in research and development (Goldberg 1998, Goldberg 2002). The combination of mutation and<br />
selection describes continuous improvement processes as they are addressed, for example, by the kaizen<br />
principle. This improvement leads to a higher efficiency of the processes that are already implemented. In<br />
that sense, Goldberg understands it as imitation. The combination of selection and re-combination, on the<br />
other hand, leads to improvement by a new arrangement of the single elements that define a technical<br />
solution. The new arrangement changes the effect of technology rather than its efficiency. Inasmuch as<br />
technology is defined by its effect, the result of the change can be considered as an innovation.<br />
Topologically, mutations and re-combinations describe two different movements in the search space.<br />
Mutations cause arbitrary small changes. Most of the original element of the search space remains intact.<br />
Re-combinations, on the other hand, bring the characteristics of two different elements together. If such a<br />
change had to be generated solely by mutations, it would require a huge number of steps. Considering the<br />
selective pressure under which the changes are taking place, it is highly improbable that mutations lead to<br />
similar results as re-combinations. At the same time, re-combinations are not very effective in generating<br />
small changes of one original alternative.<br />
In general, it is hardly possible to identify the best elements of the search space, if mutations or<br />
re-combinations are used exclusively (see Goldberg 2002). Both of them have to be used together. For<br />
some problems, it may be necessary to apply re-combinations more frequently than mutations; for other<br />
problems, the opposite may be true. In addition, the size of the population is also an important parameter<br />
for the success of Genetic Algorithms as optimization tools. Other parameters have also been researched<br />
(Reed et al. 2001). Genetic Algorithms therefore are no "free lunch" (Wolpert/ MacReady 1997); they do not<br />
provide a generally applicable solution technique. Instead, they have to be adapted to the given problem<br />
situation. In Goldberg's model, this can be understood as the need for innovation management.<br />
3. The influence of accounting on technical development<br />
3.1 The paradox of heuristics and analytics<br />
The strength of the evolutionary approach in the different fields of research is mainly due to the fact that it<br />
shows how improvement is possible without knowledge of the reasons for preference (see e.g.<br />
Beck-Sickinger/ Petzold 2009). It does not matter why certain alternatives are considered to be better than<br />
others. It is only necessary to know that they are better in order to initiate the selection procedure. Apart<br />
from the current set of alternatives and the operators to identify new alternatives, the evolutionary process<br />
does not carry any further information. Economy, on the other hand, is the expression of a systematic activity<br />
of understanding and enabling improvement. The idea of stable economic structures that do not<br />
evolve themselves over time is contradictory to the evolutionary approach. Economic structures can be<br />
considered as expressions of expertise in improvement, both in a deductive and an inductive way. The<br />
notion of quantifiable goods and their prices represents, for example, certain ideas about what can be<br />
198
Albrecht Fritzsche<br />
improved; the notion of trade and production represents certain ideas about the way how this improvement<br />
takes place.<br />
The strength of the evolutionary approach lies in the fact that it does not require any expertise to explain<br />
improvement. Quite in the contrary, it shows how arbitrary change can lead to improvement under selective<br />
pressure. Research in evolutionary economics is therefore confronted with one fundamental question: How<br />
do evolutionary adaptation and economic expertise connect? The following pages give an example of such<br />
a connection. They also show that research on Genetic Algorithms and similar search methods can help to<br />
identify various different levels on which evolutionary adaptation and economic expertise come together.<br />
Our study explores the relation between departments for management accounting and for research &<br />
development. Both departments have a strong influence on technical change, based on certain systematic<br />
activities to enable and assess improvement. At first sight, it might seem that the installation of these departments<br />
contradicts the understanding of technical development as an evolutionary process. This paper<br />
shows, however, that there is good reason to approach the interaction of these departments in evolutionary<br />
terms. In particular, the paper uncovers certain parallels between technical development as an evolutionary<br />
process and recent experiences with the application of Genetic Algorithms for optimization in complex<br />
systems.<br />
3.2 The split between entrepreneur and engineer<br />
The high frequency in which companies introduce new products to a market makes it easy to forget that<br />
technical innovation is a slow process. Cars, computers, cell phones or medicine require a long period of<br />
preparation, both for research and development of the products themselves and for the establishment of<br />
production facilities, skills and supply networks. The investments during this phase are considerable. The<br />
most important decisions about technology therefore take place a long time before it enters a market. At<br />
this time, however, there are no sales figures yet that could exert a selective pressure. Instead, the decisions<br />
are made on the basis of estimates about the expected profitability as marketing and accounting<br />
departments are able to supply them (see e.g. Trott/ Lataste 2005).<br />
For the sake of a brief illustration, the difference between the reference to the market and the reference to<br />
investment calculations in accounting can be, very roughly, described in the following way: It starts with<br />
naming the dimensions in which the profitability of the innovation will be measured. This includes, for<br />
example, the raw material and parts supply, labour, energy, maintenance and services as well as the<br />
one-time effort to implement the production facilities. In these dimensions, the innovation can be expressed<br />
as a transformation vector that shows the ratio of the quantity of the goods consumed in production to the<br />
quantity of the good produced. Marketing research then provides estimates for customer demand and the<br />
prices of the different goods. Knowing the implementation effort, the transformation vector of production<br />
and the prices, it becomes possible to calculate the expected profitability of the innovation. If the evaluation<br />
of the innovation referred directly to the market, it would not be necessary to establish these dimensions.<br />
The profit could simply be measured by looking at the bank account.<br />
As a consequence of replacing the market response as the selective pressure from outside by investment<br />
calculations, the process of technical development is split in two. In the first part of the process, market<br />
response in approximated in the dimensions of the calculation and the pricing vector. This approximation is<br />
subject to constant change, because it is impossible to consider all sources of influence on the market. It<br />
can therefore be understood as an evolutionary process by itself. Since an entrepreneur is essentially<br />
somebody who discovers new ways to make profit, the approximation of market response for investment<br />
calculations reflects the entrepreneurial approach to innovation. In the second part of the process, research<br />
and development activities change technology under the selective pressure from accounting. Inasmuch as<br />
accounting measures profitability in certain dimensions, this process resembles a multi-objective optimization<br />
problem, with slowly changing objectives over time (see Table 1).<br />
3.3 The complexity problem of profitability calculations<br />
Considering the multiple dimensions of profitability, it is clear that the improvement of a product or a production<br />
process with respect to material costs, product utility or any other single objective does not necessarily<br />
lead to an overall better solution. Quite in the contrary, an improvement with respect to one objective<br />
is most likely to have detrimental effects with respect to other objectives. The investment calculations<br />
in accounting can display these effects, but it is quite a different question whether it is also possible to<br />
predict them. Some directions of research and development may sound very promising in advance, but<br />
199
Albrecht Fritzsche<br />
later turn out to be to expensive, considering all factors involved. This phenomenon is, for example, widely<br />
known in supply chain management. It can be described as "over-complexity" (Klaus 2005).<br />
Table 1: Appearance of objectives in technical development<br />
Sequence Objective Activity<br />
1. Maximize product utility. Increase efficiency and effectiveness of the product for the<br />
customer.<br />
2. Minimize raw material costs. Reduce material consumption, waste, and expensive substances.<br />
3. Minimize labour costs. Reduce staff, reduce wages, and optimize external consulting<br />
effort.<br />
4. Minimize tax payments. Optimize distribution of costs, use government support measures.<br />
5. Optimize maintenance intervals.<br />
Adapt product to service infrastructure and competences.<br />
6. Minimize supply risks. Optimize size of supply network and tier co-operation.<br />
7. Optimize capital management. Adapt production to own capital, shareholder interest etc.<br />
8. Minimize novelty conflicts. Reduce efforts for product implementation and version changes.<br />
9. Optimize ecological impact. Reduce energy consumption, hazardous waste etc.<br />
10. Optimize intellectual property. Reduce risk of copies by competitors, dependency on outside<br />
patents, knowledge etc.<br />
Assuming such a scenario of research and development activities, technical change and the decision about<br />
its profitability must be understood as two separate procedures. If so, it makes sense to approach these<br />
activities in terms of evolution. The creation of solutions and their evaluation and selection for further use<br />
can be considered to interact as displayed by the boxes in Figure 1. It is important to note that the specific<br />
expertise in marketing and accounting and in research and development does not contradict the idea of<br />
evolutionary adaptation. In fact, the high level of sophistication of the work in the different departments<br />
must be seen as the very reason why it is necessary to use an evolutionary approach: it is impossible to<br />
combine all the expertise in both fields. The complexity of the task is so high that it requires a separation.<br />
The success of the technical development as an overall optimization depends on various parameters. In<br />
particular, the way how the change operators are applied has to be appropriate to the problem. Market<br />
demand in general is not determinate enough to reflect this issue. The dimensions of the investment calculations<br />
in accounting, however, provide very explicit information about the evaluation of the solution<br />
quality. In that sense, the role of management accounting for the selection of technical solutions is easily<br />
comparable to the role of a formal evaluation function in a computational multi-dimensional optimization<br />
problem. Recent research on the design of Genetic Algorithms can therefore convey important insight into<br />
the question of innovation management.<br />
4. Evolutionary models and engineering expertise<br />
4.1 Operators and landscapes<br />
Goldberg discusses the task of innovation management as the design of a competent Genetic Algorithm. In<br />
this context, competence means that the algorithm is able to identify good solutions. For Goldberg, this<br />
depends on the question whether the operators are able to reproduce the good characteristics of previous<br />
alternatives (Goldberg 1998). Topologically, the problem can be described as the question whether it is<br />
easy to find a path from any given alternative to the best ones. This problem is addressed by the notion of<br />
fitness landscapes.<br />
A fitness landscape is a triple (S, q, N) of a search space S, an evaluation function q and a neighbourhood<br />
structure N on that space. The neighbourhood structure is induced by the changes that an operator can<br />
cause for a given element of the search space, for example by all the mutations that can be performed on<br />
an alternative. For certain instances of the Travelling Salesman Problem, (Boese et al. 1994) have shown<br />
that the landscape induced by a mutation operator on an intuitively plausible metric leads to a structure<br />
shaped like a big valley on the search space. With respect to that landscape, local optima are situated close<br />
200
Albrecht Fritzsche<br />
to one another and a simple heuristic search algorithm can calculate best solutions quite effectively.<br />
(Reeves 1999) has extended this approach to a permutation flow shop scheduling problem using operators<br />
for mutations, shifts and inversions of parts of the sequence. With reference to these operators, the landscape<br />
on the search space provides easy paths to local global optima. (Knowles and Corne 2002) discuss<br />
different strategies of search on a global assignment problem, developing solutions intently either in the<br />
direction of better quality or somewhat "horizontally" to that direction once that good solutions have been<br />
found.<br />
In a complex multi-objective optimization problem, the direction of change that leads to better quality is<br />
generally unknown. However, the single objectives suggest certain directions in which intentional change<br />
can proceed. With respect to one dimension of the problem, these directions indicate the steepest ascent of<br />
solution quality. Under the assumption that the different objectives of the problem are not fully contradictory<br />
in their notion of good solutions, changes in these directions can be expected to have strong influence on<br />
the overall fitness, too. It therefore seems sensible to introduce operators that work explicitly on the improvement<br />
of solutions with respect to single objectives.<br />
Recent findings about the application of heuristic search methods to order scheduling problems indicate<br />
that such operators can indeed have a positive effect on the optimization procedure (Fritzsche 2009, see<br />
also Solnon et al 2008). Genetic Algorithms for order scheduling show better results, if they do not only<br />
execute arbitrary mutations and re-combinations but also constructive changes of the schedule with respect<br />
to the given objectives. For example, it is useful to apply operators which can move items in the<br />
sequences according to given constraints of minimal production distances or the due dates of the orders.<br />
With reference to such constructive operators, the notion of expertise can be introduced to the model of<br />
evolutionary adaptation.<br />
4.2 Expertise in evolutionary procedures<br />
Based on the study of fitness landscapes, the role of expertise in technical development can be explained<br />
in terms of the directions in which intentional changes are performed. Management accounting defines<br />
these directions by setting up the dimensions of the profitability model. The dimensions give orientation for<br />
the efforts to improve technical solutions. Table 2 gives an overview over the different activities that contribute<br />
to technical change with respect to the dimensions of profitability. These activities reflect the diversity<br />
of the work in research and development as well as some other departments that are closely connected<br />
to it. While Goldberg's model of technical change only involves two different operators, it seems more<br />
adequate to consider the contributions of product designers, project managers, physicists, technicians and<br />
engineers as many separate operators. All of them contribute in different ways and to a different extent to<br />
improvement with respect to the single objectives.<br />
Table 2: Expertise involved in technical development<br />
No. Objective Activity<br />
1. Maximize product utility. Product design and usability, marketing, social studies.<br />
2. Minimize raw material costs. Physics, material research.<br />
3. Minimize labour costs. Factory design, robotics, training departments.<br />
4. Minimize tax payments. Project management, infrastructure planning.<br />
5. Optimize maintenance intervals. Customer relationship management and service.<br />
6. Minimize supply risks. Modular design, cross-company research.<br />
7. Optimize capital management. Financial advice.<br />
8. Minimize novelty conflicts. Change management and engineering.<br />
9. Optimize ecological impact. Ecological auditing, systems engineering.<br />
10. Optimize intellectual property. Legal departments, R&D administration.<br />
The comparison of these activities with the constructive operators in Genetic Algorithms leads to an important<br />
conclusion about the way how these activities facilitate technical improvement. They do not have to<br />
be brought together in a systematic process of rational problem solving; technical improvement is also<br />
possible in a more liberal scenario of interaction. If all different experts are free to contribute to the solutions<br />
in whichever way they like, they can still evolve into something better, even if the single changes performed<br />
201
Albrecht Fritzsche<br />
on the solutions seem highly contradictory. Innovation management, however, has to make sure that this<br />
process takes place under a certain selective pressure of profitability. In addition, it is necessary to make<br />
sure that the changes of the solutions do not become too big, because this would inhibit the interaction<br />
between the operators. Small contributions to change, like the measures for continuous improvement in a<br />
factory, can therefore play a huge role for evolutionary adaptation.<br />
4.3 Expertise and the market<br />
Genetic Algorithms using constructive operators show how technical development can proceed as an<br />
evolutionary process in the presence of professional expertise. A fundamental problem of the whole approach,<br />
however, has so far remained unmentioned: If technical development is subject to the pressure<br />
exerted by management accounting, the solutions it creates can only be as good as the model of profitability<br />
that is used in the investment calculations. Since this model can never be fully adequate to the reality<br />
of market demand, it must be assumed that technical change always leads into a slightly wrong direction. It<br />
will not adapt to market itself, but to the dimensions in which the profitability is measured. Technical solutions<br />
will consequentially reflect all the strengths and weaknesses of management accounting. In particular,<br />
the development will be detrimental with respect to all aspects of market demand that are neglected<br />
in the investment calculations. If so, the development of technology can never fully meet the economic<br />
necessities. In fact, one would have to expect that it is often led far astray.<br />
The study of fitness landscapes, however, indicates that the evaluation function is not the only source of<br />
influence on the direction of an evolutionary process. The neighbourhood topology also plays an important<br />
role. The search algorithm can only identify solutions that are within reach of the operators applied under<br />
the circumstances set by the population size and the selection technique. In that sense, the operators do<br />
not only change the course of the search; they also set the boundaries of the areas in which the search<br />
takes place. If the search space is very big, it is quite important to reduce the search procedure to a suitable<br />
subset of the space in order to identify good solutions. If the operators perform very specific changes, the<br />
subset in which the search proceeds can be rather small. The application of such operators is justified, if<br />
they can be assumed to cover a subset of the search space that contains the elements with the highest<br />
quality. Operators that perform changes in the direction of the dimensions of the evaluation function can be<br />
expected to do so, if they are used interactively.<br />
Considering this, it becomes clear that the direction of evolutionary processes is not determined exclusively<br />
by the selective pressure under which they take place. The selection of the best elements can only cause<br />
an adaptation in the boundaries set by the operators that generate new alternatives. It therefore would be<br />
wrong to say that technical development is completely dominated by management accounting. Before it is<br />
possible to select the best technical solutions, they first have to be generated by research and development.<br />
Although the activities involved in research and development are related to certain dimensions of the<br />
profitability model in accounting, there is good reason to assume that they are also influenced by the<br />
common sense of the people involved. This influence becomes visible, for example, in the variations of<br />
technical development between different cultures. Although one can assume that the profitability calculations<br />
in competing companies are rather similar, the technical solutions that are generated in the various<br />
<strong>European</strong> countries, America or China are quite diverse. It lies at hand to attribute these variations to the<br />
fact that work in research and development is not solely guided by professional expertise in certain fields,<br />
but also by general assumptions about profitable technical solutions.<br />
5. Conclusion<br />
The purpose of this paper has been to show how different activities in a company can be brought together<br />
in an evolutionary model of technical change. Research on Genetic Algorithms gives evidence that systematic<br />
procedures to define and execute improvement do not contradict the evolutionary approach. Industrial<br />
implementations of Genetic Algorithms often use sophisticated evaluation functions that bring together<br />
many different objectives. In many cases, specific operators are used to identify better solutions with<br />
respect to one single objective. These operators can significantly increase the performance of the algorithm,<br />
if the interaction between them is organized in the right way.<br />
David Goldberg has repetitively claimed that the way how Genetic Algorithms produce change is closely<br />
connected to human skills in technical development (e.g. Goldberg 1998, Goldberg 2002). This paper does<br />
not give evidence that supports this claim, but it makes clear that the evolutionary search in Genetic Algorithms<br />
can serve as a reference for a model of technical development in a complex economic context.<br />
The concept of Genetic Algorithms gives an explanation how the investment calculations in management<br />
202
Albrecht Fritzsche<br />
accounting and the technical change procedures in research and development can interact as two separate<br />
processes. In addition, Genetic Algorithms using sophisticated operators show how it is possible that<br />
the direction of technical improvement deviates from the selective pressure of profitability. In the future,<br />
studying Genetic Algorithms may allow further conclusions about innovation management.<br />
References<br />
Beck-Sickinger, A.G. and Petzold, M. (ed.) (2009) Paradigma Evolution, Frankfurt, Peter Lang.<br />
Boese et al. (1994) "A new adaptive multi-start technique for combinatorial global optimizations", Operations Research<br />
Letters 16, Amsterdam, Elsevier, pp 101-113.<br />
Broesel, G., Keuper, F. and Woelbing, I. (2007) "Zur Uebertragung biologischer Konzepte in die Betriebswirtschaft",<br />
Zeitschrift fuer Management, 2(4), pp 436-466.<br />
Dennett, D.C. (1996) Darwin’s Dangerous Idea: evolution and the meanings of life, New York, Simon & Schuster.<br />
Dopfer, K. (2008) "Was ist Evolutionsökonomie?", Innovation zwischen Markt und Staat (Ebner, A., Heine, K. and<br />
Schnellenbach, J. ed.), Baden-Baden, Nomos.<br />
Frenken, K. (2006) Innovation, Evolution and Complexity Theory. Cheltenham/ Northampton, Elgar.<br />
Fritzsche, A. (2009) Heuristische Suche in komplexen Strukturen, Wiesbaden, Gabler.<br />
Goldberg, D.E. (2002) The Design of innovation: Lessons from and for competent genetic algorithms, Boston, Kluwer<br />
<strong>Academic</strong>.<br />
Goldberg, D.E. (1998) The Race, the Hurdle, and the Sweet Spot, (IlliGAL Report 98007), Chicago, Univ. of Illinois.<br />
Hodgson, G. M. (2005) “Decomposition and Growth: Biological Metaphors in Economics from the 1880s to the 1980s”,<br />
The Evolutionary Foundation of Economics (Dopfer, K ed.), Cambridge, Cambridge University Press,<br />
pp.105-150.<br />
Hodgson, G. M. (1993) Economics and Evolution: Bringing Life Back Into Economics, Cambridge, Polity Press.<br />
Klaus, P. (2005) "Die Frage nach der optimalen Komplexität in Supply-Chains und Supply-Netzwerken", Perspektiven<br />
des Supply Managements, (Essig, M., ed.). Berlin/ Heidelberg, Springer, pp363-376.<br />
Knowles, J. and Corne, D. (2002) "Towards Landscape Analyses to Inform the Design of Hybrid Local Search for the<br />
Multiobjective Quadratic Assignment Problem", Soft Computing Systems: Design Management and Application,<br />
Amsterdam, pp 271-270.<br />
Marshall, A. (1890) Principles of Economics, Reprint 1930, London, Macmillan.<br />
Nelson, R. R. and S. G. Winter (1982) An evolutionary theory of economic change, Cambridge, Belknap.<br />
Reed; P.M., Minsker, B. S. and Goldberg, D. E. (2001) "The Practitioner’s Role in Competent Search and Optimization<br />
Using Genetic Algorithms", Bridging the Gap. (Phelps, D. and Shelke, G. ed.), <strong>Conference</strong> Proceedings, Reston,<br />
ASCE, 1, 341.<br />
Reeves, C. R. (1999) "Landscapes, operators and heuristic search", Annals of Operations Research 86, 1999, pp<br />
473-490.<br />
Schumpeter, J. A. (1912) Theorie der wirtschaftlichen Entwicklung, Berlin 1911; Reprint 2006, Berlin: Duncker &<br />
Humblot.<br />
Simon, H. A. (2005) “Darwinism, Altruism and Economics”, The Evolutionary Foundation of Economics (Dopfer, K ed.),<br />
Cambridge, Cambridge University Press, pp 89-104.<br />
Solnon, Ch, Cung, V.D., Nguyen, A. and Artigues, Ch (2008) "The Car Sequencing Problem", <strong>European</strong> Journal of<br />
Operational Research 191(3), pp 912-927.<br />
Trott, P. and Lataste, A (2005) “The Role of Consume Market Research in New Product Decision Making: Some Preliminary<br />
Findings”, The Entrepreneurship-Innovation-Marketing Interface (Gaul and Jung ed.), Künzelsau, Würth.<br />
Veblen, T. B. (1898) "Why Is Economics Not an Evolutionary Science?", Quarterly Journal of Economics, 12(3), July,<br />
pp 373–97.<br />
Winter, S. G. (2005) “Towards an Evolutionary theory of Production”, The Evolutionary Foundation of Economics<br />
(Dopfer, K ed.), Cambridge, Cambridge University Press, pp 223-254.<br />
Wolpert, D. H. and Macready, W. G. (1997) "No Free Lunch Theorems for Optimization", IEEE Transactions on Evolutionary<br />
Computation 1.<br />
203
Case Study on Information Evaluation by GIS for Aging<br />
Society Urban Planning: GIS Application on Urban Planning-<br />
Hiroatsu Fukuda, Yupeng Wang and Kiyoshi Shinriki<br />
The University of Kitakyushu of Japan, Japan<br />
fukuda@env.kitakyu-u.ac.jp<br />
y09e0101@hibikino.ne.jp<br />
shinriki@hq.kiu.ac.jp<br />
Abstract: Japan’s aging society and shrinking population are leading to serious problems, including a declining level<br />
of community vitality in most of Japan’s secondary cities. Structural changes in Japan’s population are expected to<br />
continue for the near future, making sustainable urban design and renovation more important than ever. Proper<br />
urban design must seek and evaluate the relevant social information. This research presents a typical case study<br />
using a new information evaluation method, the Geographic Information System (GIS), against the background of an<br />
aging and shrinking society. It is carried out in Yahatahigasi-ku, a ward in the city of Kitakyusyu, Japan. The ward’s<br />
land is characterized by steep slopes, which are inconvenient for aged people and reduce their outdoor mobility.<br />
Moreover, most of the ward’s houses are over 30 years old, the average lifespan for houses in Japan. The quality of<br />
the houses is inadequate to the demands of modern life, perhaps explaining the area’s inability to retain its youth. As<br />
a result, the shopping districts of Yahatahigasi-ku, prosperous in the 1950s and 1960s, are declining in proportion to<br />
the number of deserted houses This research explores GIS design approaches as configured for an aging society. It<br />
discusses the important factors that flow into an urban design meant to address a declining population, such as the<br />
plot ratio, the mixing of commercial and residential spaces, and the placement of public facilities for the aged. The<br />
city planner must find the most suitable area for planning. Sustainable urban design requires that new elements be<br />
introduced into the traditional configuration; the constructional style of buildings in a mixed-use area will reflect the<br />
character of the district. This study considers the GIS urban planning method as a development of the GIS<br />
application. We see here the importance of geographic information to urban planning. This paper will exemplify the<br />
importance of sharing informational evaluations to the urban planning process. In this paper, a case study of Japan<br />
was investigated. It makes contribution on urban regeneration for aging society that also appropriate for cities outside<br />
Japan.<br />
Keywords: urban design, GIS, aging society, compact city, migration<br />
1. Introduction<br />
Urban design tends to correspond to the development of the city. One consequence of Japan’s declining<br />
population is that its shrinking housing density is reducing the utilization of cities’ public facilities. The<br />
results of GIS condition analyses allow us to identify the sites most in need of innovation. Information<br />
evaluation by GIS would be a crucial component of this planning method. This study proposes that<br />
reducing the size and compacting the functions of its cities would solve Japan’s social problems. Efforts<br />
to cope with reduced utilization efficiency by reducing city sizes have not been well executed but<br />
represent an important facet of dealing with our globally aging society. This proposal should be qualified<br />
to meet the special requirements of the aged, who tend to retain a traditional attitude towards the street<br />
that may no longer be suitable. Other, newer elements are also indispensable to sustainable city<br />
planning. These new elements could evolve from traditional ones, connecting different generations. This<br />
study analyzes factors such as the plot ratio, the building type, the mixing of commercial and residential<br />
spaces, and the locations of public facilities for the elderly.<br />
2. Background and goal<br />
Aging society and the diminishing population are getting more serious in Japan, leading to related<br />
problems, such as the declining vitality of communities in most of the secondary cities. This tendency of<br />
the population and its structural change is supposed to continue in near future, thus makes the<br />
sustainable urban design urgent.<br />
Japanese average life span is 78.6(men), 85.5(women) in 2007 the longest in the world. The population<br />
ratio over 65 years old is 22.8% in 2009, largest in the world. As the prospects in 2055, by the research<br />
conducted by Japanese government in 2005, the population will decreased from 127.8 million to 89.9<br />
million. The average life span will be 83.7(men), 90.3(women). The population ratio over 65 years old will<br />
be 40%. According to this result, aging society problem will grow into a serious problem in the further.<br />
The industrial structure of Yahatahigasi-ku has totally changed over the last 50 years in tandem with<br />
changing lifestyles. In 1960 to 70, Kitakyushu supported Japanese rapid economical growth, through the<br />
204
Hiroatsu Fukuda et al.<br />
aggregation of such material industries as the steel, chemical, metal and ceramic industries, which began<br />
when the government-managed Yahata Steelworks, the largest steelworks in Asia, which was founded in<br />
1901. Yahatahigasi-ku, the biggest industrial area of Japan, was deeply affected by the changing times.<br />
Today, the material industry moved to the Asian countries, high-technology industries including industrial<br />
robots, IC-related products and biotechnologies are being generated. The shopping streets of<br />
Yahatahigasi-ku, prosperous in the 1960s and 1970s, are in decline, as the proportion of deserted<br />
houses increases; a new plan for urban regeneration is expected. Yahatahigasi-ku’s problem exists<br />
throughout Japan and other developed countries. An approach to the reconstruction of these typical old<br />
towns will be discussed in this paper.<br />
3. Methods of research<br />
We will first analyze the existing condition through GIS and propose a construction site based on a<br />
comprehensive understanding of the site information evaluation and contracting studies.<br />
The date of population was obtained from web side of Kitakyushu city. The landform current situation and<br />
the distribution of living facilities of Yahatahigasi-ku were inspected by field trips. All of the dates are<br />
collected and synthesized by GIS<br />
The building typologies with new elements appropriate for the elderly will be proposed.<br />
4. Case study<br />
Yahatahigasi-ku is a ward in the city of Kitakyushu, where Yahata Steelworks is located and the people<br />
who work for the company live. The population here is decreased from 81,956 (1988) to 73,162 (2008)<br />
and the population ratio over 65 years old is 30.2% now. The land here is characterized with the steep<br />
slop which is inconvenient for aged people and reduces the chances of their outdoor activities.<br />
On the other hand, most of the houses here are used far more than 30 years, the average lifespan for<br />
houses in Japan. The quality of the house is not adequate for the modern life which may be responsible<br />
for the losing of young people.<br />
4.1 City size reduction<br />
Figure 1: Overview of Yahatahigasi-ku<br />
205
Hiroatsu Fukuda et al.<br />
As shown in the overview of Yahatahigasi-ku, most areas of this ward are mountainous, with steep<br />
slopes. Many residential areas were built for employees during the period of Japan’s industrialization.<br />
Therefore, housing developments extended to the steeply sloped areas as the city grew. The proportion<br />
of retirees has increased over the last 50 years; meanwhile, young people have been fleeing the city<br />
because of its unsuitable living conditions. Thus, the concept of “compact” is an appropriate response to<br />
the ward’s decline. First, we must designate a proper site.<br />
As shown in Figure 2, the steeply sloped areas on the mountain pose an acute problem: access by car is<br />
impossible; one can only walk after the slope begins, making it hard for inhabitants to perform daily<br />
outdoor activities. The elderly, in particular, who have to get to nursing facilities and go shopping, cannot<br />
bear the necessary effort.<br />
Figure 2: Pictures of the residences on the steeply sloped area<br />
This ward has five cultural zones. Two of those are on the steeply sloped area, as shown in Figure 2, and<br />
three lie on a flat area. Four main roads cross in the central zone; there is a railway station near the flat<br />
area that is conveniently accessed from outside the ward. There is a famous theme park, “Space World,”<br />
the biggest commercial area in this ward, and an area boasting three museums and other public facilities.<br />
City functions are already densely compacted here. Based on its merits, the central zone on the flat area<br />
was selected as the best site for innovation.<br />
4.2 Inhabitant migration proposal with information evaluation by GIS<br />
Figure 3 shows the designated migration-in promotion area, migration-out promotion area, and the buffer<br />
area. The study proposes an inhabitant migration from the sloped area to the flat area based on an<br />
examination of the aging population and the old dwellings on the slope as analyzed by the GIS. Many<br />
residences were built on the steeply sloped area. To solve the environmental problems produced by the<br />
steep slope, flat areas were chosen as the object region for inhabitant migration. The migration is<br />
expected to reduce the size of city and compact it, improving the efficiency of city services and facilities.<br />
Migration-out promotion areas are those with high slopes and high ratios of elderly—where there is,<br />
roughly, a level higher than 80m and a slope degree higher than 5. We propose to keep inhabitants from<br />
moving in and to encourage them to move out of this area.<br />
Migration-in promotion areas are flat with convenient traffic and public facilities—where there is, roughly,<br />
a level lower than 40m and a slope degree lower than 3. In addition the areas where is anticipated to rich<br />
of these facilities in the further.<br />
Buffer areas lie between migration-out promotion areas and migration-in promotion areas. Government<br />
offices and universities are situated there. Migration-out areas will become sparsely populated during the<br />
process of migrating-in. In the buffer areas, sparse landscape, community decay, and crime in streets full<br />
of abandoned houses could be alleviated by adding public facilities and encouraging more human<br />
interaction.<br />
A high density of population and buildings is required to keep public facilities efficient. Inhabitants living in<br />
migration-out promotion areas cannot access public facilities conveniently; meanwhile, those public<br />
206
Hiroatsu Fukuda et al.<br />
facilities will decay as population density decreases. Sustainable city planning must compact city<br />
functions and increase population density.<br />
Figure 3: Object region for inhabitant migration<br />
Figure 4: The situation of public facilities<br />
207
4.3 Share-ride taxi<br />
Hiroatsu Fukuda et al.<br />
The share-ride taxi is a kind of conveyance that can take a maximum of 10 people a short distance. It<br />
normally services only one or two residential areas. It does not use a designated station and can stop<br />
and collect passengers anywhere along its route. The price of a share-ride taxi is lower than that of a<br />
normal taxi and is more convenient for the aged than the bus. It can be used for daily shopping or other<br />
short errands. We can expect this form of sharing to lighten the city’s traffic. This system has been<br />
successfully introduced into many areas of Japan.<br />
Figure 5: Pictures of share-ride taxi (http://www.asagiri-town.net/q/aview/102/687.html)<br />
Two share-ride taxi routes are proposed for the central zone. Route 1 is a circle connecting a big hospital<br />
and the old residential areas with the steep slope on the mountain. Route 2 is a circle connecting a big<br />
shopping centre, three museums, and the theme park. Route 1 and 2 cross at the mixed-use area. The<br />
operational hours are to be from 9 am to 5 pm. The full route will take around 20 minutes to travel. It is<br />
now used mostly by housewives and students.<br />
Most of the facilities on the routes could not be reached by bus from the mixed-use area. Life would be<br />
made more convenient for the aged if their public facilities were connected. The GPS (Global Position<br />
System) allows users to pinpoint the location of their taxi in real time on their mobile phones.<br />
Figure 6: Routes of share-ride taxi<br />
208
Hiroatsu Fukuda et al.<br />
Figures 7 to 9 show the image of each stage of the Share-Ride Taxi Rolls. Initially, share-ride taxis will<br />
connect the mountainous areas with the flat commercial areas.<br />
Figure 7: The first stage of share-ride taxi rolls<br />
Figure 8: The second stage of share-ride taxi rolls<br />
Figure 9: The third stage of share-ride taxi rolls<br />
209
Hiroatsu Fukuda et al.<br />
After 10 years, the population density of the flat area will increase with the addition of the inhabitants from<br />
the sloped area. The taxi route in the flat area is proposed for the elderly and others (like students) who<br />
cannot drive.<br />
After 20 or 30 years, the taxis will connect several central regions in the flat areas. Given the results of<br />
the inhabitant migration program, routes connecting the mountainous and flat areas will not be<br />
necessary. Taxi routes in the flat area should enhance the convenience of the compacted city area.<br />
4.4 Site analysis<br />
Figure 10: Public facilities Figure 11: The zoning of city functions<br />
The central zone used to be a gateway into the residential area, through which employees would pass on<br />
their way home from work. We will explore the character of the central zone, analyze the relative<br />
positions of some of its facilities, and discuss the city’s zoning function.<br />
The central zone mixes commercial and residential uses. In this mixed-use area, all necessary city<br />
facilities are located within 500m of the center and are thus conveniently accessible. This mixed-use area<br />
is therefore suitable for habitation.<br />
A city axis extends from the station, goes through the mountainous area, and passes through the<br />
mixed-use area. There should also be a green axis extending from the mountainous area, in order to<br />
strike a balance between the city and nature.<br />
As an urban planning restriction, the maximum plot ratio for the mixed-use area is 500%. However, the<br />
actual plot ratio is much lower than that, as it does not take full advantage of existing city functions. To<br />
maintain a bustling city community, both the population density and plot ratio should be increased.<br />
The mountain at the south side of the site is key to the landscape. We propose to take advantage of the<br />
view of that mountain. This can be achieved through the appropriate building disposition and<br />
building-type analysis.<br />
210
Figure 12: Restriction of plot ratio<br />
4.5 Building-type analysis<br />
Hiroatsu Fukuda et al.<br />
The maximum plot ratio for the mixed-use area is 500%. To achieve a residential building style through<br />
sustainable urban design, we must consider three functions of the building under the plot ratio restriction.<br />
Figure 13: Land use zoning in 3D space<br />
211
Hiroatsu Fukuda et al.<br />
Commerce: commerce affects the activity on the street and should thus occur on the lower floors, not<br />
only for the sake of daily shopping but also for amusement. The main target is youth who live elsewhere.<br />
The plot ratio for commerce is about 100%.<br />
Service facilities for residences: considering our aging society, service facilities for the local elderly are<br />
located on the higher floor: this makes them easier to access and somewhat separate from activities on<br />
the street. The plot ratio for residential service facilities is about 100%<br />
Residences: we will draw contrasts among 5 building types in order to discuss the way the residential<br />
environment is linked to the plot ratio, the number of plies, and the construction styles. We expect to see<br />
high buildings with low plot ratios.<br />
Three types of relationship between building and street are contrasted.<br />
Figure 14: Building types<br />
Effective ventilation, solar irradiation, and privacy can be ensured by changing the residence into a slim<br />
tower; this way, the public space will be used only by residents.<br />
Type 1: face onto the street<br />
Commerce on the lower floors is close to the street. The facade of the Japanese traditional street can be<br />
maintained.<br />
Type 2: Increase greenery<br />
Set the lower floors back and introduce a green area in an open space, though this would differ from the<br />
traditional Japanese style. A slim building also increases the amount of sky an inhabitant can see from<br />
the street. Type 2 takes advantage of the city’s rich and varied local character.<br />
5. Conclusion<br />
This paper analyzed a method of urban design for an aging society through a case study that applies GIS<br />
to a Japanese city, Yahatahigasi-ku, once considered a leading city during Japan’s industrial age. A GIS<br />
information evaluation leads us to propose a method of urban reduction that requires compacting city<br />
functions, connecting access points to public facilities, and improving the residential environment.<br />
Introducing GIS is crucial to achieving the goal of designating a suitable site for inhabitant migration. The<br />
ideal site is the central zone, once an important ward gateway.<br />
212
Hiroatsu Fukuda et al.<br />
Therefore, it is imperative that the city planner find the most suitable area for planning by way of a site<br />
information evaluation. Additionally, formulating a sustainable urban design requires that new elements<br />
be introduced into the traditional configuration. Ultimately, the constructional style of buildings in<br />
mixed-use area will reflect the character of their district.<br />
Acknowledgments<br />
The authors wish to express their appreciation to “Revitalization of Local Area Project” for providing the<br />
subvention for this research.<br />
References<br />
Dewanchker B., (1997) “Comparison of the Change in Industrial Development in the Kitakyushu Industrial Zone and<br />
the EMSCHER Zone in Germany”, Journal of architecture, planning and environmental engineering.<br />
Transactions of AIJ (502) pp.51-56<br />
Dewanchker B., Takahasi N. and Ojima T., (1997) “Ecological Improvement and the Redevelopment of an Abandon<br />
Industrial Site in the City of Uozu”, Summaries of technical papers of Annual Meeting Architectural Institute of<br />
Japan. pp.821-822<br />
Fukahori H. (2006), “Disaster prevention map making with IT method”, Journal of Japan Sewage Works Association,<br />
43(527) pp.43-46<br />
Fukahori H. (2007), “Make Kitakyusyu-city an nature environment capital”, Journal of sewerage, monthly 30(2) (422)<br />
pp.38〜41<br />
Gao W., Ariyama T. and Ojima T., (1999) “Prediction on the Energy Consumption of Renewing the Residential<br />
Recycle Material”, Journal of architecture, planning and environmental engineering. Transactions of AIJ (516)<br />
pp.101-106<br />
Gao W. and Ojima T., (1995) “Study on Planning an Ecological City in SHITAMACHL Tokyo”, Journal of architecture,<br />
planning and environmental engineering. Transactions of AIJ (472) pp.63-71<br />
Gao W., (1993) “Thermal Effects of Open Space with a Green Area on Urban Environment: Part I: A theoretical<br />
analysis and its application”, Journal of architecture, planning and environmental engineering. Transactions of<br />
AIJ (448) pp.15-27<br />
Song H., Guan D., Fukahori H. and Gao W., (2009) “Study on utilization of semi-industrial districts based on GIS in<br />
Kitakyushu : Investigation of present status of semi-industrial districts in Kitakyushu Part 1”, Journal of Asian<br />
Architecture and Building Engineering, pp.529-532<br />
Urban planning general overview of Kitakyusyu city of Japan, (2007) Planning department in city urban planning<br />
bureau of Kitakyusyu, Japan<br />
213
Distortion Free Algorithm to Handle Secondary Watermark<br />
Attack in Relational Databases<br />
Sajid Iqbal, Azhar Rauf, Huma Javed and Shabir Ahmad<br />
University of Peshawar, Pakistan<br />
sajidiqbal84@gmail.com<br />
rauf.azhar@yahoo.com<br />
humajaved15@yahoo.com<br />
shabir_usmany@yahoo.com<br />
Abstract: Publishing data over the internet has greatly facilitated the ways of data transfer and file sharing but<br />
opened new threats of piracy and copyright violations. The technique of relational database watermarking is used to<br />
protect digital assets by embedding imperceptible watermarks into the data before publishing. Relational database<br />
watermarking got inspiration from digital watermarking and efforts are being made to protect databases from such<br />
attacks by embedding watermarks. Database watermarking has been an active area of research for the past few<br />
years. This process causes distortion of data and may affect the quality of information and reporting. A novel<br />
approach is proposed to embed non-visible Unicode character spaces as watermark instead of changing the least<br />
significant bits of numeric data, which avoids distortion of the data. Similarly in the previous works, watermarks were<br />
embedded in numerical data only while our technique is applicable to character type data which is another novelty of<br />
the proposed technique. Researchers have identified different types of attacks against the databases. One of them is<br />
secondary watermark addition attack. The proposed technique provides a distortion free solution to protect the<br />
database from secondary watermark addition attacks. The owner’s private key along with a timestamp issued by a<br />
trusted timestamp authority is embedded to protect the database from the threats of secondary watermark addition<br />
attack. Instead of involving the third party at a later stage, we propose its involvement at the stage of watermark<br />
construction. The third party will keep the record of timestamps, issued against owner’s encrypted key, and database<br />
for which the timestamp has been issued. Experiments have shown that our algorithm is effective even after changing<br />
90% of the contents of database by the attacker. The proposed technique not only takes care of the secondary<br />
watermark addition attack but also other types of attacks for example subset attacks and bit attacks.<br />
Keywords: relational database watermarking, secondary watermark, trusted time stamping<br />
1. Introduction<br />
In the current age of Technology, data is on the move. People can transmit data over the internet or copy<br />
it from different locations. The protection of digital assets like images and videos from piracy and<br />
copyright protection is done through watermarking. But to protect databases from piracy, little effort has<br />
been done so far.<br />
Watermarking Relational databases got its inspiration from digital watermarking. In relational databases,<br />
watermarks are embedded in the least significant bits of numeric values of some specific columns of<br />
selected records (Agarwal 2002) (Hu 2009) (Zhang 2004) (Li 2005) (wang 2008) (Cui 2008). The<br />
selection of these records and columns is based upon some hash function applied to owner’s defined<br />
key, image, voice or fingerprint (Agarwal 2002) (Hu 2009) (Zhang 2004) (Li 2005). Information hiding<br />
using watermarking technique causes some distortion but it is assumed that this distortion is acceptable<br />
(Agarwal 2002) (Hu 2009) (Zhang 2004) (Li 2005) (Wang 2008) (Cui 2008). Later on, if at some stage the<br />
real owner suspects the piracy of his database, he can use the watermark detection algorithm to recover<br />
the hidden pattern and thus prove his ownership.<br />
In this paper we present a better approach to watermark relational databases. We thus propose a novel<br />
distortion free algorithm to protect relational database assets from reversibility and secondary watermark<br />
attacks. In our proposed scheme, an identification image is used as a watermark embedding information<br />
which proves the copyright information. The proposed algorithm is resistive to different kinds of attacks<br />
including benign updates, malicious attacks, reversibility attack and secondary watermark attack.<br />
Rest of the paper is organized as: Section 2 gives a brief introduction to secondary watermark addition<br />
attack. Section 3 describes the existing work regarding secondary watermark attack in relational<br />
database watermarking. Section 4 describes the notations used and the proposed algorithm for<br />
watermark insertion and detection. Section 5 shows some experimental results and analysis and finally<br />
section 6 concludes this paper.<br />
214
2. Secondary watermark addition attack<br />
Sajid Iqbal et al.<br />
Watermarking databases and digital assets has gained enormous attention from researchers since the<br />
term was coined by Agarwal et al. (Agarwal 2002). One of the many reasons behind this effort is the fast<br />
movement of data across internet and many file sharing services for sharing and transfer.<br />
So far, different types of attacks have been observed against the relational databases causing piracy and<br />
copyright conflict including benign updates and malicious attacks. One of the important attacks in this<br />
regard that causes ownership conflict is secondary watermark addition attack. In secondary watermark<br />
attack, the attacker simply successfully discovers an unreal watermark and then fraudulently embeds his<br />
own watermark. Using that watermark he can claim his ownership, thus causing ownership conflict. It<br />
cannot be confirmed which watermark was embedded first. The scenario of secondary watermark attack<br />
can be shown as follows:<br />
The real owner of relation embeds his watermark WR to the original Relation RO using some private key K<br />
as:<br />
WE (WR, K, RO) Watermarked Relation where WE is a function that embeds watermark to the<br />
relation.<br />
An attacker finds an unreal watermark WR' although he doesn’t have access to the watermarked relation.<br />
He embeds a false secondary watermark to the database in order to prove his ownership. The addition of<br />
secondary false watermark to an already marked database can be shown by a scenario as follow:<br />
WE' (WR', K', Rw) Watermarked Relation<br />
Where WE' is the fraudulent operation that embed unreal watermark to the marked relation Rw.<br />
Now the relation contains the watermark of two different persons, the real owner and the attacker. This is<br />
hard to prove the ownership of a relation as both the persons can run their algorithm and thus detect their<br />
respective watermark.<br />
Embedding a secondary watermark to a watermarked relation can cause Destruction.<br />
3. Related work<br />
Relational database watermarking for copyright protection and piracy protection is one of the major<br />
research areas now a day. Different approaches have been adapted to tackle with piracy attacks. But so<br />
far there has been no prominent achievement to handle the secondary watermark addition attack in<br />
relational database.<br />
(Dong 2009) proposed the insertion of characteristic code as watermark information to take care of<br />
invertibility attack. In their proposed approach, the characteristic code is calculated by selecting an even<br />
number of attributes from the relation and calculating the average value of each attribute. Next the even<br />
number of attributes is divided in pairs and members of each pair are compared with each other. And<br />
finally string of bits to be watermarked is calculated by taking XOR of compared attributes. When<br />
ownership conflict occurs, owner can prove his ownership by the help of characteristic code. But a very<br />
obvious shortcoming with this approach is that in order to prove the ownership one must have original<br />
database relation which goes against one of the primary characteristics of database watermarking<br />
“Blindness” as defined by (Agarwal 2002)<br />
(Gupta 2009) proposed another watermark technique, resilient to secondary watermark addition attack.<br />
They claim that their technique is resilient to benign updates, additive attacks and bit attacks. They insert<br />
watermark information to float value attributes. In their proposed approach, initially a tuple and then an<br />
attribute is selected for the insertion of watermark information. After selection of attribute value, a bit is<br />
removed from integer part of the target value and is added to float value of the same attribute. In this way<br />
distortion is produced in both integer and float type values.<br />
An obvious shortcoming with their technique is that it is applicable on the float type attributes only. Very<br />
few relational databases have float type attributes. . They are producing distortion to both integer as well<br />
float value of the attribute causing distortion at maximum level. Furthermore, data containing float values<br />
215
Sajid Iqbal et al.<br />
have more precision and can accommodate distortion to minimum level. They propose that their<br />
technique is resilient to secondary watermark addition attack but after the addition of secondary<br />
watermark attack to their watermarked relation, there is no way to find out which one of the watermark<br />
was embedded first.<br />
(Gupta 2006) proposed an algorithm for embedding watermark information to relational databases which<br />
is based upon the difference expansion of attributes values. Their technique selects two attributes from a<br />
single record and then on the basis of their difference, watermark information is embedded to selected<br />
attributes. Insertion of watermark is based upon the difference of attribute values.<br />
One problem is that they select two attributes from a single record which causes maximum distortion at<br />
record level. Furthermore, they claim that their technique is resilient to secondary watermark addition<br />
attack but after an addition attack, there is no way to find out which watermark information was<br />
embedded first to the relation?<br />
4. Proposed algorithm<br />
In this paper, we propose a novel algorithm to handle secondary watermark insertion attack. The main<br />
idea of the proposed algorithm is as follow:<br />
Database Owner calculates his watermark string by applying one way hash function which is SHA-1<br />
(SHA-1 NSA) in proposed case to his private key ko. He then sends his encrypted key to a Trusted Third<br />
Party which acts as a Time Stamping Authority. Time Stamping Authority (TSA) appends Trusted Time<br />
Stamp τ to the owner’s key and sends back the new key k to owner which is now owner’s key plus<br />
trusted time stamp. Then on the basis of trusted time stamp appended to owner’s key and percentage of<br />
watermark ω decided by the owner, a Unicode non-visible character Hair Space is embedded to<br />
calculated attribute for specific tuple.<br />
In case of an ownership conflict, owner can prove himself as real owner by regenerating his watermark<br />
information. Time stamp appended to the key proves that the watermark of claiming owner was<br />
embedded first in case of any secondary watermark attack.<br />
Our proposed algorithm is more robust as we are providing a two-level encryption security to the owner’s<br />
key. Besides a third party is involved at watermark insertion stage.<br />
4.1 Watermark construction<br />
Watermark information (trusted time stamp appended to owner’s key) is a series of binary string which is<br />
calculated by applying one-way hash function to owner’s key which is actually an image converted to<br />
binary string in our case. Then appending trusted time stamp to the key and again applying one-way<br />
hash function.<br />
Trusted time stamp is issued by trusted third party who acts as time stamping authority. Time stamp<br />
records the occurrence of a specific event at a specific time. Steps included in construction of watermark<br />
can be summarized as follow:<br />
Apply one-way hash function to owner’s key k which is private only to the owner.<br />
k = H (ko) where ko shows the owner key.<br />
Owner after applying hash function to his key sends it to Time Stamping Authority. They append Trusted<br />
Time Stamp to it and again apply a hash function.<br />
k = H (ko $ TTS)<br />
Once trusted time stamp is appended to owner’s key and hash function is applied upon it, the final key is<br />
sent back to the owner. The watermark bit string k which is now time stamp appended to owner’s key is<br />
embedded to the relational database to be watermarked.<br />
The watermarking procedure is shown in Fig 1:<br />
216
Figure 1: Watermark construction process<br />
4.2 Watermark insertion<br />
Sajid Iqbal et al.<br />
In contrast to current approaches, we propose a unique method for watermarking which does not cause<br />
any distortion at all. Instead of changing some least significant bits (LSB) which are mostly exposed to<br />
bit-flip attack and alteration attacks, we add a non-visible character space which is Unicode Hair Space<br />
to non-numeric attributes. Unicode hair space is a character space which is invisible, smaller than a<br />
usual space and does not produce any distortion at all.<br />
In order to distribute the watermark equally along the relation, the embedding process is cell-based. Up<br />
till now approaches have been adopted to embed watermark by first selecting a tuple and then specific<br />
attribute for that tuple.<br />
We are adopting cell based approach in order to distribute our watermark uniformly. The whole relation is<br />
considered as a one-dimension array of data elements and then watermark is embedded to them.<br />
A one way hash function is applied to owner’s defined key. The resulting bit string is sent to trusted third<br />
party, where a Trusted Time Stamp (TTS) is appended to the bit string and again a hash function is<br />
applied to the resulting string. The final key which is the owner’s defined key and trusted time stamp is<br />
returned to the owner.<br />
The embedding algorithm is described as: whereas table.1 shows list of notations used in algorithm<br />
217
Table: 1. Notations used in this paper<br />
Sajid Iqbal et al.<br />
Symbol Description<br />
Η Number of tuples in relation<br />
Ε Number of attributes in relation<br />
Γ Total Number of Cells in Relation (η * ε)<br />
Ω Percent of cells to be marked<br />
N No. of cells to be marked<br />
K Final Secret Key<br />
Τ Trusted Time Stamp<br />
Algorithm 1: Watermark insertion<br />
// Insert a watermark to relation R and return watermarked relation Rw<br />
// Parameters k, τ, ω all are private to the owner<br />
1. Calculate total Number of cells γ = η * ε<br />
2. Calculate number of cells to be marked N = γ * ω<br />
3. While marked_cells ≤ N<br />
i) if bit [H(ko concatenated τ)] = = 1<br />
ii) Add double space to Cell contents<br />
iii) Else add a single space<br />
iv) Increment bit stream of [H(k concatenated τ)]<br />
v) marked_cells = marked_cells + (γ / N)<br />
5. commit<br />
6. return Rw<br />
Some of the parameters including Key ko, time stamp τ and percent of cells ω to be marked are private to<br />
the owner and are decided by him.<br />
First of all in step 1 the total number of cells ‘γ’ in a relation is calculated. It is done by multiplying total<br />
number of tuples ‘η’ in a relation across total number of available attributes ‘ε’ to be marked. In step 2 we<br />
are calculating the number of cells N to be marked out of total available cells. The calculation is based<br />
upon the percentage of watermark as decided by owner. Once the calculation is done, the loop starts<br />
marking the specified cells according to the private key. If the bit of watermark string is binary ‘1’ then<br />
double Hair Spaces are added to the contents of cell under consideration and if the bit is binary ‘0’ then a<br />
single Hair Space is added to the contents of cell. Once all the required cells are watermarked, the<br />
operation is committed and a watermarked relation Rw is returned.<br />
4.3 Watermark detection<br />
If the owner of a database suspects piracy at some stage or there is a secondary watermark attack on<br />
his relation, he can prove his ownership by simply running watermark detect algorithm. Detection<br />
algorithm will successfully retrieve owner’s key and the trusted time stamp which can prove his real<br />
ownership showing that owner’s watermark was embedded first.<br />
218
Sajid Iqbal et al.<br />
As the watermark was embedded uniformly across the relation using cell based approach, detection<br />
algorithm will retrieve the whole pattern.<br />
The detection algorithm is described as follow:<br />
Algorithm 2: Watermark detect<br />
// Search relation R, extract time stamp τ and Key k<br />
// Parameters k, τ, γ all are private to the owner<br />
1. bit_stream = ‘’ //retrieved bit stream from relation<br />
2. Calculate total Number of cells γ = η * ε<br />
3. Calculate number of marked cells N = γ * ω<br />
4. Search relation for desired character<br />
5. While cells ≤ N<br />
i) if cell[contents] contains double space<br />
ii) append ‘1’ to bit_stream<br />
iii) else<br />
iv) cappend ‘0’ to bit_stream<br />
6. Return bit_stream<br />
In case of any piracy suspect or ownership conflict owner will run his watermark detection algorithm. First<br />
of all number of cells are calculated and then on the basis of percentage of watermark decided by owner<br />
at the time of insertion cells are searched for the desired pattern.<br />
On successful completion of detection algorithm, the results returned by detection algorithm are checked<br />
against a pre-defined threshold. If detected pattern reaches the threshold α ownership can be confirmed.<br />
5. Experiments and results<br />
The database that we used in our experimentation contains 14 attributes out of which 8 were numeric<br />
and 6 were non-numeric with one primary key attribute. 30,000 tuples were selected for experiment. We<br />
ran our algorithm on Intel dual core 2 GHz with 2 GB of RAM and running windows XP as operating<br />
system.<br />
5.1 Resistance against secondary watermark attack<br />
The basic purpose of our proposed algorithm is to handle Secondary Watermark attack. An attacker can<br />
add his factitious watermark as a secondary watermark to already marked database. In order to achieve<br />
this, attacker simply follows following steps:<br />
1. Take any factitious key Kf, apply some processing on it and prepare it as watermark string.<br />
2. Randomly select tuples and attributes for the addition of watermark string. Then randomly change<br />
the least significant bit, changing ‘1’ to ‘0’ and ‘0’ to ‘1’.<br />
3. And finally claiming the ownership of the database by retrieving his own fraudulently added<br />
watermark to prove his ownership.<br />
Since the attacker adds his factitious watermark after the embedding of original watermark, real<br />
ownership can be easily found. Whenever ownership conflict occurs, the real owner can retrieve his<br />
watermark which is actually his private key and Trusted Time Stamp representing the time at which the<br />
watermark string was provided and possible time of publication of database.<br />
219
Sajid Iqbal et al.<br />
In our experiments, a secondary watermark addition attack was simulated. A secondary watermark was<br />
added to a watermarked database, producing some distortion in attributes for randomly selected tuples<br />
with % of tuples selected for watermark was greater than that of actually embedded watermark.<br />
Then actual watermark detection algorithm was run. The detection rate of actual watermark was 99%.<br />
The retrieved string was the owner’s key and trusted time stamp which was the same as provided by<br />
trusted third party and same as date of publication of database. The retrieved string finally proved the<br />
ownership of real owner.<br />
5.2 Resistance against benign updates<br />
Along with handling secondary watermark attack, our proposed algorithm is highly resistant to subset<br />
attacks such as subset deletion, subset addition and subset selection attack. We simulated all these<br />
attacks against different ratio of ω = 90% and 50%. Although the proposed technique does not produce<br />
any distortion at all and hence can be applied to entire candidate records but we did not simulate such<br />
scenario.<br />
5.3 Subset selection attack<br />
In the subset selection attack, an attacker simply selects a part of the relation R. In our experiments we<br />
selected different number of tuples from relation R which was watermarked to two different extents i.e.<br />
with ω = 50% and ω = 90%. Experiments showed as shown in Fig. 2 that at both stages our watermark<br />
pattern was found in more than 75% of tuples which was quite enough for the claim of ownership. In the<br />
first go, 90% of total relation was watermarked and then the same relation was embedded with 50%.<br />
Figure 2: Detection rate on subset selection with ω = 90% and 50%<br />
5.4 Subset deletion attack<br />
In subset deletion attack, the copy right violator deletes a subset from the relation R in order to temper<br />
the imperceptible watermark and hence cause the ownership conflict. In our experiment we simulate the<br />
subset deletion attack by deleting at maximum of 90% of tuples and the watermark was still found at a<br />
very good ration as shown in Fig. 3.<br />
5.5 Subset alteration attack<br />
Subset alteration attack is little different in nature from subset deletion and subset selection. In this kind<br />
of attack, the attacker randomly selects attribute values for tuples and changes their least significant bits,<br />
i.e. flip 1 to 0 and 0 to 1. In our simulated experiments, our algorithm proved very robust against this kind<br />
220
Sajid Iqbal et al.<br />
of attack, as we are not producing any distortion in numeric fields. Simulated results against this kind of<br />
attack showed 100% detection of embedded watermark.<br />
Figure 3: Detection rate on subset deletion with ω = 90% and 50%<br />
6. Conclusion<br />
In this paper we propose a novel approach to handle the secondary watermark attack. Up till now,<br />
approaches have been used to embed a private key only. We embed an identification image as well as a<br />
trusted time stamp in our proposed scheme. Trusted time stamp can prove the ownership of real owner<br />
in case of any reversibility attack. Experimental results have shown that the proposed scheme is robust<br />
against different kinds of attacks. The proposed technique does not cause any distortion.<br />
References<br />
Agarwal, R., Kiernan, J., (2002). Watermarking relational databases. Proceedings of the 28th VLDB conference.<br />
China: pp. 155 – 166.<br />
Chen, X., Chen, P., He, Y., Li, L. (2008). A self-resilience digital image watermark based on relational<br />
database, Proceedings of the international symposium on knowledge acquisition and modeling. Wuhan: From<br />
21 to 22 Dec. pp. 698 – 702.<br />
Cui, H., Cui, X., Meng, M. (2008). A public key cryptography based algorithm for watermarking relational databases.<br />
In International conference on intelligent information hiding and multimedia signal processing. Harbin: From 15<br />
to 17 August. pp. 1344 – 1347.<br />
Dong, X., Li, X., Yu, G., Zheng, L. (2009). An algorithm resistive to invertibility attack. In watermarking relational<br />
databases, IEEE Control and decision conference. China: From 17 to 19 June. pp. 1587 – 1592.<br />
Gupta, G., Pieprzyk, J. (2009). Database relation watermarking resilient against secondary watermarking attacks.<br />
Proceedings of the 5th International <strong>Conference</strong> on Information Systems Security. India. pp. 222 – 236.<br />
Gupta, G., Pieprzyk, J. (2008). Reversible and blind database watermarking using difference expansion.<br />
“Proceedings of the 1st international conference on Forensic applications and techniques in<br />
telecommunications, information, and multimedia and workshop”. Australia.<br />
Hu, Z., Cao, Z., Sun, J. (2009). An image based algorithm for watermarking relational databases. In International<br />
conference on measuring technology and mechatronics automation. China: From 11 to 12 April. pp. 425 – 428.<br />
Lafaye, J. (2007). An analysis of database watermarking security, The third international symposium on information<br />
assurance and security. United Kingdom: From 29 to 31August. pp. 462 – 467.<br />
Li, Y., Swarup, V., Jajodia, S., (2005). Fingerprinting relational databases: schemes and specialties, IEEE<br />
transactions on dependable and secure computing 2(1). pp. 34 – 45.<br />
“The keyed-hash message authentication code (HMAC)”, http://www.csrc.nist.gov/publications/fips/fips198/fips-<br />
198a.pdf “Secure Hash Algorithm (SHA-1),” National Security Agency (NSA) – US.<br />
Wang, H., Cui, X., Cao, Z. (2008). A speech based algorithm for watermarking relational databases. In International<br />
symposiums on information processing. From 23 to 25 May pp. 603 – 606.<br />
X. Chen, P. Chen, Y. He, L. Li, “A Self-Resilience Digital image Watermark Based on Relational Database,” IEEE 2008<br />
International Symposium on Knowledge Acquisition and Modeling<br />
Zhang, Z., Jin, X., Wan, J. (2004). Watermarking relational database using image. Proceedings of the third<br />
international conference on machine learning and cybernetics. China: From 26 to 29 Aug pp. 1739 – 1744.<br />
Zhang, Y., Niu, X., Zhao, D., Li, J., Liu, S. (2006). Relational databases watermark technique based on content<br />
characteristic. In First international conference on innovative computing, information and control. Beijing: From<br />
30 th August to September 1 st . pp. 677 – 680.<br />
221
Engineering Change Through the Domains of Enterprise<br />
Architecture<br />
Tiko Iyamu<br />
Tshwane University of Technology, Pretoria, South Africa<br />
iyamut@tut.ac.za<br />
Abstract: Change continues to be a challenge, as organizations strive hard to change with time and with new and<br />
evolutionary processes, activities and requirements. Many organizations are thus challenged by this change. Some of<br />
these organizations attempt to address the challenges of change through different strategic approaches, such as<br />
Enterprise Architecture (EA), which involves the development of an organizational capacity to respond to changing<br />
needs and conditions. The research investigated the practice of EA to understand how EA, through its domains<br />
engineer Change in the organizations in order to respond to rapid technological and business changes. Interpretive<br />
case study approach, using two organizations was adopted.<br />
Keywords: change, enterprise architecture and domains architecture.<br />
1. Introduction<br />
Enterprise Architecture (EA) comprises of several different types of architectures, each with its own type<br />
of deliverables, analysis methods, processes and participants. The EA is defined differently by different<br />
experts, both academics and practitioners. The domains of EA include Business, Data, Information,<br />
Technology, Infrastructure, as well as Service Oriented Architecture. This research adopted the EA,<br />
which consist of four domains: Business Architecture (BA), Information Architecture (IA), Technical<br />
Architecture (TA) and Application Architecture (AA) (Cook 1996; and Zachman 1996). This is mainly<br />
because the two case studies adopted the same model of four domain approach. This helps the study to<br />
reflect and align on its objective. EA is a paradigm which premises to address challenges in the<br />
computing environment. According to the IT Architecture Affinity Group, “Architecture is a blueprint that is<br />
developed, implemented, maintained, and used to explain and guide how an organization’s IT and<br />
information management elements work together to efficiently accomplish the mission of the<br />
organization”. EA is said to enable change by bridging the gap between strategic planning and<br />
implementation efforts through a strategy process that is holistic in its coverage and enterprise-wide in its<br />
scope.<br />
On one hand, organizations’ activities including service delivery practice and organizational structure<br />
keep changing. On the other hand, information technology infrastructure also keeps changing and doing<br />
so rapidly. Nothing is permanent in organizations – not even for months. As a result, there has been<br />
emphasis on the significance of change in organizations (Finkelstein 1999). Many organizations are<br />
struggling to adapt to technological and business rapid change. This research investigated how EA,<br />
through its domains could be deployed to address fast changing need in the organizations.<br />
Where EA is adopted, the processes are a critical step in the growth and development of competitive<br />
advantage particularly in large organizations. This is due to the pace of change of activities, processes<br />
and requirements. According to Mack and Frey (2002), the primary reason for creating an overall IT<br />
architecture is to build the new as well as manage transition from the old. Change is not necessarily<br />
positive; it could be prohibitive and potentially pose risk to the organization. This is the primary motivation<br />
for the study, which led to the research question: What are the elements of change in EA and how<br />
could they be used to engineer change, thereby mitigating potential risk in the organization?<br />
EA is intended to provide the required flexibility to achieve change in the fast-paced, rapidly growing IT<br />
and business environments (Cook 1996; and Watson 2000). Iyamu (2009) argued that EA defines the<br />
scope, scale and nature of required changes within an organization and that it helps to identify the<br />
resources that must become involved. EA also provides a platform to consistently address all the<br />
activities in the organization and many related concerns such as the information and technology that<br />
supports the business processes and activities. This is the argument of some individual and group<br />
experts including Zachman (1996); Armour, et al (2007); Gartner; and The Open Group Architecture<br />
Framework (TOGAF). According to Lankhorst et al (2005), EA is a valuable instrument for<br />
operationalising and implementing policies and strategies.<br />
222
Tiko Iyamu<br />
EA is therefore inevitably seen as fundamental to strategic planning, systems design, software and<br />
hardware development, and the production of multiple systems with different functionalities but all from<br />
the same basic architectures (Youngs, et al 1999). As a result, irrespective of the requirements from both<br />
the computing environment and business, EA thereby ensures uniformity, reuse and manageability of<br />
change from the current state to realization of strategic intent, and thereafter.<br />
In many of the organizations, EA serves as a tool for identifying and driving reuse throughout the existing<br />
and planned architecture. This is not as smooth as it is claimed. According to Spewak (1992), from the<br />
development to implementation stages of EA, there are challenges from both technical and non-technical<br />
perspectives. For other organizations, EA serves as an agent in managing application and technology<br />
infrastructure change within information systems supporting business transactions and information<br />
processing (Zachman 1999).<br />
Due to the complexity of the study, appropriate research method was key, in order to achieve the<br />
objective. This was to ensure a comprehensiveness and richness of the data. The adopted method is<br />
described in the next section.<br />
2. Research approach<br />
The study employed a qualitative, case study approach to investigate the elements which engineer<br />
Change in EA. It is argued that the qualitative methodology is more suitable for this type of study as it<br />
allows clarification of questions. According to Walsham (1995), qualitative research is a very useful<br />
method for complex situations and theories. It has been argued that case study materials are drawn from<br />
different means including interview questions, structured and semi-structured interviews with some of the<br />
key players, documentary sources and ad hoc observational and experience-based notes (Lee 1989; and<br />
Yin 1994). The case study method was applied primarily because it allows in-depth exploration of the<br />
complex issues involved in this study. Two case studies were conducted with a Financial institution and a<br />
Government institution. The two organizations were selected on the basis of prima facie evidence that<br />
each provides a good example of an organization subject to Change as engineered by EA, and each<br />
provides some evidence of success and failure in the process. The two case studies were adopted to<br />
gain an insightful, qualitative interpretation of the changes which were happening as a result of the<br />
practice of EA in the organizations.<br />
The Financial institution is referred to as “Company A” and “Company B” the Government institution.<br />
There were 150 and 80 employees in the computing environments of Company A and Company B,<br />
respectively. Each of the companies has been in operations for more than hundred years.<br />
Data sources included interviews and documentation. As shown in Table 1 below, the number of<br />
interviewees varied, depending on the size of the organizations. A set of balanced respondent<br />
demographics was a key factor in achieving a true reflection of the situations. Targeted respondents<br />
were from different units and were at various levels of the organizational structure within the Business<br />
and IT departments. They included Business and Systems Analysts, IT Architects, IT Managers, IT<br />
Project Managers, IT Executives and Business Managers. The interviewees, as tabulated in Table 1<br />
below, were selected on the basis of their closeness to the topics of EA, including their individual levels of<br />
experience in management, organizational and architectural issues. In the study, closeness is defined as<br />
the area of specialization, with significant amount (5 years) of stock of knowledge and experience. In<br />
Company A, the Chief Technology Officer (CTO) and in Company B, a senior manager assisted in<br />
identifying the interviewees in their various organizations. Employees also assisted in identifying relevant<br />
interviewees.<br />
Structured and semi-structured interviews were designed to elicit information on what the Change<br />
elements were in the practice of EA. Questions were asked around the understanding of EA; the<br />
perception of the impact and implication resulting from the practice of EA in the organization; and<br />
involvement and levels of commitment of actors in the organizations.<br />
The data was analyzed using the interpretive approach. Interpretative rules create signification or<br />
meaningful symbolic systems that provide ways for actors to see and interpret events. Agents reflexively<br />
apply interpretative schemes and stocks of knowledge. The interpretive approach looks at ‘reality’ from a<br />
different perspective (Walsham 1995). According to Kaplan and Maxwell (1994), in an interpretive<br />
research project, there are no predefined dependent and independent variables, but a focus on the<br />
complexity of human sense-making as the situation emerges.<br />
223
3. Analysis of the case studies<br />
Tiko Iyamu<br />
The analysis was done to understand how EA engineer change in the organizations. The results of the<br />
case studies are combined and presented in this section.<br />
As mentioned in the introduction section, both case studies adopted the model as advocated and<br />
supported by consultants and groups including Zachman (1996); Gartner; and TOGAF. The model<br />
categorizes domains of EA: Business Architecture (BA), Information Architecture (IA), Technical<br />
Architecture (TA) and Application Architecture (AA) were the domains which constituted EA in the<br />
organizations. Each of these architectures has a specific set of deliverables, analysis methods,<br />
processes and participants through which they facilitate change within the computing environment and<br />
between IT and Business departments. Thus EA has made ongoing representation of the existing and<br />
planned information systems required to support the ever-changing business strategies. This included<br />
the consequential business processes, information, and application requirements and technology<br />
selections.<br />
All the activities and processes which were happening in the computing environment of the case studies<br />
at the time were categorised into domains of BA, IA, TA and AA. This includes selection, deployment and<br />
use of technologies as well as the services within the organizations and their clients. There were many<br />
architects in each domain. Each domain was managed by a senior architect, titled architecture manager<br />
and team leader in companies A and B, respectively. The EA was managed by Chief Architect and<br />
General Manager in Company A and Company B, respectively.<br />
Each domain had its guiding Principles, Standards and Policies through which their various mandates<br />
were implemented and managed. These principles, standards, policies were defined based on IT and<br />
business requirements and interpreted by the architects. As a result, the roles of the architects were<br />
critical in the deployment of EA in the organizations. As such, not anyone could be appointed to take on<br />
the task of architect, it required special skill. For example, in Company A, the evaluation and<br />
recommendation of server virtualization and consolidation which was conducted in the organization<br />
required in-depth knowledge of specialized skill. The project was led by the architecture unit on behalf of<br />
the organization.<br />
One of the main requirements for the deployment of EA in both organizations was to bridge the gap<br />
between the business and technology components of the organization. This requirement was<br />
emphasized and prioritized in the organizations. This occurs during the process of development and<br />
implementation of EA. In order to achieve the required flexibility to handle the fast-pace of IT, change<br />
was inevitable. This was a serious challenge in both case studies. As a result, there was emphasis on<br />
business vision and requirements, how they were extracted, documented and achieved.<br />
Technical requirements were as important and critical as the business requirements in engineering<br />
change through EA. Technical requirements were defined, implemented in both technical and application<br />
domains of EA. The drivers were however derived from information and business architectures. Some of<br />
the business managers acknowledge the contribution and role of EA in their various processes. A<br />
business manager in “Company A” had appreciatively expressed as follows: “Shareholder Value would<br />
not be successfully achieved if IT didn’t support and influence aligned business goals – imagine the<br />
business driving client centricity and IT ignoring this and continuing to push silo- or product-centric<br />
solutions, thus jeopardizing or even failing to drive projects to focus on the client (like Client Relationship<br />
Management), and thus continue to compete for the client’s business and not integrating IT solutions<br />
across the organization for a single client experience and view; such unaligned behaviour is of the past<br />
and cannot drive profitability, as the client inevitably gets annoyed with the corporate organization for<br />
wasting his time and hence taking his business elsewhere”.<br />
However, available skill-set posed a serious challenge, especially in the area of business architecture.<br />
Company B didn’t have enough skilled people to carry out the tasks as dictated by EA. The other<br />
organization fared better. According to one of the interviewees, a senior manager in Company B, the cost<br />
for the retention of people with specialized skill was prohibitive. The manager admitted that in some<br />
areas, change couldn’t be implemented as a result of lack of technical know-how.<br />
According to an interviewee, a senior manager in Company A, one of the biggest challenges in managing<br />
change is gaining an accurate view of the current state. It was clear, EA is critical and significant in<br />
224
Tiko Iyamu<br />
providing a more accurate view of the current quo. Once the status quo was identified, a plan of action<br />
followed. EA further facilitates the assessment of the impact of Change on the current business<br />
environment and the conception of strategic alternatives for consideration. According to one of the<br />
interviewees, “Yes definitely, the company is now able to consolidate technologies and business<br />
solutions, as well as more rapidly engage in mergers and acquisitions – architecture is an intellectual<br />
capital investment increasing organizational maturity from tactical to strategic”.<br />
The organizations were determined to establish current-state of the environment in order to understand<br />
the gap, if any. Particularly, senior executives were keen to understand how EA could be applied to<br />
reasonably articulate their future business environments and resulting strategies and then use gap<br />
analysis to assess the impact of change and associated risks. To achieve this, the cooperation of both<br />
the computing environment and the business units of the companies were very important, including the<br />
support of the executives and other managers in the business. The IT architects realized and admitted<br />
the criticality of the support and buy-in of these managers to the success or failure of EA in the<br />
organization. Interviewee, some of them, architects in Company “A” explained as follows: “More and<br />
more business managers walk the architecture talk. I’ve often been in discussions between business<br />
managers and outsiders, where the business manager’s first question would be how proposed solutions<br />
fit with our defined architecture”. Similarly, in Company “B” stakeholders understood the importance of<br />
business buy-in. One of the interviewees, an architecture manager says: “Get the business buy-in to the<br />
architecture process first, and let them sell it to the rest of the organization. . . A closer alignment with<br />
business was formed in order to ensure a speedy and successful implementation”.<br />
The scope of EA was therefore the union of the organization, the business reengineering and<br />
development that was applied to it, and the technical domains that supported it. Thus, EA was defined<br />
through a pragmatic need: the need to design and redesign as well as continuously improve the<br />
functioning of organizations. EA can be applied in different environments and contexts to facilitate<br />
change. A business manager in “Company B” says “the business architecture has helped in the design,<br />
documentation and implementation of process-oriented activities in the organization”. This view was<br />
echoed in many quarters of the organization. Processes were easier to follow as a result of avoidance of<br />
re-inventing the wheels. Processes were reused in accordance to set rules and definition of the various<br />
domains of the EA.<br />
In the organizations used in the study, change happened in the areas of process, information flow,<br />
business function, technology innovation, as well as personnel movement. The frequency of change was<br />
a huge challenge, and the impacts and implications were of significant importance to both organizations.<br />
The changes were instigated and influenced by internal and external factors. The internal factors<br />
included business as well as IT vision, strategies, roles and responsibilities, technical know-how and<br />
budget. While the external instigating and influencing factors were technology innovation, industry trends,<br />
market competitiveness and government policy.<br />
Each component of architecture provided a unique view of the enterprise leading to its own capabilities,<br />
primarily: the BA provided the tools, models, techniques, and participants to manage the impact of<br />
change on business processes, clients and partners. Similarly, the IA enabled the management of<br />
change on information flow and exchange between applications as well as processes; the TA enabled<br />
the management of change on technology infrastructure; and the AA enabled the development and<br />
portfolio management of business application and software.<br />
However, the level of success or failure was not measured as it was out of the scope of the study. This<br />
could be an opportunity for future research work.<br />
The study revealed that EA was typically not mastery of a single innovation, but ongoing learning,<br />
adaptiveness and the development of collaborative work cultures. It was admitted that changing the<br />
culture of the organization was the highest ambition. According to one of the managers in “Company A”,<br />
“EA seem to be helping in achieving the challenges of rapid change in the organization”. Even at the<br />
demanding level, EA helped by giving all concerned a shared view of how the business works, and what<br />
it works within the context of the organization.<br />
The organizations’ strategies were the means and the domains of EA were the ends in an attempt to<br />
achieve the vision of the organizations. The study also revealed that the change elements of EA impact<br />
both technical and business components of the organizations. This was a key factor in bridging the gap<br />
225
Tiko Iyamu<br />
between the business and IT through the domains of EA. The change elements in EA had strong<br />
influence on the development and deployment of technology strategy and the amalgamation of both the<br />
IT and the business units towards one vision.<br />
Each domain had its specific template, which was designed to help achieve both business and IT<br />
objectives. The template was used to review the current state, mainly to detect deficiency, identify gap<br />
and plan for the future. The review was a major process carried out every quarter in Company A. In<br />
Company B, it was an exercise, which was done only in the technical domain on yearly basis.<br />
The templates made prediction and specifics possible in the various domains. The areas covered by the<br />
domains became stabled, uniformed and easily accessible. According to some employees in the two<br />
organizations used in the study, the templates reduced guesswork and try-and-error. Thus, encouraged<br />
and helped bridged the relationship gap between architects and other employees in the IT department,<br />
on one hand, and between the IT and business departments on another hand.<br />
4. Findings<br />
The above analysis reveals some findings. The most critical of them are presented below. The headings<br />
were reached based on what were common and critical in both organizations:<br />
4.1 Business requirements<br />
The business requirements were extracted from the companies’ strategies. They included operational<br />
goals and strategic intent, which were being pursued at the time. It focused mainly on the strategic<br />
objectives of the organization. The requirements included the drivers and trends of the organizations’<br />
vision and strategy. These drivers were influenced by factors such as time, economics and<br />
environmental trends.<br />
In order to achieve the required flexibility to handle change in the fast-paced, rapidly growing IT and<br />
business environments, EA was deployed with the intention of addressing the organizational needs and<br />
to facilitate change through the adaptation of innovations of technology to business process model, in a<br />
uniform manner. This was said to be successful only if it aligns with the business strategy and have the<br />
buy-in of the business executives.<br />
4.2 Technical requirements<br />
The technical requirements were based on the business processes and activities, which were meant to<br />
support and enable competitive advantage. The requirements drive the introduction and selection of new<br />
technology by focussing on function and scalability, compatibility and interoperability. It translated the<br />
business requirements into common terminology, which enabled consistent semantic meaning across<br />
information systems and technology in the organizations. These were the primary focus of the TA and AA<br />
domains. Through this approach, for example, technologies were certified usable or obsolete in terms of<br />
manageability. The change in technologies often had impact on business processes.<br />
4.3 Process oriented<br />
The analysis revealed that EA is a process oriented approach, not an event or project. This seemed to<br />
have been understood by many of the promoters in the organizations. For example, it required continued<br />
business and technical requirements to be translated and technology selected to respond to changing<br />
needs. It continually facilitated other activities such as concept reuse, system role definition, and the<br />
modeling of optimal information flows. It guided organizational learning and constantly enabled learning<br />
of new things and subsequent sharing of that learning. One of the examples as revealed in the study<br />
was, if the direction of the Enterprise’s vision changes, then resources and infrastructure has to be<br />
managed so that they could be shared in order to achieve the vision. This was done in a virtual circle<br />
manner and it became the norm in Company A.<br />
4.4 Skill set<br />
In EA, it was argued to have adequate specialized skill set as the execution of its functions required more<br />
than basic understanding hence skill set was critical. However in both case studies this was a challenge.<br />
Specialized skill sets were required in all the above listed findings. The issue of skill was as important as<br />
the rest of the factors in the development, as well as in the implementation of EA to engineer change in<br />
226
Tiko Iyamu<br />
the organizations. Change requires the development of an organizational capacity to respond to<br />
changing needs and conditions within scope and periodically. This is only achievable through skilled and<br />
adaptability of personnel, including the architects. The architects engineered the process of change<br />
through domains of EA, which was under their auspices. The organizations fell short in some areas of the<br />
domains, such as business.<br />
Change involves adaptations in practices, skills, and often beliefs about what is important and valuable,<br />
such as requirements of both technical and business. It required the interaction and connectedness to<br />
ensure uniformity, alignment of skills, business and technology processes. In addition, change was<br />
intended to enable sharing of power across different components of the organizations. These were very<br />
significant and were only achievable through specialized skill set.<br />
5. Interpretation of the findings<br />
The above four findings were interpreted and six critical factors were extracted. The factors are<br />
Adaptiveness, Innovation and Governance, Culture, Uniformity, Strategic Alignment and Technical Knowhow.<br />
These factors were primarily, the taproots in engineering change through EA. They were critical and<br />
prevalence in the activities of the EA domains.<br />
5.1 Adaptiveness<br />
EA is not a "quick fix" solution for the replacement of lack of planning. It requires adaptiveness. This<br />
enables it to effectively and efficiently build, maintain, and apply the entire domains of EA. It is more<br />
beneficial when it is applied uniquely to each need of the organization, depending on its business<br />
strategies, architectural maturity, priorities, corporate culture, and political environment. Its adaptiveness<br />
enabled any of the domains to address the potential risk as well as enable and support the movement<br />
from current to the desired state. Change must be adaptive to achieve its goals, either in business or<br />
technology.<br />
5.2 Innovation and governance<br />
Through its change process, EA attempts to find different and better ideas and ways of enhancing and<br />
increasing the companies’ productivity in a competitive market. Innovation was well guided, this including<br />
the selection and deployment of technology and process engineering. Through domain, EA provides<br />
governance in many activities such as data usage, software development and infrastructure deployment<br />
and management in the computing environment of the companies. For example, in Company B, EA was<br />
used to carry out investigations and implementation of server consolidation and virtualization as dictated<br />
by the requirements of the organization.<br />
5.3 Culture<br />
The rules, norms and procedures, together, created a specific culture, which were institutionalized over<br />
time in the individual company. According to Iyamu (2009), institutionalization is the process where a<br />
practice is assimilated into the norm. It is not easily disassociated, dismantled or re-designed. This<br />
affected the human actors and was also reflected in the use of technology. For example, no software<br />
could be installed on the network server without a written approval from the domain expert in architecture<br />
team. This was the norm. It helped to change individual’s attitude and approaches in carrying task in the<br />
computing environment.<br />
Culture was one of the key elements through which EA impacted change in the organizations. Within the<br />
scope of the organizational requirements, EA was developed and implemented and practiced. Certain<br />
factors of a personal nature need a particular organizational culture in which to thrive, or would feed on<br />
particular internal policies. Similarly, certain internal policies would only be possible within a particular<br />
organizational culture.<br />
5.4 Uniformity<br />
EA provided a uniform process for selecting technology infrastructures, of documenting the current state<br />
of business process, data and infrastructure; future states; and the gap between the current and future<br />
states. This applies to the autonomous business units of the organizations. This was critical to providing<br />
support for both the business processes and technologies in terms of factors such as technological<br />
enablement reuse, licensing optimization and skill management.<br />
227
5.5 Strategic alignment<br />
Tiko Iyamu<br />
The EA will not work if there is no understanding between the business and IT specialists and<br />
executives, but that it can encourage that understanding. The understanding typically concerns change<br />
of processes, management, methods, technology infrastructure design, pattern and selection. EA was<br />
instrumental in facilitating closer relationships as well as bridging the gap between the business and IT<br />
through sharing of a common view and vision of what had to be done and how. Similarly, each domain<br />
was aligned to the IT and business strategies of the organization.<br />
5.6 Uniformity<br />
The EA approach facilitated processes, activities and procedures within the computing environment of<br />
the organizations. This included all the domains of EA and began from project initiation to implementation<br />
stages. It facilitated framework to manage and share information and to ensure that the business<br />
processes and activities were supported by technology infrastructure and applications that provided the<br />
needed data, information flow and the infrastructure to enable the data.<br />
The EA facilitated change, which affected not only business processes, clients and partners, but also<br />
information, supporting business applications, and the technology infrastructure. This made it a<br />
challenging task to attempt to manage change by any means, including the adoption of EA. The evidence<br />
from the case studies revealed the elements, which made EA a change agent. On the basis that the<br />
change was the focal factor for EA deployment, these elements were also used as measurement criteria<br />
for success and failure of EA in the organizations.<br />
6. How EA engineer change through its domains<br />
The findings and the interpretation presented above are now been mapped to the domains of EA, from<br />
the perspective of change. The mapping was done to make sense of the interrelationship between the<br />
findings and the interpretation and how they engineered change through the domains of EA.<br />
As identified in the findings, the most critical factors required in engineering change through the domains<br />
of EA were Business Requirements, Technical Requirements, Process Oriented and Skill Set. This is<br />
illustrated in Figure 1 below. Based on these factors, the EA formulated principles, standards and policies<br />
within the various domains. These domains engineered change through identified elements. The primary<br />
aim was to move the organization from the current to future state.<br />
Resource EA Domain Change Element<br />
Business Requirements Business<br />
Architecture<br />
Technical<br />
Requirements<br />
Information<br />
Architecture<br />
Process Oriented Technical<br />
Architecture<br />
Skill-Set Application<br />
Architecture<br />
Adaptive<br />
Innovation &<br />
Governance<br />
Culture<br />
Uniformity<br />
Strategic Alignment<br />
Technical Know-how<br />
Figure 1: Engineering change through EA domains<br />
The domains were not independent of each other in the business processes and technical activities of<br />
the organizations. Change to one domain had impact on others. The domains collaborated with each<br />
other. This is illustrated with the up and down link arrows in Figure 1. Typically, businesses pursuits<br />
improved or reacted to change schematically.<br />
228
Tiko Iyamu<br />
The discussion that follows should be read together with Figure 1 to get a full appreciation of how<br />
elements of resources constituted themselves into the domains, through which change was engineered<br />
in the organizations.<br />
The next section provides how EA engineered changed through its domain in the organizations. The<br />
change effort is articulated on generalization. Generalization is an explanations of particular phenomena<br />
derived from the empirical interpretive study in the specific settings, which could be valuable in the future<br />
in other organizations and contexts.<br />
6.1 Business architecture<br />
Base on the organizations’ strategy, business requirements were formulated. The BA developed<br />
principles, standards and policies to provide guidance for the development of response plan. It also<br />
ensures uniformity of processes, where possible, across the organizations. In addition, the BA:<br />
Led the executive team and the line-of-business managers through the development and usage of<br />
their various strategic models, which was intended to be adaptive to changing needs.<br />
The gap between strategy development and tactical decision-making was a shortcoming of many<br />
strategic plans. BA eliminated much of the guesswork for operational managers in interpreting the<br />
impact of strategic plans on operations, and the priority that was placed on different plans. This was<br />
done by providing guidance concerning the organization information assets to knowledge workers,<br />
information processors, IT application developers, infrastructure managers, and executives.<br />
6.2 Information architecture<br />
The elements were able to effect Change mainly through their impact on the organization’s key factors.<br />
The IA defined availability and usage of information for business requirements and other key business<br />
processes and enabled the underlying processes and skill-sets for changing needs in the organization. IA<br />
primarily engineered change by:<br />
Defining the sources of information and ensure the availability and usage of this information by the<br />
key business processes and enabled by the underlying application and technical architectures.<br />
Identifying the information flows that was intended for optimization (increased velocity, density, and<br />
reach) as well as the information entities that had to be defined and used consistently across the<br />
information value chain to increase the value of information across the organizations and the external<br />
transactions.<br />
6.3 Technical architecture<br />
In the deployment of new technology, based on functional and utility advantages, the TA facilitated<br />
alignment between business and technical requirements and ensures compatibility and interoperability<br />
with the strategic infrastructure. This was a key to rapid response continual changes in the organization.<br />
It provided configuration standards and guidelines for using standard products in the environment. TA<br />
was adopted to manage technology change in two primary ways:<br />
Provided the principles, standards, configurations, process, and governance mechanisms to<br />
establish and maintain an adaptive technology infrastructure and application reuse methods.<br />
Drove the introduction of new technology based not only on functional and utility advantages, but<br />
also on compatibility and interoperability with the existing architecture. This was key to rapid<br />
implementation of infrastructure configuration standards; guidelines for using standard products; and<br />
principles for integration. These provided both a view of the recommended technology and basis for<br />
assessing the impact of new and replacement technologies within the context of the whole<br />
organization.<br />
6.4 Application architecture<br />
The application architecture defined the way applications were designed and deployed, and how they<br />
cooperated with each other. The AA provided standards and guidelines that supported Build, Reuse and<br />
229
Tiko Iyamu<br />
Buy of robust application for business productivity gains. The AA employed and managed application<br />
change through oriented process, innovation and adaptiveness. Other focus areas included:<br />
Defined the gap between the existing systems functionality and the functionality needed to satisfy BA<br />
and IA requirements.<br />
Provided a framework for migration by taking into account not only functional requirements, but also<br />
investment strategy in information systems and the “technology fit” of existing and planned systems<br />
with the TA and technology infrastructures.<br />
Responsible for reuse, this was a key component to success in the drive to facilitate change, by<br />
guiding developers toward systemic analysis, design and development of business components and<br />
objects.<br />
In summary, EA provides standards based planning, development and implementation methodology that<br />
will help IT strive to meet the rapid changes in the business processes. It allowed a translation of<br />
functional requirements to the selection of services, standards, components, configurations, their<br />
phasing, and the acquisition of products that implement them – much more than the project management<br />
and systems analysis disciplines that have been widely adopted over the last thirty years in trying to<br />
solve the same problems. Some independent consultants have in the past attempted to generically<br />
address some of these elements of Change. This has however been without empirical evidence as in this<br />
study.<br />
7. Conclusion<br />
Neither Change nor the development and implementation are simple in any organization. The return on<br />
investment depends on how well business and technical requirements are articulated, and how the<br />
people interpret the process and the technology they are using to support and enable the business. The<br />
contribution of the study arises from implications for the decision makers responsible for sponsorship of<br />
EA, including the architects. These decision makers need to understand the functions, dynamics and<br />
causes of what, why and how EA bring about Change on both business and technology, and how the<br />
impact can be ensured and measured. The other contribution of this study aims to be of significance to<br />
decision makers, professionals, including managers and employees of the organization within the<br />
computing environment, and IS researchers. It is expected that the key contribution will arise from the<br />
understanding of the fundamental elements through which EA engineer Change. Through this, a better<br />
understanding of the contribution of non-technical factors in the deployment of EA is gained.<br />
References<br />
Armour, F., Kaisler, S. & Bitner, J. (2007): Enterprise Architecture: Challenges and Implementations. HICSS, 40th<br />
Annual Hawaii International <strong>Conference</strong>, System Sciences, volume, Issue, pp. 217 – 217.<br />
Cook, M.A. (1996): Building enterprise information architectures: reengineering information systems, Prentice-Hall,<br />
Inc., Upper Saddle River, NJ.<br />
Finkelstein, C. (1999): The Pace of Change http://members.ozemail.com.au/~visible/papers/dmr02.htm<br />
Gartner. The Business of IT. Accessed 13 February 2008. http://www.gartner.com/technology/research.jsp<br />
Iyamu, T. (2009). The Factors affecting Institutionalisation of Enterprise Architecture in the Organization, Proc.<br />
<strong>Conference</strong> on Commerce and Enterprise Computing (CEC09), IEEE Computer Society, July 2009, pp. 221-<br />
225.<br />
Kaplan, B. & Maxwell, J. A. (1994). Qualitative Research Methods for Evaluating Computer Information Systems. In:<br />
J.G. Anderson, C.E.<br />
Lankhorst et al. (2005): “Enterprise Architecture at Work: Modelling, Communication and Analysis”, Springer.<br />
Lee, A.S. (1989): A scientific methodology for MIS case studies, MIS Quarterly, vol. 13, no. 1, pp. 33 – 50.<br />
Mack, R and Frey, N. (2002). Six Building Blocks for Creating Real IT Strategies. Gartner Group Inc.<br />
Spewak, S.H. (1992): Enterprise Architecture Planning: Developing a Blueprint for Data, Applications and<br />
Technology, John Wiley & Sons Inc., New York.<br />
The Open Group Architecture Framework (TOGAF). Understanding the IT architect's place in today's business<br />
environment. Accessed 27 June 2008. http://www.ibm.com/developerworks/library/ar-togaf1/.<br />
Youngs, R., Redmond-Pyle, D., Spaas, P. & Kahan, E. (1999): A standard for architecture description, vol. 38, no. 1,<br />
Enterprise Solutions Structure.<br />
Yin, R. K. (1994): Case Study Research, Design and Methods, 2 nd ed., California, Newbury Park; Sage Publications.<br />
Walsham, G. (1995). “Interpretative Case Studies in IS Research: Nature and Method”. <strong>European</strong> Journal of<br />
Information Systems, vol. 4, no. 2, pp. 4–81.<br />
Watson, R.W. (2000): An Enterprise Information Architecture: A Case Study for Decentralized Organizations. HICSS,<br />
p. 7059, 33rd Hawaii International <strong>Conference</strong>, System Sciences, vol. 7.<br />
Zachman, J.A. (1996): Enterprise Architecture: The View Beyond 2000, <strong>Conference</strong> Proceedings, Warehouse<br />
Repository Architecture Development 7th International Users Group <strong>Conference</strong>, Technology Transfer Institute.<br />
230
Evaluating Interoperability Artifacts for the Exchange of<br />
Public Service Information: Outline of a Conceptual<br />
Framework<br />
Veit Jahns<br />
University of Duisburg-Essen, Essen, Germany<br />
veit.jahns@icb.uni-due.de<br />
Abstract: Drivers to improve the interoperability of information systems in public authorities are initiatives for<br />
simplifying the access to public authorities and their services for citizens and enterprises. The spectrum of such<br />
initiatives reaches from initiatives on the regional and national level up to global initiatives. All these initiatives have in<br />
common, that they are related in some way or the other to public services. I.e., they have to deal with the challenge<br />
of describing public services either as a base for exchange information about these public services within the public<br />
administration or between public authorities and citizens and enterprises or as a base for developing information<br />
systems to support the provision of public services. As a result of these initiatives numerous artifacts have been developed,<br />
including languages and methods for modeling public services, ontologies and taxonomies for public<br />
services, etc. Although, this multitude of artifacts has the positive effect, that there is a high chance, that every aspect<br />
of the complex concept “public service” is covered by at least one of these artifacts, it is difficult to keep track on all<br />
these artifacts and how are these artifacts are related to each other, in particular what interoperability issue they<br />
address in detail, on which foundations and assumption these artifacts are based on, etc. In this paper a conceptual<br />
framework for evaluating such artifacts is proposed, which shall allow an evaluation of these artifacts with respect to<br />
the question given above. But this conceptual framework can be also useful for support of an artifact to solve a<br />
particular description problem regarding public services on the one hand, and on the other hand, it can be useful for<br />
the integration of these artifacts.<br />
Keywords: eGovernment, evaluation, public service, interoperability, artifact<br />
1. Introduction<br />
Once Garfield (1977: 175) with the numerous scientific journals and the need for literature reviews in<br />
mind stated, that maybe there are too many journals. The same can be stated for interoperability artifacts<br />
today: there are (too) many of them. Driven by private and public initiatives standards, models, as well as<br />
modeling languages and methods have been developed to address different interoperability problems<br />
and issues. This applies not only for the whole field of interoperability between information systems in<br />
public and private organizations, but also in the field of interoperability of repositories for public services.<br />
The objectives of such initiatives with respect to public service repositories are to simplify the access to<br />
public services as well as to improve the production of public services by a better integration of public<br />
authorities across administrative as well as geographical boundaries. Examples for such initiatives are<br />
projects within the <strong>European</strong> Union for the implementation of the <strong>European</strong> Service Directive, which demands<br />
a single point of contact, where a company can handles all its concerns with the public administration.<br />
But this doesn't require, that the single point of contact has to provide all these public service,<br />
but rather that this single point of contact manages the communication and transaction with the<br />
authorities providing the respective public services, and thus, needs to be interoperable regarding these<br />
public services with the respective, in particular with the respect to the exchange of information about<br />
public services.<br />
In addition, the objective of simplifying the access to public services gains more relevance by developments<br />
on the political level. E.g., it is stated in the “Ministerial Declaration on eGovernment” by the<br />
<strong>European</strong> ministers responsible for the eGovernment policy in their countries (<strong>European</strong> Union 2009),<br />
that it is the common aim of the respective states to facilitate and simplify the access to public services –<br />
not only for their own citizens and enterprises, but also for the citizens and enterprises from the other<br />
states. These efforts finds expression in many research and implementation projects that have been<br />
started in the past, e.g., TERREGOV (http://www.terregov.eupm.net), Access eGov (http://www.accessegov.org)<br />
or SemanticGov (http://www.semantic-gov.org), or are underway, e.g., Peppol<br />
(http://www.peppol.eu).<br />
Besides this Euro-centric examples many other examples for such initiatives dealing in one way or the<br />
other with interoperability issues regarding the provision of public services can be found in America, Asia-<br />
Pacific or even on the global-level. But to shorten the discussion, in all these projects, initiatives and<br />
efforts many different artifacts were created: ontologies, data models, modeling and markup languages,<br />
231
Veit Jahns<br />
etc. In general, these artifacts serve the purpose to describe public service in a unified manner, but every<br />
project has a particular perspective on public services. So this rise the questions, if they address the<br />
same interoperability issues or do they focus only on selected and distinguished interoperability issues, if<br />
these are artifacts based on the same technological and methodological basis, and how these artifacts<br />
are related to each other? Is there a reason, why there are so many different interoperability artifacts or is<br />
this field—as Frank (2000: 48) once self-critically states—just another play-ground for researchers and<br />
the purpose of the artifacts’ creation just to satisfy its creators (Frank 2000: 48)?<br />
Against this background it seems to be a promising task to review and evaluate the current state-of-the<br />
art of interoperability artifacts for modeling and describing public services, in particular to answer<br />
questions like how are these artifacts related to each other, i.e., do they compete or complement each<br />
other, which interoperability issues are addressed by the interoperability artifact, what are the technological<br />
and conceptual foundations of these artifacts, what assumptions about the domain of public<br />
services are made for the developing of these artifacts, etc. Answers to these questions will be useful to<br />
guide the further development of interoperability artifacts To this end a conceptual framework to support<br />
such a review and evaluation of interoperability artifacts will be developed and presented in this paper.<br />
The paper is organized as follows: after the introduction the two key concepts of this paper public service<br />
information and interoperability artifact will be discussed and defined. The third section is dedicated to the<br />
elaboration and justification of the conceptual framework. A conclusion with future research questions<br />
finishes this paper.<br />
2. Definition of key concepts<br />
In this paper two key concept are used, and therefore have to be defined: public service information and<br />
interoperability artifact. The concept public service information shall be defined within this paper as<br />
information about public services. In the first place, this concept seems to be clear with this definition, but<br />
looking at this concept a little bit closer this concept will be become vaguer. Then what means public<br />
service and what is information about public services? The first question is difficult to answer, because<br />
what a public service is and shall be depends on the weltanschauung by the person defining the concept<br />
of public service. Thus, providing a definition, which is acceptable for everyone (Jäckering 1991: 18–19)<br />
is a difficult task, if not impossible. Hence and for pragmatic reason, public service will be defined very<br />
generally and with the aid of Lynn (2006: 24) as a service provided by an organization assigned to it by a<br />
sovereign to provide this service. And public service information will be defined as information about this<br />
service that give indication who provides this service, how this service can be accessed, which<br />
requirements have to be fulfill to access this service and so on. But as a consequence of such a definition<br />
of the concept of public service, the manifestations of public services cover a wide spectrum.<br />
Furthermore, this spectrum possesses a high dynamic, i.e., services sometime lose their characteristic to<br />
be a public service and will be provide by private organizations, or gain it, i.e., will be provide by a public<br />
organization instead of a private organization (see e.g. Fugge 2003).<br />
The second key concept of interoperability artifact has been used before by Fustar (2009: 1–2) in his<br />
Interoperability Meaning Diagram. In this diagram he relates the concept of interoperability artifact to<br />
concepts like standard semantic model, reference model, components and so on. As the discussion in<br />
the introduction indicates, a semantic model can be also seen as an artifact: a man-made tool to address<br />
a particular interoperability issue. Furthermore, Fustar does not give a detailed definition, what an<br />
interoperability artifact shall be. Hence, an interoperability artifact shall be defined as any artifact that is<br />
able to facilitate interoperability between information systems. This concept is defined in a wider sense as<br />
Fustar's Interoperability Meaning Diagrams indicates and fits also the widely used definition of<br />
interoperability as the capability of systems to exchange information (Institute of Electrical and Electronic<br />
Engineers 1990: 114). But within this paper only such interoperability artifacts will be considered that<br />
facilitate the interoperability of information systems managing or processing information about public<br />
services. In this sense, an artifact can be any application systems, software tools, system architecture<br />
modeling language, as well as analysis and design methods. These artifacts are characterized by a<br />
remarkable complexity by the underlying information technology, by a rapid technological change, and by<br />
a notional and conceptual diversity in naming and terms. The latter applies in particular for artifacts<br />
developed and used in the practice (Frank 2000: 39). If not stated otherwise and to simplifying the<br />
language the term interoperability artifacts will be used for interoperability artifacts used to exchange<br />
public service information within the further course of this paper.<br />
232
Veit Jahns<br />
3. Elaboration of the conceptual framework<br />
Usually evaluation is defined as determine the worth or the value of something using appropriate, explicated<br />
and justified criteria (Scriven 1991 in House 1993: 1). But speaking about worth and value, raises<br />
the question, for whom and to what end? And how should a framework used to perform the evaluation be<br />
designed and what are the reasons justifying the selected criteria? As it is the concern of this paper to<br />
evaluate artifacts, it should be drawn upon here on a proposal by Frank (2009: 41–43). As he is<br />
concerned in his paper with the evaluation of artifacts in the information systems research and which<br />
requirements should be fulfilled by an framework supporting the evaluation of artifacts, he proposes the<br />
following criteria (Frank 2009: 43):<br />
Goal-orientation: The objectives of the evaluation should be explicated and justified. In particular, to<br />
avoid or at least to reduce chance of an instrumentalistic fallacy (Albert 1960: 217), where the<br />
objectives—here the objective of an evaluation—is determined by the means—here the selected<br />
criteria (see e.g. Frank 2009: 37 or for a similar argumentation Streeten 1954).<br />
Seriousness: The scientific standard should not be sacrificed in favor of a high relevance to the<br />
practice. In particular, the criteria's operationalization should be made in a cautious manner.<br />
Justification: The proposed criteria should be justified in a comprehensible manner.<br />
Regarding the goal-orientation it is referred to the explanation and justification of this paper's objective<br />
given above. In the objective it is asked about interoperability artifacts, their scope of application as well<br />
as the underlying technologies and concepts. Regarding the seriousness of an evaluation of interoperability<br />
artifacts, it seems valuable to take a close look on the nature of interoperability artifacts<br />
regarding the exchange of public service information:<br />
Interoperability artifacts are created by humans in a development process. As a result of a development<br />
process the interoperability artifacts pass through a life cycle. They were created and<br />
applied, they mature and die, i.e., that they will not be used anymore. And because the development<br />
of interoperability artifacts can be considered as an eGovernment project, the development of an<br />
interoperability artifact has to deal also with the “ongoing changing pattern of relations or interactions<br />
between government organizations, business, and citizens” (Gupta and Jana 2003: 368).<br />
They describe more or less the public services. Thus, they are mostly some kind of models, codes,<br />
taxonomies, or in short: lingual artifacts.<br />
The subject—public services—of the interoperability artifacts considered in this paper, is characterized<br />
by a “fuzziness”, which makes it difficult to find the one interoperability artifact. Instead, a<br />
wide spectrum of manifestations of interoperability artifacts—as stated above—can be found.<br />
Due to the aspects given above many different perspectives and views have to be considered and/or<br />
incorporated into the interoperability artifact. But interoperability itself is a manifold phenomenon,<br />
which cannot be reduced to one single and isolated aspect (Dorloff et al. 2011).<br />
Considering these aspects, i.e., that interoperability artifacts are difficult to grasp and to quantify, that<br />
they are primarily lingual constructs as well as having Rowlands (2003: 3) characterization of qualitative<br />
research in mind that “qualitative research tends to work with text rather than numbers”, it seems appropriate<br />
to see interoperability artifacts as qualitative data and therefore to use an qualitative approach in<br />
this evaluation. But to give the qualitative evaluation a structure it will be recurred to an approach that has<br />
its origin more in the field of quantitative evaluation: the Goal/Question/Metric (GQM) paradigm (Basili<br />
1992). But here not the complete GQM paradigm will be applied, but rather its structure will be borrowed<br />
here.<br />
The GQM paradigm the procedure is that in a first step the objective or goal will be refined by deriving<br />
questions from the objective, and subsequently these questions will be made measurable by defining<br />
metrics. So, to refine the objective of this paper the questions will be formulated. To this particular end<br />
and as the issue of exchanging public service information between public authorities is related to interoperability<br />
of information systems a step back will be taken to look at this issue and to observed it from<br />
a greater distance. The step back taken is a step to the point of view by Dorloff et al. (2011). They<br />
examine in the paper the field of interoperability problems. To structure this field they refer to an idea<br />
once used by Wand and Weber (1995) and differentiate the phenomenon of interoperability from the<br />
social, communication, and the technological dimension (Dorloff et al. 2011: 5–6):<br />
233
Veit Jahns<br />
Social dimension: Here they refer to the aspects and characteristics of interoperability related to the<br />
way how humans experience and influence interoperability. As they consider interoperability from a<br />
wide perspective. Every aspect and characteristic of the social-economic reality is gathered here.<br />
Communication dimension: They consider in this dimension the linguistic aspects and characteristics<br />
of interoperability. Primarily, the aspects and characteristics of the artifacts created to achieve<br />
interoperable information systems.<br />
Physical dimension: In this dimension Dorloff et al. (2011) consider the technology used to achieve<br />
interoperability, e.g., the used protocols and middle-ware technologies.<br />
The categorization by Dorloff et al. (2011) based on the view, that interoperability is an object. As this<br />
object has different aspects and facets they approach this object by observing it from different perspectives<br />
each on highlighting some particular aspects and hiding others.<br />
Another and orthogonal approach to understand interoperability is, to understand interoperability as a<br />
process, i.e., that interoperability is established on a certain point of time, operated for a certain time span<br />
within some kind of collaboration until the collaboration ends and the interoperability will be dissolute.<br />
Based on this view Mallek et al. (2010: 441–446) developed an interoperability requirements model,<br />
which considers interoperability from the perspective of compatibility, inter-operation and reversibility<br />
requirements:<br />
As compatibility requirements they consider functionality, capability or characteristic of an information<br />
system that is related to its interoperability and is independent of time. Compatibility requirements<br />
need to be satisfied by organizations or its respective information systems to become interoperable.<br />
Inter-operation requirements are requirements related to the interactions of organization respective<br />
their information systems. How the interaction is performed and so on. So inter-operation<br />
requirements are dependent of time and primarily of a dynamic nature.<br />
Reversibility requirements refer to the end of a collaboration. In particular, this requirements aim to<br />
maintain the autonomy of an organization and reverse it to its original state after an collaboration has<br />
ended and no need to be interoperable with the collaboration organization is given. As the<br />
implementation of interoperability requires adaption and modification of the used information<br />
systems, the reversibility requirements aim to maintain the capability of an information system to<br />
return to its original state, e.g., by limiting the impact of the interoperability implementation on the<br />
information system itself.<br />
To combine the both views and to summarize the reflections above and to word the question it is referred<br />
to table 1. In this table the questions to be used to narrow down the objective of this paper are given.<br />
After the elaboration of the question, it is needed now to define the metrics. Instead of defining metrics for<br />
these questions the concept of morphology also called morphological box (Zwicky 1973: 94–100) is used<br />
here. This approach was originally a creative technique to find new and undiscovered instances of a<br />
given type of entities (solutions, types, artifact, etc.). It is looked for the essential characteristics of an<br />
entity type. The characteristics (rows) and its possible categories (columns) are arranged in a<br />
morphological matrix describing all possible objects of the entity type. By combining the applicable<br />
category—or categories—of each characteristic it is possible to describe the object considered object. At<br />
the first glance, it may be inappropriate here to use a creativity technique for an evaluation, but as the<br />
purpose of this evaluation is Also the concept of the morphological box has been successfully used before<br />
to examine similar research questions, e.g., by Dorloff et al. (2011) with respect to e-business<br />
interoperability. Furthermore and as the questions given above suggest, it is not the intention of this<br />
paper to perform a quantitative evaluation, but a rather to perform an evaluation in a qualitative way. The<br />
concept of morphology by Zwicky offers the opportunity to do this in a structured but also in a simple<br />
manner.<br />
By perusing the questions given above, the morphological box will be constituted by the following<br />
characteristics and categories:<br />
Question (1a) can be related to the communicating pattern between actors. These types of actors are<br />
can be classified as citizens (C), business (B), non-profit organizations (N) and public administration<br />
(A). As all these actors interact amongst themselves, only those interactions where the public<br />
234
Veit Jahns<br />
administration participates are considered here (Dorloff et al. 2011: 8; Sundgren 2005: 85; Wirtz<br />
2001: 34–39).<br />
Table 1: Evaluation questions<br />
Requirement<br />
Dimension<br />
Social Dimension (1a) Which type of actors<br />
exchange public service information?<br />
(1b) Which perspective of<br />
social actors are considered?<br />
(1c) Are legal rules taken into<br />
account?<br />
Communication<br />
Dimension<br />
Physical Dimension<br />
Compatibility Inter-operation Reversibility<br />
(4a) How is the concept of<br />
public service and its subconcepts<br />
defined in detail?<br />
(4b) What is the formalization<br />
degree of the artifact?<br />
(7a) What interoperability<br />
entities at the technological<br />
level are considered in the<br />
interoperability by the artifacts?<br />
(7b) Have and existing governmental<br />
interoperability<br />
frameworks been taken into<br />
account and if so, which one?<br />
(2) Does the artifact indicators,<br />
how this artifact<br />
improves the interoperability<br />
between two organizations<br />
respective information systems,<br />
e.g., by reducing costs,<br />
and how this improvement<br />
can be measures?<br />
(5a) What spectrum of the<br />
interaction process is<br />
covered by the interoperability<br />
artifact?<br />
(5b) Does it reflect the<br />
different types of information<br />
accrued during the interaction<br />
process?<br />
(8) Does the artifact indicators,<br />
how this artifact<br />
facilitate the implementation<br />
of interoperable information<br />
systems in public<br />
administrations?<br />
(3) What assumptions are<br />
made in the artifacts' creation<br />
process regarding the<br />
duration of the interaction?<br />
(6) Has the own terminology<br />
regarding public services to<br />
be adapted to the underlying<br />
terminology of interoperability<br />
artifact?<br />
(9) How tight is the integration<br />
of these artifacts intro the<br />
information systems?<br />
For question (1b) it referred to a distinction once made by Jahns et al. (2009: 359). They distinguished<br />
ontologies by the point of view from which the ontology was developed, i.e. the view of the<br />
public administration or the citizens’ respectively the business’ point of view.<br />
As the actions of the public administration is rule by law, which impedes the implementation of<br />
eGovernment services, the question (1c) aims at this fact, by asking, if legal rules are referable in the<br />
artifacts (Alpar and Olbrich 2005).<br />
The performance evaluation of interoperability artifact is a task of its own and not within the scope of<br />
this paper. But in the other side, many creators of interoperability artifacts promise that their<br />
interoperability artifact improves interoperability in one way or the other. So, it seems reasonable to<br />
ask, if indicators and possible metrics are provided with the interoperability artifacts. Like question<br />
(2), this is a binary question and two categories can be considered here: yes and no.<br />
From question (3) the characteristics of the interaction’s stability, i.e., if the interoperability artifacts<br />
tends more to support ad hoc interactions or long-term respectively permanent interactions.<br />
The type of definition of public service concept given in the respective approach can used to provide<br />
indications about the answer of question (4a). There can be found definitions based on public tasks<br />
as defined in public regulations like laws, or definition based on activities in public organizations<br />
(Jahns et al. 2009: 359).<br />
Regarding degree of formalization ask about in question (4b), there are three possible categories no<br />
formalization, i.e., using natural language, semi-formalization, .e.g., structured natural language or<br />
graphical models without formal semantics, and complete formalization with a well and formal defined<br />
semantic.<br />
For question (5a) and (5b) a characterization by Sundgren (2005: 90–92) of data processed in public<br />
information systems can be applied here: operative data, i.e., data that is necessary to process and<br />
235
Veit Jahns<br />
to complete as task within the provision of a public service, procedural data, i.e., descriptions or<br />
representations of rules related to the provision of public services, e.g., rules emanate from laws,<br />
directive or analytical data, i.e., statistics and other indicators regarding the state of the public services'<br />
provision, e.g., performance, meta-data, i.e., data describing the data itself, e.g., what format it<br />
is used to explicate the data, the source of the data and so on, process data, i.e., information<br />
gathered in the provision of the public services they inform about, e.g., archival data, i.e., data<br />
extracted from or generated during a public service provision to reflect or document the provision.<br />
Public authorities have their own special terminology regarding their field of activity. Also the<br />
interoperability artifacts require a fixed understanding of concepts they are dealing with. But is an<br />
adequate mapping between these terminologies possible or have the public authorities to adapt their<br />
terminology to the underlying terminology of the interoperability artifact? So, two answers to question<br />
(6) are possible here: yes and no.<br />
Question (7a) Here five categories of entities can be identified: the data types as the smallest and<br />
simplest entities dividing atomic data elements into different types and determining the allowed<br />
values of a data element, the vocabulary, i.e., the data elements defined on the basis of the data<br />
types, the documents as a set of logical related data, and the process as the sequences of<br />
exchanged documents (Dorloff et al. 2011: 10; Beckmann et al. 2004: 52–56).<br />
Question (7b) refers to sets of guidelines and regulations, e.g., the <strong>European</strong> Interoperability<br />
Framework (<strong>European</strong> Communities 2004) regarding the implementation of information systems in<br />
public authorities. Such regulations can be found on a national, international and global level.<br />
Similar to question (2), one can also ask with question (8), if indicators are provided allowing the<br />
determination, if and to what degree an interoperability artifacts facilitate the implementation of an<br />
information systems. As this is also a binary question the same two categories can be identified as in<br />
question (2).<br />
Refining question (9) two options are at hand here. Firstly, annotating existing public service<br />
information or export into a separate exchange format (Jahns et al. 2009: 359). Secondly, regarding<br />
the interoperability mechanism, i.e., achieving interoperability by standardization or by conversion.<br />
(Dorloff et al. 2011: 10; Wüstner et al. 2002: 55–57). Standardization means that allowing more<br />
individuality and therefore limits the impact on the information system and makes it easier to restore<br />
the original state. In contrast, standardization tends to have a stronger impact on information systems<br />
as its inner structure and processes may have to be changed to be compliant with the standard. A<br />
conversion approach on the other side allows preserving these inner structure and processes.<br />
These characteristics and their categories are summarized in table 2.<br />
Table 2: Morphological box for evaluation interoperability artifacts<br />
Characteristics <br />
Communicating<br />
Pattern<br />
Categories<br />
A2A A2B B2A A2C C2A A2N N2A<br />
View Citizens/Business Administration<br />
Considering<br />
Legal<br />
Frameworks<br />
Yes No<br />
Duration Ad hoc Long-term<br />
Public ServiceDefinition<br />
Degree of<br />
formalization<br />
Types of<br />
data<br />
Activity-based Task-based Not Explicated<br />
Natural Semi-formal Formal<br />
Operative Procedural Analytical Metadata Process data Archival data<br />
236
Characteristics <br />
Interoperabilityentities<br />
Considering<br />
Interoperability<br />
Frameworks<br />
Facilitation<br />
Indicators<br />
Interoperability<br />
mechanism<br />
4. Conclusion<br />
Veit Jahns<br />
Categories<br />
Data types Vocabulary Document Process<br />
National International Global<br />
Yes No<br />
Standardization Conversion<br />
Based on the premise, that there are many artifacts to support the interoperability of public administration<br />
with respect to the exchange of public service information, a conceptual framework for the evaluation of<br />
these artifacts was developed. The development of this framework was completely based on theoretical<br />
reflections on this issue. What is missing up to now is the application and use of this conceptual<br />
framework by the means of real interoperability artifacts? This will be the next step will. The framework<br />
will be applied to evaluate the existing artifacts created in research and implementation projects, e.g.,<br />
those mentioned at the beginning of this paper. But performing the evaluation will not only prove—or<br />
disprove—the framework itself. As the used perspectives to develop the framework highlight particular<br />
aspects of interoperability, and obscure others, performing the evaluation can reveal further perspectives<br />
and aspects of interoperability, which should be considered in an evaluation also. Furthermore, the result<br />
of the evaluation can be helpful to integrate these interoperability artifacts and reveal new research<br />
directions, as well as can be—a minor hope of the author—a contribution to the efforts (e.g. Charalabidis<br />
et al. 2010) to establish a science of interoperability or at least give interoperability a more profound scientific<br />
base.<br />
References<br />
Albert, H. (1960) “Wissenschaft und Politik. Zum Problem der Anwendbarkeit einer wertfreien Sozialwissenschaft”, in<br />
Topitsch, E. (ed.) Probleme der Wissenschaftstheorie. Festschrift für Victor Kraft, Springer, Wien.<br />
Alpar, P. and Olbrich, S. (2005) “Legal Requirements and Modelling of Processes in eGovernment”, Electronic<br />
Journal of eGovernment, Vol. 3, No. 3, pp. 107–116.<br />
Basili, V. R. (1992) Software Modeling and Measurement: The Goal/Question/Metric Paradigm, [online], Available:<br />
http://www.cs.umd.edu/~basili/publications/technical/T78.pdf [06 Nov 2010].<br />
Beckmann, H., van der Eijk, P., Hjulstad, H., Madsen, B. N., Ondracek, N. and Schmitz, V. (2004) CEN Workshop<br />
Agreement 15045. Multilingual Catalogue Strategies for eCommerce and eBusiness, [online], Available:<br />
ftp://ftp.cenorm.be/PUBLIC/CWAs/e-Europe/eCat/CWA15045-00-2004-Jul.pdf [21 Nov 2010].<br />
Charalabidis, Y., Gonçalves, R. J. and Popplewell, K. (2010) “Developing a Science Base for Enterprise<br />
Interoperability”, in Popplewell, K., Harding, J., Poler, R. and Chalmeta, R. (ed.) Enterprise Interoperability IV.<br />
Making the Internet of the Future for the Future of Enterprise, Springer, London.<br />
Dorloff, F.-D., Jahns, V. and Schmitz, V. (2011) “E-Business Interoperability: A Systematization Attempt Based on<br />
the Morphology Concept”, [forthcoming], in Kajan, E. (ed.) Electronic Business Interoperability: Concepts,<br />
Opportunities and Challenges, IGI Global, Hershey.<br />
<strong>European</strong> Communities (2004) <strong>European</strong> Interoperability Framework for Pan-european EGovernment Services,<br />
[online], Available: http://ec.europa.eu/idabc/servlets/Docd552.pdf?id=19529, [21 Nov 2010].<br />
<strong>European</strong> Union (2009) Ministerial Declaration on eGovernment, [online], Available: http://www.se2009.eu/polopoly_fs/1.24306!menu/standard/file/Ministerial%20Declaration%20on%20eGovernment.pdf<br />
[15 Oct 2010].<br />
Frank, U. (2000) “Die Evaluation von Artefakten: Eine zentrale Herausforderung der Wirtschaftsinformatik”,<br />
Häntschel, I. and Heinrich, L. (ed.) Evaluation und Evaluationsforschung in der Wirtschaftsinformatik,<br />
Oldenbourg, München.<br />
Fugge, F. (2003) “Administrative Traditions in Western Europe”, in Peters, B. G. and Pierre, J. (ed.) The SAGE<br />
Handbook of Public Administration, Sage, London.<br />
Fustar, S. (2009) “A Generic Approach for Compliance Assessment of Interoperability Artifacts”, Grid Interop 2009<br />
Proceedings. The Road to an Interoperable Grid, Denver, CO.<br />
Garfield, E. (1977) “Why don't we have science reviews?”, in Essays of an Information Scientist, Vol. 2 (1974-1976),<br />
ISI Press, Philadelphia.<br />
237
Veit Jahns<br />
Gupta, M. P. and Jana, D. (2003) “EGovernment Evaluation: A Framework and Case Study”, Government<br />
Information Quarterly, Vol. 20, No. 4, pp. 365–387.<br />
House, E. R. (1993) Professional Evaluation. Social Impact and Political Consequences, Sage, Newbury Park.<br />
Institute of Electrical and Electronic Engineers (1990) IEEE Standard Computer Dictionary. A Compilation of IEEE<br />
Standard Glossaries, New York.<br />
Jäckering, W. (1994) “Verwaltung und öffentliche Aufgaben”, in Mattern, K.-H. (ed.) Allgemeine Verwaltungslehre, 4 th<br />
edition, Walhalla und Praetoria, Berlin.<br />
Jahns, V., Dorloff, F.-D. and Schmitz, V. (2009) “Harmonising and Exchanging Public Service Information Between<br />
Public Service Centres. Requirements, Constraints and Outline of a Solution”, in Hahamis, P. (ed.) Proceedings<br />
of the 9 th <strong>European</strong> <strong>Conference</strong> on eGovernment, London, pp. 377–384.<br />
Lynn, L. E. (2006) Public Administration. Old and New, Routledge, New York.<br />
Mallek, S., Daclin, N. and Chapurlat, V. (2010) “Towards a Conceptualisation of Interoperability Requirements”, in<br />
Popplewell, K., Harding, J., Poler and R. Chalmeta, R. (ed.) Enterprise Interoperability IV. Making the Internet of<br />
the Future for the Future of Enterprise, Springer, London.<br />
Rowlands, B. (2003) “Employing Interpretive Research to Build Theory of Information Systems Practice”, in<br />
Australasian Journal of Information Systems, Vol. 10, No. 2, pp. 3–22.<br />
Scriven, M. S. (1991) Evaluation Thesaurus, 4 th edition, Sage, Newbury Park.<br />
Streeten, P. (1954) “Programs and Prognoses”, The Quarterly Journal of Economics, Vol. 68, No. 3, pp. 355–376.<br />
Sundgren, B. (2005) “What is a Public Information System?”, in International Journal of Public Information Systems,<br />
Vol. 1, No. 1, pp. 81–99.<br />
Wand, Y. and Weber, R. (1995) “On the Deep Structure of Information Systems”, Information Systems Journal, Vol.<br />
5, No. 3, pp. 203–223.<br />
Wirtz, B. W. (2001) Electronic Business, 2 nd edition, Gabler, Wiesbaden.<br />
Zwicky, F. (1973) Entdecken, Erfinden, Forschen im morphologischen Weltbild, Droemer Knaur, München.<br />
Wüstner, E., Hotzel, T. and Buxmann, P. (2002) „Converting Business Documents: A Classification of Problems and<br />
Solutions Using XML/XSLT“, in Proceedings of the 4 th IEEE International Workshop on Advanced Issues of E-<br />
Commerce and Web-Based Information Systems, Newport Beach, CA, pp. 54–61.<br />
238
Free and Open Source Software for Public Sector Enterprise<br />
Applications in Sri Lanka<br />
Srimal Jayawardena 1 and Gihan Dias 2<br />
1 Australian National University, Canberra, Australia<br />
2 University of Moratuwa, Sri Lanka<br />
srimal.jayawardena@anu.edu.au<br />
gihan@uom.lk<br />
Abstract: This paper describes a study done on the use of Free and Open Source Software (FOSS) for enterprise<br />
applications of the public sector in Sri Lanka. The study investigates factors that affect the selection of software for<br />
public sector information systems with, an emphasis on how Free and Open Source Software (FOSS) is used and<br />
could be used in such projects. The objectives of the study are as follows. The study aims to identify factors that affect<br />
the adoption of Software in Sri Lankan government sector projects. In addition to this it identifies features that<br />
are considered as important in such large Information Systems, and how they relate to Open Source Software adoption<br />
in the context of the public sector in Sri Lanka. The findings are analyzed and suitable recommendations are<br />
presented for better selection of software in the public sector. The work is important as it identifies and highlights<br />
factors that affect the choice of software in the public sector. This is important for several categories of people. It is of<br />
importance to the strategic management and policy makers to know what drives the information system procurement<br />
decisions in order to make more relevant policies and guidelines that are congruent with the needs in government<br />
sector departments. It is useful when advocating new technologies and information systems. This is especially true<br />
when advocating the use of Free and Open Source software for the use in the public sector in a wholesome and sustainable<br />
manner. The research is also important to software vendors and solution providers to the public sector in<br />
identifying what factors need to be taken in to account when bidding for public sector IS projects. The research is of a<br />
qualitative nature. It consists of multiple case studies of selected government sector departments and projects in Sri<br />
Lanka. The study investigates information systems developed by internal staff, developed by external consultants,<br />
procured systems, system implementation via private public partnerships (PPP) and projects guided by the Information<br />
and Communication Technology Agency (ICTA) of Sri Lanka. Data was gathered through interviews of staff at<br />
different levels of selected government sector Information Systems projects. The data was comparatively analyzed<br />
on a case by case basis to identify common patterns and trends among the investigated organizations and projects.<br />
The identified factors affecting the choice of software include the cost, technical specifications, bidder’s expertise in<br />
selected technology and user’s expertise in selected technology amongst other factors described in the full paper.<br />
The study identifies patterns between the choice of software - FOSS or non FOSS, and these identified factors.<br />
Based on these, recommendations are given to adopt and benefit from the use of FOSS in public sector enterprise<br />
level software projects.<br />
Keywords: enterprise software, government, free and open source software, software adoption<br />
1. Introduction<br />
We introduce the work to the reader by describing the research problem and objectives of the research.<br />
The research methodology, scope and nature of the study are briefly highlighted before being discussed<br />
in detail in the sections which follow.<br />
1.1 The research problem<br />
Government sector Information Systems projects guided by the ICTA (ICTA 2008) are a part of the e-<br />
Development initiative of Sri Lanka (e-Sri Lanka), which is funded primarily by the World Bank. The World<br />
Bank publication by Hanna (2006) describing the experiences of envisioning and designing of the Sri<br />
Lankan e-Development program highlights that it can benefit from the use of FOSS. Apart from this, the<br />
literature surveyed has highlighted how organizations could benefit from the use of FOSS.<br />
However, to realize the benefits of FOSS, decision makers should perceive it as a viable option. Hence it<br />
is important to identify what affects the choice of software at a strategic level. This leads us to the problem<br />
statement of this research.<br />
“What factors influence the choice of software for government sector enterprise applications in Sri<br />
Lanka?”<br />
Complexity. According to (Lewin 2000), a complex system is one whose properties are not fully explained<br />
by an understanding of its parts.<br />
This problem statement is further elaborated by the research objectives explained below.<br />
239
1.2 Research objectives<br />
Srimal Jayawardena and Gihan Dias<br />
Based on the above problem statement, this research consists of the following objectives.<br />
1. To identify factors that affects the adoption of Software in Sri Lankan government sector projects.<br />
2. To identify what features are considered as important in large Information Systems, and how they<br />
relate to Open Source Software.<br />
3. To develop suitable recommendations for the better selection of software in the Sri Lankan government<br />
sector.<br />
1.3 Significance of the study<br />
The study identifies the factors that affect the adoption of software for enterprise applications. Moreover,<br />
it focuses on those factors that are peculiar to the government sector in Sri Lanka. It also identifies what<br />
aspects of ISs are taken in to account when different government sector institutions evaluate an IS. The<br />
study will help formulate polices that will aid government sector organizations to leverage on the benefits<br />
of FOSS as explained in section 2.<br />
1.4 Research methodology<br />
The research methodology involved a rigorous study of selected government organizations as out lined in<br />
the scope of the study. The research was designed based on the findings of the literature study and the<br />
research objectives. Data collection was carried out using personal interviews based on this research<br />
design. The obtained data has been analyzed in detail in section 4. The research methodology is described<br />
in depth in section 3.<br />
1.5 Scope of the study<br />
The scope of the research is limited to government sector ISs. Government projects coordinated through<br />
the Information and Communication Agency (ICTA) of Sri Lanka and other selected government organizations<br />
have been taken as case studies (Appendix A & B).<br />
1.6 Data collection<br />
The organizations covered included, those with ICTA guided projects as well as organizations without an<br />
ICTA involvement for comparison purposes (Appendix A). The data was collected by interviewing staff at<br />
different levels in the selected organizations (Appendix A). This included personnel at a strategic level<br />
and heads of the IT/IS departments. In the case of ICTA guided projects the interviews consisted of the<br />
ICTA project managers and relevant external stakeholders. The information systems (IS) projects covered<br />
in the research are indicated in Appendix B. Interviewed personnel and sample interview questions<br />
are included in Appendices C and D.<br />
1.7 Data analysis<br />
The data was comparatively analyzed on a case by case basis to identify common patterns and trends<br />
among the investigated organizations and projects.<br />
1.8 Nature and form of results<br />
The results of the research are qualitative in nature. It is a comprehensive study of the factors that affect<br />
the adoption of FOSS in the Sri Lankan government sector organizations for enterprise systems. The<br />
results also include suitable recommendations for capitalizing on the advantages of FOSS in the Sri<br />
Lankan government sector.<br />
1.9 Structure of this document<br />
The remainder of this paper is organized as follows. A background to the problem is presented which is<br />
followed by an explanation of the research methodology employed. This followed by an analysis of the<br />
investigation outcomes which are followed by conclusions and recommendations.<br />
240
2. Background<br />
Srimal Jayawardena and Gihan Dias<br />
We present a background to the work by consolidating literature surveyed on Free and Open Source<br />
Software (FOSS), on the use of FOSS in the government sector and on FOSS adoption models. FOSS<br />
is claimed to have a unique blend of benefits that have aroused the interest of organizations across the<br />
world. Some of these benefits have been reported to be as follows.<br />
A Total Cost of Ownership (TCO) study conducted by the Robert Francis Group (Orzech 2002) reveals<br />
that the use of FOSS computer operating systems like GNU/Linux can lower TCO up to 40%.<br />
Further to this, Gartner has reported a cost reduction of 15% (Maguire 2003) when using FOSS Operating<br />
Systems as opposed to Windows XP, which is a proprietary Operating System. A study done<br />
by Merill Lynch (Lemos 2003), a major financial management company, has revealed that TCO reductions<br />
by the use of FOSS are not only due to software licensing costs but from personnel and<br />
hardware costs.<br />
A recent e-Primer of the United Nations Development Program (Wong 2004) mentions enhanced<br />
security and vendor independence as other benefits of the use of FOSS in government applications.<br />
2.1 Free and Open Source Software (FOSS)<br />
According to the Wikipedia (2005), Free and Open Source Software are software that is licensed such<br />
that its human readable source code is freely available for implementers to study, change and improve its<br />
design.<br />
Research (Reddy and Evans, 2002) reveals that more than 48% of the government institutions in<br />
Chile use FOSS with at least another 11% planning to do so. Out of this 93% of the institutions use<br />
FOSS for server side applications. According to the findings of Chan (2007), several developing governments<br />
like Peru have attempted to adopt FOSS at a national level.<br />
A recent online survey done by the CIO magazine (Schindler 2008) during the period April 28, 2008<br />
to May 2, 2008 shows that more than 50% of the participants are using FOSS. Another 10% of the<br />
participants are planning to use FOSS during the next 12 months. Therefore there appears to be a<br />
growing interest in the use of FOSS across the world.<br />
10.00%<br />
53.00%<br />
Figure 1: Survey conducted by the CIO magazine<br />
37.00%<br />
241<br />
Currently<br />
using<br />
Planning to<br />
use next 12<br />
months<br />
No plans to<br />
use open<br />
source<br />
applications
2.2 Success in Enterprise Systems<br />
Srimal Jayawardena and Gihan Dias<br />
Several models have been developed to asses the success of Enterprise Systems. Out of these the<br />
popular D & M model (DeLone and McLean 2003) considers the success of the system based in the information<br />
quality, system quality and the service quality. The model strives to analyze the associations<br />
between these dimensions and the intention to use, the user satisfaction and net benefits. A pictorial representation<br />
of the updated D & M model is shown in the figure 2.<br />
Figure 2: The updated D & M model<br />
The Enterprise System Success Model (Gable et al 2003) presents a temporal analysis of the success<br />
based on what is to date and what happens in the future. The success to date is said to be influenced by<br />
the organizational impact and the individual impact. The future success is said to be influenced by the<br />
system quality and the information quality. A pictorial representation of the Enterprise Success Model is<br />
given in figure 3.<br />
2.3 Use of FOSS in the government sector<br />
Government sector software applications require long term continuity and a lower total cost of ownership<br />
(TCO) while maintaining a high degree of quality and integrity. Mitch Stoltz (1999) mentions that governments<br />
can use open source as a vehicle for promoting economic development.<br />
Some of the benefits of the use of FOSS in the government sector have been identified by Reddy and<br />
Evans (2004) as follows.<br />
2.3.1 Cost savings<br />
Significant cost savings can be made due to the lack of expensive licensing fees in FOSS (Reddy and<br />
Evans 2004). However this factor alone does not justify the use of FOSS.<br />
2.3.2 Security, stability and privacy<br />
As the source code is open for inspection, it is claimed that FOSS is more secure and stable. According<br />
to Reddy and Evans (2004), the German government was concerned that some versions of the proprietary<br />
Microsoft Windows operating system contained back doors designed to grant the US National Security<br />
Agency, access to user data. Based on this rationale, the government of Germany has adopted a<br />
strong FOSS policy.<br />
2.3.3 Independence<br />
Proprietary software solutions are often based on one or few software companies. Hence a government<br />
implementing such proprietary solutions will have to depend on those few companies for support and<br />
business continuity. Due to the open nature of FOSS, dependence on software companies is less.<br />
242
Figure 3: The enterprise success model<br />
Srimal Jayawardena and Gihan Dias<br />
2.3.4 Helping domestic industries<br />
The open nature of FOSS will help the domestic software industry to actively contribute in developing<br />
applications. This is important for the long term sustainability of government projects and in terms of national<br />
interests.<br />
2.3.5 Innovation<br />
The open nature of FOSS helps innovation. This results in more advanced and efficient software which<br />
are released faster than proprietary software which have to adhere to conventional release cycles.<br />
Figure 4: Kwan and West (2005) FOSS adoption model<br />
2.4 Open Source adoption models<br />
Kwan and West (2005) of the Silicon Valley Open Source Research Project, San José State University,<br />
have presented the following Conceptual Model for Enterprise Adoption of Open Source Software (Figure<br />
4).<br />
The model assumes a three phase process:<br />
2.4.1 Policies<br />
The firm establishes a series of policies (formal or informal) that guide its procurement decisions.<br />
243
2.4.2 Choice set<br />
Srimal Jayawardena and Gihan Dias<br />
From the range of available options, at the first decision point the firm identifies those acceptable alternatives<br />
that meet minimum requirements.<br />
2.4.3 Selection<br />
From the range of acceptable alternatives, at the second decision point, the firm makes its final selection<br />
decision.<br />
The following factors are said to influence the adoption decisions.<br />
Features<br />
These are the attributes commonly used to measure what is new or valuable about a given technology,<br />
and used by vendors to differentiate their products. Because anything could be tautologically classified as<br />
a “feature,” we define feature as the residual of the two remaining categories<br />
Risk<br />
Risk for a system is often thought of in terms of reliability — the comparative scarcity of crashes, failures,<br />
or data loss. However, risk-lowering measures might also include efforts to mitigate the effect of inevitable<br />
if rare failures (such as redundant data or 7/24 support). Firms may also consider the risks of their<br />
investments over time, such as the risk that the technology may be orphaned by the vendor (either deliberately<br />
or with the vendor’s bankruptcy), or that for some other reason the investment is rendered obsolete.<br />
Cost<br />
These may include both the initial purchase price, and the ongoing usage costs such as support contracts<br />
and upgrade fees. An attempt to calculate total cost of ownership would also include personnel<br />
cost, and (for large data centers) related equipment costs such as power, air conditioning, security, etc.<br />
The following factors have been identified by Dedrik and West (2003) as open source adoption factors in<br />
their grounded theory based research.<br />
Willingness to take risks on a new, unproven technology.<br />
Need for organizational slack to evaluate the new technology and to self-support unsponsored technologies.<br />
Tendency of open source software to be inexpensive if not free and the inherent trialability of “free”<br />
software distributed on the Internet.<br />
Availability of external sources of support and expertise.<br />
Having set the background to the research with the aid of relevant literature, we move on to describe the<br />
research methodology employed in the section which follows.<br />
3. Research methodology<br />
The study intends to investigate a selected set of government IS in considerable depth. Therefore a case<br />
study based approach was adopted. As suggested by Yin (2003) the overall evidence from a multiple<br />
case study could be considered to be more compelling and robust, when it can be afforded to. Therefore,<br />
selected enterprise IS of selected government organizations were taken as case studies. Therefore the<br />
study is mostly qualitative in nature. The data collection was done through in-depth interviews. Staff at<br />
different levels of the organization has been interviewed as appropriate in order to obtain a more holistic<br />
perspective. The selected government organizations (Appendix A), the IS projects (Appendix B), personnel<br />
interviewed (Appendix C) and the interview outline with sample questions (Appendix D) are included<br />
at the end of the paper.<br />
244
3.1 Scope of the research<br />
Srimal Jayawardena and Gihan Dias<br />
The research focuses on IS pertaining to a selected set of government organizations. The sample set<br />
consists of ICTA guided IS projects as well as non-ICTA IS projects in order to facilitate comparative<br />
analysis. The list of organizations and the IS projects included in the study are indicated in Appendix A<br />
and B.<br />
3.2 Interview design<br />
In depth interviews were carried out based on a broad outline. The structure of the interviews was designed<br />
to bring out the factors that affect the choice of software in the organizations concerned. In particular,<br />
the selection or the absence of FOSS has been investigated. The interviews investigate the<br />
source of the requirements specifications along with the nature of the technologies and factors that influenced<br />
the IS solution. The IS solution was examined in terms of its modules which included different<br />
combinations of FOSS and proprietary technologies. This ranged from totally FOSS or proprietary systems<br />
to IS with a combination of both. Issues and room for improvement in the selected IS have also<br />
been investigated. The interview outline covers the following aspects.<br />
3.2.1 A general investigation<br />
The main projects, the project objectives and components were identified in this section.<br />
3.2.2 Technology and components<br />
The software, hardware and middle ware components used for each project were identified. The use of<br />
proprietary and FOSS components were investigated and identified.<br />
3.2.3 Specification development and selection<br />
The nature of the specifications development for each IS was identified. This section investigates if specifications<br />
were developed internally, by an external consultant and in the case of ICTA guided projects,<br />
the involvement of a SAGE.<br />
3.2.4 Factors affecting the selection<br />
This section investigates the key factors affecting the selection of the IS system concerned. In the case of<br />
a procured solution the factors affecting the selection process were investigated. It was also investigated<br />
if the interviewees had any recommendations to improve the current selection process. If the IS was developed<br />
by internal staff, the reasons behind this decision was investigated.<br />
3.2.5 Issues, improvements and FOSS<br />
For ISs that have already been fully implemented, the issues and room for improvement in those systems<br />
were investigated. Also the management inclination towards using FOSS components as a viable option<br />
to resolve those issues and to make improvements was investigated. The perceived areas where FOSS<br />
could be used most advantageously have been investigated.<br />
The data collected using the described research methodology is analyzed in the next section.<br />
4. Analysis<br />
The following factors have been broadly identified to affect the selection of IS's, as a result of the in-depth<br />
interviews carried out during the study.<br />
Cost<br />
Technical Specifications<br />
Usability<br />
Bidders Expertise in selected Technology<br />
Users Expertise in selected Technology<br />
Interoperability and Open Standards<br />
245
Trilingual Support<br />
Support and Maintenance<br />
Security<br />
Srimal Jayawardena and Gihan Dias<br />
It has been attempted to analyze the factors based on how the requirements specifications were developed.<br />
The following categories have been identified during the study as possibilities of developing the<br />
requirements specifications.<br />
Developed by internal Staff<br />
Developed by External Consultant<br />
ICTA Guided<br />
Private Public Partnerships (PPP)<br />
4.1 Analysis of the factors<br />
The identified factors can be analyzed from several perspectives. An analysis of the 4 most commonly<br />
mentioned factors is shown in Figure 5. Based on this, organizations that considered ‘Bidders/Developers<br />
expertise in the selected technology’ as an important factor had a significantly lower percentage of FOSS<br />
systems as opposed to the others. Therefore, we see that this factor is a strong deterrent for using FOSS<br />
in the organizations investigated.<br />
Figure 5: Factor summary showing the use of FOSS and proprietary IS<br />
4.1.1 Cost<br />
Almost all investigated organizations have indicated the cost of the IS as a decisive factor. Almost half<br />
(43%) of investigated projects where the cost was considered as an important factor were for systems<br />
that were developed in-house. Next in line were ICTA guided projects which composed of 11% of the<br />
population. A few interesting points have been noted when analyzing the cost as a factor.<br />
Total Cost vs. IS Cost<br />
246
Srimal Jayawardena and Gihan Dias<br />
Certain projects, where the cost of the IS is small compared to the capital cost of the entire project have<br />
placed a lower importance on the cost of the IS. For example, in the CEB power systems the total cost of<br />
a power station is very large when compared to its IS component.<br />
80-20 Rule<br />
In most ICTA guided projects where the cost is a decisive factor, the cost of the projects has been given<br />
prominence in the requirements specification by using the 80-20 rule. This is where the final selection is<br />
done by allowing an 80% weight on the cost of the bid while a weight of 20% is given on technical specifications.<br />
This rule has been used to indicate to bidders that cost is an important factor in the selection<br />
process. In such situations, certain vendors have considered FOSS components in the proposed solutions.<br />
4.1.2 Technical compliance<br />
Technical compliance has been reflected in most projects by indicating a minimum level of compliance.<br />
Other factors have been considered subject to this minimum level of technical compliance. Certain projects,<br />
however, have specific technical requirements.<br />
Compliance with existing technology<br />
Certain projects impose technical constraints as it is required to migrate data that is already stored using<br />
a particular technology. For example the Department of Education has had its data in an IBM mainframe<br />
since 1974. This system was migrated to the IBM AS/400 platform in 1998 as the migration of the past<br />
data was easier within the selected platform.<br />
4.1.3 Usability<br />
Usability has been identified as a decisive factor in projects developed internally and in procured systems.<br />
However, usability was not explicitly highlighted as a decisive factor in the other categories of<br />
specification development.<br />
4.1.4 Implementers and specifications developers influence<br />
We see in figure 6 that different methods of specifications development and implementation have resulted<br />
in varying levels of FOSS and proprietary software being used in the resulting IS. The largest percentage<br />
of FOSS usage is reported from ICTA guided government enterprise IS projects. This reflects an<br />
increasing FOSS awareness within the ICTA. Other methods of implementations have lower levels of<br />
FOSS – lower than 50%. It is interesting to note that the PPP category consists entirely of proprietary<br />
software. Salient aspects of each of these categories are discussed next.<br />
PPP Projects<br />
In the Government Information Center (GIC) project, the private partner organization maintained that the<br />
single most reason for their choice for a particular set of proprietary technologies was because their expertise<br />
lay in those technologies.<br />
ICTA Guided Projects<br />
The Lanka Government Network (LGN) project is facilitated via a Korean government loan. The loan<br />
conditions dictate that the bidders of the project are required to be Korean. The expertise of these bidders<br />
has influenced the ultimate choice of technology for this project, although the ICTA has not limited the<br />
requirements to technology in particular. Also the degree on inter dependency between components of<br />
this project limit the freedom of choice given that a particular component has been implemented by a certain<br />
vendor using a proprietary solution.<br />
External Consultant<br />
The same notion is reflected under this category as well. It was found that a particular MIS at the Tertiary<br />
and Vocational Education Commission has been developed in a certain technology as the consultant<br />
247
Srimal Jayawardena and Gihan Dias<br />
concerned was familiar with it. The system has had problems in scaling up to new requirements that have<br />
evolved and a new MIS is being implemented at a larger scale.<br />
Figure 6: Specifications developer / Implementer effects on the IS<br />
4.1.5 Users expertise in selected technology<br />
The expertise of existing users in a given technology has been found to be an important factor in certain<br />
ICTA guided projects. This is particularly so in the LGN project which also relies on the expertise of the<br />
existing staff in government organizations to maintain the infrastructure installed at those organizations.<br />
4.1.6 Interoperability and open standards<br />
Compliance with open standards and interoperability has been identified as important factors in certain<br />
ICTA guided projects.<br />
4.1.7 Trilingual support<br />
Trilingual support has been highlighted as an important factor in most recent government sector projects<br />
that involve the general public as users. This was especially noted in the design of the E-Local Government<br />
project.<br />
4.1.8 Support and maintenance<br />
Support and maintainability of the IS projects have been especially highlighted as an important factor in<br />
the internally developed category. In the TVEC MIS project for example, the consultant who implemented<br />
the initial MIS had been unable to provided support and maintenance of the system after a certain point.<br />
Hence the management had decided that the future MIS should be implemented with internal expertise<br />
using an open technology, so that the system can be maintained by the internal staff.<br />
4.1.9 Security<br />
Security of the IS has been explicitly identified as an important factor in the Examination Information System<br />
of the Department of Education. In fact, security has been a strong factor for the choice of the particular<br />
technologies used to implement the system – namely IBM AS/400 based technologies.<br />
Having analyzed the data as explained above, we draw the following conclusions.<br />
248
5. Conclusions<br />
Srimal Jayawardena and Gihan Dias<br />
We conclude from the analysis that although a considerable number of government organizations perceive<br />
the benefits of using FOSS, the following factors prevent FOSS adoption.<br />
5.1 Lack of FOSS expertise in companies that bid for government projects<br />
A significant reason for not using FOSS based IS in the government sector was the fact that companies<br />
bidding for such projects did not propose such solutions. It is only recently that companies bidding for<br />
government projects have started to integrate FOSS components in the proposed solutions, to lower cost<br />
and leverage on the other benefits of FOSS.<br />
5.2 Lack of internal expertise in FOSS<br />
Certain organizations that were investigated noted that they do not have the expertise to implement and<br />
maintain FOSS IS internally. Therefore there is reluctance towards implementing FOSS solutions.<br />
5.3 Lack of FOSS support companies<br />
In several organizations, FOSS solutions have been implemented by internal software developers. A<br />
cause for concern was the lack of FOSS support companies to obtain support services. Often the FOSS<br />
expertise was from a few key developers in the staff who have had an exposure to FOSS.<br />
5.4 Use of unlicensed proprietary software<br />
The use of unlicensed proprietary software makes the cost unrealistically lower than when using legal<br />
licensed versions. This is a disincentive to consider FOSS alternatives which would have been otherwise<br />
cheaper.<br />
6. Recommendations<br />
Based on these conclusions, we make the following recommendations, in order to leverage on the benefits<br />
of FOSS in government sector enterprise software applications.<br />
6.1 Nurture a FOSS ecosystem<br />
FOSS service provider companies should be nurtured and encourage to support government IS development.<br />
This should encourage places where FOSS expertise exists, for example Universities, to provide<br />
FOSS services to government IS projects. Since, government IS will have long term benefits by leveraging<br />
on FOSS components, nurturing a FOSS ecosystem with companies that provide FOSS solutions will<br />
be a win-win situation for the government as well as for those companies.<br />
6.2 Developing internal expertise<br />
Internal expertise within the government organizations should be developed. IT personnel within government<br />
organizations should be trained to be FOSS savvy. Technical expertise within the organizations<br />
should be capable of maintaining FOSS IS. The following aspects could be considered in achieving this.<br />
6.2.1 Knowledge sharing<br />
A knowledge sharing mechanism among government (and other) organizations should be encouraged.<br />
This is to facilitate support and maintenance experience of FOSS solutions which would otherwise remain<br />
in silos.<br />
6.2.2 Government IT service<br />
With the increasing use of IT/IS within the government sector, a government IT service should be developed<br />
that could cater to the IS needs of government organizations. The personnel in this service should<br />
be trained to be FOSS savvy. There should be close liaison between this service and the private FOSS<br />
solution companies.<br />
249
6.3 Licensed software<br />
Srimal Jayawardena and Gihan Dias<br />
Licensed software should be insisted on when proprietary software components are used in government<br />
enterprise software applications. This will safeguard organizations from litigation when the legal framework<br />
develops and is capable of prosecuting the use of illegal software. Insisting on licensed software will<br />
make the costs of proprietary software components reflect realistically in the procurement process.<br />
Therefore, the monetary benefits of any comparable FOSS components will be more obvious.<br />
6.4 Policies in general<br />
Government organizations should encourage FOSS friendly policies in IS procurement policies. FOSS<br />
experience of companies is relatively young as FOSS itself is comparatively new. Government tender<br />
policies should accommodate this fact, if FOSS based solutions are to be encouraged. Also, open standards<br />
and non restrictive policies should be encouraged.<br />
List of Acronyms<br />
CIO – Chief Innovation Officer<br />
CRM – Customer Relationship Management<br />
FOSS – Free and Open Source Software<br />
HRM – Human Resource Management<br />
ICTA – The ICT Agency of Sri Lanka<br />
IS – Information System<br />
LGN – Lanka Government Network<br />
MIS – Management Information System<br />
PPP – Public Private Partnership<br />
SAGE – Software Architecture Group of Experts<br />
Appendix A<br />
Organizations investigated in the research are as follows.<br />
Table 1: Organizations investigated<br />
Organization Code<br />
Ceylon Electricity Board CEB<br />
Department of Agriculture DeptAgri<br />
Department of Education DeptEdu<br />
Department of Local Governments LocalGov<br />
ICT Agency of Sri Lanka ICTA<br />
Ministry of Public Administration PubAdmin<br />
Sri Lanka Customs Customs<br />
State Engineering Corporation SEC<br />
Tertiary and Vocational Education Commission TVEC<br />
Appendix B<br />
List of IS projects investigated are as follows. Please refer Appendix A for organization codes.<br />
250
Table 2: Investigated IS projects<br />
Appendix C<br />
Srimal Jayawardena and Gihan Dias<br />
Project Organization<br />
Accounting and Inventory System CEB<br />
Bill of Quantities System SEC<br />
Call Center DeptAgri<br />
Consumer Billing System CEB<br />
CRM System GIC/ICTA<br />
E-Learning System TVEC<br />
E-Learning System DeptAgri<br />
EHRM System ICTA/ PubAdmin<br />
eLocal Government Project ICTA/LocalGov<br />
ePensions Project ICTA/PensionsDept<br />
eSamurdhi Project ICTA/SamurdhiAuthority<br />
Examination Processing System DeptEdu<br />
Goods Processing System Customs<br />
Grama Niladari HR System PubAdmin<br />
Inventory System SEC<br />
LGN Phase 1 ICTA<br />
Mail Server SEC<br />
Management Information System TVEC<br />
Motor Vehicle Records System Customs<br />
Payroll System SEC<br />
Power System Control Systems CEB<br />
Timex BPO (Pvt) Ltd.(Public Private Partnership) PPP<br />
Warehouse Monitoring System Customs<br />
Web Information System DeptAgri<br />
Web/Email System PubAdmin<br />
People interviewed in different organizations. Please refer Appendix A for organization codes.<br />
Table 3: Interviewed personnel<br />
Name Designation Organization<br />
Dr. T.A Piyasiri Director General TVEC<br />
Mr. Ajantha Dias Director IT Customs<br />
Mr. Bandula Perera Chief Innovation Officer LocalGov<br />
Mr. Chinthaka Ranasinghe Project Manager ICTA<br />
Mr. Dammika Karangoda Chief Innovation Officer DeptEdu<br />
Mr. Janaka Jayalath Director Information Systems TVEC<br />
Mr. K.S.P Jayawardena Additional General Manager CEB<br />
Mr. Kumudu Munasinghe Head of Information Management SEC<br />
Mr. Lalith Waduge Systems Engineer PubAdmin<br />
Mr. M.G Tillakaratne Deputy General Manager CEB<br />
Mr. Moin Habeeb Manager PPP<br />
Mr. Neil Gunadasa Director of Education DeptEdu<br />
Mr. W.A.G Sissirakuimara Subject Matter Specialist<br />
Program Director- Re-engineering Govern-<br />
LocalGov<br />
Mr. Wasantha Deshapriya<br />
ment ICTA<br />
251
Appendix D<br />
Srimal Jayawardena and Gihan Dias<br />
The interviews were based on the following sample questions.<br />
1. What are the Information Systems (IS) projects in your organization and what are the components<br />
of those projects?<br />
2. What technology, hardware, software or middleware components were used in these IS? Were<br />
any FOSS components used?<br />
3. How were the specifications of the project developed?<br />
4. Who are the key people involved in the selection decision?<br />
5. Regarding the procurement process, a) What factors influence the selection decision in your procurement<br />
process? b) How do you rank these factors based on their importance for the existing procurement<br />
process? c) Do you have any recommendations to improve the selection process?<br />
6. Regarding issues and room for improvement, a) If these systems are up and running, are there<br />
any problems? b) In the said systems what areas need improvement? c) Which parts of it are you<br />
happy with and which parts are you not satisfied? d) For the unsatisfied areas, do you see FOSS as<br />
a viable option?<br />
7. In your opinion in what areas would FOSS be more advantages to you?<br />
References<br />
Chan, A. S. (2007) Retiring the Network Spokesman, Science Studies 20 :78-99.<br />
Dedrick, J. and West, J. (2003) A Grounded Theory of Innovation and Standards Adoption,” Proceedings of the<br />
Workshop on Standard Making: A Critical Research Frontier for Information Systems (MISQ Special Issue<br />
Workshop):12-14<br />
Delone, W.H. and McLean, E.R (2003) Journal of Management Information Systems 19 (4):9-30.<br />
Gable, Guy, Sedera, Darshana, Chan and Taizan (2003) Enterprise systems success: a measurement model, Proceeding<br />
of the 24th International <strong>Conference</strong> on Information Systems, Seattle, USA:576-591.<br />
Hanna N. K. (2006) From Envisioning to Designing e-Development: The Experience of Sri Lanka, World Bank Publications:179-181.<br />
ICTA (2008) Re-engineering Government, accessed on 20-01-2008,<br />
<br />
Kenneth Wong (2004) Free/Open Source Software: Government Policy, Asia-Pacific Development Information Programme,<br />
ELSEVIER, accessed on 02-05-2008, <br />
Kwan, S. K and West, J. (2005) A conceptual model for enterprise adoption of open source software,” The Standards<br />
Edge: Open Season. Sheridan Books, Ann Arbor, Michigan<br />
Lemos, R. (2003) Merrill Lynch: Linux saves money, 7 June, Cnet News.com, accessed on 27-06-2008,<br />
<br />
Maguire, J. (2003) Windows vs. Linux: TCO Feud Rages On, Newsfactor Network, accessed on 27-06-2008,<br />
<br />
Orzech, D. (2002) Linux TCO: Less Than Half The Cost of Windows, CIO Update, accessed on 27-06-2008,<br />
<br />
Reddy, B. and Evans, D. (2002) Government Preferences for Promoting Open-Source Software: A Solution in<br />
Search of a Problem, Social Sciences Research Network, accessed on 01-05-2008,<br />
<br />
Schindler Esther (2008) Open Source is Entering the Enterprise Mainstream – Survey Shows, CIO.com, accessed<br />
on 27-06-2008,<br />
http://www.cio.com/article/375916/Open_Source_is_Entering_the_Enterprise_Mainstream_Survey_Shows><br />
Stoltz, M. (1999). The Case for Government Promotion of Open Source Software, NetAction (27), accessed on 05-<br />
05-2008, <br />
Wikipedia (2005) Free and open source software, accessed on 10-01-2008,< http://en.wikipedia.org/wiki/FOSS><br />
Wong Kenneth (2004) Free/Open Source Software:Government Policy, United Nations Development Programme –<br />
Asia Pacific Development Information Programme (UNDP-APDIP), ELSEVIER.<br />
Yin R.K. (2003) Case study research: design and methods, 3rd edition, Sage Publications Inc, USA.<br />
252
Evaluating the Success of Customer Relationship<br />
Management (CRM) Systems<br />
Farnoosh Khodakarami and Yolande Chan<br />
Queen’s University, Kingston, Canada<br />
Fkhodakarami@business.queensu.ca<br />
Ychan@business.queensu.ca<br />
Abstract: Based on the Delone and McLean IS success model (1992), this qualitative study explores customer<br />
relationship management (CRM) system success factors. A case research approach was applied to examine CRM<br />
applications in three organizations. Through these case studies, five factors of CRM success were explored, namely:<br />
system quality, customer information quality, system use, user satisfaction, and system impacts. With regard to<br />
system quality, a main determinant of success was the system’s primary characteristics. CRM systems can be<br />
categorized as operational, analytical and collaborative, based on their characteristics. The study showed how<br />
system’s characteristics directly affect the purpose of system use. Operational CRM systems are mostly used for<br />
providing customer services. Analytical systems are primarily used for decision support. Collaborative systems are<br />
used both for work integration and providing service for internal and external customers. The study showed that if<br />
systems properly address users’ expectations and employees have the required skills to work with systems,<br />
employees are more willing to utilize the capabilities of systems to support CRM processes and, in turn, are more<br />
satisfied with the outcomes. Customer information quality relates to system quality; integrated CRM systems that are<br />
based on standard platforms are more capable of generating high quality customer information in a timely manner.<br />
Individual and organizational impacts of CRM systems were explored. Individual productivity, improved decision<br />
making and planning, learning and awareness were shown to be important individual benefits that CRM systems<br />
provide for employees. At the organizational level, the study showed that CRM systems help organizations to acquire<br />
and share more knowledge about their customers, and improve business processes, products and services.<br />
Keywords: customer relationship management, CRM, system evaluation, IS success, system effectiveness<br />
1. Introduction<br />
In response to increasing competitiveness in the global market, organizations are looking for means to<br />
manage their relations with customers more effectively over time. As a result, customer relationship<br />
management (CRM) has become a central part of most businesses and many organizations are<br />
implementing CRM systems. CRM systems are a category of organizational information systems that<br />
enable organizations to provide services for customers, gather and analyze customer information and<br />
manage their relationships with customers more effectively. However, despite substantial investments,<br />
many CRM systems fail to meet expectations. Besides, organizations often do not know appropriate<br />
success criteria to use to evaluate their CRM systems’ effectiveness. This study addresses these issues<br />
by drawing on a qualitative case research approach.<br />
The rest of this paper is organized as follows. First, the theoretical backround and research framework<br />
are presented. Then, the research method and findings are described. Finally, conclusions, limitations<br />
and implications are discussed.<br />
2. Theoretical background<br />
The main objective of CRM is to retain current customers through increasing their loyalty and to select<br />
new customers that provide higher profitability (Hansotia 2002). With the growth of information<br />
technology applications in organizations, CRM systems are popular. Successful CRM systems help<br />
organizations to manage customer information more effectively and increase customer satisfaction,<br />
customer loyalty and customer profitability through meaningful communication with customers (Winer<br />
2001). CRM systems fall mainly into three classes (Bose and Sugumaran 2003; Henning et al. 2003):<br />
Operational CRM systems that aim at automation of CRM processes to improve efficiency and<br />
productivity. Examples of such systems are sales and marketing automation systems (e.g., point of<br />
sales (POS) systems), call support centres and customer databases.<br />
Analytical CRM systems that aim to provide better understanding of individual customers’ behaviours<br />
and needs. Analytical systems incorporate various tools for consumer behaviour predictive modeling<br />
and purchase pattern recognition, such as data mining, data warehouse and OLAP (online analytical<br />
processing).<br />
253
Farnoosh Khodakarami and Yolande Chan<br />
Collaborative CRM systems that manage and integrate customer communication channels and<br />
networks. Email, websites, internal portals, and tele-, video- and web-conferencing applications are<br />
examples of collaborative systems that organizations use.<br />
Organizations invest in CRM systems and in turn expect the systems to positively impact their<br />
organizations and various customer related processes (Barua et al. 1995; King and Burgess 2008).<br />
Accurate measuring of the impacts of CRM systems requires “understanding systems’<br />
multidimensionality, and the development of a correspondent, standardized, validated, and robust<br />
measurement instrument” (Sedera and Gable 2004, p. 449).<br />
However, within the CRM domain, there is a lack of research that evaluates the success of CRM systems<br />
in actual use. To measure the success of information systems, Delone and McLean (1992) developed a<br />
model that includes both individual and organizational impacts of information systems. Considering the<br />
CRM systems as a category of organizational information systems, their success can be evaluated using<br />
the Delone and McLean model.<br />
This model (see Figure 1) has been widely cited and has been a valuable means to measure IS success<br />
in organizations (Gable, Sedera and Chan 2008).<br />
Figure 1: DeLone and McLean IS success model (1992)<br />
In the IS success model, two main factors, system quality and information quality, influence system<br />
usage and user satisfaction, which in turn lead to individual and organizational impacts. System quality<br />
refers to the desirable characteristics of an information system, such as system flexibility and ease of<br />
learning, while information quality refers to the desirable characteristics of the system outputs, such as<br />
management reports and Web pages. ”Use” can have different meanings. It can refer to the amount of<br />
use, frequency of use, and extent of use, as well as the nature and purpose of use (Petter et al. 2008).<br />
Examples of measures in each category (i.e., system quality, information quality, individual and<br />
organizational impacts) were gathered in a multidimensional IS success measures instrument developed<br />
by Sedera and Gable (2004) that also helped to inform our study. See Table 1.<br />
Table 1: Validated measures for IS success, Sedera and Gable (2004)<br />
For the purpose of this exploratory study, a framework was developed based on the Delone and McLean<br />
(1992) model. This framework is discussed in the next section.<br />
254
3. Research framework<br />
Farnoosh Khodakarami and Yolande Chan<br />
The impacts of information systems are often interdependent and influenced by various factors. The<br />
research framework (see Figure 2) indicates the factors that are investigated in this research. Not all of<br />
the DeLone and Mclean factors are considered in this paper because of space constraints. Some of the<br />
most important variables that affect the success of CRM systems are as follows:<br />
Figure 2: Research framework<br />
3.1 System quality<br />
System quality refers to the desirable characteristics of an information system (Petter et al. 2008).<br />
System quality is a multi-dimentional measure. Many researchers focus on the engineering<br />
characteristics and features of the system as main indicators of system quality (Ravichandran and Rai<br />
2000). Therefore system features such as integration, complexity and flexibility have been considered as<br />
system quality measures in a broad range of IS sucess studies (Petter et al. 2008). Besides, in regard to<br />
CRM systems, the primary characteristics of the system (i.e., operational, analytical and collaborative<br />
characteristics) can be considered as system quality factors. It is worthwhile to investigate how these<br />
characteristics affect systems’ success. In addition, users’ judgement is also critical for the evaluation of<br />
system quality. Measures like user requirements, ease of use and ease of learning reflect the users’<br />
evaluation of system quality (Ravichandran and Rai 2000, Petter et al. 2008).<br />
3.2 Customer information quality<br />
In the DeLone and McLean (1992) model, information quality refers to the desirable characteristics of the<br />
system output, namely the quality of information that system produces in the forms of reports, analyses<br />
and web pages (Petter et al. 2008). The main output of the CRM system is customer information, and<br />
effective management of customer information is a central characteristic of CRM systems; it helps<br />
organizations to better understand their customer needs and expectations and to respond to those needs<br />
in a timely manner (Hansotia 2002). Large amounts of data available to organizations through different<br />
communication channels are becoming a challenge for organizations. CRM success is therefore strongly<br />
associated with systems’ capability to manage the data and provide real time, meaningful, accurate and<br />
high quality customer information (Reid and Catterall 2005). Thus, availability of customer information is<br />
in fact a good indicator of the quality of the information.<br />
3.3 System use<br />
System use is an important dimension of IS success and is associated with the impact or benefits<br />
realized from IS (Petter et al. 2008). It indicates the degree and manner in which employees utilize the<br />
capabilities of an information system, and it can be described as amount of use, frequency of use and<br />
appropriateness of use (DeLone and McLean 2003; Petter et al. 2008). Most studies have measured the<br />
extent of use ( i.e., amount and frequency of use). However, the extent of use may not be the best<br />
measure of IS use since more use is not always better. To overcome this deficiency, Doll and Torkzadeh<br />
(1998) suggested that system use can be measured based on the purpose of use. They identified<br />
decision support, customer service and work integration as three main usage behaviours. Decision<br />
support refers to the extent to which the information system is used to analyze data and make decisions.<br />
Customer service is the extent to which the information system is used to provide service for both internal<br />
customers (employees) and external customers. Work integration incorporates horizontal and vertical<br />
255
Farnoosh Khodakarami and Yolande Chan<br />
integration. Horizontal integration refers to the extent to which the information system facilitates<br />
communication in work groups and helps employees to collaborate on various tasks. Vertical integration<br />
refers to the extent to which the system is used to “plan one’s work, monitor performance and<br />
communicate vertically with superiors and subordinates” (Doll and Torkzadeh 1998, p. 174).<br />
3.4 User satisfaction<br />
User satisfaction refers to the user’s level of satisfaction with system reports, outcomes and support<br />
services (Delone and McLean 2003). Scholars use various measures for user satisfaction. End User<br />
Computing Satisfaction (EUCS) (Doll and Torkzadeh’s 1994) and the User Information Satisfaction (UIS)<br />
instrument (Ives et al. 1983) have been the most widely used instruments. However, some scholars use a<br />
single measure for user satisfaction (e.g., Rai et al. 2002) and others measure positive user attitude as<br />
an indicator of IS success (e.g., Coombs et al. 2001). While some measures of user satisfaction focus on<br />
cognitive aspects of user satisfaction and the "extent to which users believe the information system<br />
available to them meets their information requirements" (Ives et al. 1983, p. 785), other measures focus<br />
on affective responses (i.e., favorable and unfavorable attitudes toward systems) as indicators of user<br />
satisfaction (Melone 1990). In this research both cognitive and affective responses are explored.<br />
3.5 System impacts<br />
System impacts refers to “the extent to which IS are contributing to the success of individuals, groups,<br />
organizations, and industries” (Petter et al. 2008, p.239). CRM systems provide benefits at different<br />
organizational levels. Individuals in different departments use CRM systems to accomplish their tasks<br />
more effectively and efficiently. Individual impacts indicate the extent to which CRM has influenced the<br />
capabilities and effectiveness of individual users. Organizations also take advantage of CRM systems.<br />
Organizational impacts indicate the extent to which CRM has promoted improvement in organizational<br />
results and capabilities (Gable et al. 2008). The Petter et al. (2008) review of literature on IS success<br />
revealed that few studies have considered organizational impacts of information systems. However,<br />
despite measurement difficulties, impacts of information systems cannot be monitored only at the<br />
individual level. Specifically, in the context of CRM, we know these systems provide benefits beyond the<br />
individual level. Within organizations, several departments such as marketing, sales, customer relations<br />
and R&D can substantially take advantage of CRM systems. Business process change, overall<br />
productivity, increased outcomes/outputs are some examples of organizational benefits of IS.<br />
4. Research method<br />
This study explores CRM system usage and evaluates systems’ effectiveness in organizational contexts.<br />
Given the exploratory nature of this study, a qualitative case study approach was used. The case studies<br />
allowed the researchers to learn about CRM systems in their business contexts and explore various<br />
factors that affected the usefulness of the systems. Case studies are specifically appropriate for<br />
information systems research; due to the rapid pace of information systems evolution, previous studies of<br />
the phenomena are limited. Case studies provide researchers with valuable insights to study phenomena<br />
and generate theories from practice (Benbasat et al. 1987). Three Canadian organizations were selected<br />
for this study. These organizations represent different industries and vary in terms of CRM capabilities,<br />
system development levels and organizational characteristics (see Table 2). The organizations use CRM<br />
systems and other information systems to support customer relations and processes. Call centres, POS<br />
and customer databases are the main operational systems used. For customer analysis, the electronics<br />
organization uses data warehouse and data mining systems, but the other two organizations largely rely<br />
on Excel analyses. In regard to collaborative systems, e-mail, company websites and social media (e.g.,<br />
Facebook, online communities and forums) are the main means of communication with customers. Each<br />
organization has several internal portals to support customer-related processes. The electronics<br />
organization has an e-support website to provide online customer services.<br />
During a four-month period in 2010, twelve telephone and in-person interviews were conducted with<br />
CRM users and with managers and employees from customer service, marketing, and IT departments<br />
within the three organizations. Sample interview questions are presented in Appendix. Interviews were<br />
recorded, transcribed and coded using the Nvivo program. The results are discussed in the next section.<br />
This research methodology is rigorous, satisfying criteria that Dube and Pare (2003) have recommended<br />
for a good IS positivist case study. Criteria for research design include: clear research questions, a priori<br />
specification of the constructs, a clean theoretical slate, and a multiple-case design. As recommended by<br />
Dube and Pare (2003), to assist with data analysis and the formulation of arguments, a case study<br />
256
Farnoosh Khodakarami and Yolande Chan<br />
database was developed as interviews were coded, field notes recorded, cross-case patterns explored,<br />
and quotes tracked.<br />
Table 2: Comparison of case characteristics<br />
Organization<br />
Organization Size<br />
Product/Service System Development Level<br />
Structure<br />
Electronics Large<br />
Health Medium<br />
Several companies,<br />
Local stores across<br />
Canada<br />
One company, Local<br />
stores across<br />
Canada<br />
Education Small One local school<br />
Various categories<br />
of electronic<br />
products<br />
Weight loss<br />
services, health<br />
and nutrition<br />
products<br />
Services<br />
(Educational<br />
programs)<br />
Well developed systems with various<br />
applications, systems are upgraded<br />
regularly<br />
Few systems that were recently<br />
launched, system development is in<br />
progress<br />
Few systems that have been used for<br />
years with minor upgrades, systems<br />
applications are not well utilized<br />
5. Results and discussion<br />
Using the case data, the relationships among system success factors were explored. Interviewees were<br />
asked to explain the benefits that systems provide for them and their organizations. Figure 3 presents the<br />
relationships that were identified through the interviews. These relationships are discussed below.<br />
Figure 3: Summary of findings<br />
257
System quality: Customer information quality<br />
Farnoosh Khodakarami and Yolande Chan<br />
System quality affects customer information quality in several ways. An important factor that determines<br />
the quality of systems is system integration. The interviews revealed that system integration greatly<br />
affects the availability of customer information. Customer information exchange is an important issue for<br />
many organizations; integrated systems facilitate the flow of information through the organization and<br />
allow the employees to get access to the right information in a timely manner. In the health organization,<br />
the disintegration of systems has increased the work overload in s, since the employees have to spend<br />
time to manually gather and combine all the information they need to do analyses and make decisions.<br />
There’s still not that direct connect between the reports. And I have to go back through so<br />
much manual work to find out how many leads each centre had and then try to correlate it to<br />
how many actually showed up and then manually again to find out how many sold.<br />
Regional Director, Health Organization<br />
In addition to system integration, system flexibility affects the availability of customer information. Flexible<br />
CRM systems allow users to define their specific information requirements and get access to that<br />
information. It is assumed that more system flexibility is beneficial for the availability of information;<br />
however an interesting observation was that the high level of flexibility will in fact restrict the availability of<br />
information in some occasions. For instance, the CRM system which is used in the education<br />
organization has a high level of flexibility that allows the employees to easily create new databases.<br />
These databases consist of custom-made fields based on individuals’ preferences. As a result,<br />
paradoxically, employees cannot easily retrieve the information they need from other employees’<br />
databases and integrate it to their own databases and this restricts the flow of information.<br />
The amount of flexibility is incredible, almost too flexible. It allows us to keep creating these<br />
separate databases as often as we can without relating any information between them and<br />
that's a big challenge.<br />
IT Associate Director, Education Organization<br />
System quality: System use<br />
System quality affects the use of systems. Both purpose and frequency of use are affected by system<br />
quality. The data suggested that purpose of use is associated with the primary characteristics of CRM<br />
systems. Within the studied organizations, operational CRM systems such as call centres and point of<br />
sales (POS) systems are mostly used for providing customer services for external clients. Analytical<br />
systems are primarily used for decision support. Employees use analytical systems (e.g., data<br />
warehouses, Excel) to analyze customer information and discover patterns of consumer behaviour.<br />
These analyses aid the decision making processes to a great extent. Collaborative systems have the<br />
highest applicability since they are used for multiple purposes. Some of these systems are used for<br />
vertical integration. For instance, in the health organization and the electronics organization, district<br />
managers have regular web conferencing and teleconferencing with the store managers to monitor<br />
stores’ performance and share customer experiences and best practices. A group of collaborative<br />
systems such as portals supports horizontal work integration. Through the internal portals, employees<br />
collaborate on projects and collectively accomplish tasks. Another group of collaborative systems<br />
facilitates internal and external customer service provision. All three organizations have e-learning<br />
websites that provide customer support training for employees. The electronics organization has a<br />
website that provides customer and product information that sales associates need. This website is<br />
updated regularly to include new products. This organization also has an e-support website that supports<br />
repair and maintenance services for external customers.<br />
There are other system quality variables that affect the frequency of use and employees’ willingness to<br />
use the systems. User requirements are one of these variables. If a system fails to address users’<br />
requirements properly, understandably employees are less likely to use it. This is particularly a challenge<br />
for the education organization. This organization uses a CRM system which was primarily developed to<br />
support sales and marketing processes. Therefore, the CRM system’s nomenclature and functionalities<br />
are not fully compatible with the particular requirements of an educational organization. As a result, the<br />
capabilities of the systems have been largely underutilized.<br />
258
Farnoosh Khodakarami and Yolande Chan<br />
A lot of examples in the training guide are sort of product related, which is not really the area<br />
that we fall under… if we go out into the more sales and marketing aspect of it [the CRM<br />
system] we might realize that it’s limited in what we’re trying to do. We’re just not using those<br />
parts.<br />
Director of Executive Development Centre, Education Organization<br />
Ease of learning is another system quality factor that influences system use. The interviews at the<br />
education organization revealed that the CRM system has sophisticated analytical capabilities and a<br />
database management tool that “scares people”. It is not easy to learn and utilize these capabilities since<br />
“it takes a different IT skill set and is not something that the average employee can do”. As a result,<br />
employees do not use these tools and merely rely on Excel spreadsheets to do basic analyses and<br />
reporting.<br />
Customer information quality: System use<br />
No direct relationship was found between customer information quality and system use. In the literature<br />
also there is little research demonstrating this relationship. The Petter et al. (2008) review of 180 papers<br />
on IS success showed that, at the time, only six papers had discussed this relationship. In general, Petter<br />
et al. (2008) did not find sufficient data to support this relationship.<br />
System quality: User satisfaction<br />
As expected, it became clear that if a system has the capability to meet the users’ requirements,<br />
employees are more satisfied with the system. For instance, the health organization has a system that<br />
allows employees to monitor clients’ weight loss progress, easily get access to their profiles and<br />
purchase behaviour, perform analyses, and provide consultancy services. This system was developed<br />
internally to address the specific requirements of the users; therefore, employees are very satisfied.<br />
My first real experience with [the new system] was just like there was a wild moment for me<br />
where I went 'Wow!' it gives me everything I need, I can actually be more effective as a<br />
regional manager by spending an effective administration day, reviewing what's going on in<br />
centres before I go there.<br />
Regional Director, Health Organization<br />
System integrity also affects user satisfaction. In many cases, lack of system integration leads to<br />
duplication in data entry and manual work in order to aggregate data from different sources; this leads to<br />
dissatisfaction. Similarly, if the CRM system is not compatible with other information systems, it can<br />
cause problems and unnecessarily increase the workload, which also leads to employee complaints.<br />
Customer information quality: User satisfaction<br />
Availability of customer information affects user satisfaction. CRM systems provide employees with the<br />
information they need for their analyses, planning and decision making processes. When CRM systems<br />
provide the right information in a timely manner, employees feel satisfied with the system. However, if the<br />
information is not properly organized, it will take a long time for employees to pull out the right information<br />
they need. This inefficient process can be tedious and time consuming, and lead to user dissatisfaction.<br />
We put a lot of information into different areas and therefore pulling out the information is<br />
difficult …. It can sometimes take a few hours to get information that if it was done properly<br />
could be probably at your fingertips.<br />
Director of Operations, Education Organization<br />
System use: Impacts<br />
CRM systems provide various benefits for organizations. Interviewees were asked to discuss some of the<br />
benefits that systems provide for their individual work and for the organization. In general, most of the<br />
interviewees admitted that CRM systems have improved their business capabilities and helped them<br />
know their customers better and make more effective decisions. The main systems’ impacts which were<br />
identified through the interviews are as follows:<br />
259
Farnoosh Khodakarami and Yolande Chan<br />
Individual impacts:<br />
Productivity<br />
Interviewees believed that CRM systems help them work more productively. They spend less time<br />
acquiring and aggregating information, they can perform more comprehensive analyses, and they provide<br />
better customer service. In addition, the feedback that organizations receive through various collaborative<br />
systems helps managers to evaluate employees’ performance and provide recommendations for training<br />
that can help employees to further improve their performance.<br />
Improved decision making and planning<br />
Systems also improve individual decision making and planning. Using the systems, employees have<br />
access to more knowledge resources and can get the information they need for decision making in a<br />
reasonable time. Exchanging ideas and feedback through the collaborative systems also helps managers<br />
with their decision-making and planning tasks.<br />
Learning<br />
Information systems provide learning opportunities for employees and customers. The organizations<br />
studied have e-learning systems and internal portals which provide documents and knowledge resources<br />
related to customers. The electronics organization also has an internal portal that provides product<br />
information for sales associates and call centre agents who can learn about the specific features of each<br />
product. This enables the sales associates and call centre agents to provide better customer service.<br />
Awareness<br />
There are several systems that increase employees’ awareness of events, market fluctuations and<br />
product changes. The health organization has an internal portal where all events and discussions are<br />
posted. Within the education organization, the instant messaging tool is broadly used for this purpose. In<br />
the electronics organization, agents benefit from increased awareness as a result of a “product alert”<br />
CRM feature. If a specific problem is identified regarding a new product, the call centre manager makes<br />
the call centre agents aware of this problem by assigning an alert to the product. It helps the agents to<br />
provide better service for customers when they have inquiries about the particular product.<br />
Organizational impacts<br />
Knowledge sharing and communication<br />
Collaborative CRM systems facilitate knowledge sharing and communication through the organization.<br />
For instance, in the electronics organization, employees can share their new ideas through an electronic<br />
suggestion box. Web-conferencing is also used in all three organizations to facilitate real-time<br />
communication in team meetings. Online communities and social media (e.g., Facebook and Twitter) are<br />
some of the systems that these organizations use to communicate with potential and actual customers.<br />
Process improvement<br />
Knowledge sharing and feedback from employees and customers help the organizations to improve<br />
various processes (e.g., marketing, advertising, and financial forecasting). The electronics organization<br />
has launched an advertising and product recommendation process which is supported by analytical CRM<br />
systems. By analyzing customer purchase history for each individual customer, the organization is able to<br />
suggest products that match each customer’s needs and complements his or her past purchases.<br />
Product /service improvement<br />
CRM systems, and specifically analytical CRM systems, help organizations to better understand<br />
customers’ need and predict their behaviour. Customer feedback which is collected through collaborative<br />
systems also reveals customers’ expectations. Having more detailed knowledge about customers helps<br />
the company to improve products and services. In addition, CRM systems enable organizations to<br />
provide a broader range of services to their customers. For instance, the health organization has<br />
launched an online system which allows customers to monitor their weight loss progress at any time,<br />
consult with health and nutrition experts, and receive supplemental resources and articles that give them<br />
the information they need to maintain their weight loss. The e-support system of the electronics<br />
organization is another effective system that helps to provide superior self-service support for customers.<br />
The system not only provides manuals and information on products but also guides customers to<br />
personally recognize product problems and fix them without the need to refer problems to repair shops.<br />
260
6. Conclusion<br />
Farnoosh Khodakarami and Yolande Chan<br />
Organizations are interested in evaluating the effectiveness of CRM systems; however the effectiveness<br />
of systems is influenced by various technological and organizational factors, and the measurement of<br />
systems’ success is challenging. Using the Delone and McLean (1992) multi-dimensional IS success<br />
model, this study explored factors that affect the success of CRM systems. The qualitative case studies<br />
provided the researchers with insights and real examples of how CRM systems are used in their<br />
organizational contexts and what variables contribute to their success. System integration was found to<br />
be an important technological variable that affects the use of systems. Systems integrity facilitates<br />
information sharing and allows employees to access accurate information in a timely manner. It was<br />
confirmed that if systems properly address the users’ requirements, they are more likely to be used.<br />
However, the level of systems’ complexity should match the employees’ IT skills; otherwise systems’<br />
capabilities will not be utilized.<br />
CRM systems help employess be more productive. They also provide learning opportunities for<br />
employees and customers. But CRM systems’ impacts go beyond the individual level. At the<br />
organizational level, systems facilitate communication and knowledge sharing throughout the<br />
organization and with customers. Furthermore, CRM systems enable organizations to provide a broader<br />
range of services for customers and improve products and business processes.<br />
In summary, CRM systems have major impacts at different levels and their effectiveness depends on<br />
several technological and organizational factors. As a result, in order to evaluate the effectiveness and<br />
success of CRM systems, organizations should draw on multi-dimensional measures and pay careful<br />
attention to the interdependency of these measures.<br />
7. Limitations<br />
Some study limitations should be acknowledged. Our findings are based on case studies and interviews<br />
in three industries. As a result, the generalizability of our findings is limited. However, our findings can<br />
serve as a starting point for other researchers exploring measures for the evaluation of CRM systems.<br />
We invite other researchers to conduct quantitative, large-scale follow-on explorations.<br />
8. Research implications<br />
This exploratory research contributes to our understanding of CRM system success measures. Previous<br />
studies on IS success have rarely focused on CRM systems. The success dimensions discussed in this<br />
paper can serve as the basis for a CRM system effectiveness construct. Future research can enhance<br />
and validate measures for the evaluation of CRM systems. The current study‘s findings can also help<br />
researchers and organizations better understand the strengths and weaknesses of current CRM systems,<br />
and collaborate on the design of future systems.<br />
9. Appendix 1: Sample interview questions<br />
On a daily basis, how often do you use CRM systems?<br />
For what purposes do you use CRM systems?<br />
What benefits do CRM systems provide for you and your organization?<br />
To what extent are CRM systems successfully and effectively supporting business processes within<br />
your department or in the whole organization?<br />
In your opinion, what are the strengths and weaknesses of your organization’s current CRM<br />
systems?<br />
What are your suggestions for improving CRM systems to match your requirements?<br />
In your opinion, what are the barriers for more extensive and effective use?<br />
How does your organization help you better utilize CRM applications?<br />
References<br />
Benbasat, I., Goldstein, D. K. and Mead, M. (1987) "The Case Research Strategy in Studies of Information Systems".<br />
MIS Quarterly, Vol 11, No.3, pp.369-386.<br />
Bose, R. and Sugumaran, V. (2003) "Application of Knowledge Management Technology in Customer Relationship<br />
Management". Knowledge and Process Management, Vol 10, No.1, pp. 3-17.<br />
261
Farnoosh Khodakarami and Yolande Chan<br />
Delone, W. and Mclean, E. (1992) "Information Systems Success: the Quest for the Dependent Variable".<br />
Information Systems Research, Vol 3, No.1, pp. 60–95.<br />
Doll, W. and Torkzadeh, G. (1994) "Developing a Multidimensional Measure of System-use in an Organizational<br />
Context". Informationand Management, Vol 33, No.4, pp.171–185.<br />
Dube, L. and Pare, G. (2003) "Rigor in Information Systems Positivist Case Research: Current Practices, Trends,<br />
and Recommendations". MIS Quarterly,Vol 27, No.4, pp. 597-635.<br />
Gable, G. G., Sedera, D., and Chan, T. (2008) "Re-conceptualizing Information System Success: The IS-Impact<br />
Measurement Model". Journal of the Association for Information Systems, Vol 9, No.7, pp. 377-408.<br />
Hansotia, B. (2002) "Gearing up for CRM: Antecedents to Successful Implementation". Journal of Database<br />
Marketing, Vol 10, No. 2, pp.121-132.<br />
Ives, B., Olson, M. and Baroudi, J. (1983) "The Measurement of User Satisfation". Communications of the ACM, Vol<br />
26, No.10, pp. 785–793.<br />
BIBLIOGRAPHY Melone, N. P. (1990) " A Theoretical Assessment of the User-Satisfaction Construct in Information<br />
Systems Research". Management Science, Vol. 36, No. 1, pp. 76-91 .<br />
Petter, S., DeLone, W., and McLean, E. (2008) "Measuring Information Systems: Models, Dimensions, measures,<br />
and interrelationships". <strong>European</strong> Journal of Information Systems, Vol 17, pp. 236–263.<br />
Rai, A., Lang, S. S. and Welker, R. B. (2002) "Assessing the Validity of IS Success Models: An Empirical Test and<br />
Theoretical Analysis". Information Systems Research, Vol 13, No.1, pp. 50-69.<br />
Ravichandran, T. and Rai, A. (2000) "Total Quality Management in Information Systems Development: Key<br />
Constructs and Relationships". Journal of Management Information Systems, Vol 16, No.3, pp.119-155.<br />
Sedera, D. and Gable, G. (2004) "A Factor and Structural Equation analysis of the Enterprise Systems Success<br />
Measurement Model", Twenty-Fifth International <strong>Conference</strong> on Information Systems, pp. 449-464, Association<br />
for Information Systems, Washington, USA<br />
Not all references cited in this article have been listed because of space constraints. Other references are available<br />
upon request from the authors.<br />
262
Toward a Novel Methodology for IT Strategic Planning<br />
Mouhsine Lakhdissi and Bouchaib Bounabat<br />
ENSIAS, Université Mohammed V, Morocco<br />
mlakhdissi@gmail.com<br />
bounabat@ensias.ma<br />
Abstract: Defining IT target state and the specific roadmap to attain it, are main focuses of IT strategic planning<br />
which is considered increasingly as the basis for IT decision-making and governance. Unfortunately, it hasn’t evolved<br />
with the same speed as other fields in the IT sphere. Most of the techniques, approaches and methods related to IT<br />
Strategic Planning date back to the 80s or 90s and are most often oriented business rather than IT. Enterprise<br />
Architecture is a really promising discipline aimed at capturing the as-is architecture of an enterprise, defining the<br />
target and the roadmap to get from existing to desired state. In that way, it is tightly related to IT strategic planning<br />
and it can provide a framework to fill the gap and contribute in structuring and formalizing IT Strategic Planning field.<br />
Enterprise Architecture benefits from a standardization effort as well as from tool support. Deliverables and artifacts<br />
are generally well defined and structured in the existing frameworks. This paper present a new methodology for IT<br />
Strategic Planning based on Enterprise Architecture.<br />
Keywords: IS strategic planning, enterprise architecture, methodology, metamodel, content framework<br />
1. Background<br />
1.1 Motivation<br />
With the complexity of today’s information systems and the necessity to make the existing IT assets more<br />
agile to provide for the constant business change, the task of governing and planning for IT assets<br />
become a key success factor for IT.<br />
Information Systems Strategic Planning is the discipline that deals with this task. Unfortunately, it hasn’t<br />
evolved with the same speed as other fields in the IT sphere. Most of the techniques, approaches and<br />
methods related to IT Strategic Planning date back to the 80s or 90s and are most often oriented<br />
business strategic planning rather than IT strategic planning (Hsu 1995). As a matter of fact, they don’t<br />
take into account the complexity of today’s information system and their diversity. Furthermore, this field<br />
lacks from a formal, rigorous and agreed upon methodology and suffers from the absence of tools to<br />
support, structure and industrialize the discipline<br />
Enterprise Architecture is a really promising discipline aimed at capturing the as-is architecture of an<br />
enterprise, defining the target and the roadmap to get from existing to desired state. In that way, it is<br />
tightly related to IT strategic planning and it can provide a framework to fill the gap and contribute in<br />
structuring and formalizing IT Strategic Planning field. Enterprise Architecture benefits from a<br />
standardization effort as well as from tool support. Deliverables and artifacts are generally well defined<br />
and structured in the existing frameworks.<br />
Existing Enterprise Architecture frameworks are of different types. While some frameworks like (Zachman<br />
1987) define a taxonomy for architecture artifacts, others like TOGAF (The Open Group 2009), tend to<br />
describe a process to produce architecture deliverables (Sessions 2007). The main concept underlying<br />
both the process and the taxonomy is the metamodel to describe architecture elements and to produce<br />
architecture deliverables.<br />
This metamodel is often either very poor to describe fully the architecture or not well structured to define<br />
the dependencies and the relationships between elements. We think that in order to define a more<br />
rigorous and structured methodology for IT Strategic Planning, it is necessary to define a rich and<br />
structured metamodel covering both architecture elements (processes, applications, data..) and<br />
transformation elements (programs, projects, budgets). This metamodel is the main focus of our work<br />
aimed at defining a new methodology for IT Strategic Planning.<br />
Our aim in this paper is to demonstrate the insufficiencies, deficiencies and inconsistencies in existing IT<br />
strategic planning methods and show how a new methodology based in part on the Entreprise<br />
Architecture practice could be proposed to address these problems. The metamodel we project to define<br />
could be used as a platform for describing the architecture, evaluating it and defining the needed<br />
transformations and planning them in term of programs/projects.<br />
263
Mouhsine Lakhdissi and Bouchaib Bounabat<br />
This paper proposes a new methodology for IT Strategic Planning based on Enterprise Architecture. One<br />
of the causes of IT project failure is poor strategic planning for them. This methodology can address the<br />
insufficiencies in existing approaches by using a rich metamodel and by combining architecture<br />
description and project definition.<br />
The second section presents IS Strategic Planning and Enterprise Architecture, compares the two<br />
disciplines and tries to bridge the gap between them. The content framework and the underlying<br />
metamodel is introduced in the the third section as a way to combine ISSP and EA. The fourth section<br />
describes the suggested metamodel and content framework. A comparison is made in the fifth section<br />
with existing metamodels before presenting future work and directions.<br />
1.2 Basic concepts<br />
The term methodology is very often misused in research as well as in practice. Many methods or<br />
frameworks are labelled “methodology” while they are not. So what is a methodology? It is defined by<br />
(Merriam 2010) as “a body of methods, rules, and postulates employed by a discipline” and by (TEAF<br />
2000) as "A documented approach for performing activities in a coherent, consistent, accountable, and<br />
repeatable manner". A method in the other hand is defined by (Merriam 2010) “a way, technique or<br />
process of or for doing something”<br />
From these definitions we can conclude that a methodology could be a structured composition of<br />
methods, rules and principles. It has a larger and more abstract scope than a method.<br />
Confusion is also often made between a framework and a methodology. (FEAF 1999) defines a<br />
framework as "A logical structure for classifying and organizing complex information". A framework is<br />
more of a structure that gives guidelines on some discipline without being prescriptive and without<br />
defining a structured process to undertake the underlying tasks.<br />
The table below summarizes some of the differences between the three concepts:<br />
Table 1: Comparison of method, framework and methodology<br />
Method Methodology Framework<br />
Scope Specific General General<br />
Abstraction Concrete Abstract Abstract<br />
Nature Prescriptive Prescriptive Descriptive<br />
Composition Simple Composite Composite<br />
Process Optional Mandatory Optional<br />
Product description Poor Average Detailed<br />
2. IT strategic planning and enterprise architecture<br />
2.1 IT strategic planning<br />
Strategy is defined by Chandler as “The determination of the basic long-term goals and objectives of an<br />
enterprise and the adoption of courses of action and the allocation of resources necessary for carrying<br />
out these goals”1 and by porter as “The art to build durable and defendable competitive advantage” 2 .<br />
One of the most complete definitions was given by (Arnoldo 1996), “A fundamental framework for an<br />
organization to assert its vital continuity, while, at the same time, forcefully facilitating its adaptation to a<br />
changing environment.”<br />
For most of the definition, strategic planning is focused on three main questions:<br />
Where we are?<br />
Where we want to go?<br />
How to get there?<br />
IS Strategic planning has been defined by (Lederer and Sethi 1992) as the process of identifying a<br />
portfolio of applications/projects that can help an organization achieve its business strategy. Its focus is<br />
on defining the IT roadmap in term of key initiatives, projects and transformations to be made on the<br />
existing information system with two main intentions:<br />
How to align information systems with business needs and overall strategy?<br />
264
Mouhsine Lakhdissi and Bouchaib Bounabat<br />
How to use information technology to change and impact the business?<br />
Due to the complexity of today’s information systems and the diversity of enterprise’s technology<br />
approaches, many methods have been defined to structure the ISSP process and techniques have been<br />
defined to address some aspects of the discipline. (Vitale 1986) classifies ISSP methods into two<br />
categories:<br />
Impact methods : trying to make It help create a positive impact and drive the change of the business<br />
Alignment methods : where the main focus is on aligning IT to respond to business needs and to help<br />
achieve strategic goals<br />
Among the methods used in IT Strategic Planning we have Critical Success Factors (CSF) (Rockart<br />
1979) which could be considered as an impact and alignment method, Business Systems Planning (BSP)<br />
(Wiseman 1988), Porter’s Competitive Forces Model (Porter 1980), Porter’s Value Chain (Porter 1985),<br />
and Scenarios (Schwartz 1991). Methods can be grouped together to constitute a methodology.<br />
Methodologies used for ISSP include those of the CCTA (1988, 1999) and Boar (2001).<br />
Many IT vendors and consultancy organizations use proprietary methods and/or methodologies, some of<br />
which are adaptations of open source approaches. Examples are Arthur Andersen’s Method/1 and<br />
Coopers and Lybrand’s Summit (Lederer and Sethi 1988, Min et al. 1999). It is also well known that<br />
organizations often develop their own in-house methodologies, often based on open or proprietary<br />
methods or approaches (Earl 1993, Lederer and Sethi 1988).<br />
2.2 Enterprise architecture<br />
ISO/IEC 42010: 2007 defines ‘‘architecture’’ as: ‘‘the fundamental organization of a system, embodied in<br />
its components, their relationships to each other and the environment, and the principles governing its<br />
design and evolution.’’. The (Open Group 2009) defines it as:<br />
“A formal description of a system, or a detailed plan of the system at component level to guide its<br />
implementation”<br />
“The structure of components, their inter-relationships, and the principles and guidelines governing<br />
their design and evolution over time”<br />
An architecture is typically made up of:<br />
A picture of the current state<br />
A blueprint, vision or detailed description for the future<br />
A road-map on how to get there<br />
Enterprise Architecture appeared in the eighties thanks to John Zachman who introduced the framework<br />
that bears his name. This framework consists of taxonomy for producing architecture artifacts from<br />
different viewpoints and perspectives. As a matter of fact, Enterprise Architecture has been defined by<br />
(Zachman, 2005) as a "set of descriptive representations (i.e. ‘models’) that are relevant for describing an<br />
Enterprise such that it can be produced to management’s requirements (quality) and maintained over the<br />
period of its useful life".<br />
Several other frameworks appeared subsequently, most of them initiated by government bodies like<br />
TAFIM (Technical Architecture Framework for Information Management), DODAF, MODAF or FEAF<br />
especially due to the requirement of the Clinger-Cohen Act.<br />
IT consulting firms created their own EA frameworks, based on the feedback from projects they<br />
undertook. Gartner as well as Cap Gemini or Accenture have their own EA frameworks which could be<br />
more accurately considered as EA practices as stated by (Session 2007)<br />
The Open Group Architecture Framework (TOGAF) started with TAFIM and reproduced practices and<br />
techniques used in other framework to constitute an EA framework of reference in the IT industry.<br />
TOGAF is with Zachman the two most used EA frameworks according to (Schekkerman 2005).<br />
TOGAF consist of:<br />
A architecture development methodology describing the process<br />
A set of guidelines and techniques supporting the methodology<br />
265
Mouhsine Lakhdissi and Bouchaib Bounabat<br />
A content framework with a metamodel describing the products (deliverables)<br />
Reference models that provide best practices to compare with<br />
A structure and description of the architecture repository (enterprise continuum)<br />
A capability framework for architecture governance and implementation<br />
TOGAF could be used in combination with Zachman where TOGAF defines the process and Zachman<br />
the deliverables. (Archimate 2009) defines a notation for architecture elements but also defines its own<br />
metamodel for architecture description.<br />
Enterprise Architecture could be used for different needs and in various contexts. It can operate as:<br />
A method to describe the enterprise as a whole with different levels and views of enterprise elements<br />
and their relationships. In this way it relates to Enterprise Modeling as was stated by (Lillehagen et<br />
al., 2005)<br />
A way to align the IS environment with the business reality and the strategic goals or to assess this<br />
alignment as described by (Bounabat 2006) (Elhari 2010)<br />
A modelling structure to define the vision for IS evolution or to describe in detail the IS to-be state<br />
A process to plan the migration between the as-is situation and the to-be state.<br />
All these EA use cases could be related to a step or a phase in term of process or deliverables of ISSP.<br />
2.3 Comparison and correlation<br />
A theoretical comparison of IT Strategic Planning and Enterprise Architecture was conducted by (Wilton<br />
2007) and (Beveridge and Perks 2003). These comparisons concluded that both ISSP and EA share the<br />
same intent and scope. (Wilton 2008) gave a more empirical comparison based on a survey which led to<br />
establishing a significant correlation between the two activities in term of topics they cover.<br />
The main difference that was highlighted by (Wilton 2008) is that ISSP tends to be process-oriented with<br />
little specification of deliverables and content while EA is product-oriented in that it defines the way the<br />
as-is and to-be state are described and modelled.<br />
We think that this difference tends to disappear due to the progress made in the field of Enterprise<br />
Architecture. As a matter of fact, with frameworks like TOGAF the gap is being bridged with a detailed<br />
process to produce architecture deliverables.<br />
Furthermore, we think that other differences are to be considered. They are summed up in the table<br />
below<br />
Table 2: Comparison enterprise architecture and IS strategic planning<br />
IS Strategic Planning Enterprise Architecture<br />
Period of development 1970-1995 1987-2010<br />
Process Linear Iterative and incremental<br />
Nature Business oriented Business and IT oriented<br />
Main deliverables Programs, projects, budgets Architecture description (as-is and<br />
to-be) and Roadmap<br />
Strategy support High Medium to low<br />
Techniques and modeling Techniques Techniques, Modeling<br />
Repository No Yes<br />
Tool support Rare Mainstream<br />
One of the main differences that still exists and that is related to the Enterprise Architecture practice is<br />
the fact that there is no concrete link between the architecture description and the programs/projects<br />
defined in the roadmap. This lack of correlation makes it difficult to address the strategic planning main<br />
objective which is planning for IT transformations with existing Enterrise Architecture frameworks.<br />
2.4 Bridging concept: Transformation<br />
A project is defined by (PMBoK) as “A project is a temporary endeavor undertaken to create a unique<br />
product, service or result.”. This definition underlines the fact that a project is intended to create a<br />
266
Mouhsine Lakhdissi and Bouchaib Bounabat<br />
product/service or result without mentioning the elements that the project will impact whether they are<br />
new elements created or existing element transformed.<br />
A project –in the context of ISSP and EA- could be defined as a set of transformations (including<br />
creations) applied on architecture elements. These elements could be business elements, application<br />
elements, data elements or technical elements or a combination of them.<br />
Elements are the basic constituents of architecture like applications, processes, servers, databases...etc.<br />
These elements are combined to create architecture models and diagrams. The transformations of these<br />
elements are combined as well to create ISSP’s projects and programs.<br />
3. Importance of a content framework and a metamodel<br />
The ISO/IEC 42010: 2007 definition of architecture as “the fundamental organization of a system,<br />
embodied in its components, their relationships to each other and the environment, and the principles<br />
governing its design and evolution.’’ highlight unequivocally the importance of the organization of<br />
elements and their relationships. This structure is defined through a metamodel of architecture elements.<br />
Enterprise Architecture is supposed to produce architecture artifacts. These artifacts are based on an<br />
architecture content framework as defined by TOGAF or an architecture map.<br />
3.1 Content framework<br />
The content framework defines the layers, views, questions and aspects that architecture description<br />
deals with. The importance of this framework is that it organizes, classifies and links architecture<br />
elements and artifacts. It is also interesting because it ensure the coherence and exhaustively of the<br />
metamodel.<br />
The content framework is classically defined as a bi-dimensional grid with lines representing layers or<br />
views and columns representing concerns and classifications.<br />
The content framework defines elements of the metamodel in a high level way emphasizing the global<br />
structure rather than the detail model.<br />
3.2 Metamodel<br />
The metamodel is the backbone of architecture description and methodology. The metamodel<br />
guarantees the exhaustiveness of overall architecture work and the coherence and alignment of<br />
architecture layers.<br />
It is similar in form to a Conceptual Data Model or a Class Diagram in UML. It is important in term of<br />
objects definition, attributes definition and relationships.<br />
Objects definition ensures the exhaustiveness and coverage of aspects like standardization,<br />
requirements and integration.<br />
Attributes provide the way to perform diagnosis and analysis on existing and future assets. Attribute<br />
can also cover aspects like security and performance necessary to the evaluation process.<br />
Relationships are very important to perform Gap Analysis inside the same layer and for alignment<br />
needs between layers.<br />
4. Existing content frameworks and metamodels<br />
Many metamodels have been defined explicitely or implicitely by EA frameworks. They are of different<br />
natures and focus depending on their intent. Some of them are poor in term of business or IS content.<br />
Others don’t take into account some aspect tightly related to EA and ISSP like :<br />
Requirements<br />
Strategy<br />
Standards<br />
Program and projects<br />
267
4.1 TOGAF metamodel<br />
Mouhsine Lakhdissi and Bouchaib Bounabat<br />
TOGAF metamodel which has been introduced only in the last version of the framework, is a theoretical<br />
metamodel somehow poor especially in the application, data and infrastructure layers. The metamodel of<br />
TOGAF describe elements of the four layers with their relationships and dependencies. The business<br />
layer is otherwise well detailed and provide for a good foundation for our metamodel. In the business<br />
layer nevertheless, only the static information is represented (no information flow) and the link between<br />
organization and data is not defined.<br />
We think that one of the contributions of this metamodel is its extensibility since TOGAF defines a core<br />
metamodel and many extensions applicable depending on the organization and project nature.<br />
Furthermore, some aspects like requirements, standards and project are not taken into consideration in<br />
the metamodel even if they are core in the framework<br />
4.2 Archimate<br />
Archimate is a structure and notation for architecture work. If used as a notation, it can be considered as<br />
complementary to frameworks like TOGAF or Zachman. Otherwise Archimate defines its proper content<br />
structure and metamodel. The structure is composed of three layers: Business, Application and<br />
Technology, with three views: passive structure, behavior and active structure.<br />
Archimate defines in a more detailed manner than TOGAF, the architecture elements of the three layers.<br />
In the business layer, it’s quite complete while still lacking definition of dynamic information flow and<br />
relationship between organization and information.<br />
In the application layer, it introduces integration elements like “application interface” and “application<br />
collaboration” but lack the underlying data integration elements (data flows, data messages…)<br />
One of the advantages of Archimate is its well defined structure combined with the detailed notation. It<br />
has intentionally -as stated in the specification document- omitted the strategy, standards and principles<br />
and strategic planning aspect. Other aspects like security and performance are also absent. They will add<br />
some of this aspect in future versions or as optional extensions.<br />
4.3 Zachman<br />
Zachman doesn’t define explicitly a metamodel. The metamodel is an implicit component appreciable<br />
through the models and artifacts defined in the Zachman grid.<br />
Through the different viewpoints and question, we can conclude that the Zachman metamodel has a<br />
large scope with detailed business and IT description. The main problem of Zachman is that it doesn’t<br />
define the relationship and dependencies between artifacts and thus between architecture elements<br />
which is a major drawback.<br />
Besides, Zachman doesn’t take into consideration Strategic Planning elements and their link with<br />
architecture elements.<br />
4.4 EA tools metamodel<br />
EA modeling tools have their own metamodel which are more often extensible. The architecture<br />
description is more or less detailed depending on the vendor positionning and history. Tools like Mega<br />
and Aris have more focus on the business layer with poorer IS capabilities, while tools like Power<br />
Designer are more relevant for IS and infrastructure description than in business modeling.<br />
Even though ISSP elements like projects can be defined and described, there is no link between the<br />
projects and architecture elements they impact.<br />
268
Mouhsine Lakhdissi and Bouchaib Bounabat<br />
5. Presentation of a new content framework and metamodel<br />
5.1 New content framework: NEOXIA Architecture Map (NAM)<br />
We introduce here a new content framework that is based on feedback from consulting projects on EA<br />
and ISSP. We call it NEOXIA Architecture Map. In this content framework we differentiate between<br />
Static element: tending to describe an element in a static most often hierarchical way<br />
Dynamic element : focusing on the dynamic view of the same element<br />
Figure 1: New content framework: Neoxia Architecture Map (NAM)<br />
We also distinguish between three natures of element:<br />
Structure elements: like organization and network<br />
269
Function elements: like services or functions<br />
Content element: like data or storage<br />
Mouhsine Lakhdissi and Bouchaib Bounabat<br />
Two additional but crucial architecture concerns emerge as intermediate planes between these elements:<br />
Security and integration.<br />
Figure 2: Architecture concerns in Neoxia Architecture Map (NAM)<br />
Between every layer and the one below we have an intermediate layer that is not shown in figure 1 but<br />
that is important to detail in some cases.<br />
As a matter of fact there is:<br />
A functional architecture layer between Business and the IS Architecture,<br />
A solution and development architecture layer between the IS and Software Architecture,<br />
A deployment and production architecture layer between the software and hardware layer.<br />
5.2 New Metamodel: NEOXIA Content Metamodel (NCM)<br />
The content metamodel is the mechanism by which we suggest to mix architecture and strategic planning<br />
element base on transformations. The content metamodel follows the overall structure of the content<br />
framework and could be illustrated as in figure 3.<br />
The suggested metamodel is composed of six layers:<br />
Strategy<br />
Requirements<br />
Business<br />
Information Systems<br />
Technology<br />
Strategic planning<br />
All layers are interrelated with static and dynamic element of the three natures: function, structure and<br />
content. Every layer is connected with the layer below with a realization link. A process is automated in<br />
an application which uses a database and are both deployed in a server. This dependence is<br />
fundamental to align the IS with the Business Architecture and the Technology with the IS Architecture.<br />
This link allows us also to analyze the gap between layers in term of coverage to make it possible to fill<br />
this gap in the strategic plan.<br />
The metamodel could be also represented as package and class diagrams. We focus on the strategic<br />
planning layer.<br />
The central concept is “Transformation” which is a migration from an as-is state to a to-be state of an<br />
architecture element. An architecture element could be any architecture object of the metamodel (ex:<br />
process, application, Hardware server…etc).<br />
270
Mouhsine Lakhdissi and Bouchaib Bounabat<br />
Figure 3: New Metamodel : NEOXIA Content Metamodel (NCM)<br />
is composed of<br />
defines<br />
is a set of<br />
originates from<br />
Program Project Transformation<br />
execute<br />
I nt er n al<br />
Or g a ni z ati o n<br />
Uni t : 2<br />
Impact<br />
Requirement<br />
originates from<br />
Strategic Objective : 2<br />
Figure 4: Class diagram of the package “IT Strategic Planning”<br />
originates from<br />
Gap<br />
concerns<br />
concerns<br />
Architecture Element<br />
A transformation is operated either:<br />
A realization of a strategic objective : this allows us to align the to-be IS situation with the strategy<br />
and to justify the strategic plan investments<br />
271
Mouhsine Lakhdissi and Bouchaib Bounabat<br />
A consequence of an IT or business requirement (principle, standard, rule, constraint) defined by the<br />
organization<br />
A result of a gap analysis: in that case the gap is observed on one or more architecture elements and<br />
the transformation is a way to fill the gap.<br />
We have as a result three types of transformations :<br />
Alignement transformation<br />
Requirement transformation<br />
Gap transformation<br />
Many transformations are grouped into projects which are managed through programs.<br />
6. Comparison with existing content framework and meta models<br />
The suggested metamodel defines the detailed structure and relations of architecture elements starting<br />
from strategy and requirement and going through the different levels (Business, Information Systems and<br />
Infrastructure) with the necessary link with Strategic Planning element like gap analysis, program and<br />
project.<br />
In addition of giving a more detailed structure for IS and infrastructure levels which are often poorly<br />
defined, the main contribution of this metamodel is the link it establishes between architecture elements<br />
and strategic planning elements.<br />
We summarize a theoretical comparison of our metamodel with existing metamodels presented in section<br />
3.<br />
Table 3: Comparison of the new metamodel with existing ones<br />
New metamodel TOGAF Archimate Zachman EA Tools<br />
NCM<br />
metamodels<br />
Requirement Yes No No.<br />
Partially Yes<br />
Undergoing<br />
Strategy Yes Yes No Yes No<br />
IT Planning Yes No No No Yes. To some<br />
Link between<br />
projects and<br />
architecture<br />
Yes No No No<br />
extent<br />
No<br />
Standards Yes No No No Yes. (Not<br />
native)<br />
Strategy definition Yes Yes No Yes Yes<br />
Business<br />
Detailed Detailed Detailed Average Poor<br />
description<br />
IS description Detailed Poor Average Detailed Detailed<br />
(Depending on<br />
Infrastructure<br />
description<br />
tool)<br />
Detailed Poor Average Detailed Detailed<br />
(Depending on<br />
tool)<br />
Tool support No Partial Yes Partial Yes<br />
Methodology<br />
support<br />
No. Undergoing Yes Yes No No<br />
Independence Yes Yes Yes Yes No<br />
7. Conclusion and future work<br />
We think that based on this proposed metamodel, a new methodology could be defined to cope with the<br />
needs of ISSP and to complement and enrich existing EA metamodels. The metamodel described was<br />
already used successfully in consulting projects in the public and private sector and was able to capture<br />
more meticulously architecture element and to support the process of IS Strategic Planning.<br />
The model could be enriched to highlight crossover architecture aspects like security, performance and<br />
integration. These aspects are very important in evaluating existing IT assets and in defining their target<br />
state.<br />
272
Mouhsine Lakhdissi and Bouchaib Bounabat<br />
A planned continuation of this work is to develop a basic modeling tool (or adapt an existing one) based<br />
on this metamodel and content framework with support to Enterprise Architecture as well as IS Strategic<br />
Planning techniques and activities. The tool will allow to put into practice the metamodel and to<br />
demonstrate the added-value of the methodology.<br />
Another extension is to formalize diagnosis and evaluation techniques into the meta model to make sure<br />
the whole IS Strategic Planning process is automated.<br />
References<br />
Archimate (2009), “Archimate 1.0 Specifications “, http://www.opengroup.org/architecture/togaf8-doc/arch/<br />
Arnoldo C. Hax (1996), “Strategy Concept and Process: A Pragmatic Approach” Barnes and Noble, bn.com<br />
Boar, B. (2001). The Art of Strategic Planning for Information Technology, John Wiley and Sons, New York, NY.<br />
Bounabat, B. (2006), “Enterprise Architecture Based Metrics for Assessing IT Strategic Alignment”, <strong>European</strong><br />
<strong>Conference</strong> On Information Technology Evaluation (ECITE), 2006.<br />
CCTA (1999). IS Strategy: process and products, Format Publishing Limited, Norwich.<br />
Chandler, Alfred A., Jr. (1962), Strategy and Structure: Chapters in the History ofAmerican Industrial Enterprise, The<br />
MIT Press, Cambridge, MA<br />
DODAF, U.S. DoD (2003). DoD Architectural Framework Version 1.0,<br />
http://www.teao.saic.com/jfcom/ier/documents/DOD_architecture_framework_volume1.doc,<br />
Elhari K., Bounabat. B (2010), “Strategic Alignement Assessement Based on Enterprise Architecture”, ICIME 2010<br />
FEAF, 1999, “Federal Enterprise Architecture Framework Specification”, http://www.whitehouse.gov/omb/e-gov/fea/<br />
Lillehagen F., Karlsen D., Enterprise Architectures – Survey of Practices and Initiatives, Rapport Computas 2005.<br />
Lederer, A. L. and Sethi, V. (1988). The Implementation of Information Systems Planning Methodologies, MIS<br />
Quarterly, September 1988, 445-461.<br />
Merriam Webster Dictionnary (2010), Dictionnary definition of Methodology and Method http://www.merriamwebster.com/dictionary/method,http://www.merriamwebster.com/dictionary/methodology?show=0&t=1290517042<br />
Pant S., Hsu C. (1995), Strategic Information Systems Planning: A Review, 1995 Information Resources<br />
Management Association International <strong>Conference</strong>.<br />
Perks, C. and Beveridge, T. (2003). Guide to Enterprise IT Architecture, Springer-Verlag, New York.<br />
Porter, M. E. (1985). Competitive Advantage: Creating and Sustaining Superior Performance, Collier Macmillan, New<br />
York, N.Y.<br />
Schekkerman J. (2005), Trends in Enterprise Architecture 2005: How are Organizations Progressing? Institute For<br />
Enterprise Architecture Development (IFEAD) report<br />
Sessions R. (2007), “A Comparison of the Top Four Enterprise-Architecture Methodologies”,<br />
http://msdn.microsoft.com/en-us/library/bb466232.aspx<br />
TEAF, U.S. Department of the Treasury (2000). Treasury Enterprise Architecture Framework Version 1,<br />
http://www.software.org/pub/architecture/teaf.asp,<br />
The Open Group (2009). “TOGAF as an Enterprise Architecture Framework”,<br />
http://www.opengroup.org/architecture/togaf8-doc/arch/<br />
Vitale, M., Ives, B. and Beath, C. (1986), “Identifying Strategic Information Systems,” Proc. 7th Int’l Conf. Inf. Sys.,<br />
San Diego, December 1986, pp. 265-276.<br />
Wilton, D. (2001). The Relationship Between IT Strategic Planning and Enterprise Architectural Practice, Journal of<br />
Battlefield Technology, 1, 18-22.<br />
Wilton, D. (2007). The Relationship between IS Strategic Planning and Enterprise Architectural Practice: a Study in<br />
NZ Enterprises, Information Resources Management Association (IRMA), Vancouver.<br />
273
Applying Innovative Information System to Improve the<br />
Effects of Community Health Promotion<br />
Chih-Yu Lin 1 , Wen-Yu Ho 2 , Chia-Hsien Wen 3 and Hui-Mei Hsu 4<br />
1<br />
National Chung Cheng University, Chiayi, Taiwan<br />
2<br />
China Medical University Beigang Hospital, Yunlin, Taiwan<br />
3<br />
Providence University, Taichung, Taiwan<br />
4<br />
National Kaohsiung Normal University, Kaohsiung, Taiwan<br />
cylin@ccu.edu.tw<br />
shiaoyin0505@gmail.com<br />
chwen@pu.edu.tw<br />
hmhsu@nknu.edu.tw<br />
Abstract: “Ottawa Charta” indicates that the importance of strengthen community actions is to achieve better health<br />
for residents. For achieving this goal, the health promotion works through community organizations is the most<br />
effective way. There are two models to promote community health in Taiwan. One is to set up a single community<br />
health department within a healthcare organization. The other one is to set up a community care center in the<br />
community by civic or government organizations. However, there is a common drawback of lacking a suitable and<br />
effective information system to evaluate the two models and serve as a basis to adjust future service content and<br />
efficiency. This study establishes an Internet information platform with Interactive Response System (IRS) to<br />
effectively evaluate the two models of community health promotion and to improve the services to upgrade the<br />
management of community health promotion efficiency. The results showed that the average satisfaction score of the<br />
Internet information platform is greater than 4 which mean that the innovative information system in this study is in<br />
accordance with the needs of community health department users and acquires a high satisfaction. The average<br />
satisfaction score of IRS is greater than 4 with the factor of “perceived usefulness” scores 5 which means residents of<br />
the community care center show a higher acceptance of using IRS in learning health education materials than that of<br />
traditional one. IRS can also improve diseases perception of community residents.<br />
Keywords: information system to improve community health promotion, community health promotion effective,<br />
interactive response system (IRS)<br />
1. Introduction and background<br />
In 1995, the Taiwan government implemented the National Health Insurance (NHI) System to provide a<br />
sound healthcare service for the citizens. According to the report of Department of Health, cancer and<br />
chronic diseases were the most serious health threats for the citizens in 2006. 78% of the funds were<br />
spent on the treatment of patients with chronic diseases. Therefore, patients who suffer from chronic<br />
diseases or have needs of long term care are the financial heavy burdens of NHI.<br />
Most citizens do not visit the doctor’s unless they are affected by a disease; this is the nature of the human<br />
beings. “Ottawa Charta” indicates that the importance of strengthen community actions is to achieve better<br />
health for residents. For achieving this goal, the health promotion works through community organizations<br />
is the most effective way. Health promotion includes promoting the cognition of health of individuals and<br />
communities, altering their attitudes for changing their behaviors of life, and seeking the methods to<br />
improve their health (Squyres, 1985).<br />
Many countries have enforced laws to protect the health of their people. For instance, Australia has<br />
enforced the Health Promotion Act in 1995; United States of America passed The Model State Public<br />
Health Act in 2003. With the rapid progress of information technology (IT), many enterprises, schools,<br />
hospitals and other organizations have employed IT to achieve their strategies and goals.<br />
There are two models to promote community health in Taiwan. One is to set up a single community health<br />
department within a healthcare organization. The other one is to set up a community care center in the<br />
community by civic or government organizations. However, there is a common drawback of lacking a<br />
suitable and effective information system to evaluate the two models and serve as a basis to adjust future<br />
service content and efficiency.<br />
The objectives of this study are to develop the community health promotion system to collect and analyze<br />
the changes of health cognition and health status of residents by using interactive response system (IRS).<br />
The data also can help the healthcare providers to understand the health conditions of the communities.<br />
274
2. Community health promotion<br />
Chih-Yu Lin et al.<br />
According to the Ottawa Charter, health promotion is the process of enabling people to increase control<br />
over, to improve their health, and to reach a state of complete physical, mental and social well-being<br />
(Ottawa Charter, 1986 ). As proposed by the Ottawa Charter, many countries have established health<br />
promotion policy such as Japan. In order to encourage all citizens to be healthy, keep away from diseases<br />
and get health information from the community, Japan announced Healthy Japan 21 in 2000. This policy<br />
focused on healthy dietary habits, promotion of physical activities, and diagnostic tests (Yajima, Takano,<br />
Nakamura & Watanabe, 2001; Dorjsuren, 2008).<br />
There are two mechanisms for health management: one is initiative model, which the health organization<br />
and government takes the initiative for the citizen’s health promotion; second is passive model, which the<br />
citizens and community takes the initiative in promoting their own health. Regardless of which model, the<br />
same objective is the prevention of diseases occurrence. Therefore, the subjects of health promotion<br />
include individuals, groups and community residents. The strategies of health promotion include health<br />
education, behavior motives, organization reengineering, economic effectiveness, and policy changes to<br />
further influence the living environment and people’s behaviors and to improve health and quality of life<br />
ultimately.<br />
Hancock and Duhlru (1986) defined community health as: continuing to create and enhance the physical<br />
and social environment of the community, while strengthening the community resources so that people<br />
can support each other, the implementation of all its living, and the maximum potential. WHO (1994)<br />
defined a healthy community as: healthy communities by healthy people, healthy environment and healthy<br />
social integration as a whole should be able to continuously improve the physical and social environment,<br />
expand community resources, support each other, and maximize potential. In brief, community health<br />
promotion is "health education" and "health policy", from the government, other non-governmental<br />
volunteer organizations, community participation, and hoping every community health issues for<br />
themselves, by residents spontaneous reaction to address community health problems, thus contributing<br />
to the community's health.<br />
The WHO (1978) proposed the concept of "National Health" that includes six principles, "strengthening<br />
community participation" is one of them. In addition, Ottawa Charter (1986) indicated that "strengthen<br />
community action" is one of the five health promotion actions. Both statements display community health<br />
promotion can be achieved through strengthening community actions. Community health building is a way<br />
to achieve community health. The community is regarding as a living body, and the whole community<br />
health is improved continuously through a planned action to change the health behaviors and decisions of<br />
community residents, and environments.<br />
The use of information technology in healthcare service has grown exponentially over the last several<br />
years (Pagliari C, Sloan D, Gregor P et al., 2005). Several studies noted that information technology was<br />
employed to improve health and health education. Heilman et al. (2011) indicated that Wikipedia is a key<br />
tool for public health promotion because the Internet has become an important health information resource<br />
for patients and the general publics. In 2004, Wikipedia editors formed a group called WikiProject<br />
Medicine to coordinate and discuss the Wikipedia’s medical content. Today, it has become the dominant<br />
online reference work. (Heilman et al., 2011). The Wantland’s study addressed important claims the use of<br />
web-based interventions compared to non-web-based interventions. The results indicated that the<br />
specified behaviors of individuals were changed by using web-based interventions, such as increasing<br />
participation in healthcare or improving body shape perception (Wantland, Portillo, Holzemer, Slaughter, &<br />
McGhee, 2004 ).<br />
3. Methods<br />
3.1 Case background<br />
The scale of case hospital is 429 beds, and the employees include 51 doctors, 195 nurses, 49 other<br />
medical staff, and 125 administrative personnel. We established a large group in this hospital for<br />
community service, which has been further divided into 22 smaller teams according to the three different<br />
structures: 1. Biopsychosocial, 2. Community- Oriented Primary Care, and 3. Integrated Delivery Systems.<br />
In order to provide a higher quality service, the care station was set up in the community.<br />
275
3.2 System design and implementation<br />
Chih-Yu Lin et al.<br />
In this study, we adopted the prototype model to develop the community health promotion system. Since<br />
system users are mainly healthcare workers, thus they do not have enough knowledge about the system<br />
development. This system includes two subsystems: (1) Community health promotion website (CHPS)<br />
which contains the diseases and health knowledge, Such as diabetes and hypertension. (2) IRS system.<br />
For example, the teachers can use this system to get the feedback from community residents on their<br />
health education progress. The system architecture has shown schematically in figure 1. The system is<br />
developed by using the PHP programming language and MYSQL database.<br />
Figure 1: The system architecture<br />
3.3 Survey instrument<br />
In order to evaluate the system performance, the system was evaluated in two parts: one is on the website<br />
effectiveness, and the other is on the IRS system. We surveyed the user's satisfaction for website and the<br />
IRS by distributing questionnaires to various members and IRS users.<br />
Both questionnaires were developed based on the literature about system satisfaction measurement and<br />
then sent to expert panel for review before being distributed to various members and IRS users. All items<br />
are structured on a Likert five-point scale ranging from 1 (strongly disagree) to 5 (strongly agree) in both<br />
questionnaires.<br />
There are two parts in the website’s questionnaire. Part one of the questionnaire collects respondents’<br />
demographics such as occupation, gender, age, work experience, and education level. Part two is the<br />
measure of end-user computing satisfaction (EUCS) developed by Doll and Torkzadeh and that is<br />
comprised of 12 questions (Doll and Torkzadeh, 1988).<br />
The questionnaire for IRS satisfaction includes four constructs such as perceived usefulness (PU),<br />
perceived ease of use (PEOU), total satisfaction, and perceived enjoyment. The items of first three<br />
constructs were modified from Wang’s study (Wang, 2003), and the items of perceived enjoyment were<br />
based on Agarwal and Karahanna’s research (Agarwal and Karahanna, 2000).<br />
4. Results<br />
4.1 The results of the CHPS website’s satisfaction<br />
The majority of users for CHPS website were hospital employees who join our health promotion activities.<br />
These users included physicians, nurses, other medical staff, and administrative personnel.<br />
276
Chih-Yu Lin et al.<br />
A total of 165 questionnaires were mailed out and 141 returned. The majority of respondents were female<br />
(66.0%), aged over 30 (67.4%), nurses (39.0%), and work experience less than ten years’ (77.3%). The<br />
details of demographics about the respondents are summarized in Table 1.<br />
Table 1: The details of demographics information about the respondents<br />
Item Category (Percentage)<br />
Male (24.1%)<br />
Gender<br />
Female (66.0%)<br />
Missing value (9.9%)<br />
21-30 years old (32.6%)<br />
Age<br />
31-40 years old (44.7%)<br />
More than 40 years (22.7%)<br />
Less than 1 year (9.2%)<br />
1-10 years inclusive (68.1%)<br />
Work Experience<br />
More than 10 years (20.6%)<br />
Missing value (2.1%)<br />
Physician (7.1%)<br />
Nurse (39.0%)<br />
Occupation<br />
Other Medical staff (28.4)<br />
Administrative personnel (25.5)<br />
Senior high school (7.8%)<br />
College (84.4%)<br />
Education<br />
Graduate school (7.1%)<br />
Missing value (0.7%)<br />
We use the EUCS instrument to evaluate the user's satisfaction for the website. The EUCS instrument<br />
comprises five constructs for measuring end-user computing satisfaction: content, accuracy, format, ease<br />
of use, and timeliness. The average score of this questionnaire is 4.3, and all of the items in the<br />
questionnaire are higher than 4.18. That means most people were satisfied to the website. The results<br />
showed that PEOU was the dimension satisfied the most people. It means the operation of the website is<br />
easy for users. The reliability of the instrument was measured by Cronbach’s alpha. The α values is 0.901<br />
which means this questionnaire has high internal consistency. The results of descriptive statistics about<br />
the user's satisfaction for website are summarized in Table 2.<br />
Table 2: The results of descriptive statistics for the user's satisfaction for website<br />
Construct Item Mean S.D. AVE.<br />
Content<br />
Accuracy<br />
Format<br />
Ease of<br />
Use<br />
Timeless<br />
Does the website provide the precise information you need?<br />
Does the information content meet your needs?<br />
Does the website provide reports that conforms to exactly what you need<br />
?<br />
Does the website provide sufficient information?<br />
4.24 0.608<br />
4.18 0.628<br />
4.23 0.569<br />
4.26 0.630<br />
Is the website accurate?<br />
Are you satisfied with the accuracy of the website?<br />
4.29<br />
4.25<br />
0.606<br />
0.638<br />
4.27<br />
Do you think the output is presented in a useful format?<br />
Is the information clear?<br />
4.26<br />
4.31<br />
0.659<br />
0.623<br />
4.29<br />
Is the website user friendly?<br />
4.36 0.668<br />
Is the website easy to use?<br />
4.44 0.669<br />
4.40<br />
Do you get the information you need in time?<br />
4.33 0.615<br />
Does the website provide up-to-date information?<br />
4.27 0.685<br />
4.30<br />
Total 4.30<br />
277<br />
4.23
Chih-Yu Lin et al.<br />
The user's satisfaction for website among groups had significant difference. According to the analysis of<br />
LSD post hoc test, most of the medical professional groups (physicians, nurses, and other medical staff)<br />
have high satisfaction than administration personnel in 4 constructs (accuracy, format, ease of use and<br />
timeless), especially in accuracy. The result implies that the website is more suitable to the medical<br />
professionals because they have more medical knowledge than administration personnel to estimate.<br />
Users with different work experience also had different satisfaction to the website. Group of “< 1 years”<br />
working experience is more satisfied to the website than user who have “1-10 years” or “> 10 years”<br />
working experience in content, accuracy, format, and timeless. The results may be reasonable because<br />
people with more working experience may have higher competence to know what they really need. The<br />
details of different satisfaction among groups are summarized in table 3<br />
Table 3: The details of different satisfaction between groups<br />
Content<br />
Accuracy<br />
Format<br />
Ease of Use<br />
Timeless<br />
Item<br />
*P < 0.05<br />
Occupation Gender Age<br />
1. Physician<br />
2. Nurse<br />
3. M.T.<br />
4. Executives<br />
1. Female<br />
2. Male<br />
1. < 40 years<br />
2. 31-40 years<br />
3. > 40 years<br />
Work<br />
Experience<br />
1. < 1 years<br />
2. 1-10 years<br />
3. > 10 年 years<br />
Education<br />
1. Senior high<br />
school<br />
2. College<br />
3. Graduate school<br />
F value Test F value Test F value Test F value Test F value Test<br />
Q1 0.923 — 8.089 — 0.728 — 1.852 — 0.977 —<br />
Q2 2.621 — 0.284 — 2.101 — 1.485 — 0.118 —<br />
Q3 2.084 — 4.506 — 1.462 — 3.369*<br />
1>2<br />
1>3<br />
0.887 —<br />
Q4 1.286 — 0.134 — 2.773 — 2.122 — 0.597 —<br />
Q5 4.461*<br />
2>4<br />
3>4<br />
0.101 — 3.420* 2>3 3.430*<br />
1>2<br />
1>3<br />
0.181 —<br />
Q6 3.119* 3>4 1.743 — 0.212 — 3.095* 1>3 1.129 —<br />
Q7 2.864*<br />
2>4<br />
3>4<br />
2.853 — 0.335 — 3.666* 1>2 0.869 —<br />
Q8 1.459 — 3.280 — 4.176*<br />
2>1<br />
2>3<br />
2.207 — 0.244 —<br />
Q9 4.376*<br />
Q10 4.767*<br />
Q11 3.049*<br />
1>2<br />
1>4<br />
3>4<br />
1>4<br />
2>4<br />
3>4<br />
1>4<br />
3>4<br />
3.669 — 0.015 — 1.021 — 0.605 —<br />
1.021 — 1.651 — 1.182 — 1.158 —<br />
5.397 — 1.130 — 0.922 — 0.470 —<br />
Q12 1.176 — 0.044 — 0.211 — 4.355*<br />
278<br />
1>2<br />
1>3<br />
0.362 —
Chih-Yu Lin et al.<br />
4.2 The results of the CHPS for IRS satisfaction<br />
The majority of users for IRS are community residents who came to care station to participate the health<br />
promotion activities, there are 27 subjects in our study.The instrument comprised four constructs for<br />
measuring the user's satisfaction for IRS: perceived usefulness, perceived ease of use, perceived<br />
enjoyment, and total satisfaction. The average score of this questionnaire was 4.74, and all items were not<br />
less than 4.22 score, which means most residents are satisfied to use the IRS in the health education<br />
program. The results of descriptive statistics for user's satisfaction of IRS are summarized in Table 4.<br />
Table 4: The results of descriptive statistics for IRS satisfaction.<br />
Construct Item Mean S.D. AVE.<br />
Perceived<br />
usefulness<br />
Perceived<br />
ease of use<br />
Perceived<br />
enjoyment<br />
Total<br />
satisfaction<br />
5. Limitation<br />
The IRS system makes it easy for you to discuss questions with your<br />
teachers.<br />
The content provided by the IRS system is easy to understand.<br />
It is easier to concentrate on the content when the IRS system is used.<br />
4.93 0.385<br />
5.0 0.0<br />
4.93 0.267<br />
The feedback from the IRS system allows the teacher to understand<br />
how well I have educated myself about the diseases.<br />
4.85 0.456<br />
I would like to request the teachers to use the IRS system as frequently<br />
as possible during the classes.<br />
If a teacher uses the IRS system during classes, I will continue to use<br />
4.52 0.509<br />
the IRS system. 4.52 0.509<br />
The remote control function of IRS system is easy to use. 4.41 0.501<br />
The operation of the IRS system is stable. 4.22 0.506<br />
Using the IRS system during classes makes the class more interesting.<br />
4.89 0.320<br />
The IRS system creates a more interactive environment in the class. 4.89 0.320<br />
4.79<br />
4.32<br />
4.89<br />
As a whole, you are satisfied with the IRS system. 4.96 0.192 4.96<br />
Total 4.74<br />
Due to the restriction of resources and time limit, we are currently using it in two different communities. By<br />
using the system we would like educate the communities on the healthy living and how to prevent<br />
ourselves from various diseases. As we are currently using them in two communities, therefore we are<br />
unable generalize the feedback of these two communities to represent the feedback of a larger population.<br />
Therefore, we are not able to identify the performance of the system on a larger scale.<br />
Most of the subjects are currently from community residents who are 60 years old and above and have a<br />
low literate rate. Therefore, it must be spent more time consuming to educate the older community<br />
residents to use our system. If a health organization would like to follow our method, we would like to<br />
recommend them to use them on people who have at least elementary education so the effectiveness of<br />
community residents’ self-health management will reduce the time consumed to understand how to use<br />
our system.<br />
6. Conclusion<br />
The purpose of IRS that applied on the community health promotion and health education is to build a<br />
friendly learning environment to improve community residents’ cognition of health promotion. Through a<br />
teaching tool of IRS, the departments related to community health promotion can be integrated<br />
horizontally and they can support the community health promotion programs continuously. In addition,<br />
building a effective management system of community residents’ health according to the learning<br />
effectiveness and developing a health promotion education program and a community website which are<br />
proper to community characteristic are in order to promote the health care of community residents.<br />
279
Chih-Yu Lin et al.<br />
The results of this study could be a reference for healthcare institutions. It can improve effectiveness for<br />
healthcare institutions to adopt IT to community health promotion activities. Therefore, healthcare<br />
institutions will involve in community health promotion activities actively.<br />
Applying IT into community health promotion activities properly can help healthcare institutions to manage<br />
self-care quality of community residents effectively. Therefore, it might reduce the incidence of chronic<br />
diseases and the medical expenditure on chronic diseases and long-term care.<br />
Combining IT in community health promotion activities in this study could be a reference model. We<br />
suggest that developing a better community health management information system and technology can<br />
improve the health of community residents continuously and conform to the spirit of sustainable<br />
development of community health promotion. In further, it also can fulfill the health policies of localized<br />
medical care, localized health service, localized aging, and preventive medicine.<br />
References<br />
Agarwal, R. and Karahanna, E. (2000) Time Flies When You’re Having Fun: Cognitive Absorption and Beliefs about<br />
Information Technology Usage. Management Information Systems Quarterly, Vol 24, No. 4, pp 665-694.<br />
Doll, W.J. and TorKzadeh, G. (1988) The Measure of End-User Computing Satisfaction. Management Information<br />
Systems Quarterly, Vol 12, No. 2, pp 259-274.<br />
Dorjsuren, B. (2008) Financing health promotion in Japan and Mongolia. Bulletin of the World Health Organization,<br />
Vol 86, No. 11, pp 896-897.<br />
Hancock, T. and Duhl, L. (1986) Promoting Health in the Urban Context.Copenhagen, WHO Regional Office for<br />
Europe (Healthy Cities Paper No.1).<br />
Heilman, J.M., Kemmann, E., Bonert, M., Chatterjee, A., Ragar, B., Beards, G.M., Iberri, D.J., Harvey, M., Thomas, B.,<br />
Stomp, W., Martone, M.F., Lodge, D.J., Vondracek, A., de Wolff, J.F., Liber, C., Grover, S.C., Vickers, T.J.,<br />
Meskó, B. and Laurent, M.R. (2011). Wikipedia: A Key Tool for Global Public Health Promotion, Journal of<br />
Medical Internet Research, Vol 13, No. 1, e14.<br />
Pagliari, C., Sloan, D., Gregor, P., Sullivan, F., Detmer, D., Kahan, J.P., Oortwijn, W. and MacGillivray, S. (2005) What<br />
is eHealth (4): a scoping exercise to map the field. Journal of Medical Internet Research, Vol 7, A9.<br />
Squyres, W.D. (1985) Patient Education and Health Promotion in Medical Care.Palo Alto, CA: Mayfield Publishing Co.<br />
Ottawa Charter (1986), WHO.<br />
Wang, Y.S. (2003) Assessment of Learner Satisfaction with Asynchronous Electronic Learning Systems. Information<br />
& Management, Vol 41, pp 75-86.<br />
Wantland, D.J., Portillo, C.J., Holzemer, W.L., Slaughter, R. and McGhee, E.M. (2004) The effectiveness of<br />
Web-based vs non–Web-based interventions: a meta-analysis of behavioral change outcomes. Journal of<br />
Medical Internet Research, Vol 6, No. 4, e40.<br />
WHO(1994). Action for health in cities. Copenhagen,WHO Regional Office for Europe.<br />
WHO(1978). Alma Ata Declartion.Geneva.<br />
Yajima, S.H., Takano, T. Nakamura, K. and Watanabe, M. (2001) Effectiveness of a community leaders’ programme<br />
to promote healthy lifestyles in Tokyo, Japan. Health Promot International, Vol 16, pp 235-243.<br />
280
Information Security Practices in Zambian Copper Mines: An<br />
Investigation Into the State-of-Practice of Information<br />
Security Within Zambian Copper Mines Based on the ISO/IEC<br />
27002 Standard<br />
Chishala Lukwesa and Christopher Upfold<br />
Rhodes University, Grahamstown, South Africa<br />
g09l2525@campus.ru.ac.za<br />
c.upfold@ru.ac.za<br />
Abstract: Information has become an important aspect of daily operations in the mining industry. This is due to the<br />
mining industry’s increased dependence on information for strategic operations and competitive advantage.<br />
Therefore, the need for reliable and accurate information to achieve this, demands enhanced information security<br />
practices in these organisations. Sound information security practices require an effective information security<br />
program backed by information security policies and frameworks. The research aims to investigate the state-ofpractice<br />
of information security in Zambian copper mines using the ISO/IEC 27002 standard as the foundation of<br />
information security, and to explore leadership perceptions regarding information security practices within these<br />
mines. Available literature on information security practices within the mining industry is used to identify barriers to<br />
effective implementation of information security practices and a conceptual framework is proposed to this effect. This<br />
framework is then verified by responses obtained by means of a survey administered to senior management and<br />
heads of Information Technology in Zambian copper mines, and an in-depth case study of information security<br />
practices in one of the said Zambian copper mines using an ISO/IEC 27002 audit tool. Identified barriers to effective<br />
information security implementation include, differing perceptions between senior management and middle<br />
management, incomplete information asset inventory, poor staff awareness, inadequate information security policies<br />
and procedures, inadequate risk management processes, and inadequate control system policies and procedures. A<br />
framework for enhanced implementation of information security in Zambian copper mines is further proposed. This<br />
framework is made up of proposed considerations and enhancements that can be used to address the barriers<br />
identified in the literature review, based on results from the leadership survey and the ISO/IEC 27002 audit<br />
instrument. This framework, based on ISO/IEC 27002 guidelines and controls, includes full management<br />
commitment to information security, adoption of an information security standard or framework, improved personnel<br />
awareness, complete and up-to-date asset inventory, defined information security governance framework, and<br />
enhanced policies and procedures for industrial control systems.<br />
Keywords: information security, leadership perceptions, state-of-practice, copper mines, ISO/IEC 27002<br />
1. Introduction<br />
The mining industry is a continuous operational and strategic industry, which contributes to both local and<br />
international economies (Teseleanu et al., 2006). This industry, like other manufacturing industries, is<br />
data intensive as it deals with information about processes, product designs, shipping, inventory and<br />
customers (Paladion, 2008). Although sometimes considered to be low-tech with little innovation and<br />
appreciation for information technologies, the industry is best understood as a scale-intensive innovative<br />
industry with its own characteristics (Upstill and Hall, 2007; Pavitt, 1984). Although argued to be slowchanging<br />
and somewhat traditional, the face of the industry has been seen to change. Part of this change<br />
has been in the timing of mining information, which is becoming indispensable to sound policy making,<br />
strategy and forecasts for future changes (Sawada, 2004). The quality of information, its production, use,<br />
flow, accessibility and credibility have been identified as key to developing trust and cooperation amongst<br />
stakeholders in the mining industry. This is coupled with openness and greater transparency in<br />
information production and dissemination throughout the mineral life cycle for enhanced decision making.<br />
(MMSD, 2002: 296). Therefore, just as quality management standards such as ISO 9000 have become a<br />
requirement for doing business in industries such as the mines, internationally recognised information<br />
security standards continue to be accepted and will eventually become a necessity for many mines<br />
(Ernst&Young, 2008). One of these information security standards is ISO/IEC 27002. This standard is<br />
universally accepted and used globally. ISO/IEC 27002 has over the years become the de facto standard<br />
for high level definition of an information security management system as it defines a comprehensive set<br />
of controls that help organisations implement a good information security program (Wessels et al., 2007:<br />
253). ISO/IEC 27002 was used in this research as it provided a framework that was suitable for Zambian<br />
copper mining organisations. ISO/IEC 27002 encompasses 11 information security domains, which are<br />
universally accepted information security domains. These domains were all applicable to current<br />
information security practices in the Zambian copper mining industry.<br />
281
2. Literature review<br />
2.1 Risk management<br />
Chishala Lukwesa and Christopher Upfold<br />
Risk management in the mining industry has been concentrated on safety, environmental, occupational<br />
health, and quality management. With the advent of the information age, the risk management process<br />
has been extended to include physical IT assets. Information that is stored, transmitted and processed<br />
within the same IT infrastructure has, however, not yet been well covered (Pironti, 2008). In view of this,<br />
although the use of information is critical to effective risk management in the mining industry, information<br />
risk management itself has not been one of the high priority risk areas. An integrated or enterprise risk<br />
management approach, which aims to identify, evaluate and manage major corporate risks in an<br />
integrated framework, is, however, now being adopted by the mining industry. This is in order to<br />
encompass all risk areas and not just focus on safety, environmental, occupational health and quality<br />
areas (Fagg, 2009; Olson and Wu, 2008: 3).<br />
2.2 Contingency planning<br />
Organisations are built on similar principles but each has unique characteristics in the way business is<br />
conducted and products and services provided. There are certain components that particular<br />
organisations cannot function without. These components are a combination of people, processes and<br />
technology. There is at least one person, certain technology or information specific business processes<br />
and some critical infrastructure components that are critical to business operations. If their continuity is<br />
not catered for in times of manmade or natural disasters, the business might fail. Therefore, an<br />
organisation should identify the critical components that need to be protected (Seldon, 2009).<br />
Contingency planning in the mining industry covers a lot of areas and issues and may involve production,<br />
distribution, Information Technology (IT) and other functional areas (Musson, 2007: 185). All functional<br />
areas and not just IT should be catered for in the business continuity plan as they need to co-exist at any<br />
given time. Recovery in mining industries also presents challenges, which may not be found in other<br />
industries. These range from government regulations to certification and re-certification requirements<br />
such as ISO 9000 (Musson, 2007: 185; Warner, 2007). It is, therefore, vital for the mining industry to<br />
deploy a business continuity process that encompasses all areas of the enterprise including supplier and<br />
customer networks. Disaster Recovery Plans (DRPs) and/or Business Continuity Plans (BCPs), should,<br />
provide a solution that will account for and ensure the recovery of all business processes cater for an<br />
alternate workspace environment for employees, facilitate ongoing operations within a consistent<br />
timeframe and cover all regulatory requirements (Warner, 2007). Warner (2007) argues that in many<br />
cases, the amount of investment in a DRP or BCP could literally mean the difference between the<br />
business and its shareholders sinking or swimming. A Meta Group report in 2000 revealed that metal and<br />
natural resource industries can lose an average of $580,588 per hour in downtime costs. It is not only<br />
money that can be lost but also reputation, current and potential customers, and customer confidence<br />
(Hinton and Clements, 2002).<br />
2.3 Information Security Governance<br />
Peltier et al. (2005: 240) define Information Security Governance (ISG) as the management structure,<br />
organisation, responsibility, and reporting processes surrounding a successful information security<br />
program. According to Von Solms and von Solms (2009: 24) information security governance is also the<br />
system by which the confidentiality, integrity and availability of an organisation’s electronic assets are<br />
maintained. Mears and von Solms (2004) and Krehnke (2007: 41-43) define the factors that make up the<br />
pillars of information security governance. They all agree that these factors include accountability and<br />
responsibility, ethics, resource allocation and management, security awareness and education,<br />
information security policies, good practice standards, risk management, compliance with legal<br />
requirements, and information sharing.<br />
Employees are a critical factor in information security as they are the first line of defence in any operation<br />
within an organisation. Security Education, Training, and Awareness (SETA) is, therefore, an important<br />
aspect of any information security program (Whitman and Mattord, 2009: 206).<br />
The Zambian government has enacted a number of laws, policies and acts, which are meant to provide<br />
for information security and to which mining organisations and its employees like every other entity in the<br />
country should adhere to. These include the National ICT Policy, Electronic Communications and<br />
282
Chishala Lukwesa and Christopher Upfold<br />
Transactions Act, 2009, and the Copyright and Performance Rights Act No. 44 of 1994 (Ministry of<br />
Communications and Transport, 2006; National Assembly of Zambia, 2009; GRZ, 1994; Campbell, 2008:<br />
260).<br />
Employees in the mining industry also make use of mobile devices such as two-way radios (EPA, 2007),<br />
iPhones and Blackberry handsets and laptops, which create a security loophole in the organisation.<br />
Security strategies, employee training and awareness, and a revised IT infrastructure are required to help<br />
achieve effective mobile security governance (Zhang, 2009).<br />
Failure in corporate governance threatens the future of every organisation (Arjoon, 2005). Information<br />
security should, therefore, be an integral part of corporate governance, aligned with IT governance and<br />
integrated into strategy, concept, design, implementation and operation (ITGI, 2006: 15).<br />
2.4 Information Security standards<br />
Implementation of best practice standards is important in order to ensure that an organisation’s<br />
information security governance practices instil confidence in the stakeholders, shareholders and<br />
customers (Brotby, 2009: 84). Adoption of an information security standard complements an<br />
organisation’s information security policy and assists in making the implementation of the policy more<br />
effective (Peltier, 2002: 70).Just as quality management standards such as ISO 9000 have become a<br />
requirement for doing business in industries such as the mines, internationally recognised information<br />
security standards will continue to be accepted and eventually become a necessity for many mines<br />
(Ernst&Young, 2008). An organisation wanting to implement security controls which are in compliance<br />
with a standard or set of standards requires support from top management and other employees in the<br />
development and implementation process. These standardised policies and guidelines must be<br />
applicable to the organisation’s culture, business and operational practices (HKSAR, 2008).<br />
3. Information Security concerns<br />
Although Zambian copper mining organisations have implemented information security measures in their<br />
organisations, there are certain aspects of information security practices that are considered to require<br />
addressing. A summary of the concerns that were identified in the literature review as barriers to effective<br />
implementation of information security in the Zambian copper mining industry is outlined in table 1 below.<br />
Table 1: Summary of identified information security concerns in the Zambian copper mining industry<br />
Concern Author(s)<br />
1. Lack of or inadequate security policies Bradley, 2008; Welander, 2007<br />
2. Lack of or inadequate security procedures Bajpai and Gupta, 2005; Welander, 2007<br />
3. Poor staff awareness Bajpai and Gupta, 2005; DTT, 2008<br />
4. Inadequate risk management process Pironti, 2008<br />
5. Increased regulation PricewaterhouseCoopers, 2009; WIPO,<br />
2009<br />
6. Dispersed data and information Copans, 2007; PricewaterhouseCoopers,<br />
2008<br />
7. Email threats such as viruses and spam DTT, 2008<br />
8. Inadequate control system policies and procedures Lowe, et al., 2007; Welander, 2007<br />
9. Cultural clashes between IT staff Bartels, 2005<br />
10. Poor asset identification and inventory PricewaterhouseCoopers, 2008<br />
11. Lack of formal mobile security governance Zhang, 2009<br />
12. Irregular or no patching of systems Bartels, 2005<br />
13. Inadequate access control procedures Bajpai and Gupta, 2005<br />
14. Inadequate email policies Bradley, 2008<br />
15. Informal change management procedures Bartels, 2005<br />
16. Informal BCPs DTT, 2008<br />
17. Lack of or irregular testing of contingency plans DTT, 2008; ICMM, 2009<br />
18. Ill-defined crisis communication procedures ICMM, 2009<br />
19. Lack of formal information security governance frameworks DTT, 2008<br />
20. Inadequate integration of information security into<br />
DTT, 2008<br />
employment policies and practices<br />
21. Poor incident response management Helberger, 2009<br />
22. Unsecure application software Tomlinson, 2007: 12; Dobelis, 2007: 46<br />
23. Ill-defined user access privileges Van Holsbeck and Johnson, 2004;<br />
Kairab, 2004: 7<br />
283
Chishala Lukwesa and Christopher Upfold<br />
Concern Author(s)<br />
24. IT-based BCPs Warner, 2007<br />
25. Ill-maintained legacy systems Zoufaly, 2009; Welander, 2009; Shelton,<br />
2009<br />
26. Human error such as social engineering Tomlinson, 2007: 12; DTT, 2008; Piccoli,<br />
2008: 431<br />
27. Incorrect dissemination of information MMSD, 2002<br />
28. Environmental threats to hardware Falco et al., 2002<br />
These concerns were mapped against the eleven (11) ISO/IEC 27007 information security domains and<br />
controls to determine to what extent the ISO/IEC 27002 security domains cater for all or part of the<br />
security concerns.<br />
4. Research methodology<br />
The research was conducted by means of an instrumental case study and includes such data collection<br />
methods as surveys in the form of questionnaires, interviews and document reviews. According to Stake<br />
(1995: 3), an instrumental case study is used when there is a need for a general understanding and there<br />
is a possibility of getting insight into the question by studying a particular case. The case study method<br />
was chosen as it is an intensive study that sheds light on a larger class of classes and uses different<br />
methods to collect various kinds of information and make observations (Gerring, 2007:20; Hamel et al.,<br />
1993: 45). It is through these methods that the object of study is understood (Hamel et al., 1993: 45).<br />
Case study methods also answer the “how” and “why” of research questions and also focus on<br />
contemporary events (Yin, 2003: 5).<br />
Descriptive statistics were used for quantitative analysis of the data. Fisher and Marshall (2009) define<br />
descriptive statistics as “the numerical procedures or graphical techniques used to organise and describe<br />
the characteristics or factors of a given sample”. Two survey instruments were administered. One set of<br />
questionnaires made up of information security concerns identified in the literature review was<br />
administered to senior executives and heads of Information Technology departments in these mines.<br />
These questionnaires were meant to explore leadership perceptions surrounding information security<br />
practices in Zambian copper mines and were administered in four of these mines. An in-depth case study<br />
was then administered to personnel in charge of infrastructure in these mines by means of a<br />
commercially available survey instrument in the form of an ISO/IEC 27002 audit tool procured from<br />
Praxiom Limited. This representation of middle management included human resource and administrative<br />
personnel, and personnel in charge of the various sections of the mining organisation’s Information<br />
Technology department. The audit tool was administered in person by the researcher. Ethical, reliability,<br />
and validity issues surrounding the collection and analysis of data were addressed.<br />
5. Findings and recommendations<br />
Research findings and recommendations that were deemed to be most significant are discussed below<br />
and outlined according to the 11 ISO/IEC 27002 domains. These domains were categorised into<br />
administrative, technical and physical domains. Recommendations were made based on the responses<br />
from the two sets of questionnaires.<br />
Administrative domains include:<br />
5.1 Information Security policy<br />
Senior management believe information security policies in their organisations adequately address<br />
information security requirements, business and organisational requirements and that they also cater for<br />
changing business needs. Middle management, on the other hand felt that the information security policy<br />
was in need of attention. 75% of senior management believe information security policies have been<br />
implemented in their organisations while 25% do not. All these respondents further believe their<br />
organisation’s information security policies meet their information security requirements, business<br />
objectives, and business requirements. Middle management on the other hand revealed that the<br />
information security policy is not kept up to date.<br />
Recommendations<br />
Information security policies provide support for information security governance and serve as a<br />
backbone for information security programs in an organisation. Thus they should be regularly reviewed at<br />
284
Chishala Lukwesa and Christopher Upfold<br />
planned intervals or when a significant information security change is made in the organisation in order to<br />
ensure that they continue to reflect the organisation’s current business and information security<br />
requirements. Users should also be notified of subsequent updates to information security policies and<br />
procedures.<br />
5.2 Organising Information security<br />
Both senior management and middle management believe only part of organisational data has been<br />
consolidated. Therefore, not all data and information in these organisations can be accounted for. 25% of<br />
senior management firmly believe that all organisational data has been consolidated and accounted for in<br />
their organisation while 50% could neither agree nor disagree as to whether this was the case in their<br />
organisations. An additional 25% on the other hand, does not believe that all their organisation’s data and<br />
information have been adequately consolidated for it to be appropriately accounted for.<br />
Recommendations<br />
Information assets should be clearly identified, defined, and responsibility allocated for each information<br />
asset. This includes definition of responsibilities for specific information security processes such as<br />
business continuity. Allocation of responsibilities should be done according to the information security<br />
policy and further documented.<br />
5.3 Asset management<br />
Both senior and middle management believe asset management still requires attention.<br />
Recommendations<br />
All organisational information assets should be identified and inventoried. This inventory should be<br />
regularly updated and should include among other things, BCPs, contracts and agreements, archived<br />
information, research information, training materials and user manuals, and audit trails.<br />
5.4 Human resources management<br />
Senior management believe information security is adequately integrated into employment policies and<br />
procedures while middle management on the other hand felt that the organisation’s security roles and<br />
responsibilities are not documented or communicated in such areas as employees’ job descriptions.<br />
However, both senior and middle management agree that personnel do not receive adequate information<br />
security education, training, and awareness.<br />
Recommendations<br />
Effective management of personnel information security roles and responsibilities ensures that their use<br />
of information processing facilities is appropriately managed and minimises insider threats. Therefore,<br />
management should undertake the responsibility of ensuring that all personnel receive the appropriate<br />
level of awareness, training, and education, which is relevant to their job role, responsibility and skills.<br />
Job descriptions should also be used to document and address employees’ information security roles and<br />
responsibilities.<br />
5.5 Business continuity management<br />
Senior management endorses business continuity strategies in these organisations and certainly the<br />
impacts that failures could have on the organisations’ processes have been evaluated. However, the<br />
inclusion of information security requirements in the BCPs and regular reviews of these BCPs still require<br />
attention. Both senior and middle management are aware that some aspects of business continuity<br />
management still require addressing. 50% of senior management firmly believe that contingency plans to<br />
deal with system failures in their organisations have been established and are regularly tested and<br />
reviewed. 25% could neither agree nor disagree as to whether this is the case, and a further 25% do not<br />
believe these contingency plans have been established or are regularly reviewed and tested. Additionally,<br />
50% of senior management believe their organisations have detailed BCPs in place which specify actions<br />
that should be taken in the event of a disaster. A further 25% could neither agree nor disagree as to<br />
whether this was the case and 25% do not believe these detailed BCPs are in place in their<br />
organisations.<br />
285
Recommendations<br />
Chishala Lukwesa and Christopher Upfold<br />
Effective business continuity management reduces the chances of interruptions to business activities and<br />
minimises the impact these interruptions would have on the business. A single framework for BCPs<br />
should be adopted and maintained to ensure that all plans are consistent and that they consistently<br />
address information security requirements, along with the identification of priorities for testing and<br />
maintenance. These plans should also be regularly tested, reviewed, and updated to ensure that they are<br />
realistic and can be used in a real-life scenario.<br />
5.6 Compliance<br />
Senior management believes their organisations adhere to regulatory requirements while middle<br />
management on the other hand, felt that requirements for intellectual property rights are in need of<br />
attention. Middle management felt that procedures that deal with intellectual property rights and<br />
proprietary software products as well as user awareness regarding the protection of these rights still<br />
require attention.<br />
Recommendations<br />
Compliance with legal, statutory, contractual, and regulatory requirements is necessary for the<br />
organisation to avoid legislative implications. All regulatory, statutory, and contractual requirements that<br />
are applicable to the organisation should be explicitly identified, documented and updated regularly. This<br />
should also be done for each information system. Appropriate procedures to ensure compliance with<br />
legislative, contractual, and regulatory requirements regarding intellectual property rights and proprietary<br />
software products should also be implemented.<br />
5.7 Information Security incident management<br />
Both senior and middle management believe the establishment and implementation of incident response<br />
plans still requires attention. 50% of senior management firmly believe that their information security<br />
incident reporting responsibilities and channels have been defined while the other 50% could neither<br />
agree nor disagree as to whether these incident reporting responsibilities and channels have been<br />
defined in their organisations.<br />
Recommendations<br />
Information security incidents should be incorporated into personnel awareness programs as examples of<br />
possible incidents, and how they could be handled and prevented in the future. Additionally, procedures<br />
that would be used to handle different information security incidents should be established. These<br />
procedures should cater for, among others, malicious code, denial of service attacks, and breaches of<br />
integrity.<br />
Technical domains include:<br />
5.8 Access control<br />
Both senior and middle management believe mobile security governance requires attention. 50% of<br />
senior management believe mobile security governance measures have been put in place to address<br />
mobile computing and communications in their organisations while 25% could neither agree nor disagree<br />
as to whether this was the case in their organisations. A further 25% do not believe these security<br />
measures have been put in place.<br />
Recommendations<br />
The mobile computing and communication policy should be used to address all mobile devices in use in<br />
the organisation coupled with personnel training to raise awareness on the risks associated with mobile<br />
computing. This policy should include requirements for physical protection, access control, backup<br />
facilities, cryptographic controls, and virus protection.<br />
286
Chishala Lukwesa and Christopher Upfold<br />
5.9 Information Systems acquisition, development and maintenance<br />
Both senior and middle management believe emphasis on information security requirements during the<br />
acquisition, development, and implementation of information systems still requires attention. 50% of<br />
senior management believe that information security requirements are emphasised during the<br />
procurement and development of application software in their organisations. 25% could neither agree nor<br />
disagree as to whether this is so as there have been cases of violations in the procurement process. A<br />
further 25% believe these information security requirements are not emphasised in their organisations.<br />
Recommendations<br />
Controls to be used for ensuring authenticity and message integrity in applications should be identified<br />
and implemented. This includes implementation of cryptographic controls and development of a policy on<br />
the use of these controls. Cryptographic techniques such as digital signatures should be used to achieve<br />
confidentiality, authenticity and integrity, and non-repudiation of information.<br />
5.10 Communications and operations management<br />
Senior management believe adequate electronic communication policies and procedures have been<br />
developed in their organisations while middle management on the other hand, felt that electronic<br />
communication procedures and controls still require attention.<br />
Recommendations<br />
Formal policies, procedures, and controls should be used to address the control of the exchange of<br />
information through all communication facilities. These controls should include:<br />
Use of cryptographic controls in information exchange to ensure the authenticity, confidentiality, and<br />
integrity of information.<br />
Restrictions on the forwarding of electronic mails to external mail destinations.<br />
Physical domains include:<br />
5.11 Physical and environmental security<br />
Adequate protection of an organisation’s assets from physical and environmental threats helps prevent<br />
loss, misuse, theft or compromise to an organisation’s activities. While senior management believe their<br />
organisations’ equipment is protected from environmental threats, middle management on the other hand<br />
felt that although this was the case, security of equipment from physical threats still requires attention.<br />
Recommendations<br />
Physical access controls should be implemented to only allow authourised access to secure areas. This<br />
should include regular review and update of access rights to secure areas. All personnel and visitors<br />
should be required to wear visible identification and personnel should notify security personnel when they<br />
encounter anyone not wearing visible identification. Measures for the protection of information processing<br />
equipment used offsite should be enhanced and should include:<br />
Disguising portable computers when travelling.<br />
Using risk assessments to identify and further implement home working controls including clear desk<br />
policies.<br />
5.12 Summary of findings<br />
Findings from senior management revealed that information security concerns deemed to be most<br />
prevalent in Zambian copper mines include:<br />
Poor staff awareness.<br />
Inadequate control system policies and procedures.<br />
Findings from the in-depth analysis of information security practices in a Zambian copper mine revealed<br />
the following information security concerns:<br />
Lack of or inadequate information security procedures<br />
287
Poor staff awareness<br />
Poor asset identification and inventory<br />
Chishala Lukwesa and Christopher Upfold<br />
Lack of formal mobile security governance frameworks<br />
IT-based Business Continuity Plans<br />
Inadequate email policies<br />
Informal BCPs<br />
Ill-maintained legacy systems<br />
These findings revealed that in most cases, senior management perceptions do differ from those of<br />
middle management. This implies that senior management may not be fully committed to the information<br />
security process within these organisations or may not be fully aware of the extent to which their<br />
organisations meet International best practice standards such as those outlined in the ISO/IEC 27002<br />
standard.<br />
6. Conclusion<br />
Zambian copper mining organisations face numerous challenges when it comes to safeguarding their<br />
information assets. These organisations have evolved from a background where business efficiency<br />
mostly affected the physical component of business operations to one where newer technologies have<br />
led to an increased dependence on information for efficient and strategic operations This evolution<br />
requires that management take up the ultimate responsibility of driving the efficient use of this information<br />
which includes effectively safeguarding information assets in order for this efficiency to be achieved.<br />
Information security programs, therefore, have to be effectively implemented and should emphasise<br />
allocation of information security responsibilities, implementation of Security Education, Training and<br />
Awareness (SETA), and adoption of an information security framework or standard. Adoption of a<br />
standard or framework will ensure that information security practices in these organisations are<br />
standardised and evidence from their information systems can be admissible in court in case of legal<br />
implications. Although the security of information can never be foolproof as the nature and use of<br />
information systems continues to evolve and new risks are identified, measures can still be put in place to<br />
reduce the effect these risks will have on business activities and the organisation’s operations.<br />
7. Directions for future research<br />
There is a need to conduct further in-depth case studies by administering the ISO/IEC 27002 Audit Tool<br />
Questionnaires in more than one copper mining organisation in Zambia. This would most likely yield<br />
further in-depth understanding of information security practices in these organisations, especially from a<br />
middle management perspective.<br />
A possible limitation of this work could relate to the ‘halo’ effect in which perhaps a slightly optimistic view<br />
of questionnaire responses could have been given by respondents. In addition, administering a similar<br />
audit tool to the users themselves would give a more complete picture of information security practices in<br />
these organisations from a senior management, middle management and user perspective.<br />
References<br />
Arjoon, S. (2005) “Corporate Governance: An Ethical Perspective”, Journal of Business Ethics, Vol 61, pp 343-352.<br />
Bajpai, S. and Gupta, J. P. (2005) “Site Security for Process Industries”, Journal of Loss Prevention in the Process<br />
Industrie,. Vol 18, No. 4-6, pp 301-309.<br />
Bartels, N. (2005) “Securing the Process Industries”, [Online], http://www.mbtmag.com/article/193165-<br />
Securing_the_process_industries.php.<br />
Bradley, C. (2008) “Top Five Steps to Proactively Manage Email Risk and Keep Compliance in Check”, [online],<br />
http://www.itsecurityjournal.com/index.php/Latest/Top-Five-Steps-to-Proactively-Manage-Email-Risk-and-Keep-<br />
Compliance-in-Check.html.<br />
Brotby, K. (2009) Information Security Governance, John Wiley and Sons, Hoboken.<br />
Campbell, C. (2008) Legal Aspects of Doing Business in Africa, Yorkhill Law Publishing, Salzburg.<br />
Copans, G. (2007) “ERP Integrator Company Provides Consolidation, Governance to Mine Operations”, [Online],<br />
http://www.miningweekly.com/article/erp-integrator-company-provides-consolidation-governance-to-mineoperations-2007-10-05.<br />
Deloitte Touche Tohmatsu (DTT) (2008) “Gaining momentum. The 2008 Energy & Resources Global Security<br />
Survey”, [online], http://www.deloitte.com/view/en_AU/au/industries/energyresources/energy/article/9f6de23a4b101210VgnVCM100000ba42f00aRCRD.htm.<br />
288
Chishala Lukwesa and Christopher Upfold<br />
Dobelis, E. (2007) Expert Systems for Business Decision on Security Requirements. In: Meersman, R., Tari, Z.,<br />
Herrero, P., Herrero, H. M. (eds). On the Move to Meaningful Internet Systems 2007: OTM 2007 Workshops,<br />
Springer, Berlin.<br />
EPA (2007) “Two-Way Radios”, [online],<br />
http://cfpub.epa.gov/safewater/watersecurity/guide/productguide.cfm?page=twowayradios<br />
Ernst & Young (2008) “Ernst & Young‘s 2008 Global Information Security Survey: Moving Beyond Compliance”,<br />
[online], http://www.ey.com/Global/assets.nsf/Russia_E/GISS_2008_EN/$file/GISS_2008_EN.pdf.<br />
Fagg, S. (2009) “Mining, Financial Services Leading Risk Charge”, [online]<br />
http://www.riskmanagementmagazine.com.au/articles/fb/0c02c6fb.asp.<br />
Falco, J., Stouffer, K., Wavering, A. and Proctor, F. (2002) “IT Security for Industrial Control Systems”, [online],<br />
http://www.isd.mel.nist.gov/documents/falco/ITSecurityProcess.pdf.<br />
Fisher, M. J. And Marshall, A. P. (2009) Understanding Descriptive Statistics. Australian Critical Care. 22: 93—97<br />
Gerring, J. (2007) Case Study Research: Principles and Practices. New York: Cambridge University Press<br />
Government of the Republic of Zambia (GRZ) (1994) “Copyright and Performance Rights Act, 1994”, [online],<br />
http://portal.unesco.org/culture/en/files/30413/11425068613zm_copyright_1994_en.pdf/zm_copyright_1994_en.<br />
pdf.<br />
Hamel, J., Dufour, S. And Fortin D. (1993) Case Study Methods. Newbury Park: SAGE Publications<br />
Helberger, C. (2009) “10 Key Information Security and Compliance Activities for 2010”, [online],<br />
http://www.itbusinessedge.com/cm/community/features/guestopinions/blog/10-key-information-security-andcompliance-activities-for-2010/?cs=34705.<br />
Hinton, W. and Clements, R. (2002) “Are You Managing the Risks of Downtime?”, Disaster Recovery Journal. Vol<br />
15, pp 3. [online], http://www.drj.com/articles/sum02/1503-12.html.<br />
Hong Kong Special Administrative Region (HKSAR) (2008) “An Overview of Information Security Standards”,<br />
[Online], http://www.infosec.gov.hk/english/technical/files/overview.pdf.<br />
International Council on Mining and Metals (ICMM) (2009) “Case Study: Communication Plans for Crisis<br />
Management”, [online], http://www.icmm.com/page/722/communication-plans-for-crisis-management.<br />
IT Governance Institute (2006) Information Security Governance: Guidance for Boards of Directors and Executive<br />
Management (2e), ITGI, Rolling Meadows<br />
Kairab, S. (2004) A Practical Guide to Security Assessments, Auerbach Publications, Boca Raton.<br />
Krehnke, D. C. (2007) Corporate Governance. In: Tipton, H. F. and Krause, M. (eds). Information Security<br />
Management Handbook (6e), Taylor and Francis group, Boca Raton.<br />
Lowe, J., Lasky, J., Gilbert, D. and Sheil, K. (2007) “Protecting Industrial Process Control, Automation, and SCADA<br />
Systems From Cyber Threats”, SPE Middle East Oil and Gas Show and <strong>Conference</strong>, [online],<br />
http://www.onepetro.org/mslib/servlet/onepetropreview?id=SPE-105696-MS&soc=SPE.<br />
Mears, L. And von Solms, R. (2004) “Corporate Information Security Governance: A Holistic Approach”, [online],<br />
http://icsa.cs.up.ac.za/issa/2004/Proceedings/Research/041.pdf<br />
Mining, Minerals, and Sustainable Development Project (MMSD) (2002) Breaking New Ground: Mining, Minerals,<br />
and Sustainable Development: The Report of the MMSD Project, IIED, London.<br />
Ministry of Communication and Transport (2006) “Zambia National Information and Communication Technology<br />
Policy”, [online], http://www.mct.gov.zm/pdf/ict.pdf.<br />
Musson, M. (2007) Business Continuity Strategies for Manufacturing and Logistics. In: Hiles, A. (ed). The Definitive<br />
Handbook of Business Continuity Management, John Wiley & Sons: Chichester.<br />
National Assembly of Zambia (2009) “The Zambian Parliament”, [online],<br />
http://www.parliament.gov.zm/index.php?option=com_docman&task=cat_view&gid=115&Itemid=113.<br />
Olson, D. L. and Wu, D. D. (2008) Enterprise Risk Management, World Scientific, Toh Tuck Link, Paladion<br />
(2008)“Manufacturing”, [online], http://www.paladion.net/Manufacturing.html.<br />
Pavitt, K. (1984) Sectoral Patterns of Technical Change: Towards a Taxonomy and a Theory. Research Policy. Vol<br />
13, pp 343-373.<br />
Peltier, T., Peltier, J. and Blackley, J.(2005) Information Security Fundamentals, CRC Press, Boca Raton.<br />
Peltier, T.R. (2002) Information Security, Policies, Procedures, and Standards, CRC Press, Boca Raton.<br />
Piccoli, G. (2008) Information Systems for Managers: Texts and Cases, John Wiley & Sons Inc, Hoboken.<br />
Pironti, J. P. (2008) “Securing Information Infrastructure:Expert Advice on Evaluating the New Risks and Structuring<br />
Your Defenses”, [online], http://www.iparchitects.com/wp-content/uploads/Securing-Information-<br />
Infrastructure.pdf.<br />
Pricewaterhouse Coopers (2008) “The Global State of Information Security® 2008: Energy and Mining Security<br />
Chiefs Have Advanced in Many—but not all—Critical Security and Privacy Arenas”, [online],<br />
http://www.pwc.com/extweb/insights.nsf/docid/A8659F991CA54234852574DB005CFA9E/$File/Global_Info_Sur<br />
vey_Energy.pdf.<br />
Sawada, K. (2004) Mining Information and Mining Policy. Journal of the Mining and Materials Processing Institute of<br />
Japan: Vol 120, No. 9, pp 532-534.<br />
Seldon, A. (2009) “You Cannot Use it if it is Gone. Hi-Tech Security Solutions: The Industry Journal for Security &<br />
Business Professionals”, [online], http://securitysa.com/article.aspx?pklArticleId=5621&pklCategoryId=106.<br />
Shelton, R. (2009) “An Integrated Approach to Legacy IT Modernization”’ [online],<br />
http://www.acs-inc.com/assets/0/24/818/824/eb1056b2-9058-4170-82cf-696257083213.pdf.<br />
Stake, R. E. (1995) The Art of Case Study Research. Thousand Oaks: Sage Publications.<br />
Teseleanu, G., Maghear, T., and Andras, A. (2006) “The Need for Innovation and Technology Transfer in Mineral<br />
Industry”, [online], http://opi4.opi.org.pl/repository/41fa6dc088dd71307f6cbcd93a3ca1b6Ot9EO9.pdf.<br />
289
Chishala Lukwesa and Christopher Upfold<br />
Tomlinson, R. F. (2007) Thinking About GIS: Geographic Information System Planning For Managers (3e), ESRI, Inc<br />
, Redlands.<br />
Upstill, G. and Hall, P. (2007) Innovation in the minerals industry: Australia in a global context. Resources Policy, Vol<br />
31, No. 3, pp 137-145.<br />
Van Holsbeck, M. and Johnson, J. Z. (2004) “Security in an ERP World”, [online], http://www.netsecurity.org/article.php?id=691.<br />
Von Solms, R. and Von Solms, S. H. (2009) Information Security Governance, Springer, New York.<br />
Warner, D. (2007) “Disaster Recovery Solutions for Business Continuity within the Manufacturing Industry”, [online],<br />
http://www.plasticsbusinessmag.com/wp/?p=35.<br />
Welander, P. (2007) “10 Control System Security Threats”, [online], http://www.controleng.com/article/269570-<br />
10_Control_System_Security_Threats.php.<br />
Welander, P. (2009) “Securing Legacy Control Systems”, [Online],<br />
http://www.controleng.com/index.php?id=483&cHash=081010&tx_ttnews[tt_news]=866.<br />
Wessels, P. L., Grobbelaar, E., Mc Gee, A. & Prinsloo, G.T.M. (2007) Information Systems in a Business<br />
Environment (4e), Lexis Nexis, Durban.<br />
Whitman, M. E. and Mattord, H. J. (2009) Principles of Information Security (3e), Course Technology, Boston.<br />
World Intellectual Property Organisation (WIPO) (2009) “What is WIPO?”, [online], http://www.wipo.int/aboutwipo/en/what_is_wipo.html.<br />
YIN, R. (2003) Case Study Research: Design and Methods. Thousand Oaks: SAGE<br />
Zhang, R. (2009) “Steps for Achieving Proper Mobile Security Governance”, [online],<br />
http://www.computerworld.com/s/article/9133117/Steps_for_Achieving_Proper_Mobile_Security_Governance?t<br />
axonomyId=&intsrc=kc_feat&taxonomyName=.<br />
Zoufaly, F. (2009) “Issues and Challenges Facing Legacy Systems”, [online],<br />
http://www.developer.com/mgmt/article.php/11085_1492531_1.<br />
290
Proposing an Optimized Change Management Process by<br />
Analyzing ITSM Frameworks<br />
Negar Madani, Ali Suzangar, Mohamad Kajbaf, Shirin Nasher and Mehrdad<br />
Kalantarian<br />
Infoamn, Tehran, Iran<br />
n.madani@infoamn.com<br />
a.suzangar@infoamn.com<br />
m.kajbaf@infoamn.com<br />
sh.nasher@infoamn.com<br />
m.kalantarian@infoamn.com<br />
Abstract: This paper proposes an optimized change management process for simplified organizational transition by<br />
evaluating and analyzing widely accepted ITSM frameworks such as ITIL, MOF v3, MOF v4 & FITS from different<br />
aspects including but not limited to the benefits and deficiencies of work flow, roles and responsibilities, key<br />
performance indicators & ARCI matrix. IT service management (ITSM) as a sub-discipline of IT management plays<br />
an important role in achieving the organizations goals and objectives. Many frameworks have focused on ITSM, and<br />
as a best practice Information Technology Infrastructure Library (ITIL) is the most widely accepted approach to ITSM<br />
and it is owned by United Kingdoms’ Office of Government Commerce (OGC). ITIL v3 is presented in five main<br />
categories and each category contains its related processes. Based on ITIL, Microsoft Corporation developed a<br />
framework for IT service management named Microsoft Operational Framework (MOF). Although MOF 3.0 is mostly<br />
in accordance with ITIL framework, MOF 4.0 is presented by three main areas and a management section that<br />
covers the activities in each area. MOF 4.0 focus is on insuring the alignment of business and IT throughout the<br />
organizations activities. One of the most important aspects in ITSM frameworks is the manner in which changes are<br />
managed in an organization. This article includes evaluating and analyzing the change management process in the<br />
above mentioned frameworks from different aspects. As identifying the role of individuals and groups is the most<br />
critical aspect of any process, the first step is to assign roles and responsibilities. This is done by using an ARCI<br />
matrix which presents the individuals and groups responsible, accountable, consulted and informed. In order to<br />
identify the deficiencies of the ARCI matrix for each framework, the matrixes are analyzed and compared and a new<br />
ARCI matrix is introduced. The next step is to define the sequence in which the actions are taken, mentioned as a<br />
work flow. The work flow of each framework is analyzed and the importance and disadvantages of each action is<br />
discussed from a practical point of view. After defining the change process, each framework applies key performance<br />
indicators for assessing its performance. These indicators are listed and compared and the most effective and<br />
efficient indicators are determined. This evaluation and the lessons learned, resulted in an optimized change<br />
management process flow with a particular ARCI matrix and revised key performance indicators which covers the<br />
identified gaps and deficiencies and can be used to benefit organizations in adopting changes and making successful<br />
transitions.<br />
Keywords: IT management, ITIL, change management, ITSM framework<br />
1. Methodology<br />
This article focuses on the comparison of change management processes in ITIL, MOF 3, MOF 4 and<br />
FITS.<br />
At first change management is introduced briefly, and the reasons for its importance and its goals and<br />
objectives have been presented in the second section.<br />
The Third section includes a table in which change management processes of ITIL, MOF3, MOF4 and<br />
FITS are presented. The processes are compared based on the main activities required for a change<br />
management process, and then the benefits, advantages and disadvantages of each process are<br />
discussed.<br />
The disadvantages of the change management processes are solved and an optimized change<br />
management process is presented in section 4, this optimized process has the advantages of all of the<br />
discussed change management processes.<br />
One of the important categories in section 4 is defining responsibility types for the optimized change<br />
management process known as RACI matrix, and assigning it to specific roles. Due to the fact that there<br />
has not been a RACI matrix in the studied processes, it is not gained by reviewing the change<br />
291
Negar Madani et al.<br />
management processes, but it is the result of a practical experience of implementing change<br />
management process.<br />
There are many other requirements for an optimized change management process, which mostly depend<br />
on the specification of each organization, therefore they are mentioned briefly as the additional<br />
requirements of a change management process. Outstanding results that have been achieved by<br />
optimizing the change management process are explained in the conclusion.<br />
2. Introduction to change management<br />
Change management plays a key role in any organization. A change may be triggered from inside or<br />
outside of an organization. The manner in which the change is managed and implemented is the subject<br />
of change management process (Duck, J. D. 2000). It is not always easy to make a change due to<br />
various reasons, for instance a single organization may have varied entities, and changing an entity may<br />
be reflected in other entities. The role of the staff is also very important for the change process to be<br />
smooth. Change management requires a broad set of skills including analytical, financial, business,<br />
system and communication skills. To make a change many different types of models, methods,<br />
techniques and tools are used, however all of the different methods follow a certain structured organized<br />
process which has proven to be successful for leaving the current state and moving to the final state<br />
(OGC 2007). The change management would help an organization and its processes to be stable.<br />
Change management is responsible for the changes made to the IT infrastructure and services, and<br />
ensures that changes are made with minimal risk or consciously taken risk. The main benefits of following<br />
standard processes or proper change management is reduction of the number of potential incidents and<br />
problems associated with each change, simple return to a stable configuration, reduction of back-outs<br />
needed, etc. Like any other process, the main elements of Change management process are as<br />
following:<br />
Process flow, which defines the steps and sequence of activities required to successfully deploy a<br />
change.<br />
Process owner, which is the person in full control of change management process, gathers process<br />
performance data and is consulted in activities related to the process.<br />
RACI matrix, which determines the responsibility types assigned to different roles, and<br />
Key performance indicators (KPI) are the most important metrics which indicate and help authorities to<br />
manage the efficiency and effectiveness of the processes.<br />
2.1 Scope<br />
There are many different approaches to the scope of change, yet changes are divided into two main<br />
categories, reactive changes and proactive changes (OGC 2007).<br />
Reactive changes are changes made in order to align the organization with the business and technology<br />
requirements (OGC 2007).<br />
Proactive changes are changes which are made in response to every day requirements like errors (OGC<br />
2007).<br />
The scope of change management is based on the specifications of each organization. In most<br />
organizations the scope of changes are defined based on the scope of configuration items (CI). If any CI<br />
is altered due to an action, the action is considered as change therefore, it should be done through<br />
change management process (Haber, L., 2003).<br />
3. Process flow<br />
In change management an important step is to define a standard process which presents the activities<br />
required to successfully deploy a change. Different processes have been presented for change<br />
management. The change management processes of ITIL, MOF4, MOF3 and FITS are compared in the<br />
table below. The first column presents the activities mentioned in ITIL V3 service transition publication.<br />
The next two columns present Microsoft Operations frameworks (MOF). Because of the different<br />
292
Negar Madani et al.<br />
approach to MOF4, there are still reasons to consider MOF3 as well. And the last column presents the<br />
British Educational Communications and Technology Agency point of view on a change management<br />
process.<br />
Table 1: Comparison between change management processes<br />
ITIL MOF4 MOF3 FITS<br />
1-Create and record the RFC<br />
(Recommended to use a tool that<br />
automatically stores the initial<br />
data)<br />
2- Review RFC and change<br />
proposal:<br />
3-Assess and evaluate the<br />
change<br />
4- Authorize the change<br />
5- Coordinate change<br />
implementation<br />
6- Review and close change<br />
1: Baseline the<br />
configuration.<br />
2: Initiate the change.<br />
3: Classify the change.<br />
4: Approve the change.<br />
5: Develop and test the<br />
change.<br />
6: Release the change.<br />
7: Validate and review<br />
the change.(update<br />
database)<br />
No steps No steps<br />
1: Identify change<br />
1-Change initiator<br />
2-Change request 2: Create request for<br />
3-Change<br />
change<br />
classification<br />
4-Change<br />
3: Approval to proceed<br />
authorization<br />
4: Plan and prepare change<br />
5-Change<br />
5: Peer review<br />
development<br />
6: Approval to implement<br />
6-Release<br />
management<br />
7-Change review<br />
7: Implement<br />
NO steps defined for<br />
validation<br />
8: Update configurationmanagement<br />
database<br />
No steps defined closure 9: Closure<br />
As presented in the above table, the firs activity in change management is base-lining and recording the<br />
current state, before taking any actions. Base-lining is clearly cited in MOF4 (1.Baseline the<br />
configuration) and is implicitly mentioned in ITIL (1- Create and record the RFC).<br />
The next activity which is alike in all processes, is initiating a change by a Request for Change (RFC). A<br />
RFC is a formal proposal for a change to be made and includes details of the proposed change, and may<br />
be recorded on paper or electronically (British Educational Communications and Technology Agency<br />
2004). In most of the above processes when the first step is determined as “initiating a change”, it is<br />
assumed that base-lining is an obvious fact and will be carried out before making any changes to the<br />
environment. The person which initiates the change is called the change initiator and is informed of the<br />
progress and the result of the change. Therefore in this activity the need for a change is recognized and<br />
is requested by a change initiator.<br />
The next step in MOF3 and MOF 4 is classifying the change, but in ITIL the change is reviewed first and<br />
then it is classified in another step. Classifying the change in MOF 4 includes identifying the priority and<br />
the category of the change (Microsoft 2008).<br />
The priority includes:<br />
Low: The change can wait until the appropriate time.<br />
Medium: Because of the impact on the change, the change should be attended as soon as possible<br />
High: The change should be made as soon as it has been tested.<br />
Emergency: The change should be made immediately; therefore many of the approval and the<br />
development steps are abbreviated.<br />
The category includes:<br />
Standard change: This change has minimal business impact and has a known set of release<br />
processes that has been proven to be successful.<br />
Minor change: This change has a low business impact and affects a small percentage of users and<br />
resources.<br />
Significant change: This change has a moderate impact on the business. It might involve downtime of<br />
services.<br />
293
Negar Madani et al.<br />
Major change: This change has the greatest impact on business, with a high risk and high cost.<br />
Emergency change: This category is high risk because of the urgency of release and the minimal<br />
time in which to test it.<br />
Unauthorized change: This level involves changes that occur outside of the agreed-to change<br />
management policies or that are specifically forbidden.<br />
MOF 3 has a similar category with a slight difference. In MOF 3 changes with the emergency priority, are<br />
assumed emergency changes. No category is defined for unauthorized changes in MOF 3 change<br />
management process (Microsoft 2004).<br />
As mentioned above, ITIL change management process has addressed classifying the RFC in step 4<br />
“Authorize the change”. ITIL has the same priority levels and category; moreover it contains a risk<br />
categorization matrix based on impact and probability (OGC 2007), as follows:<br />
Risk category 1: high impact/high probability<br />
Risk category 2: high impact / low probability<br />
Risk category 3: low impact / high probability<br />
Risk category 4: low impact / low probability<br />
The next step is to review the request and to approve it. This activity is indicated by different titles in each<br />
process, however the concept and explanations for each title declares the meaning of “reviewing and<br />
approving the RFC”.<br />
After a change has been correctly prioritized and categorized by the change manager, depending upon<br />
the category and priority, the change is routed to the appropriate authority for approval. A significant or<br />
major change is usually approved by presenting the change to a team of reviewers known as the change<br />
advisory boards (CAB) (Microsoft 2004). The CAB is composed of many different people with many<br />
different skills for the many different features of the organization. One member of the CAB will be chosen<br />
as the change owner which will be held accountable for the results of the change.<br />
It is the responsibility of the CAB to determine if the change should be approved and scheduled, be<br />
refused and ended or be returned for further clarification. Emergency changes are normally reviewed by<br />
an emergency committee of the CAB typically known as the Emergency change advisory board (ECAB).<br />
Standard changes are approved automatically and progress directly to the change development and<br />
release phase. Minor changes can be approved by the change manager without reference to the CAB.<br />
The next step of the change process is the preparation of the change. ITIL addresses preparing the<br />
change with the title of “Assess and evaluate the change” which includes: The seven Rs of change<br />
management, risk categorization, allocation of priority, evaluation of change, change planning and<br />
scheduling and assessing remediation (OGC 2007).<br />
The stated activities are mostly covered in MOF 4, MOF 3 and FITS with slight differences. For instance<br />
MOF 4 and MOF 3 address the allocation of priority in step three. In step four, MOF 3 schedules the<br />
change, appoints an owner, prepares and develops the change (Microsoft 2004), while MOF 4 designs<br />
the change, identifies configuration dependencies, builds and tests the change (Microsoft 2008).<br />
FITS on the other hand attends to the above mentioned activities in step four, five and six of the change<br />
management process: 4.Plan and prepare change, 5.Peer review, 6.Approval to implement. “Plan and<br />
prepare change” includes: details of change, impact of change, risk of change, remediation plan, change<br />
schedule, authorities names (British Educational Communications and Technology Agency 2004).<br />
The next step of the change process is making the actual change. This is usually done by passing the<br />
RFC to the relevant technical group. Therefore this step is largely a coordination role in change<br />
management.<br />
ITIL has named this step “coordinate change implementation”. MOF 3 and FITS simply forward the RFC<br />
to release management process which is responsible for releasing changes. From then on the role of<br />
change management is to observe the progress of releasing the change.<br />
294
Negar Madani et al.<br />
MOF 4 does not have an independent process for release management. In MOF 4 the release process is<br />
a combination of Deploy SMF which is a sub-process of Delivery phase, and Change and configuration<br />
SMF which is a sub-process of Manage phase. The change is applied through the steps defined in the<br />
sub-processes.<br />
The last step is to review and close the change. On completion of the change, the results should be<br />
documented and reported to appropriate authorities. To confirm that the change has met its objectives, a<br />
post implementation review (PIR) should be carried out. Depending on different types of changes, after a<br />
certain time has elapsed, changes should be reviewed (OGC 2007).<br />
ITIL addresses reviewing and closing the change in a single step with fewer details (OGC 2007), while<br />
closing the change is mentioned implicitly in MOF 4 (Microsoft 2008).<br />
In MOF 3, reviewing the change is determined in an explicit process with detailed activities and<br />
explanations for each types of change (Microsoft 2004).<br />
4. Optimized change management process<br />
The reason for presenting an optimized change management process is to solve the problems identified<br />
in the above processes and to present a process which has the advantages of all of the mentioned<br />
processes. The advantages and disadvantages of the above processes have been identified through<br />
reviewing the requirements of a successful change management process presented in ISO publications<br />
(ISO 2005), management controls (Institute of Internal Auditors 2005) and ITSM success measures<br />
(Haber, L., 2003).<br />
The optimized change management process includes:<br />
Process flow: This process flow is obtained through considering the main requirements of a<br />
successful change and, the comparison of different change management processes.<br />
RACI Matrix: The concept of the RACI matrix for change management does not exist in any of the<br />
above processes; therefore the RACI matrix has been the result of implementing change<br />
management.<br />
Key performance indicators: The KPIs of all other processes have been compared and revised in<br />
order to present a new list, which contains the appropriate KPIs for the optimized change<br />
management process.<br />
Additional requirements: Different types of organizations require different types of templates and<br />
instructions for a successful change management process; therefore it is not possible to pre define all<br />
the additional requirements. However, a brief list of templates and instructions which were required<br />
for the implementation of change management has been presented.<br />
What is unique about this optimized change management process: The outstanding achievements<br />
that have made this process unique and what distinguishes it from the other processes has been<br />
presented and thoroughly explained.<br />
4.1 Process flow<br />
In order to present an optimized change management process, main activities that are required for a<br />
successful change (ISO 2005) and the sequence in which the activities should be presented (Duck, J. D.<br />
2000) has been identified.<br />
The activities are selected based on the advantages of all the reviewed processes. And the order of the<br />
activities is arranged in a way, which will require minimal duplication.<br />
The steps and activities of the optimized change management process are selected based on the above<br />
analysis and the consideration of different environments and organizations; therefore, the steps and<br />
activities are arranged in a way to address the identified deficiencies mentioned in the above analysis.<br />
The table below presents the optimized change management process flow, which the first column<br />
indicates the main steps required to perform a change, while the column on the right indicates the<br />
activities required to fulfill each step.<br />
295
Table 2: An optimized change management process<br />
Negar Madani et al.<br />
Step Activity<br />
1-1) Create and record the RFC<br />
1-2) Review RFC<br />
1) Record and review the RFC<br />
1-3) Approval to proceed<br />
1-4) Baseline and identify CIs effected<br />
1-5) Assess and evaluate the change<br />
2) Change classification<br />
2-1) Prioritize the change<br />
2-2) Categorize based on risk<br />
2-3) Classify the change<br />
3) Authorize change 3-1) Select the authorized person/advisory board<br />
4) Change schedule and escalation<br />
4-1) Schedule change<br />
4-2) Appoint a responsible person<br />
5) Release management 5-1) Release management<br />
6) Review and close change<br />
4.2 RACI matrix<br />
6-1) Select the review team<br />
6-2) Schedule the review<br />
6-3) Review<br />
6-4) Check for Standard change<br />
6-5) Close the Change<br />
In order to determine the required roles and responsibilities for the optimized change management<br />
process, the determined roles and responsibilities in each framework has been reviewed and after<br />
considering the different environments, organizations and their requirements and also taking the<br />
optimized change management processes steps and activities into account, the bellow RACI matrix is<br />
presented.<br />
The RACI matrix has not been included in any of the mentioned change management processes;<br />
therefore it was not possible to present a comparison. The presented matrix is the result of implementing<br />
the optimized change management process.<br />
An RACI Matrix categorizes tasks into four responsibility types, which each type is assigned to specific<br />
roles per process step and activity.<br />
Responsible: This type indicates the person who actually carries out the task.<br />
Accountable: This type is usually assigned to one person through a process and that person is<br />
accountable to the correct and through completion of the task.<br />
Consulted: This type is assigned to authorities who are consulted and asked for advice in specific<br />
sections of the process.<br />
Informed: This type indicates the authorities who should be informed of the progress of a change.<br />
As determined in the table, the change initiator is accountable and responsible for requesting a change in<br />
the format of an RFC. The change manager is held accountable for the rest of the process. The first two<br />
steps are also, mostly the responsibility of the change manager or another person assigned by him. But<br />
in the rest of the process the change manager is just accountable while the responsibilities are assigned<br />
to the CAB. When the change is ready for deployment, the incident and problem management should be<br />
informed and ready to confronting any problems occurred due to the change. And from then on, it is the<br />
responsibility of the release management to deploy the change. After the change has been deployed, it<br />
has to be reviewed and checked, and the change initiator should be informed of the results.<br />
296
Table 3: Change management RACI matrix<br />
Change management Activities<br />
Negar Madani et al.<br />
1-1) Create and record the RFC A/R<br />
1-2) Review RFC A/R C<br />
1-3) Approval to proceed A/R C/I<br />
1-4) Baseline and identify CIs effected A R<br />
1-5) Assess and evaluate the change A/R<br />
2-1) Prioritize the change A/R C<br />
2-2) Categorize based on risk A/R<br />
2-3) Classify the change A/R<br />
3-1) Select the authorized person/advisory board A/R I<br />
4-1) Schedule change A R C/I<br />
4-2) Appoint a responsible person A R<br />
5-1) Release management A I R<br />
6-1) Select the review team A R<br />
6-2) Schedule the review A R<br />
6-3) Review A I R C<br />
6-4) Check for Standard change A R<br />
6-5) Close the change A I I R I I<br />
4.3 Key performance indicators<br />
The metrics determined as key performance indicators, should have specific meanings and should help<br />
with identifying trends (Haber, L., 2003). It is important that the measures indicate the rate of achieving<br />
business goals (Institute of Internal Auditors 2005). The KPIs presented in the reviewed processes have<br />
been compared. And lots of similarities have been identified between the KPI lists. The appropriate key<br />
performance indicators for the optimized change management are:<br />
Increase in the number of successful changes<br />
Reduction of the cost of changes<br />
Reduction in the number of unauthorized changes<br />
Reduction in the number of emergency requests<br />
Reduction in problems and incidents related to a change<br />
Reduction of astonished personnel due to uninformed changes<br />
Reduction in the number of unnecessary changes<br />
Increase of identified standard changes<br />
Reduction of the time of attending to a change request<br />
Reduction of the time of implementing a change<br />
4.4 Additional requirements<br />
The key elements of change management have been presented in the above sections; however several<br />
templates and documents are required for a successful change management process. The most common<br />
templates which should be prepared before lunching change management process have been cited<br />
below:<br />
A request for change template which includes fields for unique ID, requester details, date/time fields,<br />
change information, escalation and decision results, etc.<br />
A change schedule table containing the date and time of all changes.<br />
Terms and instructions for unauthorized changes<br />
297<br />
Change manager<br />
Change initiator<br />
Configuration<br />
manager<br />
CAB<br />
Incident/Problem<br />
manager<br />
Release<br />
management
Instructions for determining standard changes<br />
Instructions for selecting CAB members<br />
Instructions for the CAB agenda<br />
Negar Madani et al.<br />
The voting logic and the method by which the CAB will approve the change<br />
4.5 What is unique about the optimized change management process<br />
The main achievements and benefits of following this optimized process for implementing changes in an<br />
organization are cited bellow. These statements indicate the strengths of this process and how it adds<br />
value to the organization and the IT infrastructure from different points of view.<br />
1. Improved sequence of steps: in order to avoid unnecessary repetition of activities, the sequences<br />
of the steps are assessed and the sequence with the best results has been selected. As mentioned in<br />
classifying the change in section 2.3, In MOF3 and MOF4 before achieving an official approval for<br />
proceeding, the change is classified, which if the change is not approved it will result in the waste of<br />
resources. Plus there is no certain step for scheduling the change and appointing a responsible<br />
person for each change, which in the optimized change management process is placed in step four.<br />
The last but not the least is the final step of MOF3 and MOF4 which do not close the change<br />
properly. If a change isn’t closed properly the organization will soon face many problems due to the<br />
unfinished changes. Therefore the last step is the step where all the involved people are informed<br />
and the change is properly closed and recorded.<br />
2. Flexible and clear process activities<br />
3. Identifying standard changes: One of the main activities which is not covered in the above<br />
processes is the identification of standard changes. There should be a certain step and certain<br />
instructions for identifying standard changes. In the optimized change management process, before<br />
closing a change, it is checked to see if the change can be categorized as a standard change. This is<br />
the perfect step for identifying standard changes because up to here, the change has been<br />
thoroughly deployed and the specifications of the change are clearly understood; therefore the<br />
decision of categorizing this change as a standard change can be made based on sufficient<br />
information.<br />
4. Base-lining and identifying CIs effected: One of the disadvantages for a change management<br />
process is the lack of base-lining. Usually base-lining is not taken seriously and its importance is not<br />
recognized until a change has failed. As this common mistake is repeated in many organizations,<br />
base-lining is mentioned in the first step, right after the change is approved to proceed.<br />
5. Authorizing the change based on priority, category and classification: Clearly authorizing the<br />
change should be based on the specifications of a change. And because the activities should be<br />
chosen based on the type of the change this is a very important step and should be done carefully.<br />
Classifying the change was thoroughly discussed in section 3.3 and is presented as a major step in<br />
the optimized change management process.<br />
6. Appointing a specific owner responsible for each change: like every other process, a change<br />
management process owner is determined which is in full control of the activities, gathers process<br />
performance data and is consulted in activities related to the process<br />
7. Performing well planned and effective reviews: Reviewing a change is one of the important<br />
activities in any change. A review is consisted of a review team, review schedule and the results of<br />
the review. The optimized change management process has taken care to cover all the required<br />
steps for a successful review, which clearly is lacked in the other processes.<br />
5. Conclusion<br />
Many organizations depend on IT and therefore an appropriate IT service management is essential for<br />
any organization. There are several frameworks for ITSM including ITIL, MOF3 and MOF4, etc., with<br />
each framework focusing on a specific aspect of IT service management.<br />
The mentioned frameworks have been implemented and their advantages and disadvantages have been<br />
discovered in time. These disadvantages include, deploying change through too many steps or<br />
insufficient steps, not addressing certain requirements such as identifying standard changes, the lack of<br />
an accountable person, etc.<br />
298
Negar Madani et al.<br />
Therefore an optimized change management process has been presented in this article, which based on<br />
the detailed explanations in section 4.5, which links the final findings with the problems, will result in more<br />
success, less disruption and loss, increased revenue, reduced costs and achieving the business<br />
objectives. Also this process is designed in a way that requires a minimum amount of time, budget and<br />
effort for implementing a change.<br />
References<br />
British Educational Communications and Technology Agency (2004) Change management: Millburn Hill Road,<br />
Science Park, Coventry CV4 7JJ.<br />
Duck, J. D. (2000) Managing change: the art of balancing, Harvard Business Review, November–December.<br />
Haber, L., (2003) How do you measure ITSM success. Jupitermedia Corporation.<br />
ISO (2005) ISO/IEC 20000-1 Information Technology – Service Management Part 1: Specification,<br />
ISO (2005) ISO/IEC 20000-2 Information Technology – Service Management Part 2: Code of Practice.<br />
Institute of Internal Auditors (2005) Global Technology Audit Guide 2: Change and Patch Management Controls:<br />
Critical for Organizational Success, Institute of Internal Auditors Altamonte Springs, Florida, US.<br />
Microsoft (2008) Microsoft® Operations Framework version 4., Microsoft Publications, USA.<br />
Microsoft (2004) Microsoft® Operations Framework version 3., Microsoft Publications, USA.<br />
OGC, (2007) Service Transition. ITIL Service Management Practices v3., London: The Stationary Office.<br />
OGC, (2007) ITIL Service Management Practices v3., The Stationary Office: London.<br />
299
The use of RFID and Web 2.0 Technologies to Improve<br />
Inventory Management in South African Enterprises<br />
Sizakele Mathaba¹, Nomusa Dlodlo², Quentin Williams³ and Mathew Adigun³<br />
¹ ² ³Council for Scientific and Industrial Research, Meraka Institute, South Africa<br />
¹ ³University of Zululand, South Africa<br />
smathaba@csir.co.za<br />
sizamath@gmail.com<br />
ndlodlo@csir.co.za<br />
qwilliams@csir.co.za<br />
madigun@pan.uzulu.ac.za<br />
profmatthewo@gmail.com<br />
Abstract: Cost-effective inventory management includes balancing the cost of inventory with its profit. Most business<br />
owners fail to recognize the value of the cost of carrying inventory, which include not only the direct cost of storage,<br />
insurance and taxes but also the cost of money tied up in inventory. Running inventory using paper-based systems,<br />
Excel files and traditional enterprise software is a costly and resource-intensive approach that may not even address<br />
the appropriate issues for most businesses. It is with this in mind that this research proposes taking advantage of the<br />
Internet of Things (IoT) and Web 2.0 tools in the management of inventory .IoT promotes the communication of<br />
things/object through sensors. On the other hand Web 2.0 tools promote the communication of people through their<br />
phones or desktop computers. The collaboration of these two technologies could improve inventory management. A<br />
comprehensive literature survey is conducted on inventory management functionalities. IoT and Web 2.0<br />
technologies are then mapped to the identified inventory management functionalities. As a result the research<br />
proposes inventory management architecture. The paper looks at the architecture of a system that fully integrates the<br />
technical advantages of Radio Frequency Identification (RFID) and IoT, in collaboration with web 2.0 tool, twitter, for<br />
loss prevention and as an enabler for locating misplaced stock, anti-counterfeiting of stock, etc. The system will focus<br />
on South African Enterprises as a developing country in Africa.<br />
Keywords: Internet of things (IoT), Radio Frequency Identification (RFID), Web 2.0 tools, inventory management,<br />
South African Enterprises, Twitter<br />
1. Introduction<br />
Supply chain management is the well-organized management of the end-to-end process. It begins with<br />
the design of the product or service and ends with the time when it has been sold, consumed, and in the<br />
end, discarded by the end user. The complete process involves product design, procurement, planning<br />
and forecasting, production, distribution, fulfilment, after-sales support, and end-of-life disposal (Michael<br />
and McCathie, 2005). Inspection and anti-counterfeiting for merchandise are vital contents of supply<br />
chain management, enterprises which need to open the market completely, have to offer consumers a<br />
real-time information inquiry platforms on commodity, and to let consumers have a clear understanding of<br />
the entire life cycle of the goods they consumed and to give them a convenient way to identify the<br />
accuracy of the products. Unfortunately, few enterprises have systems in place that can monitor whole<br />
cycle of products currently. To resolve this difficulty, this paper recommends a products monitoring<br />
information system based on RFID and IoT.<br />
The paper is structured as follows. Section 2 is the literature review on the IoT. Section 3 is the research<br />
methodology, Section 4 represent the rationale of the study, while Section 5 looks Web 2.0. Technology<br />
in general. Section 6 is the proposed architecture of the inventory management system that fully<br />
integrates the technical advantages of RFID and IoT, in collaboration with web 2.0 tools twitter, for loss<br />
prevention and as an enabler for locating misplaced stock, anti-counterfeiting of stock, etc. The system<br />
will focus on South African Enterprises as a developing country in Africa.<br />
2. Background and context of the study<br />
2.1 Internet of Things<br />
The IoT is an incorporated part of the Future Internet and could be defined as a dynamic worldwide<br />
network infrastructure a with self configuring ability based on criterion and interoperable communication<br />
protocols where physical and virtual things have distinctiveness, physical characteristics, and virtual<br />
personalities and use intelligent interface, and are seamlessly integrated into the information network(de<br />
Saint- Exupery, 2009).<br />
300
Sizakele Mathaba et al.<br />
Figure 1 is an example illustrating how IoT works in warehouse management. Traditionally, operations in<br />
warehouse management are processed manually, with information gathered by hands and captured<br />
through keyboards, voice entry or barcodes and integration through human- machine -interface. The<br />
introduction of IoT, involving RFID sensors & actuators, changed the laborious processing of warehouse<br />
operations. Data entry is now automated, which results in accurate information for informed decisions.<br />
Figure 1: IoT in warehouse management (adapted from Fleisch, 2010).<br />
Figure 2 illustrates how the IoT functions.<br />
Figure 2: Illustration of Internet of Things (adapted from Shen, 2010)<br />
It consist of tags which are programmed into an object, each tag contains product information such as the<br />
expiry date, object pressure, temperature, prices etc. Sensors are embedded into these tags and<br />
301
Sizakele Mathaba et al.<br />
readers, and they enable data transfer from the tags to the readers. Readers access data from the tags in<br />
a code form. IoT middleware filters repeating and irrelevant data and sends the code to the local server.<br />
The object information will be available to the internet through the remote server. Object Naming Service<br />
(ONS) works similar to the Domain Name Service (DNS), it points out the servers storing object<br />
information.<br />
In the IoT, objects are autonomous entities in companies, information and social processes where they<br />
will be enabled to interact among themselves. They also interact with the environment by exchanging<br />
data and information sensed on the environment, while reacting autonomously to the real/physical world<br />
events and influencing it by running processes that trigger actions and create services with or without<br />
direct human intervention. The interfaces in the structure of services facilitate communications with these<br />
smart things over the Internet, query and change their state and any information related with them, taking<br />
into account security and privacy matters(de Saint- Exupery, 2009).<br />
Typical Applications of IoT:<br />
“Toyota (South Africa), Carrier tagged to streamline manufacturing and vehicle tracking. The tags are<br />
intended to remain with the vehicle throughout its life cycle” (Baudin, 2005).<br />
A car which notifies you when it must have a service.<br />
A smoke detector which is directly connected to local fire station.<br />
Washing machine twittering when it finishes washing etc. There are quite many areas where IoT can<br />
be applied.<br />
2.2 Radio Frequency Identifiers<br />
RFID tags are divided into two general types, active and passive, depending on their supply of electrical<br />
power. Active RFID tags consist of their own power source, normally an on-board battery. Passive RFID<br />
tags get power from the signal of an external reader. RFID readers also come in active and passive<br />
selections, depending on the kind of tag they read (Intermec Technologies, 2007).<br />
2.2.1 Active tags<br />
Active tags contain their own power source. They send out a stronger signal, and readers can access<br />
them from further away. The on-board power source makes them big and costly, and thus active RFID<br />
systems normally work well on big objects tracked over lengthy distances. Low-power active tags are<br />
generally a little larger than a deck of playing cards. Active tags can stay inactive until they come in range<br />
of a receiver or can continuously transmit a signal. As a result of their on-board power source, they can<br />
function at higher frequencies, normally 455 MHz, 2.45 GHz, or 5.8 GHz, relying on the application’s read<br />
reach and memory needs. Readers can converse with active RFID tags across 20 to 100 meters<br />
(Intermec Technologies, 2007).<br />
2.2.2 Passive tags<br />
On the other hand, passive tags are very economical, they can cost as little as 20 cents a piece, and new<br />
technologies are continuously manufacturing them on a low-price for integration into general materials<br />
and products. Since passive tags are inexpensive, they are expected to be the starting point in the<br />
expansion of RFID implementations especially in South Africa as a developing country. Besides to their<br />
low cost, passive tags can also be reasonably small. Recent antenna technology confines the smallest<br />
functional passive tag simply to the size of a quarter. The larger the tag, the larger the read range. At<br />
present, passive RFID tags contain about 2 Kbits of memory which is quite small to hold much more<br />
complex information than identification and history information. Technology in the nurture of RFID is<br />
continuously improving. Consequently, the amount of information and capabilities of RFID tags will<br />
increase over time, thereby allowing RFID tags to ultimately hold and transmit enough information<br />
(Weinstein, 2005).<br />
Passive-tag readers can continuously transmit its signal or transmit it when required. Once a tag moves<br />
across the reader’s range, it accepts an electromagnetic signal from the reader through the tag’s<br />
antenna. The tag then keeps the energy from the signal in an on-board capacitor. This process is called<br />
inductive coupling. Once the capacitor has made sufficient charge, it can be able to power the RFID tag’s<br />
circuitry, which transmits a modulated signal to the reader. The return signal consists of information<br />
stored in the tag. Low-frequency tags (less than 100 MHz) send information by releasing energy from the<br />
capacitor to the tag coils in altering strengths over time, which affects the radio frequency produced by<br />
302
Sizakele Mathaba et al.<br />
the tag. The reader perceives these varying waves and can use these variances to demodulate the code.<br />
(Weinstein, 2005).<br />
In higher-frequency tags (greater than 100 MHz), the tag transmits the signal by means of backscatter, in<br />
which the tag’s circuit adjust the resistance of the tag’s antenna. This change in resistance makes a<br />
transmission of Radio Frequency (RF) waves, which the reader can accept and demodulate. Passive<br />
tags usually work at frequencies of 128 KHz, 13.6 MHz, 915 MHz, or 2.45 GHz, and have read ranges of<br />
only few inches to 30 feet. Frequency selection depends on the system’s environment, kind of material<br />
the signal have to pass through , and the system’s essential read range. RFID tags can be enclosed in<br />
numerous materials. Plastics are the most familiar material for RFID, making identification cards for<br />
building entrance, credit cards, or bus fares (Weinstein, 2005 and Intermec Technologies, 2007).<br />
Figure 3 represents the RFID components and illustrates how these components works in any given<br />
environment. Tags which usually very small in size, are attached to any given object and they send out<br />
signals which are received by the reader through antenna, which is an enabler for tags and readers to<br />
transmit information. The information is sent to a back-end computer system for processing.<br />
Figure 3: RFID system components (adapted from Yan et.al, 2008)<br />
3. Methodology<br />
A comprehensive literature survey was conducted on inventory management functionalities, and to<br />
identify research work that has been covered in the IoT and web 2.0 in enhancing the management of<br />
inventory in South African Enterprises. IoT and Web 2.0 are then mapped to the identified inventory<br />
management functionalities.<br />
The research question is:<br />
How can IoT and web 2.0 technologies be utilized to improve inventory management systems in South<br />
African enterprises?<br />
Objectives of the study<br />
Review literature on inventory management<br />
Identify systems that have adopted IoT technologies that can be utilised in inventory management.<br />
Come up with inventory management architecture.<br />
4. Rationale of the study<br />
Stock outs result in much loss for enterprises. It is estimated at 30%, which affects retail sales by<br />
between 5 and 18%. In most cases, the stock maybe available but misplaced (Xin, 2009). This is a huge<br />
loss for enterprises. South Africa is not only becoming more aware of RFID, but is also progressively<br />
rising in the uptake of this technology. RFID vendors are becoming specialised in providing specific<br />
applications and catering their offerings to particular industries. Diversity of factors currently influences<br />
South Africa's demand for RFID applications namely:<br />
The need to reduce theft,<br />
303
Fraud and counterfeit products.<br />
Sizakele Mathaba et al.<br />
The RFID technology is able to keep track and trace the location of products precisely than current<br />
technologies in use, which leads to a reduction in losses. The ability of technology to increase efficiency,<br />
government support for RFID projects and international policies that require RFID applications on specific<br />
goods for import, such as beef products, are also influencing the adoption (online, url:<br />
http://dataweek.co.za/article.aspx?pklArticleId=5027&pklCategoryId=31 ).<br />
The slow adoption of RFID in South Africa presents diverse opportunities for retailers. The most vital is<br />
the availability of abundant opportunities that have not yet been exploited but are perfect for early mover<br />
advantage. Mission to roll out RFID applications in different countries are also on the rise. Supply chain,<br />
asset management and mining are amongst the areas expected to increase the demand for RFID<br />
applications, and actions in these areas are on the rise in many African countries. Despite the numerous<br />
opportunities for RFID applications, there are hindering factors to its expansion. The most remarkable is<br />
the high cost of installing the application, particularly for low value volume products. Most enterprises in<br />
South Africa are small to medium enterprises with low quantity of manufacturing. RFID is costly for such<br />
businesses to implement. Consequently, many sectors continue to use alternative solutions such as bar<br />
codes.<br />
The RFID technology is ineffective if function under certain conditions i.e. Transponders do not function<br />
well when tagged on iron objects or wet surfaces. This creates complications for manufacturing and other<br />
industries that use a variety of metallic objects (Dane et. al, 2010). The low level of alliance among<br />
industry participants is also slowing the market's development. Having agreed that RFID requires a<br />
standard of infrastructure, mostly in power and telecommunications, the demand is affected in many<br />
countries and regions facing infrastructural complications. Regardless of various technological<br />
challenges, the cost of RFID tags has started to drop. This movement is expected to carry on as more<br />
advances in the manufacturing of low cost RFID continue. The continued drop in prices will be a key<br />
issue in growing the demand for RFID in South Africa. Low cost tags will be particularly engaging to low<br />
value volume things, where the market for RFID is currently small. This will offer solution for a variety of<br />
enterprises of all sizes. Participants in market have need of working together to resolve interoperability<br />
and other concerns that have an effect on the use of RFID applications, such as discovery of a general<br />
approach to the industry principles(online, url: http:// dataweek.co.za ) .<br />
The forecast for RFID technology in South Africa are not as rosy as those for biometrics. It is doubtful that<br />
RFID in SA will see the same adoption rate as seen in the US, as local retailers are unlikely to set up<br />
RFID go-ahead like they did in the US. On top of that, the charge of importing tags and the difficulty of<br />
deployment are likely to put off many potential RFID customers, says Max Stone, distribution and partner<br />
manager at Motorola Southern Africa's Enterprise Mobility business(online, url:<br />
http://www.autoidlabs.ch/).<br />
RFID in South Africa is well established in some niche environments, such as access control i.e.<br />
Biometrics, tollgates. South African regulator ICASA only permitted suitable EPC (Electronic Product<br />
Code) RFID frequencies for SA last year, so major deployments are yet to happen. RFID has been<br />
deployed very selectively in South Africa; the cost of the tags has been too high for retail organisations to<br />
see advantage from using this type of real-time tracking. However, this may change as time elapses<br />
(online, url: http://www.iweek.co.za/ViewStory.asp?StoryID=201372 ).<br />
5. Web 2.0<br />
Web 2.0 technologies outline the beginning of the subsequent creation of web-based applications. It<br />
allows web applications to be created, that are more operationally rich and quick to respond than the<br />
usual static pages of traditional web technologies. It’s also enabling content to be produced and shared in<br />
real time, with end-users commonly able to add content to applications themselves (O’Reilly, 2007). This<br />
implies that Web 2.0 technologies support open communications and provide users the freedom to share<br />
their suggestions and opinions.Most organizations of all types and sizes and from all industry verticals<br />
have noticed the explosive growth on the web of social and community sites in the consumer space such<br />
as MySpace, YouTube, and the deluge of Web 2.0 sites. Enterprises have observed the move of major<br />
Web players such as Amazon, eBay, Live, Google, and Yahoo to include social and community<br />
elements, and the interest and demand that this has created. Now they are enthusiastically considering<br />
and in several cases constructing portals in communities and businesses for their own organizations.<br />
Web 2.0 is moving to enterprises (Kittowski et.al, 2009).<br />
304
Sizakele Mathaba et al.<br />
Organizations are interested in using Web 2.0 practices mainly in two places, which is within the business<br />
to advance competency and production, and from the organization to the customers to improve revenue<br />
and customer satisfaction. The use of Web 2.0 within organizations is called Enterprise 2.0 and is likely to<br />
be the first area where Web 2.0 will be used by organizations. Enterprises are using Web 2.0<br />
technologies to mainly communicate with customers (advertising), business partners and potential<br />
employees, allowing them to achieve the goal of true real-time collaboration among these parties. This<br />
can increase output and provides enterprises with a way to easily promote their products. The creation of<br />
online communities and blogs or wikis to initiate conversations and share knowledge is proving to be<br />
particularly interesting to enterprises. Enterprises are already using the Web 2.0 technologies inside the<br />
organization, for communication with employees and customers and marketing, etc. IoT technologies, on<br />
the other hand promote the communication among objects without any human intervention leading to a<br />
reduction of labor costs. This study proposes that IoT can be collaborated with web 2.0 technologies for<br />
communication purposes with humans. The objects will communicate with humans for updates through<br />
web 2.0 technology Twitter, to bridge the communication divide between objects and humans.<br />
6. Proposed architecture for inventory management<br />
Enterprise architecture (EA) presents a bright approach to support business strategy execution. EA is an<br />
outline that can direct organizational activities in the technological sense, and supports the business<br />
strategy from the lowest level of process. This outline helps to facilitate an organization’s business vision<br />
by encapsulating multiple views of an enterprise and provides a general understanding of what the<br />
organization aims to achieve. In addition, EA view the enterprise from a holistic and more complete<br />
architectural point of view (Chuang and Loggerrenberg 2010). Figure 4 shows the proposed architecture<br />
for inventory management. It is architecture of a system that fully integrates the technical advantages of<br />
RFID and IoT, in collaboration with the web 2.0 tools twitter, for loss prevention and as an enabler for<br />
locating misplaced stock, anti-counterfeiting of stock, availability of stock on shelves etc.<br />
In this architecture, RFID basically serves as a replacement for the bar code scanners which are normally<br />
used to track products and shipments in similar ways. This architecture fully integrates the technical<br />
advantages of RFID and web 2.0 to provide feedback on the process to end user. Since the application of<br />
FRID require no human intervention, it is essential that the end user is updated in the whole process.<br />
The RFID system consists of three fundamental components. Initially, the RFID tag is attached to an<br />
asset or product in the inventory. The tag contains information about the particular asset or product and<br />
also may include sensors. The next component is the RFID reader, which communicates with the RFID<br />
tags. The last component is the backend system, which links the RFID readers to a centralized<br />
database/server. The centralized database will store information, such as price, for each RFID tagged<br />
item. In this proposed architecture for inventory management, the passive tags will be used due to their<br />
low cost. Among the functionalities expected to be performed by this system includes:<br />
Checking the availability of stock on shelves.<br />
Identifying misplaced stock on shelves.<br />
Identifying expired/ ruined stock.<br />
Identifying counterfeit products.<br />
Sending updates to the relevant user.<br />
Electronic Product Code (EPC) is a unique global identifier of each product in IoT technologies which is<br />
used to track and trace products (Yan, 2008). The EPC RFID readers will be placed among the shelves<br />
and the products will be programmed with EPC RFID enabled tags. EPC RFID tags will send out the<br />
signal which will be received by the EPC RFID readers in the radio frequency field. The readers will<br />
receive the signal through their antennas and transmit the stored information, i.e. validation, tracking,<br />
counts, and error messages to the EPC middleware. The EPC middleware will filter out the repeating and<br />
irrelevant information. Thereafter, information will be sent to the local server. The local server computer<br />
system will twitter the information to the user i.e. reports on inventory, aggregate counts, errors occurred,<br />
misplaced stock etc. through the use of web 2.0 technology. The end user/owner will receive the<br />
notification on inventory through his/her phone. This system gives effective technical reference for<br />
enterprise managers to monitor whole process of inventory without them being physically involved in the<br />
process. The consumers of the products will benefit also in this proposed architecture. They can query<br />
information about the product on the remote server using the EPC (Electronic Product Code), the ONS<br />
305
Sizakele Mathaba et al.<br />
(Object Naming Service) is network system which works similar to the DNS (Domain Name Service).It<br />
spots out servers storing information in the internet.<br />
Figure 4: The proposed architecture for inventory management<br />
7. Conclusion<br />
Internet of Things integrated with the RFID technology and web 2.0 technologies can assist enterprise<br />
owners in managing their inventory, i.e. using them in monitoring the stock validity, stock on shelves,<br />
misplaced stock etc. The use of web 2.0 tools could play a major role in keeping enterprise owners<br />
posted about what is happening on the inventory without them being physically there and helping them to<br />
make informed decisions, and to know urgent matters which may need their attention immediately. Web<br />
2.0 tools bridge that divide of objects and humans. As a result, this study encourages South African<br />
enterprises to actively promote the development procedures of the RFID technology with web 2.0 tools<br />
and the Internet of Things to improve the inventory management in their enterprises. The IoT have the<br />
drawbacks though i.e. the cost of the technology is a major concern for developing countries like South<br />
Africa, but that can addressed by major investments and collaboration with developed countries if<br />
possible.<br />
306
8. Acknowledgments<br />
Sizakele Mathaba et al.<br />
This research was sponsored by Council for Scientific and Industrial Research (CSIR) MERAKA,<br />
Pretoria, South Africa and Internet of Things Capability Group under supervision of Dr Louis Coetzee<br />
(Group Manager).<br />
References<br />
Baudin, M. (2005) Consultant MMTI Manufacturing Management & Technology Institute: RFID Applications in<br />
manufacturingDraft ,Palo Alto, California.<br />
Chuang C, Loggerrenberg J,( 2010), Challenges Facing Enterprise Architects: a South African Perspective,<br />
Department of Informatics, and University of Pretoria, South Africa.<br />
Dane, H. and Michael, K.and Wamba, F. S., (2010) RFID Enabled Inventory Control Optimization: A Proof of<br />
Concept in Small-to-Medium Retailer, University of Wollongong, Australia.<br />
de Saint – Exupery A,(2009) Internet of Things Strategic Roadmap, CERP-IoT EC, Europe.<br />
Frost and Sullivan, (2008) “The potential for RFID applications in Africa” [Online 03 Nov 2010], Dataweek, Electronics<br />
and Communications Technology, http://dataweek.co.za/article.aspx?pklArticleId=5027&pklCategoryId=31<br />
Franz-Kamissoko, L. (2009) “SA lets its fingers do the talking”, [Online 09 Nov 10], The business technology news<br />
magazine for the empowered enterprises, http://www.iweek.co.za/ViewStory.asp?StoryID=201372<br />
Fleisch,E.(2010) “Auto-ID Labs” , [Online 12 Nov 10] University of St Gallen, http://www.autoidlabs.ch/<br />
Kittowski, F. and Klassen, N. and Faust, D. and Einhaus, J (2009) “A comparative Study on the use of Web 2.0 in<br />
Enterprises” Proceedings of I-KNOW ’09 And I-SEMANTICS’09.<br />
Michael, K. McCathie, L. (2005) , The Pros and Cons of RFID in Supply Chain Management, Proceedings of the<br />
International <strong>Conference</strong> on Mobile Business (ICMB’05) IEEE, University of Wollongong, Australia.<br />
Mtuszak G, (2007) Enterprise 2.0 Tales and Trenches”, Information Communications & Entertainment (KPMG<br />
International).<br />
Intermec Technologies Corporation (2007) “Supply Chain RFID: How It Works and Why It Pays”, whitepaper<br />
Intermec Technologies, North America.<br />
O’Reilly T, (2007), What is web 2.0, O’Reilly Media Inc.<br />
Shen, G. and Liu, B. (2010), Research on Application of Internet of Things in Electronic commerce, Third<br />
International Symposium on Electronic Commerce and Security IEEE Society, Beijing Wuzi University, China.<br />
Weinstein R, (2005), RFID: A Technical Overview and Its Application to the Enterprise, IEEE Computer Society,<br />
United States.<br />
Yan B., Huang G, (2008) ,Application of RFID and Internet of Things in monitoring and Anti-counterfeiting for<br />
Products , School of Economics and Commerce, South China University of Technology, Guangzhou, China.<br />
Xin C., (2009), RFID Application in the Retail Industry, Guangdong University of Foreign Studies, 2009 ISECS<br />
International Colloquium on Computing, communication, Control and Management ,China.<br />
307
Modeling the Genetic Schemes of Human Memory<br />
Information in the Process of Production and Reparation of<br />
Knowledge<br />
Fatemeh Mohammadi, Tahereh Mohammadi and Bahram Rezaie<br />
Faculty Member of Islamic Azad University, Shiraz, Iran<br />
faramatn@yahoo.com<br />
tmohammadi@live.com<br />
bahramrezaie@gmail.com<br />
Abstract: In this paper the theory of "genetic scheme of human information" is presented for the first time. This<br />
theory is in based on fifteen years of continuous study and research on related sciences including cognitive and<br />
meta-cognitive psychology, the system of processing information, information pathologies such as amnesia,<br />
information interference, mistakes in remembering information such as the time of tests and exams, neurology and<br />
learning, chemistry and long term memory compounds, electricity physics and electrons movements in data<br />
transferring. To make the theory more comprehensible, a conceptual model on the process of receiving knowledge,<br />
repairing knowledge and finding the relationship between different sciences to each other was created and<br />
presented. This included the use of factor analysis method and path analysis with multiple variable regressions. Its<br />
fitness degree was accepted with 86.12% by a group of 186 Faculty members of Iranian universities. The paper also<br />
offers a conceptual model for understanding the way information system of human mind performs in data saving and<br />
recalling. This model benefits from the results of previous studies and researches on the way memory performs In<br />
terms of neuroanatomy, neuropsychology, physiology, and the system of data processing in the field of<br />
metacognitive Psychology. This model was approved by a group of 97 lecturers in education and neuropsychology,<br />
who have been randomly selected. The related ratio is 85.538. The statistical method employed is factorial and path<br />
analysis. This model depicts the structure of the type of information and the way it is preserved and saved in the<br />
memory. In addition, the model makes it possible to transform the data into other forms which can be recalled and<br />
recognized in a simpler way.<br />
Keywords: information genetic scheme, memory, knowledge generation and reparation, human mind information<br />
system (HMIS), conceptual model, data saving and recalling<br />
1. Introduction<br />
The way human mind performs in saving, processing and recovering information has been studied for<br />
many years by different fields of study including psychology, neurology, and education. The result of<br />
these studies offers models, paradigms, and conceptual frameworks to simplify the understanding of<br />
complexities of human mind performance and his intelligence. (Sternberg, 2006)<br />
These authors’ primary concern was whether a simple conceptual model can show the way information<br />
system of human mind performs, and whether this model can help human to save, process, and recover<br />
information in a better and faster way?<br />
Human being’s dependency on artificial and man-made memories such as portable microcomputers and<br />
electronic note-books in twenty-first century has resulted in mind laziness and malfunction of human<br />
memory. (Mohammadi, 2010)<br />
Therefore, this paper with its emphasis on using the mentioned model, tries to demonstrate how a person<br />
can avoid forgetfulness and mental laziness and enhance his mental capacities. Based on the<br />
introduction presented, the main research questions of this study are as follows:<br />
What are the main components of the conceptual model of Human Mind Information System?<br />
Which model can be proposed to understand HMIS complexities?<br />
What is the degree of validity In university lecturers’ answers to the questionnaires.<br />
2. The significance of the research<br />
Having an appropriate model for the process of remembering and recalling information in information<br />
system of human mind is as important as using models in other experimental disciplines. If the sum of all<br />
endeavors and studies conducted on the performance of human mind, thought, memory and human<br />
intelligence is considered as a whole, then we must be able to create a concrete symbol to represent this<br />
308
Fatemeh Mohammadi et al.<br />
totality. The model presented in this paper tries to show the complexities of HMIS using simple and<br />
understandable symbols.<br />
3. Review of literature<br />
When a child is born he lacks any type of concept or meaning. No sign or symbol of the world makes<br />
sense to him (Berlo 1970, P.178). When we use a word representing a thing or a matter in the presence<br />
of a child, he mysteriously matches the two types of stimuli and then make a relationship between them.<br />
That is because the brain and its neurons are flexible. On the other hand, Berlo shows that meanings and<br />
concepts exist inside humans (Ibid PP 184-5). We know that the organism which maintains these<br />
concepts is memory (Sachs 1967, P437& wallas, 1921, P16).It seems that these concepts lie in the longterm<br />
memory (Hilgard 1975, pp 236-40).<br />
Studies on growth and evolution of memory have been the specific field of Leontiev. Having worked with<br />
normal and retorted children of various ages, he created a device by which the examinee could use<br />
secondary stimuli in order to remind a series of stimuli which were offered by the examiner. Eventually,<br />
Leontiev showed that the process of mediated remembering is long-termed and difficult. This<br />
remembrance is not regular and that there is no evidence showing if the person performs a special<br />
activity to remember previously saved information or not. Luria called this type of behavior "natural<br />
remembering", because she thinks stimulus is remembered through direct affecting process and without<br />
any mediation. But there are times when a person cannot recover and recall the related information by<br />
provided stimulus (Luria, P. 6l).<br />
Following Tief's studies, Morozova believes remembering the stimulus in relation to the response is<br />
difficult for the child. But secondary stimuli can guide the child's chosen response to a right way. She<br />
found out that the rules for mediated remembering are also true for natural remembering. Luria<br />
conducted other research studies in a different setting but with the same substructure. She shows that<br />
memory's activity can twist from a normal form of speech into complex self-folding functions. In 1929,<br />
Luria’s group started an original research on understanding the meaning of stimuli and function of<br />
mediated devices for remembering a series of words and pictures. This group showed that children at<br />
their first stage of growth are not able to create pictorial stimulus which makes the future remembering<br />
possible. But when they grow up, they project effective technique usage. It seems that older children can<br />
organize some of stimuli characteristics in order to facilitate remembering (Ghasem Zade, 2008, P. 85).<br />
In contrast to previous researchers, Vigotesky believes that mental abilities of human being can be<br />
elevated up to a full capacity. He emphasized the qualitative description of organizing information and the<br />
behavior originated from it. As a result, he founded another psychological discipline which emphasizes<br />
the human's socially organized experiences and the structure of his conscious activities. This new branch<br />
has a close relationship with neurology, but it was different from neurology regarding different controlling<br />
areas of psychological functions. Therefore, understanding the abilities of human brain is needed for<br />
neuropsychology or psychology of nervous system (Luria and Cole, 80, P.64).<br />
Vigotesky offered an approach based on his analysis of the structure of psychological function. First he<br />
considered the relationship between preliminary and elevated psychological functions, and then he<br />
determined their mental classification in a normal adult. Vigotesky's approach acted as a model for future<br />
neuropsychology researches (Luria, pp.66-70).<br />
Durkheim believes that basic mental processes cannot be considered as inner-life manifestation or<br />
outcome of natural evolution. His discussion focused on two questions:<br />
Do the content of thought and basic categories used in describing an experience differ in various<br />
cultures?<br />
And, are human’s intellectual functions different in various cultures?<br />
Berlo showed that primitive people followed different rules and functions compared to people in<br />
modern era because primitive individuals did not organize their information enough (ibid, P 73).<br />
In 1980 opinion, intellectual power in primitive cultures basically was not different from individuals living in<br />
nowadays’ technological communities, because the process of thinking, remembering necessary<br />
information, combining them, and creating consequent information has the same logical rules. But other<br />
gestalt psychologists pay attention to the common characteristics of mind in all cultures. They developed<br />
this theory that the principles of understanding and thought are similar worldwide (Werner, H. 1890-<br />
309
Fatemeh Mohammadi et al.<br />
1960). But mental characteristics are not separate from their biological and physiological structures.<br />
Human being’s nervous system has the responsibility of creating harmony between different biological<br />
functions. (Eysenck and Flanangan, 2009).<br />
Human's information is kept in memory. Physiologists found that memory is in Hippocampus area and<br />
they believe that hypo camp has a major role in recording and keeping human information. Hippocampus<br />
is located at the brain cortex. This border is called limbic. Limbic has a role in the setting of feelings,<br />
behavior, motivation and memory due to several links to other parts of nervous system. Hippocampus is<br />
extended in the length of inner edge of temporal boarder. The structure of Hippocampus reflects strata.<br />
Surface of Para Hippocampus has six layers. If we observe Hippocampus from the surface, we face three<br />
layers (out of six layers). These three layers include surface molecular layer which contains nervous<br />
fibers and small scattered neurons, the pyramidal layer which contains several large pyramidal neurons,<br />
and a multishaped layer whose structure is similar to the multishaped layer of other parts of the surface<br />
(Snell, pp.342-5).<br />
This stratified structure led to us to gain a stratified image from the first model of information system of<br />
human brain. Especially where I found that pyramidal layers of Hippocampus are made of round or oval<br />
compact neurons, which create a synapse with dendrites of pyramidal cells of Hippocampus Here, is the<br />
intersection of bringer and sender fibers of information. The study of this section on theoretical<br />
foundations was very important to find out the basic structure of HMIS.<br />
According to this evidences, Hippocampus has a role in converting the short-term memory to long-term<br />
memory. If Hippocampus is damaged, one cannot keep information in his memory for a long time. This<br />
disorder is called Anterograde amnesia. But in this case, the far-off memory which has already been<br />
formed will not be damaged. Damage to the core of Hippocampus makes severe memory disturbances<br />
(ibid P. 347). Despite these facts, Physiologists believe that what is kept in memory is not primary and<br />
raw information but it is often mixed with a meaning whose root has an inner part. (Snell, 2006, p373).<br />
Until the present time, a collection of various hypotheses has been offered in order to describe the<br />
physiological and neurological function of memory. They include, coding of detailed information in a large<br />
collection, -e.g., remembering the details of a trip, physiological coding in the variety of electrochemical<br />
activity of cells and also chemical changes of neuron bonds in message transferring (Mesulam 1981,<br />
p.250). Besides, Asratyan showed that there exists a great correlation coefficient between learning<br />
process and memory. Also, there is a direct relationship between learning and change of protein<br />
composition in the brain (Asratyan, 1983, P.43)<br />
4. Research methodology<br />
The present work is a descriptive research based on modeling the information system. The statistical<br />
population is all faculty members of Educational Science in the universities of Fars Province, Iran), for<br />
estimating the statistical sample we used Z=1.96, Confidence Coefficient = % 95 and d=0.05. Therefore<br />
our sample was n= 86. Since we estimated that about 15% of questionnaires would not be returned, we<br />
increased the number of subjects to 97, in a random basis.<br />
5. Performance of the questionnaire<br />
The study of theoretical literature and effective components on the model showed that these<br />
components, including basic and secondary components, can be primarily classified into five basic<br />
groups:<br />
Philosophy of the model<br />
Objectives of the model<br />
Theoretical foundation and confirmer of the model<br />
The model’s potential for answering and generalizing<br />
The model’s potential to adapt to facts.<br />
For evaluating the suitability of the philosophy of the model three questions, for the objectives of the<br />
model four questions, for the theoretical foundations five questions, for the potential of answering and<br />
generalizing seven questions, and eventually, for the evolution of suitability of the model adaptation to<br />
experienced facts six questions were formed. The discrete method and percentage answers were chosen<br />
since they are more logical and accountable compared to Licret spectrum. Questionnaire’s validity was<br />
confirmed by the experts and through formal validity. We studied twenty six subjects in a pilot test to<br />
310
Fatemeh Mohammadi et al.<br />
obtain their views for future changes in the questionnaire. To evaluate the inner validity of the<br />
questionnaire, test split-method calculation of intensity of relationship between two halves was used<br />
(r=%67.11). To evaluate the validity of relationship power of each question with total score of<br />
questionnaire, dotted two-threaded correlation coefficient was used with the related r value of 0.71.08.<br />
The validity of all questions was checked against Cronbach's Alpha coefficient. The result was α=0.76<br />
which shows the questionnaire was appropriate for final exertion.<br />
6. HMIS model and the findings<br />
The first question of the study is: What are the basic elements of conceptual model of mind information<br />
system?<br />
Theoretical studies show that the elements of HMIS consist of the following parts:<br />
Short-term and long-term memory<br />
Information processing system<br />
Intelligence and the part responsible for discovering the relationship between phenomena<br />
Power of adaptation and compatibility of human brain<br />
Relationship between information type and the speed of memorization and remembrance.<br />
Relationship between the types and amount of received information from the five senses<br />
Meaningfulness of the relationship between new materials and previous information.<br />
Learning methods with emphasis on Bandora, Gestalt, and Azubell.<br />
Nervous system in message transferring (neuropsychology)<br />
Communication system (communication model in message transferring from sender to receiver)<br />
Semantic flexibility in cognitive field<br />
Previous and subsequent information interference system<br />
Learning psychological affections<br />
Ultra cognition and models of self-leading in learning<br />
In answering the second question of the research i.e., which model can be suggested for understanding<br />
the complexities of human mind information system, elicited factors taken from the first question's answer<br />
were entered in the raw and primary model after lecturers' opinions were reviewed, and they were<br />
classified into five basic sections:<br />
Philosophy of the model<br />
Objectives of the model<br />
Theoretical foundations of the model<br />
Potential of the model in answering and generalizing<br />
Potential of the model in adaptation to the experience facts.<br />
7. Philosophy of the model<br />
Philosophy of the HMIS model was evaluated for understanding its complexities by answering three<br />
questions. The result of the analysis of the answers shows that the focus of the model’s philosophy is on<br />
understanding and facilitation of the memorization process and recalling information from long term<br />
memory. This process follows a neuropsychological mechanism which is very complex and its<br />
understanding is very difficult. In order to facilitate understanding of this complex process, the main<br />
structure of the model is based on the type of information introduced to short-term memory and then to<br />
the long-term memory as well as their placement and saving.<br />
To simplify the model, we put the name of the first upper layer in short-term memory, the "current layer".<br />
Current layer is a layer that keeps received and recalled information for a short period of time. This layer<br />
consists of all of information types. Data analysis of answers to the basic component of the model<br />
philosophy showed that %78.62 of statistical sample accepted this philosophy for suggested model of<br />
human mind information system.<br />
311
8. Objectives of the model<br />
Fatemeh Mohammadi et al.<br />
We explain the objectives of the HMIS model by four questions and by regarding components elicited<br />
from the review of the results of research questionnaire and opinion polls of lecturers. We consider the<br />
following objectives for the model:<br />
Understanding the complexities of the performance of memorization and remembering in our<br />
information system.<br />
Engendering meaningful communication for constant saving and quick remembering.<br />
Familiarizing the lecturers with teaching active methods and involving various senses of students for<br />
effective learning.<br />
Arranging the information in the mind, identifying the reasons of forgetting information, and<br />
preventing information forgetfulness.<br />
Statistical descriptive analysis showed that % 89.15 of sample agreed with these model objectives.<br />
9. Theoretical foundations of the model<br />
The current model’s theoretical foundations were evaluated by five questions. The combination of<br />
theoretical studies and acquired results of previous researches as well as data analysis of the<br />
questionnaire and lecturers’ opinions showed that elicited data from the five senses are not saved and<br />
recovered by the same speed and quality. We resulted that, that specific information is remembered<br />
sooner or it is closer to the current layer or simply chemical combinations act more quickly for the<br />
analysis. Therefore, the conclusion is that information is memorized earlier. In other words, they entered<br />
to their special layer of long-term memory sooner. According to the results of sample answers and<br />
background studies, the following order for saving all types of data is offered:<br />
The nearest type of information to the current layer is the information which is received and saved<br />
through experiment or experience. The reason may be the involvement of more senses for learning<br />
or memorizing. For each of the five senses, chemical links opens. Senses of smell touch and taste<br />
function in this layer.<br />
In the second layer, visual and pictorial information exist. As theoretical studies show, the speed of<br />
image memorization and remembering is more than all other types of information received through<br />
hearing and reading. The bulk of received information from visual sense is more than other senses. It<br />
means that %75 of information is received from the sense of sight.(Mohammadi, 2010)<br />
In the third layer, information received through reading and writing is located. The reason for<br />
durability of read and written information compared to heard information and also their careful and<br />
quick remembering can be the use of empirical senses in memorizing and remembering read and<br />
written information .In the case of sighted-talent people while reading, the form and characteristics of<br />
page and written images are facilitating factors which come closer to the tested layer. In a blind<br />
person, the sense of touch makes perception easier and is closer to test and experience layer<br />
compared to heard information.<br />
Finally in the lowest layer the information from the sense of hearing is saved. The bulk of saved<br />
heard information is less than the information received through visual data. Remembering heard<br />
information is also more difficult and memorizing them lasts only for a short time. We know that %75<br />
of heard information can be lost in 24 hours if they are not repeated. (Mohammadi, 2010)<br />
The analysis of answers to questions of the model showed that surface bulk and type of received<br />
information from the five senses are not equal. Therefore, the layers cannot be at the same level<br />
regarding their setting. So, we should design the surface ratio of each layer according to theoretical<br />
foundations. Studies show that on average % 74 of healthy people gain information through sight sense,<br />
%15.67 through sense of hearing, and % 6.9 from reading and writing. %2.999 of the information is also<br />
perceived through experiment and experience; besides, %0.006 of human being’s information is in the<br />
current layer. So, the surface of each layer changes according to the arrangement of layers’ information,<br />
quality and the bulk of each layer. The following simplified model is suggested for better understanding.<br />
The information analysis of the samples' viewpoints about the component of theoretical basis showed<br />
that %86.09 of them accepted the component.<br />
312
Figure 1: Two types of memory<br />
Fatemeh Mohammadi et al.<br />
Current data layer (Short Term Memory)<br />
Long term Memory<br />
Experienced information layer<br />
Seen information layer<br />
Written information layer<br />
Heard information layer<br />
Figure 2: Types of information layers<br />
10. Model’s potential in answering and generalizing<br />
The elements of model’s potential in answering and generalizing were considered by seven questions.<br />
Since theses elements are more important, the number of questions is more than other parts of the<br />
model. Based on the results of information analysis, the samples' viewpoints, and also the answers for<br />
the first research question, we showed with %95 certainty that the bulk of different information types are<br />
not equal for saving & recovering. The more continuous and detailed the parts from large groups, the<br />
easier they can be recalled them. For instance, obtained information from a journey experience has a<br />
great bulk. It consists of so many small fragments and without any attempt; they are saved in the memory<br />
and can be remembered more easily. Yet, a short sentence or a scientist's birth date which is heard via<br />
radio can hardly be recalled. Such information is shown with a big empty circle, because it cannot be<br />
decomposed. Thus, they cannot be recalled easily unless they can appear in another shape or through<br />
association with other information.<br />
In 1986, Mohsenian Rad in an interactive model on man to man relationship presented a meaning unit as<br />
a ball in human mind. According to him and based on the speed of memorizing & remembering data,<br />
‘listened to’ information can be explained by the smallness and roughness of their shapes. The smaller<br />
parts make an apparent shape which move in a limited space, the greater chance of memorizing and<br />
recalling them. In neuropsychology theories, it is said that as all kinds of information are made of<br />
electrochemical energy, they can change from one form to another and potentiality can be re-read and<br />
remembered. So, there should be a way to get to the surface (the current layer). Each group of<br />
information which can enter to the current layer simultaneously and in the form of a group of information<br />
can be remembered more quickly. The combination of the two, displays a complete picture of the model.<br />
Experienced and experimented information layer<br />
Seen information layer<br />
Written and read information layer<br />
Heard information layer<br />
Figure 3: Symbols of any type of information<br />
313
Fatemeh Mohammadi et al.<br />
The analysis of data obtained from the answers of the sample group on "the power of answering and<br />
generalizing of the model" showed that %89.06 of them agree with the mentioned component.<br />
11. Model's power of adaptation to experienced facts<br />
This element of model's power of adaptation to experienced facts was checked by six questions. The<br />
results of the samples’ viewpoints and statistical analysis of their answers to questions indicated that the<br />
given model has %87.06 of adaptation to the experienced facts. %48.91 of faculty members of the<br />
sample group graded the component, appropriate to completely suitable in the information system of<br />
human mind. The model’s adaptation to experienced facts means that by the help of the model, one can<br />
explain the process of learning level discrepancy from one method to another, based on the movement<br />
from one layer to another one. For instance, when somebody is listening to a song or psalm in order to<br />
recite it, he writes it on a piece of paper because by this method, he can recall the order of the psalm<br />
lines by its position on the paper. In other words, instead of just using the Layer responsible for hearing,<br />
he goes to upper layer and uses the Layer which stores read information, written layer and also the layer<br />
that save pictorial information, In fact while he memorizes the song's lines from their position on the<br />
paper, he creates a link between two above layers.<br />
Current data layer (Short Term Memory)<br />
Experienced information layer<br />
Seen information layer<br />
Written information layer<br />
Heard information layer<br />
Figure 4: Information layers format<br />
As people experiment at a particular time, the amount of experienced information is more remembered<br />
and recalled than the heard information. One can conclude that the passage of transferring information<br />
from current layer to other layers and vice versa cannot have the same expanse. For example, the<br />
expansion of experienced information input and output transferring band is much wider than the<br />
expansion of heard or read information transferring band. Similarly, the width of other layers' bands<br />
cannot be the same. Therefore, the suggested model of human's mind information system will reform in a<br />
way that it can adjust itself to the experienced facts. Thus, the channels of transferring information have<br />
different band expansion according to their bulk and type. The theoretical findings based on the existing<br />
literature and the results of faculty members’ viewpoints were approved by %84.91 of faculty members as<br />
Figure 5: Fully Information layers<br />
314<br />
The current information layer<br />
Experienced information layer<br />
Seen information layer<br />
Written and read information layer<br />
Heard information layer
Fatemeh Mohammadi et al.<br />
12. Experimental evidence of model justification<br />
Experiment and research show that permanence of heard information is less than the other types of<br />
information. Thus remembering heard information is more difficult than the other types of information.<br />
This difficulty includes using more energy in recalling, or a need for a stronger motivation to intrigue<br />
information-saving cells to receive and recall the information. Based on the mentioned results we placed<br />
the heard information in the lower layer, far from the current layer. Similarly, here is evidence for the<br />
placement of other types of information in their layers. For example, the information received through<br />
experiments and experiences is rarely forgotten. Pictures are remembered faster than heard information.<br />
The information that are actively dealt with or are repeated successively like a story, remain longer and<br />
more precisely in long term memory.<br />
13. Ratio of the model<br />
The third question in this research is: "What is the validity or ratio of the information system model of the<br />
human mind from the viewpoint of the faculty members?" The result of the information analysis showed<br />
that faculty members’ viewpoints were approved the HMIS model with %85.538 validity.<br />
Table 1: Answer sheet information of sample group according to the ratio degree of the model<br />
Main elements Completely<br />
agree<br />
Agree No<br />
Comment<br />
Opposite Ratio<br />
Model's Philosophy %56 %38 %5 %2 %87.62<br />
Model's Objectives %68 %21 %8 %3 %89.1<br />
Model's Theoretical<br />
Foundations<br />
%63 %29 %7 %1 %86.09<br />
Model's Potential of Answering and<br />
Generalizing<br />
%71 %23 %5 %1 %89.06<br />
Model's potential of Adaptation to Experience<br />
Facts<br />
%61 %28 %9 %2 %84.91<br />
Total View %85.538<br />
Table 2: Descriptive Information of Model Ratio Degree<br />
Total Sample<br />
Model<br />
Man Woman<br />
Mean of Percentage Number Mean of Percentage Number Mean of Percentage Number<br />
Ratio<br />
Ratio<br />
Ratio<br />
85.538 %100 97 81.132 %52.6 51 90.044 %47.4 46<br />
14. Discussion and conclusion<br />
In order to understand the performance of human mind's information system in memorizing and<br />
remembering information, a conceptual model is presented. In this model, the results of previous studies<br />
and researches on the performance of the mind, including neuroanatomy, neuropsychology, physiology,<br />
learning psychology, and processing the system of information in the field of metacognitive were used.<br />
This model is approved by a group of 97 faculty members of education and neuropsychology that were<br />
randomly selected with the ratio of 85.538.<br />
This model shows the structure of information types and the way of locating and saving them in the mind.<br />
It also shows the reason for forgetting heard information among physically and mentally healthy people.<br />
Indent The reason is the form of recognition of heard information in relation to other types of information.<br />
The size and capacity for saving heard information on the one hand and the low speed of remembering<br />
and recalling them from memory on the other hand cause 75 percent of heard information to be forgotten<br />
after 24 hours. This result is compatible with the results of Bach-and-Rita, Zametkin, Raven, and<br />
Raritan's researches. This model also shows how using different senses in learning lead to deeper<br />
saving of information and faster remembering them. Furthermore, this model makes it possible to change<br />
information to other form that can be recognized and remembered simply. This result is also compatible<br />
with Sharma, Orton and Kolb researches result. Also this model can help us to find out the Process of<br />
Production and Reparation of Knowledge.<br />
315
Acknowledgement<br />
Fatemeh Mohammadi et al.<br />
I would like to convey my gratitude to the editor of this article, Maryam Nasseri and my dear family.<br />
References<br />
Bach, Y. Rita (1990) Brain Plasticity as a basis for recovery of function in humans, Journal of Neuropsychologia,<br />
volume 28, pages 547-554.<br />
Berlo, D.K. (1970). The process of communication: an introduction to theory and practice. New York,<br />
Berlo, N. & R. Naugle, (1985)"Case Studies in Cerebral Plasticity", International Journal of Clinical Neuropsychology,<br />
P7<br />
Eysenck, Michael and Cara Flanangan (2009) Psychological Psychology, Translated By: Seifollah Bahary and<br />
Others, (In Persian), Islamic Azad University, Rudhen Branch, Iran.<br />
Finucci, Whitehouse, Isaacs & Childs(1984) Mattis, French & Rapin (1975)Miles,1986, Derivation and Validation of<br />
Reading Disability, Journal of Developmental Medicine and Child Neurology, Volume 26, Pages 143-153<br />
Galaburda AM, Sanides F, Geschwind N.(1978) Human brain. Cytoarchitectonic left-right asymmetries in the<br />
temporal speech region. Arch Neurol, 35(12):812-7.<br />
Galaburda AM, Sherman GF, Rosen GD, Aboitiz F, Geschwind N.(1985) Developmental dyslexia: four consecutive<br />
patients with cortical anomalies. Ann Neurol,18 (2):222-33.<br />
Ghasem Zade, Habibollah (2008) proceeding of Iranian Cognitive Neuropsychology Symposium, (In Persian) Tehran<br />
University. Tehran, Iran.<br />
Hilgard, E. (1975), Psychology background, Translated by Barahani, M. pp 236-40).<br />
Holt, Rinehart and Winston, Inc. Byron, P. Rourke (2005) Neurology of Learning Disabilities: Past and Future,<br />
Journal of learning disabilities volume 28, pages 111-114<br />
Khoda panahi, M.(2007) An Approach to Neuropsychology Research Method(in persion), Allame Tabatabaee<br />
University, Tehran, Iran.<br />
Kolb B. (1989) Brain Development, Plasticity and Behavior. Journal of American Psychologist, Vol. 44, Pages 1203-<br />
1212.<br />
Leontiev Alexis (1959) Problems of Mental Development, Moscow, Akad, Pedag, Nauk.<br />
Luria A. R and Shilawle A. (1980) the Making of Mind, A Personal Account of Soviet Psychology, Edited by Michael<br />
Cole and Sheila Cole, P.64<br />
Mesulam, M. M. (1981) A Cortical Network for Directed Attention and Unilateral Neglect, Annual of Neurology, vol.<br />
10, 309-325<br />
Mohammadi, Fatemeh (2010) human Relationship in organization, (In Persian), Faramatn Publication, Shiraz, Iran.<br />
Mohsenian Rad, Mehdi (1991) Human Communications, (In Persian) Sorush Publication, Tehran, Iran, pp 315-6.<br />
Orton, S. (1928) Specific Reading Disability, Journal of the American Medical Association, Vol 90, Pages 1095-1099<br />
Raven, J. C. (1965) Guide to Using the Colored Progressive Matrix. San Antonio, TX: Psychological Crop.<br />
Reitan, R. M. and Wolfson, D.(1985) the Halstead-Reitan Neurological Test Battery. Tucson, AZ: Neuropsychology<br />
Press.<br />
Richard Snell (2006) Clinical Neuroanatomy, <strong>6th</strong> edition,? Publication, pp.342-5).<br />
Sharma Lata Prem (2006) Learning Readiness, Sarup Teaching Learning Series, No 19. S.T.L.S.<br />
Sternberg Robert, (2006) Cognitive psychology, Thomson, Wadsworth,4th edition, P.43.<br />
Vygotsky, L S. (1978) Mind in Society, Ed, Michael Cole, Vera John Steiner, Sylvia Scribner, and Ellen Souberman,<br />
Cambridge, Harvard University Press.<br />
Werner, H., & Kaplan, E. (1952). The acquisition of word meaning: A developmental study<br />
Zametkin, A. J. et al (1990) Cerebral Glucose Metabolism in Adult with Hyperactivity of Childhood Onset. New<br />
England Journal of Medicine, Volume 323, Pages 1361-1366<br />
316
Copyright Protection for GIS Vector Map Based on Wavelet<br />
Transform<br />
Amer Sedeeq Mustafa<br />
Al-Mustansiriyah University, Baghdad, Iraq<br />
ahea.aur@gmail.com<br />
Abstract: The copyright of GIS vector map has become an increasingly complex topic. The main technologies which<br />
are used for copyright protection are the copyright marking techniques, and these techniques are delegated by<br />
watermarking. In this paper, a new watermarking algorithm for GIS vector map is presented. The magnitude of the<br />
DWT coefficients of vertices extracted from the map is altered to embed the watermark. The selection of the hiding<br />
coefficients will be depended on the threshold value; and a key used to select the exact locations which embed the<br />
watermark. The watermark bits are embedded in coefficients which should satisfy certain conditions. The Insertion of<br />
the watermark in these selected coefficients is done by changing the specific bits of the coefficients like spatial<br />
domain. The watermarks can be blindly detected without the original map and watermark information. The<br />
experimental results demonstrate the high robustness of the proposed algorithm to various attacks like noise<br />
additions etc, and by evaluating the quality of the watermarked map via the WPSNR with respect to the original map,<br />
we show that the watermarked map is very similar to the original map, which indicates that the proposed<br />
watermarking scheme can achieve a better tradeoff between the robustness and the transparency.<br />
Keywords: 2D vector graphics; digital watermarking; wavelet transform; geographic information system (GIS)<br />
1. Introduction<br />
Nowadays, GIS has been widely applied in many fields. The rapid development of computer<br />
communication and the Internet makes it very easy to expediently exchange data via networks. On the<br />
other hand, it also becomes crucial to protect the digital copyright of various digital Medias. GIS data<br />
represents a high material value due to its high cost of acquisition, which is either performed manually or<br />
through analysis of aerial photography Wang (2007), Endoh (2002).<br />
The preparation of maps and services is the first stage involved in the Author module to enable the<br />
display of maps via internet. Each map layer is prepared base on the data and user requirement.<br />
Appropriate symbols and color are assigned to ensure the displaying of the result which fulfils the criteria<br />
and features required. Next, the map layers will be compiled and stored in a service to ease users<br />
accessing to the application server. With the emergence of digital maps, copyright protection of maps<br />
encountered a new challenge. New technology should be developed to solve this problem. Digital<br />
watermarking technology provides a new solution for digital map copyright protection .Despite the fact<br />
that the protection of vector images can be of great importance in many applications; copyright protection<br />
of vector images still drew little attention within the watermarking fields. There are only a few published<br />
works done on watermarking vector images. Indeed, the research on vector map digital watermark has<br />
just started and authenticity of a vector image is becoming a hot research area Busch (2004).<br />
M. Sakamoto et al proposed a watermarking scheme by modifying the location relationship of vertices<br />
Takashima (2000). Ohbuchi et al presented several algorithms for 2D vector map watermarking. In<br />
spatial domain, a scheme based on correlation detection was proposed in Endoh (2002). Micheal Voigt<br />
and Christoph Busch proposed an algorithm which can embed multibit data based on direct sequence<br />
spread spectrum and modified correlation detection Busch (2003). These algorithms of watermarking are<br />
belong to spatial domain. The watermarking algorithm for vector maps in DFT domain was proposed by<br />
N. Nikolaidis, I. Pitas, and V. Solachidis Solachidis (2000), Pitas (2000) they proposed a blind<br />
watermarking scheme which embeds a single bit into a polyline by modifying the discrete Fourier<br />
coefficients of polyline’s coordinate sequence.<br />
2. Digital watermarking<br />
Digital watermarking is the process of embedding information into digital multimedia content such that the<br />
information (which we call the watermark) can later be extracted or detected for a variety of purposes<br />
including copy prevention and control. Digital watermarking has become an active and important area of<br />
research, and development and commercialization of watermarking techniques is being deemed<br />
essential to help address some of the challenges faced by the rapid proliferation of digital content Wang<br />
(2009).<br />
317
Amer Sedeeq Mustafa<br />
The detection and analysis of digital information can ensure the integrity and reliability of the digital<br />
information; therefore serve as a means to protect the rights and interests of the author. Robust<br />
watermark refers to the watermark’s ability to resist regular signal process and malicious attacks, it is<br />
mainly used to confirm the author, creator, copyright owner, distributor, and authorized consumer of the<br />
digital media. Fragile watermark is a class of watermark which is less robust. It is mainly used to protect<br />
the authenticity and integrity of the document Solachidis (2000).<br />
3. Wavelet transform<br />
Wavelet analysis has been turned into research hot spot in the signal processing in recent years because<br />
it is a powerful analysis tool. The wavelet theory has good performance and it is been widely applied in<br />
voice signal processing, image analysis, data compression and etc.. Eui (2002). With the wide application<br />
of wavelet technology, watermark technology based on wavelet has gained more and more attention.<br />
Wavelet transform can get image multi-scale details step by step via flexing and parallel moving<br />
operation. As scale becomes smaller, every part gets more accurate, and ultimately all image details can<br />
be focalized accurately. If planar discrete wavelet analysis is made to input image, it will produce three<br />
high frequency parts (HL, LH and HH) and one low frequency part (LL). The low frequency part can<br />
denote optimal approach to original image. The majority of image energy concentrates in here.<br />
4. Vector map<br />
The vector map represents real-world features as strings of x,y pairs representing spatial information<br />
about the features. A point is a single pair of values with an x value representing the longitude of the point<br />
in some coordinate space and a y value representing the latitude of the point. Linear features may consist<br />
of a set of connected points or some mathematical function describing the beginning point of the feature<br />
and the formula that constructs it. Polygon features are closed loop of X-Y coordinate pairs, last pair are<br />
the same. The example of vector map is shown in Figuer 1.<br />
Figure 1: Example of vector map<br />
5. The proposed algorithm<br />
The block diagram of the proposed watermarking algorithm is shown in Figuer 2.<br />
5.1 Vertex extraction and extension<br />
All vertices are extracted from the 2D vector graphics V, where N means the number of vertices. Let {Vk }<br />
be vertex sequence, Vk = (xk , yk ) means vertex coordinates, xk and yk means longitude and latitude<br />
coordinates respectively. The number of vertices may be not in multiples of 32, so extension of vertices in<br />
a minimum scope is proposed. The increased number of vertices can be calculated by:<br />
M = min {x | (N + x) mod 32 = 0} (x =1, 2….) (1)<br />
318
Amer Sedeeq Mustafa<br />
Figure 2: Block diagram of the proposed watermarking algorithm<br />
The value of increased vertex coordinates is the same as the final vertex coordinate. After this step, the<br />
extracted vertices will be sorted in vector of one dimension as follow:<br />
V1 V2 ……………………………….. V32<br />
Each vertex Vk is composed of the x and y coordinates, so each group has 64 coordinates, and then<br />
structure 8 × 8 matrix based on these 64 coordinates can be constructed as follow:<br />
5.2 Wavelet transform<br />
x1 y1 x2 y2 x3 y3 x4 y4<br />
..<br />
..<br />
..<br />
..<br />
..<br />
..<br />
.. .. .. .. .. ..<br />
.. .. .. .. .. .. x32 y32<br />
The proposed method for hiding watermark is based on the Discrete Wavelet Transform (DWT). The<br />
wavelet transform is used to generate watermark and decompose the 8 × 8 matrix into regions to help in<br />
selecting the suitable locations to hide the watermark in vector map.<br />
We applied one level discrete wavelet decomposition on the 8 × 8 matrix. The Haar wavelet basis is<br />
chosen due to its simplicity. The 2×2 Haar matrix that is associated with the Haar wavelet is:<br />
319
Amer Sedeeq Mustafa<br />
⎡1<br />
1 ⎤<br />
H2= ⎢ ⎥<br />
2)<br />
⎣1<br />
−1⎦<br />
The DWT separates 8 × 8 matrix into a lower resolution (LL) as well as horizontal (HL), vertical (LH) and<br />
diagonal (HH) detail components. The 2×2 Haar matrix is applied in the following way:<br />
G'(i, j)= H2 [G(i, j)] (3)<br />
Where G(i, j) refers to matrix values and will be modified to G′ (i, j) wavelet coefficients matrix after DWT.<br />
5.3 Watermark generation<br />
In this step we will generate watermark bits from wavelet coefficients matrix G ′ (i, j), and generate an 8 ×<br />
8 binary matrix A according to wavelet coefficient matrix G′ . The values of matrix A can be defined as :<br />
⎧ 1 G'(i,<br />
j) > M<br />
⎨<br />
A(i, j)= ⎩ 0 G'(i,<br />
j) ≤ M<br />
(4)<br />
Where M is the mean value of G′ and i,j=1,2…8. Watermark bits are generated as follows:<br />
P= A(1, 1) ⊕ A(2, 1) ⊕ ………….. A(8, 1) (5)<br />
Where ⊕ refers to XOR.<br />
5.4 Watermark embedding<br />
After separates 8 × 8 matrix into four channels LL, HL, LH, and HH. Since LL coefficient modification can<br />
easily lead to major changes in vertex coordinates thereby reducing the transparency of the watermark,<br />
while the HH coefficients are too sensitive, we choose the HL and LH regions of the coefficients. The<br />
hiding coefficients will be selected depending on the threshold value and a secret key. This key is used to<br />
select the exact locations which embed the watermark.<br />
The threshold value is calculated for HL and LH regions by using the following equation Eui (2002):<br />
T=2 (log2max (Ci))-1 (6)<br />
Where Ci represents the largest integer coefficient HL and LH regions. The watermark is embedded only<br />
for the selected coefficients.<br />
The watermark bits are embedded in coefficients that are in HL and LH subbands. The host coefficients<br />
should satisfy the following condition:<br />
|Coefficient Magnitude | > Threshold value.<br />
Inserting the watermark in these selected coefficients is done by changing the specific bit of the<br />
coefficients like spatial domain. We use inverse DWT transform to produce watermarked matrix, and then<br />
transform the matrix into one-dimensional sequence and remove extended vertices to get the<br />
watermarked vertex coordinates. Then get the watermarked map.<br />
5.5 Watermark extraction<br />
Watermark extraction is the inverse process of watermark embedding, only need to know the threshold<br />
value and the secret key. The steps are as follows:<br />
Extract all the vertices of test vector map and then extend to set every group 32 vertices.<br />
8×8 matrix is generated.<br />
Make DWT transform on the matrix.<br />
320
Amer Sedeeq Mustafa<br />
Generate an 8 × 8 binary matrix A′ according to this DWT coefficient matrix as in Eq.(4). Extract<br />
watermark bits p′ as follows:<br />
p′ = A′ (1, 1) ⊕ A′ (2, 1) ⊕ ……….. A′ (8, 1).<br />
Calculate threshold value for HL and LH regions as in Eq. (6).<br />
According to the secret key and the threshold value the selected hiding locations where watermark<br />
embedded can be determined and watermark bits W’ can be gotten.<br />
Since we select two middle DWT coefficients (HL and LH) to embed watermark repeatedly, we obtain<br />
W and W . If p′ = W and p′ = W , then this map is copyrighted.<br />
'<br />
1<br />
'<br />
2<br />
'<br />
1<br />
6. Experimental results<br />
'<br />
2<br />
In order to evaluate the proposed watermarking algorithm, two vector maps are used in Figuer 3.<br />
Figure 3: Watermarked vector maps (map1 and map2)<br />
So as to test the robustness of the proposed algorithm, we try to apply different kinds of operation to the<br />
watermarked data, and extract the watermarking information after that. After extracting watermark, we<br />
use similar measurement of the extracted and the referenced watermarks as validation; it can be defined<br />
by the Normalized Correlation (NC). The formula which calculates similarity is defined as:<br />
∑<br />
∑<br />
*<br />
* 2<br />
NC= ( W . W ) / ( W ) , k=1,2,…..L (7)<br />
k<br />
k<br />
k<br />
k<br />
k<br />
*<br />
Where, W and W are referenced and extracted watermarks. L is length of watermark bits.<br />
k<br />
k<br />
Meanwhile, the visual quality of watermarked and attacked images is measured by using the Peak Signal<br />
to Noise Ratio (PSNR). A ’PSNR’ derived from digital image fidelity evaluating is presented in this paper,<br />
which is used as a measure of the map quality:<br />
PSNR( dB)= 10log10<br />
2 ⎡max<br />
⎤<br />
∀(n,<br />
m) f (i, j)<br />
⎢<br />
⎥<br />
⎢⎣<br />
MSE ⎥⎦<br />
Where MSE is the mean square error:<br />
MSE=<br />
∑∀(n, m)<br />
(fw(i,<br />
j) − f(i, j))<br />
nm<br />
2<br />
321<br />
(8)<br />
(9)
Amer Sedeeq Mustafa<br />
where n, m are the dimensions of the input map and f, fw are the original and the watermarked maps.<br />
However, PSNR declines from the perceived subjective quality because the Human Visual System (HVS)<br />
does not correlate well with the square of the error. For this reason, the Weighted Peak Signal to Noise<br />
Ratio (WPSNR) that takes into account the properties of human visual system is also used as follows<br />
Zhang (2009):<br />
WPSNR(dB)=10 log10<br />
2 ⎡max<br />
⎤<br />
∀(n,<br />
m) f (i, j)<br />
⎢<br />
⎥<br />
⎢⎣<br />
MSE * NVF ⎥⎦<br />
2<br />
2<br />
φσ ( i,<br />
j))<br />
σ ( i,<br />
j)<br />
Where Noise Visibility Function NVF (i, j) =1/(1+ x , x denotes the local variance of the<br />
map in a window centered on the vertex with coordinates (i, j).<br />
2<br />
2<br />
θ is a tuning parameter corresponding to the particular map and is given as θ =D/ σ , where σ is<br />
the maximum local variance for a given map and D is an experimental value, range from 50 to 100. Table<br />
1 shows the watermark extracting results of the proposed algorithm.<br />
Table 1: Result of test<br />
Attack<br />
x max<br />
Result table for illustration the performance of watermarking algorithm<br />
Map 1 Map2<br />
PSNR<br />
WPSNR<br />
20% JPEG compression 34.1 36.7 0.9720 27.5 35.2 0.9654<br />
40% JPEG compression 32.2 33.8 0.9684 27.5 35.2 0.9521<br />
Salt and Pepper Noise 27.75 30.3 0.94 27.5 35.1 1.00<br />
Gaussian lowpass filter 44.96 47.9 0.988 41.2 43.8 0.986<br />
1/4 cropping 11.11 13.5 0.862 10.4 18.0 0.861<br />
7. Conclusions<br />
In this paper, we propose a watermarking algorithm which can be used for GIS vector map. Experimental<br />
results show that the degradation by embedding the watermark is too small to be visualized. In addition,<br />
the robustness of this algorithm is strong enough to some attack like noise, compression, and cropping.<br />
Therefore, the proposed watermarking scheme can achieve a better tradeoff between the robustness and<br />
the transparency.<br />
Future work will focus on making the watermarks robust to more attacks and optimizing the algorithms to<br />
provide higher capacity and efficiency.<br />
References<br />
Nikolaidis, N, Pitas, I and Solachidis, V. (2000) “Fourier Descriptors Watermarking of Vector Graphics Images”, Proc.<br />
of the International <strong>Conference</strong> on Image Processing, vol.3, pp.10-13.<br />
Ohbuchi R, Ueda H, and S. Endoh (2002) “Robust Watermarking of Vector Digital Maps”, Proc. of IEEE International<br />
<strong>Conference</strong> on Multimedia and Expo (ICME02), Lausanne, Switzerland, pp.577-580.<br />
Qingtang SU, Xianxi LIU, Steve Zhang (2009) ”Quality Evaluation of Digital Image Watermarking”, Fifth International<br />
<strong>Conference</strong> on Intelligent Information Hiding and Multimedia Signal Processing.<br />
Sakamoato, M, Matsuura, Y and Takashima, Y (2000) “A Scheme Of Digital Watermarking for Geographical Map<br />
Data”, Proc. of the Symposium on Cryptography and Information Security, Okinama, Japan, pp.26-28.<br />
322<br />
NC<br />
PSNR<br />
WPSNR<br />
(10)<br />
x max<br />
NC
Amer Sedeeq Mustafa<br />
Sung. K. Je, Cheol. K. Kim, and Eui (2002). Y. Cha “An Adaptive Watermarking Algorithm Using Wavelet Transform”,<br />
Neural Network and Real-World Applications Lab, Department of Computer Science, Pusan National University,<br />
Gumjung-gu, Pusan 609-735, Korea.<br />
Solachidis, V, Nikolaidis, N and Pitas, I. (2000) “Watermarking Polygonal Lines using Fourier Descriptors”, Proc. of<br />
the IEEE International <strong>Conference</strong> on Acoustics, Speech and Signal Processing, Istanbul, Turkey, vol. IV,<br />
pp.1955-1958.<br />
Voigt, M and Busch, C. (2003) “Feature-based Watermarking of 2D-vector Data”, Proc. of the SPIE, Security and<br />
Watermarking of Multimedia Content, Santa Clara, USA, vol.5020, pp.359-366.<br />
Voigt, M, Yang, B and Busch, C. (2004) “Reversible Watermarking of 2D-vector Data”, Proc. of the Multimedia and<br />
Security Workshop on Multimedia and Security, Magdeburg, Germany, pp.160-165.<br />
Xia M. Niu, ChengY. Shao, and Xiao T. Wang (2007) “GIS Watermarking: Hiding Data in 2D Vector Maps”, Studies in<br />
Computational Intelligence, vol.58 , pp.123-155.<br />
Zheng, Liangbin, Wang, Yulu Jia, Qun. (2009) “Research on Vector Map Digital Watermarking Technology”, Proc. of<br />
the IEEE Education Technology and Computer Science. ETCS '09. First International Workshop.<br />
323
Evaluation of IT Investment Methods and Proposing a<br />
Decision Making Model<br />
Shirin Nasher, Mehrdad Kalantarian, Ahmad Akbari, Ali Suzangar, Mohammad<br />
Kajbaf and Negar Madani<br />
Infoamn IT Consultancy CO., Tehran, Iran<br />
shirin_nasher@vu.iust.ac.ir<br />
mehrdad_kalantarian@vu.iust.ac.ir<br />
akbari@iust.ac.ir<br />
a.suzangar@infoamn.com<br />
m.kajbaf@infoamn.com<br />
n.madani@infoamn.com<br />
Abstract: Information technology investment decision making is one of the significant issues. Since the IT<br />
investment evaluation is not just based on direct and tangible factors and many other intangible and indirect<br />
qualitative criteria influence this evaluation. Generally, there are two different approaches in evaluation methods with<br />
their own advantages and disadvantages: tangible methods such as Discounted Cash Flow, Net Present Value,<br />
Information Economics, etc. and intangible methods such as Value Analysis, Multi Objective Multi Criteria, Critical<br />
Success Factors, etc. But a more effective and precise road map is to guide decision makers to choose an<br />
appropriate multi criteria model that consider both tangible and intangible factors together. In this paper by literature<br />
review of these mentioned methods, various tangible and intangible factors were determined from different academic<br />
papers and practitioner resources and then were classified in two domains and a number of sub-domains. In order to<br />
obtain complete and applicable criteria, five reduction factors were defined, i.e. clarity, completeness, non–<br />
redundancy and operationality. Then this criteria list was delivered among a number of mangers and information<br />
technology specialists. According to the given answers, a new criteria list was obtained with eliminating nonapplicable<br />
criteria. As a consequence, to assess the importance of each criterion for creating the model, the rating<br />
scores, one to five were defined and added to the new lists. Then these new criteria lists were conducted among<br />
CIO, CEO, CFO and other related specialists in different Iranian companies to customize these criteria according to<br />
their business strategies and requirements. Score one represents not important and five represents very important.<br />
The results are assumed as the minimum level of criteria with maximum coverage in different information technology<br />
projects. In the next step, based on the results, we developed an analytic hierarchy decision making model. The<br />
results of this research indicate that this model is applicable and can be easily expanded by aggregating new sub<br />
criteria to be customized for different IT investment evaluations.<br />
Keywords: IT investment, evaluation of IT project, intangible benefits, decision making model, economic factors,<br />
rational decision making<br />
1. Introduction<br />
Information technology is actively applied in various segments of society that has a strong impact on the<br />
global performance of the businesses. In the past few years, the rapid growth of IT investment is one of<br />
the most incredible issues in different business units and organizations. On the other hand, its adoption is<br />
based on tangible and intangible aspects. But there is a strong persistency to concentrate on tangible<br />
aspects, but the fact is that intangible factors often are the most important one associated to investment.<br />
Indeed, it is not easy for administrations to evaluate the real return of IT investment, so convincing<br />
organization chief managers to allocate their budget to information technology projects is difficult. Even<br />
though organizations are eager to spend considerable amount of time and money to select appropriate IT<br />
project, but they may not include all relevant criteria in evaluating the IT investment (Wu and Ong<br />
2008).Hence even nowadays aligning IT investment strategies with business policies is an important<br />
issue in organizations (Borenstein and Baptista Betencourt 2005), (Goh and Kauffman 2005), (Joseph<br />
Wen, Yen and Lin 1998).<br />
Although there are different methods and tools developed for IT investment evaluation, but the majority<br />
only take easily measured financial aspects into account without any attention to strategic or operational<br />
aspects. Among these methods, the multi criteria multi objective method combines various criteria in<br />
general specific and applied to a unique situation. Hence the existence of a model based on the validated<br />
tangible and intangible criteria, independent from the context related to the specific IT investment seems<br />
very important to support a decision (Chen, Zhang and Lai 2009), (Schniederjans and Hamaker 2003),<br />
(Schniederjans , Hamaker and Schniederjans 2004).<br />
324
Shirin Nasher et al.<br />
The main goal of this study is to identify and structure a set of relevant and validated minimal measurable<br />
set of criteria and sub criteria to formulate a hierarchy decision making model for a specific IT investment.<br />
These criteria were raised from different research and practitioner resources and validated through the<br />
evaluation of specialist to support and develop decision makers confidence in the selection of the most<br />
suitable project for a particular case (Goldstein, Katz and Olson 2003), (Mahmood and Szewczak 1999).<br />
In this paper, first in the literature review section, different tangible or intangible IT investment evaluation<br />
methods will be introduced. Then the advantages and disadvantages of these methods will be compared<br />
to each other.<br />
Second, the methodology of this research will be described in the consequence. At last in the conclusion<br />
the benefits of the proposed model will be presented.<br />
2. Literature review<br />
The best IT investments are those which help to maximize the value of the firm. They also contend that to<br />
maximize the value of the firm, IT investment decisions need to be able to maximize IT benefits while<br />
minimize IT risks (Joseph Wen, Yen and Lin 1998).<br />
There are three reasons why being familiar with IT benefits and risks is important. First, some IT benefits<br />
are lost through inappropriate management, while some are lost because they are not recognized by the<br />
management in the early IT planning stage. Second, it is important to identify the benefits to be measured<br />
prior to selecting applicable evaluation methods. This is because some methods are suitable for<br />
evaluating tangible benefits while others are more suitable for intangible benefits. Finally, the recognition<br />
of risk as an important component in IT investment decision making has long been recognized (Epstein<br />
and Rejc 2004-2005), (Joseph Wen, Yen and Lin 1998).<br />
2.1 Information technology benefits and risks<br />
In general, the benefits of IT investments can be classified into five broad classes, the purpose of which<br />
is to (1) increase productivity and operating process performance; (2) facilitate management support; (3)<br />
gain competitive advantages; (4) provide a good framework for business restructure or transformation<br />
and (5) provide clearance of expenditures (Epstein and Rejc 2004-2005).<br />
IT has provided many benefits to corporations over the years. Since we are living in a global information<br />
society with a global economy, which is increasingly dependent on the creation, management and<br />
distribution of information resource. However, investments in information technology are subject to higher<br />
risks than any other capital investments for several reasons. First, their components are comparatively<br />
fragile. Second, information systems are likely to be the target of disgruntled workers, protester, and even<br />
criminals. They can also fall in the hands of the competitors. Finally, the decentralization of information<br />
systems and the use of distributed processing have increased the difficulty of design, development,<br />
management, and protecting information systems. IT risks are classified into two general classes: (1)<br />
physical risks; and (2) managerial risks (Epstein and Rejc 2004-2005).<br />
2.2 Information technology investment evaluation methods<br />
Given countless existing methods, a broader framework for categorizing and understanding evaluation<br />
techniques seems highly desirable. Such a framework by dividing IT evaluation approaches into two<br />
broad categories based upon their underlying assumptions: objective/rational or subjective/political. In the<br />
objective/rational category, they further divided the objective/rational category into two zones: efficiency<br />
i.e., doing things correctly and effectiveness i.e., doing the correct things. The subjective/political<br />
category described as the understanding zone i.e., discovering why things are done (Tuten 2009).<br />
Traditional IT evaluation practice operates from an objective/rational point of view, focusing on the<br />
efficiency and effectiveness of solutions. Such evaluation approaches are grounded in a positivist<br />
epistemology—an epistemology that, when applied to this context, holds that information systems are<br />
inherently objective and rational. Therefore, practitioners should evaluate information systems using<br />
objective/rational methods (Tuten 2009).<br />
Overall, researchers have tended to describe traditional evaluation methods as formal, overt, ritualistic,<br />
mechanistic, quantitative, and/or prescriptive in their efforts to determine the costs, benefits, and risks<br />
associated with IT investments (Tuten 2009).<br />
325
Shirin Nasher et al.<br />
Nevertheless, researchers have suggested that formal evaluation frequently fails to be undertaken with<br />
rigor and is completely avoided by practitioners in many cases. In a recent study IT evaluation practices<br />
in <strong>European</strong> companies, researchers found that only one third of the organizations surveyed conducted<br />
formal evaluations. Yet, when organizations perform IT evaluation, they tend to employ traditional<br />
methods that hold considerable legitimacy with executives and managers. This finding found that<br />
quantitative evaluation methods were widely used by the organizations conducting formal evaluations<br />
(Tuten 2009). Table1 shows these methods based on mentioned categorize.<br />
As Table1 shows, these rational/objective effectiveness methods may be subcategorized into one of<br />
three groups of methods: economic, non-economic, and hybrid.<br />
Table 1: The information technology investment evaluation methods (Tuten 2009)<br />
Economic<br />
Discounted Cash Flow<br />
Cost Benefit Analysis<br />
Payback Period<br />
SEASAME<br />
Return on Management<br />
Return on Investment<br />
Options Theory<br />
Risk/Sensitivity Analysis<br />
Effectiveness Zone<br />
Non-economic<br />
User Information<br />
Satisfaction<br />
Hybrid<br />
Balanced Score Card<br />
Critical Success Factors<br />
Information Economics<br />
Multi-Criteria Approaches<br />
Value Analysis<br />
Efficiency Zone<br />
Simulation<br />
TQM/Software<br />
Metrics<br />
2.2.1 Economic methods<br />
In particular, the researcher discussed each of the following widely cited methods: Discounted Cash Flow<br />
(DCF) techniques, Cost/Benefit Analysis (CBA), payback period, Systems Effectiveness Study and<br />
Management Endorsement (SESAME), Return on Management (ROM), Return on Investment (ROI),<br />
options theory, and risk sensitivity analysis (Tuten 2009).<br />
2.2.2 Non-economic method<br />
In the rational/objective literature stream, techniques for measuring user satisfaction, particularly the User<br />
Information Satisfaction (UIS) method, provide the notable exception (Tuten 2009).<br />
2.2.3 Hybrid methods<br />
Hybrid approaches may utilize financial/economic factors and/or non-economic dimensions to evaluate<br />
information systems. All of the following methods have been associated with the rational/objective stream<br />
of IT evaluation techniques. These approaches vary considerably with respect to their degree of apparent<br />
objectivity, as demonstrated by either their reliance on quantitative measures or empirically observable<br />
outcomes. For example, in practice Information Economics relies heavily on their quantitative enhanced<br />
ROI metric. In contrast, Critical Success Factors (CSF) method utilizes a dialogic approach to uncover<br />
executives’ explicit and implicit goals and objectives. In this sense, the term hybrid provides an apt<br />
description for this group’s diversity of methods and measures. These methods are such as Balanced<br />
Score Card (BSC), Critical Success Factors (CSF), Information Economics (Parker and Besnson 1988)<br />
and etc. (Tuten 2009).<br />
Note that, the focus of this paper is on the effectiveness zone methods, so it has preferred to eliminate<br />
describing other zone. Since there are differences between these more applicable methods, Table2 has<br />
summarized the previous discussions based on methods attributes.<br />
In Table2, the IT benefit factors column shows that which of the mentioned above methods concentration<br />
is on tangible or intangible criteria. Therefore these benefit factors can determine the criteria that are<br />
required for IT investment evaluation and decision making lastly.<br />
326
Shirin Nasher et al.<br />
Table 2: Differences among information technology investment evaluation methods (Joseph Wen, Yen<br />
and Lin 1998)<br />
Evaluation<br />
category<br />
ROI<br />
Model/procedure<br />
examples<br />
NPV, DCF, pay-back<br />
period formulas<br />
Measures of IT<br />
benefits factors<br />
tangible<br />
Measures of<br />
IT risks<br />
discount<br />
rates,<br />
surrogate<br />
measures<br />
CBA cost/benefits formulas tangible factors same as ROI<br />
ROM<br />
IE<br />
MOMC<br />
VA<br />
CSF<br />
Productivity based<br />
formulas<br />
same as ROI<br />
supplemented with<br />
ranking and scoring<br />
math models and<br />
multistage iterative<br />
processes<br />
multistage, evolutionary<br />
process<br />
multistage,<br />
evolutionary process<br />
RO Multistage process<br />
PA Financial models<br />
Delphi<br />
approach<br />
3. Methodology<br />
multistage,<br />
evolutionary process<br />
tangible, labor<br />
value-added as<br />
intangible<br />
tangible and<br />
some intangible<br />
tangible and<br />
intangible<br />
tangible factors<br />
user’s surrogate<br />
measures<br />
tangible and<br />
intangible<br />
factors<br />
measures for<br />
cost<br />
savings<br />
user’s surrogate<br />
measures<br />
Not<br />
addressed<br />
surrogate<br />
measures,<br />
risks with<br />
ranking and<br />
scoring<br />
several<br />
measures<br />
of utility and<br />
risks<br />
not<br />
addressed<br />
user’s<br />
surrogate<br />
measures<br />
surrogate<br />
measures<br />
for risks<br />
and costs<br />
direct<br />
measures<br />
of risks<br />
user’s<br />
surrogate<br />
measures<br />
Major<br />
advantages<br />
mainly<br />
quantitative<br />
focus on<br />
efficiency<br />
mainly<br />
quantitative<br />
focus on<br />
effectiveness<br />
mostly qualitative<br />
measures of<br />
efficiency<br />
qualitative and<br />
quantitative<br />
measures<br />
mainly<br />
quantitative,<br />
multiple and<br />
conflicting<br />
objectives<br />
tangible factors<br />
intangible<br />
factors,<br />
centered on<br />
effectiveness<br />
many intangible,<br />
centered<br />
on effectiveness<br />
higher efficiency<br />
tangible and<br />
intangible<br />
factors<br />
Major limitations<br />
no intangible,<br />
reliance<br />
on accounting<br />
data<br />
surrogate<br />
measures<br />
for intangible<br />
factors<br />
limited<br />
quantitative<br />
measures,<br />
assumptions<br />
hard to meet<br />
major<br />
simplifying<br />
assumptions<br />
and<br />
models<br />
relatively new<br />
in MIS, still in<br />
development<br />
prototyping,<br />
need<br />
several<br />
revisions to<br />
final results<br />
highly<br />
qualitative<br />
process<br />
highly<br />
subjective and<br />
qualitative<br />
mainly<br />
quantitative<br />
highly<br />
qualitative<br />
In this paper, IT investment decision making problem has been presented. The methodology of problem<br />
solving with the purpose of building the model to evaluate IT investments would be described here. This<br />
methodology has three steps that would be explained as follows:<br />
Gathering and reducing criteria<br />
Weighting<br />
Proposing model<br />
3.1 Gathering and reducing criteria<br />
This stage has comprised of two steps. At first by reviewing pervious research studies concerning IT<br />
investment evaluation, many of the criteria and sub criteria were identified (Renkema and Berghoutb<br />
1997), (Kraemer and Dedrick 1994), (Joseph Wen, Yen and Lin 1998), (Henderson and Venkatraman<br />
1990). More than 250 criteria were enlisted in an excel file that were concerned both tangible and<br />
intangible factors. After that these criteria were classified into four domains in the separated excel sheets,<br />
327
Shirin Nasher et al.<br />
i.e., operational, tactical, strategic and risk. Each domain has its own criteria and relative sub criteria. The<br />
definitions and characteristics of the criteria were described in each sheet. The first sheet was considered<br />
for the brief description about this research.<br />
Second, because this raw list has many unrealistic and non-applicable criteria, a reduction process was<br />
conducted to identify clear and operational criteria list. Therefore, reduction factors were defined as<br />
follows: the clarity, completeness, non–redundancy and operationality. In the consequence, these factors<br />
have been added for each criterion in the new columns with check boxes to that excel sheet.<br />
This file has delivered among ten managers and information technology specialists. Based on the<br />
gathered results, these actions were done:<br />
1. If the most specialists had voted that the criterion is not clear, it must be defined clear.<br />
2. If the most specialists had voted that the criterion is not complete, it must be completed.<br />
3. If the most specialists had voted that the criterion is redundant, it must be eliminated.<br />
4. If the most specialists had voted that the criterion is not operational, it must be changed or at last<br />
eliminated.<br />
After applying reduction factors to the set of 250 criteria and these actions, several criteria were<br />
eliminated and several were changed or completed. At last the new minimal measurable list of criteria<br />
was obtained. This new list has two domains, i.e. “strategic” and “operational and tactical” with eleven<br />
criteria and forty two sub criteria. A brief description of each eleven criteria is as follows:<br />
1. Strategic alignment: This criterion is related to this crucial point that which investment can align<br />
with organizational strategy and also the possibility of IT providing opportunities for businesses.<br />
2. Competitive advantages: It refers to the organization competitive position among market and<br />
supplier and other related items that mentioned in this criterion.<br />
3. Organizational productivity: It is about the organizational productivity improvement.<br />
4. Market assessment: This criterion deals with the pervious background of the specific IT project<br />
and mentions its former customers and their credit in the market. Also, investigates the ability of<br />
project in the market anticipating.<br />
5. Compliance: It seeks to identify the compliance between proposed investment with the current<br />
internal or external requirements and standards.<br />
6. Quality: It tends to focus on the IT performance characteristics and the impact of it on the services<br />
improvement to support business processes.<br />
7. Security: This criterion and its related items refer to three main aspects (Confidentiality, Integrity<br />
and availability) of security in information systems.<br />
8. Flexibility and compatibility: This criterion and its related items refer to assess program<br />
reliability.<br />
9. Cost: It is related to the different costs associated with a particular investment.<br />
10. Satisfaction: It considers external customers or internal user’s satisfaction about the new<br />
proposed investment.<br />
11. Effects: This criterion tends to focus on the effects of investment on the internal organizational<br />
culture and communications.<br />
The primary model based on these minimal measurable validated criteria was shown in figure1.<br />
3.2 Weighting<br />
In this step to determine domains, criteria and sub criteria importance in order to proposing a model, 5<br />
importance degrees were defined in new lists for each criterion i.e., very important, important, relatively<br />
important, less important and not important. Then these new lists were conducted among twenty of IT<br />
experts with minimum 10 years professional experience in Iran. They determined importance degree for<br />
each criterion based on their own organizational strategies and business requirements. The scores one<br />
to five were defined that score one represents not important and five represents very important. After<br />
answer sheet gathering, first the stability of given answers was calculated by SPSS. It was 91.2 percent<br />
which indicates that IT expert’s answers were at minimum deviation.<br />
328
Shirin Nasher et al.<br />
Figure1: The multi criteria model for IT investment evaluation<br />
3.3 Weighting<br />
In this step to determine domains, criteria and sub criteria importance in order to proposing a model, 5<br />
importance degrees were defined in new lists for each criterion i.e., very important, important, relatively<br />
important, less important and not important. Then these new lists were conducted among twenty of IT<br />
experts with minimum 10 years professional experience in Iran. They determined importance degree for<br />
each criterion based on their own organizational strategies and business requirements. The scores one<br />
to five were defined that score one represents not important and five represents very important. After<br />
answer sheet gathering, first the stability of given answers was calculated by SPSS. It was 91.2 percent<br />
which indicates that IT expert’s answers were at minimum deviation.<br />
Second, the mean value of each criterion and domains were calculated. The cut off value of 3.0 within<br />
values was assumed and those criteria which their mean values are greater or equal 3.0 were identified<br />
and ignored. The 3.0 value is almost moderate importance of each criterion. Based on this assumption,<br />
just one of the sub criteria has been eliminated and excluded from the model. This criterion was<br />
“improvement in organizational communication” from the “effects” criteria. This condition shows that all<br />
these criteria are important and reduction process in the previous step was successful to capture a<br />
minimal set of attributes for IT investment evaluation.<br />
Third, as a consequence based on mean values the entire criterions were ranked. Table3 shows the<br />
gathered results. According to this table, security criteria are the most important and “Improvements<br />
internal and external organizational culture” criterion has less importance within all criteria.<br />
After that, the overall weighting of criteria and domains were calculated with dividing each mean value by<br />
the total sum of its related criteria and sub criteria mean values. This is called prioritizing. By prioritizing<br />
them, the level of each criterion would be determined in its criteria and in an analysis hierarchy model<br />
they would be shown at decreasing level. Figure2 illustrates the hierarchy model based on the<br />
decreasing prioritization.<br />
According to mentioned model, strategic domain has more impact on the investment final decision<br />
making than operational domain. Indeed, nowadays it is the fact organization strategy and IT investment<br />
alignments with strategies are so important for chief managers.<br />
329
Table 3: Evaluation criteria and statistical results<br />
Domain<br />
Strategi<br />
c<br />
0.516<br />
Operati<br />
onal<br />
and<br />
Tactical<br />
0.484<br />
Criteria<br />
Strategic Alignment<br />
0.287<br />
Competitive<br />
Advantages<br />
0.167<br />
Organizational<br />
productivity<br />
0.267<br />
Market Assessment<br />
0.148<br />
Compliance<br />
0.131<br />
Quality<br />
0.217<br />
Security<br />
0.247<br />
Flexibility<br />
and<br />
compatibility<br />
0.133<br />
Cost<br />
0.183<br />
Satisfaction<br />
0.117<br />
Effects<br />
0.103<br />
Shirin Nasher et al.<br />
Sub Criteria<br />
Impact on achieving IT strategic<br />
objectives<br />
Impact on aligning IT with organization<br />
strategy<br />
Organization governance<br />
Improved efficiency of current business<br />
process<br />
Improved efficiency of changed business<br />
process<br />
Improved efficiency of new business<br />
process<br />
Customers<br />
Partners<br />
Suppliers<br />
Competitors<br />
New incoming<br />
Increase organizational productivity<br />
The number and credit of previous<br />
customers<br />
Competitor opinion<br />
Define market<br />
Market anticipating<br />
Internal standards and requirements<br />
External standards and requirements<br />
Response Time<br />
Turnaround Time<br />
Error Rates<br />
Program Code Quality<br />
Quality Assurance<br />
Customization of Outputs<br />
Outputs Format<br />
Comparability of output produced<br />
Easy output format<br />
Security( Confidentiality, Integrity,<br />
Availability)<br />
Flexibility against technological change<br />
compatibility<br />
Ability to customize<br />
Update Feature<br />
Facilitate changes in the code<br />
Buy<br />
Implementation<br />
Human Resources Training<br />
Maintenance<br />
Time<br />
Internal satisfaction<br />
External satisfaction<br />
Improvements internal and external<br />
organizational culture<br />
Importance<br />
Average<br />
4.5<br />
4.65<br />
4.14<br />
4.1<br />
3.72<br />
3.7<br />
4.55<br />
4.35<br />
3.4<br />
4.3<br />
3.1<br />
4.63<br />
3.91<br />
3.55<br />
3.85<br />
3.83<br />
3.11<br />
3.2<br />
4.23<br />
3.90<br />
4.31<br />
3.75<br />
4.05<br />
3.73<br />
3.89<br />
3.18<br />
4.17<br />
4.85<br />
4.25<br />
4.4<br />
4.2<br />
3.68<br />
3.35<br />
4.33<br />
4.6<br />
3.8<br />
4.15<br />
3.95<br />
3.65<br />
3.88<br />
3<br />
Rank<br />
6<br />
2<br />
17<br />
18<br />
30<br />
31<br />
5<br />
8<br />
35<br />
11<br />
40<br />
3<br />
21<br />
34<br />
25<br />
26<br />
39<br />
37<br />
13<br />
22<br />
10<br />
28<br />
19<br />
29<br />
23<br />
38<br />
15<br />
1<br />
12<br />
7<br />
14<br />
32<br />
36<br />
9<br />
4<br />
27<br />
16<br />
20<br />
33<br />
24<br />
41<br />
Weight<br />
0.181<br />
0.184<br />
0.166<br />
0.165<br />
0.150<br />
0.149<br />
0.231<br />
0.221<br />
0.172<br />
0.218<br />
0.157<br />
1<br />
0.258<br />
0.234<br />
0.254<br />
0.253<br />
0.493<br />
0.507<br />
0.120<br />
0.111<br />
0.122<br />
0.107<br />
0.115<br />
0.106<br />
0.110<br />
0.090<br />
0.118<br />
1<br />
0.214<br />
0.221<br />
0.211<br />
0.185<br />
0.168<br />
0.208<br />
0.221<br />
0.182<br />
0.199<br />
0.190<br />
0.483<br />
0.516<br />
The security criteria are the most important criteria and the impact of it on the choosing IT project is much<br />
more. Also, the prioritizing sub criteria can help to have a scale to understand the impact of them on the<br />
investment decision. Note that, the sum of sub criteria weights is one and it shows their overall impacts<br />
are considered on the final investment decision making.<br />
330<br />
1
Figure 2: Analytic hierarchy decision making model<br />
3.4 Proposing model<br />
Shirin Nasher et al.<br />
Since the model wants to illustrate the criteria importance level on the final decision making process,<br />
another hierarchy model was proposed. In this model, two domains were determined on the top of the<br />
model. This model is comprised hierarchical level based on five unit intervals that the mean value of each<br />
criterion determine its own level. By comparing these values, they have been assigned to their<br />
appropriate level as shown in Figure3.<br />
There was an unanimous consensus that this proposed model has improved the decision process in<br />
order to have consistent and complete criteria based on different tangible and intangible factors for IT<br />
investment evaluation. Note that this proposed model is general model and independent from the specific<br />
IT investment to aid investment decision making process.<br />
4. Conclusion<br />
Adoption of the new information technology requires investment justification. Also, IT investment decision<br />
making is based on evaluating tangible and intangible criteria. But the intangible criteria are complex to<br />
evaluate and therefore this action may take a long time and efforts. So many managers prefer to neglect<br />
them and just focus on the financial factors. They think that IT investment affects on their organization<br />
and enhances productivity efficiency in situ. But indeed return on IT investment has a potential latency to<br />
create value just because of these intangible factors. So it shows the importance of these criteria to have<br />
an appropriate investment evaluation. The existence of the pre-defined criteria hierarchy model based on<br />
the consensus of many specialist, can structured the decision making process systematically. Hence this<br />
article proposed criteria list and model can significantly decrease times and efforts for managers and<br />
organizations.<br />
So in this paper, we first determine these tangible and intangible criteria and then categorize them in<br />
different criteria and sub criteria. As a consequence, we give importance rank to them according to the<br />
experiment of chief managers in Iran. Based on these ranking, at last we proposed the hierarchy model in<br />
331
Shirin Nasher et al.<br />
order to consider all these tangible and intangible criteria in IT investment evaluation. Note that since<br />
there are many sub criteria and they may be changed from one IT evaluation to another, we try to<br />
consider general criteria to create a generic model. So this proposed model is applicable and can be<br />
easily expanded by aggregating new sub criteria to be customized for different IT investment evaluations.<br />
Figure 3: Final hierarchy decision making model<br />
References<br />
Borenstein, D. and Baptista Betencourt, P. R. (2005) “A Multi-criteria Model for the Justification of IT Investments”,<br />
INFOR Journal, Vol. 43, No. 1, Feb, pp. 1-21.<br />
Chen, T., Zhang, J. and Lai, K. (2009) “An Integrated Real Options Evaluating Model for Information Technology<br />
Projects under Multiple Risks”, International Journal of Project Management, Vol. 27, Issue 8, pp.776–786.<br />
Epstein, M. and Rejc, A. (2004-2005) “Measuring the payoffs of IT investments”, CMA Management Journal, Vol. 78,<br />
No. 8, pp. 20-25.<br />
Goh, K.H. and Kauffman, R.J. (2005) “Towards a Theory of Value Latency for IT Investments”, Proceedings of the<br />
38th Hawaii International <strong>Conference</strong> on System Sciences, pp. 1-9.<br />
Goldstein, Ph., Katz, R.N. and Olson, M. (2003) Understanding the Value of IT, EDUCAUSE Quarterly magazine,<br />
Number 3, pp. 14-18.<br />
Henderson, J. and Venkatraman, N. (1990) Strategic Alignment: A model for Organizational transformation Via<br />
information technology,<br />
Joseph Wen, H., Yen, D.C. and Lin, B. (1998) “Methods for measuring information technology investment payoff”,<br />
Human Systems Management Journal, Vol. 17, No. 2, January, pp. 145–153.<br />
Kraemer, K. and Dedrick. J. (1994) “Payoffs from Investment in Information Technology: Lessons from the Asia-<br />
Pacific Region”, World Development Journal, Vol. 22, No. 12, pp. 1921-1931.<br />
Mahmood, M.A., Szewczak, E.J. (1999) Measuring Information Technology Investment Payoff: Contemporary<br />
Approaches, Idea Group Publishing, London.<br />
Parker, M.M. and Benson, R.J. (1988) Information Economics: linking business performance to information<br />
technology, Prentice-Hall.<br />
Renkema, Th. J.W. and Berghoutb, E.W. (1997) “Methodologies for information systems investment evaluation at the<br />
proposal stage: a comparative review”, Information and Software Technology Journal, Vol. 39, Issue 1, pp. 1-<br />
13.<br />
Schniederjans, M. and Hamaker, J. (2003) “A new strategic information technology investment model”, Management<br />
Decision Journal, Vol. 41, No. 1, pp. 8-17.<br />
Schniederjans, M.J., Hamaker, J.L. and Schniederjans, A.M. (2004) Information Technology Investment Decision-<br />
Making Methodology, World Scientific Publishing Co., London.<br />
Tuten, P. M. (2009) “A Model for the Evaluation of IS/IT Investments”, A dissertation submitted in partial fulfillment of<br />
the requirements for the degree of Doctor of Philosophy in Information Systems, Graduate School of Computer<br />
and Information Sciences Nova Southeastern University, USA.<br />
Wu, L.C. and Ong, C.S. (2008) “Management of information technology investment: A framework based on a Real<br />
Options and Mean–Variance theory perspective”, Technovation Journal, Vol. 28, Issue 3, March, pp. 122-134.<br />
332
Critical Factors in the use of Mobile Learning by “Digital<br />
Natives” on Portuguese Teaching<br />
Mário Carrilho Negas and Paulo Ramos<br />
Universidade Aberta, Portugal<br />
mnegas@univ-ab.pt<br />
rcjp@sapo.pt<br />
Abstract: The purpose of this paper is to discuss the mLearning as a useable resource in formal education (in a<br />
classroom context), in order to improve the learning experience of young people. It is intended to enumerate the<br />
possible success factors that can contribute to the mLearning and to show that, when properly integrated into school<br />
curricula, they could be a valid learning experience for youth. Young people are referred to as the "digital natives"<br />
because of their constant use of digital resources in their day-to-day. Some factors of significant importance are<br />
identified, namely the fact that young people see the mLearning as a valid learning tool, and that they accept its use<br />
in teaching, provided it is not to replicate contents taught in the classroom. This article highlights the important role<br />
that mobile learning can have on academic life, and the factors which underpin its success, such as the widespread<br />
use of mobile devices by young people. The "Digital Natives" use mobile technology with great frequency, particularly<br />
to keep in touch with family and social networks as they see the technology as an essential part of their lives. Mobile<br />
technology solutions may help transforming the way to create, teach and learn; in this sense, mobile phones may<br />
have an important role to play in tomorrow’s school.<br />
Keywords: mobile learning, eLearning, moving learning capabilities; mobile devices; wireless network; digital<br />
natives; teaching<br />
1. Introduction<br />
The constant evolution of the technologies, as it is the case of the mobile technologies and particularly<br />
mobile phones, has made possible a significantly wide set of solutions that help to change the way we<br />
create, teach and learn. The ubiquity of mobile technology in our day-to-day (Green et al, 2001, Katz and<br />
Aakhus, 2002; Levinson, 2004), and this fact has contributed to the emergence of a new educational<br />
paradigm, the mobile learning (mLearning). For Prensky (2001) and Kukulska-Hulme (2005) the concept<br />
of mobile learning was an evolving subject that is no longer associated exclusively with the<br />
characteristics of portability and mobility of the equipment. On the contrary, it must be understood<br />
nowadays as a concept associated with the mobility of the user.<br />
The ancestral human search for knowledge has led to the search for newer and more sophisticated ways<br />
to store, process and transmit information, leading to the evolution of tools which enable that search.<br />
Palfrey, J. and Gasser, U. (2008) state that although several decades of constant technology evolution<br />
and the resulting saturation of the digital technologies in some cultures, there wasn’t yet a generation that<br />
has lived exclusively in the age of the digital technology.<br />
In 2007, the Portuguese government launched a major public initiative – especially given its extent – in<br />
order to distribute laptops with Internet access to children and youth. This helped creating the right<br />
conditions for the use of information and communication technologies (ICT) in education since ICT<br />
learning occurs at a younger age.<br />
In the 60´s, Herbert Marshall McLuhan predicted that the planet would become a teaching room and our<br />
virtual address.<br />
Learning environments have acquired new characteristics; they are more open, more interactive and<br />
flexible. These characteristics may allow combining different learning styles. The use of mLearning in this<br />
context is an excellent opportunity for affirmation of this tool in the context of teaching-learning<br />
environment.<br />
The evolution of mobile devices may enable – if used properly by education agents – the access to new<br />
knowledge and the review of previously acquired knowledge.<br />
333
Mário Carrilho Negas and Paulo Ramos<br />
2. The "Digital Natives" and the teaching-learning<br />
Palfrey and Gasser (2008) stated that the first generation of "Digital Natives" is the current youths. In fact,<br />
this generation is the first one that was born and has lived in the digital age since birth. According to the<br />
authors, this will contribute to a major overhaul in the way the world is organized.<br />
The way young people perceive the world will have implications on issues of political, cultural and family<br />
life, which will be changed forever. Students today are questioning the current practice and breaking with<br />
the so-called traditional guidelines (Boyle, 2008), questioning some of the practices of the 20th Century.<br />
In comparison to the "Digital Immigrants" ("Digital Immigrants"), the “Digital Natives” have some<br />
differences in speech, social interaction, and are fluent in using new ICT. It is possible to identify three<br />
generational groups: Baby Boomers (1946-1964), the "Generation X" (1965-1977) and the "Digital<br />
Natives" (1978-1994). The latter, in contrast to previous generations, tend to use information they gather<br />
from various online sources.<br />
The use of mobile technologies by young people tends to overcrowding. It is relevant to know to what<br />
extent the use of mLearning can contribute as a motivational factor in the process of teaching and<br />
learning of young people. In this sense the institutions and teachers are aware of this reality: how young<br />
people react to this technology, how they adopt it, and if the educational programs of schools enable the<br />
use of these tools.<br />
The discussion of new technologies in education is a constant in the adoption of new methodologies;<br />
however, to digital natives this is not an issue, because younger generations begin academic life with no<br />
doubts or fears (Tori, 2010).<br />
Young people tend to identify the use of mLearning as a possible tool for learning. They also tend to<br />
accept the use of mLearning in their courses, as long as that is not only a repetition of educational<br />
content taught in the classroom, and preferably not just an isolated initiative, but rather a practice<br />
supported by several teachers. The learning content to be provided must be diversified, classified<br />
according to the subjects being taught, and adapted to the possibility of using mobile phones at low cost<br />
and to the prevalence of information in audio and video formats instead of text. The fact that the "Digital<br />
Natives" use technology from an early age eases the process of constant adaptation to technological<br />
change that generally occurs. It's with recognized ability that "Digital Natives" the hardware to their goals<br />
of recreation, using the online environment more than any other preceding generation.<br />
According to Tori (2010), the "Digital Immigrants" and "Digital Natives" have points of convergence. Even<br />
though that may be true, the author mentions that the "Digital Immigrants" contributed significantly to the<br />
“Digital Natives” empathy with ICT.<br />
Despite having been taught by the “Digital Immigrants”, the "Digital Natives" began to adopt new<br />
practices in the use of ICT. Generally, for “Digital Natives”, a day’s time is divided between online and<br />
offline interaction, "nick name" and name, virtual friendships on social networks and real life friendships.<br />
Even though this quick change in reality may be confusing for those who are not "Digital Natives", it is<br />
something intrinsic to the present reality for those who belong to that group. When entering the job<br />
market, this generation carries the practices associated with the digital world. This young generation has<br />
a significant set of distinctive skills in comparison with the current generation, the Generation X.<br />
Virtual social networks are far superior in dimension than the neighborhood, academic or working social<br />
networks, and may also be viable sources of information to be used in collaborative learning<br />
environments: they cover very different subjects, are very heterogeneous and culturally diverse, which<br />
allows the obtainment of diversified and relevant information. However it is also questioned if their<br />
contribution as an educational resource in teaching-learning process may constitute a source of<br />
distraction and waste of time, taking into account the objectives of educational programs (Economist<br />
debates: Social networking ", 2008).<br />
Boyle, L. and Stanford, J. (2008) sustain that “digital natives” are native speakers of technology, fluent in<br />
the digital language of computers, video games and the Internet. They also state that the students, as<br />
digital natives, will continue to evolve and change so rapidly that they will have, in general, a much better<br />
idea of what the future will bring. Since the natives are already busy adopting new systems for<br />
334
Mário Carrilho Negas and Paulo Ramos<br />
communications (instant messaging), sharing (blogs), buying and selling (eBay), exchanging (peer-topeer<br />
technology), creating (Flash), meeting (3D worlds), collecting (downloads), coordinating (wikis),<br />
evaluating (reputation systems), searching (Google), analyzing (SETI), reporting (camera phones),<br />
programming (modding), socializing (chat rooms) and even learning (Web surfing), they have a better<br />
understanding of their surroundings than previous generations.<br />
3. The mobile learning in distant teaching<br />
Pachl et al. (2009) examined the characteristics identified as important for mobile devices in order to<br />
make them so attractive to users in the education field. Portability is one of those factors, as the use of<br />
portable equipment in various life situations became socially acceptable, and naturally they have a vast<br />
potential for providing access to culture, content and features, such as the possibility to record<br />
information in audio, video or text. Another very interesting factor is the number of applications or<br />
services that are can be accessed through a mobile device, some of which being accessible also through<br />
the Internet.<br />
The mobile learning (mLearning) is presented as one of the possible tools of distance learning,<br />
specifically in eLearning. Pachl et al. (2010) state that learning with the use of mobile devices is an<br />
emerging and growing field of research due to the increasing use of mobile equipment in schools,<br />
universities and workplaces. The use of mobile networks is gaining greater importance in unconventional<br />
places. These and other factors have contributed to the growing interest of researchers, particularly in the<br />
area of educational research.<br />
Mobile Learning is an educational perspective, a new dimension in education that could meet immediate<br />
learning needs with great flexibility and interactivity.<br />
Undoubtedly, one of the great advantages of mobile devices is their accessibility, which allows constant<br />
contact with other users (either because the device can always go with the user, or because of the instant<br />
access it allows to the communication network or information requested).<br />
Being easily carried and used in periods of dead time (such as travel or waiting periods), these<br />
applications become an excellent way to access learning content and information in general.<br />
The lifelong learning is increasingly seen as a constant practice. Therefore, there is a great expectation<br />
about whether mobile technologies have the potential and feasibility to be used in lifelong learning<br />
systems in a large scale.<br />
4. The role of mobile technologies in the increase of the sustainability of<br />
distance teaching<br />
When it was created, the cell phone was meant to keep people in touch. Now it has become a true all-inone<br />
device. It enables us to connect to the Internet, e-mail, listening to music, watching TV or even<br />
checking our bank accounts online. Wireless mobile networks have suffered quick qualitative changes<br />
throughout times, since the first analogic generation (1G) became digital (2G). It became a massive<br />
service by the third generation (3G), which integrated wireless digital transactions and services that were<br />
traditionally available in landlines (Harte, Levine & Kikta, 2002), and included better control of information<br />
packages, connections speeds of up to 2 Mbps, multiple radio channels, among other features.<br />
The Long Term Evolution technology (LTE), commonly named 4G, will be the future tendency to fulfill the<br />
requirements of the Digital Natives, and will enable faster connections, with maximum debits of 100/150<br />
Mb on downloads and 50 Mb on uploads. LTE presents a greater efficiency in using the radioelectric<br />
space, and a larger mobile networking capacity than the current 3G, which presents maximum download<br />
speeds of 7,2 Mb and upload speeds of 384 Mb.<br />
It is expected for LTE to be available in 2011, progressively in FDD mode on Europe. Greater speed and<br />
efficiency will also serve as a countermeasure against the expected increase in data traffic. Teachers<br />
may include the technologies available in their teaching methodologies, so that they can provide an<br />
enriching experience to their students (Guy, 2009).<br />
Logistically speaking, LTE may sustain the academic experiences, since it keeps up several bandwidths,<br />
which has special importance in a time when mobile broadband has been growing exponentially. LTE will<br />
also enable the opening of new possibilities, such as services and functionalities that require high<br />
335
Mário Carrilho Negas and Paulo Ramos<br />
transmission debits, like the advanced diffusion of television via internet, online games and web<br />
conferences, whether there is a high or low mobility associated.<br />
The hardware, software and networking quality are critical factors. A recent study (Guy, 2009) by<br />
Schmidt, Hodge, Dickerson, and Ellis (2008) reveals that students often enroll in online courses due to<br />
the schedule flexibility, the information available and its formats (audio, video, among others), instead of<br />
conventional classes, even when the universities and academic campus are geographically close.<br />
As there is an increase in the amount and virtual space of the information that is transacted online,<br />
devices and networks need to adjust to the Digital Natives’ impatience – who want it all on the spot<br />
without waiting. The development of new technologies (hardware, software, telecommunications,<br />
networks and Internet) happens at a frantic rate and the lifestyle of the Digital Natives in their schools and<br />
surrounding social environments has also been acquiring an increasingly accelerated pace. Mobile<br />
technology is more present in our day-to-day academic or professional life (Guy, 2009) with the huge<br />
advantage of not needing a conventional physical space, electricity connection of cable connection to the<br />
Internet. The sales of informatics equipment – notably laptops – have ascended to the 88,3 millions on<br />
the first trimester of the current year worldwide, according to the news, which means that a computer<br />
went from being a superfluous item to become a first necessity tool.<br />
5. Opportunities and challenges of mobile learning in Portugal<br />
Studies about the young users of the technologies of communication and information began in 2006<br />
(Cardoso et al., 2007), with the study entitled “E-Generation - The Uses of Media for Children and Youth<br />
in Portugal”. The methodology used by that study was the application of an online inquiry to young<br />
Internet users. This inquiry was completed 1377 times during the month it was online, from which 1353<br />
entries were considered valid, allowing a characterization of young Portuguese Internet users and their<br />
relationship with technologies and the means of communication. In 2008 this study was conducted again<br />
to deepen the knowledge and the comprehension of the data previously obtained in 2006. The study<br />
highlighted several behavioural characteristics common to children and youths, namely the fact that the<br />
socialization space is now online, in chats, blogs, and social networks. While trying to strengthen their<br />
social networks, cell phones have played an important role as a mean of socialization, but also as an<br />
agent of personal education. According to these studies, most of the young people inquired began using<br />
the Internet at an early age, between 10 and 12 years old, and it is expectable that they tend to do it<br />
increasingly early (Cardoso, et al, 2007, 395). Therefore, young people develop more and more early a<br />
mentality focused on the use of the technologies of information and communication, as well as a range of<br />
skills centered on multitasking (McMahon and Pospisil, 2005, p. 421). The technology and the means that<br />
young people use to communicate during their school years tend to influence the way they relate to the<br />
information and consequently the method by which they prefer to learn. Several authors attribute to these<br />
young people the ability to stay in contact through SMS and watch television, whilst simultaneously<br />
playing computer games and listening to music (Frand, 2000, p. 18; Oblinger, 2003; Rickard and<br />
Oblinger, 2003). The study entitled E-Generation, 2006 (Cardoso et al., 2007) points out that 96,6% of<br />
people between eight and eighteen years old (with a sampling universe of 1353 inquiries) has a cell<br />
phone, which reveals its high utilization rate. In Portugal, the average age at which the inquired people<br />
acquired a cell phone is of 11,8 years old. It is, without a doubt, important for Portugal, but only the near<br />
future will reveal if this fact has contributed to the enhancement of young people’s skills in a more<br />
generalized way. A study conducted by Pew internet & American Life Project reveals that 55% of the<br />
American young Internet users between the age of 12 and 17 are registered on social networks. (Lenhart<br />
e Madden, 2007). Valkenburg, Peter and Schouten (2006) also refer that in Holland about a quarter of<br />
teenagers is currently a member of one or more social networking websites. By analyzing the Top 5 of<br />
the most researched expressions on the Internet in Portugal in 2003, 2005 and 2007, we find, in<br />
decreasing order of hits, in 2003: Portugal; Download; Lisbon; Free; Oporto; in 2005: Sex; Emule; Google<br />
earth Sapo; Benfica; in 2007: Youtube; HI5; Gmail; Wikipedia; Games. It is possible to verify that the<br />
most researched expressions on the Internet in Portugal in 2007 include 5 communication-related<br />
expressions. A study entitled e-Generation 2008 (Cardoso et al., 2008), focused specifically on youth,<br />
from 8 to 18 years old, (with the sampling universe composed of 489 validated inquiries), yielded the<br />
following results: 78,8% of the people who were inquired use e-mail; 70,5% use the IM network; 63,3%<br />
use social networking sites; 43,1% send SMS through the Internet, and 12,4% of the inquired make<br />
phone calls through the Internet. Although the Internet is more and more used by young people for<br />
communication purposes, the mobile phone remains the most used equipment, followed by the TV, and<br />
only then the Internet. In voice communication the mobile phone remains the most widely used<br />
equipment by young people to maintain their social contacts, and it is also the most used to keep them in<br />
336
Mário Carrilho Negas and Paulo Ramos<br />
touch with family members, and as a way to organize tasks and daily meetings (using SMS and phone<br />
calls).<br />
6. The "e-escolinha" and "e-escola" as a driver of mobile learning in Portugal<br />
Guy (2010) affirms in his book that the overall trends in mobile learning technology are becoming clear,<br />
and that devices will continue to have greater processing capacity, greater storage capacity, faster<br />
network access, and will be carried and accessed more and more frequently due to the students’<br />
necessity for research capabilities. Although Portugal is a developing country on many levels, there have<br />
been many physical, logistical and behavioural advances in what refers to the ICT, allowing the country to<br />
minimize or even dissolve the distance to some other more developed countries on this subject. With this<br />
goal in mind, it was created a program to distribute laptops “Classmate PC” at reduced cost for school<br />
children. These laptops were called Magalhães (Magellan) and were manufactured in Portugal with an<br />
Intel processer.<br />
The Program was created to promote the access to the Information Society and to foster info-inclusion,<br />
through the availability of laptops and internet broadband connections, and it has registered an adhesion<br />
of 1 million people in the first two years.<br />
The Program is divided into three parts: e-schools for children from the 5th to the 12th year, and small<br />
schools for children in the first year of school. The Technological Plan for Education has provided<br />
students (especially the most financially needy) with the opportunity to possess a computer, allowing<br />
them to heuristic development, autonomy, and to serve as a complement to their personal development.<br />
This initiative was launched in July 2007 in partnership with the major Portuguese telecommunication<br />
operators, such as Optimus, Vodafone and TMN, including students from the 1st to the 12th grade,<br />
teachers and adult students as well. Several protocols were also signed with software and hardware<br />
designers, to enable the creation of a richer educational experience.<br />
Consequently, the Program was recognized by its effort and awarded on November 2007 with the “Best<br />
<strong>European</strong> Project Award” by Toshiba Europe, represented by the president of the company Noriaki<br />
Hasimoto. In January 2009, in London, Microsoft indicated the Program e-escolas as an example to be<br />
followed by other governments.<br />
The final results of these plans can be seen as an improvement in the students performance and facilitate<br />
the access to reasonable hardware and Internet access. Overall, this project intends to achieve several<br />
objectives, namely: giving opportunities to children and youth from the 5th to the 12th year, younger<br />
students aged 6-10, and developing the technology and the software industries. It should be noted that,<br />
in recent years, Portugal has taken measures to develop these industries, such as the case of mobile<br />
software. Several measures were developed to create a larger offer of Internet services, according to the<br />
2010 annual report from Bareme Internet. Besides this increase, the Portuguese municipal facilities often<br />
include free internet services in libraries and other public areas. In what concerns the private sector, there<br />
is an emergence of internet cafés in which Wi-Fi services offer convenient and inexpensive access to the<br />
internet. There are also conditions for the students to reduce mobile internet monthly payments.<br />
7. Use of mobile learning on Portuguese teaching<br />
Distance learning or distance education (EaD) in Portugal has existed for several years (Cardoso and<br />
Machado, 2001). Currently, resorting to the ICT, the use of the ELearning has become a teaching tool<br />
that can be adapted to the agitated daily life of our time, and that represents a solution for students that<br />
are geographically distant from education institutions or have mobility problems.<br />
According to Leal and Amaral (2004), with the dawn of the 90’s appeared different teaching-learning<br />
practices, supported by the Mobile Learning technologies, that allowed the online participation of each<br />
student/teacher in either synchronous or asynchronous mode. Since then, the Internet utilization rate has<br />
grown exponentially (in Tek, 13 de Outubro de 2010), as well as the number of users. In fourteen years,<br />
the penetration rate grew by 920 per cent, an average of 20 per cent a year. In the present year, 57,1 %<br />
(4 749 million) of the Portuguese periodically access the Internet. Judging by these numbers, it is<br />
possible to perceive the importance of this technology to research information.<br />
337
Mário Carrilho Negas and Paulo Ramos<br />
Figure 1: Number of internet users in Portugal (source: Marktest - Bareme Internet (2010))<br />
Young internauts with ages from 15 to 17 are the group more often connected to the Internet (99,2%) (<br />
Marktest - Bareme Internet, 2010).<br />
The project “mLearning – The role of mobile learning in <strong>European</strong> Education” developed by TecMinho,<br />
Minho University, in association with other institutions from northern Europe, is a reference in the<br />
implementation and development of mobile learning. This project is managed by Ericsson (from Ireland)<br />
and is inserted in the Program Sócrates of the <strong>European</strong> Union, which has been financing mLearning<br />
activities in Portugal. The main outcome of this program will be:<br />
The development of a document that may serve as guideline for politics about the role of mLearning.<br />
The analysis of the initiatives and the role that mLearning currently plays on the 25 members of the<br />
EU;<br />
The study of good practices and current projects concerning mLearning worldwide.<br />
Analysis of the pedagogic aspects of mLearning;<br />
Adaptation of existing learning contents to mLearning;<br />
Development of several formation courses in mLearning by the partner entities involved in the<br />
project;<br />
A scientific analysis of the advantages of mLearning for handicapped people and of the strategies for<br />
presenting learning contents to these students.<br />
According to Junior and Coutinho (2008), of the 1166 individuals who answered the electronic enquiry,<br />
88% claimed to use a mobile phone and 11% stated they also had a PDS or a Smartphone besides their<br />
mobile phone. However, when enquired about mLearning, more than a half of the people didn’t know this<br />
concept of learning, and only 39% said that they recognized the concept applied to teaching and learning<br />
activities. However, 25% had used mobile phone, 15% the tablet computer, 7% the pocket PC and 5%<br />
the Palm or PDA, and the great majority, 48%, revealed they never used any kind of device.<br />
As stated by the study “E-Generation - The Uses of Media for Children and Youth in Portugal”, only a<br />
small minority of young people (5,3%) possesses a laptop and also a minority (6,9%) uses internet<br />
through wireless devices, like the the mobile phone or a laptop (Cardoso et al., 2007).<br />
In spite of being a part of young people’s daily lives, even in the classroom, the eventual potential of<br />
mobile technologies has yet to be explored in eLearning environment.<br />
8. Conclusions<br />
It may be said that the cell phone is a piece of technology which is relatively inexpensive to the average<br />
citizen when compared with other mobile devices. Additionally it is easy to use and is very widespread<br />
338
Mário Carrilho Negas and Paulo Ramos<br />
nowadays. Young people constitute a large group of cell phone users, and they use it in several activities<br />
during the day, for various social and family contacts and for schedule management. In this sense the<br />
utilization of mobile learning as an aid for teaching-learning is a strategy that is gaining supporters as<br />
mobile devices evolve. However, it is known that the “Digital Native” community doesn’t easily accept<br />
mobile learning as a learning tool if it only duplicates the contents that are taught in the classroom. As far<br />
as “Digital Natives” are concerned, mobile learning should be a complement to the classroom learning, in<br />
order to make the learning experience more interactive. Mobile learning allows the enrichment of the<br />
pedagogic offer in a way that fits the students experience outside the classroom, making it faster, easier<br />
and more flexible to access and share information. Even so, there are some difficulties to implement the<br />
mobile learning, namely the fact that the young public is by nature eager to know new things: through the<br />
utilization of the resources already available in the Web 2.0, they acquire innovative and enriching<br />
experiences that may drive them to place their expectations higher than it is possible for teachers and<br />
education institutes to keep up with. If it is accepted that “Digital Natives” have access to technology and<br />
are acquainted with mobile learning, using it constantly, it must be considered, on the other hand, the<br />
teachers’ necessity to also acquire the necessary competences to enable them to provide systematic<br />
answers to the various challenges, namely the creativity, cooperation, communication and involvement<br />
that the mLearning requires (Oblinger, G. (2003). According to Facer, Faux, and McFarlane (2005) the<br />
deficit of technical preparation and the lack of confidence and practice of mLearning that is common to<br />
teachers can have a negative effect on the involvement and pedagogic use of this resource by the<br />
“Digital Natives”. In this sense, Sharples, Taylor and Vavoula (2005, p. 4) identify some of the pedagogic<br />
difficulties concerning mLearning, namely the fact that the pedagogic practices in a classroom are<br />
different and it is intended to cover both formal and informal components of teaching and learning. For<br />
Bryant (2006) and Thomas (2005), the involvement of the “Digital Natives” in the preparation of<br />
pedagogic contents in a cooperative way, “user-led-education”, may be stimulating for the students. Over<br />
the past few years, Portugal has been undertaking efforts to broaden the use of the ICT by young people.<br />
However, the teacher’s instruction and the allocation of funds to support the development and application<br />
of the mLearning on Portuguese schools has been very insufficient. The fact that thousands of laptops<br />
“Classmate PC” have been distributed to children and youth from the 5 th to the 12th grade and also in<br />
elementary schools for children in the first grade, combined with the decreasing costs of mobile<br />
communications, constitutes an important measure, if associated with the high penetration rate of mobile<br />
devices near young people.<br />
References<br />
Ally, M. (2009). Mobile Learning Transforming the Delivery of Education and Training, AU Press, Athabasca<br />
University.<br />
ANACOM (2006). Serviços de Comunicações Electrónicas Móveis, 2.º Trimestre de 2006. (Retrieved November 15,<br />
2010, from: http://www.anacom.pt/streaming/smrp_2t06.pdf)<br />
ANACOM (2009). Serviços de Comunicações Electrónicas Móveis, 2.º Trimestre de 2009. (Retrieved November 15,<br />
2010, from: http://www.anacom.pt/render.jsp?contentId=970592)<br />
Bareme Internet. (Retrieved November 10, 2010, from:<br />
http://www.marktest.pt/marktest/default.asp?c=1210&n=2085)<br />
Boyle, L. and Standford, J. (2008). Digital Natives. Morlacchi Editore, Perugia.<br />
Bryant, T. (2006). Social software in academia. EDUCAUSE Quarterly, 2.( Retrieved November 22, 2010, from:<br />
https://www.educause.edu/ir/library/pdf/eqm0627.pdf)<br />
Cardoso, E., and Machado, A. (2001). Problemática da Adopção de Plataformas eLearning nas Instituições de<br />
Ensino Superior - a fase de iniciação. Associação Portuguesa de Sistemas de Informação, nº 15, 49.<br />
Cardoso, G., Espanha, R. and Lapa, T. (2007). E-Generation: Os Usos de Media pelas Crianças e Jovens em<br />
Portugal. (Retrieved September 27, 2010, from:<br />
http://www.cies.iscte.pt/destaques/documents/CriancaseJovens_AsuarelacaocomasTecnologiaseosMeiosdeCo<br />
municacao.pdf)<br />
Economist debates: Social networking (2008). Economist.com. (Retrieved September 25, 2010, from<br />
http://www.economist.com/debate/overview/123).<br />
Facer, K., Faux, F., and McFarlane, A. (2005). Challenges and opportunities: Making mobile learning a reality in<br />
schools. Proceedings of mLearn 2005. (Retrieved November 5, 2010, from<br />
http://www.mlearn.org.za/CD/papers/Facer%20-%20Faux%20-%20McFarlane.pdf)<br />
Frand, J. (2000). The information age mindset: Changes in students and implications for higher education.<br />
EDUCAUSE Review, 35(5), pp.15-24.<br />
Green, N., Harper, R., Murtah, G., and Cooper, G. (2001). Configuring the Mobile User: Sociological and Industry<br />
Views. In Personal and Ubiquitous Computing (pp. 146-156). London: Springer.<br />
Guy, R. (2009). The Evolution of Mobile Teaching and Learning. Informing Science Press, California, pp.57- 58.<br />
Guy, R. (2010). Mobile Learning: Pilot Projects and Initiatives. Informing Science Press. California, pp.1-7.<br />
Harte, L., Levine, R. and Kikta, R. (2002). 3G Wireless Demystified. McGraw-Hill. NY.<br />
339
Mário Carrilho Negas and Paulo Ramos<br />
Junior, J. and Coutinho, C. (2008). The use of Mobile Technologies in Higher Education in Portugal: an exploratory<br />
survey. (Retrieved November 30, 2010<br />
(http://repositorium.sdum.uminho.pt/bitstream/1822/8468/1/Elearn_Tec_Moveis.pdf)<br />
Katz, J. and Aakhus, M. (2002). Perpetual Contact: Mobile Communication, Private Talk, Public Performance.<br />
Cambridge: Cambridge University Press.<br />
Kukulska-Hulme, A. (2005). Introduction. In Kukulska-Hulme, A. and Traxler, J. (eds).. Mobile Learning: A Handbook<br />
for Educators and Trainers. London: Routledge, pp.1-6.<br />
Leal, D. and Amaral, L. (2004). Do ensino em sala ao eLearning. Guimarães: Universidade do Minho.<br />
Lenhart, A. and Madden, M. (2007). Social Networking Websites and Teens: An Overview. Pew Internet and<br />
American Life Project. (Retrieved November 10, 2010.<br />
http://www.pewinternet.org/PPF/r/198/report_display.asp)<br />
Levinson, P. (2004). Cellphone: The story of the world's most mobile medium and how it has transformed everything.<br />
Nova Iorque: Palgrave Mcmillan.<br />
McLuhan, H. (1964). Understanding Media: The Extensions of Man. NY, McGraw-Hill.<br />
McMahon, M. and Pospisil, R. (2005). Laptops for a digital lifestyle: Millennial students and wireless mobile<br />
technologies. Proceedings of ASCILITE 2005. (Retrieved November 10, 2010, from:<br />
http://www.ascilite.org.au/conferences/brisbane05/blogs/proceedings/ 49_McMahon%20&%20Pospisil.pdf)<br />
Oblinger, G. (2003). Boomers and gen-Xers, millennials: Understanding the “new students”. EDUCAUSE Review,<br />
38(4), pp.37-47.<br />
Pachler, N., Bachmair, B. and Cook, J. (2010). Mobile Learning: Structures, Agency, Practices. New York: Springer.<br />
Palfrey, J., and Grasser, U., (2008). Born Digital: Understanding the First Generation of Digital Natives. NY, Basic<br />
Books.<br />
Prensky, M. (2001). Digital Natives, Digital Immigrants. On the Horizon, 9(5), pp.1-5.<br />
Rickard, W. and Oblinger, D. (2003). Higher education leaders symposium: The next-generation student. Redmond,<br />
WA. (Retrieved November 10, 2010, from http://www.educause.edu/ir/library/pdf/NLI0425a.pdf).<br />
TecMinho – mLearning –The role of mobile learning in <strong>European</strong> Education. (Retrieved November 10, 2010, from:<br />
http://www.tecminho.uminho.pt/showPage.php?url=el_projs_mLearning.html&static&zid=6).<br />
Thomas, S. (2005). Pervasive, persuasive eLearning: Modeling the pervasive learning space. Proceedings of the 3rd<br />
international conference on pervasive computing and communications workshops (PERCOMW’05). Kauai<br />
Island, Hawai’i: IEEE Computer Society), pp. 332-336<br />
Thomas, S. (2005). Pervasive, persuasive eLearning: Modeling the pervasive learning space. Proceedings of the 3rd<br />
international conference on pervasive computing and communications workshops (PERCOMW’05). Kauai<br />
Island, Hawai’i: IEEE Computer Society, pp.332-336.<br />
Tori, R. (2010). Educação sem distância: As tecnologias interactivas na redução de distâncias em ensino<br />
aprendizagem. Editora Senac - São Paulo, pp.218.<br />
Valkenburg, P., Peter, J. and Schouten, A. (2006). Friend networking sites and their relationship to adolescents’ wellbeing<br />
and social self-esteem, CyberPsychology and Behavior, 9, pp.585-590.<br />
340
A Framework for Information Systems Innovation: A Case of<br />
Competitive Intelligence in Organisations<br />
Phathutshedzo Nemutanzhela and Tiko Iyamu<br />
Tshwane University of Technology, South Africa<br />
phathuts@gmail.com<br />
iyamut@tut.ac.za<br />
Abstract: Knowledge is used as a focal factor for competitive advantage, through effective and efficient<br />
performances by employees in many organisations. As a result, knowledgeable employees are expected to share<br />
their knowledge with others to increase innovation within the organisation. Unfortunately, this is not always the case.<br />
Generally, employees behave differently within an organisation. No organisation has total control of its employees’<br />
behaviour and actions. The behaviour and action, which has impact on how knowledge is shared is influenced by<br />
many factors such as fear of losing jobs, power to negotiate personal interest. These factors influence the<br />
deployment of CI were found and interpreted and conclusion was made from those interpretations. The study aimed<br />
at establishing the impact of Competitive Intelligence (CI) on Information systems (IS) innovation products and<br />
services in organisation. A case study was conducted, using a financial organisation. Innovation-decision process<br />
was applied in the data analysis.<br />
Keywords: competitive intelligence (CI), diffusion of innovation (DoI), Information systems (IS), innovation<br />
1. Introduction<br />
Companies keep competing against each other using products and services, making competitive<br />
intelligence (CI) an important tool in the development of strategy in the organisations. The importance of<br />
competitive intelligence is attributed to its contribution to technological knowledge and intelligence, and it<br />
use for the analysis of information systems innovation in organisations. It should be pointed out early that<br />
innovation and technology are often taken in a similar light as asserted by Rogers (2003: p12) that ‘we<br />
often use the word “innovation” and “technology” as synonyms.’ And as such CI is primarily intended to<br />
be used for the state of art, technological trends and challenges, with a strategic vision on<br />
competitiveness and customers (Ashton & Klavans, 1997; Fleisher & Bensoussan, 2003). Competitive<br />
intelligence therefore can be considered as a tool for innovation process, observation of market, analysis<br />
of strategic behaviors of both competitors and customers, including their values, expectations and needs<br />
(Krücken-pereira et al., 2001).<br />
An innovation according to Rogers (2003: p12) will be considered as ‘an idea, practice, or object that is<br />
perceived as new by an individual or other unit of adoption.’ A unit of adoption in this case could be the<br />
organisation, a society and also a target market.<br />
Competitive Intelligence offers a real strategic advantage for many businesses (Stephen, 2006). Gilad<br />
(2000) argues that some of the largest corporate organizations have a dedicated CI department, while<br />
smaller businesses often practice CI on an ad hoc basis. This they do so by informally collecting<br />
information from a variety of internal and external sources, such as the Internet, trade shows conferences<br />
and networking meetings. Competitive Intelligence is of importance to many businesses mainly because<br />
it helps to formulate strategy, as well as make informed decisions.<br />
The CI is deployed with the intention to better, coordinate internal processes and activities of<br />
organizations, primarily, to reach market more effectively. Gathering people, the logic and the physical<br />
architecture around common purposes provide individuals with the information they need to expand their<br />
own knowledge (Malhotra, 2000; Hoven, 2001). This approach help to build high performance teams in<br />
the organisation. This indeed, is the foundation of the integrated organisation, where the information<br />
technology is capacitating technological innovation.<br />
2. Research methodology<br />
The main focus of the study was to investigate and understand the impact of competitive intelligence on<br />
information systems innovation products and services in organisations. The approaches and methods<br />
employed in the study include the case study, qualitative research method, and semi-structured interview<br />
approach. The data was analysed, using the Innovation-decision process from the perspective of DoI<br />
theory<br />
341
Phathutshedzo Nemutanzhela and Tiko Iyamu<br />
The case study was adopted primarily because it is an approach that assists to achieve a deep<br />
understanding of a specific phenomenon. According to Cooper and Schindler (2006), the case study is an<br />
approach which combines individual and (sometimes) group interviews with record analysis and<br />
observation; used to understand events and their ramifications and processes. Hofstee (2006) argued<br />
that the case study approach is useful when detailed knowledge is required of any particular case. A<br />
pseudonym name, “Divhetsheleni” was used to represent the case, the organisation used in the study.<br />
The participants were codified from DV_LA001 to DV_LA012 to adhere to the ethical consideration as<br />
agreed with the organisation and the university.<br />
The qualitative research method was selected because of the nature of the study. The qualitative study<br />
allows data to be gathered from multiple sources (Yin, 2009). In other words it is not limited to one source<br />
of information.<br />
The research followed a structured interview schedule so that the process does not lose focus. Wellman<br />
and Kruger (2001: 160) assert that “In a structured interview, the interviewer puts a collection of<br />
questions from a previously compiled questionnaire to a respondent face to face”. This was<br />
advantageous for data collection because it made it possible to explain the questions that were not<br />
understood by the respondent and there was chance to further probe responses.<br />
2.1 Diffusion of Innovation (DoI)<br />
The Innovation-decision process from the perspective of DoI theory was employed in the data analysis. In<br />
DoI theory, technological Innovation is communicated through particular channels, over time, among the<br />
members of a social system (Rogers, 2003). The theory is concerned with the manner in which a new<br />
technological idea, artefact or technique, or a new use of an old one, migrates from creation to use.<br />
According to Rogers (2003), the Innovation-decision process involves five steps. As shown in Figure 1<br />
below, the process include: (1) knowledge, (2) persuasion, (3) decision, (4) implementation, and (5)<br />
confirmation. These stages typically follow each other in a time-ordered manner. The stages are briefly<br />
described below.<br />
Figure 1: Diffusion of Innovations (Rogers, 2003)<br />
342
Phathutshedzo Nemutanzhela and Tiko Iyamu<br />
The Innovation decision process characterized as a process that occurs while individuals participate in a<br />
series of actions related to decisions (Rogers, 1995). Knowledge occurs when individuals are aware of<br />
the Innovation and gain understanding of its functions. Persuasion is when individuals or decision-making<br />
units exhibit favourable or unfavourable behaviour toward the Innovation. Decision indicates when the<br />
individual or unit decides to adopt or reject the Innovation. Implementation occurs when the individual or<br />
unit decides to use the Innovation. Confirmation occurs when decision makers confirm or reject their<br />
decision to adopt the Innovation (Rogers, 1995).<br />
3. Data analysis<br />
Using the Innovation-Decision Process (Rogers, 2003), the analysis of the data is presented as follows:<br />
3.1 Knowledge<br />
Knowledge involves management efforts - from identifying needs to delivering intelligence products and<br />
services to consumers and clients. Similarly, innovations are evaluated, monitored and controlled using<br />
knowledge. In the organisation, knowledge was applied in the development of innovation, which was<br />
supported by policy, and generated new requirements. Through knowledge the organisation could<br />
innovate, as well as plan how to utilise the products and services from innovation. According to<br />
DV_LA003 (p4:30-31), Knowledge in the organisation is assessed by questions, and, they are<br />
categorised into needs, such as intelligence problem, decisions, risk metrics (risk tolerance level) and<br />
measurement methods. The organisation was challenged by the knowledge to innovate, primarily, for<br />
competitive advantage.<br />
The organisation focuses on the innovations of security tools and approaches. DV_LA004 (p13:100-101)<br />
explained that there is a growing use of Competitive Intelligence in the financial services sector and<br />
mentioned that the organisation has spent considerable resources to protect their businesses from<br />
infiltration of their systems by other forces.<br />
According to DV_LA003 (p7:45-46), Knowledge is a vital tool, as such, and the organisation conducts<br />
surveys to assess employees' knowledge on the deployment of CI. The organisation worked hard on<br />
communicating with staff, ensuring strategies were clear, addressing performance and remuneration<br />
issues and motivating people. A number of employees, within the organisation, did not have the<br />
knowledge regarding some of their Information Systems innovation Product and Service offerings. This<br />
was attributed to factors such as inadequate skills; no clear consequences of not meeting performance<br />
standards; lack of alignment with strategy and a reward system that fails to motivate properly.<br />
If employees are not informed, or are knowledgeable about products and services it creates risk to both<br />
individuals and the organisation. DV_L003 (p7:62-63) explained that everyone in the organisation are<br />
aware of the strategy and possesses knowledge of the products and services which are deployed in the<br />
organisation.<br />
3.2 Persuasion<br />
At the Persuasion stage, a person’s main type of thinking is affected by, or related to, feelings. Some<br />
employees in the organisation develop an attitude towards certain Innovation and are psychologically<br />
involved with Information Systems Products and Services. According to DV_LA002 (p5:55-56), there is a<br />
marketing team which responsible innovation, and persuades the Board of Directors for approval. Post<br />
approval, the rest of the employees is persuaded on implementation and use of the innovation.<br />
Organisations use incentives and bonuses as tools to persuade employees to perform their tasks, and<br />
align with the organisational innovative strategy. An employee stated that when I am working on any<br />
project, I know that I should do my ultimate best to make sure the project becomes a success. The lack of<br />
some knowledge, on my part, in being able to deliver successfully on the project, affects my incentives<br />
and can, sometimes, also affect everyone involved on the project. DV_LA006 (p16:172-173).<br />
Some employees felt that only using the marketing team is exploitative because of the rest of the team’s<br />
limited perception of the innovation, and reality. Not having a say in the matter may serve as an indicator<br />
of the message’s influence on this stage. However, DV_LA001 (p2:14-15) believes that, by having one<br />
team focusing on finding new ideas to innovate, it eliminates time waste and also cuts cost; inevitably<br />
speeding up the delivery process. DV_LA004 (p8:83-84), commented that there is a perception that<br />
people who work in the field might have first information of latest innovation, but often choose to remain<br />
343
Phathutshedzo Nemutanzhela and Tiko Iyamu<br />
silent because it’s not their place to come up with new ideas. The employee added that it then becomes<br />
more difficult or easier to persuade these people to adopt innovation in the organisation, as they already<br />
know more about it. This depended on the type of findings they have about the innovation, as a result of<br />
their interaction with clients.<br />
Another interviewee, DV_LA005 (p9:148-149) commented that Key Performance Indicators (KPIs) is<br />
good approach to periodically assess the performance of the individuals, business units and the<br />
organisation at large. The employees further explained that in the organisation, KPIs is defined in a way<br />
that is understandable, meaningful, and measurable. They are rarely defined in such a way such that<br />
their fulfilment would be hampered by factors seen as non-controllable by the organisations or individuals<br />
responsible. The KPI ensure that each employee knows the organisation’s expectations, as well as the<br />
output required by the customers.<br />
3.3 Decision<br />
Many of the individuals in an organisation do not adopt innovated Information Systems Products and<br />
Services without proper testing and evaluation. The latter was carried out in relation to the usefulness<br />
and fit in the organisation’s environment. Some individuals used partial trial to gain better understanding<br />
of innovation and deployment of products and services in the organisations.<br />
One of the interviewees, DV_LA008 (p6:17-18) provided the following explanation; some employees in<br />
the organisation are policy decision-makers and initiate requirements for CI products and services. These<br />
decision-makers are the recipients of the end Products and Services of CI through Information Systems<br />
Innovation. The decisions are based on information and knowledge, and sometimes lead to the levying of<br />
more requirements; thus triggering the Intelligence Cycle. After finishing a cycle, a new set, or improved<br />
set, of template will be produced that will be used in monitoring and identifying risk.<br />
According to a Director, DV_LA002 (p5:43-44), It is common to find deployments filled with projects that<br />
have attractive returns on investment but do not move the needle on performance parameters that matter<br />
in the marketplace. In selecting Products and Services to innovate and build a portfolio of improvement<br />
initiatives, clear linkage to strategic priorities needs to be established; not just at the outset, but on an ongoing<br />
basis. When making decisions, one of the most seductive pitfalls is to become comfortable with<br />
embracing innovations that are, in themselves, attractive without paying sufficient attention to overall<br />
optimization.<br />
Although finance could be viewed as the artery of a firm and an important indicator of management<br />
direction, management mindset encompasses more variables than only financial decisions. The<br />
organisation’s brand, strategy, employees and customers play a major role when making a decision to<br />
innovate. The brand has to define the organisation through the innovation. The innovation has to align<br />
with strategy, the employees’ need for knowledge and skills to deploy the CI innovation, and the<br />
customer has to get the best Product and Services that will help solve their problems. DV_LA001 (p3:19-<br />
20).<br />
3.4 Implementation<br />
Each division at Divhetsheleni has its own strategic team that deals with implementing the strategy of the<br />
division, making sure it aligns with the organisational strategy. All CI implementation is defined by the<br />
same vision and goals in order to improve innovation, as well as to create a competitive brand. The<br />
strategy team ensures that Information Systems Innovation aligns with the organisation’s strategy.<br />
However, the organisation had a Marketing Department, which was responsible for innovations that<br />
needed to be implemented in all the various divisions of the organisation. According to one of the<br />
employees, (DV_LA001 p2:18-19), Implementation of a new innovation was about being prepared to<br />
measure the organisation's performance consistently, constantly, to recognise weaknesses and, indeed,<br />
be willing to address those weaknesses. This begins with initial measurement, which serves as an<br />
indication of where the organisation is, so that it can determine where it wants to be. The organisation still<br />
follows old traditional ways of implementation, while Information Systems Innovation challenges the<br />
employees to acquire more knowledge, and understanding, of the products and services they deploy.<br />
The organisation readiness to implement becomes a challenge during the innovative process.<br />
A forward–looking organisation seeks to provide value-added offerings, through Information Systems<br />
Innovation, at every stage of its life in order to improve the culture of the organisation. The organisation’s<br />
344
Phathutshedzo Nemutanzhela and Tiko Iyamu<br />
high-performance culture was characterised by openness and trust, with clear accountability for<br />
execution, and the freedom to take calculated risks; thereby constantly raising individual and<br />
organisational performance. This was expressed by DV_LA005 (p15:204-205).<br />
According to DV_LA002 (p6:64-65), most of the organisation’s successful implementations of a<br />
remuneration strategy, and the alignment of benefits across all staff, are based on job function rather<br />
than grades.<br />
The employee further expressed that, Performance scorecards reflect an integrated approach to<br />
sustainability and are balanced across economic, environmental, social, transformational and cultural<br />
elements. Our commitment to driving high performance is evident in our new organisational structure,<br />
with the separation of support structures from the client-facing clusters to ensure that the business is able<br />
to focus on improving economic performance regardless of economic conditions.<br />
3.5 Confirmation<br />
Confirmation was the recognition, by the organisation, of the benefits of using Information Systems<br />
Innovation; its integration into the organisation’s on-going routine, or promotion of the innovation to<br />
others; as well as the identification of Products and Services to be deployed. This gives the organisation<br />
an opportunity to provide feedback immediately when decisions for tomorrow’s solutions are being made.<br />
According to one of the employees, we establish a team work with our clients from the beginning of the<br />
deployment, DV_LA006 (p18:248-249). This team work between the organisation and the customers<br />
made it possible to test and confirm the proposed innovation.<br />
In addition, another employer, DV_LA001 (p3:29-30) commented that in cases whereby many solutions<br />
required testing, the customer gave the organisation access to people who were experts on the<br />
applications. The collaboration between hardware and software providers, at this early stage, eliminated<br />
the risk normally associated with new technology deployment.<br />
4. Findings<br />
There were some findings from the above analysis of the data. The findings are discussed as follows:<br />
Culture<br />
When an organisation is operating globally, culture always becomes a challenge. The organisation’s<br />
high-performance culture is characterised by openness and trust, with clear accountability for execution,<br />
and the freedom to take calculated risks; thereby constantly raising individual and organisational<br />
performance. For the organisation to deliver on its high performance culture, it must have created an<br />
environment where information is shared openly, whereby its’ people are reward for their skills. In<br />
addition, trust must be created to ensure that employees do not feel used.<br />
Each innovation has performance scorecards that reflect an integrated approach to sustainability. These<br />
are balanced across economic, environmental, social, transformational and cultural elements. In addition,<br />
these innovation performance scorecards assist in building a new organisational structure; separating<br />
support structures from client-facing clusters.<br />
Knowledge Sharing<br />
One of the challenges of an organisation is that of getting people to share their knowledge. In today's<br />
organisation, where so much depends on teamwork and collective knowledge, it is only a handful of<br />
people who have the kind of knowledge with which they can hold their peers (including bosses) to<br />
ransom. Such individuals were Directors or Managers, who were careful and caution of losing trade<br />
secrets. It may be a particular Specialist who has been in the organisation for many years and has built<br />
up his, or her, own unique way of achieving success without perhaps even understanding the deep tacit<br />
knowledge of how they do it. All these poses threat to organisation, when they want to innovate, because<br />
such individuals holds on to the expertise that they have.<br />
Organisation is intended to create a commitment to culture, change, challenge, competition and<br />
cooperation. If, as is often the case, time pressure leads to poor knowledge sharing, then there must be a<br />
345
Phathutshedzo Nemutanzhela and Tiko Iyamu<br />
commitment to allow time for it to happen. Commitment to knowledge sharing must be demonstrated<br />
throughout the organisation.<br />
Repositioning of Brand<br />
The organisation shifts its focus from repositioning its brand to building the brand and the brand promise.<br />
When an organisation innovates, the brand needs to sell the idea. Therefore, the customer has to start<br />
seeing the innovation from brand perspective.<br />
The organisation continues to inspire, motivate and challenge people to make a difference, whilst striving<br />
to become a leader in client-employee understanding and care; across all market segments. The<br />
introduction of a new brand expression, accentuates the intensified, and increased, nationwide above<br />
and below-the-line advertising campaigns to position the organisation’s brand in a more relevant and<br />
approachable manner.<br />
Education and Awareness<br />
Education and Awareness in the organisation encourages and enhances people's participation in<br />
activities aimed at conservation, protection and management of the environment; essential for achieving<br />
sustainable development. The organisation seems to ignore the importance of educating their staff about<br />
Information Systems Product and Services and the path taken by the organisation to deploy Competitive<br />
Intelligence. There appears to be the assumption that seems that all employees are aware of the<br />
deployment process. Employees who were not aware of the innovations which were taking place in the<br />
organisation, could be considered and classified more as expense, than an asset. Their inability to<br />
transfer, or acquire skills, from one other is evident since they are not cognisant of the innovation taking<br />
place within their organisation. Spending the necessary time to educate employees, and making them<br />
aware of CI, will cut organisation costs and reduce time spent doing feasibility studies on the process of<br />
adopting and implementing innovation by an outside partner.<br />
Improve Innovation<br />
The organisation improves innovation across all business clusters in order to create a single view of its<br />
client base. It enables a closer relationship with clients, and it facilitates better co-ordination all actions.<br />
The organisation strives to provide more transactional service offerings, and the driving of primary<br />
investor status, by focusing on further enhancements.<br />
Some divisions, however, were planning on introducing a number of new products, services and business<br />
ventures. A dedicated unit, focused on driving innovative products, has been established. The<br />
organisation will be building on the innovative products that they launch every year; aiming to have<br />
improved its innovation delivery capability significantly.<br />
Inclusiveness<br />
Some of the innovation for process improvement lies with the process owners. This is not a new concept,<br />
but it bears emphasising; as organisations are still not taking this aspect seriously when innovating.<br />
Frequently, process owners are frontline people who do not have the chance to participate in<br />
improvement efforts. While they may not be formally trained in quality tools, their closeness to the<br />
process is a vantage point second to none. The organisation uses their marketing team to come up with<br />
innovation. The same team runs the marketing campaigns and interacts with the customer, forgetting the<br />
very people who actually deploy the process.<br />
Prioritization<br />
One of the most seductive pitfalls is for an organisation to become comfortable with doing Information<br />
Systems innovation projects that are in themselves attractive, without paying sufficient attention to overall<br />
optimisation. It is common to find deployments filled with Information Systems innovation projects that<br />
have attractive returns on investment, but do not move the needle on performance parameters that<br />
matter in the marketplace. In selecting Information Systems innovation projects and building a portfolio of<br />
improvement initiatives, clear linkage to strategic priorities needs to be established, not just at the outset<br />
but on an on-going basis.<br />
346
Phathutshedzo Nemutanzhela and Tiko Iyamu<br />
Creativity is not generally associated with defining the portfolio of Information Systems innovation<br />
projects; however, ongoing re-evaluation can uncover opportunities for innovation. The sum, or where<br />
time and attention is placed, defines an organisation's strategic direction.<br />
The next section presents the interpretation of the findings from the case study.<br />
4.1 Interpretation of the findings<br />
The findings from the analysis are regarded as the factors which influence the Innovation of Competitive<br />
Intelligence (CI) within the Information Systems environment. The Products and Services of CI manifest<br />
into many other components, as depicted in Figure 2 below. The discussion that follows explains each of<br />
the components.<br />
Figure 2: Factors influencing competitive intelligence on information systems innovation<br />
Accessibility – This refers to the organisation’s ability to access information needed for the making of<br />
public decisions as well as its ability to share the economic status of the organisation with its<br />
employees. It also speaks to the sharing of knowledge and expertise by employees, with their peers,<br />
without hiding any information for fear of any threat that might arise.<br />
Power - This is often a cop out by managers, or change agents, who are not adequately addressing<br />
the human factors or motivational aspects. In today's organisation, where so much depends on<br />
teamwork and collective knowledge, it is only a handful of people who have knowledge for which they<br />
can hold their peers (and bosses) to ransom. It might be the owner or manager of a small<br />
organisation not wanting to lose trade secrets; it may be a particular specialist who has been in the<br />
organisation, many years, and built up his or her own unique way of achieving success without<br />
properly understanding the deep tacit knowledge of how they do it.<br />
Enrolment - An individual may have knowledge used in one situation but be unaware that other<br />
people, at other times and places, might face similar situations. Additionally, knowledge derived for<br />
one purpose may be helpful in totally different contexts; or it may be a triggered for innovation. Many<br />
innovative developments come from making knowledge connections across different disciplines and<br />
organisational boundaries.<br />
Management - There is pressure on productivity, on deadlines, and it's a general rule that the more<br />
knowledgeable you are, the more there are people waiting to collar you for the next task.<br />
Management ensures that lessons learnt by individuals and groups are captured into the knowledge<br />
database for sharing. .<br />
Competitive Advantage – This is the strategic advantage one organisation entity has over its rival<br />
entities within its competitive industry. Achieving Competitive Advantage strengthens and positions<br />
an organisation better within the business environment.<br />
In order for an organisation to develop a competitive advantage it must have resources and capabilities<br />
that are superior to those of its competitors. Without this superiority the competitors can simply replicate<br />
what the organisation was doing, and any advantage disappears quickly.<br />
Framework for Information Systems Innovation<br />
From the analysis and findings, the case study provides insight to the different characteristics of<br />
Competitive Intelligence (CI) Products and Services Innovations within the Information Systems<br />
347
Phathutshedzo Nemutanzhela and Tiko Iyamu<br />
environment. This is illustrated in Figure 3. To have a full understanding, the discussion that follows<br />
should be read with the figure 3.<br />
Figure 3: Competitive intelligence products and services innovations within the information systems<br />
environment<br />
Information Systems innovation begins with the organisational environment, whereby there is top-down<br />
support for project management practices and investment, by the organisation, in the innovation process.<br />
In rolling out Products and Services, implementation by the organisation focuses on doing so correctly,<br />
rather than simply doing what it takes to complete the CI deployment as quickly as possible, at a<br />
minimum cost.<br />
Organisations that are strong in Adaptability and Involvement have an edge in innovation and creativity,<br />
while organisations excelling in Mission and Consistency have a high measure of stability, return on<br />
investment and return on sales. Organisations measuring high in all components have a dramatic<br />
financial advantage over organisations that are weak in these areas. Organisations at the bottom perform<br />
just as one would expect: They are sluggish, wasteful and out of touch with their customers.<br />
There is reliance, by the organisation, on the Key Performance Indicators (KPI’s) to identify or measure<br />
the acceptance of innovation by a client. KPI, for an organisation, is a tool that persuades decision<br />
makers as a result of the reaction of the customer towards the organisation’s innovation. It also measures<br />
an employee’s performance on certain tasks related to the innovation.<br />
If successful implementation of Products and Services contributes directly to strategic objectives and the<br />
bottom line, Information Systems Innovation should lead to organisational success. Innovation requires<br />
an investment in resources, therefore responsibility lies with senior management to ensure that<br />
investments are made wisely and add value to the organisation as a whole. Unsuccessful CI<br />
deployments translate into poor investment and, ultimately, senior management is accountable for<br />
projects not reaching the objectives of the organisation.<br />
In conclusion, the brand of an organisation brand is a key intangible attribute by which they compete. Its<br />
main objective is to support differentiators; and reposition the company as being significantly different<br />
from any other financial organisation. In so doing, potential clients are drawn to the brand which,<br />
ultimately, encourages them to choose their organisation above that of their competitors. The role of<br />
competitive intelligence on information systems innovation is widely revealed. It was better understood<br />
that while competitive intelligence is overemphasized as revolutionary customers focused, information<br />
systems products and services still remain challenging. It was also understood that not all organisations<br />
that deploy competitive intelligence are responsible for the innovation. Lack of knowledge sharing and the<br />
organisational culture were found to be important factors for the deployment of competitive intelligence<br />
products and services in the organisations.<br />
From the findings and interpretation of the case study on the impact of Competitive Intelligence on<br />
Information Systems products and services innovations in organisations, a Framework was developed,<br />
Figure 4. The Framework is aimed at providing direction for organisations during Information Systems<br />
348
Phathutshedzo Nemutanzhela and Tiko Iyamu<br />
innovation, the stages on the Framework encompasses the activities involved in Information Systems<br />
innovation.<br />
Figure 4: Framework for information systems innovation<br />
The Information Systems it’s composed of five different stages that organisations need to follow for<br />
deployment of CI products and services. Each Stage has columns that are categories by different<br />
aspects of innovation on a row level. All innovation stages or phases actives, roles and responsibilities,<br />
requirements and information systems inputs for the deployment of a successful competitive intelligence<br />
products and services has been illustrated in this framework.<br />
5. Conclusion<br />
Knowledge sharing is critical to Information System Innovation in the organisation that deploys it, making<br />
it very useful for competitive advantage. The study helps managers to gain better understanding of how<br />
knowledge sharing influences products and services in the organisations.<br />
Sometimes the effect of Competitive Intelligence on the innovative process is not obvious, but it does<br />
exist because the companies compete, and are challenged with customers’ needs. Therefore it is<br />
strategically important to equip as many as possible employees with enough knowledge to carry out the<br />
customers’ demands. Regardless of how knowledge is acquired, Competitive Intelligence deployment<br />
relies on the knowledge individuals and group have about the information Systems products and<br />
services.<br />
References<br />
Ashton, W. B. Klavans, R. A. (1997). Keeping abreast of science and technology: technical intelligence for business.<br />
United States: Battelle Press.<br />
Cooper, D. & Schindler, P. (2006). Business Research Methods, McGraw-Hill, New York.<br />
Den Hoven, J. Van (2001). Information Resource Management: Foundation for knowledge management.<br />
http://www.brint.com/members/01060524/ irmkm/irmkm_1.html.<br />
Fleisher, C.S. (2003). "Competitive Intelligence Education: Competencies, Sources and Trends," Information<br />
Management Journal, March/April, 56-62.<br />
Gilad, B. (2000). "An Ad Hoc, Entrepreneurial CI Model," Competitive Intelligence Magazine, vol. 3, no. 4, pp. 33-39.<br />
Gregor, S. & Jones, K. (l999). Enabling Electronic Commerce in Agribusiness: The Initial Stages of a Diffusion<br />
Process (where - Collecter or Bled?).<br />
Grover, V. (1997). An extension of the Tri-core model of information systems innovation: Strategic and technical<br />
moderators. <strong>European</strong> Journal of Information Systems, 6(4), 232-242.<br />
349
Phathutshedzo Nemutanzhela and Tiko Iyamu<br />
Haag Stephen, (2006). Management Information Systems for the Information Age. Third Edition. McGraw-Hill<br />
Ryerson.<br />
Hofstee, E. (2006). Constructing a Good Dissertation: A Practical Guide to Finishing a Master's, MBA or PhD on<br />
Schedule, EPE Publishers, South Africa.<br />
Hovav, A., Patnayakuni, R. & Schuff, D. (2001). Internet Technology Diffusion: Adoption ofIPV6 Global Cooperation<br />
in the New Millennium, 9th ECIS, Bled, Slovenia.<br />
Krücken-Pereira, L.; Debiasi, F.; Abreu, A. F., 3, jul (2001).Technological innovation and competitive intelligence: an<br />
interactive process. READ - Electronic Journal of Directors, v.7, Available at:<br />
. Accessed 19 September 2003.<br />
Malhotra, Y. (2000). Knowledge Management & New Organization Forms: A framework for Business Model<br />
Innovation. Information Resources Management Journal. Jan-Mar., vol.13, no.1, pp.5 – 14.<br />
Rogers, E. M., (1995). Diffusion of Innovations. 4 th ed. New York: Free Press.<br />
Rogers, E.M., (2003). Diffusion of innovations (5th ed.). New York: Free Press.<br />
Welman J. C. & Kruger S.J, (2001). Research Methodology: for the Business and Administrative sciences, 2 nd Ed.<br />
Oxford University Press, South Africa.<br />
350
RFID: A Review of its Relevance and Application in South<br />
African Retailing Systems<br />
Jonathan Oni, Edward Dakora and Vesper Owei<br />
Cape Peninsula University of Technology, Cape Town, South Africa<br />
jo45ng@yahoo.com<br />
naadakora@gmail.com<br />
OweiV@cput.ac.za<br />
Abstract: The advent of Radio Frequency Identification (RFID) has enabled identifying an object using a radio<br />
signal. RFID has become the talk of the day when world retail giants including Wal-Mart, Tesco, Target, and<br />
Albertsons announced their implementation of the technology to ensure efficient supply chain management, and,<br />
naturally, demanded that their suppliers meet RFID requirements. Today, these retailers boast some of the best<br />
supply chain systems and management in the industry worldwide. The increase in adoption and application of RFID,<br />
therefore, seems inevitable. It has been envisaged that the near future will see almost all new products sold to<br />
consumers having an RFID tag that can allow them to be remotely and uniquely identified and tracked. The potential<br />
impact of RFID on the South African retail industry is significant. This technology is mostly employed in supply chain<br />
management and customer relationship management. In supply chain management, RFID can benefit retailers in<br />
many ways apart from the mere identification of products and suppliers, such as monitoring the production process<br />
and the sequence of activities from the material production to delivery of the final products. Preliminary studies reveal<br />
that while South African retailers are aware of RFID and are interested in its adoption, few are actively engaged in<br />
pilot tests or planned implementations. Although, they are aware of its benefits, the relevance and its application in<br />
the South African retail sector is still limited. This paper reviews the relevance of RFID and the application in South<br />
African retail system.<br />
Keywords: RFID, RFID relevance, RFID application, South Africa, retail system<br />
1. Introduction<br />
RFID have led to benefits such as cost reduction, low shrinkage, and increase in value and quality of<br />
products. Fitzek (2003) noted that the benefits of RFID may manifest throughout the supply chain, the<br />
primary benefits are evident at the retail end, including marketing and customer service provision.<br />
Although, the South African retail system is experiencing a slow pace in adoption and application, the<br />
benefits of RFID is enormous.<br />
This paper reviews the relevance of RFID and the application in South African retail system. The paper<br />
further explores in greater depth the factors that influence the application of RFID in the South African<br />
retail system. The findings and their implications will be discussed, and the limitations and ideas for future<br />
research will be outlined. The paper concludes with a discussion on the application of RFID and its<br />
relevance to the South Africa retail system. The rest of the paper is structured as follows: The next<br />
section is a literature review of RFID global perspective and adoption in South Africa. This is followed by<br />
the research method in section 3. Section 4 discusses the analysis and findings of the research method.<br />
The discussions and limitations of the research are presented in section 5. The paper concludes in<br />
section 6 with a summary of the paper and some positive thoughts on the way forward for a successful<br />
application of RFID in the retailing sector of South Africa.<br />
2. Literature review<br />
2.1 ICT in emerging economies<br />
The ICT economy report hailed this technology as what fuelled the “strong wave of innovation that<br />
transformed the global economy during the last quarter of the 20 th century” (UNCTAD, 2007). The<br />
Internet and World Wide Web as part of ICTs are being used by businesses of all kind around the world<br />
in the management of their day-to-day operations and transactions, and, therefore, forms an integral part<br />
of their marketing strategies (Sagi, Carayannis, Dasgupta & Thomas, 2004). The growth of the ICT sector<br />
has transformed many economies around the world, making the drivers of growth to be more information<br />
and knowledge based rather than the traditional natural resource dependency (ITU, 2007). However, the<br />
said report also acknowledges the disadvantages of developing countries including limited infrastructure,<br />
human capital, policy incentives, investments and high cost of services which hamper their growth in<br />
ICTs.<br />
351
Jonathan Oni et al.<br />
Yet, it has been argued that the arrival of ICT revolution in Africa represents one aspect of globalisation<br />
on the continent (Carmody, 2009). Among ICTs, mobile phones have been the most accepted and<br />
adopted on the continent, with growth and penetration rates reaching alarming proportions, although this<br />
rate may vary from country to country. The high penetration rate of mobile technology and mobile phones<br />
in African countries represent a significant contribution to economic development, and poverty reduction.<br />
The global IT report (Geiger & Mia, 2009) affirms the positive impact of Mobile phones penetration on<br />
economic growth, and its potential to reduce poverty and bridge the digital divide. Moreover, the<br />
completion of the SEACOM’s fibre optic cable system, connecting Africa to the rest of the world and<br />
providing faster internet connectivity is a major boost to the development of ICTs on the continent.<br />
However, there is still a limitation on individual ownership of ICTs equipment and Internet connectivity in<br />
developing economies, and, therefore, ICTs supported business model that are proven to be successful<br />
the developed world may not be readily application in the developing world situation (Roztocki &<br />
Weistroffer, 2009). ICT based innovations that are relevant to the developing market situation might be<br />
necessary. For example, in the case of RFID technology and its application retailing, the passive RFID as<br />
will be discussed later, might be more user friendly to retailers in developing country markets like South<br />
Africa. In South Africa, retailers are seeking more efficiency in the mist of the economic downturn by<br />
implementing new ICTs including Web-based portals, messaging platforms and wireless technologies in<br />
their operations (Dos Santos & Baccialoni, 2009).<br />
2.2 Global adoption of RFID<br />
With the hype of RFID, Chen (2004); Riggins and Slaughter (2006) suggest that the market is predicted<br />
to grow by round about 20% each year and will reach 2-3 billion US dollar in 2008.Though, after many<br />
years of hyping the technology, the actual adoption lags behind the optimistic expectations. According to<br />
Matta and Moberg (2006), besides all attention which has been paid to the RFID technology, the pace of<br />
actual or planned RFID adoptions by companies in their supply chains remain low to moderate. Various<br />
and recent studies have reported adoption rates only between 7-15% (Brown & Bakhru, 2007);<br />
(Leimeister & Knebel, 2007); (Matta & Moberg, 2006).<br />
It has been envisaged that the near future will see almost new products sold to consumers having an<br />
RFID tag that can allow them to be remotely and uniquely identified and tracked (Cazier, Jensen & Dave,<br />
2008). RFID tags are said in Cazier’s et al. work to bear products information including but not limited to<br />
product type, price, place of origin, and place of purchase. This information can be read using any<br />
appropriate scanner. Two types of RFID tags exist according to Angeles, (2005); Cazier et al.:<br />
Active tags rely on a power source (battery) to function, and this power source makes it possible for<br />
the tags to broadcasts “their own signal over varying distances, depending on the potency of their<br />
power source and range of frequency.” Nevertheless, active tags are said to be very expensive and<br />
are therefore, not popular in business environments; their applications are more prominent in military<br />
and other such environments.<br />
Passive tags use no battery, “sit idle until passed near a reader that emits radio waves” (Haag, 2007)<br />
and are economical to produce. They are not capable of broadcasting their signal due to their lack of<br />
power source; they are common with consumer goods.<br />
Figure 1: A typical RFID system (Robert, 2006)<br />
Since the biggest benefits of RFID can be seen in the supply chain sector (Curtin et al., 2006), it is logical<br />
to study its adoption from an organizational perspective as opposed to individual adoption. Applying<br />
various database, Schmitt and Michahelles (2009) presented a result of studies conducted on the<br />
adoption of RFID. The reviews were adapted and summarized in Figure 2. The first column indicates the<br />
352
Jonathan Oni et al.<br />
author(s) followed by the research topic. The third column shows the used research methodology while<br />
the fourth column summarizes the results of the study especially the identified adoption factors.<br />
Figure 2: RFID adoption studies adapted from Schmitt and Michahelles (2009)<br />
2.3 The South Africa retailing system<br />
South Africa appears to have the oldest formal retailing system in Africa, with an early experience of the<br />
supermarket revolution in the developing world, the first wave in the 1990s. This system has advanced<br />
353
Jonathan Oni et al.<br />
over the years, with a lot of takeovers, mergers and acquisitions, conglomerations and expansions. To<br />
date, the country has retail companies grown into chains and groups, and become regional giants. For<br />
example, Shoprite has grown to become the largest food retailer in the continent over the past decade,<br />
due to its Africa-wide expansion drive.<br />
Moreover, the South African retail system also meets international standards, and, in a few instances,<br />
South African retailers have won internal awards for innovation. Woolworths (SA) has won twice in a roll,<br />
the coveted award as the worlds “Responsible Retailer of the Year” at the World Retail Congress 2010;<br />
beating Cargills Food City, Kingfisher, New Look, 0 Boncario and Teknosa to it this year (The Word Retail<br />
Congress, 2010). This is due to Woolworths “Good Business Journey” initiative, for sustainable farming<br />
and organic food production, locally. At the same event, Truworths International (SA) won the Emerging<br />
Market Retailer of the Year. All these retailers have employed numerous technologies and innovative<br />
strategies from, customer relationship management to supply chain management. Most of the clothing<br />
retailers like Truworths and the Foschini have credit account facilities allowing customers to shop now<br />
and pay later, in some cases it take just an SMS to open such an account. There are substantial use of<br />
scanning systems and devices at the stores and distribution centres some of which would be RFID.<br />
However, it appears the use adoption and use of RFID scanning systems for tracking products through<br />
the supply chain to stores is limited in the South African retail system. The annual reports of the major<br />
retailers make mention of the use of sophisticated computers systems and supply chain management<br />
systems, especially in the case of the supermarket chains, but they are silent on the specific use of RFID.<br />
2.4 Relevance of RFID to South African retailing<br />
The potential impact of RFID on the retail industry is significant. This technology is mostly employed in<br />
supply chain management and customer relationship management. In supply chain management, RFID<br />
can benefit retailers in many ways apart from the mere identification products and suppliers, such as<br />
monitoring the production process and the sequence of activities from the material production to delivery<br />
of the final products (Chang, Hung, Yen & Chen, 2008). Chang et al. also point out that retailers in<br />
developed countries such as America and Japan have employed some unique characteristics of RFID to<br />
enhance their global supply chain management and efficiency. This also leads to other benefits such as<br />
cost reduction, and increase in value and quality of products, Chang et al. indicates.<br />
With South Africa’s already sophisticated retail system, and their drive to expand into the rest of the<br />
African continent, the management and movement of products to stores in different countries have a<br />
critical issue. RFID appear to be the technology needed to help trace, manage the handling of products<br />
all the way from the manufacturer to the point of sale in retail store. Despite the potential usefulness of<br />
RFID, the adoption and application of this technology among South African retailers seem very limited.<br />
Brown and Russell (2007) in their study of RFID adoption in South African retailers found no traces of its<br />
adoption; the most progressive retailer had only began talking about it for possible future exploration. To<br />
date, the Annual reports of the major retailers in South Africa are still silent on the adoption and<br />
application of RFID.<br />
Moreover, the benefits of RFIDs are not limited to supply chain, but are enjoyed by other parts of<br />
retailing, including marketing and customer service provision. In this line of thinking, Eckfeldt (2005) and<br />
Uhrich et al., (2008) identified some three ways in which RFID can provide value to consumers including<br />
convenience, improved service, and increased customer satisfaction, which together can result in<br />
competitive advantage and stimulate demand.<br />
Figure 3 illustrates a typical produce supply chain, and with the application RFID, the retail at the store at<br />
can track the goods right from the point of production through to the final consumer, even after purchase.<br />
3. Research methodology<br />
The character of this paper calls for the participation of different stakeholders within the South African<br />
retail sector. This starting point directs the research primarily towards the use of a qualitative research<br />
method to be applied. An accurate description of qualitative research is that it is concerned with analysing<br />
human action in terms of meanings (Ezzy, 2002). Qualitative research is noted by the significance of<br />
producing detailed or thick descriptions of the social settings being investigated. Neuman (2000) agrees<br />
with Myers (1997) that qualitative data are in the form of text, written words, phrases, or symbols<br />
describing or representing people, actions, and events in social life. In line also with the various<br />
354
Jonathan Oni et al.<br />
description of qualitative research, Blaikie (2000 p.251) assert that qualitative researchers “give a great<br />
deal of attention to describing what might appear to be minute and trivial details of social activity. It is<br />
argued that this is necessary to provide a backdrop within which actions and interactions can be<br />
understood”.<br />
Figure 3: Produce supply chain (source: GS1 (2009:8))<br />
In this paper, the semi-structured interview approach was employed. This involves commencing<br />
interaction as an unstructured interview by presenting the primary overall question, and then using certain<br />
questions to elicit information. The interview questions are designed to provide the in-depth analysis<br />
required and also to allow the authors to probe for further information. One advantage of interviews as<br />
data collection techniques is that they allow the researcher the opportunity to use individual quotes and<br />
refer to specific situation as implied by the interviewee.<br />
3.1 Document analysis<br />
Extensive academic literatures on RFID adoption and implementation were reviewed (see figure 1.2);<br />
including journals on previous related research, as well as white papers. The articles were reviewed and<br />
analyzed and the gaps identified. The aim being to ascertain to what extent RFID is being applied to the<br />
South African retailing sector and its relevance.<br />
3.2 Interviews<br />
The interviewed managers were asked about the relevance of RFID and their application across different<br />
chain stores. They were also asked about their experiences with RFID application in the retail sector of<br />
South Africa. Specifically, they were asked factors that influence the application across various retail<br />
outlets as well as challenges being faced.<br />
The study targeted some retail industry, namely the fast moving consumer goods (FMCG) retailers in<br />
South Africa. The first series of interviews was conducted with Four (4) retail and store managers. Three<br />
(3) additional IT Consultants’ with several years of network technologies and application experience were<br />
interviewed. The interviews were conducted between September and November 2010. Although, two of<br />
the consultants could not be reached for a face-to-face interview due to time constraints, they were<br />
however asked to complete the interview questions at their private time and return their response via<br />
email. To ensure participants of their anonymity, each participant was provided with an informed consent<br />
form outlining the aims of the paper and stating that they will remain anonymous at all times.<br />
The authors were aware of shortcomings of the interview technique such as interview bias, interviewer<br />
effect, and the potential impact of the interviewer characteristics. The interview commenced only after<br />
355
Jonathan Oni et al.<br />
they agreed to participate. Information gained in the interviews were recorded, transcribed and analysed<br />
afterwards.<br />
4. Analysis of findings<br />
Our findings identified several factors that influence the application of RFID across South African retail<br />
sector. The relevance RFID was also outline by the various respondents in the interview. The response of<br />
the interviews are presented and discussed below.<br />
4.1 Cost<br />
Although there are compelling reasons to adopt RFID to replace barcodes as noted by several<br />
respondents, issues of cost have been identified. The ever changing technological know-how brings new<br />
challenges. The software cost, integration and implementations, and supporting infrastructure costs are<br />
very high. The additional cost of ongoing training of staff coupled with the risk involved in loosing these<br />
staff makes the adoption and application less attractive. With the slow pace in buying power of<br />
customers, investing on RFID is considered ‘suicidal’ as noted by one of the respondent.<br />
Cost concerns are consistent with various literatures (Brown & Russel, 2007); (Huber et al., 2007).<br />
Thus, this supports the respondents’ view that costs associated with RFID negatively influence its<br />
adoption and application.<br />
4.2 Technology<br />
The pace of technology adoption in Africa is considered limited due to the cost involved. Although,<br />
innovative organizations encourage adoption of new technologies, the fear of perceived benefits<br />
discourages some, whether they are from within or outside of the organization. In line with this, Asif and<br />
Mandviwalla (2005), states that technology can add to value by reducing costs or enhancing quality. The<br />
technological environment of most retailing sector in South Africa does not encourage the adoption of<br />
RFID. One respondent noted the following, “We do not have the basic tools that will encourage the<br />
application of RFID, even if we are able to afford it, we still do not have the man-power to implement it”.<br />
Other respondent also feel that RFID technology can not be compatible will their organizational<br />
structures. The uncertainties that surround future Government policies regarding storage and use of<br />
information among retailers also hinder the application of RFID.<br />
4.3 RFID readiness - environment<br />
Most of the responders draw attention to the environmental conditions which have impacted on adoption.<br />
The readiness to adopt and implement RFID will greatly depend on the business environment. Real<br />
advantage of RFID can only be gained in the supply chain when suppliers and customers collaborate.<br />
This implies that willingness of both parties to improve their environment and business processes might<br />
facilitate the adoption of a technology that will bring benefits to both of them (Fawcett et al., 2006).<br />
One respondent cited “external forces, internal resources, technical and behavioural issues” as some of<br />
the potential factors influencing RFID application in the retail sector. It can thus be implied that the<br />
relevance of RFID is seen among the different retails outlets but the enabling environment has imparted<br />
negative on its full application.<br />
4.4 IT skills<br />
In South Africa, skills shortage across a range of market sectors especially in information technology and<br />
engineering is a pressing issue (Anderson, 2008; Mulder, 2007). Hence, the non-availability of necessary<br />
IT skills to implement, integrate and maintain RFID greatly impacts on his adoption and application. One<br />
respondent noted “…we can only implement this technology (RFID) if we can boost of the necessary IT<br />
skills, which at the moment, we can no”. Another respondent said “Our head office will already short staff<br />
in the IT department; embracing RFID will compound our already existing problem of lack of skills and<br />
staff”. As reported by Dainty et al. (2005), effort to create sustainable future supply of indigenous skills<br />
must begin with a robust campaign to promote the industry, its occupations and its careers. This process<br />
will help in building and maintaining a talent pipeline through job experience that is not only confined to<br />
home market but also to the global market (Hall & Sandelands, 2009).<br />
356
4.5 Improved delivery and reliability<br />
Jonathan Oni et al.<br />
The respondents’ acknowledge that RFID has enabled substitution of manual processes into an<br />
electronic means. This in turn has resulted in enhanced delivery of goods and services with accuracy and<br />
faster processing capability. Another area of its relevance is the complete acquisition of outgoing<br />
shipments both within the organisation and outside correspondence. One respondent noted that “this<br />
technology has enabled them to trace missing materials from warehouse to the retail outlets”. Perhaps,<br />
this improved delivery and reliability would be apparent with Wal-Mart’s bid to take over Massmart in<br />
South Africa; presenting some retailers an opportunity for RFID to be introduce in a significant way in the<br />
South African retailing system, if, of course, the takeover goes through successfully.<br />
4.6 Increased efficiency in the supply chain<br />
The goal of cost efficiency and fast handling of stocks with the supply chain has been enhanced with<br />
RFID. As large volumes of goods are handled, savings in time in handling these goods become an<br />
important advantage. For example, it has been noted that expediting the loading and unloading of trucks,<br />
reduces the risks of contamination of goods (Shulman, 2001). One respondent agrees that “increased<br />
efficiency in the supply chain using RFID has also enabled them to track products not purchased before<br />
their sell-by date”. This agrees with reports stating that in the <strong>European</strong> grocery sector, products that are<br />
not purchased before their sell-by date are estimated to cause yearly costs of billions of dollars (Leicester<br />
University & Cranfield University, 2001).<br />
4.7 Improved service<br />
Some South African retails stores are already beginning to weigh how RFID-based systems might<br />
improve the overall customer-service and consumer shopping and -spending experience. This for<br />
example can be seen in Prada’s experimenting with high-tech futuristic stores that provide real-time<br />
inventory to sales associates, on-demand multimedia displays of sexy fashions, and dressing rooms<br />
showing customers runway images of models wearing the clothes they’re trying on (Eckfeldt, 2005).<br />
Airlines in South Africa like their counter parts, including Delta Air Lines, are beginning to use RFID<br />
tracking systems to ensure that baggage arrives on time at the correct, specified destinations. These<br />
show that there is improved service satisfaction on both the organisation implementing RFID and its<br />
customers.<br />
5. Conclusion<br />
Although there is substantial interest in the adoption and application of RFID technology by major<br />
retailers around the world, this interest appear comparatively limited in the South African retail system<br />
due to several constraints as discussed earlier. But the situation is not unique to South Africa as most<br />
developing economies have general ICT related issues.<br />
This study has shown that cost, technology, RFID environment readiness, IT skills are factors influencing<br />
the full application of RFID in South African retailing sector. It has also identified that RFID is being<br />
utilized in various sectors to improve delivery and reliability, increase efficiency in the supply chain, and<br />
improve service among organisations and customers especially in the case of global retail chains like<br />
Wal-Mart. As RFID technology progresses in its development, and application and becomes readily<br />
available in the near future, there will be vast opportunities for different organisations to not only embrace<br />
the technology, but also to examine its implications and benefits in a larger scale. Hence, this study might<br />
be useful in promoting the relevance and application of RFID not only in the retail sector of South Africa<br />
but also in other sectors. The information gathered will assist organisations that want to implement this<br />
technology.<br />
To successfully adopt and implement RFID in the retail sector of South Africa, enterprises will have to<br />
resolve issues of cost, technology development, IT skills and environment readiness. As these issues are<br />
addressed, it is certain that RFID will continue to innovate and diffuse, and it will quickly be assimilated<br />
into our daily lives. Though, limited South African firms are employing RFID to improve the efficiency of<br />
their operations and to gain a competitive advantage, it is believed that the near future will see an<br />
increase in this.<br />
6. Limitations<br />
The issues covered in this research are a reflection of the different interviews conducted with various<br />
retail and store managers as well as IT consultants. Consequently, some of these issues may differ from<br />
357
Jonathan Oni et al.<br />
one organisation to another and may not directly reflect a particular RFID adoption and application<br />
approach. A wider and more quantitative study might be necessary in the future to allow for<br />
generalisation.<br />
7. Contribution<br />
The experience gathered from this study will be applied to programs focused on RFID adoption and<br />
application as well as relevance in the South African context. It will assist decision makers in identifying<br />
potential factors that could influence RFID adoption. It is hope that by sharing this research study, it can<br />
stimulate the use of RFID as a tool to improve and gain competitive advantage in the retail sector. This<br />
study has not included all the factors that could influence RFID application or its relevance to the South<br />
African context such as organisation perception, competitive pressure, industry and trade, etc. Thus, this<br />
can be an area for future research.<br />
References<br />
Anderson, B. (2008). "Dire skills gap", Finweek, No.17 April, pp.81.<br />
Angeles, R. (2005). RFID Technologies: spply-chain applications and implementation issues. Information systems<br />
management, Winter: 51-63.<br />
Asif, Z. and Mandviwalla M. (2005). Integrating the Supply Chain with RFID: A technical and business analysis.<br />
Communications of the Association for Information Systems, 15, 393 - 427.<br />
Blaikie, N. W. H. (2000). Designing social research: the logic of anticipation, Polity Press, Malden, MA.<br />
Brown, A., Bakhru, A. (2007). Information Systems Innovation Research and the Case of RFID. In: McMaster, T.,<br />
Wastell, D., Ferneley, E., DeGross, J. (Hrsg.). Organizational Dynamics of Technology-based Innovation:<br />
Diversifying the Research Agenda, Volume 235, 363-376, Boston, MA: Springer.<br />
Brown, I. & Russell, J. (2007). Radio frequency identification technology: An exploratory study on adoption in the<br />
South African retail sector. International Journal of Information Management, 27: 250-265.<br />
Carmody, P. (2009). A new socio-economy in Africa? Thintegration and the mobile phone revolution. IIIS discussion<br />
paper no. 279. URL.<br />
Cazier, J.A., Jensen, A.S. & Dave, D.S. (2008).The impact of consumer perceptions of information privacy and<br />
security risks on the adoption of residual RFID technologies. Communications of the Association of Information<br />
Systems, 23(14): 235-256, September.<br />
Chang, S.I., Hung, S.Y., Yen, D.C. & Chen, Y.J. (2008). The determinants of RFID adoption in the logistics industry -<br />
a supply chain management perspective. Communications of the Assocation for Information System,<br />
23(12):197-218.<br />
Chen, Y.-H. (2004). Getting ready for RFID. OR/MS Today, 31 (3), 30-35.<br />
Curtin, J., Kauffman, R. J., & Riggins, F. J. (2006). Making The ‘Most’ Out Of RFID Technology: A Research Agenda:<br />
For The Study Of The Adoption, Usage And Impact Of RFID. Paper presented at the Information Technology<br />
and Management.<br />
Dainty, A.R.J., Ison, S.G., Root, D.S. (2005). "Averting the construction skills crisis: a regional approach", Local<br />
Economy, Vol. 20 No.1, pp.79-89.<br />
Dos Santos, L.A. & Bacchialoni, A. (2009). ICT in the South African retail and wholesale vertical. White<br />
paper.http://www.bmi-t.co.za/?q=content/sa-ict-retail-and-wholesale-vertical-sector [20 September 2009].<br />
Eckfeldt, B. (2005). What does RFID do for the consumer? Communications of the ACM, 48(9): 77-79, September.<br />
Ezzy, D. (2002). Qualitative analysis: Practice and innovation, Allen & Unwin, Crows Nest, N.S.W.<br />
Fawcett S.E., Ogden J.A., Magnan G.M. and Cooper M.B. (2006). Organizational commitment and governance for<br />
supply chain success. The International Journal of Physical Distribution & Logistics Management, 36, 22 - 35.<br />
Fitzek, D. (2003). Application of RFID in the grocery supply chain: Universal solution for logistics problems in the<br />
CPG industry or a mere hype? Technical Report, Universitat St. Gallen.<br />
Geiger, T. & Mia, I. (2009). Mobile telephony: a critical enabler of networked Readiness?.In Dutta, S. & Mia, I. (eds).<br />
The global information technology report 2008-2009: mobility in a networked world.<br />
http://www.weforum.org/pdf/gitr/2009/gitr09fullreport.<br />
GS1. (2009). Traceability for fresh fruits and vegetables implementation Guide.<br />
http://www.gs1.org/docs/gsmp/traceability/Global_Traceability_Implementation_Fresh_%20Fruit_Veg_i1.pdf [22<br />
November 2010].<br />
Haag, S., Baltzan, P. & Phillips, A. (2007). Management information systems for the information age, Boston, MA:<br />
McGraw-Hill/Irwin.<br />
Hall, J., Sandelands, E. (2009). Addressing South Africa's engineering skills gaps", Education + Training, Vol. 51 Iss:<br />
3, pp.215 – 219<br />
Huber, N., Michael, K., McCathie, L. (2007). Barriers to RFID Adoption in the Supply Chain. IEEE RFID Eurasia,<br />
Istanbul, 1-6.<br />
ITU: International Telecommunication Union. (2007). WORLD INFORMATION SOCIETY REPORT. www.itu.int/wisr.<br />
[04 January 2011].<br />
Jones, P., Clarke-Hill, C., Shears, P., Comfort, D., Hillier, D. (2004). Radio Frequency Identification in the UK:<br />
Opportunities and Challenges. International Journal of Retail & Distribution Management, 32 (3), 164-171.<br />
Lai, K.-H., Ngai, E.W.T., Cheng, T.C.E. (2005). Information Technology Adoption in Hong Kong's Logistics Industry.<br />
Transportation Journal, 44 (4), 1-9.<br />
358
Jonathan Oni et al.<br />
Leicester University and Cranfield University. (2001), Shrinkage – Introducing a Collaborative Approach to Reducing<br />
Stock Loss in the Supply Chain, ECR Europe, research report.<br />
Leimeister, J.M., Knebel, U., Krcmar, H. (2007). RFID as Enabler for the boundless real-time Organisation: Empirical<br />
Insights from Germany. International Journal of Networking and Virtual Organisations, 4 (1), 45-64.<br />
Matta, V., Moberg, C. (2006). The Development of a Research Agenda for RFID Adoption and Effectiveness in<br />
Supply Chains. Issues in Information Systems, 7 (2), 246-251.<br />
Mulder, C. (2007). "Nurture the minority who show potential and desire", Finweek, No.8 February, pp.60.<br />
Myers, M. (1997). 'Qualitative Research in Information Systems', MIS Quarterly, vol. 21, no. 2, pp. 241-242.<br />
Neuman, W. L. (2000). Social research methods: qualitative and quantitative approaches, 4th edn, Allyn and Bacon,<br />
Boston.<br />
Roberts, C.M. (2006). Radio frequency identification (RFID). International Journal of Compute r s and S e c u r i t y, 2<br />
5 (2006) 18 – 26.<br />
Riggins, F.J., Slaughter, K.T. (2006). The Role of Collective Mental Models in IOS Adoption: Opening the Black Box<br />
of Rationality in RFID Deployment. Proceedings of the 39th Hawaii International <strong>Conference</strong> on System<br />
Sciences (HICSS), Hawaii, HI.<br />
Sagi, J., Carayannis, E., Dasgupta, S. & Thomas, G. (2004). ICT and business in the new economy: globalisation<br />
and attitudes towards ecommerce. Journal of global information management, 12(3): 44-64, July-September.<br />
Schmitt, P. and Michahelles, F. (2009). Status of RFID/EPC Adoption. Auto-ID Labs White Paper WP-BIZAPP-048.<br />
March 2009<br />
Sharma, A., Thomas, D., Konsynski, B. (2008). Strategic and Institutional Perspectives in the Evaluation, Adoption<br />
and Early Integration of Radio Frequency Identification (RFID): An Empirical Investigation of Current and<br />
Potential Adopters. Proceedings of the 41st Hawaii International <strong>Conference</strong> on System Sciences (HICSS),<br />
Hawaii, HI.<br />
Shih, D.-H., Chiu, Y.-W., Chang, S.-I., Yen, D.C. (2008). An empirical Study of Factors affecting RFID’s Adoption in<br />
Taiwan. Journal of Global Information Management, 16 (2), 58-80.<br />
Shulman, R. (2001), "Perishable systems take center stage", Supermarket Business, Vol. 56 No.4, pp.47-8.<br />
Roztocki, N. & Weistroffer, H. R. (2009). Information and Communications Technology in Developing, Emerging and<br />
Transition Economies: An Assessment of Research. AMCIS 2009 Proceedings. Paper 652.<br />
http://aisel.aisnet.org/amcis2009/652<br />
Strüker, J., Gille, D. (2008). The SME Way of Adopting RFID-Technology – Empirical Findings from a German crosssectoral<br />
Study. 1<strong>6th</strong> <strong>European</strong> <strong>Conference</strong> on Information Systems (ECIS 2008), Galway.<br />
The World Retail Congress. (2010). Day three programme. Daily Magazine, in association with Retail Week. October<br />
25-27, Berlin.<br />
Uhrich, F., Sandner, U., Resatsch, F., Leimeister, J.M. & Krcmar, H. (2008). RFID in retailing and Customer<br />
Relationship Management. Communications of the Association for Information Systems, 23: 219-234.<br />
United Nations. UNCTAD Secretariat. (2007). The information economy report 2007-2008: science and technology<br />
for development: the new paradigm of ICT. http://www.unctad.org/en/docs/sdteecb20071_en.pdf. [20June<br />
2008].<br />
Whitaker, J., Mithas, S., Krishnan, M.S. (2007). A Field Study of RFID Deployment and Return Expectations.<br />
Production and Operations Management, 16 (5), 599-612.<br />
Wyld, D.C., Jones, M.A., Totten, J.W. (2005). Where is my Suitcase? RFID and Airline Customer Service. Marketing<br />
Intelligence & Planning, 23 (4), 382-394.<br />
359
Towards a Comprehensive Evaluation Framework for ICT for<br />
Development Evaluation – an Analysis of Evaluation<br />
Frameworks<br />
Caroline Pade-Khene and Dave Sewry<br />
Rhodes University, Grahamstown, South Africa<br />
c.khene@ru.ac.za<br />
d.sewry@ru.ac.za<br />
Abstract: Information and communication technologies (ICT) can support rural development activities through<br />
providing supportive information and knowledge, and creating essential interconnectivities between rural areas and<br />
more developed regions. However, rural ICT for development (ICT4D) is still at best a ‘working hypothesis’, faced<br />
with barriers and challenges associated with implementation and use in the rural environment; which threaten the<br />
success, sustainability or relevance of ICT interventions in developing countries. The evaluation of rural ICT projects<br />
is indispensable as it determines the need, effectiveness, impact, sustainability and extent of the awareness of the<br />
contribution such projects or programmes can make in poverty alleviation and development. Even so, existing ICT for<br />
development evaluations are confronted with shortcomings and challenges which influence the reliability of<br />
evaluation conclusions. These shortcomings highlight the need to embark on a more comprehensive evaluation<br />
approach, sensitive to the rural environment. The paper adopts a theoretical multi-method approach to determine the<br />
multiple variables and components associated with rural ICT evaluation, and then to determine how these variables<br />
interrelate. The approach is founded on programme evaluation, ICT for development evaluation, and information<br />
systems evaluation. Programme evaluation informs this research study of the fundamental domains of evaluation<br />
and the structure that shapes an evaluation. An analysis of ICT for development evaluation frameworks and<br />
information systems frameworks compares and contrasts key characteristics that form the structure of a<br />
comprehensive evaluation that can be applied in ICT for development projects. The analysis and a review of<br />
programme evaluation enables the development of a Rural ICT Comprehensive Evaluation Framework (RICT-CEF)<br />
that encompasses the key components essential for a comprehensive evaluation of rural ICT projects. The<br />
theoretical framework aims to inform ICT intervention to improve and support rural development, through the<br />
application of fundamental and interconnected evaluation domains sensitive to the rural environment, throughout the<br />
project’s lifecycle.<br />
Keywords: ICT for development, information systems evaluation, rural development, programme evaluation<br />
1. Introduction<br />
Rural development is a significant focus for developing countries toward reaching their goals in<br />
development and poverty alleviation. ICTs are known to be key tools that have the potential to enhance<br />
rural development activities, as the availability of information and knowledge can support the operation<br />
and effectiveness of rural development activities. However, the growing establishment of ICT projects in<br />
rural areas signifies the need for accountability and an evaluation of the effectiveness, impact, and<br />
sustainability of such projects. There are cases where ICTs are known to have social and economic<br />
benefits in communities, but there are also cases where the implementation of ICT projects has actually<br />
not made a difference, or the effects have been harmful in communities (Tacchi, Slater and Lewis 2003).<br />
ICT4D actors need to understand that ICT4D is, at best, a “working hypothesis” where many key<br />
questions remain largely unanswered, with no concrete or credible data to support a wide range of claims<br />
concerning the use of ICT for development (Wagner, Day, James, Kozma, Miller, and Unwin 2005;<br />
Moodley 2005). Therefore, the evaluation of ICT4D initiatives is essential. Nevertheless, Pade and Sewry<br />
(2009) illustrate that current ICT4D evaluations are confronted with shortcomings and challenges which<br />
influence the reliability of evaluation conclusions. Embarking on a multi-method approach of evaluation<br />
theory can provide a foundation and guideline for developing a more comprehensive framework that<br />
addresses and considers the shortcomings and challenges of current ICT4D evaluation.This research<br />
paper introduces programme evaluation as a foundation for defining evaluation. A summary of key<br />
aspects that are considered across all the evaluation frameworks are explored as part of the process of<br />
constructing a comprehensive evaluation framework which is sensitive to rural ICT environments.<br />
Subsequently, a skeleton view of the Rural ICT Comprehensive Evaluation Framework (RICT-CEF) is<br />
presented. It is concluded that in order to obtain a better understanding and application of the RICT-CEF,<br />
a real-life case study investigation can reveal the lessons learned (shortcomings and suitability) from<br />
applying such a framework in a rural environment.<br />
360
Caroline Pade-Khene and Dave Sewry<br />
2. The foundation of programme evaluation<br />
Programme evaluation can be defined as (Rossi et al. 2004): “The use of social research methods to<br />
systematically investigate the effectiveness of social intervention programs in ways that are adapted to<br />
their political and organisational environments and are designed to inform social action to improve social<br />
conditions”. In contrast to basic research, the main objective of programme evaluation is not necessarily<br />
the discovery of new knowledge, but rather the “study of the effectiveness with which existing knowledge<br />
is used to inform and guide practical action” (Clarke 1999). This research study adopts Rossi et al.’s<br />
(2004) defined domains of programme evaluation, which include: Needs Assessment, Programme<br />
Theory Assessment, Process Assessment, Outcome Assessment, Impact Assessment, and Efficiency<br />
Assessment. The purpose and goals specific to a domain of evaluation attach a unique evaluation<br />
structure for each domain.<br />
3. A comprehensive evaluation<br />
Customarily, most researchers define the evaluation of a programme around assessing the impact it has<br />
on the target population. However, other aspects need to be considered when determining the<br />
effectiveness of a programme, which are related and build on each other to produce a more profound<br />
evaluation. Each evaluation domain applies a different approach to analyse an aspect of a programme,<br />
developing results that may build onto other evaluation types. In this case, a comprehensive evaluation,<br />
which is the ideal, becomes essential. Rossi et al. (2004) indicates that a comprehensive evaluation<br />
constitutes a process evaluation, which builds into an impact evaluation. On the other hand, Babbie and<br />
Mouton (2001) provide a more broad definition, where a comprehensive evaluation includes the<br />
application of all five domains of evaluation to assess the effectiveness of a programme. The<br />
development of an evaluation framework to apply to a particular programme intervention requires that<br />
there be a frame of reference (theory or model), that provides a base to understand the programme’s<br />
goals and context in which it operates (Babbie and Mouton 2001). Analytical frames of reference, such as<br />
theories and models, fulfil the following important functions: provide definitions of key concepts, guide<br />
data-collection and analysis, and provide an explanatory (causal) framework for the interpretation of<br />
empirical findings (Babbie and Mouton 2001). Consequently, the development of a comprehensive<br />
evaluation framework for rural ICT projects requires a collective analysis of the various theories, models,<br />
and aspects that may constitute the evaluation of a rural ICT project.<br />
4. An analysis of evaluation frameworks<br />
The factors associated with ICT project implementation in rural areas suggest the need to embark on a<br />
multi-method approach to determine the common multiple variables associated with rural ICT evaluation,<br />
and then to determine how these variables interrelate (Lannon 2007; Hudson 2001b). Two evaluation<br />
areas provide useful insight for the development of a comprehensive evaluation framework: the<br />
evaluation of ICT for development, and Information systems evaluation. ICT4D evaluation frameworks<br />
point out essential factors that should be built into evaluation based on the experience and lessons<br />
learned that have been applied to develop and design such frameworks. An analysis of IS evaluation<br />
frameworks also provide an outlook of aspects that can be applied in ICT4D projects to assess its value<br />
and usefulness, in volatile rural environments. A template based on foundational literature on programme<br />
evaluation, the need for ICT4D evaluation, and the shortcomings and challenges of current ICT4D<br />
evaluation was applied to each evaluation framework to provide a comparative analysis. The evaluation<br />
template produces a review of key frameworks that can contribute to the development of a<br />
comprehensive framework. The ICT4D evaluation frameworks selected are listed in Figure 1 and the IS<br />
evaluation frameworks selected are listed in Figure 2.<br />
A summary of an exploration of key aspects across the evaluation frameworks is discussed below. The<br />
review focuses on fundamental characteristics and approaches that can contribute to the development of<br />
a comprehensive evaluation framework for ICT4D evaluation.<br />
4.1 The project lifecycle stage at which the evaluation is conducted<br />
Lifecycle stages differ in five common ways:<br />
Pilot stage and/or proposal stage (frameworks 1 & 14): Essential to determine whether the ICTD pilot<br />
project is relevant or worth being taken to scale.<br />
361
Caroline Pade-Khene and Dave Sewry<br />
Annually (framework 2): A less time consuming and costly approach to evaluation, and allows a<br />
comparison to be made throughout the years of implementation and operation of the ICTD project.<br />
However, it is still important to regard the lifecycle stage<br />
Figure 1: List of ICT4D evaluation frameworks analysed<br />
Figure 2: List of IS evaluation frameworks analysed<br />
Only implementation and post-implementation stages (frameworks 6, 9, 12 & 13): Evaluation aimed<br />
at applying formative (process improvement) and summative (impact) evaluation. Lessons learned<br />
indicate that such frameworks expand their approach to evaluating other project stages. When ICT4D<br />
evaluation occurs during post-implementation, the aim is not to close the project. Rather, a project’s<br />
process/implementation will be assessed to learn lessons and improve or enhance the approach to<br />
implementing similar projects in other rural areas, once a successful or failed project has been fully<br />
implemented.<br />
Post implementation (framework 5): A typical stage conducted in ICTD evaluation. Impact<br />
assessment timing is crucial, and it is infeasible and unrealistic to assume that an ICT initiative can<br />
impact a community over a short period of time. A possible solution is to apply an iterative approach<br />
to impact assessment (as part of a comprehensive framework); identifying impacts that result and<br />
build up over time (Hudson 2001b; Pade and Sewry 2009).<br />
Throughout the life of the project (framework 3, 4, 7, 8, 14, & 15): Common evaluation domains<br />
associated with this approach include a needs assessment, process assessment and impact<br />
assessment across the lifecycle stages of the project. Some frameworks additionally consider<br />
efficiency assessment or programme theory assessment. Frameworks that appear to be or state they<br />
are applied throughout the life of the project, does not guarantee the evaluation is comprehensive. All<br />
domains of evaluation need to be applied throughout the project life cycle.<br />
4.2 The project environment<br />
The environments that most ICT projects target are rural, remote or marginalized areas in developing<br />
countries. An evaluation, like a project (Schwalbe 2006), cannot operate in isolation and needs to be<br />
sensitive to the complexities and factors of their environments. The project environment for an<br />
362
Caroline Pade-Khene and Dave Sewry<br />
information systems evaluation is usually not specified. However, in two particular frameworks (12 &15),<br />
the project environment is indicated, as the authors view evaluation as being partly shaped by the<br />
environment in which it is applied. Evaluation is seen as a continuous learning process as the operation<br />
and use of the system is sensitized to needs of its environment, through improvements applied and<br />
suggested from preceding evaluations (framework 12; framework 15).<br />
4.3 Domains of evaluation<br />
The evaluation domains (related to programme evaluation) include:<br />
Needs Assessment: Only four ICTD frameworks (frameworks 3, 4, 5 and 10) and two IS evaluation<br />
frameworks (frameworks 14 & 15). It aims to understand and set the desired priorities of the rural<br />
community and propose solutions that can be supported by ICT. The assessment should also<br />
accommodate demand driven needs that emerge over-time. IS evaluation suggests that a needs<br />
assessment should explore the project’s contribution and alignment with business strategy (rural<br />
development strategy). Furthermore, the assessment should be based on 1) the need for a<br />
development solution, and 2) an awareness of potentially new ICT initiatives relevant to the rural<br />
context.<br />
Programme Theory Assessment (framework 14): This domain specifically assesses whether the<br />
conception of an ICT4D project is actually designed to support rural development programmes.<br />
Where an assessment reveals the programme theory is flawed, responsible stakeholders need to<br />
redesign the programme. The domain is not applied or considered across the ICT4D evaluation<br />
frameworks and only in one IS framework; yet it is a crucial evaluation domain in such projects. The<br />
evaluation exercise described in framework 14 initially focuses on the project proposal stage, where a<br />
decision has to be made on whether to go ahead with the IS project or not. The types of theory<br />
assessed, relate to some of the tasks of the framework, e.g. implementability, risk, analysis of<br />
benefits, etc.<br />
Process Assessment: commonly applied across most of the evaluation frameworks analysed<br />
(frameworks 1, 2, 5, 7, 8 and 15). Aspects evaluated include project objectives, output and planning,<br />
sustainable ICT characteristics; all targeted at stakeholders involved in managing a project locally.<br />
Process assessment is generally seen as a procedure that results in suggested improvements to a<br />
project’s implementation, but most importantly, it should be seen as a continuous learning process (or<br />
iterative), given the complexity of rural environments.<br />
Impact and Outcome Assessment: a popular evaluation domain which all the evaluation frameworks<br />
apply. They are to an extent interchangeable, as most evaluation frameworks tend to describe them<br />
as similar aspects; however, Rossi et al. (2004), define them as separate domains. Some<br />
frameworks describe their specific outcome/impact assessment goals or approaches and provide<br />
useful concepts that should be included in the assessment of ICT for development projects, e.g.<br />
strengthening existing social and organisational community structures, gender sensitivity, etc. Impact<br />
assessment can be applied in two ways, i.e. 1) comparing before implementation and after<br />
implementation data, 2) assessment only post-implementation using retrospection, and ‘contrary to<br />
the fact questions (framework 8). Framework 14 sites that two elements need to be recognised in the<br />
assessment of outcome, that is, firstly the two stage benefits realization process (short-term<br />
(outcome) & long-term (impact)), and secondly, when developing impact theory, to make realistic<br />
assumptions concerning the project’s impact.<br />
Efficiency Assessment (frameworks 2, 11 and 14): Usually consists of a cost-benefits analysis and a<br />
cost-effectiveness analysis. This adds rigour to an impact assessment, and should be considered as<br />
part of a more comprehensive evaluation, rather than an isolated evaluation approach. A variety of<br />
costs for ICT4D can be determined, and the literature presents a variety of perspectives or<br />
approaches to financial assessment for cost benefit analysis (Heeks and Molla 2008).<br />
Additional evaluation domains identified include:<br />
A Baseline Study (frameworks 4 and 8): Serves to inform external stakeholders of the social, cultural,<br />
economic, political and technological (readiness) status of a community, as well as introduce<br />
community members to the project. It also acts as a foundation for a needs assessment and adds<br />
value to an impact assessment (comparing the before and after status of community).<br />
Flexibility Assessment (framework 14): An important domain in the context of rural ICT projects,<br />
whose environment is unpredictable at times (demand driven needs also emerge). Two aspects that<br />
should be considered in assessing flexibility include: 1) Examine the future changes that the ICT<br />
363
Caroline Pade-Khene and Dave Sewry<br />
project may be subjected to in the rural environment, and then ensure that these changes can be<br />
easily accommodated, and 2) an assessment needs to be done to determine whether the proposed<br />
ICT4D project is building in constraints to flexibility (for example, the project should be viewed<br />
holistically, considering how other traditional information flows (word of mouth, community meetings<br />
etc.) may be affected negatively or positively by the project).<br />
Scalability Assessment (framework 1): Determines whether a pilot project should be brought to scale.<br />
It assesses the conditions that are needed to replicate or scale up the project, and what might be the<br />
result of large scale application. ICT projects should begin at pilot level, to observe their success or<br />
potential sustainability.<br />
4.4 The evaluation model/approach/process<br />
The approach/model applied in each framework is characterised by the purpose of the evaluation, the<br />
project’s environment, and in some cases, the evaluation domain. The approaches applied can be<br />
categorized into:<br />
A process approach (frameworks 1, 3, 7, 10, 11, 12 and 13): This is the most common approach<br />
applied across the evaluation frameworks. This approach describes the step by step procedure to<br />
conduct an evaluation. It is not necessarily dependent on the lifecycle stage of a project, like an<br />
evaluation lifecycle approach, but rather focuses on the process or procedure to conduct an<br />
evaluation of any particular domain. In this case, a generic procedure is applied, which can be<br />
applied to assess various domains on which the evaluation framework intends to focus.<br />
A dimension focused approach (frameworks 2, 9 and 14). Evaluation tasks or elements that shape<br />
the process to conduct an evaluation are applied. There is not necessarily a standard sequence for<br />
the tasks/elements stated to carry out the evaluation, but some logic is required to determine an<br />
appropriate sequence to carry out the evaluation of particular projects. The evaluation tasks may<br />
therefore occur simultaneously, independently, or sequentially.<br />
An evaluation lifecycle approach (frameworks 4, 8 and 15): This approach applies the stages of an<br />
evaluation, based on the domains of evaluation. These domains are conducted in sequential order,<br />
depending on the lifecycle stage the project is on.<br />
A case study approach (framework 5 and 6): This evaluation process is a simplistic approach, made<br />
up of qualitative or quantitative methodologies and procedures used to collect the data, and the<br />
results reported. Other aspects of the frameworks (evaluation questions, data collection instruments,<br />
etc.) provide useful insight for ICT4D evaluation.<br />
4.5 Structure of the evaluation<br />
4.5.1 Evaluation questions and indicators<br />
The types of evaluation questions vary, depending on the purpose of the evaluation. Not all the<br />
frameworks specify their evaluation questions, specifically framework 11, 14 and 15. Here, evaluation<br />
questions can be proposed based on their evaluation goals. The questions from the frameworks provide<br />
a useful guideline for the kind of questions that can be asked under each evaluation domain, but are not<br />
necessarily the standard exhaustive list. Some questions may be generic and applicable to any ICT4D<br />
project, but some are for specific types of projects such as education, e-governance, etc. Therefore,<br />
generic questions identified under each domain of evaluation can be used across different ICT4D project<br />
evaluations.<br />
4.5.2 Nature of the evaluator-stakeholder relationship<br />
Of the evaluation frameworks analysed, there is an almost equal distribution of independent,<br />
participatory, and empowerment evaluation:<br />
Independent evaluation (frameworks 1, 2, 5, 6, 9 and 11). An independent evaluator is chosen to plan<br />
or design the evaluation, conduct the evaluation, and disseminate results; often stipulated by the<br />
evaluation sponsor. It is assumed that community (end-user) involvement is made adequate through<br />
the application of data collection methodologies to communicate their views, for instance, focus<br />
groups discussions and interviews. The nature of the evaluator-stakeholder relationship may depend<br />
on the domain of evaluation conducted, e.g. a cost-benefit analysis may need to be conducted by an<br />
evaluator experienced in financial modelling & analysis.<br />
364
Caroline Pade-Khene and Dave Sewry<br />
Participatory or Collaborative Evaluation (frameworks 4, 7, 8, 12, 13 and 15): The evaluator<br />
collaborates with participating stakeholders to plan, conduct and analyse the evaluation. Participatory<br />
research is applied, placing communities at the centre of the evaluation, through understanding the<br />
community’s existing social institutions, processes, and information ecology. For example, local<br />
stakeholders can participate in questionnaire design, specifying topics of local importance,<br />
suggesting culturally acceptable questions for ascertaining age and income levels, and identifying<br />
potential user groups to be included in the sampling frame and focus groups (framework 8). It aims to<br />
produce results for key stakeholders such as practitioners, funders and local participants or endusers.<br />
Empowerment Evaluation (frameworks 3, 8, 10 and 14): Occurs when participatory or collaborative<br />
evaluation additionally focuses on developing the capabilities of participatory stakeholders to conduct<br />
the evaluation. This contributes to an evaluation that is informed by local knowledge, develops local<br />
skill, encourages local ownership, builds demand for ICT services through the community incitement<br />
by local MCT staff, and promotes local sustainability (Scharffenberger 1999; Hudson 2001a). Target<br />
groups in a rural community, who have the potential to engage in and use ICTs to support their rural<br />
livelihood activities, can be selected and trained (empowered) to conduct the evaluation. The leading<br />
evaluator is not necessarily absent in conducting the evaluation when an empowerment approach is<br />
used, but plays a key role in supervising and guiding local stakeholders in the process.<br />
4.5.3 Methods and procedure applied<br />
Most frameworks apply a mixture of methodological approaches, that is, qualitative or quantitative, rather<br />
than focusing on one. Although there is no standard methodology or procedure across the frameworks,<br />
common methodologies can be identified which are assumed to be more effective for research in rural or<br />
marginalized environments. The mix and match of qualitative and quantitative approaches provides a<br />
deeper analysis of evaluation findings from different stakeholder or project perspectives. The Acacia<br />
Evaluation and Learning System framework (framework 8), suggests a useful procedure for mixing and<br />
matching different methodological approaches, called the ‘Rolling Design’ approach (action research),<br />
which integrates multiple data collection instruments that are sequentially implemented to obtain relevant<br />
data from a wide range of perspectives (Hudson 2001a). Examples of the various quantitative<br />
methodologies used include: questionnaires, quantitative data on technology, log sheets/records, and<br />
cost benefit analysis methodologies. Examples of the various qualitative methodologies used include:<br />
tracer studies, participatory appraisals, most significant change, interviews, photography, audio recording<br />
and video, focus groups, workshops, observation and participant observation, and document analysis.<br />
5. A comprehensive evaluation framework for rural ICT projects in developing<br />
countries<br />
An analysis of ICT4D frameworks and Information Systems frameworks, in addition to the concepts of<br />
programme evaluation, provides the fundamental elements necessary to shape and design a<br />
comprehensive evaluation framework for ICT4D projects. A high level (skeleton) illustration of the Rural<br />
ICT Comprehensive Evaluation Framework (RICT-CEF) is presented in this section.<br />
5.1 Defining comprehensive evaluation for rural ICT projects<br />
The comprehensive nature of a rural ICT evaluation is determined by the various domains applied to<br />
evaluate a project, as well as the key aspects that structure the evaluation process. Each evaluation<br />
domain applies a different approach to analyse an aspect of a project, developing results that may build<br />
onto other evaluation domains. Essentially, these different domains which are all key to the evaluation of<br />
a programme are interlinked, and where they are all applied interdependently, they contribute to a<br />
comprehensive evaluation of a rural ICT project (Babbie and Mouton 2001; Rossi, et al. 2004).<br />
Evaluation, in this case, is performed throughout the life of the rural ICT project, from Project Idea<br />
Generation to Post-implementation Review (in reference to the Rural ICT Project Life Cycle (RICT-PLC)<br />
proposed by Pade, Mallinson and Sewry (2008)). Therefore, comprehensive rural ICT evaluation is a<br />
continuous endeavour encapsulating different stages of ICT utilisation by target groups and individuals in<br />
rural communities. Comprehensive evaluation throughout the life of the project requires the evaluator to<br />
be sensitive to the rural environment, and especially aware of any social, political, cultural, or economic<br />
factors that can influence the evaluation form and scope at different stages of the rural ICT project.<br />
Accordingly, the evaluation approach should be multidisciplinary, comparative, participatory and<br />
multicultural to contribute to research on different elements associated with the contribution of ICTs to<br />
rural development. In this research, comprehensive rural ICT evaluation is therefore defined as: The<br />
365
Caroline Pade-Khene and Dave Sewry<br />
comprehensive application of all appropriate domains of evaluation throughout the life of a rural ICT<br />
project, in ways that are adapted to the social, political, cultural, and economic rural environment, and<br />
designed to inform ICT interventions to improve and support rural development.<br />
5.2 The structure of the Rural ICT Comprehensive Evaluation Framework (RICT-CEF)<br />
The RICT-CEF, depicted in Figure 3, is structured around three key components that guide the<br />
comprehensive evaluation of a rural ICT project or programme. The components include the overall<br />
evaluation plan, the comprehensive evaluation lifecycle, and the iterative design of the framework.<br />
Figure 3: The Rural ICT Comprehensive Evaluation Framework (RICT-CEF)<br />
5.2.1 The overall evaluation plan<br />
The Overall Evaluation Plan is applied to the overall evaluation process to guide the various evaluation<br />
domains. Providing a high level view of the evaluation plan, it comprises the key attributes that need to be<br />
366
Caroline Pade-Khene and Dave Sewry<br />
taken into account for planning the overall evaluation (considering the triple constraint of scope, time, and<br />
cost by Schwalbe (2006), and guidelines that should be applied across all the domains of evaluation<br />
(Clarke 1999; Hudson 2001a; APC 2008, Scharffenberger, 1999; Rossi et al. 2004; Wagner et al. 2005;<br />
Dymoke-Bradshaw and Cox 2003; Hedman and Borell 2005). However, each evaluation domain has its<br />
own local level evaluation plan that is specific to meeting the requirements and purpose of that particular<br />
domain. The overall evaluation plan provides a sketch and guideline for the whole comprehensive<br />
evaluation process, and acts as the initiation stage of the comprehensive evaluation framework.<br />
5.2.2 The evaluation lifecycle<br />
The Evaluation Lifecycle forms the central component of the framework, which describes the process and<br />
domains that a rural ICT comprehensive evaluation should go through. The lifecycle has domains that<br />
connect from the beginning to the end of the overall evaluation, and hence provide a gauge of the<br />
evaluation’s progress towards contributing to a comprehensive evaluation of the ICT intervention<br />
(Schwalbe 2006). The domains that make up the comprehensive evaluation cycle are:<br />
1. The Baseline Study: assesses the existing status (socio-economic and readiness) of the rural<br />
community. (frameworks 4 and 8)<br />
2. Needs Assessment: assesses the desired priorities and information needs for rural development,<br />
so as to propose appropriate solutions that can be supported by ICT. (frameworks 3, 4, 5, 6, 10, 13<br />
and 15)<br />
3. Programme Theory Assessment: assesses the conceptual design of the ICT project to support<br />
rural development programmes. (Rossi et al. 2004)<br />
4. Process Assessment: assesses how well the project is operating to implement its intended<br />
functions, as stipulated in the project plan. It also assesses the flexibility of the project. (frameworks<br />
1, 3, 4, 7, 8, 14 and 15)<br />
5. Outcome and Impact Assessment: assesses the intended and unintended outcomes or effects of<br />
the ICT project on the conditions of the targeted population. (frameworks 1, 2, 3, 4, 5, 10 and 14)<br />
6. Efficiency Assessment: assesses ICT intervention costs associated with project effects or impact.<br />
(frameworks 1 and 11)<br />
7. Scalability Assessment: assesses whether a rural ICT pilot project should be taken to scale.<br />
Therefore, this domain is only applied to ICT pilot projects. (frameworks 1 and 12)<br />
Each domain is described according to the following structure:<br />
a).The Purpose of the evaluation<br />
b).The Domain Process: The domain process either adapts one particular process from the<br />
frameworks analysed, or a combination of key processes and practices from different frameworks,<br />
that suit the requirements and purpose of the domain.<br />
c).The Evaluation Plan: Guides the evaluation domain process. It is tailored around three features,<br />
adopted from the evaluation structure of Rossi et al., (2004): 1) The Evaluation Questions, 2) The<br />
Methods and Procedures, and 3) The Nature of the Evaluator-Stakeholder Relationship. The<br />
framework provides examples of commonly used questions or methods that have been applied in<br />
other frameworks or ICT projects, which an evaluator may possibly choose from or adapt for other<br />
ICT4D project evaluations. The nature of the evaluator-stakeholder relationship, conversely, is the<br />
recommended approach for the particular evaluation domain.<br />
d) Pilot Project Aspects: Highlights additional aspects or factors that should be considered when<br />
evaluating a pilot project.<br />
e) Expected Outputs: The expected outputs of the particular domain.<br />
f) The Dissemination of Results: Specifies the target audience for evaluation results and the<br />
approach for documenting and presenting the results.<br />
g) Application Stage of the RICT-PLC: The evaluation framework makes use of the RICT-PLC<br />
proposed by Pade, Mallinson and Sewry (2008) to associate the different evaluation domains with the<br />
rural ICT project phases that may directly use the results of the different evaluations. Every ICT4D<br />
project lifecycle is unique. Even though various ICT projects may have different lifecycles or<br />
associated phases, the RICT-PLC provides a general picture for stakeholders of various ICT projects<br />
interventions to determine where best to apply a particular evaluation domain within their projects.<br />
367
Caroline Pade-Khene and Dave Sewry<br />
Nevertheless, the comprehensive evaluation of the project may initiate during the ongoing<br />
implementation of the project, and not from the inception of the project. In this case, the different<br />
evaluation domain results still serve to improve different aspects of the project that can be revised.<br />
h) Domain Dependence: Each domain of evaluation is interlinked to preceding or subsequent<br />
domains within the evaluation lifecycle (see Figure 3). This section explains the interdependent<br />
relationship between the domains.<br />
5.2.3 The iterative design of the framework<br />
This Iterative Design reflects the iterative nature of rural ICT projects, which adapt over time due to the<br />
progressive integration of technology in rural livelihoods, and the changing nature of the rural<br />
environment, as a result of development activities. In this case, particular evaluation domains are<br />
repeated over-time, building and developing the scope of the evaluation as the project’s intervention and<br />
impact progresses in the rural community (frameworks 12, 15 and 8). This is particularly essential to<br />
address the time-dependent nature of rural ICT projects, as many changes or impacts brought about by<br />
ICT initiatives are long-term and often indirect. ICT4D initiatives are a form of human development<br />
infrastructure which is highly time-dependent. The comprehensive evaluation lifecycle should be<br />
incorporated into the rural ICT project lifecycle, where evaluation results continuously contribute to<br />
improving and guiding the iterative (progressive) development of the ICT project. Within the iterative<br />
design of the evaluation framework, evaluators should document changes and identify evaluation aspects<br />
(needs, process (service utilisation), impact, etc.), that result over time, and also consider barriers that<br />
could hinder long-term impacts. Each iteration (iteration0, iteration1…) is initiated and driven by the<br />
changing needs of the community. The subsequent domains of evaluation are therefore influenced by<br />
changing needs, as the results of the needs assessment build into the evaluation of other aspects of the<br />
rural ICT project, such as, the process, outcome, impact and efficiency (see Figure 3).<br />
Iteration0 is associated with assessing a pilot project, and hence the scalability of the project to determine<br />
whether or not it should be taken to scale. The subsequent iterations (which comprise of only needs<br />
assessment through to efficiency assessment) contribute to evaluating the progressive development of<br />
the rural ICT project over-time. The duration or life of the ICT intervention is not always obvious, and in a<br />
number of cases, they are not sustainable in developing countries. Therefore, the iterative evaluation<br />
design may be appropriate for sustainable on-going ICT projects, but in cases where the ICT services are<br />
not being developed further or new ICT initiatives are no longer being introduced to the rural community,<br />
the evaluation design changes. In this case, the evaluation design in the long-term is centred only on<br />
assessing the impact of the project over time.<br />
6. Conclusion<br />
The evaluation of ICT4D projects is essential to determine their value in supporting a change in social<br />
conditions, and identifying ways to enhance or improve the project. Programme evaluation provides a<br />
foundation to understand the concept of evaluation for social development programmes. The types of<br />
domains of evaluation vary at different stages of a programme’s life cycle. An evaluation framework that<br />
applies all five domains in the evaluation of a programme is considered a comprehensive evaluation<br />
framework. Embarking on a multi-method evaluation approach is essential to develop an appropriate<br />
framework which is sensitive to the various factors associated with rural ICT project implementation. The<br />
results of the framework analysis and a review of programme evaluation highlighted the fundamental<br />
elements necessary to shape and design a comprehensive evaluation framework for ICT4D projects. A<br />
skeleton presentation of the RICT-CEF illustrates the overall evaluation plan, evaluation lifecycle, and<br />
iterative design fundamental for the comprehensive evaluation of rural ICT projects in developing<br />
countries. In order to obtain a better understanding and application of the RICT-CEF, a real-life case<br />
study investigation can reveal the lessons learned (shortcomings and suitability) from applying such a<br />
framework in a rural environment.<br />
References<br />
Association for Progressive Communications, The (2008) “Gender Evaluation Methodology”. [online],<br />
http://www.apcwomen.org/gem/?q=about.<br />
Babbie, E and Mouton, J. (2001). The Practice of Social Research, Oxford University Press, Cape Town.<br />
Batchelor, S. and Norrish, P. (2006). Framework for the Assessment of ICT Pilot Projects, The World Bank, InfoDev,<br />
Washington DC.<br />
Carlsson, S. A. (2003) “Advancing Information Systems Evaluation (Research): A Critical Realist Approach”,<br />
Electronic Journal of Information Systems Evaluation, Vol 6, No.2, pp 11-20.<br />
368
Caroline Pade-Khene and Dave Sewry<br />
Clarke, A. (1999) Evaluation Research: An Introduction to Principles, Methods and Practice (1e), London: SAGE<br />
Publications.<br />
Dymoke-Bradshaw, C. and Cox, B. (2003) “Developing Organisation-technology Fit: The Role of Evaluation in IS<br />
Innovation Projects”, Electronic Journal of Information Systems Evaluation, Vol 2, No.1.<br />
Fitzerald, G. (1998) “Evaluating Information Systems Projects: A Multidimensional Approach”, Journal of Information<br />
Technology, Vol 13, pp 15-27<br />
Gigler, B. (2004) “Including the Excluded – Can ICTs Empower Poor Communities? Towards an Alternative<br />
Evaluation Framework Based on the Capability Approach”, Paper read at 4 th International <strong>Conference</strong> on the<br />
Capability Approach. University of Pavia, Italy, September.<br />
Harris, R. W. (2001) “Using Stories for Evaluating Community Based Information Communication Technologies in<br />
Developing Countries”, Part of the report of the PANAsia Telecentre Learning & Evaluation Group (PANTLEG),<br />
[online], http://rogharris.org/Using_Stories.pdf.<br />
Hedman, J. and Borell, A. (2005) “Broadening Information Systems Evaluation through Narratives”, Electronic<br />
Journal of Information Systems Evaluation, Vol 8, No.2, pp 115-122.<br />
Heeks, R. and Molla, A. (2008) “Compendium on Impact Assessment of ICT-for-Development Projects”, Working<br />
Paper Series, No. 36, Institute for Development Policy and Management, Manchester.<br />
Hudson, H. E (2001a) The Acacia Programme: Developing Evaluation and Learning Systems for African Telecentres.<br />
in: Latchem, C. and Walker, D. (eds). Telecentres: Case Studies and Key Issues, The Commonwealth of<br />
Learning, Vancouver.<br />
Hudson, H. E (2001b) Telecentre Evaluation: Issues and Strategies. in: Latchem, C. and Walker, D. (eds).<br />
Telecentres: Case Studies and Key Issues. The Commonwealth of Learning, Vancouver.<br />
Kozma, R., McGhee, R., Quellmalz, E. and Zalles, D. (2004) “Closing the Digital Divide: Evaluation of the World<br />
Links Program”, International Journal or Educational Development, Vol 24, pp 361-381.<br />
Lannon, J. (2007) “Evaluating Online Human Rights Resource Centres”, Paper read at 16 th International <strong>Conference</strong><br />
on Information Systems Development (ISD2007),Galway, Ireland, August.<br />
Macome, E. and Cumbana, C (2002) “Evaluation Study of the Manhica and Namaacha Telecentres after 3 years of<br />
operations”, [online], Eduardo Mondlane University Informatics Centre,<br />
http://portal.unesco.org/ci/en/files/13411/10679419181Needs_assessment_MOZ.pdf/Needs%2Bassessment%2<br />
BMOZ.pdf.<br />
Moodley, S (2005) “The Promise of E-Development? A Critical Assessment of the State of ICT for Poverty Reduction<br />
Discourse in South Africa”, Perspectives on Global Development, Vol 4, No.1.<br />
Pade, C. and Sewry, D. (2009) “The Practice and Need for Rural ICT for Development Evaluation: An Experience of<br />
the Siyakhula Living Lab Baseline Study”, Paper read at 3 rd International Development Informatics Association<br />
(IDIA) <strong>Conference</strong>. Digitally Empowering Communities: Learning from Development Informatics Practice.<br />
Kruger National Park, South Africa, October.<br />
Pade, C., Mallinson, B. and Sewry, D. (2008) “Project Management Practice for Rural ICT Project Sustainability in<br />
Developing Countries”, Paper read at 2 nd <strong>European</strong> <strong>Conference</strong> on Information Management and Evaluation<br />
(ECIME), University of London, Royal Holloway, UK, September.<br />
Rossi, P. H., Lipsey M. W. and Freeman, H. E. (2004) Evaluation: A Systematic Approach (7e), Sage Publications,<br />
London.<br />
Scharffenberger, G. (1999) Telecentre Evaluation Methods and Instruments: What Works and Why? In Gomez, R.<br />
and Hunt, P. (eds), Telecentre Evaluation: A Global Perspective. International Development Research Centre<br />
(IDRC), Ottawa.<br />
Schwalbe, K. (2006) Information Technology Project Management (4e), Thomson Course Technology, Boston,<br />
Massachusetts.<br />
Tacchi, J., Slater, D. and Lewis, P. (2003). Evaluating Community Based Media Initiatives: An Ethnographic Action<br />
Research Approach. Paper read at the Information Technology for Development <strong>Conference</strong>, Oxford, United<br />
Kingdom, July.<br />
Wagner, D. A., Day, B., James, T., Kozma, R. B., Miller, J. and Unwin, T. (2005) Monitoring and Evaluation of ICT in<br />
Education Projects: A Handbook for Developing Countries, Washington DC: The World Bank (InfoDev).<br />
369
User Adoption of the New Health Information System -<br />
Individual and Organizational Determinants<br />
Bahlol Rahimi<br />
Urmia University of Medical Sciences, Iran<br />
bahlol.rahimi@umsu.ac.ir<br />
Abstract: The computer has become an integral part of healthcare delivery by providing tools for improvements in<br />
the quality of care. However, the failure rate for information systems implementation in healthcare organizations will<br />
pose a great challenge in the area. In this paper, we aim to discuss user adoption of the new health information<br />
system in healthcare organization. In this study, we review the literature in order to illustrate previous studies that<br />
present failures in implementing new IT-based systems in healthcare settings. Next, we discuss and apply theory to<br />
interpret the results. The results demonstrate some important aspects in the implementation of new health<br />
information system such as considering user involvement in the design and implementation phase of the system,<br />
taking into consideration and differentiating the different professional needs when developing the system, user<br />
training methods, and to consider work routine changes due to implementation of the new system. Our findings show<br />
that evaluations of health IT-based systems seem to have not been successfully implemented in a number of cases,<br />
and have failed to demonstrate improvements in patient care, operating costs savings, and improvements in<br />
productivity. This study shows that, in fact, we have not learned from the past. It can be concluded that the<br />
application of a participatory design approach in health IT-based systems may be fruitful for the organization. The<br />
involvement of the end users in the design and implementation process will allow for the collection of their ideas, thus<br />
benefiting the developing system.<br />
Keywords: user adoption, health information system, user involvement, participatory design<br />
1. Introduction<br />
The computer has become an integral part of healthcare delivery by accelerating organizational and<br />
provider-patient communication as well as providing tools for improvements in the quality of care (Rahimi<br />
& Vimarlund 2007). However, the failure rate for information systems implementation in healthcare<br />
organizations will pose a great challenge in the area (Kaplan 2000). This is in part because the big efforts<br />
that policy and decision makers do for implementation of those system such as huge investment (Poon et<br />
al 2004). In particular, the majority of attempts to introduce information systems initiatives into both the<br />
private and public sectors, including healthcare information systems that require data entry by healthcare<br />
practitioners, have been unsuccessful (Van Der Meijden et al 2003; Adi 2000).<br />
However, it cannot be ignored that a characteristic of many public sector reforms is that they are<br />
introduced into organizations that are not willing to accept the reform, have the technical ability to<br />
understand and implement the change, or possess the ability to maintain the change once it has been<br />
introduced. As a result, the reforms are often severely delayed or create distortions that have damaging<br />
effects, leading to changes that are often ultimately abandoned (Diamond & Khemani 2005).<br />
In light of the abundant experience available, it is difficult to understand why the frequency of<br />
implementation failures is so high. For example, it has been determined that half of all computerized<br />
information systems fail largely due to user resistance (Anderson 1997). Likewise, Paré et al (1998)<br />
showed that many healthcare organizations have devoted significant amounts of money and frustrated<br />
countless people in wasted efforts to introduce new information systems.<br />
These failures are alarming, since the introduction of new, large healthcare information systems require a<br />
major investment (Adi 2005). In order to have successful and sustainable information technology in a<br />
healthcare setting, many factors need to be considered when introducing the new technology. One factor<br />
that has been indicated as crucial for success is the involvement of clinicians and other staff who will use<br />
the healthcare information systems throughout the planning and preparation stages of the design<br />
(Sengstack & Gugerty 2004; Birkmann et al 2006; Nykänen & Karimaa 2006). In general, regular and<br />
open communication between designers and users allows for the understanding of user preferences,<br />
their needs, their problems, and confusions. Recently, participatory design has been employed in<br />
different fields (Joyes & Chen 2007); however, as for health informatics, which is a more recently<br />
developed field, the use of participatory design has a shorter history (Pilemalm & Timpka 2007). Due to<br />
obvious problems in introducing integrated computer-based health applications, utilizing new approaches<br />
such as participatory design can make a difference where traditional methods have failed, in particular,<br />
with regards to the social factors that are of importance for the successful implementation of an<br />
370
Bahlol Rahimi<br />
information system (Pilemalm & Timpka 2007). Participatory design, which explicitly advocates active<br />
user participation throughout the design process, is a proactive design method.<br />
In this paper we aim to discuss the underlying factors, mostly individual and organizational, for user<br />
adoption of new health information systems in healthcare organization.<br />
2. Theoretical background<br />
2.1 Participatory design<br />
The participatory design approach was introduced in Scandinavia in the 1970s (Schuler & Namioka<br />
1993). It started in the field of computer software design, but similar approaches have been used in other<br />
fields as well (Sjöberg & Timpka 1998). Various studies have found that involving users in the system<br />
development process is an effective way to increase their adoption of the system (Faber 2003). In this<br />
respect, participatory design can be an appropriate approach for promoting user involvement.<br />
Participatory design was initiated as a resource to empowering system users by involving them in the<br />
design of an important part of their work setting. By involving of healthcare staff in the design and<br />
development processes of health information systems, it is said that the usability of the systems as well<br />
as their success increase (Van der Meijden et al 2001). Furthermore, it is important to consider the<br />
attitude of end-users -instead of one single viewpoint on a system’s requirements- in the process of<br />
development in order to increase final acceptance (Vimarlund & Timpka 1998).<br />
In addition, users are able to redesign and evaluate their work routines by applying the experience<br />
obtained during the participation process, and are even provided with the opportunity to improve their<br />
understanding of computers (Vimarlund & Timpka 1998). However, another objective for participating in<br />
the process of participatory design is based on recognizing the fact that it gives designers new and better<br />
ways of gaining an understanding of the users’ everyday working practices. Additionally, users are<br />
expected to be more willing to accept the final system once it is introduced, when they have assisted<br />
developers to arrive at a more accurate and realistic model, and reduces their aversion to the new<br />
system. The final overall effect of this participation can be considered to be the main incentive for<br />
improving work efficiency and productivity. It is argued that participation helps users to enhance their<br />
skills, thereby increasing the quality of services provided.<br />
To have an effective and credible health information system, user involvement in the design and<br />
development processes influences the perception, reception, and ultimately, the success of the new<br />
system (Wild et al 2004).<br />
2.2 Social capital<br />
The political scientist Robert D. Putman (1993) explained social capital as a theory about the<br />
complexities of trust and participation (Putman 1993a). The concept refers to features of social<br />
organization such as norms and networks that can improve the efficiency of society by facilitating<br />
coordinated actions (Putman 1993b). The social capital vocabulary has expanded, and today it<br />
addresses broadly the value of social networks both between similar and different groups of people by<br />
norms of reciprocity. The narrowest concept of social capital however, is still associated with Putman<br />
(Putman 1993a, 1993b). He views it as a set of “horizontal associations between people”, and argues<br />
that social capital encourages the formation of social networks and associated norms that have an effect<br />
on the productivity of the community and it facilitate the coordinated cooperation for the mutual benefit of<br />
the members of this “association” (Putman 1993a). Prior to Putman’s study, the concept of social capital<br />
had been used in community studies to describe relational resources embedded in personal ties within a<br />
community (Cohen & Prusak 2001). Nowadays, it is being applied in a wide range of studies in different<br />
disciplines, from anthropology and economy to information systems. However, despite the variety of<br />
definitions available for social capital, there is a consensus about the fact that social capital consists of<br />
those specific processes among people and organizations, which arise from working collaboratively in an<br />
atmosphere of trust, leading to the accomplishment of a goal with mutually shared benefits (Putman<br />
1993b). The benefits of social capital have, however, been identified and considered as “optimal” when<br />
social capital is created by individuals who learn to trust one another so that they are able to make<br />
credible commitments and rely on generalized forms of reciprocity, contributing to an increase in society’s<br />
productive potential (Ostrom 2000).<br />
371
Bahlol Rahimi<br />
To summarize, social capital consists of individuals forming and utilizing social networks in communities.<br />
Like financial capital, social capital is a resource that helps to sustain a community and encourages the<br />
collaboration and cooperation between members of groups for mutual benefits (Preece 2002).<br />
Introducing a new, healthcare information systems in which all users are involved and where networks<br />
are used for collaboration at different levels of the healthcare setting environment, are examples of such<br />
assets. Furthermore, the networks are usually groupings of people with common interests.<br />
2.3 Human capital<br />
Human capital is defined as “the skills, knowledge, and experience possessed by an individual or<br />
population, viewed in terms of their value or cost to an organization or country” (McKean 2005) and “the<br />
skills, general or specific, acquired by an individual in the course of training and work experience”<br />
(Smullen & Hand 2005). The main components of human capital, as stated in the literature include<br />
qualifications and knowledge acquired through formal education as well as skills, competencies, and<br />
expertise acquired through on-the-job training. Several authors have recognized the importance of<br />
human capital as a firm’s only appreciable asset due to the fact that the individuals’ productivity improves<br />
with literacy. It is said that workers may increase their productivity at the workplace by learning new skills<br />
and improving old ones (Vimarlund & Timpka 1998). The introduction of new technologies is likely to<br />
require the employment of a highly educated and skilled staff. There is, however, substantial evidence<br />
that education and training have strong positive effects on the accumulation of physical capital (Blundell<br />
et al 1999). In the case of implementing an healthcare information systems, acquiring knowledge and<br />
experience through formal and informal training becomes important and any shortfalls in the program<br />
raise many problems such as users’ complaints and resistance against the system.<br />
3. Methods<br />
In this study, I first review the literature in order to illustrate previous studies that present failures in<br />
implementing new IT-based systems in healthcare settings. Next, I discuss and apply the concepts just<br />
discussed to interpret the results.<br />
A literature review of published evaluation studies of IT-based systems in healthcare during the last five<br />
years (from 2003 to 2007) was performed. Linköping University’s database was used to gain access to<br />
articles on this subject. The keywords used in the search were: patient records, medical records, health<br />
records, information technology, medical informatics, healthcare information, health informatics, hospital<br />
information system, patient care information system, information system/technology, and evaluation<br />
study. I included papers that demonstrated a failure in implementing IT-based systems as well as some<br />
studies that discussed success and failure factors affecting the systems.<br />
4. Context of the study<br />
Study context of the papers including in this study has been mostly the Östergötland County council in<br />
Sweden, where tax-financed healthcare services are provided to the residents by the county council.<br />
Sweden has a decentralized healthcare system, with 20 county councils and 290 municipal councils as<br />
principals and care providers. Their responsibility as principals includes the provision of adequate care<br />
services and the requirement to develop, finance, and assure quality of all care activities. Implementation<br />
of a new integrated computerized patient record system was initiated in 2007 as a pilot project at a<br />
healthcare centre in the west part of the county, Motala. The implementation process continued from the<br />
west part (Motala) to the east part (Norrköping) of the county and was finished by the end of 2008. This<br />
new integrated system, developed commercially, provides a comprehensive overview of the patient’s<br />
health conditions and care. The system provides an infrastructure for sharing patient data between all<br />
healthcare care providers within the county council.<br />
5. Results<br />
A literature review was performed for evaluation studies of IT-based systems in healthcare. In our review<br />
we found that during the period 2003– 2008, most of the evaluation studies aimed to include issues such<br />
as the effectiveness of the systems, the quality of care, user and patient satisfaction, and the system’s<br />
usability. Based on the review literature, introduction and use of the computerized patient record systems<br />
was found to have positive effects such as economic benefits, high acceptance score and satisfaction<br />
among the users in the implemented sites and also improvements in management and work process.<br />
Moreover, introducing the telemedicine systems was found to have positive effects such as reducing<br />
spent time per patient during the visiting by clinical staff, economic benefits, and also quality of care.<br />
Regarding to the decision support system’s studies, it can be seen that introducing a clinical decision<br />
372
Bahlol Rahimi<br />
support systems in healthcare organizations had positive effects such as improved quality of care,<br />
satisfaction among users in the implemented sites, and also improvements in management and work<br />
process.<br />
The data captured in this study to show importance of evidence-based information system<br />
implementation in healthcare, an explorative case study design based on a single case was used for data<br />
collection and analysis. In this study the data were collected during a period of four months through<br />
interviews and document analyses. In the first step, representatives from all professional categories<br />
(physician, nurse, social worker, administrator, and pharmacist) using the new integrated computerized<br />
patient record system were interviewed by a member of the studied county council. Then 34 interviews<br />
conducted to validate our first results. Each interview session lasted about two hours. All documents<br />
published by the county council and the local magazines and newspapers that mentioned the system<br />
were also reviewed.<br />
In this study, different actors’ perspectives were considered when collecting data and analyzing them. We<br />
have used physicians, pharmacist, and nurses and other healthcare staff’s perspectives for our analysis.<br />
Considering different actors’ perspectives are needed to study in order to understand the likely<br />
accomplishment of new ICT goals in which it built in the area of health informatics. We categorized our<br />
finding into three groups that are: informatics skills, design interaction, and expectation.<br />
5.1 Informatics skills<br />
The data showed the nurses and other non-physician staffs were particularly unsatisfied, because they<br />
felt that the training sessions were based mostly on physicians’ needs. It was found that a failure to give<br />
all groups of users’ adequate training in using the integrated computerized patient record system<br />
negatively impacted the outcome of the implementation process. For instance, because the nurses had<br />
not learned to use the system functions properly, they found that the new practice routine was time<br />
consuming. Once the system was implemented, ongoing support was reported to be crucial for the<br />
success of the newly implemented system. They asked for the option of further training in order to<br />
overcome day-to-day problems.<br />
5.2 Interaction design<br />
The first technical problem was that logging on to the integrated system was perceived as consuming too<br />
much time. Then, after logging on to the system, several functions were found to be unintuitive and not<br />
user-friendly, causing dissatisfaction and disappointment. With the new system, calling up a specific file<br />
consumed more time than the previous system. The integrated system also required use of new terms<br />
and concepts, and the users emphasized that learning these took time.<br />
5.3 Expectation<br />
In the case study setting, the users expressed that more user participation in the design and<br />
implementation phase of the system would have provided a better fit into workflows and work practices.<br />
There was a general unwillingness to adapt clinical routines to the new system. The main adjustment of<br />
the implementation process that the users – especially physicians – asked for was “more involvement in<br />
the decision procedures”. More user involvement would both have helped define the system<br />
requirements in more detail and revise work practices to better integrate the new system. The<br />
respondents also made complaints about the timing of the implementation at the pilot site. They felt that<br />
the policy-makers had decided to implement the system in too short a time period, causing problems with<br />
adjustments, mainly in learning terms and navigation routines.<br />
5.4 What do end-users say is missing?<br />
The consequences of this trial seem to indicate that medical informaticians must learn from the past in<br />
order to communicate and adapt their own practices according to experience and research. The results<br />
demonstrate the lack of some important aspects in the implementation of a new system:<br />
1. User involvement in the design and implementation phase of the system would provide better<br />
insight into existing workflows and work practices. From the users’ point of view, to better integrate<br />
the new system, involving users in the system design and implementation phase would have avoided<br />
some human-computer interaction problems as well as training problems.<br />
373
Bahlol Rahimi<br />
2. In fact, perhaps the most important question is concerned with the way in which the<br />
implementation process can be adapted to different professional needs. It is clear that different staff<br />
members in the hierarchy of a healthcare organization demand different interfaces and modules. For<br />
example, taking into consideration and differentiating the nurses’ needs from the physicians’ needs,<br />
when developing the system, become important in order to achieve a good adoption rate for the new<br />
system. To understand how the implementation of a new system can be adapted to a large variety of<br />
user needs and expectations seems to be an important step towards facilitating user adoption of<br />
ICPRs.<br />
3. Staff complained that the training provided was not based on their need.<br />
4. The changes made to the current work routine due to the new system were also a source of<br />
resistance against the adoption of the system.<br />
5. Additionally, the results show that one of the important factors influencing the outcome of health ITbased<br />
systems implementation is the design of the user-training programs. Users’ inability to<br />
effectively interact with the new system caused secondary issues, leading to opinion that the health<br />
IT-based systems were not appropriately designed.<br />
6. Discussion<br />
In this study we found that, most of the evaluation studies aimed to include issues such as the<br />
effectiveness of the systems, the quality of care, user and patient satisfaction, and the system’s usability.<br />
A key requirement for analytical work is a clear definition of what constitutes an IT-based application.<br />
Today there are different terminologies relating to similar systems. Analysis of the effects has<br />
predominantly taken place at an organizational level. However, this does not take account of the<br />
differences in the settings in which analysis is taking place.<br />
It has been difficult to find generalized models and methods to evaluate IT-based applications in clinical<br />
settings that cover all such aspects as economic and both inter-organizational and intra-organizational<br />
approaches. The review reported in this study found that no standard framework exists for evaluating or<br />
developing evaluations and obtaining clearer and more exact feedback about the implemented systems’<br />
effects, or about the outputs of implementation and the use of ICT in healthcare settings. It also found<br />
that no previous studies have explored the impact of ICT on the healthcare systems’ productivity and<br />
effectiveness.<br />
A significant barrier to investment in ICT in healthcare is the widely recognized fact that any cost saving<br />
resulting from technology changes is not always seen by the implementer, but is rather passed on to a<br />
third party. In essence, benefits appear at one site and in one budget, while a large share of the cost<br />
commitments appear at another site and in another budget. To our best knowledge, the evaluation<br />
studies performed do not include any discussion about this important issue, or how lack of incentive to<br />
adopt systems can influence the organization and its personnel. Further, the potential effects of the<br />
implementation of IT-based applications are identified without analyzing them from an inter-organizational<br />
and economic perspective (Nykänen & Karimaa 2006).<br />
The case study comparisons with the literature review found both a recurrence of previously reported<br />
implementation problems and the development of new issues specific to the integrated system context.<br />
Possibly, the most important challenge in the case study setting concerned the way in which the<br />
integrated computerized patient record system implementation process could be adapted to the needs of<br />
different user groups. The second important challenge associated with it was the human-computer<br />
interaction consequences of the large-scale technical integration of sub-systems into a homogenous<br />
infrastructure. In this study re-experiences of known implementation problems were found. The fact that<br />
users’ training were based on physicians’ needs and not adjusted to nurses and other non-clinicians was<br />
a major source of complaints. From the case study setting, requests were expressed for user involvement<br />
in the design and implementation phase of the system, in order to provide better insights into existing<br />
workflows and work practices (Joyes & Chen 2007).<br />
A HIS system with integrated clinical decision support can be an advantage for the busy clinician who<br />
must combine and manage an increasing body of clinical knowledge. However, such support will not be<br />
optimal if clinicians begin to trust these systems without questioning the assistance (Pilemalm & Timpka<br />
2007). Recent research on safety in man-machine interaction suggests that the presence of<br />
environmental cues reflecting hazards increases alertness among decision-makers and reduces the risk<br />
of mistakes. From this perspective, it is positive that the system users had doubts about the reliability and<br />
374
Bahlol Rahimi<br />
completeness of the support provided by the system. Alertness is particularly important in light of the<br />
results of several previous studies that reported that HIS systems led to a number of errors and adverse<br />
drug events.<br />
Our findings show that evaluations of health IT-based systems seem to have not been successfully<br />
implemented in a number of cases, and have failed to demonstrate improvements in patient care,<br />
operating costs savings, and improvements in productivity. This study shows that, in fact, we have not<br />
learned from the past. Weng et al. (2006), in their paper aiming to apply participatory design for the<br />
creation of a highly usable collaborative protocol writing system for a national cancer clinical trial protocol<br />
authoring organization, show that after applying participatory design, users liked the system design. The<br />
users also liked the fact that some features of the system could speed up the reviewing process, the<br />
interoperability between the new system and the old tools, the integration, and the “all-in-one” design.<br />
The system’s usability was demonstrated in their results. Their study concluded that they could reduce<br />
usability errors of healthcare information systems with concrete approaches to participatory design.<br />
Therefore, before draining further the scarce resources available to healthcare organizations, it is<br />
necessary to develop strategies that will involve the end-users of the new system and ask for their<br />
assistance in accelerating the organizational change as well as work routines. Vimarlund et al (2001), in<br />
their paper discuss the economic impacts from the use of participatory design for the development of<br />
public-health information systems. They argue for the need of a method that minimizes the information<br />
asymmetry in the development process, thus avoiding market failures. They conclude that participatory<br />
design is beneficial for the public-health organizations involved in information system development<br />
processes by avoiding the risk of rejection of the final system, while increasing knowledge, decreasing<br />
information asymmetry, updating habits and work routines in a “natural manner”, avoiding unnecessary<br />
sunk-costs, and contributing to the efficient use of human and economic resources (Weng et al 2007).<br />
The challenges we have described first and foremost highlight a lack of collaborative approaches in the<br />
implementation of HISs. For instance, many of the most significant failures emerge from the absence of<br />
feedback from end users to developers during the development process. The logical conclusion is to<br />
increase the involvement of end users in the implementation process in order to provide:<br />
Designers with new and better ways of gaining an understanding of users’ everyday work practices.<br />
Users with the opportunity to redesign and evaluate their work routines by applying experience<br />
obtained during the participation process<br />
Users with the opportunity to appreciate the HIS as a useful tool in their everyday working life<br />
In fact, what is needed is the use of an implementation methodology that minimizes the information<br />
asymmetry in the implementation process, and allows the accumulation of the knowledge needed to<br />
prevent rejection of the final system. To compare with another field, a participatory design approach has<br />
recently been used in individual education, particularly with an increasing interest in experiential- and<br />
learner-centered education; users take responsibility for their own learning. In other words, there has<br />
been some recognition that learners as well as teachers should participate in deciding what and how the<br />
teachers are expected to teach (Joyes & Chen 2007).<br />
Failure in information systems implementation may originate from different sources, such as development<br />
failures and information system usage problems. User participation, better modeling tools, and better<br />
management of the developing project are means that could be used to reduce these risks in information<br />
systems projects. When user participation was emphasized during the design process, this showed to be<br />
a factor for a successful design and resulted in positive user satisfaction (Nykänen & Karimaa 2006).<br />
Using a participatory design approach in health IT-based systems design and implementation allows for a<br />
group of people with a common interest to exchange ideas and experiences within the group. It seems<br />
that during their collaboration and exchange of experiences and skills, the staff members widen their<br />
knowledge, or, in other words, increase the human capital. The participatory design approach, through<br />
the involvement of end-users in the design and implementation processes, provides them the opportunity<br />
to learn how to work and collaborate with one another in a new and common subject. They learn how to<br />
use each other’s ideas and viewpoints to improve the quality of work. Moreover, by exchanging their<br />
experiences with each other, they obtain new experiences. In a broader context, they might employ this<br />
new experience and knowledge in their future work in the organization. It can be claimed that in this way<br />
the human capital is increased.<br />
375
Bahlol Rahimi<br />
From another perspective, despite the importance of information and communication technologies for<br />
knowledge sharing, it has been known for a long time that the creation and diffusion of knowledge within<br />
organizations relies on the development of social networks, a shared system of meanings, and cultivation<br />
of shared values and norms. As part of this emphasis on the social aspects of knowledge creation and<br />
sharing, considerable attention has recently been directed toward the role played by social capital<br />
(Vimarlund et al 2001; Spender 1996). In fact, users may not have the time to exchange information<br />
during the ordinary work life, yet this kind of involvement brings the users together and creates the<br />
possibility for building a network. From a social capital perspective, the individuals’ involvement in the<br />
development process can become more important because such a system allows on-the-job IT-training<br />
as well as the creation of a new social network, thereby incorporating new knowledge and skills or<br />
improving old ones, while working. This harbors both, present and future organizational benefits, from the<br />
experience and specialization of employees (Vimarlund & Timpka 1998).<br />
The benefits of social capital has, however, been identified and considered as “optimal” when it is created<br />
by individuals who learn to trust one another so that they are able to make credible commitments and rely<br />
on generalized forms of reciprocity, contributing to an increase in a society’s productive potential. In our<br />
case, social capital is obtained through mutual relationships that are used to accomplish different joint<br />
activities. Like financial capital, social capital is a resource that helps to sustain a community and<br />
encourages collaboration and cooperation between members of groups for their mutual benefits<br />
(Huysman & Wulf 2004). To summarize, the introduction of a new health information systems in which all<br />
users are involved and where networks are used for collaboration at different levels of the healthcare<br />
setting environment, are examples of such assets. Further, the networks are usually groups of people<br />
with common interests. We believe there is a particular need to involve users and other experts in the<br />
early concept phase, where important decisions are made. It can be concluded that using a participatory<br />
design approach to involve end-users in the design of their own system or program, benefit both the<br />
individual and the organization. It can be concluded that the application of a participatory design<br />
approach in health IT-based systems may be fruitful for the organization. The involvement of the endusers<br />
in the design and implementation process will allow for the collection of their ideas, thus benefitting<br />
the developing system as well as the development of the work routines. In addition, a feeling of<br />
ownership could be reached through their participation in the design and implementation of the system.<br />
7. Conclusion<br />
The study was related to the topic of factors that influence the implementation and use of integrated<br />
HISs. The reason to focus on this topic was that though there had been a strong increase in<br />
implementation and use of information systems in healthcare setting, issues such as adoption, HISs<br />
effects, and factors influencing the implementation and use of integrated HISs still need to receive more<br />
attention and publication of such studies contribute to the emergence of an evidence-based health<br />
informatics which can be used as a platform for decisions by policy makers, executives, and clinicians<br />
Based on the result, it can be concluded that there is an increasing need to share knowledge and to find<br />
methods for evaluating the impact of investments and formulating indicators for success. It is therefore<br />
interesting to develop or extend evaluation methods that can be applied to this area with a multi-actor<br />
perspective in order to understand the effects, consequences, and prerequisites for the successful<br />
implementation and use of IT in healthcare.<br />
It can also be concluded that HISs be introduced to fulfil a high number of organizational, individualbased,<br />
and socio-technical goals at different levels. It is therefore necessary to link the objectives that<br />
these systems are designed to achieve with organizations’ short-term, middle-term, and long-term<br />
strategic goals. Another conclusion is that implementers and vendors have to direct more attention to<br />
what has been published in the area to avoid more failures in the future. The other conclusion is that if we<br />
want more evidence-based practice, we need more practice-based evidence.<br />
Acknowledgements<br />
I wish to thank Toomas Timpka, Vivian Vimarlund and Anna Möberg for their valuable help to conduct<br />
this study.<br />
References<br />
Adi, A. (2000) Healthcare information systems: challenges of the new millennium. Hershey, PA, Idea Group<br />
Publishing, USA.<br />
376
Bahlol Rahimi<br />
Anderson, J. G. (1997) “Clearing the way for physicians? Use of clinical information systems”, Communications of<br />
the ACM. Vol 40. No. 8, pp 83-90.<br />
Blundell, R. Dearden, L. Meghir, C. and Sianesi, B. (1999) “Human capital investment: The returns from education<br />
and training to the individual, the firm and the economy”. Fiscal Studies Vol 20, No. 1, pp 1–23.<br />
Birkmann, C. Demski, H. and Engelbrecht, R. (2006) “Introducing patient cards in clinical routine: evaluation of two<br />
research projects”. Methods of Information in Medicine. Vol 45, No. 1, pp 73-8.<br />
Cohen, D. and Prusak, L. (2001) “In good company: How social capital makes organizations work.” Harvard<br />
Business School Press, Boston Massachusetts.<br />
Diamond, J. and Khemani, P. (2005) Introducing financial management information systems in developing countries.<br />
IMF Working Paper, Fiscal Affairs Department, WP/05/196.<br />
Faber, M.G.(2003) “Design and introduction of an electronic patient record: how to involve users?” Methods of<br />
Information in Medicine. Vol 42, No. 4, pp 371-5.<br />
Huysman, M. H. and Wulf, V. (2004) “Social capital and information technology”. MIT Press: Cambridge.<br />
Joyes, G. and Chen, Z. (2007) “Researching a participatory design for learning process in an intercultural context”,<br />
International Journal of Education and Development using Information and Communication Technology, Vol 3,<br />
No. 3, pp 78-88.<br />
Kaplan, B. (2000) “Culture counts: How institutional values affect computer use”, M. D. Computing. Vol 17, No.5, pp<br />
23–26.<br />
McKean, E. (2005) “Human capital” in The New Oxford American Dictionary, second edition. Oxford University<br />
Press. Oxford Reference Online.<br />
Nykänen, P. and Karimaa, E. (2006) “Success and failure factors in the regional health information system design<br />
process--results from a constructive evaluation study”, Methods of Information in Medicine. Vol 45, No. 1, pp<br />
85-9.<br />
Ostrom, E. (2000) “Social capital and development projects.” Progress in Planning Vol 54, No. 2, pp 73-132.<br />
Paré, G. and Elam, J. J. (1998) “Introducing Information Technology in the Clinical Setting: Lessons Learned in a<br />
Trauma Center”, International Journal of Technology Assessment in Health Care, Vol 14. No.2, pp 331-343.<br />
Pilemalm, S. and Timpka, T. (2007) “Third generation participatory design in health informatics-Making user<br />
participation applicable to large-scale information system projects”, Journal of biomedical informatics, Vol 41,<br />
No. 2, pp 327-39.<br />
Poon, E. G. Blumenthal, D. Jaggi, T. Honour, M. M. Bates D. W. and Kaushal, R. (2004) “Overcoming barriers to<br />
adopting and implementing computerized physician order entry systems in U.S. hospitals”, Health Affaires<br />
(Millwood). Vol 23, No.4 , pp 184-90.<br />
Preece, J. (2002) “Supporting community and building social capital.” Communications of the ACM. Vol 45, No. 4, pp<br />
37-39<br />
Putman, R. D. (1993a) “The Prosperous Community: Social Capital and Economic Growth.” American Prospect, Vol<br />
13, No. 1, pp 35-42.<br />
Putman, R. D.(1993b) “Making democracy work: Civic traditions in modern Italy”, Princeton University Press,<br />
Princeton, NJ.<br />
Rahimi, B. Moberg, A. Timpka, T. and Vimarlund, V. (2008) “Post-implementation survey on using of integrated<br />
computerized patient record system”, MIE 2008 conference Proceeding, Gothenburg, Sweden, May.<br />
Rahimi, B. and Vimarlund, V. (2007) “Methods to evaluate health information systems in healthcare settings: A<br />
literature review”, Journal of Medical Systems. Vol 31, No. 5, pp 397–432.<br />
Schuler, D. and Namioka, A. (1993) “Participatory design: Principles and practices. Lawrence Erlbaum Associates,<br />
Hillsdale, NJ.<br />
Sengstack, P. P. and Gugerty, B. (2004) “CPOE systems: success factors and implementation issues”. Journal of<br />
Healthcare Information Management, Vol 18, No. 1 pp 36-45.<br />
Sjöberg, C. and Timpka, T. (1998) “Participatory design of information systems in health care”, Journal of the<br />
American Medical Informatics Association, Vol 5, No. 2, pp 177-83.<br />
Smullen. J. and Hand. N. (2005) “Human capital” in A Dictionary of Finance and Banking. Oxford University Press.<br />
Oxford Reference Online.<br />
Spender, J.C. (1996) “Organizational knowledge, learning and memory: three concepts in search of a theory.”<br />
Journal of Organizational Change Management, Vol 9, No. 1, pp 63-78.<br />
Van der Meijden, M. J. Tange, H. Troost, J. and Hasman, A. (2001) “Development and implementation of an EPR:<br />
how to encourage the user?” International Journal of Medical Informatics. Vol 64, No. 2-3, pp 173-85.<br />
Van Der Meijden, M. J. Tange, H. J. Troost, J. and Hasman, A. (2003) “Determinants of success of inpatient clinical<br />
information systems: A literature review”. Journal of the American Medical Informatics Association. Vol 10, No.<br />
3, pp 235–243.<br />
Vimarlund, V. Eriksson, H. and Timpka, T. (2001) “Economic motives to use a participatory design approach in the<br />
development of public-health information systems.” Medinfo. Vol 84, pp 768-72.<br />
Vimarlund, V. and Timpka, T. (1998) “Participatory design in economic terms: an economic discussion.” PDC 98<br />
proceeding of Participatory design conference. Washington USA. pp 11-17.<br />
Weng, C. McDonald, D. W. Sparks, D. McCoy, J. and Gennari, J. H. (2007) “Participatory design of a collaborative<br />
clinical trial protocol writing system.” International Journal of Medical Informatics. Vol 76, Suppl 1, pp 245-51.<br />
Wild, E. L. Hastings, T. M. Gubernick, R. Ross, D.A. and Fehrenbach, S. N. (2004) “Key elements for successful<br />
integrated health information systems: lessons from the States”. Journal of Public Health Management Practice,<br />
November (Suppl), pp 36-47.<br />
377
Relationship Between Organizational Factors and RFID<br />
Adoption in Iranian Gas Company Central Warehouses<br />
Farnaz Rahimi 1, 2 and Gholamabas Arabshahi 2<br />
1 Alzahra University, Tehran, Iran<br />
2 Gas company, Mashhad, Iran<br />
farnaz12200@yahoo.com<br />
a_arabshahi@ent.ut.ac.ir<br />
Abstract: This paper examined organizational factors that influence adoption of RFID (Radio Frequency<br />
Identification) technology in IRAN Gas company central warehouses. RFID has the distinct advantage of collecting<br />
data at some distance from the actual product, with no direct line of sight and without labor. Such a data automation<br />
system for real-time tracking, safety monitoring and overall warehouse operation leads to real-time visibility and<br />
tracking of assets and inventory that is necessary for a warehouse. In this paper we examine organizational factors<br />
(Structural: Size and resources and Centralization and Cultural: Innovativeness) associated with the adoption of<br />
RFID in gas company warehouses. Using a survey questionnaire we collected data from all 30 Gas company central<br />
warehouses employees(our unit of analyses) in IRAN .We found that organizational size do not show any<br />
relationship With RFID adoption and Centralization finantional recourses and Innovativeness are positively and<br />
significantly associated with the adoption of RFID . The implications of these and other findings are discussed.<br />
Keywords: organizational factors, warehouse, RFID<br />
1. Introduction<br />
Using radio waves, RFID technology can automatically identify objects. With RFID tags, objects can be<br />
tracked automatically by radio readers providing greater inventory visibility and improved business and<br />
control processes (Ngai et al.2007 Nystrom et al.2006) RFID’s main ability is to capture more data<br />
automatically without human intervention, in almost real time, so it can provide a more dynamic control<br />
environment for an organization.<br />
While RFID has been discussed in the literature as a technology that can provide several advantages:<br />
enhanced warehouse operations efficiency , better of inventory control and monitoring (McFarlane and<br />
sheffi 2003, Higgins,2006) , reduced labor costs (Kinsella,2003) , inventory obsolescence material<br />
handling costs reduction (McFarlane and sheffi 2003), quality control (Hoffman,2006), reductions in outof-stock<br />
and delivery safety stock (Kinsella,2003)uncertainty of product availability reduction (Asif and<br />
Mandviwalla 2005), to its adopters ,the RFID adoption rate is not growing as fast as expected (Nystrom et<br />
al.2006. Boucher 2007).This suggests more effort is necessary to understand the process of adoption of<br />
the technology and to identify factors affecting the RFID adoption decision (Ngai et al2008).<br />
According to Rogers et al. (1996) companies that used more warehousing information technologies had<br />
better performance in the areas of quality and productivity improvements and cycle time reductions than<br />
the low user group. Past researches show that, organizational factors often have significant influence on<br />
the process of technology adoption.(Ko et al.2000; Lefebvre et al. ,1991 ; Masters et al. ,1992<br />
;O'Callaghan et al. ,1992 ;Sullivan ,1990; Premkumar et al .1997).<br />
In this paper we study the relationships between organizational factors (size, financial resources and<br />
centralization) as structural characteristic and (organizational culture) and RFID adoption in the Gas<br />
company warehouses.<br />
The rest of the paper is organized as follows: Section 2 introduces RFID, and the literature review of<br />
RFID adoption studies .In Section3, we present the research model and hypotheses. This is followed by<br />
the description of the research methods used in data collection. Section 5 the findings will be presented.<br />
2. Literature review<br />
2.1 RFID<br />
RFID is a common term for technologies and systems that use radio waves to automatically identify<br />
people or objects.<br />
378
Farnaz Rahimi and Gholamabas Arabshahi<br />
RFID typically consists of three basic components: tags (transponders), readers, and middleware (Asif et<br />
al.2005 ; Preradovic.2007).<br />
A tag usually containing a microchip and an antenna that are attached to or embedded in an object<br />
(Curtin 2007) The microchip contains identification information and may have other application data (e.g.<br />
price, cost, location, and manufacture date, etc.)(Chao 2005)<br />
A reader (also known as an interrogator) sends out a radio signal and abets the tag to broadcast the data<br />
contained on its chip. The reader then change the radio waves returned from the tag into digital data and<br />
forward them to a computer system (High jump Software, 2004; Zebra Tech- neologies, 2004).<br />
2.2 Reviews of related studies<br />
Earlier studies on RFID adoption have focused on a variety of factors associated in adoption process:<br />
Hoske (2004) study the cost factor .Jones, Clarke-Hill, Shears, Hillier, and Comfort (2004b) looked at<br />
privacy and public policy with regards to RFID.<br />
Adam (2004) investigated benefits, costs, standards, and environment; four main factors associated<br />
RFID adoption.<br />
Jones M .et al. (2005) identified benefits, costs, standards, privacy, and the power of retailers as factors.<br />
In their study to identify the factors may influence RFID adoption in South African retail organizations,<br />
Brown and Russell (2007) found that the RFID adoption intention was explained by technological factors<br />
(i.e., relative advantage ,compatibility complexity ,and cost) ,organizational factors (i.e., top management<br />
attitude ,information technology expertise, organization size organizational readiness), and external<br />
factors(i.e., competitive pressure, external support, and existence of change agents.)<br />
Reviewed prior studies, Schmitt et al. (2007) categorized 25 adoption factors in the technology,<br />
organizational, and environment groups. They concluded five most important factors affecting RFID<br />
adoption and diffusion in the automotive industry: compatibility, costs, complexity, performance, and top<br />
management support. Most of these factors belong to the technology characteristics group.<br />
Based on previous literature, there are variety of factors related to technology adoption that we can<br />
categorize them into four groups: environmental, technological, inter-organizational and organizational.<br />
Each group contains different variables in different studies.<br />
In this research we analyze organizational factors comprising four variables discussed below:<br />
3. Research model and hypotheses<br />
3.1 Organizational size<br />
Organizational size has an important impact on the adoption of technological adoption (Kimberly and<br />
Evanisko, 1981; Fama and Jensen, 1983; Kennedy, 1983; Damanpour,1992)<br />
In some studies a positive relationship between firm size and adoption has been concluded (Tornatzky,<br />
1990;Grover, 1993;Premkumar1999; Damanpour, 1992; Moon and Bretschneider 1997).Because of the<br />
cost of RFID tags and other related equipments ,only large companies have the financial resources to<br />
invest in RFID(Wang et al,.2006.) In addition, greater formalization (Pugh et al 1969), greater task<br />
specialization (Blau1970), and more complex forms of communications (Haveman 1993) are<br />
characteristics; help larger organization adopt new technologies.<br />
Although counterarguments exist that smaller organizations have more flexibility due to fewer levels of<br />
bureaucracy, thus increasing their ability to adopt technology more quickly (Patterson et al 2003), few<br />
empirical results have shown that smaller firms tend to be more innovative (Salavou, et al. 2004)<br />
In this study, organizational size is defined as work force size (number of employees) (Kimberly and<br />
Evanisko 1981;Teo and Tan ,1998) and the number of items categorized in each warehouse.<br />
We expect that organization size will have a positive influence on RFID adoption:<br />
379
Farnaz Rahimi and Gholamabas Arabshahi<br />
H1: There is a positive relationship between size of organization in terms of number of employees and<br />
number of items in each warehouse and RFID adoption.<br />
3.2 Financial resources<br />
Financial resources of an organization are another predictors of innovation (Hage and Aiken 1967; Hurley<br />
and Hult 1998).<br />
Financial resources facilitate the adoption of technology and provide organizations with the capacity to<br />
accept risk inherent in investing in new technology (Grover and Goslar 1993).<br />
The measure we used in this study is the budget allocated for new technology acquisition.<br />
H2: There is a positive relationship between the amount of financial resources and RFID adoption.<br />
3.3 Centralization<br />
Centralization has been defined as the degree that power is distributed among all the positions within an<br />
organization (Hage and Aiken 1967; Schminke et al.2000).<br />
It consists of two dimensions :(Hage and Aiken 1967; Schminke, et al.2000)<br />
1)Participation in decision making which shows how much members of the organization, participate in<br />
organizational decisions(Hage and Aiken 1967), and 2) Authority hierarchy which refers to" the amount of<br />
autonomy employees have in relation to deferring to superiors when making decisions about day to day<br />
tasks" (Tata and Prasad 2004, p. 251)<br />
Rogers (1995) suggested that centralization is negatively associated with organizational adoption. This is<br />
empirically supported by prior studies (e.g Hage, 1969; Moch and Morse, 1977; Kimberly and Evanisko,<br />
1981; Pierce and Delbecq, 1977;Grover 1993; Grover and Goslar 1993)<br />
A reason for this relationship is that in a centralized structure, top management is not aware of<br />
operational- level problems and con not suggests relevant innovations to meet their needs. Another<br />
reason is that the more power is concentrated at the top by a few strong leaders, the more new ideas and<br />
innovations are restricted .( Yoon,2009)<br />
However, some researchers found a positive relationship between centralization and technology adoption<br />
.(Ettlie et al. 1984).They found that centralized decision making would facilitate faster and more efficient<br />
adoption because lower level management resistance would not impede top management support.<br />
Evidence has also supported that decentralized organizations are more likely to adopt technological<br />
innovations.<br />
In this study we ask the respondents to state how much they are interfered to every day decision making<br />
and how much their boss intervene to their tasks.<br />
Based on these arguments and previous findings, the following hypothesis is proposed in this research:<br />
H3: There is a negative relationship between the degree of centralization of the organization and<br />
adoption of RFID.<br />
3.4 Organizational culture<br />
There are over 54 definitions of organizational culture in literature. Verbeke et a l.(1998). “A system of<br />
shared norms and behaviors that are learned by members of the organization and shape their way of<br />
doing” (p. 313). Sutherland and Morieux (1988) defined organizational culture as “a set of emotional<br />
conditions, views of the past and standards of behavior belonging collectively to an organization (p. 43).<br />
Organizational culture can explain why some organizations adopt new technology quicker than their<br />
competitors. (Glisson and James 2002).<br />
380
Farnaz Rahimi and Gholamabas Arabshahi<br />
We ask the respondents about management’s attitude towards innovations and new ideas and also<br />
respondent willingness to search and seek innovations. (Hurley and Hult1998)<br />
H4: There is a positive relationship between innovative organizational culture and adoption of RFID.<br />
4. Research design<br />
The propositions in last Section formed the basis for carrying out the investigation. This study is an<br />
applicable-descriptive research. We use questionnaire which developed by adopting measures from<br />
several sources (Hurley and Hult 1998;Hage and Aiken 1967; Jaworski and Kohli1993; Matsuno et al.<br />
2002; Kimberly and Evanisko1981).<br />
A5-point Likert scale was used for each item, anchored by Strongly Agree at one end to Strongly<br />
Disagree at the other. The target population for this study was Gas company central warehouses<br />
employees in all the states in IRAN. Due to resource constraints, we chose a sample, using random<br />
sampling and mail the questioners to 145 respondents.120 questioner were returned and used for our<br />
analysis.<br />
4.1 Reliability<br />
Cronbach’s Coefficient Alpha in a small sample size was 0/93 which means the questioner has a high<br />
degree of reliability.<br />
5. Data analysis<br />
Using SPSS software, we specify Correlation Coefficient between RFID Adoption and other factors<br />
(organizational size, centralization, financial resources, and innovative culture) as follows:<br />
Table 1: Results of analyzed data<br />
RFID Adoptionorganizational<br />
size<br />
RFID Adoption -<br />
centralization<br />
RFID Adoption –<br />
financial resources<br />
RFID Adoption –<br />
innovative culture<br />
sample<br />
size<br />
120<br />
120<br />
120<br />
120<br />
Correlation<br />
Coefficient<br />
-0.069<br />
0.461<br />
0.208<br />
0.599<br />
Statistic<br />
0.75<br />
5.64<br />
2.31<br />
8.13<br />
6. Findings<br />
As mentioned in the table, the results are as followed:<br />
P-value<br />
0.452<br />
0/000<br />
0.023<br />
0.000<br />
α<br />
0.05<br />
0.05<br />
0.05<br />
0.05<br />
result<br />
there is no<br />
relation between<br />
RFID Adoptionorganizational<br />
size<br />
there is a positive<br />
relationship<br />
between RFID<br />
Adoptionorganizational<br />
size<br />
there is a positive<br />
relationship<br />
between RFID<br />
Adoptionorganizational<br />
size<br />
there is a positive<br />
relationship<br />
between RFID<br />
Adoptionorganizational<br />
size<br />
Organizational size does not have impact on RFID adoption. All other structural variables (finantional<br />
resources, centralization) and also innovative culture of the organization have a positive relationship with<br />
RFID adoption. So the propositions number 2 and 4 (financial resources and innovative culture) are<br />
proved and the other 2 hypothesis (size and centralization) rejected.<br />
These findings can guide the decision makers. It shows that consideration of innovativeness and<br />
allocating financial recourses to new technology can speed up the rate of adoption. Because Gas<br />
381
Farnaz Rahimi and Gholamabas Arabshahi<br />
Company is governmental in Iran, centralization is unavoidable. In such situations, decision makers<br />
should concern financial resources and innovative culture more.<br />
Acknowledgements<br />
We highly appreciate Khorasan Razavi Gas Co for their technical and financial support of this research.<br />
References<br />
Adam ,K.(2004). The South African retailers’ propensity to adopt radio frequency identification technology.<br />
Unpublished Honours. Technical Report, University of Cape Town<br />
Asif Z.,M .Mandviwalla, Integrating the supply chain with RFID: a technical and business analysis, Communication of<br />
AIS1 5(24), 2005,pp.393–426.<br />
Blau, P.M., A formal theory of differentiation in organizations. American Sociological Review 35,1970,pp.201–218.<br />
Boucher, R.,RFID adoption is lagging,eWeek.24(1)(2007)25.<br />
Brown, I. Russell, J. ,Radio frequency identification technology :an exploratory study on adoption in the South African<br />
retail sector ,Int. J. Inform .Manage.27.(4) (2007) 250-265.<br />
Chao ,C.C. Yang,J.M. Jen,W.Y. ,Determining technology trends and forecasts of RFID by a historical review and<br />
bibliometric analysis from1991to2005.Technovation27(5)(2007)268–279.<br />
Curtin ,J. Kauffman, R.J. Riggins, F.J., Making the ‘MOST’ out of RFID technology :a research agenda for the study<br />
of the adoption, usage and impact of RFID Inform.Technol.Manage.8(2)(2007)87–110<br />
Damanpour, F.,1992. Organizational size and innovation. Organization Studies 13,375 – 402.<br />
Ettlie, John E., William P. Bridges and Robert D. O’Keefe (1984), “Organizational Strategy and Structural Differences<br />
for Radical versus Incremental Innovation Management Science, 30 (6), 682-95.<br />
Fama,E. ,Jensen, M.,1983. Separation of ownership and control. Journal of Law and Economics 26,301–325.<br />
Glisson, Charles and Lawrence R. James (2002), “The Cross-level Effects of Culture and Climate in Human Service<br />
Teams,” Journal of Organizational Behavior, 23,767-94.<br />
Grover, V. (1993) An empirically derived model for the adoption of customer-based interorganizational systems,<br />
decision science , 24, 3, 603-640.<br />
Grover, V., and Goslar, M.D. (1993) The initiation, adoption, and implementation of telecommunications technologies<br />
in U.S. organizations, journal of management information systems, 10, 1, 141.<br />
Hage, Jerald and Michael Aiken (1967), “Relationship of Centralization to Other Structural Properties,” Administrative<br />
Science Quarterly, 12, 72-92.<br />
Hage, J. (1969) Routine technology, social structure, and organizational goals ,administrative science quarterly, 14,<br />
June, 1969, 368-379.<br />
Haveman ,H. A., Organizational size and change :diversification in the savings and loan industry after deregulation,<br />
Administrative Science Quarterly 38(1),1993,pp.20–50.<br />
Higgins.L.N. Cairney,T., RFID opportunities and risks ,Journal of Corporate Accounting &Finance 17(5),2006,pp.51–<br />
57.<br />
High jump Software,2004.The true cost of radio frequency identification (RFID).High jump Software Free<br />
Report.URL: <br />
Hoffman,W. Metro moves onRFID,TrafficWorld270(4),2006,p.18.<br />
Hoske, M.(2004).RFID: Adoption increases despite costs. Control Engineering, 51(7),46–47.<br />
Hurley, Robert F. and G. Thomas Hult (1998), “Innovation, Market Orientation, and Organizational Learning: An<br />
Integration and Empirical Examination,” Journal of Marketing, 62 (July), 42-54<br />
Jaworski, Bernard J. and Ajay K. Kohli (1993), “Market Orientation: Antecedents and Consequences,” Journal of<br />
Marketing, 57 (July), 53-70.<br />
Jones ,M. ,Wyld,D. ,&Totten,J. (2005).The adoption of RFID technology in the retail supply chain. The Costal<br />
Business Journal, 4(1)29-42.<br />
Jones,P.,Clarke-Hill, C.,Shears,P ,Hillier,D .,&Comfort,D .(2004). Radio frequency identification in retailing and<br />
privacy and public Policy issues.Management Research News, 27(8/9),46–56.<br />
Kennedy,A.M. ,1983. The adoption and diffusion of new industrial products: a literature review .<strong>European</strong> Journal of<br />
Marketing17,31–88.<br />
Kimberly, John R. and Evanisko , Michael J. (1981), “Organizational Innovation: The Influence of Individual,<br />
Organizational, and Contextual Factors on Hospital Adoption of Technological and Administrative Innovations,”<br />
Academy of Management Journal, 24 (4), 689-713.<br />
Kinsella.B.,The Wal-Mart factor,Industrial Engineer35(11),2003 ,pp. 32–36.<br />
Ko,E. Kincade, DH ,Brown JR. Impact of business strategy on the adoption of quick response technologies :the<br />
apparel industry experience .IntJOper ProdManage2000;20(9):1093-111.<br />
Lefebvre LA, Harvey J, Lefebvre E. Technological experience and technology adoption decisions in small<br />
manufacturing firms .R&D Manag 1991;21 (3):241-9.<br />
Masters JM ,Allenby GM, Lalonde BJ, Maltz A .On the adoption of DRP .JBus Logist1992;13(1):47-67.<br />
Matsuno, Ken, John T. Mentzer and Aysegul Ozsomer (2002), “The Effects of Entrepreneurial Proclivity and Market<br />
Orientation on Business Performance Journal of Marketing, 66 (3), 18-32<br />
Mc Farlane ,D.Sheffi,Y.,The impact of automatic identification on supply operations chain,International Journal of<br />
Logistics Management14(1),2003,pp.1–17.<br />
Moch, M., and Morse, E. (1977) Size, centralization and organizational adoption of innovations,Amer.social.Rev, 92,<br />
Oct, 1977, 716-725.<br />
382
Farnaz Rahimi and Gholamabas Arabshahi<br />
Moon,M.J.,Bretschneider,S.,Can state government actions affect innovation and its diffusion?:an extended<br />
communication model and empirical test Technol.Forecast.Soc.Change54(1)(1997)57–77.<br />
Ngai.E.W.T. ChengT.C.E. Au,S. Lai,K. ,Mobile commerce integrated with RFID technology in a contained<br />
depot,Decis.SupportSyst.43(1)(2007)62–76.<br />
Ngai,E.W.T. Moon,K.K.L. Riggins,F.Yi.J.,C.Y.,RFID research: an academic literature review(1995–2005) and future<br />
research directions, Int .J.Prod.Econ.112.520-510(2008)(2).<br />
Nystrom,N.C.Lin,H.A.Yu,H.C. Wu,M.A,Challenges to global RFID adoption, Technovation 26 (12) (2006)1317–1323.<br />
O'Callaghan R ,Kaufmann PJ, Konsynski BR. Adoption correlate and share effects of electronic data interchange<br />
systems in marketing channels. JMark 1992;56(2):45-56.<br />
Patterson, Kirk A., Curtis M. Grimm, and Thomas M. Corsi (2003), “Adopting New Technologies for Supply Chain<br />
Management,” Transportation Research Part E 39 (march),95-121.<br />
Pierce, J.L., and Delbecq, A.L. (1977) Organizational structure, individual attributes and innovation, journal of<br />
management information system, 2, 1, January, 27-37.<br />
Premkumar,G.,Ramamurthy,K.,Crum,M.,1997.DeterminantsofEDI adoption in the transportation industry .<strong>European</strong><br />
Journal of InformationSystems6,107–121.<br />
Premkumar,G.,Roberts,M.,Adoption of new information technologies in rural small<br />
businesses,Omega27(4)(1999)467–484.<br />
Preradovic ,S. ,N.C.Karmakar ,Modern RFID readers ,Microwave Journal 50(9)(2007).<br />
Pugh,D.S. Hickson,D.J.Hinings,C.R.Turner,C. ,The context of organizational structure,Administrative Science<br />
Quarterly 14 1969,pp.91–114.<br />
Rogers,dale.daugherty.s.pj.Elllinger.a.e."the relationship between information technology and warehousing<br />
performance".logestics and transportation review.vol 32.Desember 1996.<br />
Rogers, E. (1995),diffusion of innovations, (4th ed.) The Free Press, New York.<br />
Salavou, H., G. Baltas and S. Lioukas (2004), “Organisational Innovation in SMEs <strong>European</strong> Journal of Marketing, 38<br />
(9/10), 1091-112.<br />
Schmitt,P., Thiesse ,F.,E.Fleisch ,Adoption and diffusion of RFID technology in the automotive industry ,in: H.Osterle,<br />
J.Schelp ,R. Winter(Eds.),Proceedings of the15th <strong>European</strong> <strong>Conference</strong>on Information Systems, St<br />
Gallen,Switzerland,2007.<br />
Schminke, Marshall, Maureen L. Ambrose and Russell S. Cropanzano (2000), “The Effect of Organizational<br />
Structure on Perceptions of Procedural Fairness Journal of Applied Psychology, 85 (2), 294-304<br />
Sullivan PC. A study of the adoption of quick response in the United States apparel manufacturing industry<br />
.Unpublished doctoral dissertation .New YorkUniversity;1990.<br />
Sutherland, Ewan and Yves Morieux (1988), “Effectiveness and Competition – Linking Business Strategy,<br />
Organizational Culture and the Use of Information Technology,” JIT, 3 (1), 43-47<br />
Tata, Jasmine and Sameer Prasad (2004), “Team Self-Management, Organizational Structure, and Judgments of<br />
Team Effectiveness,” Journal of Managerial Issues16 (2),251.<br />
Tan, M., and Teo, T.S.H. (1998) Factors influencing the adoption of the Internet,international journal of electronic<br />
commerce, 2, 3, 5-18.<br />
Tornatzky .M.L.G., The processes of technological innovation, Lexington Books,Fleischer,1990.<br />
Verbeke, Willem, Marco Volgering, and Marco Hessels (1998), “Exploring the Conceptual Expansion within the Field<br />
of Organizational Behaviour Organizational Climate and Organizational Culture,” Journal of Management<br />
Studies, 35 (3), 303-29<br />
Wang,W., Yuan, Y., Wang, X., Archer, N.,RFID implementation issues in China: Shanghai port cast study ,J.<br />
InternetComm. 5(4) (2006) 89–103.<br />
Yoon,Tom.(2009).an empirical investigation of factors affecting organizational adoption of virtual world. ProQuest.<br />
Zebra Technologies, 2004 .RFID :the next generation of AIDC . Zebra TechnologiesWhitePaper.URL:<br />
<br />
383
Business Intelligence as Decision Support in Business<br />
Processes: An Empirical Investigation<br />
Ari Riabacke 1 , Aron Larsson 1,2 and Mats Danielson 1<br />
1<br />
Stockholm University, Kista, Sweden<br />
2<br />
Mid Sweden University, Sundsvall, Sweden<br />
ariria@dsv.su.se<br />
aron.larsson@miun.se<br />
mad@dsv.su.se<br />
Abstract: Our concern in this paper is the role of business intelligence systems and the perceived business value of<br />
implemented systems and their contribution to facilitate the fulfilment of organizational goals. The study builds upon a<br />
survey answered by 43 respondents from different large companies in Scandinavia. The survey used questions on<br />
how visions, objectives, strategies are supported by BI systems, on how business values are derived from such<br />
systems, and on how design and implementation issues affect the solutions. The overall conclusion of the study is<br />
that there are markedly different levels of problems in the areas, most problems being found in integration of BI<br />
information and decision processes, and that there is room for large improvements and further work within everything<br />
from implementation to requirements engineering for business intelligence decision support systems.<br />
Keywords: business intelligence, decision support, business value, empirical study<br />
1. Introduction<br />
During the latter decades organizational decision making has become increasingly complex due to the<br />
amount of accessible information that managers wish to incorporate in decision making activities has<br />
increased dramatically. Given the integration of information systems into business processes,<br />
organizational decision making may often experience a substantial amount of environmental pressure<br />
due to, for instance, information overload and sharper time constraints (Rasmussen 2000, Speier and<br />
Morris 2000). At present, one popular approach to deal with the increasing amount of data in<br />
organizations is to build data warehouses and BI-systems.<br />
Systems or methods for supporting decisions come in many forms. For example, decision analytical<br />
systems, focusing on supporting problem structuring and evaluations of alternative courses of action,<br />
optimization tools focusing on finding an optimal solution for a well-defined system of constrained<br />
variables, or business intelligence systems (BI) aiding in the gathering and analysis of business<br />
information. See (Turban et al. 2010) for a comprehensive treatment of business intelligence. In this<br />
respect, business intelligence systems may be viewed as information systems with special focus on<br />
providing accessible business data, i.e., they can be viewed as type of decision support system with the<br />
capability of (easily and quickly) providing reliable and up-to-date information or key figures about the<br />
organization.<br />
Since the 1990s, such BI systems typically employ three different technologies for supporting decision<br />
making in an organization: data warehouses (DW) for the gathering of business data, data mining (DM),<br />
and on-line analytical processing (OLAP) for data analysis (Shim et al. 2002). The benefits of a decision<br />
support system may further vary. Already in 1981, Keen summarized the potential benefits of a decision<br />
support system (DSS) as they enable the following: facilitate a more rational treatment of decisions<br />
problems; yield a better understanding of the business; yield faster response to unexpected situations;<br />
facilitate ad-hoc analyses; yield new insights and learning; improve communication; enhance control;<br />
enhance team work; reduce time spent on planning; and better use of resources (Keen 1981). This<br />
basically remains true even today. However, the decision support function enabling better decision<br />
making and other more advanced uses of a BI system such as facilitating business strategies is hard to<br />
measure (Watson and Wixom 2007). The results of a previous study presented in (Stenfors et al. 2007),<br />
according to managers the strongest incentives for investing in a DSS are to increase the effectiveness of<br />
operational processes and to support strategic work and decision making. These benefits are indeed of a<br />
more qualitative kind, and the decision to invest in a DSS is therefore often based rather on estimated or<br />
perceived values than on actual costs. Further, the results from the abovementioned study indicate that<br />
managers view DSS and BI systems as support instruments to business process management. However<br />
enabling this support does call for the BI system to be systematically aligned or integrated with business<br />
processes to a greater extent (Bogdana et al. 2009). The more recent and wider concept of BPM<br />
(business performance management) is related to business intelligence as the utilization of BI systems<br />
384
Ari Riabacke et al.<br />
can be viewed as an element of BPM, but where the use is supposed to be aligned with organizational<br />
strategy and determined by a program for how to implement this strategy (Frolich and Ariyachandra<br />
2006).<br />
But how does this need for a DSS manifest itself? Recent years have witnessed a remarkable increase of<br />
companies investing in BI systems, see (Hannula and Pirttimäki 2003). This makes it interesting to study<br />
the use and knowledge of the effects on the businesses of the companies utilizing, or having the ability to<br />
utilize, such systems. Further, studies of BI systems as a support for organizational decision making are<br />
relatively scarce. With respect to information systems in general, several studies have been carried out in<br />
order to identify factors behind a widespread use and end-user acceptance of a system within an<br />
organization. These factors are often user-related, such as user attitudes, their previous experiences, and<br />
user participation in development, but also context-related, such as top management support and<br />
conditions facilitating the use of information systems; see (Bajaj and Nidumolu 1998, Sabherwal et al.<br />
2006, Riabacke 2006).<br />
It is well-known that up to 50%, some say even 70%, of all BI implementations fail in one or more<br />
respects, not delivering the value-for-investment expected at the outset. While there are studies on the<br />
pitfalls of developing and realizing BI systems, this study focuses on systems already in operation, their<br />
organisational context, and hurdles on the path to efficient and effective use of BI information. We will, in<br />
the paper conform to the common use of the concept of system, denoting the technical part of a BI effort.<br />
From a systems science perspective, the system boundaries are indeed much wider, encompassing the<br />
organisation and its control mechanisms. Not to interfere with colloquial interpretations of the system<br />
concept, we will refer to this wider concept as the BI scope.<br />
BI is not about technology, it is much more. It is about providing an organisation with support for better<br />
decision making. Our hypothesis, from literature and interviews, is that far too much focus in the<br />
development and implementation of BI solutions in organizations is put on the technical parts of the<br />
system. Much of the effort is thus on the right hand side of the model shown in Figure 1, i.e. on the<br />
design and building of data warehouses, and on creating different types of reports. In comparison less<br />
effort seems to be put on structuring and understanding business processes and gaining insights into the<br />
decision making activities involved in these processes that are to benefit from the BI systems. In other<br />
words, the level of integration of BI systems into business processes could be much higher, yielding a<br />
better business value of the investments made in the system.<br />
Figure 1: Objects of interaction in the use of BI applications and systems<br />
Thus, the main area of concern in this paper is to what extent this view is supported by studying the<br />
perceived values of BI systems use and the views on system design and implementation.<br />
2. Study outline and method<br />
The study was conducted using a quantitative method based upon a questionnaire. Of 1085 persons<br />
considered for the survey and accessible through a large and collaborative consultancy firm, 105 were<br />
selected for participation. Of these 105, 43 completed the questionnaire, yielding an effective completion<br />
response rate of 41%. Thus, 43 respondents from different large Scandinavian companies employing BI<br />
systems for decision making support, having a leading role in these, fully answered the questionnaire in<br />
which they had to estimate how well different statements regarding their BI solutions corresponded to<br />
385
Ari Riabacke et al.<br />
their own perception of the reality in their own business. The respondents came from various business<br />
segments according to Table 1.<br />
Table 1: Respondents’ business segments<br />
Business segments Count<br />
IT 5<br />
Logistics 4<br />
Banking and insurance 15<br />
Real estate and construction 1<br />
Consumer products 2<br />
Energy 3<br />
Pharma and health care 1<br />
Industry 6<br />
Public sector 4<br />
Other 2<br />
TOTAL 43<br />
Table 2: Respondent’s role (some have more than one role)<br />
Role Count<br />
CEO 3<br />
CIO 3<br />
CFO 6<br />
Other management 3<br />
Head of Business Segment 4<br />
IT Management 7<br />
System owner 1<br />
Requirements officer 3<br />
System architect 4<br />
Developer 4<br />
Other 10<br />
Total 48<br />
The questions were divided into six categories 1 to 6, each comprising of six questions. The categories<br />
are:<br />
1: Visions, objectives, and strategies<br />
2: Business values from BI systems<br />
3: Requirements analysis and needs<br />
4: Change management<br />
5: Technical solutions<br />
6: Decision making support<br />
Thus, each category consists of 6 statements which the respondents were asked to assess. The<br />
respondents indicated how much they agreed with each statement on a 1-to-7 Likert scale, where 1 was<br />
strongly disagreeing and 7 was strongly agreeing. The statements were designed so that a higher value<br />
indicated better integration of BI into the business processes. One example of a statement in category 1<br />
is “BI has a role in our strategic processes”. See the appendix for the complete set of statements. For<br />
each category, the aggregated results (modal scores) of all respondents are shown in two spider<br />
diagrams (one for each category cluster). Thus, the area in the diagram represents the degree of BI<br />
integration in the category according to the responses.<br />
386
3. Results<br />
Ari Riabacke et al.<br />
The results are presented in two formats. The first format compares, for each category and question, the<br />
number of respondents who assessed a score of below 4 (the neutral score) to the number of<br />
respondents who assessed a score of above 4. These results are presented in three bar chart diagrams,<br />
one for each pair of categories. The other format is in the form of spider diagrams showing the modal<br />
score of each question, i.e. the most common assessment for each question. As there were more than<br />
one mode for some questions, the darker areas represent the result obtained from minimum modal<br />
scores and the brighter areas the maximum.<br />
3.1 Below neutral score vs. above neutral score<br />
2<br />
2<br />
1<br />
1<br />
5<br />
0<br />
Figure 2: Number of responds below (red) and above (green) the neutral score of 4 for categories 1-2<br />
30<br />
25<br />
20<br />
15<br />
10<br />
5<br />
0<br />
3.1 3.2 3.3 3.4 3.5 3.6 4.1 4.2 4.3 4.4 4.5 4.6<br />
Below 4<br />
Above 4<br />
Figure 3: Number of responds below (red) and above (green) the neutral score of 4 for categories 3-4<br />
35<br />
30<br />
25<br />
20<br />
15<br />
10<br />
5<br />
0<br />
1. 1. 1. 1. 1. 1. 2. 2. 2. 2. 2. 2.<br />
5.1 5.2 5.3 5.4 5.5 5.6 6.1 6.2 6.3 6.4 6.5 6.6<br />
Below 4<br />
Above 4<br />
Figure 4: Number of responds below (red) and above (green) the neutral score of 4 for categories 5-6<br />
387<br />
Below<br />
Above
3.2 Modal scores<br />
Ari Riabacke et al.<br />
1.1<br />
7<br />
6<br />
5<br />
4<br />
3<br />
2<br />
1<br />
0<br />
Figure 5: Spider diagram of modal scores for categories 1-2<br />
2.4<br />
2.5<br />
2.3<br />
2.6<br />
2.2<br />
2.1<br />
3.1<br />
7<br />
6<br />
5<br />
4<br />
3<br />
2<br />
1<br />
0<br />
Figure 6: Spider diagram of modal scores for categories 3-4<br />
4.4<br />
4.5<br />
4.3<br />
4.6<br />
4.2<br />
4.1<br />
5.1<br />
7<br />
6<br />
5<br />
4<br />
3<br />
2<br />
1<br />
0<br />
Figure 7: Spider diagram of modal scores for categories 5-6.<br />
6.4<br />
6.5<br />
6.3<br />
6.6<br />
6.2<br />
6.1<br />
388<br />
1.2<br />
1.6<br />
3.2<br />
3.6<br />
5.2<br />
5.6<br />
1.3<br />
1.5<br />
3.3<br />
3.5<br />
1.4<br />
3.4<br />
5.3<br />
5.5<br />
5.4
4. Analysis<br />
Ari Riabacke et al.<br />
The results differ markedly between categories of statements. In category 1, Visions, objectives, and<br />
strategies, there is good agreement for all except two statements. There seems to be a fairly good<br />
grounding of BI systems in the organisations’ control documents and visions, showing a conceptual<br />
integration of BI as decision support and with an understanding of how the system may provide support.<br />
1.2 and 1.6 did indicate the opposite, these two having modal scores of 3 and 2 respectively. Those two<br />
latter statements referred to specific management issues which were perceived to be less integrated with<br />
BI visions.<br />
In category 2, Business values from BI systems, there was also a good agreement with the statements in<br />
general. There is a clear indication that the needs are being catered for and that perceived business<br />
values are derived from business intelligence information. The only exception here is statement 2.3 which<br />
deals with the source of promotion for BI solutions, indicating that the need of a BI system is not primarily<br />
perceived by the benefiting organisation but rather promoted by external sources.<br />
In category 3, Requirements analysis and needs, all indicators show that there was good agreement with<br />
the statements. This implies that the most businesses indeed have a fairly good control over the design<br />
and implementation of BI systems from an organisational perspective.<br />
In category 4, Change management, there was much less agreement with the statements. In particular,<br />
statements 4.1 – 4.3 indicated stronger disagreement with the statements, implying problems with<br />
integrating the information in decision making processes. The result for statement 4.5 is clearly positive,<br />
dealing with user involvement in general, whereas 4.4 and 4.6 indicate a neutral to agreeing position with<br />
the statements. The latter three are statements more typical to information systems in general.<br />
In category 5, Technical solutions, most statements are agreeable for a majority of respondents. This<br />
category deals with technical issues on implementation as opposed to category 3 which deals with<br />
organisational issues. Like category 3, this is not overly complicated to deal with for accomplished<br />
organisations. There is, however, strong disagreement regarding statement 5.3, again dealing with top<br />
management. On the other hand, there is also disagreement among researchers on the involvement of<br />
top management in these issues in contrast with involvement in organisational issues which are<br />
considered much more pivotal to success.<br />
Finally, in category 6, Decision making support, we find the strongest area of disagreement, with all six<br />
indicators being disagreed with on an aggregate level. This clearly implies that the largest set of problems<br />
with BI systems lies in integrating them with decision making processes in the businesses, indicating<br />
problems with BI scope rather than the BI systems themselves.<br />
The same picture is painted by the modal scores, see Figures 5-7.<br />
5. Conclusion<br />
Of the six problem areas studied in this paper, clearly Change management and Decision making support<br />
are those that contribute most to BI systems not being as efficiently functional in an organisational sense<br />
as anticipated. The area Visions, objectives, strategies is mixed, with some problems, and Technical<br />
solutions, Business values from BI systems and Requirements analysis and needs are less problematic<br />
than the others. This indicates that the perceived value from a BI system does not lie in its decision<br />
support function and that there is little understanding in how to use it as decision support, i.e. a problem<br />
within the wider concept of BI scope.<br />
In other words, when looking back at the triangle in Figure 1, we have found that most efforts done<br />
towards the integration of business intelligence systems in organizations focuses on the right hand side<br />
of the triangle in Figure 1, i.e. on the technical parts of the BI solution.<br />
6. Discussion and outlook<br />
While almost every for-profit organization builds business cases or similar around new investments in<br />
general, this is much more seldom done for BI systems, where the inability to measure its impact renders<br />
standard measurement models unusable. There is of course a reason for BI systems winding up in the<br />
state they are. One hypothesis that need to be further examined in a more qualitative investigation is that<br />
this is a result of that BI systems are treated as just another kind of traditional information systems. In<br />
389
Ari Riabacke et al.<br />
contrast to operational information systems which directly support operational processes that often are<br />
rather well defined, BI systems have the capability to support follow-up and decision making processes<br />
which typically are less defined in an enterprise. Hence, for an organization, a BI system is<br />
complementary to an operational information system, serving as a support for managing the business at<br />
several different levels. Surely, decision processes could also be modelled using process modelling<br />
techniques, but not the same as operational business processes. However, when lacking another<br />
approach, it is understandable that BI systems are treated as “just the next information system” when it is<br />
being designed, implemented, and maintained. At present, the main objective when implementing a BI<br />
system seems to be that of simply getting it to operate, with too little emphasis on actually getting it to<br />
work within its context and thus provide value. This also supports the statement that the practice of BPM<br />
concepts is not widespread or mature in the responding organizations, or that the BI system is neglected<br />
in this practice. Therefore, too many organizations nowadays express a neutral or unhappy view on their<br />
BI systems. Avoiding this calls for another approach in the design phase, approaches stimulating how the<br />
information is going to be used and by whom, as opposed to approaches focusing on simply getting it<br />
implemented and usable.<br />
In order to develop good BI solutions that support organizations in reaching their overall goals, we must<br />
strive for a more holistic view on the implementation of BI systems, for instance by not implementing BI<br />
systems without having a proper treatment and understanding of the business processes and the related<br />
decision making activities. This means that the solutions must be based on knowledge from a number of<br />
areas in addition to database management and computer engineering, i.e. the balance between decisionmakers,<br />
business models and the technical support system must be better managed. In most current BI<br />
solutions there is a lack of balance in terms of focus on the different corners in the triangle. This calls for<br />
requirements specification and design processes focusing on and identifying the business and decision<br />
processes that are to benefit from the information provided through the system as well as the<br />
organizations overall strategy.<br />
7. Appendix<br />
7.1 The questionnaire<br />
The following three sub-sections show the statements in the questionnaire as they appear in the results<br />
section. They have been grouped into three clusters for presentational purposes.<br />
7.1.1 Categories 1-2<br />
1.1 The BI-systems support the work towards fulfilling the organizations objectives, strategies, and<br />
visions.<br />
1.2 The BI strategy is well established at the organization’s top management.<br />
1.3 BI is an important part of the organization’s strategy process.<br />
1.4 We have a clear understanding of how the BI systems will aid us in reaching stipulated<br />
objectives.<br />
1.5 We ensure that the technical solution, the organization, and the business objectives are<br />
consistent in our solution for BI.<br />
1.6 BI and BI-related questions are a common issue on the regular management meetings.<br />
2.1 The BI systems are tied to the needs of the organization.<br />
2.2 The BI systems provide information relevant to the different business units and their needs.<br />
2.3 It is clear who promote the needs for BI systems.<br />
2.4 The BI systems support the organization’s core processes.<br />
2.5 The development of our BI systems is solely driven by our business needs.<br />
2.6 We always create a “business case” prior to investments in BI systems.<br />
390
7.1.2 Categories 3-4<br />
Ari Riabacke et al.<br />
3.1 Our BI systems provide information that is easy to understand and use in decision making,<br />
3.2 The BI systems support the organization in reaching its overall business objectives.<br />
3.3 We have a clear distribution of roles and responsibilities when designing our BI systems.<br />
3.4 We have a clear distribution of roles and responsibilities when implementing our BI systems.<br />
3.5 The ownership of BI system implementation projects lies on the using part of the organization.<br />
3.6 Technical experts, users, and management are always involved in the design and implementation<br />
of BI systems.<br />
4.1 We secure that the foreseen users utilize the information provided by the BI systems.<br />
4.2 We have well established routines for how to prepare the organization and its users to use new<br />
technology and new information.<br />
4.3 We have clear ideas on how new available information will fit with existing decision processes.<br />
4.4 We allocate funds for education of personnel when investing in BI systems.<br />
4.5 We secure that the users are involved in development and implementation of BI systems.<br />
4.6 We secure that the organization and its individuals know how and why they are to use the BI<br />
systems and the information they provide.<br />
7.1.3 Categories 5-6<br />
5.1 We secure that the technical parts of the BI systems support stipulated business objectives.<br />
5.2 The business put clear requirements on what is desired from the technical parts of the BI<br />
systems.<br />
5.3 Corporate management is an active part in the technical requirements specification process for<br />
the BI systems.<br />
5.4 The users are an active part in the requirements specification process for the BI systems.<br />
5.5 There is a clear and structured user representation in the selection of technical solutions for our<br />
BI systems.<br />
5.6 The users of our BI systems is an active part when deciding on presentation formats of the<br />
information provided by the systems.<br />
6.1 We measure the business value generated by our BI systems in a quantitative fashion.<br />
6.2 There are clear processes for how the information provided by BI systems are be used in<br />
operational decision making.<br />
6.3 We use essentially all information accessible from our BI systems.<br />
6.4 There are clear guidelines and processes for how to cope with information anomalies or<br />
deviations from expectations.<br />
6.5 We utilize the information provided from our BI systems in a majority of the business decisions<br />
made in our company.<br />
6.6 We continuously secure that our BI systems provide relevant information for our business units.<br />
391
References<br />
Ari Riabacke et al.<br />
Bajaj, A. and Nidumolu, S. R. (1998) “A Feedback Model to Understand Information System Usage”, Information &<br />
Management, Vol. 33, pp. 213-224.<br />
Bogdana, P. I., Felicia, A., and Delia, B. (2009). “The Role of Business Intelligence in Business Performance<br />
Management”, Annals of Faculty of Economics, Vol. 4, No. 1, pp. 1025-1029.<br />
Bryman, A. (1989) Research Methods and Organizational Studies, Routledge, New York.<br />
French, S. and Ríos Insua, D. (2000) Statistical Decision Theory, Arnold, London, 2000.<br />
Frolich, M. N. and Ariyachandra, T. R. (2006). “Business Performance Management: One Truth”, Information<br />
Systems Management, Vol. 23, No. 1, pp. 41-48.<br />
Hannula, M. and Pirttimäki, V. (2003) “Business Intelligence Empirical Study on the top 50 Finnish Companies”,<br />
Journal of American Academy of Business, March 2003, pp. 593-599.<br />
Keen, P. G. W. (1981) “Value Analysis: Justifying Decision Support Systems”, MIS Quarterly, Vol. 5, No. 1, pp. 1-15.<br />
Lee, D., Newman, P. and Price, R. (1999) Decision Making in Organisations, Calays Ltd., Ives plc.<br />
Pomerol, J.-C. and Adam, F. (2004) “Practical Decision Making – From the Legacy of Herbert Simon to Decision<br />
Support Systems”, Proceedings of Decision Support in an Uncertain World: The IFIP TC8/WG8.3 International<br />
<strong>Conference</strong>, pp. 647-657.<br />
Rasmussen, J. (2000) “Human Factors in a Dynamic Information Society: Where are we Heading?”, Ergonomics,<br />
Vol. 43, No. 7, pp. 869-879.<br />
Sabherwal, R., Jeyaraj, A. and Chowa, C. (2006) “Information Systems Success: Individual and Organisational<br />
Determinants”, Management Science, Vol. 52, No. 12, pp. 1849-1864.<br />
Riabacke, A. (2006) “Managerial Decision Making Under Risk and Uncertainty”, IAENG International Journal of<br />
Computer Science, Vol 32., No. 4, pp. 453-459.<br />
Shim, J. P., Warkentin, M., Courtney, J. F., Power, D. J., Sharda, R. and Carlsson, C. (2002) “Past, Present and<br />
Future of Decision Support Technology”, Decision Support Systems, Vol. 33, No. 2, pp. 111-126.<br />
Speier, C. and Morris, M. G. (2000) “The Influence of Query Interface Design on Decision making Performance”, MIS<br />
Quarterly, Vol. 27, No. 3, pp. 397-423.<br />
Stenfors, S., Tanner, L., Syrjänen, M., Seppälä, T. and Haapalinna, I. (2007) “Executive Views Concerning Decision<br />
Support Tools”, <strong>European</strong> Journal of Operational Research, Vol. 181, pp. 929-938.<br />
Watson, H. J. and Wixom, B. H. (2007) “The Current State of Business Intelligence”, Computers, Vol. 40, No. 9, pp.<br />
96-99.<br />
Turban, E., Sharda, R., and Delen, D. (2010). Decision Support and Business Intelligence Systems, 9 th ed., Prentice<br />
Hall Press, Upper Saddle River NJ.<br />
Yin, R. K. (2003) Case Study Research: Design and Methods, 3 rd ed., Thousand Oaks: Sage Publications.<br />
392
Assessing Future Value of Investments in Security-Related IT<br />
Governance Control Objectives – Surveying IT Professionals<br />
Waldo Rocha Flores, Teodor Sommestad, Hannes Holm and Mathias Ekstedt<br />
Royal Institute of Technology, Stockholm, Sweden<br />
waldorf@ics.kth.se<br />
teodors@ics.kth.se<br />
hannes.holm@ics.kth.se<br />
mathiase@ics.kth.se<br />
Abstract: Optimizing investments in IT governance towards a better information security is an understudied topic in<br />
the academic literature. Further, collecting empirical evidence by surveying IT professionals on their relative opinion<br />
in this matter has not yet been explored to its full potential. This paper has tried to somewhat overcome this gap by<br />
surveying IT professionals on the expected future value from investments in security-related IT governance controls<br />
objectives. The paper has further investigated if there are any control objectives that provide more value than others<br />
and are therefore more beneficial to invest in. The Net Present Value (NPV) technique has been used to assess the<br />
IT professionals relative opinion on the generated future value of investments in 19 control objectives. The empirical<br />
data was collected through a survey distributed to professionals from the IT security, governance and/or assurance<br />
domain and analyzed using standard statistical tools. The results indicate that the vast majority of investments in<br />
control objectives is expected to yield a positive NPV, and are beneficial to an organization. This result implies that<br />
investments in control objectives are expected to yield positive Net Present Value for the firm, which is an important<br />
finding since many of the benefits from an investment are indirectly related an may occur well into the future. The<br />
paper moreover contributes in strengthening the link between IT governance and information security.<br />
Keywords: IT governance, control objectives, Information security, net present value<br />
1. Introduction<br />
Contemporary enterprises are largely dependent on Information Technology (IT) as it supports many<br />
critical business and administrative functions. This dependency has unfortunately led to an increase in<br />
potential threats to the enterprise and its information assets. Enterprises are therefore compelled to<br />
invest in information security to protect information assets, to avoid negative consequences that poor<br />
information security can lead to, and create value for their shareholders. Value for an enterprise’s<br />
shareholders is created by making beneficial real investment decisions. The meaning of real investments<br />
is expenditures that generate cash in the future and, as opposed to financial investments, like stocks and<br />
bonds, are not financial instruments that are traded in the financial markets (Grinblatt and Titman 1998).<br />
To aid a decision maker in determining whether an security investment create value for the shareholder a<br />
responsible manager need to argue and communicate the rational with the choice of the investment<br />
alternative (Johnson 2006). By focusing on communicating security issues concerning business issues<br />
and providing related business benefits, risks and benchmarks, the importance of a security investment<br />
can be heightened. Financial metrics such as net present value (NPV) improves the effectiveness of the<br />
communication and can therefore be used to evaluate different security investment alternatives so that<br />
the most beneficial investment alternative is identified (Theo, Renkemaa and Berghoutb 1997). The use<br />
of NPV, explicitly, as an evaluation tool for security investments at the proposal stage has also been<br />
suggested by (Gordon and Loes 2002).<br />
IT governance provides value by harmonizing decisions about the management and use of IT with<br />
desired business behaviors. This implies that IT governance investments pay off (Van Grembergen and<br />
De Haes 2008) (Weill and Ross 2004) (Simonsson, Johnson and Ekstedt 2010). A core objective for IT<br />
governance is information security (Calder and Watkins 2008), and it is therefore reasonable to believe<br />
that investments in IT governance also are beneficial for an organizations information security. A problem<br />
is, however, that benefits to the organization are often indirectly related to the investment and may occur<br />
well into the future. It may therefore be difficult for managers in an organization to demonstrate the<br />
expected future value of investing in control objectives. And to effectively direct investments endeavors, a<br />
crucial question is if there exist any differences in how effective these investments are in generating<br />
value. The need for decision support regarding investment opportunities in security is therefore needed.<br />
In this article we have investigated the value in terms of reduced negative consequences from security<br />
incidents generated from investments in IT governance control objectives (time, people or money etc.).<br />
393
Waldo Rocha Flores et al.<br />
To answer this question we have surveyed IT professionals in the security, governance and/or assurance<br />
domain and asked them on their opinion regarding the value of investments in IT governance control<br />
objectives. The Net Present Value (NPV) technique was used in order to assess the IT professional’s<br />
relative opinion on the value these investments is expected to provide.<br />
The structure of the paper is outlined as follows. Section 2 presents some related work to motivate our<br />
contribution in respect to previous research. In section 3 the control objectives from COBIT Security<br />
Baseline and the Net Present Value technique is presented. Section 4 presents the methodology used in<br />
the study. Section 5 presents and analyses the results. Section 6 discusses the results and finally,<br />
section 7 concludes the paper.<br />
2. Related works<br />
In the IT governance field there exists several studies on the impact of different components on overall<br />
goals. COBIT (ISACA 2007a) describes what processes that are primarily and secondarily related to<br />
several information criteria in its appendices. (Simonsson, Johnson and Ekstedt 2010) have assessed the<br />
impact of IT governance maturity on the external quality of delivered IT services experienced by the<br />
business. The authors further also investigated the relative importance of different IT governance<br />
processes. In addition, (Debreceny and Gray 2009) assessed the IT governance maturity of<br />
organizations with respect to predefined IT governance maturity attributes.<br />
The use of an information security standard as an IT governance framework has been proposed in<br />
(Calder and Watkins 2008). In this work, the authors discussed the relevance of the information security<br />
standard management standard ISO 27000 as a framework for fulfilling general objectives with IT<br />
governance.<br />
This study examines how control objectives associated with IT governance support information security<br />
and generate future value in terms of reducing negative consequences from security incidents. There are<br />
several studies that assess the importance of different security goals, for instance (Bartolini and Sallé<br />
2004) (Su, Bolzoni and Eck 2007) (Neubauer, Klemen and Biffl 2005). These authors have assessed the<br />
priority or impact of security objectives on information security. This article, however, assess the impact<br />
of investment in IT governance objectives on security.<br />
The impact of objectives related to security can be assessed in a number of ways. Economic methods<br />
are often used to evaluate different security strategies. For example, (Gordon and Loes 2002) (Mercuri<br />
2003) (Sonnenreich, Albanese and Stout 2006) have described methods where economic variables were<br />
used to evaluate security alternatives regarding the cost of investments and the value they provide.<br />
Security per se does not provide business value. The benefit of security investments are instead<br />
assessed in terms of the reduced cost from security incidents. Therefore, economic techniques such as<br />
the NPV or return on investment (relabeled as return on security investment) have been suggested by<br />
(Gordon and Loes 2002) (Sonnenreich, Albanese and Stout 2006).<br />
In the present paper, NPV technique is used to empirically assess the relative value investments in 19 IT<br />
governance control objectives are expected to provide. To the authors knowledge this has not yet been<br />
presented in the academic literature.<br />
3. COBIT security baseline and the Net Present Value<br />
This section describes the best-practice guideline that served as a basis for the survey and the evaluation<br />
technique used to evaluate generated future value of investment in control objectives.<br />
Control Objectives for Information and related Technology Standards (COBIT) is currently the most wellestablished<br />
best-practice guideline for IT governance. It outlines 34 IT processes, their purpose and the<br />
controls that should be used to govern them (ISACA 2007a)(Debreceny and Gray 2009) (Simonsson,<br />
Johnson and Ekstedt 2010). A core part of COBIT concerns information security and the 19 most<br />
essential IT governance control objectives toward better information security has therefore been<br />
extracted from COBIT and presented in COBIT Security Baseline (ISACA 2007b). In this paper COBIT<br />
security baseline (ISACA 2007b) was used as the basis for developing the survey used for collecting<br />
data.<br />
COBIT security baseline describes 19 security-related control objectives that are cross-referenced to<br />
related COBIT processes and related control sections in ISO/IEC 27002:2005 (ISO/IEC 2005). The<br />
394
Waldo Rocha Flores et al.<br />
control objectives are grouped into four domains: Plan and Organize (PO), Acquire and Implement (AI),<br />
Deliver and Support (DS), and Monitor and Evaluate (ME) (c.f. Figure 1). The domains map to IT´s<br />
traditional responsibility areas of plan, build, run and monitor. Within the COBIT framework, these<br />
domains, as shown in Figure 1, cover the following:<br />
Plan and Organize (PO): This domain provides direction to solution delivery and service delivery. It<br />
covers strategy and tactics, and concerns the identification of the way IT can best contribute to the<br />
achievement of the business objectives.<br />
Acquire and Implement (AI): This domain provides the solutions and passes them to be turned into<br />
services. To realize the IT strategy, IT solutions need to be identified, developed or acquired, as well<br />
as implemented and integrated into the business process. In addition, changes in and maintenance<br />
of existing systems are covered by this domain to make sure the solutions continue to meet business<br />
objectives.<br />
Deliver and Support (DS): This domain receives the solutions and makes them usable for end users.<br />
It is also concerned with the actual delivery of required services, which includes service delivery,<br />
management of security and continuity, service support for users, and management of data and<br />
operational facilities. All IT processes need to be regularly assessed over time for their quality and<br />
compliance with control requirements.<br />
Monitor and Evaluate (ME): This domain serves the purpose of monitoring all processes to ensure<br />
that the direction provided is followed. Further, the domain addresses performance management,<br />
monitoring of internal control, regulatory compliance and governance.<br />
Monitor and Evaluate<br />
17.Regularly monitor the performance of information<br />
security.<br />
18.Gain confidence and trust in security through<br />
reliable and independent sources.<br />
19.Ensure that information security functions comply<br />
with applicable laws, regulations and other external<br />
requirements.<br />
Deliver and Support<br />
11.Define and manage security aspects of service<br />
levels.<br />
12.Manage security aspects of third-party services.<br />
13.Ensure that the enterprise is capable of carrying on<br />
its day-to-day automated business activities with<br />
minimal interruption from a security incident.<br />
14.Ensure that all configuration items are<br />
appropriately secured and security risks minimized by<br />
ensuring the enterprise’s awareness of its IT-related<br />
assets and licenses.<br />
15.Ensure that all data remain complete, accurate and<br />
valid during input, processing, storage and<br />
distribution.<br />
16.Protect all IT equipment (hardware) from damage.<br />
IT Governance Control Objectives toward<br />
better information security<br />
Plan and Organize<br />
1.Identify information and services critical to the<br />
enterprise and consider their security requirements.<br />
2.Define and communicate information security<br />
responsibilities.<br />
3.Define and communicate management aims and<br />
directions with respect to information security.<br />
4.Ensure that security functions are staffed properly<br />
with people who possess the necessary skills to fulfill<br />
the role.<br />
5.Discover, prioritize, and either contain or accept<br />
relevant information security risks.<br />
Acquire and Implement<br />
6.Consider security when identifying, automated<br />
solutions.<br />
7.Consider security when acquiring and maintaining<br />
the technology infrastructure.<br />
8.Consider security when enabling operational use.<br />
9.Ensure that all changes, including patches, support<br />
enterprise objectives and are carried out.<br />
10.Ensure that all new systems and changes are<br />
accepted only after sufficient testing of security<br />
functions.<br />
Figure 1: IT governance control objectives drawn from COBIT security baseline<br />
The NPV method was used to assess the generated future value of the 19 control objectives. The starting<br />
point in the NPV method is the cost of capital for an investment, otherwise known as the required return.<br />
The cost of capital is the amount that an investor requires to compensate her for the time value of money<br />
tied up in the investment and for taking on risk in the investment. Thus, it represents the costs for her to<br />
take on the investment alternative and it is the minimum return that investors expect for providing capital<br />
395
Waldo Rocha Flores et al.<br />
to an investment, thus setting a benchmark that an investment has to meet. For an investment to add<br />
future value, it must earn more than the cost of capital. When calculating the NPV for an investment the<br />
following formula is used:<br />
NPV<br />
∞<br />
∑<br />
Rt<br />
t=<br />
( 1+<br />
i<br />
= t<br />
0 )<br />
The cost of capital rate ( ) is used as the discount rate. This is the rate of return that could be earned on<br />
an alternative investment with similar risk. is the net cash flow, i.e. the amount of cash, inflow minus<br />
outflow, at time To calculate the NPV for an investment each cash inflow/outflow associated with it is<br />
discounted back to its present value. The NPV is the sum of all present values from the time series of<br />
cash flows. If the NPV is larger than zero it is beneficial to make the investment and the investment is<br />
therefore a good real investment opportunity. If NPV is less than zero it will not add value and thus be a<br />
less attractive investment opportunity. Applied to our case a good real investment opportunity is an<br />
investment in security that generates future value in terms of reduced impacts from security incidents and<br />
provides value to the shareholder. To estimate the value of investments in control objectives the NPV of<br />
all 19 control objectives was assessed by surveying IT professionals.<br />
4. Method<br />
This study utilizes a survey as a measurement tool, this due to the obvious strengths in terms of<br />
statistical analysis and cost efficiency. The aim of the study is not to reach in-depth information regarding<br />
each of the 19 control objectives; it is simply to gain an understanding of the relative impact of control<br />
objectives on information security and can thus be categorized as an exploratory study.<br />
4.1 Population and sample<br />
When constructing a survey it is important to specify a population with favorable attributes and choose a<br />
representative sample from that population (Saunders, Lewis and Thornhill 2009). A natural prerequisite<br />
for respondents is that they can relate to the questions of the survey (Blair 2005). In this case it meant a<br />
population of experienced professionals in the IT security, governance and/or assurance domain.<br />
A strategic sample consisting of a conference with approximately 100 participants satisfying the<br />
population constraints was used. Of this population 22 answered the survey. 15 respondents were<br />
professionals in IT Security, 9 were professionals in IT Assurance and 11 were professionals in IT<br />
governance. The most common combination was IT Security and IT governance. Four respondents came<br />
from the banking industry, six from the energy industry, three from the public administration sector, two<br />
from the telecommunication sector, two from transport logistics and one consultant. One respondent did<br />
not specify his or her sector.<br />
The majority of the respondents were CISA (Certified Information System Auditor) and/or CISM (Certified<br />
Information Security Manager); twelve respondents were CISA, six CISM and three CGEIT (Certified in<br />
the Governance of Enterprise IT). Only four respondents did not hold any certification. Thus, the<br />
respondents originate from many different branches and a diverse number of roles. Although slightly<br />
diminutive this sample is considered representative of the population.<br />
4.2 The survey<br />
A face-to-face survey was carried out since all respondents were at the same geographical location (Blair<br />
2005). The survey consisted of two pages of which the first according to recommendation by (Blair 2005)<br />
introduces the concepts of COBIT Security baseline, NPV, and a description of how to answer the<br />
questions. Furthermore, the first page also includes three questions used to assess background<br />
information of respondents. The first question concerned the domain the IT professional work in, the<br />
second if the respondent holds any certification and the third in which industry the IT professionals are<br />
active in. The second page of the survey consisted of 19 questions utilized in order to gain information<br />
regarding the significance of the IT governance control objectives in the COBIT security baseline. All of<br />
the 19 questions included in the survey were taken directly from COBIT security baseline without any<br />
manipulation.<br />
396<br />
(1)
Waldo Rocha Flores et al.<br />
For each of the 19 control objectives the respondents were asked to indicate the size of the<br />
corresponding NPV on a scale. In figure 2, the 5 questions belonging to the Plan and Organize domain<br />
are presented to provide information of the question format. The questions were answered using ratio<br />
scale, with 0 NPV in the middle, positive NPV to the right and negative NPV to the left. The results were<br />
interpreted using a ruler with a millimeter scale, rounded to the nearest complete mm. The outcome was<br />
then rescaled to the interval [-5,+5] in order to generate more pedagogical results.<br />
Please estimate how investments in each of the areas below would add additional value (NPV) to an organization in<br />
the industry you work in.<br />
4. Control objectives (investment opportunities)<br />
Yields the following NPV<br />
Negative Positive<br />
4.1. Identify information and services critical to the enterprise and<br />
consider their security requirements.<br />
4.2. Define and communicate information security responsibilities.<br />
4.3. Define and communicate management aims and directions<br />
with respect to information security.<br />
4.4. Ensure that security functions are staffed properly with people<br />
who possess the necessary skills to fulfill the role.<br />
4.5. Discover, prioritize, and either contain or accept relevant<br />
information security risks.<br />
Figure 2: Question format<br />
4.3 Analysis methods<br />
Three tools were utilized to analyze the results of the survey: box plots, tests for normality and statistical<br />
measurements.<br />
Box plots (sometimes referred to as “box and whisker plots”) were first used to evaluate the results. Box<br />
plots were chosen to be used due to the favorable non-parametric distribution requirement by the tool.<br />
The box plot provides a graphical representation of a dataset according to percentiles. The bottom of a<br />
box (the lower quartile, Q1) is the 25 th percentile and the top of the box (the upper quartile, Q3) is the 75 th<br />
percentile. The band in the box represents the median (50 th percentile). Data in the 1 st -24 th percentile and<br />
the 76 th -100 th percentile are represented by whiskers and/or dots (outliers). More information regarding<br />
box plots can be found in (McGill, Tukey och Larsen 1978).<br />
Tests for normality were performed using QQ-plots (Warner 2008). Outliers who did not meet the<br />
normality assumption were removed according to the recommendation by (Montgomery 2005) in order to<br />
improve the overall data quality. The evaluation of the data set with box plots gave an understanding of<br />
the data before manipulation; therefore test for normality was performed after analyzing the box plots.<br />
Standard statistical measurements such as mean, standard deviation and confidence intervals were<br />
finally used to present the survey results. Descriptions of these measures can be found in (Montgomery<br />
2005).<br />
5. Results and analysis<br />
A control objective with a positive NPV should be interpreted as a control objective that adds value and is<br />
beneficial to invest in. If the NPV is less than zero it will not add value and is therefore not an attractive<br />
investment opportunity. Further, the control objective with highest NPV will be the most attractive<br />
investment opportunity when compared to other investments in control objectives with similar risk. Figure<br />
3 shows a box plot over the 19 control objectives and table 1 describes their mean, standard deviation<br />
and confidence intervals. The results from the survey indicate that the vast majority of security<br />
investments have a positive net present value, and are thus beneficial for an organization to invest in.<br />
The mean of all NPVs but one is positive. A 95 percent confidence interval over the mean values of<br />
respondent assessments indicate that the true mean (the mean of the population) lies between the upper<br />
397
Waldo Rocha Flores et al.<br />
(UB) and lower (LB) in the 95 % of the cases. In the remaining part of this section, the results will be<br />
discussed on a domain-level and thereafter each domain will be discussed.<br />
Figure 3: Assessments of control objectives investment potential<br />
Table 1: Control objectives ranking<br />
Objective Mean Std dev LB (95%) UB (95%)<br />
CO1 2.0 1.4 1.4 2.6<br />
CO2 1.5 1.3 1.0 2.1<br />
CO3 1.7 1.5 1.1 2.3<br />
CO4 1.1 1.5 0.5 1.7<br />
CO5 2.0 1.1 1.5 2.5<br />
CO6 1.2 1.3 0.6 1.7<br />
CO7 1.3 1.3 0.8 1.9<br />
CO8 1.3 1.2 0.8 1.9<br />
CO9 1.6 1.5 1.0 2.3<br />
CO10 1.5 1.6 0.8 2.2<br />
CO11 0.5 1.8 -0.3 1.2<br />
CO12 1.0 2.2 0.0 1.9<br />
CO13 1.6 1.8 0.8 2.4<br />
CO14 0.8 1.6 0.1 1.5<br />
CO15 1.3 1.7 0.5 2.0<br />
CO16 -0.2 1.9 -1.0 0.7<br />
CO17 0.9 1.7 0.1 1.6<br />
CO18 0.1 1.2 -0.4 0.6<br />
CO19 1.3 1.5 0.6 1.9<br />
Mean 1.2 1.5 0.5 1.9<br />
398
5.1 Domains<br />
Waldo Rocha Flores et al.<br />
A box plot of the results aggregated to the domain level can be seen in Figure 4. Plan and organize (PO,<br />
CO1-CO5) have the highest median, followed by acquire and implement (AI, CO6-CO10). Deliver and<br />
support (DS, CO11-CO16) and monitor and evaluate (ME, CO17-CO19) have relatively similar medians<br />
and ranges, although DS have a higher degree of spread of data. Table 2 shows the mean of the domain<br />
experts’ assessments of NPV for control objectives within each domain. As can be seen here the highest<br />
mean is associated to the domain Plan and organize and the lowest is associated with Deliver and<br />
Support and Monitor and evaluate. Also, the mean of the NPV is positive for all four domains.<br />
Figure 4: Box plots per domain<br />
Table 2: Control domains mean and spread<br />
Domain Mean Std dev LB (95%) UB (95%)<br />
PO 1.6 1.1 1.2 2.1<br />
AI 1.3 1.1 0.8 1.8<br />
DS 0.8 1.3 0.3 1.4<br />
ME 0.8 1.0 0.3 1.2<br />
5.1.1 Plan and organize<br />
As can be seen in the box plot in Figure 3, investments in all control objectives in this domain are<br />
presumed to yield relatively high NPVs. QQ-plots indicate that the outliers need to be removed in order to<br />
satisfy the assumption of normality. The variables with the highest median, smallest box and lowest<br />
range is CO1 (“Identify information and services critical to the enterprise and consider their security<br />
requirements”) and CO5 (“Discover, prioritize, and either contain or accept relevant information security<br />
risks”).<br />
5.1.2 Acquire and implement<br />
Investments in control objectives within this domain are generally believed to render a positive NPV. The<br />
two control objectives with highest means are CO9 (“Ensure that all changes, including patches, support<br />
enterprise objectives and are carried out”) and CO10 (“Ensure that all new systems and changes are<br />
accepted only after sufficient testing of security functions”) .The boxes in this domain are approximately<br />
the same size, with the exception of which is slightly larger than the rest. CO10 does however have a<br />
strong outlier on the negative side which increases its standard deviation (cf. figure 3 and table 1).<br />
399
5.1.3 Deliver and support<br />
Waldo Rocha Flores et al.<br />
There is some variation between the expected NPV of investments in this domain. For instance, the<br />
mean NPV associated with CO16 (“Protect all IT equipment from damage”) is negative, but CO13<br />
(“Ensure that the enterprise is capable of carrying on its day-to-day automated business activities with<br />
minimal interruption from a security incident”) has the fourth largest mean NPV. Compared to the other<br />
domain there seems to be a higher disagreement among respondents on the NPV associated with<br />
investments in this domain. The assessments of NPV associated with CO12 (“Manage security aspects<br />
of third-party services”) is for example spread between highest return assessed to well below zero.<br />
5.1.4 Monitor and evaluate<br />
The monitor and evaluate domain consist of three control objectives. Their mean values are associated<br />
with a comparably low NPV. For instance is CO18 (“Gain confidence and trust in security through reliable<br />
and independent sources“) assessed to yield a NPV close to zero. Also within this domain is the spread<br />
of assessments comparably high. A 95 percent confidence interval over the respondents’ assessments<br />
does for example stretch between 0.1 and 1.6 for CO17 (“Regularly monitor the performance of<br />
information security”).<br />
6. Discussion<br />
The result from this survey is intended to reflect the opinion from a number of professionals within the IT<br />
security, governance and/or assurance domain. The reliability and validity of the instrument used to<br />
measure their opinion is described in section 6.2. The section below will discuss the results of this<br />
measurement.<br />
6.1 Value of investments in control objectives<br />
This paper has provided empirical data from a panel of IT professionals regarding the value in terms of<br />
reduced negative consequences from security incidents generated from investments in IT governance<br />
control objectives (time, people or money etc.). 16 out of 19 control objectives are associated with a 95<br />
percent confidence interval that only span over the positive side of the scale. This indicates that the panel<br />
agrees that investments in IT governance control objectives strengthen security objectives and are<br />
therefore beneficial. This is a finding that can be useful for a practitioner as it may be difficult to<br />
demonstrate the expected NPV of investing in control objectives since the benefits to the firm are often<br />
indirectly related to the investment and may occur well into the future.<br />
When it comes to individual control objectives the results gives some indications that C05 (“Discover,<br />
prioritize, and either contain or accept relevant information security risks”) and CO1 (“Identify information<br />
and services critical to the enterprise and consider their security requirements”) are beneficial to engage<br />
in. The less beneficial control objectives are CO16 (“Protect all IT equipment from damage.”) and CO18<br />
(“Gain confidence and trust in security through reliable and independent sources”). The respondents’<br />
answers regarding individual control objectives vary, the box plot in Figure 3 and the 95 percent<br />
confidence interval in Table 1 indicate this. However, the variations are small. Two explanations for these<br />
small variations are possible:<br />
Positive lopsidedness<br />
Respondent experience.<br />
These are now further discussed.<br />
Positive lopsidedness: The mean of the respondents assessments result in a positive NPV for all control<br />
objectives but one. This reflects a general belief among respondents in the benefit associated with<br />
investments on control objectives associated with security. As earlier described the respondents are<br />
individuals which have a vested interest in IT /security governance and practice such activities on a daily<br />
basis. With this as a basis the positive lopsidedness in NPVs could be attributed to an expected bias in<br />
the respondent’s judgment. While the positive lopsidedness could be an effect of the respondents biased<br />
opinions it could also reflect the situation in enterprises today. The software security market has grown in<br />
recent years (Gartner 2009) (Gartner 2010) and forecasts of the security market such as (DefenseNews<br />
2010) predict an increase in coming years. Such data indicate that many organizations do see a benefit<br />
in investments that strengthen security objectives.<br />
400
Waldo Rocha Flores et al.<br />
Respondent experience: The respondents have different backgrounds and experiences. Some variation<br />
among their opinions can therefore be expected. The assessments made of NPV potentials were also<br />
made for organizations in the industry where the respondent is a practitioner. Some variation can thus be<br />
explained by differences among industries. For example, the regulations that apply to organizations vary<br />
with industry. A regulation might require certain investments to be made and in that way influence the<br />
NPV that can be obtained from future investments associated with a control objective.<br />
Based on the spread in this data it appears as the accuracy of IT professionals assessments of individual<br />
investment opportunities in the IT governance domain is associated with a significant uncertainty. Future<br />
work could investigate the reasons for the spread among the respondents’ rankings of such investment<br />
opportunities.<br />
6.2 Instrument validity<br />
This section first address the instrument’s validity in terms of: content-, construct-, external- and internal<br />
validity (Boudreau, Gefen and Straub 2001) (Brewer 2000).<br />
Content validity is the degree to which items in an instrument reflect the content universe to which the<br />
instrument will be generalized (Boudreau, Gefen and Straub 2001). E.g., if this study aimed to draw<br />
conclusions on profitability of the COBIT security baseline using only results from control objectives 1-4<br />
the content validity would be low. Normal methods of measuring content validity are through literature<br />
review and expert panels/judges (Boudreau, Gefen and Straub 2001). All control objectives of the COBIT<br />
security baseline are included in the survey. Furthermore, all survey questions are taken directly from the<br />
framework without any manipulation. However, while the baseline concepts are thoroughly covered there<br />
might be slight problems with the prioritization technique employed (NPV). The NPV of a potential<br />
investment is a commonly used indicator on how lucrative investments are for an organization, but when<br />
making investment decisions other factors could also be considered. In particular, the risk of the<br />
investment and the opportunity costs associated with it could be of essence. These factors are only<br />
implicitly measured in this survey and one should therefore be somewhat careful when interpreting the<br />
relative ranking associated with different investments. Finally, the survey was studied by two academic<br />
experts in the area who assessed the tools content validity as high.<br />
Construct validity is the extent to which an operationalization measures the concepts that it purports to<br />
measure (Boudreau, Gefen and Straub 2001), i.e. whether the survey measures the benefit of investing<br />
in different control objectives and domains. First of all, neither the COBIT security baseline nor<br />
prioritizations of its elements are abstract in such a manner that could cause significant problems with the<br />
construct validity. Second, the baseline and prioritization potency of NPV are generally thought of as<br />
highly valid theory. Finally, the tool was tested (and discussed) by two academic experts and one<br />
respondent which were thought to be representative of the specified population. These pilot studies both<br />
hinted toward high construct validity.<br />
Another important validation concept is external validity; the degree to which the results of a study can be<br />
generalized [30]. This study is on one hand built upon a sanctioned theoretical framework and a to some<br />
extent representative sample. However, it is to the authors’ knowledge also the first study performed<br />
relating to the COBIT security baseline and NPV; therefore it is not possible to compare the results of this<br />
study to the results of previous research. Also, the sample is too small to draw any certain general<br />
conclusions.<br />
An additional well regarded attribute of validity is internal validity (Brewer 2000). However, since no<br />
evaluation of causal relationships was made in this study, internal validity is not applicable here.<br />
7. Conclusion<br />
Optimizing the investments of resources in information security is a complex task. The study presented in<br />
this article has tried to somewhat make the investment decision more rational by investigating the value<br />
in terms of reduced negative consequences from security incidents generated from investments in IT<br />
governance control objectives (time, people or money etc.). By surveying IT professionals in the security,<br />
governance and/or assurance domain regarding the NPV that could be obtained from these investments<br />
the study found that investments in control objectives are beneficial for a firm to engage in. The paper<br />
has therefore provided data that supports the theory that investments in IT governance also are beneficial<br />
for the information security in an organization.<br />
401
Waldo Rocha Flores et al.<br />
The mean NPV vary between the control objectives, and there are some indications of control objectives<br />
that provide more value. However, this study cannot assert which control objective that is most beneficial<br />
to invest in. Interesting future work includes therefore more research into the impact of prioritization<br />
between control objectives. Research could investigate if the provided benefit for a firm depends on<br />
proper prioritization, and if so, identify which control objective/objectives that should be prioritized.<br />
References<br />
Bartolini, C, and M Sallé. "Business Driven Prioritization of Service Incidents." DSOM, LNCS 3248. 2004. 64-75.<br />
Blair, Czaja R. Designing surveys. Sage Publications Inc, 2005.<br />
Boudreau, Marie-Claude, David Gefen, and Detmar W Straub. "Validation in Information Systems Research: A Stateof-the-Art<br />
Assessment." MIS Quarterly 25, no. 1 (2001): 1-16.<br />
Brewer, Marilynn B. "Research Design and Issues of Validity." In Handbook of research methods in social and<br />
personality psychology, by Harry T Reis and Charles M Judd, 3-17. Cambridge: Cambridge University Press,<br />
2000.<br />
Calder, A, and S Watkins. IT governance A manager´s guide to Data Security and ISO 27001/ISO 27002. Kogan<br />
Page, 2008.<br />
Cronobach, L J. "Coefficient alpha and the internal structure of tests." Psychometrika, ([14] L.J. Cronbach,<br />
"Coefficient alpha and the internal structure of tests," Psychometrika, vol. 16, 1951, pp. 297-334.) 16 (1951):<br />
297-334.<br />
Debreceny, R, and G.L Gray. "IT Governance and Process Maturity: a Field Study." In Proceedings of the 42nd<br />
Hawaii International <strong>Conference</strong> on System Sciences. 2009.<br />
DefenseNews. IT Security Market Heats Up - Demand Growing for Guidance on Cyber Protection. DefenseNews,<br />
2010.<br />
Gartner. Gartner says worldwide security software market on pace to grow 8 per cent in 2009. Gartner, 2010.<br />
Gartner. Gartner says worldwide security software revenue grew 18.6 per cent in 2008. Gartner, 2009.<br />
Gordon, L.A, and M.P Loes. "The Economics of Information Security Investment." ACM Transactions on Information<br />
and System Security Vol. 5, No. 4, 2002: 438–457.<br />
Grinblatt, M, and S Titman. Financial Markets and Corporate Strategy. McGraw-Hill International Edition, 1998.<br />
Hardy, G. "Using IT governance and COBIT to deliver value with IT and respond to legal, regulatory and compliance<br />
challenges." information security technical report, 2006: 55-61.<br />
ISACA. COBIT Security Baseline: An Information Security Survival Kit 2nd edition. ISACA, 2007b.<br />
ISACA. Control Objective for Information and Related Technology Standards. ISACA, 2007a.<br />
ISO/IEC. "Code of Practice for Information Security management, ISO/IEC 27002:2005." Switzerland, 2005.<br />
Johnson, E.C. "Security awareness: switch to a better program." Network Security, 2006: 15-18 .<br />
McGill, Robert, John W Tukey, and Wayne A Larsen. "Variations of Box Plots." The American Statistician 32, no. 1<br />
(1978): 12-16.<br />
Mercuri, R.T. "Analyzing Security Costs." Communications of the ACM Vol. 46, No. 6, 2003: 15-18.<br />
Montgomery, Douglas C. Introduction to Statistical Quality Control. Vol. 5. New Jersey: John Wiley & Sons, Inc.,<br />
2005.<br />
Neubauer, T, M Klemen, and S Biffl. "Business Process-based Valuation of IT security." EDSER. St. Louis, Missouri,<br />
USA, 2005.<br />
Saunders, Mark, Philip Lewis, and Adrian Thornhill. Research Methods for Business Students. Edinburgh: Pearson<br />
Education Limited, 2009.<br />
Simonsson, M, P Johnson, and M Ekstedt. "The effect of IT Governance Maturity on IT Governance Performance."<br />
Information Systems Management, December 2010: 10-24.<br />
Sonnenreich, W, J Albanese, and B Stout. "Return On Security Investment (ROSI) - A practical Quantitative Model."<br />
Journal of Research and practice in Information Technology, Vol. 38, No. 1, 2006.<br />
Su, X, D Bolzoni, and P.V Eck. "Understanding and Specifying Information Security Needs to support the Delivery of<br />
High Quality Security Services." The international <strong>Conference</strong> on Emerging Security Information, Systems and<br />
Technologies. 2007.<br />
Theo, J.W, E.W Renkemaa, and Berghoutb. "Methodologies for information systems investments evaluation at the<br />
proposals stage: a comparative review." Information and Software Technology Vol. 39 No. 1, 1997: 1-13.<br />
Van Grembergen, W, and S De Haes. Implementing information technology governance: Models, practices, and<br />
cases. IGI Pub., 2008.<br />
Warner, R. M. Applied Statistics. SAGE Publications, 2008.<br />
Weill, P, and J.W Ross. IT governance: How top performers manage IT decision rights for superior results. Harvard<br />
Business School Press, 2004.<br />
Whitman, M.E. "In defense of the realm: understanding the threaths to information security." International Journal of<br />
Information Management, February 2004: 43-57.<br />
Yin, R K. Case study research: Design and Methods. Washington D.C: SAGE Publications, 2003.<br />
402
Evaluation Information Extraction for Health Text Categories<br />
Using C4.5 and Naïve Bayes<br />
Klaokanlaya Silachan and Panjai Tantasanawong<br />
Silpakorn University, Thailand<br />
Klao_99@yahoo.com<br />
Panjai@su.ac.th<br />
Abstract: Healthcare information technology, HIT or HICT supports health information seeking tasks. Data mining is<br />
an information extraction technique for identifying relevant information, structuring this information and subsequently<br />
providing the means to add textual information through mining the natural language. In the healthcare field,<br />
techniques summarise patient records, extract terms, ascertain new knowledge and provide diagnostics for various<br />
conditions. In this study we are interested in health information particularly with regard to the relationship between<br />
symptoms and cause in cardiovascular disease (CVD). Data filtered through a discovery pattern can be used for<br />
many tasks, including mining and evaluation to the accuracy of information extraction. In this paper an approach and<br />
framework of classification analysis has been proposed for evaluating health information from personal and medical<br />
data extracted from records and documents. The information extraction, as a feature of the selection from textual<br />
data, is significant in that it uses sliding window and TF-IDF techniques for a multiword check as well as for statistical<br />
matrix term weighting and rate of recurrence. Diagnosis of a probable cardiovascular disease is done by looking at<br />
the symptoms and mined data by way of the ML algorithm. The outcome of the specific medical conditions is<br />
considered and determined the probability of cardiovascular disease in order to classify the documents and<br />
cardiovascular disease by the use of rule induction. This model generates a classifier from the patient medical data<br />
training set. Two machine learning techniques, C4.5 and Naïve Bayes, have been applied where the domain concept<br />
is ‘diseases’. Evaluation of the accuracy for the disease term and document classification process for our theoretical<br />
patient’s symptom and diagnosis were measured precisely using recall and F-measure for comparison as well as to<br />
demonstrate the process of learning. The results show that in predicting the categories of cardiovascular disease, the<br />
C4.5 classifier has 99.8% rate of prediction accuracy (percentage correct) and performed better performance than<br />
the Naive Bayes.<br />
Keywords: information extraction, health text categories, C4.5, Naïve Bayes, evaluation<br />
1. Introduction<br />
The main goal of text mining is to enable the user to extract relevant information from textual resources<br />
and manage operations such as classification and summary support (Dalila 2004). Healthcare<br />
information technology, also known as HIT or HICT, extracts health information through a data mining<br />
technique that aims to identify relevant information, structure the information, and at the same time<br />
provide a means to allow the addition of semantics. It does this by processing the natural language and<br />
looking for patterns in unstructured text. Information extraction(IE) for healthcare has been designed to<br />
summarise patients’ medical records by extracting terms or call terms and then identify emerging patterns<br />
in information from the text on symptoms, conditions and diagnoses. The patient’s medical data are<br />
largely maintained in free-text form. Thus, using information extraction techniques as a reliable and<br />
efficient method to extract structured information for further data mining form, this free-text may greatly<br />
benefit research endeavours.<br />
This study is focused in health information and particularly the relationship between symptoms and<br />
causes in cardiovascular diseases (CVD). These disorders of the heart and blood vessels includes<br />
coronary heart disease (heart attacks), cardiovascular disease (stroke), raised blood pressure<br />
(hypertension), peripheral artery disease, rheumatic heart disease, congenital heart disease and heart<br />
failure. The major causes of cardiovascular disease are excessive smoking, physical inactivity, and an<br />
unhealthy diet, resulting in severe illness, disability and death. Data extraction has important implications<br />
for the classification of documented information. (K.S. Kavitha et al. 2010 and Reddy 2007). The<br />
probability of cardiovascular disease is based on symptoms diagnosis. For classification of diseases and<br />
medical document, cardiovascular medical data is obtained from the extracted text. Then the diagnosis of<br />
disease and classification will be evaluated for accuracy.<br />
It is necessary to measure the effectiveness of classifying data sets, medical data, for more precision and<br />
predictive accuracy in the classification of cardiovascular disease. In this paper an approach and<br />
framework is proposed that focuses on the evaluation of health information for the classification and<br />
analysis of cardiovascular disease case studies. The disease data can be extracted from a patient’s<br />
medical records and other documents. This step uses sliding windows and Term Frequency-inverse<br />
403
Klaokanlaya Silachan and Panjai Tantasanawong<br />
Document (TF-IDF) techniques for a multiword check as well as for statistical matrix term weighting and<br />
incidence in information extraction with the patient’s textual medical term data. The next step involves<br />
diagnosis of a probable cardiovascular disease by looking at the symptoms and mining the databases for<br />
classification with an ML algorithm. This model generates a classifier from the medical data training set<br />
based on the characteristics of texts already classified. Two machine learning techniques, C4.5 and<br />
Naïve Bayes, can be applied where the domain concept is ‘diseases’. Finally, the evaluation of accuracy<br />
of the classified disease for a theoretical patient treatment and the subsequent diagnosis can be<br />
measured precisely using recall and F-measure with an alternate machine learning algorithm (ML) for<br />
comparison and to demonstrate the process of learning.<br />
2. Related work<br />
Many researchers have studied text categorisation, medical health information and evaluation, such as<br />
Maria Grineva, et al. 2009, proposed a method for the extraction of key terms from textual documents. In<br />
this study the method is based on a combination of the following two techniques: a Wikipedia-based<br />
semantic relatedness measure of terms and an algorithm for detecting community structure of a network.<br />
Parikshit Sondhi et al. 2010 studied a shallow information extraction problem that involved extracting<br />
sentences of a given set of topic categories from medical forum data. Given a corpus of medical forum<br />
documents, which the research goal is to extract two related types of sentences that describe a<br />
biomedical case, i.e.: medical problem descriptions and medical treatment descriptions. Such an<br />
extraction task directly generates medical case descriptions that can be useful in many applications.<br />
Experimental results show that the accuracy was up to 75%.<br />
The Dittapol et al. 2009 study involved the application of methods of knowledge discovery from<br />
databases using decision tree algorithms for respiratory system diagnosis to classify patients from the<br />
Pranakornsri Ayudthaya Hospital into three groups. From the hospital’s medical records the study<br />
methods were knowledge discovery with the employment of ID3, C4.5 and CART decision tree<br />
algorithms and determination of the effectiveness of the three algorithms.<br />
Keku E 2005 proposed that trained interviewers should conduct annual follow up telephone interviews to<br />
ascertain any significant health events, including subsequent diagnostic tests, hospitalisations or death.<br />
Information on cohort hospitalisations and deaths is transmitted to the medical record abstraction unit<br />
who review death certificates and hospital records. Interviews with the next of kin and completed<br />
questionnaires by physicians and medical examiners or coroners were used to obtain information on<br />
deaths within the cohort.<br />
3. Methodology<br />
The proposed framework and method for information medical text extraction data for text term<br />
classification is explained in the diagram, Figure 1, below.<br />
Figure 1: Framework for information extraction and evaluation for health categories<br />
404
Klaokanlaya Silachan and Panjai Tantasanawong<br />
Textual data was extracted to classify the cardiovascular disease using the principles of extract term,<br />
sliding window and TF-IDF for matrix term weight and frequency. Groups of cardiovascular patients were<br />
classified using rule to recognize the relationship between the symptoms of the patient in the case where<br />
the conditions were symptomatic of cardiovascular disease. This Healthcare base was also used for<br />
building the cardiovascular diagnostic system. The features of C4.5 were applied to classify<br />
cardiovascular disease and relevant documents. Details of these steps include:<br />
3.1 Data set<br />
The construction of a corpus of medical text samples were collected from clinics and referenced from<br />
websites, such as the World Health Organization (WHO), and others given that a standard data collection<br />
should be used in the experimental evaluation. The corpus contains 300 cases. This text corpus is useful<br />
for training and testing with NPL and data mining purposes.<br />
3.2 Pre-processing<br />
In this step the data was checked for missing and incorrect values. Some of the data was alphanumeric.<br />
We found some inconsistencies in the patient’s records which we corrected manually by taking an<br />
external reference.<br />
3.3 Information extraction<br />
To understand the relationships between a patient’s cardiovascular symptoms and their treatment, we<br />
used the health service knowledge database and extracted those terms describing symptoms commonly<br />
found in the texts of patients’ medical records. Extraction of the symptom terms for classification was by<br />
single word or in phrases. As a result, a sliding window technique was used to retrieve phrases and then<br />
match it to a medical dictionary to control the vocabulary of the health information as well as to classify<br />
the type of heart diagnosis. (Dalila et al. 2004; Zhi-Hua etc al. 2003). The process is shown below.<br />
Sliding window for Information Extraction (IE) has the following steps:<br />
1. Extract key term with the number stored in the wordlist.<br />
2. Multiword created using sliding windows<br />
If(word != “ “)<br />
Word += “ “<br />
Word += WordList[j];<br />
3. Check for word or multiword.<br />
4. Count the frequency of the words.<br />
5. Check the word.<br />
6. Stored number of terms found.<br />
3.4 Feature selection<br />
The main idea of feature selection is to select subsets of features from the original documents. It is<br />
performed by keeping the highest score on words according to a predetermined measure of their<br />
importance. The term frequency-inverse document frequency (TF-IDF) is the ability to separate the words<br />
in the document. The frequency of words in the document and the ratio of the number of documents<br />
containing a particular word appear as an equation (Aurangzeb et al. 2010) and help evaluate how<br />
important a term is in proportion to the number of times a term appears in the document, which is then<br />
divided by the frequency of that word in the corpus (T.Gruber 1995; S.Izumi 2007). A terms weighting<br />
was calculated as shown:<br />
405
Klaokanlaya Silachan and Panjai Tantasanawong<br />
W(fd) = tf(fd) log (D / df(f)) (1)<br />
freqij<br />
tf(f i,dj) = (2)<br />
Max k freq Kj<br />
D<br />
idf = log (3)<br />
#(f 1 )<br />
Where, Term frequency (TF), invert Document Frequency(IDF) , W(fd) feature f d (current transaction),<br />
Tf(fd) feature f d normalize, feature d), D (training set), Df(f) feature f.<br />
Disease term weight-frequencies based on TF-IDF has the following steps:<br />
1. Store of words and the frequency in the table.<br />
2. Loop all terms found in each document.<br />
3. Check all terms found in each document.<br />
4. Check for this term in any other documents.<br />
5 Show the number of documents in which these words are found.<br />
6 Calculate values Log2 (N / nj).<br />
7. Loop all documents.<br />
8. Calculate frequency in all documents.<br />
9. Check for term found in each document or not.<br />
10. Calculate TF-IDF.<br />
11. Check store weight value.<br />
3.5 Medical data<br />
When IE is conducted from a medical text and the data is stored in a relational database as medical data.<br />
It consists of symptom words and phrases including terminology and conditions. The next step is to focus<br />
on analyzing the medical data alongside medical diagnostic data using classification rules in data mining<br />
to ascertain patterns of terms and to substantiate classification of diseases. Medical diagnostic data<br />
ought to be used in the experimental evaluation.<br />
3.6 Classification<br />
Classification of categories is the most widely used technique in medical data mining (Se-Chul Chun<br />
2008), and the C4.5 classification algorithms were used for the classification of diagnoses following<br />
feature extraction of medical data in order to make the classification. Results show this is consistent with<br />
the performances of these classification algorithms for other applications (Aurangzeb 2010).<br />
3.6.1 Rule base for probable diagnosis cardiovascular model<br />
To understand the rules for establishing the symptoms that are likely to cause cardiovascular disease, a<br />
method of constructing a rule-based decision tree from the medical terms was used to create a specific<br />
rule (Zhi-Hua Zhou et al. 2003). The rule was used to learn rules from the new training data set based on<br />
a technique of collecting “if . . . then . . .” rules (Kent W. Bridges 1992). The symptoms are variable,<br />
selected for the classification of cardiovascular disease and other diseases.<br />
406
Klaokanlaya Silachan and Panjai Tantasanawong<br />
3.6.2 Classification of document and diseases model<br />
In the present study we used two different methods for the classification of cardiovascular disease and<br />
medical document. They differ from traditional approaches by generating domain models from data. In<br />
the proposed system, we use two different machine learning methods for classifying: Decision Tree<br />
(C4.5) and Naïve Bayes. (J.R. Quinlan 1993; B.H. Jun et al.1997)<br />
C4.5 is a commercially available classification algorithm used to generate a decision tree using the idea<br />
of information entropy and was used for the analysis of a text categorisation system based on the<br />
decision tree model. The data set acquired from this study was used as an input to generate the decision<br />
tree using C4.5 algorithm. This research assumed uniform misclassification and performed a 10-fold<br />
validation for the classifier method.<br />
C4.5 defined the information gain formally as follows: (Anthony et al. 2004 and J.R Quinlan 1996).<br />
n<br />
Entropry(S) = Σ - Pi * log 2 Pi<br />
I=1<br />
Where: S = Data set, Sv = A partition of S according to the value of attribute A, n = The number of<br />
attributes A, |Sv| = The number of dataset in the partition, |S| = Total number of dataset in S.<br />
Gain(S,A) = Entropry(S)-Σ |S | Entropy(S )<br />
v v<br />
Value(A) |S|<br />
Where: X = Data set, n = The number of partition in S,pi = Proportion of Sv to S.<br />
Gain Ratio = Gain / Split Information<br />
Where split information is a value based on the column sums.<br />
A Bayes classifier is a simple probabilistic classifier based on Bayes theorem. The goal is to build a<br />
model. Given a set of data X = x1, x2, x3,…..,xn, a data mining problem is designed to uncover the<br />
properties of the distribution from which the set comes. The Bayes rule can be expressed as follows:<br />
P (h1|xi)= P (xi|h1)P (h1)/P (xi|x1)P (h1) + P (x1|h2)P (h2) (7)<br />
and called the posterior probability, while P(h1) is the prior probability. P(xi ) is the probability of the<br />
occurrence of data value and P(xi|h1) is the conditional probability ( Buntine W.1992 and Margaret H.<br />
2002).<br />
3.7 Performance evaluation<br />
Many different metrics are used in machine learning and data mining to build and evaluate models. The<br />
four performance measures including precision, recall, F-measure and accuracy are employed to<br />
evaluate models. A distinguished confusion matrix is obtained to calculate the four measures. The<br />
confusion matrix is a matrix representation of the classification results.<br />
Table1: A confusion matrix for a two-class classification<br />
Actual class<br />
Recognised<br />
Predicted as Positive Class Predicted as Negative Class<br />
Actual Positive class True Positive (TP) False Negative(FN)<br />
Actual Negative class False Positive(FP) True Negative(TN)<br />
Text classification rules are typically evaluated using performance measures from medical data. Common<br />
metrics for disease classification evaluation include recall, precision, accuracy F. Given a test set of N<br />
documents, the cells contain the counts for true positive (TP), false positive (FP), true negative (TN) and<br />
407<br />
(4)<br />
(5)<br />
(6)
Klaokanlaya Silachan and Panjai Tantasanawong<br />
false negative (FN), respectively. Clearly, N = TP + FP + TN + FN. The metrics for binary-decisions, the<br />
precision, recall, F-measure are defined as: (Gundersen et al.1996).<br />
4. Results<br />
TP<br />
Recall = (8)<br />
TN<br />
TP<br />
Precision = TP+FP<br />
(9)<br />
2*TP<br />
F-measure =<br />
2*TP+FP+FN<br />
(10)<br />
Accuracy = TN+TP (11)<br />
TN+ FN+FP+TP<br />
The medical dataset for our set of experiments is shown as follows:<br />
4.1 Extracted terms<br />
After extracting the symptom and disease terms, these can be reviewed as records and the<br />
corresponding terms as attributes of the record from the training dataset group. The example below<br />
shows the vocabulary used.<br />
Table 2: Sample of diagnosis single and multi-word<br />
weakness, paleness, chills, vomiting, cough, nausea, angina, syncope<br />
Single word<br />
Change in consciousness, shortness of breath, cardiac arrest, feeling of indigestion<br />
Multi-word<br />
Table 3: Samples of extracted symptom terms<br />
Documents Extracted Symptom Terms<br />
D-1<br />
chronic vapour, chest pain, low fever, panting, tired<br />
D-2<br />
headache, high fever, vapour, prickly heat<br />
D-3 encephalitis, have pains, pain and ache<br />
..... .......<br />
D-33 Angina, acute myocar(myocardium),backache, dizziness<br />
D-52 syncope, acute myocar<br />
Table 4: Sample document-term frequency data matrix<br />
Doc<br />
D-4<br />
D-33<br />
D-n<br />
TF1<br />
Angina<br />
TF2<br />
Syncope<br />
1 1<br />
...<br />
TFn<br />
Acute myocar<br />
1 - 1<br />
1 1 1<br />
408
4.2 Classification<br />
Klaokanlaya Silachan and Panjai Tantasanawong<br />
The implemented of C4.5 algorithm identified classifiers with weka tool from medical data. The medical<br />
data is divided into two sets of training set and test set. When the classification of diseases already will<br />
be able to see the pattern of the document and that the diagnosis of diseases The result of classification<br />
by disease in training set are shown as seen in Tables 5 and 6.<br />
Table 5: Sample of cardiovascular disease document classification<br />
Documents Cardiovascular Symptom Terms<br />
D-33<br />
D-52<br />
Angina, acute myocar, backache, Dizziness<br />
Syncope, Acute myocar<br />
D-87 Weakness, shortest of breath, angina<br />
Table 6: Sample of Cardiovascular disease classification<br />
Class Documents Symptom Terms<br />
Other<br />
Cardiovascular<br />
4.3 Evaluation performance<br />
D-1<br />
chronic vapour, chest pain, low fever, pant, be tired<br />
D-2 Have a headache, High fever, vapour, prickly heat, ..<br />
D-3 encephalitis , have pains, pain and ache, fever, itch<br />
........ ........<br />
D-33<br />
Angina, Acute myocar, backache, Dizziness<br />
D-52 Syncope, Acute myocar<br />
D-87 Weakness, shortest of breath, angina<br />
.. ........<br />
The data mining method used to build the model is classification. The instances in the dataset are<br />
representing the results of different types of testing to predict the accuracy of the cardiovascular disease<br />
document. The C4.5 and Naïve Bayes algorithms were used to predict the accuracy of cardiovascular<br />
medical dataset. This paper evaluation reports on the classifier are conducted with a Weka tool. The<br />
performance of the classifiers is evaluated and their results are analyzed. The results of the comparative<br />
analysis are based on 10 fold cross-validations. According to the attributes, the dataset divided into two<br />
parts such that 70% of the data is used for training and 30% is used for testing. For the purposes of this<br />
project, the data set was divided into a training set of 210 instances and a test set of 90 instances. See<br />
the performance comparison experiment in Table 7-8.<br />
Table 7: Evaluation on classification result using the training set<br />
C4.5 Naïve Bayes<br />
Precision 0.93 0.98<br />
Recall 0.95 0.95<br />
F-measure 0.97 0.97<br />
Accuracy 95.2 94.8<br />
Table 8: Evaluation of the accuracy of prediction results using the test set<br />
C4.5 Naïve Bayes<br />
Precision 0.96 0.95<br />
Recall 0.98 0.98<br />
F-measure 1 0.97<br />
Accuracy 99.8 99.6<br />
A comparison of the two algorithms with Precision, Recall and F-measure on medical training set and test<br />
set in a graph pattern seen in Figure 2 and 3.<br />
409
Klaokanlaya Silachan and Panjai Tantasanawong<br />
Figure 2: Precision, recall, f-measure of<br />
training set<br />
Figure 3: Precision, recall, F-measure of test<br />
set<br />
The results in Table 7-8 and Figure 2-3 show that the C4.5 algorithm can build the training set classifier<br />
better than Naive Bayes by a value 95.2%. In predicting the relationship of the symptoms to categories of<br />
cardiovascular disease, the test set has 99.8% prediction accuracy (percentage correct), and therefore is<br />
better than that of the Bayesian algorithm, with a precision rate of 0.96% and the recall at 1.0%. Finally,<br />
for the F-measure, which showed the relationship between the best queries and answer, recall is 0.98%.<br />
All results are acceptable.<br />
5. Conclusion<br />
This paper illustrates an approach where information extraction and features the selection of text medical<br />
data using sliding windows, TF-IDF techniques, with a multiword check and term weight-frequency.<br />
Information extraction(IE) is drawn from a medical text, and the data extraction is stored in a relational<br />
database of medical data. It is necessary to employ medical diagnostic data in the experimental<br />
evaluation. Data extraction is important for the classification of the documented information, including<br />
classification of whether or not cardiovascular disease is featured in major health information, called<br />
health text categories. It has been shown in this paper that classification, summary and probable<br />
diagnosis of cardiovascular disease can be carried out through data mining using rule induction. The<br />
commercially available C4.5 algorithm and Naïve Bayes algorithm were used as classifiers .The overall<br />
reliability was then checked and evaluated through a confusion matrix. precision, recall, F-measure, and<br />
accuracy were used as criteria for evaluating the performance of the training set with the accuracy values<br />
at 95.2%. The test set classifier for predictive accuracy of the C4.5 algorithm was marginally better than<br />
that of Bayes algorithm with the accuracy value at 99.8%. The follow-up to these test results would be to<br />
gain permission to evaluate the system against real data from a wide range of medical data records.<br />
References<br />
Anthony J. Myles, Robert N. Feudale, Yang Liu, Nathaniel A.Woody and Steven D. Brown. (2004) “An introduction to<br />
decision tree modeling”, [online] , Journal of chemometrics, Publish in Interscience.<br />
Appelt D, Israel D.(1999) ”Introduction to Information Extraction Technology”, In a tutorial for IJCAI-1999.<br />
Aurangzeb Khan, Baharum Baharudin, Lam Hong Lee, Khairrullah khan. (2010) “A Review of Machine Learning<br />
Algorithms for Text-Documents Classification”, Journal of Advance in information technology,VOL.1,No.1.<br />
B.H. Jun, C.S.Kim, H.Y.Song, J.Kim. (1997) “A new criterion in selection and discretization of attributes for the<br />
generation of decision tree”, IEEE Transaction on Pattern Analysis and Machine<br />
Intelligence,Vol.19,No.12,pp.1371-1375.<br />
Buntine W. (1992) “Learning Classification Trees”, Statistics and Computing.<br />
Dalila Bekhouche,Yann Pollet, Bruno Grilheres, and Xavier Denis. (2004) “Architecture of a Medical Information<br />
Extraction System”, NLDB2004, LNCS3136 Springer- Verlag Berlin Heidelberg, pp. 380-387, 2004.<br />
Dittapol Muntham, Lily Ingsrisawang. (2009) ”An Application of Decision Tree Algorithms for Diagnosis of the<br />
Respiratory System: A case Study of Pranakorn Sri Ayudthaya Hospital”, Journal of Health System, volume 1.<br />
Gundersen ML, Haug PJ, Pryor TA (1996) “Development and evaluation of a computerized admission diagnoses<br />
encoding system”, Comp Biomed Res, 29(5), 351–72.<br />
J.R Quinlan. (1996) ”Induction of Decision Tree”, Machine Learning, Vol, 1, No.1, pp.81-106.<br />
J.R Quinlan.(1993) ”C4.5: Programming for machine learning”, San Mateo,Calif,Morgan Kaufmann.<br />
Keku E, Rosamond W, Taylor HA Jr, Garrison R, Wyatt SB, Richard M, Jenkins B, Reeves L, Sarpong D. (2005)<br />
”Cardiovascular disease event classification in the Jackson Heart Study : methods and procedures”, Ethnicity &<br />
Disease, Volume 15, pp.73-76.<br />
Kent W. Bridges. (1992) “Rule-base Decision Making : A way to determine which alien species to control”, Alien Plant<br />
Invasions in Native Ecosystems of Hawaii: Management and Research.<br />
410
Klaokanlaya Silachan and Panjai Tantasanawong<br />
K.S. Kavitha, K.V.Ramakrishnan, Manoj Kumar Singh.(2010). ”Modeling and design of evolutionary neural network<br />
for heart disease detection”, IJCSI international journal of Computer science, Issue, Vol. 7, Issue 5.<br />
Margaret H., dunham. (2002) ”Data mining Introductory and Advanced Topics”, Chapter3, p.52.<br />
Maria Grineva, Maxim Grinev and Dmitry Lizorkin (2009) ”Effective Extraction of Thematically grouped Key Terms<br />
From Text “, Association for the Advancement of Artificial Intelligence.<br />
Parikshit Sondhi and Manish Gupta and ChengXiang Zhai and Julia Hockenmaier. (2010) “Shallow Information<br />
Extraction from Medical Forum Data”, Coling 2010: Poster Volume, pages 1158–1166, Beijing.<br />
Reddy,K,S. (2007) “India wakes up to the threat of cardiovascular diseases”. J.Am.Coll, Cardiol, 50, 1370-1372.<br />
S.Izumi, N.Yasuda, G.Itabashi, Y.Kato, K.Takahashi, T. Suganuma,N. Shiratori(2007) “A Health Advice Derivation<br />
System based on an Ontology”, 6 th International and 2nd Asian Semantic Web <strong>Conference</strong><br />
(ISWC2007+ASWC2007).<br />
Se-Chul Chun, Jin Kim, Ki-Baik Hahm, Yoon-Joo Park, Se-Hak Chun . .(2008) ”Data mining technique for medical<br />
informatics detecting gastric cancer using case-based reasoning and single nucleotide polymorphisms”, Expert<br />
Systems Volume 25, Issue 2, pages 163–172.<br />
T.Gruber (1995). “Toward Principles for the Design of Ontologies for Knowledge Sharing”, International Journal of<br />
Human - Computer studies, p.43, 907 – 928. WHO(1992). ” International Statistical Classification of Disease<br />
and Related Health Problem”, tenth revision (ICD-10), Geneve.<br />
Zhi-Hua Zhou, Yuan Jiang. (2003) “Medical diagnosis with C4.5 rule preceded by artificial neural network ensemble”,<br />
Information Technology in Biomedicine, IEEE Transactions on.Issue1. p.31-42.<br />
411
Optimizing Information Technology Value Governance<br />
Framework Based on Val IT<br />
Ali Suzangar, Mehrdad Kalantarian , Shirin Nasher , Mohammad Kajbaf and Negar<br />
Madani<br />
Infoamn IT Consultancy CO., Tehran, Iran<br />
a.suzangar@infoamn.com<br />
m.kalantarian@infoamn.com<br />
sh.nasher@infoamn.com<br />
m.kajbaf@infoamn.com<br />
n.madani@infoamn.com<br />
Abstract: Nowadays, information technology investment is crucial for organizations. Four IT investment principles<br />
are “risk”, “modeling”, “management” and “governance and compliance”. There are different frameworks such as<br />
ITIL, COBIT, Val IT and Risk IT which provide best practices to those IT investment principles. The Risk IT<br />
framework is about business risks related to the use of information technology. The connection to business is<br />
founded in the principles on which the framework is built, i.e. effective enterprise governance and management of IT<br />
risk. Val IT is a governance framework and complementary of COBIT from the business and financial perspective.<br />
Since Val IT framework only provides enterprise with the mechanism that it requires for monitoring and optimizing the<br />
IT business value, therefore there is no tool to calculate and evaluate information technology investment return. ITIL<br />
financial management process has three main aspects that provides guidance on how controlling spending and cost.<br />
In order to apply all direct and indirect cost evaluating in new information technology investment, a hierarchy decision<br />
making model is necessary. Since these frameworks activities are considered based on three principles of investing,<br />
i.e. risk, management and governance, none of them has a process or approach to create a model based on the<br />
investment costs in order to make final decision. So a comprehensive framework that based on all four investment<br />
principles, can be as a good solution for evaluating new information technology investments. In this paper, a RMMG<br />
framework that considers IT investment evaluation decision making model and financial process among the value<br />
governance and compliance activities, and also evaluating IT investment risk will be proposed to aid those<br />
organizations that confront by IT value investment issues.<br />
Keywords: IT value governance, Val IT, ITIL, Risk IT, decision making model, RMMG framework<br />
1. Introduction<br />
IT governance has been defined by ISACA as “the responsibility of executives and the board of directors,<br />
and consists of the leadership, organizational structures and processes that ensure that the enterprise’s<br />
IT sustains and extends the organizations strategies and objectives” (ISACA 2009a), (ITGI 2007b).<br />
IT governance can thus be pictured as focusing primarily on the following five areas (ISACA 2009a):<br />
Strategic alignment: achieving the goals and strategies of an enterprise through the coherent<br />
undertaking of activities by the different governance structures or management levels within an<br />
organization.<br />
Value delivery: creating new value for the enterprise through IT, maintaining and increasing value<br />
derived from existing IT investments, and eliminating IT initiatives and assets that are not creating<br />
sufficient value for the enterprise.<br />
Risk management: addressing IT-related risks. IT risk is the business risk associated with the use,<br />
ownership, operation, involvement, influence and adoption of IT within an enterprise.<br />
Resource management: ensuring that the right capabilities are in place to execute the strategic plan<br />
and sufficient, appropriate and effective resources are provided.<br />
Performance measurement: tracking the achievement of the objectives of the enterprise’s IT-related<br />
services and solutions and compliance with specific external requirements (ISACA 2009a).<br />
Three IT governance frameworks have been developed by ISACA: COBIT, Val IT and Risk IT. The<br />
integrated application of processes defined by COBIT, Val IT and Risk IT can help enterprises<br />
significantly improve IT governance and manage IT-related risks (ITGI 2007b).<br />
The best IT investments are those which help to maximize the value of the firm. They also contend that to<br />
maximize the value of the firm, IT investment decisions need to be able to maximize IT benefits while<br />
minimize IT risks (Joseph Wen, Yen and Lin 1998).<br />
412
Ali Suzangar et al.<br />
There are four main important issues about IT investment and return benefits as follows:<br />
1. Management of business risk is an essential component of the responsible administration of any<br />
enterprise. Almost every business decision requires the executive or manager to balance risk and<br />
reward (ISACA 2009c). The recognition of investment risk as an important component in IT<br />
investment decision making has long been recognized (Epstein and Rejc 2004-2005), (Joseph Wen,<br />
Yen and Lin 1998).<br />
2. Since organizations are often unable or unwilling to justify the expenditure to improve and develop<br />
services (OGC 2007), it is important to identify some methods and models to calculate the investment<br />
direct cost and indirect cost (tangible and intangible cost criteria).<br />
3. Always IT organizations looking for controlling the spending of money within the organization<br />
(annual budgeting), set of processes to account the budget that is allocated to information technology<br />
and also charging mechanisms to evaluate IT services delivered costs. As a result IT organizations<br />
need the financial management included budgeting, accounting and charging processes (OGC<br />
2007).<br />
4. IT organizations like to have governance framework for value management in a manner that is fully<br />
integrated with overall enterprise governance, provide strategic direction for the investment decisions,<br />
define the characteristics of portfolios required to support new investments and resulting IT services,<br />
assets and other resources and improve value management on a continual basis (ITGI 2008b).<br />
All above issues can be considered as four main IT investment principles as follows:<br />
1. Value risk assessment<br />
2. Value modeling<br />
3. Value management<br />
4. Value governance and compliance<br />
Since no existence value frameworks can cover all mentioned principles, this paper will tend to propose a<br />
comprehensive framework based on these principles.<br />
In literature review, first Risk IT framework will be defined. Then, an analytic hierarchy IT investment<br />
decision making model based on multi criteria multi objective method will be introduced. Note that since<br />
the designing methodology of an analytic hierarchy IT investment decision making model is not the main<br />
concern of this paper, it would be introduced briefly in 2.2. But the complete approach has been<br />
submitted as a separate paper to this conference titled “Evaluation of IT investment methods and<br />
proposing a decision making model”. After that, ITIL financial management process will be defined in<br />
three parts to control spending costs in the organization. In the consequence, Val IT framework will be<br />
described as a mechanism for many organizations to assess their current IT value state and recognize<br />
their pain points, and other symptoms in creating IT enabled value.<br />
After comparing these frameworks, the RMMG proposed framework will be defined with its related layers<br />
and processes. In conclusion, the advantages of proposed RMMG will be described.<br />
2. Literature review<br />
2.1 Risk IT framework<br />
The all-encompassing use of IT can provide significant benefits to an enterprise, but it also involves risk.<br />
Due to IT’s importance to the overall business, IT risk should be treated like other key business risks,<br />
such as strategic risk, environmental risk, market risk, credit risk, operational risks and compliance risk,<br />
all of which fall under the highest ‘umbrella’ risk category. While these other risks have long been<br />
incorporated into corporate decision-making processes, too many executives tend to relegate IT risk to<br />
technical specialists outside the boardroom (ISACA 2009c).<br />
The Risk IT framework is about IT risk, in other words, business risk related to the use of IT. The<br />
connection to business is founded in the principles on which the framework is built, i.e., effective<br />
enterprise governance and management of IT risk. The three domains of the Risk IT framework are as<br />
follows: Risk Governance, Risk Evaluation and Risk Response (ISACA 2009c).<br />
413
Ali Suzangar et al.<br />
The risk management process model groups key activities into a number of processes. These processes<br />
are grouped into three domains. The process model will appear familiar to users of COBIT and Val IT<br />
(ISACA 2009c).<br />
2.2 Analytic hierarchy it investment decision making model<br />
Since the intangible criteria are complex to evaluate, therefore this action may take a long time and<br />
efforts. So many managers prefer to neglect them and just focus on the financial factors (Borenstein<br />
2005). They think that IT investment affects on their organization and enhances productivity efficiency in<br />
situ. But indeed return on IT investment has a potential latency to create value just because of these<br />
intangible factors. So it shows the importance of these criteria to have an appropriate investment<br />
evaluation. The existence of the pre-defined criteria hierarchy model based on the consensus of many<br />
specialist, can structured the decision making process systematically. Figure1 shows this proposed<br />
model.<br />
For this model, first tangible and intangible criteria were identified and then categorized in different criteria<br />
and sub criteria. As a consequence, an importance rank was determined to them according to the<br />
experiment of chief managers. Based on these ranking, at last this hierarchy model in order to consider<br />
all these tangible and intangible criteria in IT investment evaluation was proposed. Note that this<br />
proposed model is general model and independent from the specific IT investment to aid investment<br />
decision making process.<br />
Figure 1: Analytic hierarchy decision making model<br />
So there was a consensus that this proposed model has improved the decision process in order to have<br />
consistent and complete criteria based on different tangible and intangible factors for IT investment<br />
evaluation.<br />
414
2.3 ITIL financial management process<br />
Ali Suzangar et al.<br />
Financial Management is the sound stewardship of the monetary Resources of the organization. It<br />
supports the organization in planning and executing its business objectives and requires consistent<br />
application throughout the organization to achieve maximum efficiency and minimum conflict.<br />
Within an IT organization it is visible in three main processes as follows: Budgeting, IT accounting, and<br />
charging (OGC 2007).<br />
2.3.1 Budgeting<br />
Is the Process of predicting and controlling the spending of money within the organization and consists of<br />
a periodic negotiation cycle to set budgets (usually annual) and the day-to-day monitoring of the current<br />
budgets.<br />
The aim of Budgeting is that the actual costs match the budget (predicted costs). This budget is usually<br />
set by negotiations with the Customers who are providing the funds. Good Budgeting is essential to<br />
ensure that the money does not run out before the period end. Where shortfalls are likely to occur the<br />
organization needs early warning and accurate information to enable good decisions to best manage the<br />
situation (OGC 2007).<br />
2.3.2 IT accounting<br />
IT Accounting is the set of processes that enable the IT organization to account fully for the way its<br />
money is spent (particularly the ability to identify costs by Customer, by service, by activity). It usually<br />
involves ledgers and should be overseen by someone trained in accountancy (OGC 2007).<br />
2.3.3 Charging<br />
Charging is the set of processes required to bill Customers for the services supplied to them. To achieve<br />
this requires sound IT Accounting, to a level of detail determined by the requirements of the analysis,<br />
billing and reporting processes (OGC 2007).<br />
2.4 Val IT<br />
Executives, even if they are aware of the need for more effective governance and management of<br />
information technology, may not recognize that many of the day-to-day business challenges they face<br />
involve issues of value management. Val IT provides proven value management principles, processes<br />
and practices to enable enterprises to maximize the delivery of business value from investments involving<br />
IT (ISACA 2010), (ITGI 2008c).<br />
There are some internal trigger (major IT project failures, serious budget overruns and etc) and external<br />
events (a major shift in the marketplace and etc) that are likely to convince executives to improve their<br />
enterprise value management practices (ISACA 2009b).<br />
After identifying these challenges within organization, the Val IT framework provides useful guidance on<br />
proven processes and practices that enable effective governance of investments involving IT.<br />
This guidance is found in the three domains of Val IT (ITGI 2008b):<br />
Value Governance (VG): ensuring that value management practices are embedded in the enterprise<br />
Portfolio Management (PM): ensuring that the enterprise secures optimal value across its portfolio of<br />
investments involving IT<br />
Investment Management (IM): ensuring that the enterprise’s individual investments each contribute<br />
the value expected of them.<br />
Val IT Processes are a collection of interacting activities undertaken in accordance with management<br />
practices (ITGI 2008b).<br />
3. Comparing<br />
According to previous discussion, it can be considered four main principles in IT investment discussing,<br />
i.e. “value risk assessment”, “value modeling”, “value management” and “value governance and<br />
compliance”. Each mentioned frameworks focuses on one or more than one of these principles. So in<br />
415
Ali Suzangar et al.<br />
Table 1, these frameworks are compared in order to show how much they cover the aspects of<br />
investment evaluation. The main target of the frameworks and also the processes and domains goals that<br />
are included in those frameworks are assumed as this table comparison criterion.<br />
Table1: comparison table<br />
Val IT<br />
ITIL<br />
Risk IT<br />
Value Risk Assessment<br />
Low<br />
Low<br />
High<br />
Value Modeling<br />
Non<br />
Non<br />
Non<br />
Value Management<br />
Medium<br />
High<br />
Low<br />
Value Governance and compliance<br />
High<br />
Medium<br />
Low<br />
Based on this table, all frameworks have concentrate on each specific principle. It means that e.g. Val IT<br />
focus on the governance aspect is high, since management and risk aspect is not considered in the high<br />
level. The main point of this table is that none of them concentrate on the value modeling principle.<br />
4. Proposing RMMG framework<br />
The proposed framework scenario is defined as follows:<br />
Assume that, a new IT project is proposed to an organization to invest on it or the organization decides to<br />
invest on a new IT project based on its own objectives and strategies. By this trigger, some readiness<br />
activities should be done providing the safe investment and appropriate evaluation for managers. As an<br />
important principle, organization requires assessing risks before and after new investment to monitor and<br />
evaluate all investment opportunities and threats in order to able mitigating and responding them. Then<br />
beside the risk assessment, a model based on tangible and intangible benefit factors of proposed IT<br />
project can help managers to make a final decision on allocating their budget or not.<br />
Additionally, the organization requires a management process that can predict and control its costs and<br />
also account fully for the way its money is spent. So this process in order to provide requirement of IT<br />
investment justification is demanded. And at last, value governance and compliance is applied to take<br />
full advantage of proposed IT project, maximizing benefits, capitalizing on opportunities and gaining<br />
competitive advantages.<br />
By this scenario, the organization will require an overall comprehensive framework that is consisted of<br />
value risk assessment, value modeling, value management and value governance and compliance.<br />
Figure 2 illustrates this proposed framework that is called Risk, Modeling, Management and Governance<br />
(RMMG).<br />
The RMMG consists of four layers as Figure 2 illustrates. Each provides the guidance necessary for a<br />
comprehensive framework as required by the mentioned investment principles:<br />
Value Risk Assessment<br />
Value Modeling<br />
Value Management<br />
Value Governance and Compliance<br />
RMMG framework structure has been fully changed against Risk IT, ITIL, and Val IT frameworks. But<br />
those framework processes can be located in different layers of the RMMG as follows based on the main<br />
goal of the processes, and also the domains targets that are included in those frameworks.<br />
Each layer addresses capabilities having direct impact on the new information technology project<br />
investments. The structure of the layer is in the form of a lifecycle. It is iterative and multidimensional. The<br />
layer is expected to provide structure, stability and strength to investment capabilities with durable<br />
principles, methods and tools. This serves to protect investments and provide the necessary basis for<br />
value risk, management and governance. The RMMG can be adapted for use in various business<br />
environments and organizational strategies. The layers of RMMG definitions are as follows:<br />
416
Figure2: RMMG framework<br />
4.1 Value Risk Assessment<br />
Ali Suzangar et al.<br />
The Value Risk Assessment layer, as the core of the framework, provides how assessing investment<br />
risks before and after investing and implements activities to analyze and react to those risks. Value Risk<br />
Assessment is useful in the context of Value Modeling, Value Management, and Value Governance and<br />
Compliance. Value Risk Assessment defined processes that are as the same as Risk IT framework<br />
related processes as follows:<br />
Establish and maintain a common risk view (Risk IT)<br />
Integrate with ERM (Risk IT)<br />
Make risk-aware business decisions (Risk IT)<br />
Collect data (Risk IT)<br />
Analyze risk (Risk IT)<br />
Maintain risk profile (Risk IT)<br />
Articulate risk (Risk IT)<br />
Manage risk (Risk IT)<br />
React to events (Risk IT)<br />
4.2 Value Modeling<br />
The Value Modeling layer is considered in two parts, i.e. direct evaluation and indirect evaluation in order<br />
to offer the analytic hierarchy decision making model based on these two evaluation types of investment.<br />
In the indirect evaluation part, it covers all existence investment intangible criteria for converting them to<br />
quantitative cost. The scope of this layer includes developing a decision making model for managers to<br />
understand the weight of each investment criteria according to their objectives and strategies. The<br />
complete approach has been submitted as a separate paper to this conference titled “Evaluation of IT<br />
investment methods and proposing a decision making model”.<br />
417
4.3 Value management<br />
Ali Suzangar et al.<br />
The value management layer practices includes controlling the spending of money within the organization<br />
(annual budgeting), set of processes to account the budget that is allocated to information technology<br />
and also charging mechanisms to evaluate IT services delivered costs. It provides the strategic direction<br />
for investments, the desired characteristics of the investment portfolio, and the resources and funding<br />
constraints within which portfolio decisions must be made. Also, defines potential programs based on<br />
business requirements, determines whether they are worthy of further consideration, and develops and<br />
passes business cases for candidate investment programs to portfolio management for evaluation. One<br />
of the processes of this layer is from ITIL and the others are Val IT related processes as follows:<br />
Financial Management (ITIL)<br />
Establish strategic direction and target investment mix (Val IT)<br />
Determine the availability and sources of funds (Val IT)<br />
Manage the availability of human resources (Val IT)<br />
Evaluate and select programs to fund (Val IT)<br />
Monitor and report on investment portfolio performance (Val IT)<br />
Optimize investment portfolio performance (Val IT)<br />
Develop and evaluate the initial program concept business case (Val IT)<br />
Understand the candidate program and implementation options (Val IT)<br />
Develop the program plan (Val IT)<br />
Develop full life-cycle costs and benefits (Val IT)<br />
Develop the detailed candidate program business case (Val IT)<br />
Launch and manage the program (Val IT)<br />
Update operational IT portfolios (Val IT)<br />
Update the business case (Val IT)<br />
Monitor and report on the program (Val IT)<br />
Retire the program (Val IT)<br />
4.4 Value Governance and Compliance<br />
The Value Governance and Compliance layer establishes the overall framework, including defining the<br />
portfolios required to manage investments and resulting IT services, assets, and resources. This layer<br />
monitors the effectiveness of the overall governance framework and supporting processes, and<br />
recommends improvements as appropriate. The processes of this layer are as same as Val IT value<br />
governance related processes as follows:<br />
Establish informed and committed leadership (Val IT)<br />
Define and implement processes (Val IT)<br />
Define portfolio characteristics (Val IT)<br />
Align and integrate value management with enterprise financial planning (Val IT)<br />
Establish effective governance monitoring (Val IT)<br />
Continuously improve value management practices (Val IT)<br />
5. Conclusion<br />
Organizations apply different IT governance and management frameworks in order to reach their<br />
objectives and strategies such as improving their services and gaining competitive advantages, through<br />
new information technology investment decisions. According to this paper, we can consider four main<br />
principles to evaluate a new IT project investment, i.e. Value Risk Assessment, Value Modeling, Value<br />
Management and Value Governance and Compliance. There are different frameworks based on these<br />
principles. Since each of these frameworks considers one or two mentioned evaluating principles, so the<br />
existence of the comprehensive framework can be applicable for those organizations that want to<br />
implement an appropriate processes and activities to have an appropriate and exact evaluating in all new<br />
418
Ali Suzangar et al.<br />
investment decisions. Hence in this paper, a new conceptual and comprehensive framework was<br />
identified and illustrated by four layers in order to cover and consider all those principles that was called<br />
RMMG framework.<br />
This proposed framework, in addition to assess value risk, implement value management and<br />
governance activities, it evaluates direct and indirect investment costs to create an analytic hierarchy<br />
decision making model to help organization in their investments. It states the RMMG strength point.<br />
References<br />
Borenstein, D. and Baptista Betencourt, P. R. (2005) “A Multi-criteria Model for the Justification of IT Investments”,<br />
INFOR Journal, Vol. 43, No. 1, Feb, pp. 1-21.<br />
Epstein, M. and Rejc, A. (2004-2005) “Measuring the payoffs of IT investments”, CMA Management Journal, Vol. 78,<br />
No. 8, pp. 20-25.<br />
ISACA (2009a) Implementing and Continually Improving IT Governance, Rolling Meadows, IL 60008 USA.<br />
ISACA (2009b) IT Value Special Compilation, Rolling Meadows, IL 60008 USA.<br />
ISACA (2009c) The Risk IT Framework, Rolling Meadows, IL 60008 USA.<br />
ISACA (2010) Value Management Guidance for Assurance Professionals: Using Val IT 2.0, Rolling Meadows, IL<br />
60008 USA.<br />
ITGI (2007b) IT Governance Roundtable: IT Governance Trends, Rolling Meadows, IL 60008 USA.<br />
ITGI (2008a) Enterprise Value: Governance of IT Investments Getting Started With Value Management, Rolling<br />
Meadows, IL 60008 USA<br />
ITGI (2008b) Enterprise Value: Governance of IT Investments the Val IT Framework 2.0, Rolling Meadows, IL 60008<br />
USA<br />
ITGI (2008c) IT Governance Global Status Report, Rolling Meadows, IL 60008 USA.<br />
Joseph Wen, H., Yen, D.C. and Lin, B. (1998) “Methods for measuring information technology investment payoff”,<br />
Human Systems Management Journal, Vol. 17, No. 2, January, pp. 145–153.<br />
OGC (2007) Service Design ITIL service management practices v.3, London: the stationary office.<br />
419
Developing an Outpatient Electronic Medical Record System<br />
in Taiwan From a Physicians’ Perspective<br />
Hsiao-Ting Tseng, Pei-Ting Chang and Ray Chang<br />
National Chung Cheng University, Chia-yi, Taiwan<br />
appleapple928@gmail.com<br />
chang.payting@gmail.com<br />
misicc.pig@gmail.com<br />
Abstract: Electronic medical record systems (EMRS) are regarded as an innovative administrative tool in healthcare<br />
to integrate patient information. Physicians can use an EMRS to receive integrated information to provide better<br />
outpatient service. Many hospitals worldwide, including in Taiwan, have either implemented or plan to implement<br />
EMRS to reduce or replace the traditional paper-based charts and patient records. The key reason for some failures in<br />
implementation of EMRS is that system design was not user-friendly, i.e. it did not take into account physicians’ real<br />
needs. This research employed a case study to explore the development of an Outpatient Electronic Medical Record<br />
System (OEMRS) that was developed on the basis of physicians’ expectations. A qualitative interview confirmed that<br />
physicians’ participation can significantly affect their satisfaction and acceptance toward OEMRS use.<br />
Keywords: physician’s perspective, electronic medical record system, outpatient, satisfaction, acceptance<br />
1. Introduction<br />
According to the annual report of the Institute of Medicine for the year 1999, an estimated 98,000 people<br />
died because of medical negligence directly or indirectly. With IT applications, the casualty figure can be<br />
reduced to 55% (Drew, Carolyn and Robert, 2004). IT applications can improve the quality of medical care<br />
significantly. Since then, application of information technology to improve healthcare efficiency and to<br />
reduce healthcare expenditure has become an important trend among healthcare institutions.<br />
Electronic medical records systems are increasingly being used to keep record of medical procedures,<br />
interactions, treatment and procedures by physicians, medical staff and even patients. Using<br />
computer-based technologies to convert traditional medical records into electronic file formats, and to<br />
allow medical personnel to access this system via computers is called electronic medical records or EMR.<br />
According to the Medical Records Institute, electronic medical records are defined as electronic<br />
information of an individual’s lifelong health records. The coverage of Electronic Medical Records (EMR)<br />
includes automatic formatting of paper-based records and other health information, such as medication<br />
records, examinations charts and images of X-rays, etc.<br />
EMR have been used at all-levels in healthcare institutions to cope with a variety of medical service<br />
demands, such as teaching, evaluation, telemedicine and chronic care. According to a survey by the<br />
Bureau of National Health Insurance in 1999, about 98% of Taiwan’s medical institutions have introduced<br />
at least one (of several) computer-based technology to assist medical practitioners. EMR can reduce<br />
operational costs in hospitals and save space required for storage of physical paper-based records;<br />
physical space availability is very limited in Taiwan. Furthermore, patients’ medical records in different<br />
healthcare institutions can be integrated by EMR. It can not only reduce wasteful consumption of medical<br />
resources but also provide patients with continuous and high-quality medical services.<br />
The main users of EMR are healthcare staff and physicians. Active participation of physicians in the<br />
process of promoting electronic medical records can significantly enhance application and introduction of<br />
EMR in the healthcare industry (citation). Peter et al. (2006) showed that active participation of<br />
physicians in the process of introduction of EMR can help make the interface of EMR systems more<br />
suitable for physicians’ needs. Moreover, with participation of physicians, system implementation can<br />
become a much smoother process. It is important for a EMR system to meet physicians’ needs and yet<br />
there are not many studies on doctors’ views on efficacy of EMR systems. In this study, we examine the<br />
working of a hospital which has developed an Outpatient Electronic Medical Records System (OEMRS) as<br />
a case study to explore and understand the development and evaluation of the system from physicians’<br />
perspectives.<br />
2. Background<br />
Electronic medical records include records of all healthcare related information required by clinical workers<br />
for providing effective and continuous healthcare. Electronic medical records not only constitute an<br />
420
Hsiao-Ting Tseng et al.<br />
electronic record of diagnoses and prescribed treatments, etc., but they also facilitate medical information<br />
integration. The goals of this system are storage and management of medical history of individual<br />
patients, as well as medical knowledge used for their treatments at different times. Hence, EMR can help<br />
healthcare personnel make related decisions and improve the quality of medical care, besides making<br />
hospital management more efficient. Electronic medical records should focus on one objective: to record<br />
long-term medical care information contiguously to provide continuous, useful and high-quality medical<br />
care.<br />
In the system’s development phase, participation of users is regarded as the key factor for success. Lee<br />
and Kim (1999) showed that good communication between information workers and users can have<br />
positive effect on eliminating perception differences (Lee and Kim, 1999). The literature also indicates that<br />
in the process of system development, the more there is users’ participation, the deeper is the influence<br />
on the final product (Zmud and Cox, 1979).<br />
In an outpatient medical care system, physicians are responsible for providing medical care services like<br />
prescriptions and examinations. Prior studies have showed that if physicians participate in the<br />
development process of the OEMRS, the interface of OEMRS can meet the physicians’ needs, and thus<br />
the need for educating and training them to use the OEMRS can be reduced (Peter et al., 2006). Using an<br />
electronic medical record system, physicians’ work efficiency, data processing procedures, and time<br />
management can be greatly improved (Richard, Elaine and Don, 1997). However, the most important<br />
concerns of physicians about OEMRS are patient safety and the working process. These factors can<br />
result in impedance to introduction of an OEMRS (Robert, 2006). Thus, if we physicians act as advocates<br />
or consultants for an electronic medical system, resistance can be reduced (Peter et al., 2006).<br />
3. Material and methods<br />
Most previous research about electronic medical records has focused on technical perspectives,<br />
technology combination, sharing mechanism, related standards and legal perspectives such as privacy<br />
concerns. Few studies have tried to understand the role physicians’ participation can play in the<br />
development process. Past research has showed that lower participation of medical staff can be<br />
considered a dilemma while planning introduction of electronic medical records, whether in small clinics or<br />
large hospitals. Therefore, this research has chosen to examine the case of a hospital that invited<br />
physician participation in developing an OEMRS.<br />
3.1 Case hospital<br />
The subject 1,000-bed hospital is located in southern Taiwan. The outpatient department of the hospital<br />
serves approximately 3,600 people daily and offers over 40 medical specialties. A new building was<br />
recently constructed to offer more outpatient services and preventive medical treatment.<br />
Both computerized and paper-based medical records are used in the outpatient department of the<br />
hospital. While a computerized physician order entry system is in use, paper-based medical records can<br />
still be obtained for use by physicians as references for patients’ histories. Outpatient consultation,<br />
emergencies, admission records, examinations images and pathology reports are available in physical<br />
form. Past records cannot be displayed on the same page when using the same program for browsing. If<br />
physicians need to query some related information, they need to access a sub-system or alternate<br />
windows. Due to the inconvenience of this process, most physicians still use paper medical records. In<br />
response to expansion of outpatient department services in the future, and to address difficulties in<br />
cross-referencing of medical records, it is necessary to develop an integrated outpatient electric medical<br />
records system.<br />
A high compliance rate was observed among information department staff and clinic physicians at the<br />
case hospital. The medical vice-president and the administrative vice-president showed great interest in<br />
this project and provided all support, hoping that the OEMRS can solve the difficulties of cross-referencing<br />
of medical records. A project team was formed with the deputy superintendent of the hospital as the<br />
convener for implementation of the OEMRS at the case hospital. Representatives from the administrative<br />
department were the president’s special assistant and five members from the information department.<br />
One physician was chosen from each major department as the consultant physician. All consultant<br />
physicians were active and showed strong interest in OEMRS development. In addition, they interacted<br />
well with the information department staff. Consultant physicians said they would like to participate in the<br />
421
Hsiao-Ting Tseng et al.<br />
development process of the OEMRS to further solve the difficulties of cross-referencing of medical<br />
records.<br />
3.2 System development<br />
In the process of development of OEMRS, the clinical processes that physicians wanted covered were<br />
listed and incorporated in the OEMRS to make the clinical process smooth. Secondly, in-depth interviews<br />
and questionnaires were used for examining the outpatient department’s work flow and practices. System<br />
development steps are shown in Figure. 1.<br />
Figure 1: System development steps<br />
The prototyping life-cycle, brought by Boar (1984) was employed as a system development method for<br />
this study. A visible and operable prototype was used as a communication tool between developers and<br />
related users (e.g. physicians). As a result, the time and cost of system analysis can be reduced. There<br />
are 3 factors affecting the decision whether an information system is suitable for the prototyping method<br />
(Table 1): (1) degree of environmental change, (2) degree of structuring of the decision-making process,<br />
and (3) similar previous design experiences (Kendall and Kendall, 1988).<br />
Table 1: Factors of system usage<br />
422
Hsiao-Ting Tseng et al.<br />
The main reasons for choosing the prototyping method to develop the OEMRS are as follows.<br />
Demands from physicians are often not clear, or are vague, and these demands cannot be listed easily.<br />
Meanwhile, users’ demands may also change over time. In addition, outpatient department’s physicians in<br />
the case hospital have long experience in clinical work but lack experience in OEMRS development and<br />
usage. Required features and specifications of an electronic medical records system may not be easily<br />
identified. Using the prototyping life cycle system, the system can be modified repeatedly to fit users’<br />
demands. Also, suitability of system interface can be confirmed with users to obtain better suggestions for<br />
further modifications. Based on a deep study of working of the case hospital’s outpatient department and<br />
interviews with related users, the users’ demands were identified as shown in Table 2.<br />
Table 2: System demands<br />
The needed software/hardware are tabulated in Table 3.<br />
Table 3: System develop environment and tools<br />
423
Hsiao-Ting Tseng et al.<br />
The OEMRS integrates other systems as shown in Figure 2. After patient’s registration, his/her basic<br />
information is recorded (or retrieved from existing records). If physiological or biochemical or radiological<br />
examinations are required, examination sheets are signed by the concerned department. Data are then<br />
transmitted back to the OEMRS through PACS. Physicians can see all records of patients’ diagnoses and<br />
determining treatments.<br />
Figure 2: OEMRS process diagram<br />
The main functions of OEMRS are shown in Figure 3. They are: (1) Logon portal; (2) Medical record main<br />
page; (3) History of visit records; (4) Special medication records (offering special chemical treatment<br />
medicine record and general medicine query function); (5) Examination and check reports; (6) Case<br />
management records (offering case management records including cancer cases, chronic kidney disease<br />
and diabetes cases); (7) Special check records (offering special check reports).<br />
Figure 3: OEMRS function infrastructure diagram<br />
424
Hsiao-Ting Tseng et al.<br />
The OEMRS integrates the above functions in a window on the computer screen which can be adjusted to<br />
meet the physician’s needs. Major treatment records and related information for a patient are put together<br />
on a first screen with specific font colors or alert windows to remind physicians of patient previous<br />
treatment histories.<br />
Physicians can log on to OEMRS from the original computerized physician order entry (CPOE) or PACS.<br />
After logging on, users can input patients’ medical record numbers to find related information and view<br />
PACS images. If data are input into the OEMRS via the original interface, medical record numbers of<br />
patients are directly used by the OEMRS to show the related information. Discharge medical note shows<br />
information about a patient’s diagnostic records and examinations from admission to discharge, disease<br />
history, and discharge medication records. Operations notes provide patients’ surgery related records to<br />
meet surgeons’ needs. As a result, timely viability of surgery records and operations notes queries can be<br />
fulfilled.<br />
Outpatient physicians can browse patients’ history to check outcomes of previous visits and compare<br />
different records to observe changes of chief complaints in order to adjust medication. Special<br />
medication records answer any queries for special chemical treatments which are distinguished from<br />
general medicine. Furthermore, it can be sorted and arranged by medicine name, department, or date of<br />
prescription to meet physicians’ needs. Outpatient physicians can quickly read check reports through<br />
Examination by choosing specific data bases to browse information they need. Pre-established<br />
sequencing patterns are listed by date, or by users needs. Extreme Values higher than standard are<br />
shown as red (H), and those of lower values are shown as red (L) to alert physicians. This function can<br />
fulfill physicians’ needs for collection of related continuous examination data, neurological examination<br />
data, and hints of extraordinary examination results. Physicians usually have to compare different<br />
examination reports of different dates in the outpatient consulting process for determining further<br />
treatment. Reports of the same items of examination and tests carried out at different times are compiled<br />
in the same window. Special items are shown in red fonts to alert the physicians. Examination statistics<br />
can be converted into line charts by physicians to help patients understand the changes in their<br />
examination data. Such graphical presentations of examination reports help pediatric doctors explain<br />
patients’ condition to their parents. Case management records and collects data that a case manager<br />
needs about patient-related information such as cancer case management, chronic kidney disease, and<br />
diabetes case management. Special examination records can be searched from the main page tag.<br />
4. System evaluation<br />
Since physicians usually have higher control over medical systems, and are always critical in deciding<br />
whether the new system can be promoted smoothly or not, the case hospital developed the OEMRS from<br />
a physicians’ perspective. During system planning and development phase, modifications are made after<br />
discussions on the OEMRS between hospital personnel from different departments and the system<br />
vendor. These modifications and discussions were analyzed, developed and practiced by the information<br />
staff. After the system went online, factors influencing system usage and feedback from users were taken<br />
for future improvements to realize the system’s benefits. Hence, in-depth interviews were used in this<br />
research to get individual opinions that the questionnaire cannot get. This research used semi-structured<br />
interviews and selected pediatric physicians as interviewees since that was the most supportive division in<br />
the outpatient department of the subject hospital. Usages of OEMRS by the interviewers are shown in<br />
Table 4.<br />
Most interviewees started to use the OEMRS in 2008, and were using the system actively at the time of<br />
the interview. The reason to use OEMRS was the hospital’s policy stance and the consulting physician<br />
having joined the OEMRS implementation team. Except the director of the division, none of the physicians<br />
tried to influence others to use the system. Table five shows individual evaluations of the OEMRS.<br />
Table five shows that the quality of the process of information storage and retrieval by the OEMRS was<br />
still below the physicians’ expectations. Most interviewees reported “partial” in stead of positive to the<br />
OEMRS information quality. System quality of OEMRS was assured by most of the interviewees which<br />
contributed to continuous use of the OEMRS.<br />
425
Table 4: Usage of OEMRS by interviewees<br />
Table 5: Evaluation of OEMRS<br />
Hsiao-Ting Tseng et al.<br />
426
5. Conclusion and discussion<br />
Hsiao-Ting Tseng et al.<br />
The objective of this research is to use a case study to explore the implementation of an OEMRS with<br />
physician’s participation. The interviews show that one of the motivations for physicians to use OEMRS is<br />
to accept and promote hospital policy. In the development phase of the OEMRS, cooperation and support<br />
of consultant physicians of the case hospital and communication between physicians and information staff<br />
were critical factors for the success of system development and operation. The system development<br />
experiences can serve as references for other hospitals in developing their own OEMRS.<br />
It is clear that physicians’ satisfaction will be improved if the system interface is easy to use and<br />
corresponds to physician needs. Usage frequency of OEMRS is directly influenced by user experience.<br />
More improvements and modifications need to be done to increase physicians’ acceptance and OEMRS<br />
usage. In summary, physicians’ attitudes toward OEMRS development are generally positive. How to<br />
upgrade OEMRS to assist more physicians’ clinical decision making is the task to be considered. For<br />
instance, integrating automated health examination systems (Kuo and Fuh, 2010) to enhance inference<br />
quality and wireless/mobile devices to benefit point of care (Chung, Bell and Lee, 2006) are feasible<br />
directions.<br />
References<br />
Boar B. (1984) Application Prototyping: A Requirements Definition Strategy for the 80s, John Wiley & Sons, USA, New<br />
York.<br />
Chung, K., Bell, R. and Lee, D. (2006) “A Point of Care Clinical Documentation System for Hospice Care Providers”,<br />
Journal of Medical Systems, Vol. 30, No. 1, pp 33-37.<br />
Drew E. A., Carolyn C. and Robert J. B. (2004) “Improving Patient Safety — Five Years after the IOM Report.” The<br />
New England Journal of Medicine, Vol. 35, No. 1, pp 2041-2043.<br />
Lee, J. N. and Kim, Y. G. (1999) “Effect of Partnership Quality on IS Outsourcing: Conceptual Framework and<br />
Empirical Validation.” Journal of Management Information Systems, Vol. 15, No. 4 pp 29-62.<br />
Kendall, E. K. and Kendall, E.J. (1988) Systems Analysis and Design, Englewood Cliffs, Prentice-Hall, Boston.<br />
Kuo, K.L. and Fuh, C.S. (2010) “A Health Examination System Integrated with Clinical Decision Support System.”<br />
Journal of Medical Systems, Vol. 34, No. 5, pp 829-842.<br />
Peter G. R, Howard A., Robert W. and Peter B. (2006) “Active Physician Participation Key to Smooth MOE MAR<br />
Rollout.” Healthcare Quality, Vol. 10, sp, pp 58-64.<br />
Richard S. D., Elaine B. S. and Don E. D. (1997) The Computer-based Patient Record: An Essential Technology for<br />
Health Care. National Academy Press, Washington.<br />
Robert H. M. and Ida S. (2004) “Physicians' Use of Electronic Medical Records Barriers and Solutions.” Health Affairs,<br />
Vol. 23, No. 2, pp 116-126.<br />
Robert M. W. (2006) “Expected and Unanticipated Consequences of the Quality and Information Technology<br />
Revolutions.” Journal of the American Medical Association, Vol. 296, No. 20, pp 2780-2783.<br />
Rogers, E. M. (1995) Diffusion of Innovations (4th ed.), Free Press, New York.<br />
Zmud, R. W. and Cox, J. F. (1979) “The implementation process: A change approach”. MIS Quarterly, Vol. 3, No. 2,<br />
pp 35-43.<br />
427
MTN Foundation's Digital Library Project in Nigerian<br />
Universities: An Evaluative Study<br />
Ngozi Blessing Ukachi<br />
University of Lagos, Akoka, Nigeria<br />
ukachingozi2001@yahoo.com<br />
Abstract: This survey study was conducted with the main objective of evaluating the use and service provision of<br />
MTN Foundation’s digital library project in Nigerian universities with regards to the objectives of establishing them.<br />
The initial plan to study two out of the three already commissioned libraries was forestalled by temporary closure of<br />
the second one hence the study concentrated on one, situated at University of Lagos, Nigeria. Questionnaire<br />
constituted the major instrument for data collection even though the librarian in charge of this library was interviewed<br />
and documentary records consulted to obtain such data as list of electronic resources/database available in the<br />
library. A total number of 1000 questionnaires were distributed to the library users. Out of this number, 852 were<br />
returned while 812 questionnaires were found usable. The data collected was subjected to descriptive statistical<br />
analysis using simple percentages and frequency counts. The result of the study revealed that this project is<br />
appreciated by users considering their level of acceptance of the fact that; the resources in this library has enriched<br />
their awareness on the existence of various kinds of e- resources (76.6%); the use of the resources has assisted<br />
them greatly in doing their assignments (68.8%); and resources from this library have made their research work very<br />
easy (54.4%). The study also identified some inhibitors to the optimal utilization of the library resources. These<br />
include; lack of awareness and training on the use of the library, low bandwidth, limited access to computer<br />
terminals, etc. The study recommends, among other things, that the bodies responsible for funding the library should<br />
ensure that the Internet bandwidth is upgraded to enhance its speed, users should be trained or given proper<br />
orientation on how to use the resources provided by this library and, the acquisition of more computer systems to<br />
augment the ones provided by MTN foundation<br />
Keywords: digital libraries, electronic resources, library resources use, universities, MTN foundation, Nigeria<br />
1. Introduction<br />
The poor availability and obsolescence of books, journals and other learning materials have always<br />
posed a problem that impairs the ability of higher education institutions in Africa to adequately achieve<br />
their purpose of teaching, research and community service for their expanding population of students<br />
(UNESCO, 2003). Even though Nigeria has more than three hundred higher institutions presently, the<br />
precarious state of the economy has negatively affected the institutions hence scholars and students can<br />
only access a small portion of the necessary library facilities in these institutions. The pitiable state of<br />
library facilities necessitated the Federal Government of Nigeria to make some interventions through<br />
such bodies as, the World Bank Federal Universities Development Sector Adjustment Credit, the<br />
Petroleum Special Trust Fund, National Education Materials Procurement Programme and the Education<br />
Tax Fund. These bodies assisted institutions with the provision of such facilities as computers, books and<br />
journals, and other office equipment for the libraries. Subsequently, at the instance of the former<br />
President, His Excellency, Chief Olusegun Obasanjo, UNESCO facilitated the development of a plan that<br />
sets out a road map to efficiently deliver local and international content to all Nigerian Higher Education<br />
Institutions, staff and students, in the context of their peculiar environments, by means of a Virtual Library<br />
Database provision to augment the existing library resources (UNESCO, 2003). The Virtual Library<br />
Database is meant to be a full text databases covering all fields of study including indigenous content.<br />
Though these steps had some level of positive effect on libraries in Nigeria, the digital divide gap with its<br />
associated problems were still very much evident in Nigerian libraries. It was against this background that<br />
Ogunsola and Aboyade, (2005) observed the urgent need for Nigerian higher institutions to move from<br />
the traditional role of teaching, learning and research to those driven by the information technology. They<br />
concluded that any college or university in Nigeria that does not have full Internet connectivity with<br />
reasonable speed and relevant ICT facilities in the next few years will not be in a position to fulfil the<br />
purposes for its establishment. Ordinarily, given the limited funding available to academic libraries in<br />
Nigeria with the escalating cost of information materials alongside Information Technology (IT) facilities, it<br />
is obvious that the acquisition of these facilities will end up being a day dream if abandoned to be<br />
handled solely by the library. The need to ensure that users of academic libraries in Nigeria have access<br />
to relevant, timely, up-to-date and appropriate information in their fields of interest necessitated the<br />
continued struggle to have in place a digital library that will give users access to electronic information<br />
resources from varying sources.<br />
428
Ngozi Blessing Ukachi<br />
It is in light of the above that MTN Nigeria had in 2005 commenced its Universities Connect project which<br />
provides computers, Internet access and other infrastructures to Nigerian Universities. This project was<br />
embarked upon to achieve MTN’s laudable objectives of bridging the knowledge and digital divide,<br />
enhancing educational infrastructure development and, providing educational resources for effective<br />
learning. However, even though this laudable project has succeeded in making available and also<br />
providing access to diversified digital resources from enormous numbers of libraries globally, there is still<br />
need to evaluate the use and service provision of these libraries vis-a-vis the objectives of establishing<br />
them. This it is believed will expose lapses and give room for amendments where necessary. According<br />
to Nadiri and Mayboudi (2010) library services evaluation help librarians understand users perceptions<br />
and it contributes in improving library service quality and meeting users’ needs better.<br />
2. Objectives of the study<br />
The general objective of this study is to evaluate the use and service provision of these libraries with<br />
regards to the objectives of establishing them.<br />
The specific objectives are:<br />
1. To find out the extent of electronic resources made accessible to users in these libraries<br />
2. To establish if users are given training on how to utilize the resources<br />
3. To determine the extent to which the resources provided by these libraries are being utilized by the<br />
users<br />
4. To find out how the use of the resources of this library has positively impacted on the students<br />
5. To identify factors inhibiting the optimal utilization of the resources provided by these libraries<br />
6. To make recommendations based on the outcome of the study<br />
3. Background information of MTN Nigeria<br />
MTN Nigeria is one of the global mobile telecommunication service providing companies in Nigeria. The<br />
MTN Foundation is the vehicle through which MTN Nigeria implements its corporate social responsibility<br />
programme. The Foundation was established in July 2004 with a focus on giving back to society and<br />
impacting the quality of lives in Nigeria in a meaningful way. The MTN Foundation, which is the<br />
arrowhead of MTN's social investment drive, is playing a key role in deploying various programmes<br />
designed to help reduce poverty and foster sustainable development in Nigeria.<br />
This digital library project in Nigerian Universities by MTN is popularly referred to as MTN Foundation<br />
Universities Connect Project. The institution of this project was brought about by the intense need to<br />
bridge the digital and knowledge divide problem being experienced in the nation by majority of the<br />
citizens who even though they attend Universities, cannot boast of having utilized electronic information<br />
resources in their research activities. This innovative project is designed to provide digital access to<br />
information and infrastructure to Federal Universities in Nigeria. The project facilitates access to a<br />
collection of digital resources from over 5,500 libraries to assist university students and lecturers with<br />
their research works. The project is implemented in partnership with Netlibrary Nigeria Limited. The<br />
present beneficiary institutions of the Universities Connect project are:<br />
For Phase 1- University of Lagos, Akoka, Lagos State<br />
For Phase 2- Ahmadu Bello University, Zaria, Kaduna State<br />
For Phase 3- University of Nigeria Nsukka, Enugu State<br />
For phase 4- University of Benin, Benin City, Edo State. (This particular one has been approved but<br />
not yet commissioned)<br />
MTN Nigeria’s specific objectives of establishing this Education Portfolio are as follows:<br />
To empower through the provision of information and technology resources<br />
To enable skills acquisition, transfer and development<br />
To raise national literacy levels<br />
To provide educational resources for effective learning<br />
To enhance educational infrastructure development<br />
429
The scope of the project is as follows:<br />
Ngozi Blessing Ukachi<br />
128 networked computers, 3 servers, 2 high capacity printers and 100KVA generator<br />
VSAT equipment & internet connectivity bandwidth with 2 years subscription<br />
2 years subscription to electronic resources through the NetLibrary network (Journals, Books, etc)<br />
Conducive study environment through space renovation, provision of adequate lighting, furniture, and<br />
alternative power supply 75b<br />
Technical training for 12 members of existing library staff to work with NetLibrary over 2 years to build<br />
capacity<br />
One-week library awareness to be held annually for students and lecturers<br />
One-week awareness workshops to be held annually for approximately 600 students and 120<br />
lecturers<br />
Initiate a 5–year maintenance contract with the University to ensure a conducive environment with a<br />
clear sustainability strategy<br />
Two years comprehensive insurance cover to take care of theft & fire<br />
An interactive students website - “www.universitiesconnect.com”<br />
4. Methodology<br />
A descriptive research approach was adopted for this study. The purpose of descriptive research is to<br />
describe the current state of affairs at the time of the study without manipulating the variables or having<br />
control over them (Salkind, 2006). The study employed survey procedure for the collection of data.<br />
According to Nworgu (1991) survey method of investigation is one in which a group of people or items<br />
are studied by collecting and analyzing data from only a few people or items considered to be<br />
representative of the entire group. The choice of this research design is considered adequate for the<br />
study because it will enhance and give a better understanding of the situation on ground in relation to the<br />
extent of services provided by the library and also, extent of usage by the users.<br />
This study which was initially programmed to evaluate two out of the three presently existing functional<br />
digital libraries established by MTN Nigeria, ended up evaluating only one, situated at University of<br />
Lagos. This is because the second library at the time of this study was temporarily locked up as a result<br />
of some logistic issues that are yet to be resolved between the institution and the Netlibrary. The<br />
population of this study comprises both the librarian and users of this library.<br />
A self-developed questionnaire tagged Electronic Resources Availability, Extent of Use, Impacts and<br />
Inhibitors (ERAEUII) scale for users was used for data collection. The questionnaire is divided into four<br />
parts. Part A deals with the extent of electronic resources available in the library. Part B concentrated on<br />
the extent of use of this library and its resources. Part C sought responses on the positive impacts of<br />
utilizing this library while Part D concentrated on identifying inhibitor to the optimal utilization of the<br />
resources provided in this library. To complement the questionnaire, the researchers also interviewed the<br />
librarian in-charge of this library. The data collected was subjected to descriptive statistical analysis using<br />
percentages and frequency counts.<br />
5. Findings of study<br />
Questionnaire distribution pattern<br />
A total of One thousand (1000) questionnaires were distributed to the library users. Out of this number,<br />
852 were returned while 812 questionnaires were found usable.<br />
Table 1: List of electronic resources available in this library<br />
1 E- RESOURCES/ DATABASES<br />
2 Sabinet Online Reference Database<br />
3 EBSCOhost Reference Database<br />
4 Jstor<br />
5 HINARI<br />
430
Ngozi Blessing Ukachi<br />
6 MIT<br />
7 SSRN<br />
8 Nation <strong>Academic</strong> Press Database<br />
9 E-Print in physics, Mathematics, computer science and Quantitative Biology<br />
10 Digital Library for earth system education Online<br />
11 Virtual Libraries online<br />
12 Science Direct Online<br />
13 Law Journals Online<br />
14 Oxford Journals Online<br />
15 Ebooks Online Library<br />
16 Linguistics Database Online<br />
17 Guteng Online Library<br />
18 The Encyclopedia<br />
19 The Wiki Service<br />
20 Free Library Online<br />
21 Google Search Engine<br />
22 E-Books Collection<br />
23 File Management<br />
6. Training on the use of the library<br />
On the issue of whether the users are given training on how to use the resources, they all unanimously<br />
answered no. This implies that they are not given any form of training on the use of this library. As a<br />
result of this, majority of the users of this library are only aware of the existence of a few of the database,<br />
thereby underutilizing the resources.<br />
During the interview with the librarian in charge of this library, the researcher was made to know that the<br />
library has not actually been training the users on the use of this library neither have they been holding<br />
one-week library awareness as indicated in the scope of the project by MTN foundation. The reason for<br />
this according to him was that the library lack adequate manpower to be able to perform that function as it<br />
has just four (4) regular staff members.<br />
7. Extent of use of the library<br />
To determine the extent to which the resources provided by these libraries are being patronized by the<br />
users, they were asked to indicate how frequently they utilize the resources in this library. Options such<br />
as; “daily, 2-3 times weekly, weekly, and once in a while”, were given. Their response is represented in<br />
the pie chart below:<br />
Table 2 indicates that 66.7% of the respondents responded that they utilize the library once in a while, the<br />
other 33.3% responded that they utilize it 2-3 times weekly, while none of the respondents attested to be<br />
utilizing it daily or weekly.<br />
Table 2: Response on extent of use of the library<br />
Extent of use Frequency count Percentage<br />
Daily 0 0.0<br />
2-3 times weekly 268 33.3<br />
Weekly 0 0.<br />
Once in a while 544 66.7<br />
431
Ngozi Blessing Ukachi<br />
The librarian was equally asked during the interview to comment on his observation regarding the<br />
utilization of the resources of this library. It was gathered from his response that the library is not being<br />
adequately utilized.<br />
Table 3 illustrates the ways and extent to which the use of the resources provided by this library has<br />
benefitted the students. The findings revealed that 622 (76.6%) of the respondents indicated that the use<br />
of the resources in this library has enriched their awareness on the existence of various kinds of e-<br />
resources. This was closely followed by 559 (68.8%) acceptance of respondent that the use of the<br />
resources has assisted them greatly in doing their assignments. While the option that ‘resources from this<br />
library have made my research work very easy’ received 458 (54.4%) acceptance, ‘my general academic<br />
performance has improved as a result of using MTN Digital library resources in augmenting my<br />
classroom works’ received 320 (39.4%) respondent’s acceptance.<br />
Table 3: Response on positive impacts of utilizing MTN digital libraries<br />
POSITIVE IMPACTS Frequency count Percentage<br />
Enriched my awareness on the existence of various kinds of e- resources 622 76.6<br />
Assisted me greatly in doing my assignments 559 68.8<br />
Resources from this library have made my research work very easy 458 54.4<br />
My general academic performance has improved as a result of using<br />
MTN Digital library resources in augmenting my classroom works<br />
320 39.4<br />
Table 4 shows factors inhibiting the optimal utilization of the resources provided by this library. This<br />
revealed low bandwidth attestation by 786 respondents, which is 96.7% of the study’s user population. A<br />
number of 527 (64.9%) of the respondents indicated that limited access to computer terminals is one of<br />
their inhibitors to optimal utilization of MTN Digital library resources even as Poor attitude of library staff<br />
was indicated by 386 (47.5%) respondents. Irregular power supply received 278 (34.2%) indication by the<br />
respondents while insufficient IT knowledge received 267 (32.8).<br />
Table 4: Response on inhibitors to optimal utilization of MTN digital library<br />
Frequency count Percentage<br />
Inhibitors<br />
Low bandwidth 786 96.7<br />
Irregular power supply 278 34.2<br />
Poor attitude of library staff 386 47.5<br />
Limited access to computer terminals 527 64.9<br />
Insufficient IT knowledge 267 32.8<br />
8. Discussion of findings<br />
The findings of the study show that the necessary information communication and technology facilities in<br />
the form of computers, printers, scanners, Internet access, database, etc, required for service provision in<br />
this library are all available. It also revealed that these facilities are in good working condition. From<br />
personal observation, the library has cconducive study environment. There is adequate space and<br />
furniture provision and also alternative power supply.<br />
On the subject of positive impacts of using the library, high percentage (76.6%) of the respondents<br />
attested that the use of this library have enriched their awareness on the existence of various kinds of e-<br />
resources. A number of 559 respondents being 68.8% of the user population stated that it has assisted<br />
them greatly in doing their assignments while 54.4% confirmed that resources from this library have<br />
made their research work very easy. This result shows that adequate utilization of these resources by<br />
students will enhance their academic performance and assist in making them assets to the society. This<br />
finding corroborates with Tella, Tella, Ayeni and Omoba’s (2007) findings in their work on Self Efficacy<br />
and Use of Electronic Information as Predictors of <strong>Academic</strong> Performance. This work identified that<br />
electronic resources has many functions and benefits that are capable of positively influencing the<br />
academic performance of both students and researchers in the university as well as other higher<br />
educational institutions.<br />
432
Ngozi Blessing Ukachi<br />
On the other hand, the analysis of the results reveals some shortfalls which if not taken care of could<br />
affect the actualization of the goal of establishing this library. The entire user population of the study<br />
unanimously stated that no form of training is given them on the use of this library and its resources. The<br />
repercussion could be seen from the result of the extent of use of the library where not even a single user<br />
accepted to have been using the library on a daily basis rather a very high percentage (66.7%) of the<br />
user population indicated that they use it “once in a while”. It is a common knowledge that an individual<br />
will utilize facilities more when he possesses adequate skills necessary for their use while on the other<br />
hand he cannot effectively utilize what he does not have any training on.<br />
Low bandwidth and limited access to computer terminals were identified as the major inhibitors to optimal<br />
utilization of this Library. This finding agrees with those of Aduwa-Ogiegbaen and Iyanu (2005), Edom<br />
(2007) and Ukachi (2008) who noted inadequate number of ICT facilities, limited access, among other<br />
factors as inhibitors to optimal utilization of this kind of libraries. The forgoing indicates the slow pace at<br />
which the movement towards bridging the digital divide is being made. Poor attitude of library staff was<br />
also identified as another factor inhibiting the optimal utilization of these resources. This may perhaps be<br />
caused by the problem of inadequate staffing being experienced by the library which is likely to be<br />
affecting their service provision.<br />
A survey research on the use of ICT equipment carried out by Daniel et.al (2003) shows 70%<br />
respondents’ acceptance of incessant electric power failure as a major impediment while this particular<br />
work has 34.2% acceptance of this factor. This provision of standby power generator is a pointer that<br />
institutions are now realizing the place of power supply in digital/electronic service provision.<br />
9. Conclusion<br />
This study without doubt has led to the establishment of facts as regards the main objective of the work,<br />
which is to evaluate the use and service provision of MTN Foundation’s University digital library projects<br />
in Nigeria. The result of the study showed that this project is appreciated by users considering their<br />
testimonies concerning the positive impacts of utilizing the resources being provided. At the same time, it<br />
should be noted that mere provision of this resources without creating corresponding awareness on its<br />
existence tantamount to waste of resources as availability is only one of the factors that influence usage<br />
but does not necessarily guarantee optimal usage on its own.<br />
10. Recommendation<br />
Based on the findings of the study, if the digital/electronic libraries must live to continually fulfill their<br />
primary goal which is, the provision of information and services geared at satisfying the information needs<br />
of users, the following recommendations meant to check the usage inhibiting factors should be<br />
considered and implemented.<br />
The bodies responsible for funding the library should ensure that the Internet bandwidth is upgraded<br />
to enhance its speed. This should also be emulated by similar libraries.<br />
To curb the problem of limited access to computer systems, the institution should acquire more<br />
computer systems to augment the ones provided by MTN foundation. Meanwhile, considering the<br />
funding situation of the libraries in Nigeria which Okiy (2005) termed to be inadequate, the library on<br />
its own should also develop some fee-based electronic reference services such as, Selective<br />
Dissemination of Information(SDI), Abstracting and Indexing services to individuals, etc. This will help<br />
them generate funds internally to augment what is being allocated to them for the acquisition of<br />
additional ICT facilities to enable larger number of users to access them.<br />
Users should be trained or given proper orientation on how to use the resources provided by this<br />
library more effectively. The relevance of library orientation should never be underestimated be it in<br />
the traditional or electronic library environment.<br />
To actualize the above, more staff members should be employed in the library to assist in the service<br />
provision of the library.<br />
References<br />
Aduwa- Ogiegbaen, S. E. and Iyanu, E. O. S (2005) Using Information Technology in Secondary Schools in Nigeria:<br />
problem and prospects. Educational Technology and Society. 8(1). P.104-112.<br />
Edom B. O. (2007) An assessment of the problems of the use of ICT facilities in information sourcing and retrieval by<br />
the academic staff in Nigerian universities. The InformationTechnologist: 4(2).p.113-127<br />
433
Ngozi Blessing Ukachi<br />
Daniel, J. O. et al (2003) Forty years of information and communication technology (ICT) Library services to the<br />
nation: in S. O. Lanlokun (ed.) forty years of library services in Nigeria. The Guardian Newspaper, March 24.p.2.<br />
http://www.mtnonline.com/mtn-foundation/...http://www.mtnonline.com/MTN-Foundation/index.asp<br />
Nadiri, H. and Mayboudi, S. M. A. (2010) Diagnosing university students’ zone of tolerance from university library<br />
services. Malaysian Journal of Library & Information Science, 15(1). p.1-21.<br />
http://majlis.fsktm.um.edu.my/document.aspx?FileName=842.pdf. Accessed on 22/01/2011<br />
Nworgu, B. G. (1991) Educational research: basic and methodology. Ibadan, Wisdom Publishing<br />
Ogunsola, L. And Aboyade, W. (2005) “Information and Communication Technology in Nigeria: Revolution or<br />
Evolution” Journal of Social Sciences, 11(1): 7-14<br />
Okiy, Rose B. (2005) Funding Nigerian libraries in the 21 st century: will funding from alternate sources suffice? The<br />
Bottom Line: managing library finances, 18(2). p.71-77<br />
Salkind, N. J. (2006) Exploring research. 6 th ed. Pearson: New Jersey. Samba, A. 2007<br />
Tella, A., Tella, A., Ayeni, C. O. and Omoba, R. O. (2007), Self Efficacy and Use of Electronic Information as<br />
Predictors of <strong>Academic</strong> Performance. Electronic Journal of <strong>Academic</strong> and Special Librarianship. 8(2).<br />
http://southernlibrarianship.icaap.org/content/v08n02/tella_a01.html Accessed on 22/06/2010<br />
Ukachi, N. B. (2008) Utilisation of Information Communication Technologies (ICTs) in Reference Services of<br />
<strong>Academic</strong> Libraries: Threats and Challenges. The InformationTechnologist. 5(2).p62-69<br />
UNESCO, (2003),“Feasibility Study on the Development of a Virtual Library by Institutions of Higher<br />
Education in Nigeria”, Consultancy Support Services Ltd., Abuja, Nigeria, available at:<br />
http://www.portal.unesco.org/en/files/39276/...Final.pdf/FS4VL_Final.pdf. Accessed on 18/11/2009<br />
434
Cloud Computing-Based IT Solutions for Organizations With<br />
Multiregional Branch Offices<br />
Harris Wang<br />
Athabasca University, Canada<br />
harrisw@athabascau.ca<br />
Abstract: One of the most significant phenomena of the new century is globalization. As business goes global,<br />
multiregional branch offices are needed and networked computing and information services must then be established<br />
for those branch offices. In this paper we investigate a cloud computing based approach to the rapid deployment of<br />
computing and information services for organizations with multiregional branch offices. We first take a look at the<br />
general process leading to the deployment of Computer and Information Technology (CIT) services for organizations,<br />
and then present some cloud computing-based solutions for organizations with multiregional branch offices, followed<br />
by discussions about their key features as well as issues and concerns surrounding the proposed IT solutions. An<br />
important contribution of this paper is a generalized view of the cloud computing-based approach, which may be<br />
used as guidance in implementing and deploying such IT solutions.<br />
Keywords: cloud computing, globalization, information technology infrastructure<br />
1. Introduction<br />
In today’s world, the reach of computer and information technology is vast and deep. Computer and<br />
information technology are prevalent in every corner of our society from health care and military<br />
deployments, to that of the corporate business world; using computers and information technology has<br />
become a fundamental part of daily operations for almost all organizations. Information technology drives<br />
business and communications through virtually removing borders and distance limits and making it<br />
possible for instant commerce and communication globally between organizations or internally.<br />
The expectation that surrounds the use of computer and information systems on this global network has<br />
become one that society expects instant communications with instant gratification and results from all<br />
businesses at all times. In order to provide this type of service it is necessary to have a robust and<br />
reliable information technology infrastructure underpinning the communications and information systems.<br />
Traditionally, the information technology infrastructure that supports these systems needed for<br />
businesses with multiregional branch offices usually starts internally with a core network and data center<br />
for the originating site or head office and extends to branch offices worldwide. This traditional approach<br />
often provides more reliable and better guaranteed computing and information services. With the<br />
advances in computing and information technology, we now may have more options in providing the IT<br />
needs to businesses with branch offices worldwide. In this paper, we will explore a cloud computing<br />
based approach to the rapid deployment of computing and information technology services for<br />
businesses with offices worldwide.<br />
2. General management process towards an IT solution for businesses<br />
In general, any new branch office will require a data center, network core and infrastructure,<br />
telecommunications systems, 3-tier application systems, web servers, database servers and client<br />
computers, and all these elements require a high degree of information security considerations to be<br />
taken. No matter how these elements will be implemented, some general Project Management (PM)<br />
steps must be taken (PMI 2004).<br />
In order to determine exactly what IT infrastructure and systems are needed and how to implement the IT<br />
infrastructure and determined systems for a new business expansion, the first step is defining the needs<br />
of the business. In this case a Chief Technical Architect (CTA) or the like will need to work with the<br />
functional representatives from the business and defines what will be needed to support the daily<br />
operations of the business; from an ERP (for Enterprise Resources Planning) system and Email services<br />
to Computer Numerical Controlled (CNC) systems and Computer-Aided Design (CAD) systems, this<br />
should be determined in the project conception stages. This Initiation phase of the project allows the<br />
business to communicate the technical needs or concepts to be evolved.<br />
The next phase of the project is Project Planning (PP), which in the project lifecycle is the next logical<br />
progression for the project. It is important in this phase to determine the full extent of the work packages<br />
or work breakdown structure (WBS), the sequencing of activities, resources required, risks and risk<br />
435
management plans, cost estimating and budget, change and quality control, procurement plans, plan and<br />
hold a formal project kick off meeting and define the finalized project plan (PMI 2004).<br />
The last phase of the project is Project Execution where the work begins to occur is signified by acquiring<br />
the project team, receiving responses from any request for proposal, selection of the vendor based on<br />
proposal, execution of the work packages, perform quality assurance, and the PM distributes and<br />
controls project information and implements administrative closure procedures.<br />
This 3-phase process of project management is applicable when using a cloud computing based IT<br />
solution to meet the computing and information technology needs of businesses with multiregional branch<br />
offices.<br />
3. Cloud computing based solutions<br />
Cloud computing is defined as “a pay-per-use model for enabling available, convenient, on-demand<br />
network access to a shared pool of configurable computing resources (e.g., networks, servers, storage,<br />
applications, services) that can be rapidly provisioned and released with minimal management effort or<br />
service provider interaction” (Pearson 2009). In literature and practice, there are three cloud computing<br />
models: Infrastructure as a Service (IaaS), Platform as a Service (PaaS), and Software as a Service<br />
(SaaS) (Fouquet 2009, Gruman 2008).<br />
In the IaaS model, users subscribe for uses of certain components of provider’s IT infrastructure.<br />
Although the subscribers don’t have control of the entire cloud infrastructure, they do have control over<br />
selected portions of it, such as firewalls, operating system, deployed applications and storage.<br />
In the PaaS model, a combination of applications, which forms a platform, is subscribed as a service by<br />
users. For example, a combination of software tools may be used as a programming and software<br />
platform.<br />
The SaaS model can be seen as special case of PaaS, where a single application can be subscribed as<br />
a service. Such services are often accessed through a Web browser.<br />
As for cloud, there are private cloud, public cloud and hybrid cloud based on the relationship between<br />
cloud owners and cloud users (Buyya 2008, Dikaiakos 2009). Private Cloud is usually owned and used<br />
by the same organization such as a corporation. It often refers to a proprietary computing infrastructure<br />
owned by the organization, and provides computing and information services to its employees behind the<br />
organization’s firewall. Public cloud often refers to computing and IT infrastructure that is owned by an<br />
organization but provide computing and information services to external users or subscribers. By<br />
subscribing services provided by other well established companies, new start-ups, for example, can<br />
quickly realize their computing and information technology needs without investing so much money and<br />
time to implement their own computing and IT infrastructure.<br />
Hybrid cloud is a mixture of public cloud and private cloud. It can be useful for some well established<br />
corporations who already have their own computing and IT infrastructure, but need additional computing<br />
and IT services for expansion in new areas. By subscribing needed computing and information services<br />
available in public cloud, they can make use of the services more quickly without investing big money and<br />
lengthy time on implementing their own CIT infrastructure. Accordingly, we derive from the above the<br />
following cloud computing based solutions for businesses with multiregional branch offices, to meet their<br />
computing and information technology needs.<br />
3.1 Private cloud based solution to meet the IT needs of multiregional branch offices<br />
For organizations who have well established computing and information technology infrastructure, such<br />
as some high-tech companies in the IT business, a completely private cloud based solution may be a<br />
better choice to provide computing and IT services to new branch offices in other cities or other countries.<br />
In such a case, the data centre, servers and all major computing and IT devices reside behind the firewall<br />
on the organization’s enterprise network, located on site of the head office, while users in branch offices<br />
access the computing and information services through VPN or a Web browser, if a Web interface has<br />
been made available for accessing the service. Figure 1 depicts the architecture of our private cloud<br />
based solution.<br />
436
Branch Office A<br />
Branch Office C<br />
Figure 1: Architecture of private cloud based solution<br />
In private cloud based solution since everything is under control of the very same organization, this<br />
solution gives the organization the total freedom and autonomy in managing all the components of the<br />
computing and information technology infrastructure.<br />
3.2 Federated cloud based solution to meet the IT needs of multiregional branch offices<br />
Federated cloud can be seen as a variant of private cloud. Same as in private cloud, in federated cloud<br />
computing and information technology infrastructure is still privately owned by the sole organization, but<br />
the equipment, servers and services are distributed among head office and branch offices. This may be<br />
necessary when different branch offices have different missions and each needs more dedicated<br />
computing and information services. For example, one branch may be working on data mining, while<br />
another branch may be more on server development. The architecture of federated cloud based solution<br />
is showed in figure 2.<br />
Branch Office A Branch Office B<br />
Branch Office C<br />
Figure 2: Architecture of federated cloud based solution<br />
3.3 Public cloud based solution to meet the IT needs of multiregional branch offices<br />
A public cloud based solution is for the organization to subscribe all needed computing and information<br />
technology services from providers in the public cloud. This solution is suitable for organizations that<br />
have no resources or interest to implement their own CIT infrastructure, or small start-ups. By subscribing<br />
the needed computing and IT services readily available in the public cloud, a start-up can quick get its<br />
business going and have the creative ideas tested. If the business doesn’t fly, it can easily get off the<br />
boat with less to lose. Figure 3 depicts the architecture of public cloud based solution.<br />
Figure 3: Architecture of public cloud based solution<br />
Private cloud owned by and hosted<br />
at the head office of the cooperation<br />
Branch Office E Branch Office F<br />
Branch Office G<br />
Branch Office A<br />
Branch Office C<br />
Branch Office E<br />
Public Cloud<br />
Head Office of the<br />
Cooperation<br />
437<br />
Branch Office D<br />
Branch Office B<br />
Branch Office D<br />
Branch Office A<br />
Branch Office D<br />
Branch Office F
3.4 Hybrid cloud based solution to meet the IT needs of multiregional branch offices<br />
As in the common definition, hybrid cloud involves both private and public cloud. A hybrid cloud based<br />
solution may be applicable for an organization that has an established CIT infrastructure sufficient for its<br />
current needs, but doesn’t want to invest big money and time to expand its current CIT infrastructure for<br />
new business. In this case, it would rather choose to get the needed CIT services from a reliable source<br />
in the public cloud. By doing this, it may be more easily turn around if the new adventure doesn’t go well.<br />
The architecture of hybrid cloud based solution is shown in figure 4.<br />
Branch Office A<br />
Private<br />
Branch Office A<br />
cloud owned<br />
by and<br />
hosted at<br />
Branch Office A<br />
the head<br />
office of the<br />
cooperation Branch Office A<br />
Branch Office A<br />
Figure 4: Architecture of hybrid cloud based solution<br />
Public Cloud<br />
4. Advantages and challenges of cloud computing based solution<br />
Cloud computing based IT solution for organizations with multiregional offices brought both advantages<br />
and challenges. The advantages are as follows:<br />
Cloud computing provides organizations with more agile solutions to meet their IT needs. With cloud<br />
computing, they can quickly get the IT services needed to run their business. This is especially<br />
suitable for new business with a brilliant but risky adventure. Because the new business idea is<br />
brilliant, they need to get the idea tested as soon as possible before it is stolen by competitors;<br />
because it is a risky adventure, it is very likely to fail. By subscribing CIT services available from the<br />
public cloud, they could avoid big loss in CIT infrastructure investment (Cohen 2004).<br />
By subscribing services from a third-party, an organization can save money on capital investment. It<br />
can also save on maintenance because the CIT infrastructure is owned and maintained by the thirdparty.<br />
Compared to CIT infrastructure developed from scratch, cloud computing based solutions can be<br />
more reliable by subscribing well established computing and information technology services<br />
provided by trusted third-parties (Stantchev 2009).<br />
Because additional CIT services can be easily subscribed from providers in the cloud, a cloud<br />
computing based IT solution can be more scalable. It can also be more easily scale down if<br />
necessary, because unsubscribing unneeded services are usually less painful than disposing some<br />
expensive CIT equipments (Buyya 2009, Brandic 2009).<br />
Generally speaking for the entire IT industry, cloud computing provide better IT solutions to utilize IT<br />
infrastructure and equipments, as these often expensive infrastructure and equipments can now be<br />
shared among more users. This also means the overall power consumption will be lower, and hence<br />
the carbon footprint will be smaller. So, it is good for the environment too (Buyya 2008).<br />
In cloud computing based solutions, because important data centers and servers are often owned<br />
and managed by big well established companies who usually have much better expertises and<br />
resources to secure their equipments and services, cloud computing based solutions are often<br />
secure in this regard (Stantchev 2009).<br />
There are few challenges or concerns surrounding could computing based IT solutions. There challenges<br />
or concerns are:<br />
1. The operational costs may be high. We mentioned that the capital cost can be much lower when a<br />
cloud computing based solution is chosen. However, its operational cost may be higher than running<br />
its own CIT infrastructure because it may be required to pay big money for the subscribed services.<br />
438
This all depends on the relationships and contract between the organization and the services<br />
provider or providers (Li 2009).<br />
2. Reliability may be questionable if the CIT services provided in the cloud are not carefully chosen.<br />
In theory, a cloud computing based IT solution should provide reliable CIT services if the providers<br />
are carefully chosen. There are two factors that may cause the CIT services obtained from cloud<br />
computing unreliable. These factors include:<br />
a. Unpredictable failure of network external to the organization. Since those portions of network are<br />
not under control of the organization, it cannot really do anything to avoid such failure, or to ensure a<br />
quick recovery once such a failure has occurred.<br />
b. Services broken down at the service provider. It is very important to choose a reliable and trusted<br />
service provider. Even if this has been done, there is still no guarantee as things may change over<br />
time. The provider may have gone bankrupt; it might be sold to another party which is unfamiliar or<br />
even hostile to the subscriber. The original owner of the service provider may turn to be unfriendly for<br />
some reasons.<br />
3 .In a cloud computing based solution, the security of data and subscribed CIT services are in thirdparties’<br />
hand. They care secure only if the third parties are doing their good job. However, even if<br />
great effort has been taken in choosing the service providers, security may still be a problem fir the<br />
following reasons (Pearson 2009, Kaufman 2009, Kandukuri 2009):<br />
a. They may be an oversight when evaluating the security measures implemented at the service<br />
providers;<br />
b. There may be new threats that the service providers are unprepared for. If the providers do not<br />
implement anti-threat measures in time and effectively, the security of the data and CIT services may<br />
be at great risk;<br />
c. The service providers may change their security practice or even side (due to pressure from the<br />
top, for example).<br />
d. Because in a cloud computing based solution, users access data and subscribed services through<br />
the Internet, security breach may occur during transmission between users and the service providers.<br />
5. Conclusions<br />
Cloud computing brings us both opportunities and challenges. With the advancement of computer and<br />
network technologies, more reliable and powerful computing and information technology services have<br />
become available on the Internet, and high speed reliable internet access to these services is becoming a<br />
reality for many users in the world. Therefore, it is possible for organizations with multiregional branch<br />
offices to get the needed computing and information technology services through cloud computing.<br />
In this paper, we presented four cloud computing based solutions: private cloud based, federated cloud<br />
based, public cloud based, and hybrid cloud based. The subtle differences between these solutions, as<br />
shown in the figures, are important because they have crucial impact to the details of implementation and<br />
even administration, including accounting and security management. For example, in the private or<br />
federated cloud computing based solutions, the accounting can be simple because there is no external<br />
body involved, but the system administration may be heavy especially in the case of federated cloud<br />
computing because of the equipments and services are distributed amongst the branch offices. This may<br />
increase the complexity of the entire system. In case of public cloud computing based solution, the<br />
organization may have less to do in security, but the accounting will have to consider the operational cost<br />
of subscribed CIT services.<br />
We also pointed out in this paper some advantages of using cloud computing based solutions, as well as<br />
some issues. It is important for an organization to carefully study both these advantages and issues when<br />
considering a cloud computing based solution for its needs.<br />
References<br />
Buyya, R., Yeo, C. S. and Venugopal S. (2008) Market-Oriented Cloud Computing: Vision, Hype, and Reality for<br />
Delivering IT Services as Computing Utilities. Department of Computer Science and Software Engineering, The<br />
University of Melbourne, Australia. pp. 9.<br />
Buyya, R., Ranjan, R. and Calheiros, R.N. (2009) Modeling and simulation of scalable Cloud computing<br />
environments and the CloudSim toolkit: Challenges and opportunities, International <strong>Conference</strong> on High<br />
Performance Computing & Simulation, HPCS '09. Page(s):1 – 11<br />
439
Harris Wang<br />
Brandic, I. (2009) Towards Self-Manageable Cloud Services, Annual IEEE International Computer Software and<br />
Applications <strong>Conference</strong>, 2009. COMPSAC '09. 33rd Volume 2, Page(s):128 - 133<br />
Cohen, D., Lindvall, M. and Costa, P. (2004) An introduction to agile methods, Advances in Computers, New York,<br />
Elsevier Science, pp. 1-66.<br />
Dikaiakos, M.D., Katsaros, D., Mehra, P. Pallis, G. and Vakali, A. (2009) Cloud Computing: Distributed Internet<br />
Computing for IT and Scientific Research, Internet Computing, Issue 5, Volume 13, IEEE.<br />
Gruman, G. (2008) What cloud computing really means. InfoWorld (http://www.infoworld.com), retrieved 2010-01-17<br />
from http://www.infoworld.com/d/cloud-computing/what-cloud-computing-really-means-031<br />
Fouquet, M. and Niedermayer H., Carle G. (2009) Cloud computing for the masses, in Proceedings of the 1st ACM<br />
workshop on User-provided networking: challenges and opportunities, pages 31-36.<br />
Kandukuri, B.R., Paturi, V.R. and Rakshit, A. (2009) Cloud Security Issues, IEEE International <strong>Conference</strong> on<br />
Services Computing, SCC '09. Page(s):517 – 520<br />
Kaufman, L.M. (2009) Data Security in the World of Cloud Computing, Security & Privacy, IEEE, Volume 7, Issue<br />
4, Page(s):61 – 64<br />
Li, X., Li Y., Liu T., Qiu J. and Wang F. (2009) The Method and Tool of Cost Analysis for Cloud Computing, IEEE<br />
International <strong>Conference</strong> on Cloud Computing, CLOUD '09. Page(s):93 – 100<br />
Pearson, S. (2009) Taking account of privacy when designing cloud computing services, Proceedings of the 2009<br />
ICSE Workshop on Software Engineering Challenges of Cloud Computing, p.44-52.<br />
Project Management Institute (2004) Guide to the Project Management Body of Knowledge, 3rd Edition.<br />
Stantchev, V. (2009) Performance Evaluation of Cloud Computing Offerings, Third International <strong>Conference</strong> on<br />
Advanced Engineering Computing and Applications in Sciences, ADVCOMP '09, Page(s):187 – 192<br />
440
Case Study on Information Evaluation by GIS for Aging<br />
Society Urban Planning: Information Evaluation of Shrinking<br />
Cities<br />
Yupeng Wang, Hiroatsu Fukuda and Kiyoshi Shinriki<br />
The University of Kitakyushu, Japan<br />
y09e0101@hibikino.ne.jp<br />
fukuda@env.kitakyu-u.ac.jp<br />
shinriki@hq.kiu.ac.jp<br />
Abstract: An understanding of the sites upon which they are working is crucial to city planners at every stage of the<br />
urban design process. Usually, the information (such as population, landform, building conditions, and economic<br />
data) is too complex to be processed and reported for analysis through the ordinary statistical approaches. The GIS<br />
(Geographic Information System) offers software that can collect related databases for different fields onto one map,<br />
allowing both the individual analysis diagram and the compound analysis to be easily achieved. Thus, GIS provides<br />
planners a more comprehensive understanding of the site upon which they are working. Our research is an actual<br />
urban design project in Japan, an example of the utilization of GIS in urban planning. It is a typical case, with an<br />
intricate landform and typical population composition. First, GIS analyzes the basic data, covering landform,<br />
construction age, and population. These analyses occur in two stages: first, a comparative study of the population<br />
structure over the last 10 years in different landform areas, then the forecasting of its evolution. Using these<br />
analyses, the proposal plan tries to improve the residential environment and prevent decline by retaining the local<br />
population and attracting outsiders. The introduction of GIS allows digital data to be composited with draft data,<br />
resulting in clearer and more precise analyses. This method produced accurate dates for Yahatahigasi-ku, therefore<br />
assisting the evaluation of data reflecting the current situation and orienting the next step of the urban design. It has<br />
thus been demonstrated that GIS enhances the efficiency of site analysis. We consider this study of GIS in urban<br />
planning to be a development of GIS application. This paper presents a new information evaluation method for urban<br />
planning that provides more social information on the target area than previous methods and shares the results of<br />
information evaluations.<br />
Keywords: urban design, GIS, population, landform, building state<br />
1. Introduction<br />
New urban proposal plans usually become necessary as a city develops, in order to deal with variations<br />
in population and transformations in lifestyles. Studying the site from different perspectives and<br />
developing a comprehensive understanding of the site are the key steps to a successful plan. The GIS<br />
can help planners gather both the abstract and concrete data into a single graphic. This paper proposes<br />
the introduction of the GIS in the site analysis stage. It discusses the GIS analysis method through a case<br />
study of Yahatahigasi-ku, Japan. Based on these analyses, the proposal plan tries to improve the<br />
residential environment and prevent urban decline by retaining the local population and attracting<br />
outsiders.<br />
2. Goal, research objective, and methods of research<br />
2.1 Goal<br />
Japan’s aging society is becoming a serious problem, which must be analyzed from different<br />
perspectives if it is to be understood and solved. This paper aims to find a new approach that involves the<br />
GIS in the analysis stage.<br />
2.2 Research objective<br />
This study is an urban regeneration proposal plan for Yahatahigasi-ku, Japan. This area became the<br />
biggest industrial area of Japan after the Second World War but is now struggling against depopulation<br />
and an aging community. Furthermore, the houses built during Japan’s industrial age are becoming<br />
decrepit, and most of the land lies on a steep slope. The proposal plan explores ways of solving these<br />
problems and providing residents with a better living environment.<br />
Landform<br />
441
Yupeng Wang et al.<br />
To accommodate employees, dwellings were built on the mountain around the industrial area. As<br />
employees aged, the inconvenient landform made it difficult for them to access the facilities near their<br />
houses. Young people are moving away in search of more comfortable residential environments, while<br />
few people move in (though the local population is increasing). Meanwhile, industrial styles have<br />
changed: factories are smaller, and some of the flat areas can be redeveloped.<br />
Construction age<br />
Dilapidated dwellings in this area also pose a problem. Traditional wood dwellings constitute the majority<br />
of the street. The number of empty houses is increasing, though, especially in the steeply sloped area.<br />
An accurate average date of dwelling construction would enable an analysis of the ward’s actual<br />
condition.<br />
Population<br />
Yahatahigasi-ku had its glory days during the development of the West Japan Steel Factory from 1950 to<br />
1960. The city’s population was at its highest at that time. However, as the industry innovated, local<br />
residents began to leave. The current population of 60 thousand will decrease to 20 thousand in about 40<br />
years. Though depopulation and collective aging are problems throughout Japan, the problem in<br />
Yahatahigasi-ku is particularly serious.<br />
2.3 Methods of research<br />
This paper uses GIS to collect data on landform, construction dates, and population in an analysis of the<br />
ward’s actual condition and future. This is a new approach to site analysis.<br />
3. Data analysis through GIS<br />
3.1 Landform<br />
Population affects the size of city. As a city grows, dwellings and other buildings are built. Given a<br />
haphazard layout, a residential environment becomes inconvenient, energy is wasted, and social<br />
problems emerge. A contour line with a 2m pitch and building layout are put into one picture (see<br />
Figure1), showing the connection between the landform and buildings. Factorial facilities are built in the<br />
beach area. A great number of dwellings are on the steeply sloped area of the mountain. This condition<br />
flows from the industrial age, when the city grew through a rapid increase in population. It is necessary to<br />
remediate access methods for the elderly in a sustainable urban design.<br />
Figure 1: Landform and building distribution in Yahatahigasi-ku (2005)<br />
442
Yupeng Wang et al.<br />
To express the steepness of these areas, “slope degree” is introduced as a standard of measure.<br />
Differences in height and average slope in each 100m×100m mesh can be calculated by counting the<br />
number of contour lines in each mesh area. As shown in Figure 2, the slope degree is the same as the<br />
number of contour lines.<br />
Figure 2: Definition of “slope degree” in each mesh<br />
Figure 3: Image of layer overlapping<br />
Using the overlapping mesh layer and contour line layer seen in Figure 3, we found the slope degree of<br />
each mesh. Consequently, the overlap of mesh upon building layer and the slope degree of each building<br />
can be expressed in different colors according to the different slope degrees (as shown in Figure 4).<br />
Slope degree of each mesh → Layer overlapping → Slope degree of each building<br />
Figure 4: Explanation of building slope degree<br />
443
Yupeng Wang et al.<br />
Figure 5 shows that some buildings have a low slope degree in the high-level area, namely those<br />
buildings built on the flat area of the mountain. A comparison with Figure1 shows that buildings closer to<br />
the top of mountain have a higher slope degree. Poor approach planning makes daily life inconvenient for<br />
city dwellers.<br />
Figure 5: Slope degree of each building<br />
3.2 Construction age<br />
Unlike the stone masonry buildings of Europe, the traditional wooden dwellings of Japan last about 30<br />
years, after which they become decrepit. Moreover, during the immediate postwar period of 60 years ago,<br />
a lack of resources in Japan reduced the quality of Japanese building materials and shortened the life<br />
spans of Japanese dwellings.<br />
In Yahatahigasi-ku, the percentage of dwellings more than 30 years old is higher than 60% in almost all<br />
of the towns. In most towns, 80% of the dwellings are more than 30 years old. In about half the towns,<br />
20% of the dwellings are older than 60. The layout showing construction age appears in Figure 6. A great<br />
number of old buildings lie in the steep slope area, some of which have been abandoned by their owners<br />
and remain empty. The number of empty dwellings will increase as the population decreases.<br />
The average construction age of towns is shown in Figure 7. Almost all of the towns are over 30 years<br />
old. Significantly, the average construction age of some towns is over 60, implying that the whole street is<br />
dilapidated and that a macro-scale reconstruction conforming to the needs of modern lifestyles is needed<br />
here.<br />
3.3 Population<br />
Local activity creates a bustling community and conditions the atmosphere of the street. Thus, population<br />
and the construction of population affect local revitalization. Population layout and construction should be<br />
analyzed before urban planning begins.<br />
As shown in Figure 8, the population density of towns on the sloped area is lower than that in the flat area<br />
because most of dwellings on the slope are isolated houses. Apartments have recently been built close to<br />
the railway station. The population of the flat area is increasing slowly, but the total population of this<br />
ward is still decreasing, especially in the towns on the steeply sloped area.<br />
444
Yupeng Wang et al.<br />
Figure 6: Construction age of buildings in Yahatahigasi-ku<br />
Figure 7: Average construction age of each town<br />
An aging society is that in which the percentage of the population over 65 is higher than 25%. A<br />
percentage higher than 50% produces a limitative colony, a social community that is difficult to maintain.<br />
As Figure 9 shows, almost all of the towns satisfy the criterion of an aging society. The percentage of the<br />
population that older than 65 is higher on the steeply sloped area, where it is even higher than 50% (thus<br />
meeting the criterion of a limitative colony). All of the limitative colonies are situated in the highly sloping<br />
area of the mountain.<br />
445
Yupeng Wang et al.<br />
Furthermore, as Figure10 shows, the percentage of the population older than 75 is higher than 20% in<br />
many towns, suggesting that collective aging is a major problem in this ward.<br />
Figure 8: Population density of Yahatahigasi-ku (2008)<br />
Figure 9: Percentage of population older than 65 (2008)<br />
446
Figure 10: Ratio of population older than 75 (2008)<br />
Yupeng Wang et al.<br />
We discover future population and population construction trends and predict the ward’s population size<br />
in 10 years (2018) by applying an equation to the trend line of the last 10 years (1998 ~ 2008).<br />
P18=P18T-(P08N-P08)×P18T/P08T<br />
Where<br />
P18:The population of 2018 (prediction)<br />
P08:The population of 2008<br />
P18T:Transitive the population of 2008 to 10 years later.<br />
P08T:Transitive the population of 1998 to 10 years later.<br />
Figure 11: Population of each age in Yahatahigasi-ku (1998)<br />
447
Yupeng Wang et al.<br />
Figure 12: Population of each age in Yahatahigasi-ku (2008)<br />
Figure 13 shows the population size 10 years from now, based on the trend of the last 10 years. By then,<br />
the population will have decreased from 73162 to 63045. The percentage of the population that older<br />
than 65 is going to increase from 30.2% to 35.4%, far higher than the criterion of 25% that defines an<br />
aging society.<br />
Thus, if current conditions continue, the population is going to decrease, and the percentage of elderly is<br />
going to increase. Some countermeasures should be introduced to this area immediately to maintain its<br />
vitality.<br />
Figure 13: Population of each age in Yahatahigasi-ku (2018-prediction)<br />
(The date in the dark color means the population is higher than a thousand.)<br />
4. Conclusion<br />
This paper discusses an urban regeneration proposal plan for Yahatahigasi-ku as a case study and<br />
proposes the GIS method of site analysis. The GIS information evaluation of Yahatahigasi-ku reveals that<br />
many dwellings in the highly sloped area suffer from a poor living environment because of their advanced<br />
ages. Poor approach planning forces inhabitants to walk everywhere, which is hard for the elderly who<br />
constitute a high proportion of Yahatahigasi-ku’s population.<br />
The introduction of GIS allows digital data to be composited with draft data, resulting in clearer and more<br />
precise analyses. This method produced accurate data for Yahatahigasi-ku, therefore assisting the<br />
evaluation of data reflecting the current situation and orienting the next step of the urban design. It has<br />
thus been demonstrated that GIS enhances the efficiency of site analysis.<br />
Acknowledgments<br />
The authors wish to express their appreciation to “Revitalization of Local Area Project” for providing the<br />
subvention for this research.<br />
References<br />
Dewanchker B., (1997) “Comparison of the Change in Industrial Development in the Kitakyushu Industrial Zone and<br />
the EMSCHER Zone in Germany”, Journal of architecture, planning and environmental engineering.<br />
Transactions of AIJ (502) pp.51-56<br />
Dewanchker B., Takahasi N. and Ojima T., (1997) “Ecological Improvement and the Redevelopment of an Abandon<br />
Industrial Site in the City of Uozu”, Summaries of technical papers of Annual Meeting Architectural Institute of<br />
Japan. pp.821-822<br />
Fukahori H. (2006), “Disaster prevention map making with IT method”, Journal of Japan Sewage Works Association,<br />
43(527) pp.43-46<br />
Fukahori H. (2007), “Make Kitakyusyu-city an nature environment capital”, Journal of sewerage, monthly 30(2) (422)<br />
pp.38〜41<br />
Gao W., Ariyama T. and Ojima T., (1999) “Prediction on the Energy Consumption of Renewing the Residential<br />
Recycle Material”, Journal of architecture, planning and environmental engineering. Transactions of AIJ (516)<br />
pp.101-106<br />
448
Yupeng Wang et al.<br />
Gao W. and Ojima T., (1995) “Study on Planning an Ecological City in SHITAMACHL Tokyo”, Journal of architecture,<br />
planning and environmental engineering. Transactions of AIJ (472) pp.63-71<br />
Gao W., (1993) “Thermal Effects of Open Space with a Green Area on Urban Environment: Part I: A theoretical<br />
analysis and its application”, Journal of architecture, planning and environmental engineering. Transactions of<br />
AIJ (448) pp.15-27<br />
Song H., Guan D., Fukahori H. and Gao W., (2009) “Study on utilization of semi-industrial districts based on GIS in<br />
Kitakyushu : Investigation of present status of semi-industrial districts in Kitakyushu Part 1”, Journal of Asian<br />
Architecture and Building Engineering, pp.529-532<br />
Urban planning general overview of Kitakyusyu city of Japan, (2007) Planning department in city urban planning<br />
bureau of Kitakyusyu, Japan<br />
449
The Impact of Software Test Constraints on Software Test<br />
Effectiveness<br />
Grafton Whyte and Donovan Lindsay Mulder<br />
University of the Western Cape, Cape Town, South Africa<br />
drgwhyte@aol.com<br />
donovanmulder@hotmail.com<br />
Abstract: Software testing is the one of the primary methods used in the validation and verification of output in the<br />
software development industry. It is seen as a key method for achieving software quality, reliability, fitness for<br />
purpose and customer satisfaction. Software testing is however an expensive process accounting for as much as<br />
50% of the cost of developing software based systems. In recent years, software testing as a discipline has come<br />
under pressure due to time, cost and skills constraints. These constraints impact negatively upon software test<br />
effectiveness. Therefore it is critical to identify and implement test tools that reduce the negative impact of software<br />
test constraints on software test effectiveness. In this paper the researcher examines some of the most popular<br />
software testing tools such as test case prioritisation, test suite reduction and test selection criteria, to identify: Which<br />
individual test tools are most likely to yield optimal test effectiveness and, Which combination of test tools is most<br />
likely to yield optimal test effectiveness and mitigate the effect of test constraints An extensive review of the software<br />
testing literature was conducted and used to construct a survey instrument as the basis for examining the impact of<br />
test constraints on software test methodology. The survey was issued to expert software test practitioners from<br />
various locations globally; the sample consisted of 43 test cases. The main findings were that no one approach to<br />
testing would yield satisfactory results but a combination of two or more test types from Automated testing, Smoke<br />
testing, Test case prioritisation and Regression test selection could yield effective software testing results and<br />
mitigate the effects of test constraints.<br />
Keywords: software test tools, software test effectiveness, software test constraints, test selection methodology, test<br />
case selection criteria<br />
1. Introduction<br />
As the significance of software increases aspects such as quality, reliability and customer satisfaction<br />
have become strategic goals for software development organisations (Chin-Yu, 2005). Software testing is<br />
a costly (Bryce and Colbourn, 2005) and unavoidable task (Bertolino, 2007). It is complex and arduous<br />
and can consume more than 50% of the cost of a software development project (Shahamiri, et al., 2009;<br />
McMinn, 2009; Srikanth and Williams, 2005). Engel and Last (2007) state that inadequate execution of<br />
software and systems verification, validation and testing account for losses that can eclipse more than<br />
10% of a company’s turnover.<br />
Software testing has been the method of choice for software validation and verification (Bertolino, 2007)<br />
and remains the method of choice through which confidence in the end product is realised (Ramler and<br />
Wolfmaier, 2006). In today’s software development environment testing has come under pressure due to<br />
shorter time to market, shrinking budgets and higher quality demands (Ramler and Wolfmaier, 2006).<br />
Shahamiri (2009) and Chin-Yu (2005) states that improvement of software quality and reliability through<br />
defect detection is a goal of software testing as is risk mitigation (Frank and Weyuker, 2000).<br />
Software testing drives are commonly beleaguered by constraints such as time, cost, and skill. These<br />
constraints risk the realisation of software testing goals and reduce test effectiveness. Understanding<br />
how to mitigate this risk is a key-factor for achieving successful software testing. Therefore the research<br />
question asked here is<br />
What test approaches are most likely to reduce the negative impact of test constraints on test<br />
effectiveness?<br />
Hence this research aims to identify software test tools or approaches likely to increase test effectiveness<br />
in the presence of test constraints.<br />
450
2. Literature review<br />
2.1 Definitions and terminology<br />
Grafton Whyte and Donovan Lindsay Mulder<br />
Software testing is defined as the observation of the execution of software based systems in order to<br />
verify that the system behaves as expected and to identify defects in the system under test (Bertolino,<br />
2007).<br />
Software testing effectiveness is defined as the number of defects found through software testing<br />
divided by the total number of defects (Vallespir and Herbert, 2009).<br />
For the purposes of this paper the researcher defines a:<br />
Test tool as any methodology, process or know-how that contributes to software testing and as such<br />
is not limited to software based test tools.<br />
Test approach as one or a combination of test tools used to implement software testing.<br />
Test constraint as any factor that inhibits the software testing process from achieving the desired<br />
levels of performance.<br />
Test Design refers to test case content and test case size.<br />
2.2 Test constraints<br />
According to a study by Chin-Yu (2005), defect detection rate is impacted by the skill level of test<br />
personnel and size of the project. As does testability (Berner, et al., 2005) of the system. Mirarab and<br />
Tahvildari (2008) argue that time constraints impact upon regression testing and negatively impact on the<br />
cost effectiveness of test case prioritization methodologies. Bryce and Colbourn (2005) state that due to<br />
time and cost constraints entire test suites in most instances are not run.<br />
2.3 Test design<br />
Rothermel, et al. (2004) discusses four regression test methodologies from a test design perspective.<br />
Two design factors are considered: test case size and test case content. The four methodologies are<br />
“retest-all, regression test selection, test case prioritisation, and test suite reduction”. All four are shown to<br />
have increased cost effectiveness, defect detection capability with reduced test execution time with the<br />
application of relevant test case design. It is shown that large test cases increases the cost effectiveness<br />
and reduce test execution time in the retest-all, regression test selection, and test case prioritisation<br />
methodologies. This was not shown to be true for test suite reduction. It was shown that functional<br />
grouping within test suite reduction could lead to greater defect detection capability. This however might<br />
not be true for difficult-to-find defects in all four methodologies. It is also shown that large test cases have<br />
a greater defect detection capability when compared to smaller test cases. Smaller test cases are more<br />
prone to finding hard to detect defects and lend itself to more effective test case prioritisation and<br />
regression test selection.<br />
2.4 Test selection methods<br />
One of the most common test selection methodologies is test case prioritisation. Test cases are ordered<br />
according to explicit test selection criteria. This is to enhance the defect detection rate early in the test<br />
process (Rothermel, et al., 2004; Srikanth and Williams, 2005) and decrease test execution time<br />
(Sampath, et al., 2008) with respect to the realisation of specific testing goals such as test all high risk<br />
functionality. TSC could be code coverage, possibility of defect existence and defect detection potential<br />
(Sampath, et al., 2008; Do, et al., 2008). Other TSC used to detect high risk defects quickly are<br />
“requirement volatility, customer priority, implementation complexity and fault proneness” (Srikanth and<br />
Williams, 2005).<br />
Regression test selection increases cost effectiveness however there are trade-offs one of which is<br />
between safety and efficiency. Regression test selection methods assume that test cases that do not test<br />
changed functionality will not detect defects (Engström 2010). The trade off here isthe risk of not finding<br />
all defects versus cost effectiveness and short test execution cycles.<br />
Another common test selection methodology is test suite reduction. The goal of test suite reduction is to<br />
reduce the cost of regression testing (Parse, et al., 2009) by satisfying all test requirements with the least<br />
amount of test cases (Zhang, et al., 2008a, b). Another aim of test suite reduction is to permanently<br />
451
Grafton Whyte and Donovan Lindsay Mulder<br />
remove redundant test cases from a test suite. This leads to reduced cost and time of test execution<br />
verification and test suite management (Rothermel, et al., 2004). Code and functional coverage and<br />
defect detection potential are commonly used as test selection (reduction) criteria (Parse and Khalilian<br />
and Fazlalizadeh, 2009). Discarding test cases can quickly result in a significant decrease in the defect<br />
detection capability of reduced test suites.<br />
2.5 Test oracles<br />
Test oracles are defined as an accepted dependable source of specified input and expected output of<br />
software behaviour and a means of reconciling expected and actual behaviour (Shahamiri, et al., 2009).<br />
From the definition it is clear that test oracles have a significant impact on the defect detection capability<br />
and cost effectiveness of a test suite which is a view supported by Memon, et al. (2003) as well as<br />
Memon and Xie (2004, 2005). A strong test oracle will have many verification points thus increasing test<br />
development, execution and verification time but increasing defect detection capability. A weak test<br />
oracle could result in reduced test execution time though this might be due to misleading or incomplete<br />
oracle information. In this case there are risks of defects not being detected. Memon and Qing (2005)<br />
also conclude that test cases lose their defect detection capability substantially, through decrepit test<br />
oracles; comprehensive test oracles employed at the end of the execution of a test case yields the best<br />
cost benefit ratio; some test oracles only detect defects during a small window of opportunity and<br />
comprehensive and regular application of comprehensive test oracles can counteract the impact of small<br />
test cases. This is supported by Rothermel, et al. (2004) who also states that large test cases minimises<br />
the impact of weak test oracles on test effectiveness.<br />
2.6 Smoke testing<br />
Smoke testing is used to detect defects early in the software development lifecycle (Memon, et al. 2003).<br />
It is widely accepted that smoke testing leads to lower defect resolution costs, lower costs of formal<br />
testing, enhanced software quality and risk minimization.Memon and Xie (2004) conclude that smoke<br />
tests detect greater than 60% of defects; small smoke tests execute a large percentage of code; larger<br />
smoke tests are able to detect more defects; test oracles have a substantial impact upon the defect<br />
detection capability of smoke tests.<br />
2.7 Automated testing<br />
Shahamiri and Kadir and Mohd-Hashim (2009) argue that test automation has been the one method used<br />
to decrease the costs of software testing. This is supported by Ramler and Wolfmaier (2006) and Karhu,<br />
et al. (2009) who also states that automated testing can be used in place of manual testing when time is<br />
a constraint. This is further supported by Zhu, et al. (2006) where it is stated that in order to reduce the<br />
costs and improve software test effectiveness it is critical to automate the testing process.<br />
2.8 Literature review conclusion<br />
The Software Test Tool Model (figure 1) serves as a summary of the literature review. This model<br />
summarises the relationship between Test Tools (Test Selection Criteria, Test Selection Methods, and<br />
Test Execution Methods) and Test Effectiveness as conceptualised for the purpose of this research.<br />
Derived from the ideas and concepts discussed in the literature review, this model provides a visual map<br />
of how all the test tools and test selection criteria fit together to form a software testing approach and how<br />
these factors collectively can increase test effectiveness.<br />
The Software Test Tool model works as follows: the Test Selection Criteria impacts upon the<br />
effectiveness of the Test Selection Methods which in turn impacts Software Test Effectiveness. Test<br />
Selection Criteria applied to Test Cases which in turn impacts upon Software Test Effectiveness. Test<br />
Design and Test Oracles impacts upon the defect detection rate and capability of Test Cases which in<br />
turn impacts upon Software Test Effectiveness. Test Selection Criteria impacts the test execution time of<br />
Test Execution Methods which in turn impacts upon Software Test Effectiveness.<br />
From the literature it is concluded that:<br />
Skill of human resources, project size, inadequate requirements, software testability, time, cost and<br />
test design are test constraints falling into the categories of time, cost and skill.<br />
Retest-all, regression test selection, test case prioritisation, test suite reduction, smoke testing, test<br />
automation, test oracles and test design are determined from the literature to be test tools used to<br />
minimise the impact test constraints on test effectiveness.<br />
452
Grafton Whyte and Donovan Lindsay Mulder<br />
Test selection criteria such as code coverage, possibility of defect existence, defect detection<br />
potential, test case design (test case content and size);and requirement volatility, customer priority,<br />
implementation complexity and fault proneness are identified as test selection criteria which can<br />
support test case prioritisation.<br />
These test tools and test selection criteria were found to directly affect test effectiveness through four<br />
aspects which featured prominently in the reviewed literature; these are (a) defect detection<br />
capability, (b) defect detection rate, (c) cost effectiveness and (d) test execution time.<br />
Figure 1: Software effectiveness model<br />
3. Research methodology<br />
The ideas discussed in the literature review embody the Goal-Question Metric Approach template<br />
proposed by Basili et al. (1994): They propose analysing software testing tools, software test selection<br />
criteria and software test constraints for the purpose of knowing which software testing tools and test<br />
case selection criteria or a combination thereof increases software test effectiveness in the presence of<br />
software test constraints with respect to their usage and possible software testing improvement potential<br />
from the view point of software testing experts and professionals.<br />
There are a plethora of test tools and test selection criteria in the current literature that can arguably<br />
minimise the negative impact of test constraints on test effectiveness. The usage of these test tools and<br />
test selection criteria differ widely, while information about the actual usage in the presence of test<br />
constraints in project situations seems to be missing. To assess the extent to which test tools minimise<br />
the negative impact, a research instrument was needed that could be measure the perceived impact of<br />
test tools on test effectiveness in the presence of test constraints. No such instrument was found in the<br />
literature. Most if not all of the research in the literature dealt with test tools used under experimental<br />
conditions and not in real projects. Also it is not certain how these test tools and test selection criteria<br />
would contribute to the overall test effectiveness in live projects. Therefore a research instrument was<br />
developed based on test tools and test selection criteria found in the literature and that could be used to<br />
measure the perceived impact on actual projects.<br />
453
Grafton Whyte and Donovan Lindsay Mulder<br />
A survey questionnaire consisting of closed questions was used as the research instrument. Quantitative<br />
data was collected. The research constructs are test tools, test case selection criteria, test constraints<br />
and test effectiveness. The survey was designed to elicit data from practitioners’ perspectives and<br />
experience. The aim was to identify the perceived impact (a) test tools have on test effectiveness, (b) test<br />
selection criteria on test effectiveness, (C) test constraints on test effectiveness and (d) actual test<br />
effectiveness of testing activities. The research instrument was piloted using 2 case studies. The pilot<br />
study was interview based in order to gauge first-hand the effectiveness of the survey instrument. The<br />
instrument yielded coherent results with minor modifications.<br />
The research population was the world of software testing professionals in which one of the authors was<br />
a recognised member. Convenience sampling was used to derive the sample from the population. Each<br />
respondent was asked to provide data for three projects. Each project was treated as a single case. In<br />
total 43 cases was reported on by the respondents. The survey questionnaire was sent to 14 members of<br />
the sample of which 2 did not respond effectively achieving a response rate 85.7% with respect to the<br />
initial sample. However the sample size grew to 18 with a total of 16 responses (88.9%). Of the possible<br />
47 possible cases 43 were deemed acceptable as the value of software test effectiveness was omitted in<br />
4 cases. Respondents were located in geographically dispersed locations (USA, England, Ireland,<br />
France, South Africa, India and Australia). The predominant roles were test managers and test leads (see<br />
figure 2). The case studies covered a range of industries with financial services and telecoms dominating<br />
(see figure 3).<br />
3.1 Validity<br />
Construct Validity: Observations were based on the respondents experience and as such might have led<br />
to some differences in interpretation to mitigate, definitions of all test variables were defined in the<br />
questionnaire.<br />
External Validity: 43 case studies were included in this research. The results can be generalised to a<br />
certain extent as the case studies are from different industries and from geographically dispersed<br />
locations .Although convenience sampling was used to select respondents due to the dispersed nature of<br />
the sample it can be considered a fair representation of the population.<br />
Internal Validity: The survey questionnaire was designed with a certain goal in mind; there were no<br />
extraneous questions therefore, it is reasonable to conclude that the instrument had internal validity.<br />
3.2 Reliability<br />
Joppe (2000) defines reliability as: “The extent to which results are consistent over time and an accurate<br />
representation of the total population under study is referred to as reliability and if the results of a study<br />
can be reproduced under a similar methodology, then the research instrument is considered to be<br />
reliable.” The survey tool is based upon the Goal-Question Metric Approach template proposed by Basili,<br />
et al. (1994), the tool demonstrated acceptable levels of reliability in their research therefore, reliability is<br />
assumed.<br />
Figure 2: Distribution of roles<br />
454
Figure 3: Distribution of industries<br />
4. Research results<br />
Grafton Whyte and Donovan Lindsay Mulder<br />
According to the Goal-Question Metric approach (Basili et al.1994) for measurement to be useful from an<br />
organisational view point it must be goal driven and specific. The purpose of measurement in this<br />
research is to identify the testing tools most likely to overcome organisational constraints and yield<br />
optimal test effectiveness.<br />
From the outset two goals were set for this research:<br />
1. Identify individual Test Tools (TT) and Test Selection Criteria (TSC) most likely to yield best Test<br />
Effectiveness (TE). It was determined that the best statistical approach for identifying these<br />
constructs would be correlation analysis.<br />
2. Identify the combination of Test Tools (TT) and Test Selection Criteria (TSC) most likely to yield<br />
best Test Effectiveness (TE). It was determined that the best statistical approach for identifying these<br />
constructs would be multiple regression analysis.<br />
4.1 Descriptives<br />
Constructs for this research were organised into three construct groups:<br />
Software Test Tools (TT) – Table 1.0lists each tool as presented to respondents, a definition was also<br />
added to ensure a common understanding. Tools are listed here with group descriptions where<br />
appropriate. Respondents were asked to rate for each project whether the test tool increased test<br />
effectiveness. Their responses were graded on a Likert scale of 1 to 5 denoting ‘complete disagreement’<br />
to ‘complete agreement’, respectively. Sample size (N=43) in every case.<br />
Table 1: Software Test Tools<br />
Type Description<br />
Test Design<br />
Variable<br />
Label<br />
Mean<br />
Std. Dev.<br />
Test case content TT1 3.77 1.269<br />
Test case granularity TT2 3.33 1.375<br />
Test oracles<br />
Test selection method<br />
TT3 2.63 1.865<br />
Regression test selection TT4 3.35 1.526<br />
455
Grafton Whyte and Donovan Lindsay Mulder<br />
Type Description Variable<br />
Label<br />
Mean<br />
Std. Dev.<br />
Re-test all TT5 2.65 1.462<br />
Test suite reduction TT6 2.19 1.577<br />
Test case prioritisation TT7 3.91 1.151<br />
Test execution method<br />
Automated testing TT8 2.70 1.859<br />
Manual testing TT9 4.19 1.220<br />
Smoke testing TT10 3.60 1.788<br />
Software Test Selection Criteria (TSC) – Table 2.0 list each test selection criteria. Respondents<br />
received them with definitions and were asked to rate projects on each criterion; whether these test<br />
selection criteria increase software test effectiveness. Responses were graded on a Likert scale of 1 to 5<br />
denoting ‘no increase’ to ‘significantly increased’, respectively.<br />
Table 2: Software Test Selection Criteria<br />
Type Description Variable<br />
Mean<br />
Std. Dev.<br />
Label<br />
Test Case<br />
Code coverage TSC1 1.40 1.400<br />
Functional coverage TSC2 3.81 1.118<br />
Defect detection capability TSC3 3.09 1.702<br />
Setup time TSC4 2.74 1.706<br />
Implementation complexity TSC5 2.86 1.207<br />
Execution time TSC6 3.58 1.239<br />
Requirements<br />
Customer priority TSC7 3.63 1.543<br />
Implementation complexity TSC8 2.88 1.562<br />
Fault proneness TSC9 2.51 1.907<br />
Volatility TSC10 2.79 1.767<br />
Test Case Composition<br />
Test case content TSC11 3.53 1.182<br />
Test case granularity TSC12 2.88 1.219<br />
No test selection criteria TSC13 1.74 1.853<br />
Test Constraints (TC) – Table 3 list the main test constraints encountered by test projects, again<br />
respondents were asked to rate projects on each criterion, whether these test constraints hindered<br />
software test effectiveness. Responses were graded on a Likert scale of 1 to 5 denoting ‘no impact’ to<br />
‘significantly impacted’, respectively.<br />
Table 3: Test Constraints<br />
Type Description<br />
Test Constraints<br />
Variable<br />
Label<br />
Mean<br />
Std. Dev.<br />
Time Time 3.72 1.260<br />
Cost Costs 2.93 1.534<br />
Skills Skills 2.81 1.332<br />
The relationship between these variables and test effectiveness can be represented in the following linear<br />
function<br />
Test effectiveness (TE) = Test Tools (TT) + Test Selection Criteria (TSC) – Test Constraints (TC)<br />
Subsequent correlation and regression tests explored the applicability of this function.<br />
4.2 Goal 1 - correlation tests<br />
Pearson correlation was run for each of the constructs independently (assuming two of the three<br />
constructs were zero) to see which of the underlying variables correlates with the construct Test<br />
Effectiveness (TE).<br />
456
Grafton Whyte and Donovan Lindsay Mulder<br />
The results from Table 4 suggest that regression testing selection (TT4) and smoke testing (TT10)<br />
correlate strongly with test effectiveness (TE), with 54% and 44% of the variability in TE explained by<br />
these two variables, with both achieving significance levels in the 99 percentile, suggesting these results<br />
are very reliable. Two further variables, test case prioritisation (TT7) and automated testing (TT8)<br />
correlate less strongly with TE (35% and 38%, respectively), but still achieve significance levels in the 95<br />
percentile.<br />
Table 4: Test Tools (TT) correlated with TE<br />
Type Description Variable<br />
Pearson<br />
Significance<br />
Test Design<br />
Label<br />
Correlation<br />
(2-tailed)<br />
Test case content TT1 -0.036 0.821<br />
Test case granularity TT2 0.099 0.527<br />
Test oracles<br />
Test selection method<br />
TT3 0.206 0.185<br />
Regression test selection TT4 0.536 0.000**<br />
Re-test all TT5 0.016 0.919<br />
Test suite reduction TT6 0.106 0.501<br />
Test case prioritisation<br />
Test execution method<br />
TT7 0.347 0.023*<br />
Automated testing TT8 0.384 0.011*<br />
Manual testing TT9 -0.129 0.409<br />
Smoke testing TT10 0.444 0.003**<br />
From Table 5 we note that one variable, defect detection capability (TSC3), achieved correlation of 39%<br />
at significance levels in the 99 percentile (0.010). Other variables that achieved high significance levels in<br />
the 95 percentile are functional coverage, customer priority, test case content and no test selection<br />
criteria (test all) with respective correlations of 31%, 35%, 31% and 34%.<br />
Table 5: Test Selection Criteria (TSC) correlated with TE<br />
Type Description Variable<br />
Pearson<br />
Significance<br />
Test Case<br />
Label<br />
Correlation<br />
(2-tailed)<br />
Code coverage TSC1 0.219 0.159<br />
Functional coverage TSC2 0.315 0.039*<br />
Defect detection capability TSC3 0.390 0.010**<br />
Setup time TSC4 0.270 0.080<br />
Implementation complexity TSC5 0.196 0.207<br />
Execution time<br />
Requirements<br />
TSC6 0.071 0.649<br />
Customer priority TSC7 0.354 0.020*<br />
Implementation complexity TSC8 0.017 0.914<br />
Fault proneness TSC9 0.155 0.320<br />
Volatility<br />
Test Case Composition<br />
TSC10 0.106 0.499<br />
Test case content TSC11 0.306 0.046*<br />
Test case granularity TSC12 0.167 0.285<br />
No test selection criteria TSC13 0.343 0.024*<br />
There were no significant correlations with any variables in the test constraint construct (TC).<br />
4.3 Goal 2 – multiple regression testing<br />
Having identified variables to emerge from independent tests of association with Test Effectiveness,<br />
attention was turned to testing the linear function in totality using multiple regression analysis to see if the<br />
variables that emerged in the correlation test would hold or would new variables emerge as predictors of<br />
test effectiveness.<br />
457
Grafton Whyte and Donovan Lindsay Mulder<br />
Data for the model were analysed using linear multiple regression analysis. The procedure estimates the<br />
coefficients (beta) of one or more independent variables to predict (R²) the value of a single dependent<br />
variable. Variables are systematically entered and removed from the equation using the stepwise method<br />
to determine the line of best fit.<br />
Table 6: Linear Regression model<br />
Dependent Variable Independent Variables Beta R Adjusted R² Significance<br />
Test Effectiveness<br />
(TE)<br />
0.718 0.479 .000<br />
Regression test selection<br />
(TT4)<br />
0.465 .000<br />
Smoke testing (TT10) 0.340 .005<br />
No test selection criteria<br />
(TSC13)<br />
0.337 .004<br />
The model suggests that 48% of the variability in the construct Test Effectiveness (TE) is explained by<br />
the variables TT4, TT10 and TSC13. The proportion that each of the independent variables explain is<br />
indicated by their beta percentages, 46%, 34% & 33% respectively.<br />
5. Discussion<br />
Test tools and test selection are applied quite extensively in the case studies investigated. Given the<br />
diversity of the population it can be assumed this is likely to be the case for the Software Testing industry<br />
globally. This indicates that software testing practice is becoming more rigorous and formalised and to a<br />
certain extent scientific, which is a good sign for the software development industry and business as a<br />
whole.<br />
The results suggest regression test selection (TT4) and smoke testing (TT10) significantly, and to a<br />
lesser extent test case prioritisation (TT7) and automated testing TT11) correlated with Test<br />
Effectiveness. These Test Tools can each be applied individually to testing activities in order to increase<br />
Test Effectiveness.<br />
In practice these Test Tools are most associated with increased Test Effectiveness and their use is more<br />
likely to increase the probability that a software development project will meet its strategic goals.<br />
Smoke testing is used to detect defects before more expensive formal testing. Automated testing<br />
significantly reduces test execution time and also increases functional coverage during testing in shorter<br />
test cycles. Individually the application of these Test Tools will lead to early detection of defects, reduced<br />
cost of testing, quicker time to market and increased software quality.<br />
Regression test selection and test case prioritisation are dependent upon adequate test selection criteria.<br />
One test selection criterion that correlated strongly with Test Effectiveness is defect detection capability<br />
(TSC3). Other test selection criteria that correlated with slightly less significance are functional coverage<br />
(TSC2), customer priority (TSC7), test case content (TSC11) and no test selection criteria (TSC13).<br />
Defect detection capability (TSC3) is an imperative design consideration. It is the make or break factor in<br />
the realisation of software quality through increased test effectiveness. It plays a pivotal role in reducing<br />
the risk of software going to market with undetected defects. Maximised functional coverage (TSC2)<br />
during test execution reduces the risk of parts of a software system being untested after it has been<br />
released to market. Test case content (TSC11) directly relates to functional coverage and as such is<br />
used to measure functional coverage of software testing, thus giving the business a risk based view that<br />
can be used to determine the readiness of a software system before release to market. Customer priority<br />
(TSC7) allows software development businesses to quickly meet the most important goals of the<br />
customer. No test selection criteria (TSC13) translates into testing of the entire software system thereby<br />
mitigating all risk posed by inadequate functional coverage. However this is an expensive approach to<br />
software testing and requires no or very little time constraints.<br />
Individually these Test Selection Criteria increase software quality, customer satisfaction thereby possibly<br />
enhancing and preserving a business’s good reputation.<br />
458
Grafton Whyte and Donovan Lindsay Mulder<br />
Applying multiple regression analysis to the combined Test Tools (TT) and Test Selection Criteria (TSC)<br />
identified the combined variables of regression test selection (TT4), smoke testing (TT10) and no test<br />
selection criteria (TSC13) to have the most significant impact upon Test Effectiveness.<br />
In practice this result suggest that smoke testing (TT10) should be used for early detection of major<br />
defects after which regression test selection (TT4) should be applied to the software under test and that<br />
no test selection (TSC13) be applied to test cases, therefore test everything. Testing everything is an<br />
approach best used for major software release prior to go-to-market. Though this is expensive it mitigates<br />
the risk of defects ‘making it out into the wild’ resulting in damaged reputation and possible significant<br />
economic loss and even loss of life. During minor releases where it does not make economic sense to do<br />
exhaustive testing with full functional coverage one or more of the Test Selection Criteria identified<br />
through the correlation analysis should be applied. This will lead to reduced test execution costs and time<br />
with increased defect detection rate in system under test.<br />
The data analysis strongly suggested a saturation point in the application of the number of Test Tools<br />
and Test Selection Criteria, exactly at which point this achieved is debatable but initial indications from<br />
this research suggests after a 80% level of Test Effectiveness has been achieved. The total number of<br />
Test Tools and Test Selection Criteria in this study is twenty-three (23). Correlation and multiple<br />
regression analyses reduced this to nine and three respectively, suggesting a high-degree of over-lap<br />
between tests. In business terms it would seem not to make sense to try and achieve levels of Test<br />
Effectiveness beyond the 80 percentile point.<br />
Interestingly, neither correlation nor multiple regression analyses revealed a significant relationship<br />
between Test Constraints and Test Effectiveness. Given that organisational constraints are always<br />
uppermost in practitioner’s minds. One reason for this absence could be due to the fact that the sample<br />
was made up of seasoned test experts, who through their experience have learnt how to mitigate the<br />
impact of Test Constraints on Test Effectiveness.<br />
6. Conclusion<br />
The aim of this paper was to identify from the plethora of Test Tools used in practice to determine<br />
1. Which individual test tools are most likely to yield optimal test effectiveness and,<br />
2. Which combination of test tools is most likely to yield optimal test effectiveness and mitigate the<br />
effect of test constraints<br />
Data from forty three cases across various industries and countries were collected to identify the current<br />
application of Test Tools in practise. The research was designed using the Goal-Question Metric<br />
approach and the data analysed using correlation analysis to identify individual Test Tools and Test<br />
Selection Criteria most closely associated with Test Effectiveness and; multiple regression analysis to<br />
identify the combination of Test Tools and Test Selection Criteria that would lead to optimum Test<br />
Effectiveness.<br />
The correlation analysis identified nine variables; Four Test Tools (TT): regression test selection (TT4),<br />
smoke testing (TT10), test case prioritisation (TT7) and automated testing (TT11) and, five Test Selection<br />
Criteria (TSC): defect detection capability (TSC3), functional coverage (TSC2), customer priority (TSC7),<br />
test case content (TSC11) and no test selection criteria (TSC13), allowing test practitioners to pick and<br />
mix the various approaches given the specific constraints.<br />
The multiple regression analysis identified the combined variables of regression test selection (TT4),<br />
smoke testing (TT10) and no test selection criteria (TSC13) to have the most significant impact upon Test<br />
Effectiveness.<br />
The main value of this research is that it begins to spell out for the growing community of test<br />
practitioners who have limited experience in the field of software testing and are confronted by a<br />
bewildering array of test options, which individual and combined test techniques will most likely help them<br />
achieve optimal test effectiveness.<br />
Areas of further research:<br />
A saturation point in the application of Test Tools and Test selection criteria is alluded to in this<br />
paper, more evidence to support this idea is required and to pinpoint the threshold. This will enable<br />
459
Grafton Whyte and Donovan Lindsay Mulder<br />
Of the research sample the majority of test cases were overwhelmingly successful, further research<br />
needs to be conducted on cases which are not successful to ascertain if the emergent variables<br />
remain consistent.<br />
References<br />
Basili, V.R. Caldiera, G., Rombach, H.D. 1994, 'The Goal Question Metric Approach', Encyclopedia of Software<br />
Engineering<br />
Berner, S., Weber, R., Keller, R.K. 2005, 'Observations and Lessons Learned from Automated Testing', Zühlke<br />
Engineering AG Zürich-Schlieren Switzerland, 2005.<br />
Bertolino, A 2007, 'Software Testing Research: Achievements, Challenges, Dreams', no. Future of Software<br />
Engineering (FOSE'07).<br />
Bryce, R.C., Colbourn, C.J. 2005, 'Test Prioritization for Pairwise Interaction Coverage', A-MOST’05, Copyright 2005<br />
ACM 1-59593-115-5/00/0004, St. Louis, Missouri, USA.<br />
Chin-Yu, H. 2005, 'Performance analysis of software reliability growth models with testing-effort and change-point',<br />
The Journal of Systems and Software, vol 76, pp 181 - 194.<br />
Do, H., Mirarab, S., Tahvildari, L. 2008, 'An Empirical Study of the Effect of Time Constraints on the Cost-Benefits of<br />
Regression Testing', SIGSOFT 2008/FSE-16, November 9–15, Copyright 2008 ACM 978-1-59593-995-1,<br />
Atlanta, Georgia, USA.<br />
Engel, A., Last, M. 2007, 'Modeling software testing costs and risks using fuzzy logic paradigm', The Journal of<br />
Systems and Software, 2007.<br />
Engström, E 2010, 'Regression Test Selection and Product Line System Testing', 2010 Third International<br />
<strong>Conference</strong> on Software Testing, Verification and Validation, IEEE Computer Society, Paris.<br />
Frank, P. G. , Weyuker, E. J. 2000, 'Testing software to detect and reduce risk', The Journal of Systems and<br />
Software, 2000.<br />
Joppe, M. 2000. 'The Research Process', Available at [Accessed 08March<br />
2009]<br />
Karhu, K., Repo, T., Taipale, O., Smolander, K. 2009, 'Empirical Observations on Software Testing Automation',<br />
2009 International <strong>Conference</strong> on Software Testing Verification and Validation, IEEE Computer Society, Denver<br />
Colorado.<br />
Memon, A.M., Bananerjee, I., Hashmi, N., Nagarajan, A. 2003, 'DART: A Framework for Regression Testing<br />
Nightly/daily Builds of GUI Applications', Proceedings of the International <strong>Conference</strong> on Software Maintenance<br />
(ICSM’03), IEEE Computer Society.<br />
Memon, A.M., Xie, Q. 2004, 'Empirical evaluation of the fault-detection effectiveness of smoke regression test cases<br />
for GUI-based software', Proceedings of the 20th International <strong>Conference</strong> on software Maintenance (ICSM<br />
2004), Chicago, USA.<br />
Memon, A.M., Xie., Q. 2005, 'Studying the Fault-Detection Effectiveness of GUI Test Cases for Rapidly Evolving<br />
Software', IEEE Transactions on Software Engineering, 2005.<br />
Parse, S., Khalilian, A., Fazlalizadeh, Y. 2009, 'A New Algorithm to Test Suite Reduction Based on Cluster Analysis',<br />
2009.<br />
Ramler, R., Wolfmaier, K. 2006, 'Economic Perspectives in Test Automation: Balancing Automated and Manual<br />
Testing with Opportunity Cost"'.<br />
Rothermel, G., Elbaum, S., Malishevsky, A.G. , Kallakuri, P., Xuemei Q. 2004, 'On Test Suite Composition and Cost-<br />
Effective Regression Testing', ACM Transactions on Software Engineering and Methodology, 2004.<br />
Sampath, S., Ren´ee C. B., Gokulanand, V., Vani K., A. Gunes¸ K. 2008, 'Prioritizing User-session-based Test Cases<br />
for Web Applications Testing', International <strong>Conference</strong> on Software Testing, Verification, and Validation, IEEE<br />
Computer Society.<br />
Shahamiri, S.R., Kadir, W.M.N.W, Mohd-Hashim. S.Z. 2009, 'A Comparative Study on Automated Software Test<br />
Oracle Methods', Fourth International <strong>Conference</strong> on Software Engineering Advances, IEEE Computer Society.<br />
Srikanth, H., Williams, L. 2005, 'On the Economics of Requirements-Based Test Case Prioritization', EDSER'05,<br />
Copyright 2005 ACM 1-59593-118-X/05/0005, St. Louis, Missouri, USA.<br />
Srikanth, H., Williams, L., Osborne, J. 2005, 'System Test Case Prioritization of New and Regression Test Cases',<br />
International Symposium on Empirical Software Engineering, IEEE Computer Society.<br />
Vallespir, D., Herbert, J. 2009, 'Effectiveness and Cost of Verification Techniques', IEEE Computer Society, Mexican<br />
International <strong>Conference</strong> on Computer Science.<br />
Zhang, R., Jiang, J., Yin, J., Jin, A., Lou, J., Wu, Y. 2008, 'A New Method for Test Suite Reduction', The 9th<br />
International <strong>Conference</strong> for Young Computer Scientists, IEEE Computer Society.<br />
Zhang, X., Xu, B., Chen, Z., Nie. C., Li, L. 2008, 'An Empirical Evaluation of Test Suite Reduction for Boolean<br />
Specification-based Testing', The Eighth International <strong>Conference</strong> on Quality Software, IEEE Computer Society.<br />
Zhu, H., Horgan, J.R., Cheung, S.C., Li, J.J. 2006, 'The First International Workshop on Automation of Software<br />
Test', ICSE’06, ACM 1-59593-085-X/06/0005, Shanghai, China.<br />
460
Strategic Considerations for Effective Mapping of<br />
Educational Technology to <strong>Academic</strong> Outcomes<br />
Hossein Zadeh, Arthur Shelley and Rod McCrohan<br />
RMIT University, Melbourne, Australia<br />
hossein.zadeh@rmit.edu.au<br />
arthur.shelley@rmit.edu.au<br />
rod.mccrohan@rmit.edu.au<br />
Abstract: Advancements in IT and web-based technologies in the past two decades have the potential to<br />
revolutionize higher education, but realizing this potential seems to remain an elusive goal. Educators in diverse<br />
contexts are constantly exploring innovative technologies that claim to add value to traditional classroom knowledge<br />
delivery and impact the course delivery and design in colleges and universities. Even though the use of technology<br />
in higher education has substantially increased over the past few years, this use has been primarily limited to content<br />
management and delivery. There are, however, much broader strategic issues around alignment between the suite<br />
of learning technology tools promoted by universities and the academic plan, learning and teaching priorities, and<br />
expected graduate capabilities. To date many universities have not undertaken a full strategic assessment of the<br />
issues and implications of the introduction and sustaining of educational technologies and how these can best be<br />
utilized to fulfill academic policy and add value to student learning experiences and outcomes. We conducted a focus<br />
group of a number of innovative and early technology-adopter university teachers in which they discussed how<br />
educational technology tools (Ed-Tools) can be embedded into courses such that they develop students’ analytical,<br />
sense-making and synthesis skills whilst enhancing their decision-making and delivery capabilities. This study<br />
highlights the importance of effectively using educational technology to support academic staff with their teaching<br />
and learning. A distinguishing factor of this study is that it looks at the issue from a holistic point of view, and not<br />
merely from the teachers’, or the administration, or the students’ point of view. We have compiled a list of barriers<br />
and also a list of recommendations. Findings of this exploratory study are presented intentionally in the form of a<br />
discussion paper rather than a traditional academic paper. It is hoped this presentation format opens up the issues<br />
for in-depth consideration by learning and teaching policy makers at universities and colleges.<br />
Keywords: educational technology, ed-tech, ed-tools, teaching and learning, graduate capabilities<br />
1. Introduction<br />
Advancements in IT and web-based technologies in the past two decades have the potential to<br />
revolutionize higher education, but realizing this potential seems to remain an elusive goal. Educators in<br />
diverse contexts are constantly exploring innovative technologies that claim to add value to traditional<br />
classroom knowledge delivery and impact the course delivery and design in colleges and universities<br />
(Barnett, Keating, Harwook, & Saam, 2004 as cited in Ajjan & Hartshorne, 2008). Numerous past studies<br />
have shown that technology use in higher education has increased (Maloney, 2007). However; this use<br />
has been primarily limited to content management and delivery, such as accessing course materials.<br />
Dalsgaard (2006) noted that whether the focus is on distance education or campus based education,<br />
universities all over the world are using information management systems to support and improve<br />
learning within their institutions. Learning, however, is far more than transfer of content. In the<br />
experience of the authors, the extent to which these technologies are actually applied varies to a great<br />
extent; some courses use technology merely as content management while others engage in a variety of<br />
more innovative approaches to actively engage students in interactive learning experiences.<br />
To be effective, learning and teaching requires preparing the students to cope with making decisions in<br />
an unpredictable future environment. The Web, for example, has grown into a global information space<br />
with more than a billion users and has entered a new, more social and participatory phase where people<br />
are engaging in different types of online collaborative activities (Anderson, 2007). This new form of the<br />
Web, often referred to as Web 2.0, allows people to communicate knowledge, share resources and<br />
participate in social networks (Ding, Jacob, Caverlee, Fried & Zhang, 2009). With the emergence of Web<br />
2.0, students no longer access the Web only for course information; instead they access and create<br />
collective knowledge through social interactions (Maloney, 2007).<br />
1.1 Learning support systems<br />
Almost all digital object support systems (online or local) share many of the same functions, but differ in<br />
the environments in which they are used. Course management systems (CMSs), learning content<br />
management systems (LCMSs) and virtual learning environments (VLEs) are generally used to facilitate<br />
461
learning in an academic setting. While both the terms CMS and VLE can be used interchangeably, these<br />
terms seem to be region specific. CMS is more commonly used in North America and VLE is more<br />
prevalent in <strong>European</strong> countries (Weller 2007 as cited in Daniels, 2009). These eLearning acronyms are<br />
further complicated with the more recent introduction of the PLE or personal learning environment,<br />
intended to shift greater control of learning environments to the learners by giving them greater flexibility<br />
in choosing their learning tools (Sclater 2008 as cited in Daniels, 2009).<br />
Learning management systems (LMS) typically are used to train employees in business environments.<br />
CMSs are widely used by colleges and universities to manage and deliver course content presented via<br />
the web. For example, MIT uses Stellar; the University of Michigan uses CHEF; and Stanford uses<br />
CourseWork. Often these systems designed for a particular institution are not publicly available.<br />
Educational institutions without adequate resources to develop a personalized CMS can employ an open<br />
source CMS such as Moodle, LRN, or Sakai. For institutions needing additional technical support,<br />
proprietary learning systems are available from companies such as Blackboard and Desire2Learn. The<br />
primary functions of CMSs are to organize and distribute course content, administer learning exercises or<br />
quizzes, and track student progress. While CMS software is often used to manage distance learning<br />
courses, it is just as popular in supporting face-to-face instruction. It should be noted that all these<br />
information management systems are based on web 2.0 technologies (Daniels, 2009).<br />
1.2 Web 2.0<br />
Web 2.0 also referred to as ‘social software’ covers a range of tools that allows users to interact and<br />
share data with each other via the internet. Blogs, wikis, podcasts, social networking web sites (facebook,<br />
twitter, MySpace etc.) and social bookmarking sites, such as Delicious, and 3D environments such as<br />
Second Life are examples of some of the tools that are being used to share and collaborate in social,<br />
educational and business contexts. The Web 2.0 “read/write” idea is not new as there were listservs,<br />
groupware, and web-based communities linking people with common interests even before the advent of<br />
Web 2.0 (Alexander, 2006 as cited in Ajjan & Hartshorne, 2008).<br />
In the educational context Web 2.0 has reshaped both the way instructors teach and the way students<br />
learn. These new technologies have simplified content sharing and transformed knowledge creation,<br />
usage and distribution (Dearstyne, 2007). With the use of Web 2.0, students no longer access the web<br />
only for course information; instead they access and create collective knowledge through social<br />
interactions (Maloney, 2007 as cited in Ajjan & Hartshorne, 2008). Table 1 lists some of the more<br />
commonly used Web 2.0 educational technology tools in use today. Dalsgaard (2006) argues that Web<br />
2.0 technologies support self-governed, problem- based and collaborative learning processes unlike<br />
other educational viewpoints where the responsibility rests with the teacher to deliver knowledge while<br />
the learner passively receives it. Ferdig (2007, as cited in Ajjan & Hartshorne, 2008) also notes that Web<br />
2.0 applications can support pedagogical approaches such as active learning, social learning, and<br />
student publication, by providing environments and technologies that promote and foster these<br />
interactions.<br />
With the emergence of the new learning environment, it is imperative for universities to ensure that<br />
educational technologies have positive impact in assisting the current learning and teaching practices<br />
Table 1: List of commonly used Web 2.0 educational technology tools<br />
List of commonly used<br />
Ed-Tools<br />
Audio blogs, Video Blogs & Podcasts<br />
Wikis, Discussion boards, Instant<br />
Messaging (IM)<br />
E-portfolios & blogs<br />
Social bookmarks<br />
Tele and Video Conferencing<br />
Function<br />
Sharing ideas, information and content/creations<br />
Working and collaborating with others in a shared work area<br />
Collection and/or presentation of evidence of experiences, thinking over<br />
time, productions, etc.<br />
To exchange information, ideas, resources, materials<br />
Conversations for geographically spread resources<br />
462
Some examples of benefits of the use of Ed-Tools are presented in table 2<br />
Table 2: Examples of benefits from implementation of Ed-Tools<br />
Theme Comment Reference<br />
Interaction,<br />
communication and<br />
collaboration<br />
Knowledge creation<br />
Ease of use and flexibility<br />
Writing skills & motivation<br />
Insight<br />
A sense of community and<br />
resource sharing amongst users.<br />
Active and student-centered<br />
learning. Educators as facilitators<br />
User friendly and flexible learning<br />
technologies<br />
Peer assessment & feedback.<br />
Global exposure and writing<br />
proficiency<br />
Educators gain an insight in to the<br />
world of students<br />
Alexander, 2006; Brown & Adler, 2008;<br />
Hartshorne & Ajjan, 2009; Richardson, 2009;<br />
Dalsgaard, 2006<br />
Williams & An, 2010<br />
Alexander, 2006; Brown & Adler, 2008;<br />
Hartshorne & Ajjan, 2009; Richardson, 2009;<br />
Dalsgaard, 2006<br />
(Ajjan & Hartshorne, 2008)<br />
Williams & An, 2010<br />
These practices should support the seven principles of good practice in undergraduate education,<br />
namely:<br />
Good practice encourages interaction between students and faculty<br />
Good practice encourages interaction and collaboration between students<br />
Good practice uses active learning techniques<br />
Good practice gives prompt feedback<br />
Good practice emphasizes time on task<br />
Good practice communicates high expectations<br />
Good practice respects diversity – talents, experience, and ways of learning<br />
Furthermore, many universities consider as priority engagement with communities, enterprises and<br />
industry, and strive to produce graduates who are work-ready professionals.<br />
In order to prepare such graduates, universities need to have educational technologies adopted and<br />
embedded into course learning and teaching activities. So far the adoption does not appear to be<br />
universal across universities (or even within individual universities). In order to elicit and address the<br />
issues affecting the adoption and embedding of learning and teaching technologies, a group of<br />
academics who have already embedded a variety of learning and teaching technologies into their<br />
teaching practice were invited to participate in a focus group. A summary of the findings of the focus<br />
group is presented in this report.<br />
2. Process<br />
Leading educational technology adopters as well as key learning and teaching representatives from<br />
business schools in one of the largest universities in Australia were invited to a focus group. The focus<br />
group was facilitated by an external consultant specilising in Ed-Tools. The aim of the focus group was to<br />
identify key strategic themes and issues regarding the use and adoption of Ed-Tools in learning and<br />
teaching. Examples of barriers to adoption of Ed-Tools, as reported in the literature, were noted by the<br />
focus group (see Table 3). Specifically, the goal of the focus group was to address the question: How can<br />
educational technologies be adopted & embedded into course learning and teaching activities?<br />
The external facilitator guided the group through a series of questions designed to discover:<br />
How teaching & learning technologies are currently being used at the university;<br />
How industry use of ICT relates to student experience;<br />
The barriers to embedding teaching & learning technologies into the curriculum; and<br />
How the university could support staff in their adoption of teaching & learning technologies into the<br />
curriculum.<br />
463
Table 3: Examples of barriers to adoption of Ed-Tools<br />
<strong>Academic</strong>s<br />
Students<br />
Theme Comment Reference<br />
Technology adoption ‘one-at-a time’ approach to adoption and<br />
integration.<br />
McGee & Diaz, 2007<br />
Lack of integration with<br />
CMS<br />
Learners' changing<br />
expectations<br />
Institutional commitment<br />
and Support<br />
Uneasiness with nature of<br />
e-tools<br />
Technical issues<br />
Time and effort<br />
Emerging e-tools not integrated into CMS<br />
Changing student cohorts and expectations of<br />
instructional strategies<br />
Review and change/upgrade of CMS.<br />
Investment in E-tools<br />
Some students uncomfortable with openness<br />
of e-tools<br />
Technical glitches and insufficient technical<br />
support with e-tools<br />
E-tools take time away from learning subject<br />
matter content<br />
McGee & Diaz, 2007;<br />
Dalsgaard, 2006<br />
McGee & Diaz, 2007<br />
Williams & An, 2010<br />
Williams & An, 2010<br />
Williams & An, 2010<br />
Data was collected and analyzed by the external facilitator. Only annonymized results were made<br />
available to the researchers.<br />
3. Outcomes<br />
Outcomes of the focus group are presented in this section. Suggestions for addressing the issues are<br />
outlined in the next section.<br />
3.1 Current use of teaching and learning technologies in the university<br />
The focus group participants reported various uses of teaching and learning technologies. The most<br />
commonly used technologies are:<br />
Various Blackboard functions (blogs, discussion forums, collaborative workspaces, online testing)<br />
Video objects embedded and/or streamed online<br />
Slide presentations embedded online (for example using Camtasia)<br />
Lectopia<br />
Other technologies that are being used to varying degrees include:<br />
eBooks - to which academics can contribute comments<br />
Commercial pay per use online quizzes<br />
Pebblepad - for recognition of prior learning (ePortfolios)<br />
Wikis – both internal and external<br />
Business simulation tools<br />
Web conferencing (via Elluminate)<br />
Delicious (bookmark sharing)<br />
Google Docs (collaborative development)<br />
Facebook (for semi-formal learning)<br />
Video conferencing between campuses and also with individual students through Skype<br />
KeePad response systems<br />
iPhone Apps<br />
The focus group reported that while some academics are embedding educational technologies into their<br />
courses, there is evidence to suggest that this is not a widespread practice. Such evidence can be found<br />
in the uptake of PebblePad and Lectopia both within the business schools and the wider university.<br />
464
3.2 Industry use of information and communication technology (ICT)<br />
As universities are to create work-ready professionals, it is important to compare the exposure to, and<br />
use of technologies in the university courses to the common use of technologies in industry. There is a<br />
large range of technologies being used in industry – some of them specific to the industry itself. Of the<br />
more common technologies, the following were considered in use across a range of industries (in no<br />
particular order):<br />
Wikis<br />
Search engines<br />
Standard desktop tools: spreadsheets, document processing etc<br />
Website development tools<br />
Communications software<br />
Presentation software<br />
Skype<br />
Microblogging<br />
Social networking (eg LinkedIn, Facebook)<br />
Instant messaging<br />
Simulations<br />
Videoconferencing<br />
Content management systems<br />
Filing systems<br />
Project management tools<br />
Blogs<br />
Email<br />
Phone - SMS, Smart phones<br />
Discussion boards<br />
In comparing these to the technologies used in the university, it was considered that some academics<br />
use some of these, but many academics don’t use any of these technologies at all.<br />
3.3 Barriers to embedding teaching and learning technologies into the curriculum<br />
From the previous deliberations it was considered that there are particular barriers to academics adopting<br />
and embedding technologies on a wider scale. The group listed and considered a large range of barriers<br />
that might exist. As a small group, it was not possible to consider solutions to all of these, so the issues<br />
were sorted, grouped and a decision was made that the following are most pressing barriers:<br />
1. The value placed on teaching and issues of career progression<br />
2. Time issues<br />
3. Logistics and large class sizes<br />
4. Workplan issues<br />
Suggestions in addressing these issues are presented in the next section.<br />
Other barriers that were listed and should be considered included:<br />
Technology is used as an ‘add on’, not embedded into the course<br />
Lack of appreciation for pedagogy/instructional design<br />
Very limited understanding of available technologies and how these can support and enhance<br />
learning and teaching<br />
Kudos is given for the adoption of a technology rather than good teaching practice, neither for the<br />
effective and sustainable use of the new technology<br />
465
Many academics do not attend learning and teaching forums/PD<br />
Over focus on Good Teaching Score (GTS)<br />
<strong>Academic</strong> preferences and cultural resistance - not interested in adoption of technology<br />
Startup costs<br />
Issues of technical support for both the adoption of a new technology and on-going support.<br />
Lack of internal budget to support innovative use of technology<br />
Poor user interfaces eg previous versions of Blackboard<br />
Fear of losing IP if it is put online<br />
Access to materials is limited to one semester<br />
Universities focus on raising their international rankings impacts on what activities are supported<br />
3.4 Suggestions<br />
The group considered the four most pressing barriers to adoption and provided the following suggestions<br />
as to what the university could do to support staff in their adoption of educational technology tools in the<br />
curriculum.<br />
3.4.1 Value of teaching /career progression/motivation<br />
It is believed that teaching is not valued or rewarded by many universities, particularly in Australia where<br />
the main academic focus seems to be only on “research”. The current funding model used by the<br />
Australian Federal Government and the increased focus on raising rankings impact on what activities are<br />
supported and do not provide incentive to engage in development of teaching. Research and publication<br />
receive more recognition and facilitate progress in one’s career. This is unlike the U.S. system where one<br />
can choose a teaching career path in academia. This is a de-motivator for academics to focus on good<br />
teaching practice and continuously improve their teaching skills.<br />
Australian universities could provide support in the following ways:<br />
Teaching should be supported as a career option. This would include recognition of leadership, the<br />
scholarship of teaching, and innovation in teaching practices. Benchmarks need to be developed for<br />
good teaching practice to be measured against. The university needs to enable more flexibility by<br />
providing alternative career paths for academics, for example, teaching or research.<br />
Recognition and reward should be for developing good learning experiences not for merely adopting<br />
a new technology. The university needs to reinforce the effective use of new technologies, not just<br />
the adoption of them. Managers need to acknowledge the benefits of developing teaching skills and<br />
the value this brings to student experiences and learning outcomes.<br />
Australian universities have a strong focus on Good Teaching Score (GTS) and if innovations do not<br />
contribute to improved GTS, then they are not encouraged. As GTS is often negatively impacted in<br />
the first trial of a new technology, this is a disincentive for academics to be innovative. Universities<br />
need to redefine what constitutes good GTS and alternative recognition schemes for teaching need<br />
to be investigated. Some schools at the university have internal schemes with additional recognition.<br />
These schemes should be investigated and, if appropriate, adopted more widely throughout the<br />
university.<br />
<strong>Academic</strong>s need to develop skills in how to use technology effectively as a teaching tool.<br />
Unfortunately, many learning and teaching forums are not well attended. The university needs to find<br />
a way to reward academics for attending and contributing to learning and teaching forums and to<br />
support and encourage the development of skills in the design of technology supported learning<br />
activities.<br />
Innovation sometimes fails, however there are always lessons to be learned that can improve<br />
teaching practice and adoption of technology in the future. Universities need to develop a process for<br />
building upon these foundations and acknowledging those academics who have been early adopters.<br />
Adopting new teaching practices and technologies incurs additional time and cost. There should be<br />
discretionary budget allocated to schools to provide resources for innovations. Some schools already<br />
466
allocate small budgets for the trials of innovations. Where this model has been implemented, it should<br />
be investigated with a view to adopting it more widely within the university.<br />
Currently, academics feel that there is no recognition and support for day-to-day innovation of<br />
teaching practice. Successes are largely legitimised when there is some form of external grant<br />
associated with the development. The university needs to develop a process and culture of<br />
supporting and recognising innovation internally, that is, amongst one’s peers.<br />
There appears to be a strong cultural resistance from academics to moving away from face-to-face<br />
teaching. Technology is often viewed as an ‘extra’ which means it is viewed as taking extra time and<br />
resources. This view needs to be addressed.<br />
There appears to be a lack of understanding around copyright and IP. On the one hand, academics<br />
fear losing their IP if it is digitised, and on the other, they are not encouraged to use content which is<br />
publicly available. Universities needs to ensure that academics and copyright advisors are fully aware<br />
of their digital rights and obligations.<br />
3.4.2 Time issues<br />
Time issues relate to both academics and students and also to time pressures associated with the<br />
amount of content to be covered in various curricula. <strong>Academic</strong>s need to be supported and encouraged<br />
to be creative and experiment with technologies, that is, time allocated and no repercussions for less than<br />
successful experimentation.<br />
In this regard, universities could provide support in the following ways:<br />
It takes time to become familiar, and stay up to date, with teaching and learning technologies. It<br />
would help academics if time was allocated for this.<br />
The work plan is critical to success, therefore it should include and reward the adoption and effective<br />
use of technologies.<br />
In the early adoption phase academics need technical support. It is essential that educational<br />
designer and media production support teams understand educational needs and provide support in<br />
a timely way.<br />
As it takes time to develop resources, there should be more encouragement/incentive for academics<br />
to share resources in order to leverage better practices.<br />
To some extent, technology is being seen (and used) as an add-on to teaching and learning which<br />
merely takes more of an academic’s scarce time. Universities need to support and encourage the use<br />
of technology as an integral part of the teaching and learning process that will save time.<br />
Interestingly, and independently, these findings reflect those of McGee & Diaz (2007) who argued that<br />
under ideal circumstances, a faculty member may require anywhere from three to four terms to adopt a<br />
learning technology tool and even more time to produce positive results in teaching and learning.<br />
3.4.3 Logistics and large class sizes<br />
In addition to the time issues outlined above, many of the very large classes in business schools are<br />
compulsory, making the audience difficult to manage as they don’t really want to be there.<br />
Internationalization brings its own set of challenges and in order to service overseas students well, the<br />
use of technology is essential.<br />
Support in the form of the following can help to address these issues:<br />
Provide more staff to manage the workload. The team could include: teaching assistants,<br />
tutors/sessional lecturers and educational designers.<br />
Educational designer and media production staff who are well versed in the use of technologies<br />
should be made available to teaching staff on an ‘as needs’ basis.<br />
There is a need to understand how to build an effective teaching team for very large subjects. The<br />
university should support and encourage this development.<br />
The inflexibility of IT policy is usually a deterrent to exploring new technologies. While it is understood<br />
that security is paramount, there needs to be some leeway for academics to experiment with a range<br />
of teaching and learning technologies.<br />
467
3.4.4 Work plan issues<br />
It was considered that work plans are key to many of the issues impacting upon the adoption and<br />
embedding of technology. There are differences across schools in terms of how innovations are<br />
incorporated in the work plan.<br />
The following are suggestions for improvement:<br />
The work plan could list the range of innovations which are used and this could be used to assess<br />
academics for promotion.<br />
Work plans could be attached to the resource structure and a direct course budget.<br />
The structure of work plans should be examined with a view to making them less linear.<br />
4. In conclusion<br />
Interactive eLearning tools and web 2.0 offer great potential to engage students in richer and more<br />
interactive learning experiences that are aligned with their learning preferences. Teaching approaches<br />
can adapt their instructional strategies to take advantage of these new developments to optimize the<br />
learning outcomes for both the students and the teaching staff. A shift from content delivery and recall to<br />
facilitated interactive learning experiences is easier using new technologies, but there is a resistance<br />
amongst many teaching staff to move in this direction. Whilst some early adopters have innovated and<br />
applied some new technologies to achieve great outcome (with evidence to show this generates better<br />
capability development), these cases are still in the minority.<br />
This study looks to understand the underlying barriers to embedding the new technologies into an<br />
integrated learning experience and find pragmatic ways to remove them. It is hoped that such an<br />
approach will enhance the teaching practices and learning experiences of both teachers and students<br />
and provide graduates with stronger capabilities to perform in the workplace.<br />
Web 2.0 technologies create an environment where a teacher becomes a facilitator of learning rather<br />
than a distributor of knowledge (Williams & An, 2010). <strong>Academic</strong>s need support and encouragement to<br />
assume this new role and to use technology effectively for teaching and learning. Technology is not a<br />
generic solution and in order to implement it effectively, universities need to provide an environment that<br />
encourages and supports innovation and good teaching practice. It also needs to demonstrate<br />
commitment to the use of technology by embedding them in practice, for example, academics should be<br />
using Ed-Tools themselves to build competency and show leadership around the value of these tools.<br />
Acknowledgement<br />
The authors wish to thank Mr Stuart Whitman, managing director of ScienceAdvantage for his valuable<br />
insight through this project and Ms Carol Skyring for focus group facilitation and data analysis. The<br />
authors wish to also thank <strong>Academic</strong> Development Group—College of Business, RMIT University for<br />
financing this work through a Research in Learning and Teaching (RiLT) funding award.<br />
References<br />
Ajjan H & Hartshorne R, 2008, “Investigating faculty decisions to adopt Web 2.0 technologies: Theory and empirical<br />
tests”, Internet and Higher Education, Vol. 11, pp. 71-80. (Accessed 9 th September, 2010)<br />
Alexander B, 2006, “Web 2.0: A new wave of innovation for teaching and learning?”, EDUCAUSE Review, Vol. 41,<br />
no. 2, pp. 32-44. (Accessed 9 th September, 2010)<br />
Anderson, P. (2007), “What is Web 2.0? Ideas, technologies and implications for education”, [online], JISC<br />
Technology & Standards Watch, Retrieved from www.jisc.ac.uk/media/documents/techwatch/tsw0701b.pdf.<br />
Barnett M, Keating T, Harwook W, & Saam J, 2004, “Using emerging technologies to help bridge the gap between<br />
university theory and classroom practice: Challenges and successes”, School Sciences & Mathematics, Vol.<br />
102, no.6, pp. 299−314. (Accessed 9 th September, 2010)<br />
Berkman Center for Internet & Society at Harvard Law School, 2010, Available at: http://cyber.law.harvard.edu/<br />
(Accessed 9 th September, 2010)<br />
Brown J S & Adler R P (2008), “Minds on fire: Open education, the long tail, and learning 2.0”, EDUCAUSE Review,<br />
Vol. 43, no. 1, pp. 17-32. (Accessed 9 th September, 2010)<br />
Dalsgaard C (2006), “Social software: eLearning beyond learning management systems”, <strong>European</strong> Journal of Open,<br />
Distance and ELearning, No. 2006, available at:<br />
www.eurodl.org/materials/contrib/2006/Christian_Dalsgaard.htm (Accessed 9 th September, 2010)<br />
Daniels P (2009), “Course Management Systems and Implications for Practice”, International Journal of Emerging<br />
Technologies & Society, Vol. 7, no. 2, pp. 97-108. (Accessed 9 th September, 2010)<br />
468
Hossein Zadeh et al.<br />
Dearstyne B W (2007), “Blogs, Mashups, and Wikis: Oh my!”, Information Management Journal, Vol. 41, no.4, pp.<br />
24−33. (Accessed 9 th September, 2010)<br />
Ding, Y., Jacob, E. K., Caverlee, J., Fried, M., & Zhang, Z. (2009), “Profiling social networks: A social tagging<br />
perspective”, [online], D-Lib Magazine, 15. Retrieved from http://www.dlib.org/dlib/march09/ding/03ding.html.<br />
Hartshorne R & Ajjan H (2009), “Examining student decisions to adopt Web 2.0 technologies: theory and empirical<br />
tests”, Journal of Computing in Higher Education, Vol. 21, no. 2. (Accessed 9 th September, 2010)<br />
Hew K F, & Sim J W S (2010), “The use of weblogs in higher education settings: A review of empirical research”,<br />
Educational Research Review, Vol. 5, pp. 151-166. (Accessed 9 th September, 2010)<br />
Maloney, E. J. (2007), “What Web 2.0 can teach us about learning”, Chronicle of Higher Education, 53 (18) B26-B27.<br />
McGee P & Diaz V (2007), “Wikis and Podcasts and Blogs! Oh, My! What Is a Faculty Member Supposed to Do?”,<br />
EDUCAUSE Review, Vol. 42, no. 5, pp. 28-41. (Accessed 9 th September, 2010)<br />
Minocha S (2009), “Role of social software tools in education: a literature review”, Education + Training, Vol. 51, no.<br />
5/6, pp. 353-369. (Accessed 9 th September, 2010)<br />
Penntags project at the University of Pennsylvania (2008), Available at: http://tags.library.upenn.edu/<br />
Proserpio, L and Gioia, DA (2007), “Teaching the virtual generation”, Academy of Management Learning and<br />
Education, Vol. 6, No. 1, pp 69-80<br />
Richardson W (2009), Blogs, wikis, podcasts, and other powerful web tools for classrooms, 2nd Ed, Thousand Oaks,<br />
CA: Corwin Press.<br />
Williams K & An Y (2010), “Teaching with Web 2.0 Technologies: Benefits, Barriers and Lessons Learned”,<br />
International Journal of Instructional Technology and Distance Learning, Vol. 7, no. 3, pp. 41-48. (Accessed 9 th<br />
September, 2010).<br />
469
470
PhD<br />
Research<br />
Papers<br />
471
472
Evaluating the Performance of ERP Systems in Saudi<br />
Arabian Higher Education: A Stakeholders’ Perspective<br />
Mona Althonayan and Anastasia Papazafeiropoulou<br />
Brunel University, UK<br />
mona.althonayan@brunel.ac.uk<br />
Anastasia.Papazafeiropoulou@brunel.ac.uk<br />
Abstract: Enterprise resource planning (ERP) systems are complex and comprehensive software designed to<br />
integrate business processes and functions. Despite the difficulties and risk adopting ERP systems is expanding<br />
rapidly. Furthermore, universities make large investments in information systems (IS) expecting positive impacts<br />
(return on investment). Moreover, universities are facing serious challenges implementing a new technology. Meeting<br />
stakeholders’ expectation in higher education is one of those challenges; this refers to the universities as unique<br />
organizations. Since the effectiveness of post implementation of ERP systems has become an essential indicator of<br />
success; effective selection, development and improvement of information systems requires a systematic evaluation<br />
tool. Although there are a variety of IS success evaluation studies, there is no consensus on the appropriate manner<br />
of evaluation of IS success to help organisations return on investments in information systems. Previous studies<br />
focused on information systems and user performance, but they highlight the need for more focus on ERP systems<br />
and stakeholders’ performance. Despite the importance for IS evaluation, there is a lack of accepted frameworks on<br />
IS evaluation in general and ERPs in specific in higher education. This paper focuses on three information system<br />
models: Delone and McLean’s IS success, Task Technology Fit (TTF) by Goodhue and End User Computing<br />
Satisfaction (EUCS) by Doll and Torkzadeh. To avoid the weakness in these three models, this paper attempts to<br />
integrate the three models to produce a new construct which has a more comprehensive view of the most important<br />
factors pertinent to evaluation of stakeholders’ performance on ERP systems in higher education, combining the<br />
impact and the quality of the system. This framework will be applied at a later study the stakeholders’ performance in<br />
Saudi Arabian higher education.<br />
Keywords: ERP systems; performance; stakeholders; evaluation; higher education, individuals, framework<br />
1. Introduction<br />
Organizations spend millions of dollars on information systems to improve organizational or individual<br />
performance. According to Petter et al., (2008) information systems (IS) are developed using information<br />
technology (IT) to help individual performance. Practitioners and researchers seek to understand and<br />
measure the success of these investments (Goodhue, 1995). Despite the importance of IS success,<br />
Sedera et al., (2003) indicate the success of large IS, particularly enterprise resource planning (ERP)<br />
systems is arguably difficult to measure, since ERPs generate substantial and intangible benefits and the<br />
systems involve many users (stakeholders) ranging from top executives to data entry operators, who can<br />
define its success differently. In addition, there are many applications that span the organization, with a<br />
diversity of capabilities and functionality. Thus, measuring ERP systems’ success is a complex<br />
endeavour.<br />
ERP has played a significant role both in the recent history of information technology management in<br />
higher education, and the history of higher education itself. It is important to define the concept of ERP<br />
systems in higher education: “multiple in scope, tracking a range of activities that include human<br />
resources systems, student information systems and financial systems” (Robert et al., 2004).<br />
It is clear that there are many similarities between implementing ERP system software in the education<br />
sector and other organisations/sectors. This uniqueness is based on different combinations of certain<br />
characteristics, which were determined by Okunoye and Folick (2006) as: “complexity of purpose, limited<br />
measurability outputs, both autonomy and dependency from wider society, diffuse structure of authority,<br />
and internal fragmentation”.<br />
In higher education institutions, the stakeholders are considered a fundamental factor that distinguishes<br />
education institutes from other organizations. The major proponents of ERP systems, such as Swartz and<br />
Orgill (2000) argue that there are many encouraging reasons to implement ERP systems, for example, to<br />
improve information access and the effectiveness of workflow. Moreover, other reasons to consider ERP<br />
are capability to improve controls and the ability of stakeholders to use ERP easily. In contrast, Bradley<br />
and Lee (2007) identified that universities have similar problems to other organizations, such as<br />
coordinating resources, controlling cost and motivating and facilitating ERP amongst the faculty and the<br />
staff; therefore, IT evaluation is important, especially with the large budget universities spent on IT/ERP<br />
473
Mona Althonayan and Anastasia Papazafeiropoulou<br />
systems project. ERP systems literature has considered manufacturing industries, but few studies<br />
discuss ERP in academic institutes. Despite the rapid current growth of the ERP in higher education<br />
sector, there is lack of scholarly publication discussing ERP implementation in higher education (Rabaa`i<br />
et al., 2009).<br />
Since the value from IS evaluation and the impacts of ERP system on both the organization and<br />
individuals, is to justify the value of productivity, quality, and competitiveness of the organization, the aim<br />
of this paper is to integrate the three models D& M IS success, Task Technology Fit (TTF) Goodhue and<br />
End User Computing Satisfaction (EUCS) Doll and Torkzadeh, to produce a new construct which has a<br />
more comprehensive view of the most important factors by which to evaluate stakeholders’ performance<br />
on ERP systems in higher education.<br />
First evaluation of stakeholders’ performance is discusses on and the three selected models the D&M,<br />
TTF and EUCS are introduced. The Proposed framework including the final factors chosen from the three<br />
models to evaluate stakeholders’ performance on ERP systems in higher education is then discussed.<br />
2. Evaluation of stakeholders’ performance<br />
The following subsections discusses three IS models to evaluate ERP system stakeholders’<br />
performance.<br />
2.1 Task-Technology Fit (TTF)<br />
Task Technology Fit (TTF) is defined by Goodhue (1995) as “the extent that technology functionality<br />
matches task requirements and individual abilities”.<br />
On the other hand, Goodhue and Thompson (1995) identified TTF as “the degree to which a technology<br />
assists an individual in performing his or her portfolio of tasks”. The concept was derived originally from<br />
work adjustment theory (Dishaw and Strong, 1999).<br />
Chang (2008) explained that the TTF model considers the degree to which the capabilities of the<br />
technology match the demand of the task, and has four main constructs: task characteristics, technology<br />
characteristics, which together affect the third construct of TTF which in turn affects the outcome variable,<br />
either utilization or performance (Dishaw et al., 2002). In addition, Goodhue et al., (2000) stated that TTF<br />
presumes that the performance impacts upon the fit between three constructs: technology characteristics,<br />
task requirements, and individual abilities. The TTF model posits that IT will be used if the functions<br />
available to the user support (fit) the activities of the user.<br />
Task<br />
characteristics<br />
Technology<br />
characteristics<br />
Individual<br />
characteristics<br />
Task<br />
Technology Fit<br />
Performance<br />
Impacts<br />
Figure 1: The model of Task Theology Fit (Goodhue, 1995)<br />
Since the measurement of system success is difficult, many management information systems (MIS)<br />
researchers depend on user evaluation of systems as a surrogate for MIS success, that means the<br />
assessment made by a user (Goodhue, 1995).<br />
The TTF instrument is conceptually based on the Task Technology Fit theory in which the<br />
correspondence between information systems functionality and task requirements leads to positive user<br />
evaluations and positive performance impact (Goodhue, 1998).<br />
474
Mona Althonayan and Anastasia Papazafeiropoulou<br />
Figure 2: TTF model and user evaluation (Goodhue, 1995)<br />
Goodhue and Thompson (1995) argued that greater use leads to better performance only when there is<br />
task-technology fit. Goodhue and Thompson designed a new model to demonstrate the link between the<br />
information technology and individual performance. This connection draws on insights from two<br />
complementary researches (user attitudes as predictors of utilization and task technology fit as predictor<br />
of performance). This new model is called the Technology-to-Performance Chain (TPC). This new model<br />
emphasizes that for the information technology to have a positive impact on individual performance; the<br />
technology must be utilized and must be a good fit with the tasks it supports.<br />
Although Kositanurit et al., (2006) assert that the TTF model does not answer the question of what<br />
characteristics of the IS lead to the highest levels of user performance, it suggests some constructs that<br />
are relevant to investigation. Researchers have attempted to integrate the TTF model with other models.<br />
For instance, Dishaw and Strong (1999) conducted a study to integrate between the technology<br />
acceptance model (TAM) and TTF. They aimed to provide a theoretical basis for exploring the factors<br />
that explain software utilization and its link with user performance. Dishaw and Strong (2002) extended<br />
their work by investigating the relationship between computer self-efficacy and the combined TAM\TTF<br />
model. Continuing their effort to combine models, Dishaw and Strong (2004) integrated two models,<br />
namely, TTF and the unified theory of acceptance and use of technology (UTAUT).<br />
Gros et al., (2005) supported the combination between TTF and other models by saying it is important to<br />
consider how systems help or impede the user, but it is not sufficient alone to explain system’s success<br />
or satisfaction with it.<br />
2.2 Information Systems Success model<br />
Delone and McLean’s (1992) IS success model is widely cited, and has been a valuable contribution to<br />
the literature on IS success measurement, because it was the first study that tried to impose some order<br />
to develop a comprehensive IS model and instrument for a particular context (Gable et al., 2008).<br />
Delone and McLean (1992) conducted a large number of studies found in the academic literature<br />
covering the period of 1981-1987 attempting to identify those factors that contribute to information<br />
systems success. Based on these studies, they defined six major dimensions or categories of IS<br />
success: systems quality, information quality, use, user satisfaction, individual impact and organisational<br />
impact. The main contributions Delone and McLean make to our understanding of IS success evaluation<br />
are: (1) a method for classifying the large number of IS success measures that have been used in the<br />
literature into six categories; (2) their approach begins to identify relevant stakeholder groups in the<br />
process of evaluation; (3) they suggest a model of “temporal and causal” interdependencies between<br />
these categories (Seddon and Kiew, 1994; Myers et al., 1997; Seddon, 1997).<br />
System quality<br />
Information<br />
quality<br />
Task<br />
Technology<br />
Individual<br />
Use<br />
User<br />
Satisfaction<br />
Figure 3: D&M (1992) I /S success model<br />
User evaluations<br />
(surrogate for TTF)<br />
Task Technology Fit<br />
(TTF)<br />
475<br />
Individual<br />
impact<br />
Performance<br />
Organizational<br />
impact
Mona Althonayan and Anastasia Papazafeiropoulou<br />
Delone and McLean (1992) proposed an IS success model, but did not empirically tested it (Sabherwal et<br />
al., 2006). However, many studies have attempted to test and modify Delone and McLean’s model and<br />
called for further development and validation of their model. For instance, Seddon and Kiew (1994) were<br />
the first to empirically test the causal/process nature of the model by testing part of Delone and McLean’s<br />
model, and Seddon (1997) re-specified and extended the Delone and Mclean model and presented an<br />
alternative model of IS success. Rai et al., (2002) and Sabherwal et al., (2006) further built on Delone<br />
and McLean’s (1992) model.<br />
Seddon (1997) argued that Delone and McLean attempted to do too much by combining both the process<br />
and causal explanation of IS success in their model, and that as a result the model was confused and<br />
miss specified. The importance of Seddon’s (1997) study is that he adopted a theoretical approach to<br />
modify Delone and McLean’s model, and he distinguished between actual impact and expected impacts,<br />
and incorporated the additional construct of perceived usefulness (Sabherwal et al., 2006).<br />
Seddon (1997) believed that successful systems will provide benefits such as helping the user do more<br />
or better work in the same time, or take less time to achieve as much work of the same quality as was<br />
done in the past.<br />
The focus in this paper is on individual impact, which was defined by Delone and McLean (1992) as “the<br />
effect of information on the behaviour of the recipient of all the measure of IS success”. Gable et al.,<br />
(2008) identified individual impact as the “measure of the extent to which the (IS) has influenced the<br />
capabilities and effectiveness, on behalf of the organization, of key users”. The reason for choosing the<br />
individual impact is practicality, because impact is closely related to performance, meaning, according to<br />
Delone and McLean (1992) that impact could be an indication that the information system has given the<br />
user a better understanding of the decision context, improved his or her decision-making and productivity<br />
and has brought about a change in user activity or changed the decision-maker’s perception of the<br />
importance or usefulness of the IS. The usefulness is the degree to which a person believes that using<br />
particular systems would enhance his or her job performance (Seddon, 1997).<br />
Pitt et al., (1995) augmented Delone and McLean’s model to include service quality as a measure of IS<br />
success. They argued that Delone and McLean’s model needs to be expanded to reflect the IS<br />
department’s service role. Moreover, the basis of Delone and Mclean’s categorization theory is<br />
communication; thus, the IS department is not just a provider of products, it is also a service provider.<br />
According to Petter et al., (2008) many researchers have suggested that service quality is an important<br />
factor to be added to the Delone and McLean’s IS success model, because it is salient to IS success. In<br />
addition, there is a danger of miss -measuring IS effectiveness if researchers do not include an<br />
assessment of service quality.<br />
Petter et al. (2008) defined service quality as “the quality of the support that systems users receive from<br />
the IS department and IT support personnel”. Moreover, it measures the service quality of IT departments<br />
as opposed to individual IT applications, by measuring and comparing user expectation and their<br />
perceptions of the IT department (Petter et al., 2008). There is support for this argument in the IS<br />
literature. For example, Conrath and Mignen (1990) reported that the second most important component<br />
of user satisfaction, after general quality of service, is the match between user’s expectations and actual<br />
IS service. Rushinek and Rushinek (1986) added that fulfilled user expectations have a strong effect on<br />
overall satisfaction.<br />
Pitt et al. (1995) propose that service quality can be assessed by measuring customer expectations and<br />
perceptions of performance level for a range of service attributes. Then the difference between<br />
expectation and perceptions of actual performance can be calculated and averaged across attributes.<br />
A decade later, Delone and McLean (2003) reviewed and evaluated this argument before updating the<br />
information systems success model, based on a review of more than100 articles of the empirical<br />
conceptual literature on IS success that was published during the same period. Petter et al., (2008)<br />
explained the utility of the update to the Delone and McLean IS model and evaluated its usefulness in<br />
light of dramatic changes in IS practice, especially the explosive development of e-commerce. The<br />
update study was conducted on six dimensions: systems quality, information quality, service quality, use,<br />
user satisfaction and net benefits.<br />
476
Information<br />
quality<br />
System quality<br />
Service quality<br />
Mona Althonayan and Anastasia Papazafeiropoulou<br />
Intention to use Use<br />
User satisfaction<br />
Figure 4: Updated D&M IS success model (2003)<br />
This research focuses on service quality, which is considered an important dimension in IS measurement<br />
success, because suggests that there is a correlation between stakeholders’ expectation of the service<br />
quality and their performance level. Therefore, considering the service quality dimension for measuring<br />
stakeholders’ performance of ERP systems in higher education is really essential. Table 1; illustrates<br />
Delone and McLean model the service quality factors proposed in Pitt et al. (1995) and confirmed in,<br />
D&M update (2003).<br />
Table 1: Service quality factors, D&M information system update model (2003)<br />
Service quality<br />
-Reliability<br />
-Assurance<br />
-Empathy<br />
-Tangibles<br />
-Responsiveness<br />
Sedera and Gable (2004) attempted to build upon their previous work (2003) to derive a standardized<br />
instrument for measuring enterprise systems success, based on the research cycle developed by<br />
Mackenzie and House (1979) and McGrath (1979) that which entails two main phases: (1) exploratory<br />
phase and (2) confirmatory phase. Results from confirmatory factor analysis utilizing structural equation<br />
modelling techniques confirmed the existence of four distinct and individually important dimensions of<br />
ERP systems: individual impact, organisational impact, system quality and information quality individual<br />
impact.<br />
System<br />
quality<br />
- Ease of use<br />
- Ease of<br />
learning<br />
- User<br />
requirements<br />
- System<br />
features<br />
- System<br />
accuracy<br />
- Flexibility<br />
- Sophistication<br />
- Integration<br />
- Customization<br />
Enterprise<br />
systems<br />
success<br />
Figure 5: Sedera and Gable (2004) Validated Measures of ERP<br />
Individual<br />
Information impact<br />
quality - Learning<br />
-Availability<br />
- Usability<br />
- Understand<br />
ability<br />
- Relevance<br />
- Format<br />
- Conciseness<br />
477<br />
- Awareness/<br />
Recall<br />
- Decision<br />
effectiveness<br />
- Individual<br />
productivity<br />
Net Benefits<br />
Organization<br />
al impact<br />
- Organizational<br />
cost.<br />
- Staff<br />
requirements.<br />
- Cost reduction.<br />
- Overall<br />
productivity<br />
- Improved<br />
outcomes/outputs<br />
- Increased<br />
capacity<br />
- E-government
Mona Althonayan and Anastasia Papazafeiropoulou<br />
Petter (2008) believed that Sedera and Gable’s (2004) modification model, particularly the instrument for<br />
evaluating IS success, is unique. His point of view is based on two different reasons: (1) Sedera and<br />
Gable’s model captures the multidimensional and complex nature of IS success by measuring four<br />
dimensions; (2) another strength in this model is that the instrument was tested within the context of ERP<br />
systems to ensure its validity. In contrast, Darmawan (2001) argued that combining two or more levels<br />
into a single level analysis causes the aggregation or disaggregation of data collected at the lower<br />
(individual) to the higher (organisation) level, and could introduce bias, meaning an over- or underestimation<br />
of the significance of effects associated with variables that are aggregated or disaggregated.<br />
The reason for selecting the individual impact dimension from Delone and McLean’s IS success model<br />
(1992) and service quality impact from the update model is because both dimensions can help this<br />
research, to discover whether or not ERPs improve the stakeholders’ performance and how it meets their<br />
requirements.<br />
The reason for selecting the individual impact dimension from Delone and McLean’s IS success model<br />
(1992) and service quality impact from the update model is because both dimensions can help this<br />
research, to discover whether or not ERPs improve the stakeholders’ performance and how it meets their<br />
requirements.<br />
2.3 End user computing satisfaction<br />
The end user computing satisfaction model designed by Doll and Torkzadeh (1988) is a potentially<br />
measurable surrogate for utility in decision-making, while users interact directly with the application<br />
software to enter information or prepare output reports. The end user application’s utility in decisionmaking<br />
is enhanced when the output meets the user’s requirements (Doll and Torkzadeh, 1988).<br />
End user computing satisfaction is conceptualized as the effective attitude towards a specific computer<br />
application by someone who interacts with the application directly (Doll and Torkzadeh, 1988).<br />
The EUCS instrument consists of five factors: content, format, timeliness, ease of use/efficiency and<br />
accuracy. Although this model is focused on end user computing satisfaction, it includes factors which<br />
could be useful for IS/ERP systems measurement.<br />
EUCS was designed for a conventional computing environment. Therefore, performance was excluded,<br />
since as Doll and Torkzadeh explain, performance-related behaviours may be application specific, and<br />
development of generalizable measure of EUCS success is difficult (Amoli and Farhoomand, 1996).<br />
Content<br />
Accuracy<br />
End- User<br />
Computing<br />
Satisfaction<br />
Format<br />
Timeliness<br />
Figure 6: End user computing satisfaction model (Doll and Torkzadeh, 1988)<br />
In the ERP systems domain Somers et al., (2003), adopted the EUCS model to measure end user<br />
satisfaction in ERP systems, their study examined further the theoretical meaning, structure,<br />
dimensionality, reliability and validity of EUCS when adopted to measure EUCS with ERP software<br />
applications. Their study confirmed that the EUCS instrument can be better understood and applied as a<br />
standardized measure of advanced information technology, for instance in ERP systems application.<br />
478<br />
Ease of<br />
use
Mona Althonayan and Anastasia Papazafeiropoulou<br />
<strong>Academic</strong>s and practitioners have continually sought a reliable and valid measure of information systems<br />
success. Ideally, success measures were focused on user behaviour or measuring decision outcome,<br />
other than what users consider value in a system which is linked to how it helps them to achieve their<br />
goal (Torkzadeh et al., 2005).<br />
The selection factors which are illustrated in Table 2 are derived from the D&M model (individual impact<br />
factors), TTF model and EUCS model, and are based on their relevance to the performance and system<br />
quality. (*) Refers to the factors most related to performance, (Q) Refers to the factors most related to the<br />
quality of the system, while (R) refers to repeated factors in TTF, EUCS and D&M models.<br />
Table 2: The selection factors from the three models<br />
Information systems success factors Task technology Fit End User<br />
Computing<br />
Satisfaction<br />
-Time taken to complete a task *<br />
-Improved personal productivity *<br />
-Time efficiency of task accomplishment *<br />
-Interpretation accuracy *<br />
-Computer awareness *<br />
-Confident on performance*<br />
-User confidence*<br />
-Task performance*<br />
-Confidence in decision*<br />
-Ability to identify solutions*<br />
- Ability to identify strategic opportunities or problems*<br />
-Ability to forecast firm performance*<br />
- Time to solve problem*<br />
- Accuracy of problem solution*<br />
- Efficiency of effort *<br />
- Time savings*<br />
-Personal effectiveness*<br />
- Problem identification*<br />
-Immediate recall of information*<br />
-Delayed recall of information*<br />
3. Proposed framework<br />
-Lack of<br />
confusion(Q)<br />
-Level of detail<br />
-Locatability<br />
-Meaning<br />
-Right data(Q)<br />
-Accessibility (Q)<br />
-Assistance(Q)<br />
-Authorization(Q)<br />
-Ease of use (Q)<br />
-Flexibility (Q)<br />
-System<br />
reliability(R)<br />
-Training (Q)<br />
-Accuracy (Q)<br />
-Compatibility (Q)<br />
-Currency (Q)<br />
-Presentation<br />
-Content(Q)<br />
-Format (Q)<br />
-Timeliness(Q)<br />
-Ease of use(R)<br />
-Accuracy(R)<br />
The factors in Table 2 derived from the TTF and EUCS models, when combined together can help to<br />
evaluate the performance from the technical aspects. Moreover, the selected factors are the most<br />
suitable in the ERP systems environment, and aim to measure how the ERP systems enhance individual<br />
performance. This study excludes some of the factors in the TTF model. Goodhue (1998) asserted that<br />
presentation and level of detail have similar meaning in EUCS model, which are content and format. In<br />
contrast, locatability and meaning will not help to evaluate ERP systems from the stakeholders’<br />
performance together with the individual impact because Goodhue (1998) believed that TTF measures<br />
are intended to evaluate all systems and service of IS department, whilst the EUCS focuses on individual<br />
application.<br />
Furthermore, the derivation of the factors in Table 3‘Time taken to complete task, improve stakeholders<br />
productivity, immediate recall of information, stakeholders confidence and performance and ability to<br />
identify problem and solutions’ was initially based on a comprehensive study conducted by Delone and<br />
McLean (1992) under the dimension of individual impact. All the factors were interpreted by the<br />
researcher because they all give similar meanings. Computer awareness is an important factor to<br />
measure the stakeholders’ knowledge about the system in general and ERP systems in particular. When<br />
the individual has appropriate knowledge and experience required to use the system, better performance<br />
should result. It is important to add service quality from Delone and McLean’s updated information<br />
system success model (2003) to the factors to measure the quality of support that ERP systems<br />
stakeholders received from the IT department and its impact on the individual performance.<br />
479
Mona Althonayan and Anastasia Papazafeiropoulou<br />
Table 3: The author has gathered the factors with the same meaning into one factor<br />
Information Systems Success Factors Factors<br />
- Time taken to complete a task<br />
Time taken to complete task<br />
-Time efficiency of task accomplishment<br />
-Task performance<br />
-Improved personal productivity<br />
Improve stakeholders’ productivity<br />
- Efficiency of effort<br />
-Personal effectiveness<br />
-Effectiveness in supporting decision<br />
-Immediate recall of information<br />
Immediate recall of information<br />
-Delayed recall of information<br />
-Confident on performance<br />
Stakeholders confidence on performance<br />
-User confidence<br />
-Confidence in decision<br />
- Ability to identify strategic opportunities or problems<br />
-Ability to identify solutions<br />
- Accuracy of problem solution<br />
- Time to solve problem<br />
- Time savings<br />
Ability to identify problem and solutions<br />
-Computer awareness Computer awareness<br />
None of the three models, TTF, EUCS and D&M, discussed previously, when applied separately do not<br />
provide effective evaluation of stakeholders’ performance, since TTF and EUCS evaluate the technical<br />
aspects of the systems, and the individual impact in the D&M model focuses on the human / social<br />
aspects. However, when the three models are integrated, this will effectively evaluate the stakeholders’<br />
performance.<br />
Gable et al., (2008) argued that a holistic measure for evaluating an IS should consist of dimensions that<br />
together look both backward (impact), representing the net benefits, and forward (quality), representing<br />
the best surrogate measure of probable future impact. The combination of impact and quality represents<br />
a complete measure of the information system.<br />
Impact<br />
(impacts to<br />
date)<br />
IS impact<br />
Quality<br />
(impacts<br />
anticipated)<br />
Figure 7: Gable et al (2008) the conceptual model<br />
Therefore, to overcome the shortcomings of previous models, this research aims to integrate all the three<br />
models to create a new synthesized model which has a more comprehensive view of the most important<br />
factors that affect stakeholder performance it does so by adopting the conceptual model developed by<br />
Gable et al., (2008) which combines the impact and the quality, and selecting the appropriate factors.<br />
The outcome is the factors the study gathered from Delone and Mclean’s (1992) IS success model; it<br />
considers (impact) as a half measure. The factors gathered from the TTF model and EUCS consider<br />
(quality) as a half measure, which will be used to evaluate stakeholders’ performance.<br />
Since individual performance is an essential indicator of organizational performance, consequently,<br />
studying the impact of ERP systems on stakeholders’ performance is a significant way to assess the<br />
utility of this software on higher education and how it contributes to performance efficiency and<br />
effectiveness, as shown in Table 5.<br />
480
Mona Althonayan and Anastasia Papazafeiropoulou<br />
ERPS impact<br />
Figure 8: ERPs impact adopted from Gable et al (2008) the conceptual model<br />
Table 5: The final chosen factors from three models<br />
Information Systems Success Task-Technology Fit (TTF) End User Computing<br />
Satisfaction<br />
-Time taken to complete task<br />
-Improve stakeholders’ productivity<br />
-Immediate recall of information<br />
-Stakeholders’ confidence and performance<br />
- Ability to identify problem and solutions<br />
- Computer awareness<br />
- Service quality<br />
(Reliability Assurance-Empathy-Responsiveness-<br />
Tangible)<br />
4. Conclusion<br />
-Lack of confusion<br />
-Right data<br />
-Accessibility<br />
-Assistance<br />
-Authorization<br />
-Ease of use<br />
-Flexibility<br />
-Training<br />
-Accuracy<br />
-Compatibility<br />
-Currency<br />
-Content<br />
-Format<br />
-Timeliness<br />
Several widely used instruments have been tested for user satisfaction in IS in general and ERP systems<br />
in particular, although few studies have discussed ERP systems from stakeholders’ point of view. The<br />
literature review highlights the importance of ERP from technical and organisation levels, while social and<br />
individual levels and their aspects have been ignored. Since stakeholders are the central elements<br />
creating value through their interaction with ERP system, there is a need for studying and evaluating ERP<br />
systems by focusing on how the human factors influence success and how ERP system can improve<br />
stakeholders’ performance.<br />
The aim of this study was to develop integrated framework derived from the three widely used models:<br />
Delone and McLean’s IS success, Task Technology Fit (TTF) and End User Computing Satisfaction<br />
(EUCS), which measure different aspects of factors that have impact on individual performance on ERP<br />
system environment. This study has proposed a theoretical framework that aims to evaluate the impact of<br />
ERP systems on stakeholders’ performance.<br />
The selection of the most appropriate and related factors among these models in this study will help<br />
researchers and practitioners to evaluate the stakeholders’ performance in ERP systems by focusing on<br />
the Saudi Arabian higher education.<br />
References<br />
D&M<br />
Information systems<br />
success<br />
Task-Technology Fit (TTF)<br />
End User Computing<br />
Satisfaction (EUCS)<br />
Amoli, J. E., Farhoomand, A. (1996) “A structural model of end user computing satisfaction and user performance’’,<br />
Information &Management, 30. Pp. 65-73<br />
Bradley, J., Lee, C. (2007) “ERP Training and User Satisfaction: A case Study”, Journal of Enterprise information<br />
systems, 3 (4). Pp. 33-50<br />
481
Mona Althonayan and Anastasia Papazafeiropoulou<br />
Chang, H. H. (2008) “Intelligent agent’s technology characteristics applied to online auction task: A combined model<br />
of TTF and TAM”, Technovation, 28 Pp.564–577<br />
Darmawan, I.G.N. (2001) “Adoption and implementation of information technology in Bali’s government: A<br />
comparison between single level path analyses using PLSPATH 3.01 and Amos4 and multilevel path analyses<br />
using MPLUS2.01”, International Education Journal, 2 (4) Pp. 100-125<br />
Delone, W. H., Mclean, E. R. (1992) “Information Systems Success: The Quest for the dependent variable”,<br />
Information Systems Research, 3. (1) Pp.60-95<br />
Delone, W. H., Mclean, E. R. (2003) “The Delone and McLean model of information systems success: A ten- year<br />
update”, Journal of Management Information Systems, 19 (4) Pp. 9-30<br />
Dishaw, M.T., Strong, D. M. (1999) “Extending the technology acceptance model with task – technology fit<br />
constructs”, Information &Management, 36. Pp.9-21<br />
Dishaw, M.T. Dishaw, M.T.Strong,D., Bandy,B.D., (2002) “Extending the Task Technology Fit model with self-<br />
efficacy constructs”, Eighth American <strong>Conference</strong> of Information Systems, Pp. 1021- 1027<br />
Dishaw, M.T.Strong,D., Bandy,B.D., (2004) “The impact of task technology fit in technology acceptance and<br />
utilization model”, The Tenth American <strong>Conference</strong> on Information Systems, New York: Pp. 3306-3311<br />
Doll, W., Torkzadeh, G. (1988) “The measurement of end user computing satisfaction”, MIS Quarterly, 12. (2) Pp.<br />
250-273<br />
Gable, G., Sedera,D., Chan,T., (2008) “Re- conceptualizing Information System Success: The IS – Impact<br />
Measurement Model”, Journal of the Association for Information Systems, 9. (7) Pp. 376-408<br />
Goodhue, D. (1995) “Understanding user evaluation of information systems”, Management science, 41(12) Pp. 1827-<br />
1995<br />
Goodhue, D. Thompson, R. (1995) “Task – technology Fit and individual Performance”, MIS Quarterly, 19(2) Pp.<br />
213-236<br />
Goodhue, D. (1998) “Development and Measurement Validity of a Task Technology Fit Instrument for User<br />
Evaluations of Information Systems”, Decision Sciences, 29 (1) Pp. 105- 138<br />
Goodhue, D., Klein,B.D., March,S.T., (2000) “User evaluation of IS surrogates for objective performance”,<br />
Information and Management, 38. Pp.87-101<br />
Gros, C.D., Mueller,H., Christian,L., (2005) “Evaluating user interaction with clinical information systems: A model<br />
based on human computer interaction models”, Journal of Biomedical informatics, 38. Pp.244-255<br />
King, P., (2002) “The promise and performance of enterprise systems for higher education”, Research Study from the<br />
EDUCUSE Center for Applied Research.: Pp. 1-7<br />
Kositanurit, B. Ngwenyama,O., Osei-Bryson,KM.,(2006) “An exploration of factors that impact individual performance<br />
in an ERP environment: an analysis using multiple analytical techniques”, <strong>European</strong> Journal of Information<br />
Systems, 15 .Pp. 556-568<br />
Myers, B. (1997) “A comprehensive model for assessing the quality and productivity of the information systems<br />
function: Toward a theory for information systems assessment”, Information Resources Management Journal,<br />
10. 1. Pp.6-25<br />
Okunoye, A., Folick, M. (2006) “ERP Implementation in higher education: An Account of Pre- Implementation and<br />
implementation Phases”, Journal of Cases on Information Technology, 8 (2) Pp.110- 132<br />
Petter, S., Delone,W., McLean,E., (2008) “Measuring information systems success: model, dimension, measures,<br />
and interrelationships”, <strong>European</strong> Journal of Information Systems, 17.Pp.236-263<br />
Pitt, L F., Watson, R., Kavan, C.B., (1995) “Service Quality: A measure of information systems effectiveness”, MIS<br />
Quarterly, 19.2. Pp.173-187<br />
Rabaa`i, A., Bandara, W., Gable, G., (2009) “ERP system in the higher education sector: A descriptive case study”,<br />
20 TH Australian <strong>Conference</strong> on Information Systems: Melbourne, Pp. 456-470<br />
Sabherwal, R., Jeyaraj, A., Chowa, C., (2006) “Information Systems: individual and organizational determinants”,<br />
Management science, 52. (12) Pp. 1849-1882<br />
Seddon, P. (1997) “A respecification and extension of the Delone and Mclean of IS success”, Information Systems<br />
Research, 8 (3) Pp. 240-253<br />
Seddon, P., Staples,S., Patnayakuni,R., Bowtell,M., (1999) “Dimension of Information Systems success”,<br />
Communication of the Association for Information Systems, 2 (20) Pp. 2-60<br />
Sedera, D., Gable, G., Cgan, T., (2003) “Measuring Enterprise System Success: A Preliminary Model” America<br />
<strong>Conference</strong> on Information Systems (AMCIS) Pp. 476-485<br />
Sedera, D. Gable, G. (2004) “A factor and structural equation analysis of the enterprise system success<br />
measurement model”, Twenty- fifth International <strong>Conference</strong> on Information Systems, Pp. 449-464<br />
Somers, T.M., Nelson, K., Karimi, J.,(2003) “Confirmatory Factors Analysis of the End User Computing Satisfaction<br />
Instrument: Replication within an ERP Domain”, Decision Sciences, 34 (3) Pp. 595-621<br />
Swartz, D., Orgill, K. (2000) “Higher education ERP: Lessons Learned”, EDUCAUSE 2000 in Nashville: Pp. 1-12<br />
Torkzadeh, G., Xenophon, K., Doll, W., (2005) “Confirmatory factor analysis and factorial invariance of the impact of<br />
information technology instrument”, OMEGA, 33. Pp. 107-118<br />
Torkzadeh, G., Doll, W.J. (1999) “The development of a tool for measuring the perceived Impact of information<br />
technology on work”, OMEGA, 27. Pp.327-339<br />
482
The Benefits of ICT Adoption: An Empirical Study of Nigerian<br />
SMEs<br />
Idisemi Apulu and Ann Latham<br />
University of Wolverhampton, UK<br />
i.apulu@wlv.ac.uk<br />
A.Latham@wlv.ac.uk<br />
Abstract: The successful operation of companies in most industries is becoming increasingly dependent on their<br />
ability to adopt and utilize ICT systems (Chibelushi, 2008). The rapid development of ICT is suddenly changing the<br />
conventional way of doing business in organizations. In recent times, organizations of all types are adopting ICT in<br />
order to improve organizational efficiency, productivity, communication and to strengthen their competitiveness. In<br />
the present knowledge economy, there is a need for organizations, including SMEs, to develop some forms of<br />
competitive advantage with the use of ICT in order to be successful. SMEs play a vital role in the Nigerian economy<br />
and greatly contribute to the country’s industrial development. Despite the growing number of studies on the benefits<br />
of ICT adoption in SMEs, the benefits of ICT adoption in the Nigerian context is still under researched. Thus, this<br />
paper is drawn from on-going PhD research and aims at investigating the benefits that are derived from the adoption<br />
of ICT within Nigerian SMEs. The analyses of data collected for this study would provide valuable information to SME<br />
owner-managers, policy makers and academic researchers.<br />
Keywords: ICT, SMEs, adoption, benefits, Nigeria<br />
1. Introduction<br />
In recent years, the growth of ICT has had a substantial impact on the way organizations function. Sigala<br />
(2003), states that the widespread use of ICT is changing the way people and companies work. ICT can<br />
be viewed as a collective term for a wide range of software, hardware, telecommunications and<br />
information management techniques, applications and devices, and are used to create, produce, analyse,<br />
process, package, distribute, receive, retrieve, store and transform information (Barba-Sánchez et al.,<br />
2007). Also, Gupta et al. (2008) define ICT as technologies such as the Internet, Intranets, Extranets,<br />
ERP and other such technologies that cover a wide spectrum from basic infrastructure implementation to<br />
technologies that improve services and operations in organizations. ICT helps to support human activities<br />
that enhance organizational or personal efficiency and effectiveness (Cohen et al., 2002). ICT helps to<br />
execute activities faster, support autonomous decision making processes, and enable distributive<br />
operations in order to become more efficient (Pokharel, 2005). It has revolutionized the manner in which<br />
firms interact and do business. ICT has a valuable potential for developing SMEs as the effective use and<br />
better integration of ICT in business processes could assist them to make more efficient decisions<br />
relevant to their performance (Sigala, 2003; Pokharel, 2005; Ongori and Migiro, 2010). Although there is<br />
a growing number of studies on the benefits of ICT adoption in SMEs, the benefits of ICT adoption in the<br />
Nigerian context as the area of study is still under researched. The paper is drawn from on-going PhD<br />
research and aims at investigating the benefits that are derived from the adoption of ICT within Nigerian<br />
SMEs. The analyses of data collected for this study would provide valuable information to SMEs ownermanagers,<br />
policy makers and academic researchers.<br />
2. Information and Communication Technology (ICT)<br />
Fulantelli and Allegra (2003), state that ICT offers enterprises a wide range of possibilities for improving<br />
their competitiveness and provides mechanisms for getting access to new market opportunities and<br />
specialized information services in organizations. According to Apulu and Latham (2009a), ICT helps<br />
organizations to exchange information and build closer relationships with their customers, suppliers and<br />
business partners. It also helps companies to provide immediate customer feedback that allows them to<br />
react fast to customer demands and recognize new market niches. Fulantelli and Allegra (2003) argue<br />
that the use of ICT brings about transparency in business processes in terms of dealing with<br />
stakeholders, which in turn, could lead to the development of better business practices that will help to<br />
meet customers’ service levels. ICT is an effective tool that can be used for improving external<br />
communications and delivering quality service to customers. It improves a firm’s ability to plan better for<br />
the future by providing the management with easy access to relevant information about different activities<br />
and/or operations within and outside the firm (Parida et al., 2009). Akkeren and Cavaye (1999) also state<br />
that ICT improves the ability for SMEs to compete with larger organizations and enables them to operate<br />
on an international scale. In the present knowledge economy, there is a need for organizations including<br />
SMEs, to develop some forms of competitive advantage with the use of ICT.<br />
483
Idisemi Apulu and Ann Latham<br />
3. Small and Medium Sized Enterprises (SMEs)<br />
SMEs form a significant proportion of the economy in many countries (Sarosa and Underwood, 2005).<br />
Wielicki and Arendt (2010) state that SMEs’ importance is magnified by the fact that for any country,<br />
regardless of its stage of economic development, SMEs generate the bulk of economic output (usually<br />
well over 90%). SMEs play a vital role in the Nigerian economy and greatly contribute to the country’s<br />
industrial development as they assist in creating jobs thereby improving the socio-economic development<br />
of the country. Nigerian SMEs are known as cornerstones responsible for Nigeria’s economic growth and<br />
stability (Ojukwu, 2006). They are also viewed as sources of flexibility and innovation that makes<br />
significant contributions to economies, both in terms of the number of SMEs and the proportion of the<br />
labour force employed by these firms. Lawrence (2008), states that SMEs contribution to the economy of<br />
many countries cannot be over emphasized.<br />
The definition of SMEs varies from country to country but is often based on employment, assets or a<br />
combination of the two. The Small and Medium Sized Development Agency of Nigeria (SMEDAN)<br />
defines SMEs based on the following criteria: a micro enterprise as a business with less than 10 people<br />
with an annual turnover of less than 5 million Naira, a small enterprise as a business with 10-49 people<br />
with an annual turnover of 5-49 million Naira and a medium enterprise as a business with 50-199 people<br />
with an annual turnover of 50-499 million Naira. In Nigeria, SMEs cover the entire range of economic<br />
activity within all sectors (SMEDAN, 2005). Zhouying et al. (2009), state that the use of ICT in SMEs can<br />
offer SMEs several benefits that could assist in increasing their capacities and competitiveness.<br />
4. Benefits of ICT in SMEs<br />
ICT has a significant positive impact on organizational performance (Maldeni and Jayasena, 2009) and is<br />
vital to SMEs. Without the use of ICT, modern businesses are not possible as ICT has a significant<br />
impact on SMEs’ operations and is claimed to be crucial for the survival and growth of economies in<br />
general (Berisha-Namanil, 2009). ICT provides opportunities for business transformations (Chibelushi,<br />
2008) and provides SMEs the opportunity to conduct business anywhere (Jennex et al, 2004). Brady et<br />
al. (2002), state that ICT plays an important role in enhancing the productivity and effectiveness of certain<br />
activities or functions made by SMEs. ICT can create new, strong linkages between internal activities,<br />
and even coordinate these actions more closely with their consumers and suppliers to facilitate<br />
integration within the company (Leenders and Wierenga, 2002). According to Prasad et al. (2009), ICT<br />
enhances companies’ abilities to coordinate activities regionally, nationally and globally and creates<br />
interrelationships between companies. It also expands the scope of industries in which companies<br />
compete to achieve competitive advantage (Porter and Millar, 1985).<br />
Love et al. (2004) ascertain that the use of ICT offers many benefits to SMEs at different levels<br />
(operational level, tactical level and strategic level). The benefit of ICT cuts across all sectors of the<br />
economy and all the fields of human activities. The <strong>European</strong> Commission (2008), states that SMEs<br />
could use ICT to grow and to become more innovative. Swift (2009) argues that SMEs benefit from the<br />
use of ICT as it connects them more easily and cheaply to external contacts. Through the use of ICT,<br />
SMEs can engage in e-commerce which will aid them in increasing their efficiency in their day-to-day<br />
business operations and sustain their business growth through the opening of new market channels<br />
(Ongori and Migiro, 2010). ICT enables SMEs to have access to robust business information that leads<br />
to organizational effectiveness (Irani, 2002). Akkeren and Cavaye (1999) state that ICT provides a cost<br />
effective way for SMEs to market their business, launch new products, improve communications, gather<br />
information and identify potential business partners. Hence, ICT can be regarded as a vital resource that<br />
enhances the competitiveness of SMEs in any business environment.<br />
5. Methodology<br />
The research question for this study is “Does ICT help SMEs in Nigeria?” The study adopts a case study<br />
approach and aims at determining whether or not ICT helps SMEs in Nigeria. Case study research<br />
enables the researcher to discover more effectively the context of a situation. Case studies assist to gain<br />
a deeper understanding of the studied phenomenon (Yin, 2003). In this study, case studies utilizing semistructured<br />
interviews, observation, and document review were employed. Multiple sources of data were<br />
used to address the ethical need to increase the reliability and validity of the research processes (Yin,<br />
2003). According to Orlikowski and Baroudi (1991); Remenyi and Williams (1996), case study is one of<br />
the most frequently used research methods in information systems research. Also, Sauer (1993) is of the<br />
opinion that research in information systems is best done by case study.<br />
484
Idisemi Apulu and Ann Latham<br />
Semi structured interviews that focused on the in-depth understanding of the benefits of ICT adoption<br />
within SMEs in Nigeria were used as tools for collecting data during the study and enabled the<br />
researchers to elicit respondents views and experiences in their own words. 25 SMEs in Apapa, Lagos<br />
were selected in accordance with the definition of SMEs in Nigeria (10 to 199 employees) and based on<br />
their decision to deploy ICT in their business processes. All the 25 interviewees were either ownermanagers,<br />
heads of departments or IT officers. Interviews were carried out using an interview guide, with<br />
each interview lasting between 1-1½ hours and tape recorded.<br />
Although the interviews were used in order to aid better interpretation of results, the use of documents<br />
and observations were also important. The researcher examined relevant documents (e.g. annual<br />
reports, project reports) that were gathered from the participating organizations. In addition, observing the<br />
management operations and procedures of the SMEs assisted in providing better information than just<br />
relying upon company reports and key informants. This multi-method approach enabled triangulation to<br />
take place. The data analysis process of the case study involved identifying patterns in the case study<br />
data. This includes arranging data in a chronological order, writing up the data according to phases and<br />
themes, in addition to grouping the paragraphs of the same themes and phases together.<br />
Furthermore, Lagos was chosen for the purpose of this study as it is considered the commercial nerve<br />
centre in Nigeria because of its strategic location, peculiar demographics and contribution to national<br />
Gross Domestic Product (GDP). According to Lawal (2002), over 60% of industries are located in Lagos<br />
hence; Lagos could serve as the industrial characteristics of Nigeria.<br />
The primary details of the SMEs that participated in the interviews are shown in the table below.<br />
Table 1: Primary details of the SMEs that participated in the interviews<br />
SME (Case study) Type of Business Number of<br />
Employees<br />
Business Age<br />
1 Engineering 43 (Small firm) 12 years<br />
2 Medical/Health 20 (Small) 13 years<br />
3 Telecommunication/Communication 26 (Small) 18 years<br />
4 Wholesale/Retail 90 (Medium) 9 years<br />
5 Hospitality 120 (Medium) 5 years<br />
6 Telecommunication/Communication 16 (Small) 3 years<br />
7 Telecommunications/Communication 16 (Small) 3 years<br />
8 Transport/Clearing and Forwarding/Haulage 15 (Small) 2 years<br />
9 Financial Service/Service Industry 35 (Small) 5 years<br />
10 Manufacturing 80 (Medium) 10 years<br />
11 Transport/Clearing and Forwarding/Haulage 15 (Small) 21 years<br />
12 Transport/Clearing and Forwarding/Haulage 15 (small) 9 years<br />
13 Others 25 (Small) 2 years<br />
14 Communication/Telecommunication 65 (Medium) 8 years<br />
15 Manufacturing 89 (Medium) 7 years<br />
16 Financial service or service industry 25 (small) 4 years<br />
17 Pharmacy/Medical/Health 13 (small) 14 years<br />
18 Others 18 (small) 8 years<br />
19 Oil and Gas/Maritime 14 (small) 2 years<br />
20 Oil and Gas/Maritime 30 (small) 5 years<br />
21 Engineering 29 (Small) 11 years<br />
22 Financial Service/Service Industry 22 (small) 17 years<br />
23 Education 29 (small) 7 years<br />
24 Legal Practice 17 (small) 5 years<br />
25 Oil and Gas/ Maritime 15 (small) 1 year<br />
Interviewees were asked to comment on the benefits derived from the use of ICT solutions in their<br />
various firms. Respondents identified several benefits of using ICT which include:<br />
5.1 Efficiency<br />
Thirteen respondents identified efficiency as a benefit gained from adopting ICT. The respondents stated<br />
that ICT brings about efficiency in terms of conducting their various businesses. For example one<br />
respondent states that:<br />
485
Idisemi Apulu and Ann Latham<br />
“ICT has helped us to become more efficient in terms of carrying out our duties as a small<br />
company. In the past, we were using bill cards to get our stock update, which was a manual<br />
process and usually takes some days. However, since we started using ICT we can get our<br />
stock update under some hours and it has made our company to become more efficient”.<br />
5.2 Speed<br />
Some interviewees were also of the opinion that ICT brings about speed. According to one respondent:<br />
“ICT now makes us to deliver quality service to our customers within a very short time.<br />
Unlike the manual method, the use of ICT makes our work to move very fast”.<br />
5.3 Productivity<br />
Increase in productivity was identified as another benefit of using ICT in the companies. Respondents<br />
believe that the adoption of ICT in their various companies has enabled them to become more<br />
productive. According to one company:<br />
“With ICT our business is more productive in terms of meeting our targets and ICT has<br />
helped us to become more creative”.<br />
5.4 Communication<br />
Some respondents stated that ICT improves communication. According to one respondent:<br />
“ICT helps us to communicate, in terms of exchanging information. ICT is mainly used in our<br />
organization for communication with our customers, suppliers, to get quotation for various<br />
products by searching the internet and also to get information from other stake holders in the<br />
business”.<br />
Another respondent states that:<br />
“ICT helps us to communicate better with our clients. We usually receive feedbacks from our<br />
clients within a short time via email and it also helps us to manage our data”.<br />
5.5 Easy Access to Information<br />
Easy access to information was described by respondents as a benefit of using ICT. Respondents<br />
believe that the use of ICT has enabled information to become readily available in their various<br />
companies. One respondent commented:<br />
“ICT makes it very easy for us to access information. ICT is a key to success in every<br />
organization”.<br />
Another respondent states that:<br />
“ICT brings about the availability of data, that is, information availability on a daily basis”.<br />
5.6 Awareness and Increase in profit<br />
Respondents identified some other benefits of using ICT such as helping to create awareness and<br />
assisting their companies to make or increase profit. One company comments:<br />
“ICT is interesting, it has added a lot to the body of knowledge. ICT helps organizations to<br />
create awareness because it provides the opportunity for companies to advertise their<br />
products and services online. ICT has helped us to save cost and also, to increase profit.<br />
Virtually every aspect of the whole life is now embracing ICT which makes it more lucrative<br />
for companies to get more benefits and gain, such as increase in their annual returns. ICT is<br />
wonderful”.<br />
Another company comments:<br />
“ICT has helped to increase our company’s profit. Every organization is a profit oriented<br />
organization and it implies that we are here to make profit through the use of modern<br />
technologies”.<br />
486
5.7 Customer satisfaction<br />
Idisemi Apulu and Ann Latham<br />
All the respondents’ mentioned that ICT helps to provide better customer service to their customers<br />
thereby increasing customers’ satisfaction. One respondent commented:<br />
“ICT helps us to satisfy our customers by providing them with their required service”.<br />
Also, another respondent commented:<br />
“ICT is what is in vogue nowadays. You cannot do without technology. If you are competing<br />
with other organizations that are ICT compliant, it is very important that you are also able to<br />
compete, as this will help to satisfy your customers. Information has a lot to do with<br />
organizational advantage in terms of customers’ satisfaction”.<br />
5.8 Competitive advantage<br />
Eighteen respondents believe that one of the benefits of using ICT is the fact that it helps organizations to<br />
have some form of competitive advantage. According to one respondent:<br />
“The adoption of ICT in this company has enabled us to have some sort of competitive<br />
advantage against our competitors”.<br />
The other respondents also made similar comments while identifying the benefits of using ICT in their<br />
different companies.<br />
5.9 Customer growth and good reputation<br />
Two respondents stated that ICT has assisted them to attract more customers and to establish good<br />
reputation with their existing customers. One of the respondents commented:<br />
“The world globally is changing; technology is now at the forefront of every organisation.<br />
There is no way as an SME; we can meet up with the current trend of technology if we do<br />
not use ICT because everything is going digital. As a company, using manual systems<br />
nowadays can hinder good patronage from people. Hence, we had to come out and do what<br />
others are doing in order to meet up with the standard of others. The use of ICT has assisted<br />
us to attract more customers and has also helped to build a good reputation for this<br />
company.<br />
5.10 Planning and focusing<br />
One company stated that ICT helps them to plan and be more focused. ICT also helps to provide their<br />
business with the services they require. The company, a telecommunication company, comments:<br />
“The benefits of using ICT are much and based on my opinion, ICT is indispensible. For<br />
example if you do not have a computer, you cannot produce things like recharge cards.<br />
Companies are also doing away with manual accounting these days. They are moving from<br />
paper ledger to automated ledger. It is only with the use of ICT infrastructures that we can<br />
reasonably handle these developments and put things in a good format. ICT helps us to plan<br />
and to be more focused. Also, ICT helps us in the area of database management and<br />
inventory production. It constitutes the life wire of the efficient management of our business”.<br />
5.11 User friendliness<br />
User friendliness was also identified by a respondent as a benefit of ICT. The respondent stated that:<br />
“ICT is user friendly. For example as a construction company, we can design an estate<br />
within a week using ICT applications such as AutoCAD”.<br />
5.12 Confidence<br />
Another respondent believes that ICT brings confidence to business.<br />
According to the respondent:<br />
“One of the benefits we have derived from the use of ICT applications is the fact that it helps<br />
to build confidence in our customers especially when they see the use of modern<br />
technologies in this company. People can also do business with us online and this increases<br />
487
Idisemi Apulu and Ann Latham<br />
confidence. When a company have a website for example, it brings about confidence.<br />
People can go to the website and get the information they require. ”<br />
The table below provides a summary of the Findings.<br />
Table 2: Summary of the findings<br />
Benefits of ICT in Nigerian SMEs Number of Respondents<br />
Efficiency 13<br />
Speed 17<br />
Productivity 3<br />
Communication 3<br />
Easy access to information 7<br />
Awareness and Increase in profit 4<br />
Customer satisfaction 25<br />
Competitive advantage 18<br />
Customer growth and reputation 2<br />
Planning and focusing 1<br />
User friendliness 1<br />
Confidence 1<br />
6. Discussion<br />
The evidence demonstrates that there are several benefits associated with the use of ICT in SMEs. The<br />
responses obtained from the interviewees have enabled the researchers to identify some major benefits<br />
associated with the use of ICT in Nigerian SMEs. The introduction of ICT in SMEs can bring about<br />
modifications in the way they work. On the whole, ICT applications can provide several benefits across a<br />
wide range of intra- and inter- firm business operations and transactions. Parida et al. (2009), in their<br />
research with Swedish firms, discovered that ICT helps firms to access information. The authors stated<br />
that easy access to information is precious for firms operating in a dynamic environment and that the<br />
firms need to be updated on new technology, innovations amongst others, which might influence a firm’s<br />
future competitiveness. According to Ion and Andreea (2008), with ICT high efficiency is obtained in<br />
business transactions due to a fast and accurate processing of information. ICT applications can assist to<br />
improve information and knowledge management within an organization, can reduce transaction costs<br />
and can increase the speed and reliability of transactions in any organization (Apulu and Latham, 2009b).<br />
More specifically, Barba-Sánchez et al. (2007), argue that SMEs can obtain a wide range of benefits<br />
from the use of ICT. They also state that ICT can reduce business costs, improve productivity and<br />
strengthen growth possibilities.<br />
Based on the case studies and literature review, it can be said that ICT helps to improve customer<br />
service in SMEs. Parida et al. (2009), state that ICT is a tool that provides customers with better services<br />
in terms of quicker responses and closer interactions. ICT also improves communication and maintain<br />
close relationships with business partners. Most of the respondents in the study stated that ICT brings<br />
about competitive advantage. Apulu and Latham (2009b) advocate that appropriate use of ICT can assist<br />
SMEs gain competitive advantage by reducing costs and improving core business processes. Therefore,<br />
from the literature review and case studies, it is certain that one of the major benefits derived from the<br />
use of ICT in Nigerian SMEs, is the fact that it helps them to remain competitive.<br />
Some respondent also mentioned that ICT helps to increase efficiency in their various organizations.<br />
According to Cohen et al. (2002), ICT plays a supportive role in terms of enhancing organizational<br />
efficiency and effectiveness. Fink and Disterer (2006) also argue that ICT helps organizations to become<br />
more efficient, effective and competitive. Furthermore, Pavic et al. (2007) state that organizations have<br />
the opportunity to achieve a competitive advantage from the advances in ICT through market efficiency<br />
gains, better quality, innovation and customer responsiveness.<br />
Increase in profit is identified as another benefit that is derived from adopting ICT in Nigerian SMEs.<br />
Respondents stated that ICT is used as a marketing tool to help increase profit. Also some respondents<br />
mentioned that ICT is a modern technology and it is impossible for companies to cope without utilizing<br />
some form of modern technologies especially in terms of competing with other organizations that are ICT<br />
compliant. According to Ongori and Migiro (2010), the impact of globalization has enabled companies to<br />
build strong relationships with their counterparts from any part of the world by using ICT. This has obliged<br />
many SMEs to adopt ICT in order to survive in the present competitive era.<br />
488
Idisemi Apulu and Ann Latham<br />
Easy access to information was described as another benefit. Respondents stated that the use of ICT<br />
has enabled information to become readily available in their various companies. Similarly, Irani (2002)<br />
ascertains that ICT helps SMEs to have access to robust business information that leads to<br />
organizational effectiveness. Also, Funteli and Allegra (2003) argue that organizations can exchange<br />
real-time information and build closer relationships with customers, suppliers and business partners with<br />
the use of ICT. Customers can also receive immediate feedback from companies and move from manual<br />
data processing to automated data processing. Improved communication was also described by<br />
respondents as a benefit of using ICT within Nigerian SMEs. It was noted that most SMEs decided to<br />
adopt ICT to enable them communicate with their customers, suppliers, and stakeholders in their field of<br />
business. According to Kollberg and Dreger (2006), ICT provides new ways of storing, processing,<br />
distributing and exchanging information within companies and customers. Also, Akkeren and Cavaye<br />
(1999) state that ICT helps to improve communication, gather information and identify potential business<br />
partners in organizations.<br />
Ashrafi and Mutarza (2008), argue that organizations of all type around the globe are currently utilizing<br />
ICT in order to provide better customer service. In terms of customer satisfaction, all the respondents in<br />
the study stated that ICT has enabled them to provide better customer service to their customers.<br />
Respondents also mentioned that ICT has assisted them to fully compete with their competitors.<br />
According to Chowdbury and Wolf (2003), ICT helps to increase business competitiveness and enhance<br />
enterprise performance through indirect cost savings such as labour cost and increased labour<br />
productivity and direct cost such as reduction of firms input. From the case studies, it was identified that<br />
the use of ICT increases productivity as it helps SMEs to meet their desired targets and also increases<br />
the speed of delivery in terms of responding to customers’ demands. All these help to bring about<br />
customers’ satisfaction. Jones et al. (2003), state that ICT brings quick response time to markets,<br />
customers, suppliers, increases flexibility and reduces delivery time and processing of payment.<br />
Having online presence has been identified by some respondents as another benefit of adopting or using<br />
ICT in their companies as ICT provides an opportunity for businesses to offer products and services in<br />
the global markets. Some respondents also commented that having online presence helps to create<br />
awareness. For example, a respondent stated that one of the benefits of using ICT is the fact that it helps<br />
to build their customers confidence in company, especially with the use of modern technologies.<br />
Therefore, this implies that ICT helps to inform potential customers’ about a company’s service.<br />
ICT is being applied in a wide range of areas within many organisations in recent times. This has radically<br />
changed the manner in which businesses are conducted in organizations (Cohen and Kallirroi, 2006).<br />
According to Zhouying et al. (2009), ICT can help firms to overcome spatial and institutional barriers, cut<br />
time, cost of production, increase flexibility, take new marketing channel and employ new ways of doing<br />
business. This means that ICT brings about increased competitiveness.<br />
Furthermore, Parida et al. (2009) argue that ICT improves a firm’s ability to plan better for the future. A<br />
company stated that ICT helps them not only to plan and become more focused, but also to provide their<br />
business with the services they require. Two other companies mentioned that ICT helps them to build a<br />
brand and to satisfy their customers thereby assisting them to build a good reputation. In order words, the<br />
integration of ICT in SMEs business processes can greatly improve their performance.<br />
7. Conclusion<br />
The findings of this study imply that ICT helps SMEs in Nigeria. Some benefits associated with the use of<br />
ICT in Nigerian SMEs have been identified, which support the findings of other research. For example,<br />
ICT enhances organizational speed/efficiency (Cohen et al., 2002; Sigala, 2003; Pokharel, 2005; Ongori<br />
and Migiro, 2010), improves the ability for SMEs to compete (Akkeren and Cavaye, 1999; Fulantelli and<br />
Allegra, 2003), enhances productivity (Brady et al., 2002), enables SMEs to have easy access to robust<br />
information (Akkeren and Cavaye, 1999; Irani, 2002), helps to plan for the future (Parida et al., 2009) and<br />
improves communication (Akkeren and Cavaye, 1999). Furthermore, the study identifies some distinctive<br />
benefits that are associated with the use of ICT in Nigerian SMEs, such as customer satisfaction,<br />
customer growth and reputation and confidence in business.<br />
The identified benefits show that the deployment of ICT applications in these SMEs can assist in<br />
providing better customer service and enable them to effectively compete with their competitors. Based<br />
on the literature review and the case studies, it is certain that the utilization of ICT in Nigerian SMEs<br />
would assist in increasing their competitiveness. Therefore, it can be concluded that Nigerian SMEs<br />
489
Idisemi Apulu and Ann Latham<br />
stand to gain some added advantage from the use of ICT as it can help to further develop their business<br />
processes. The findings of this study could encourage SME owner-managers and IT professionals to<br />
intensify their efforts in deploying ICT in their various organizations as the study has provided some<br />
explanation on the benefits associated with the deployment of ICT within SMEs using the Nigerian<br />
context.<br />
References<br />
Apulu, I. and Latham, A. (2009a) “Knowledge Management: Facilitator for SMEs competitiveness in Nigeria”, UKAIS<br />
2009 <strong>Conference</strong>, St Anne’s College, Oxford United Kingdom.<br />
Apulu, I. and Latham, A. (2009b) “A review of factors affecting ICT adoption in SMEs in Developing Countries: a<br />
Case Study of Nigeria”, Proceedings of the ISBE 2009 <strong>Conference</strong>, University of Liverpool, United Kingdom.<br />
Ashrafi, R. and Murtaza, M. (2008) “Use and Impact of ICT on SMEs in Oman”, Electronic Journal Information<br />
Systems Evaluation, Vol 11, No. 3, pp 125-138. [online] http://www.ejise.com/volume-11/volume11issue3/ashrafiAndMurtaza.pdf<br />
[Accessed on 10th October, 2009]<br />
Barba-Sánchez, V., Martínez-Ruiz, M and Jiménez-Zarco, A. (2007) “Drivers, Benefits and Challenges of ICT<br />
Adoption by Small and Medium Sized Enterprises (SMEs): A Literature Review, Problems and Perspectives in<br />
Management, Vol 5, No. 1, pp 103-114.<br />
Berisha-Namanil, M. (2009) “The role of information technology in small and medium sized enterprises in Kosova,<br />
Fulbright Academy 2009 <strong>Conference</strong> Small Places Can Change the World. [online] Availble:<br />
http://www.fulbrightacademy.org/file_depot/0-<br />
10000000/2000030000/21647/folder/82430/Berisha+Paper+IT+in+SMEs+in+Kosovo.pdf [July 15, 2009].<br />
Brady, M.; Saren, M. and Tzokas, N. (2002) “Integrating Information Technology into Marketing Practice – The IT<br />
Realize of Contemporary Marketing Practice”, Journal of Marketing Management, Vol 18, pp 555-577.<br />
Chibelushi, C. (2008) “Learning the hard way? Issues in the adoption of new technology in small technology oriented<br />
firms”, Education + Training, Vol 50, Issue 8/9, pp 725-736.<br />
Chowdhury, S. K., and Wolf, S. (2003) “Use of ICTs and the economic performance of SMEs in East Africa”,<br />
Discussion Paper No. 2003/06.<br />
Cohen, S. and Kallirroi, G. (2006) “e-Commerce Investments from an SME perspective: Costs, Benefits and<br />
Processes”, The Electronic Journal Information Systems Evaluation, Vol 9, Issue 2, pp 45 – 56.<br />
<strong>European</strong> Commission (2008) “Making SMEs more competitive” [online] Available:<br />
http://ec.europa.eu/enterprise/sme/competitive_en.htm [October 30, 2008].<br />
Fink, D., and Disterer, G. (2006) “International case studies: to what extent is ICT infused into the operations of<br />
SMEs”, Journal of Enterprise Information, Vol. 19, 608-624.<br />
Fulantelli, G., and Allegra, M. (2003) “Small company attitude towards ICT based solutions: Some key-elements to<br />
improve it”, Educational Technology and Society, Vol 6, Issue 1. [online] Available<br />
http://www.ifets.info/journals/6_1/fulantelli.html [March 10, 2009].<br />
Gupta, B., Dasgupta, S., and Gupta, A. (2008) “Adoption of ICT in a government organization in a developing<br />
country: an empirical study”, The Journal of Strategic Information Systems, Vol 17, No 2, pp140-154.<br />
Ion, P., and Andreea, Z. (2008) Use of ICT in SMEs management within the sector of services. [online] Available:<br />
http://steconomice.uoradea.ro/anale/volume/2008/v4-management-marketing/085.pdf [September 5, 2009].<br />
Irani, Z. (2002) “Information systems evaluation: navigating through the problem domain”, Journal of Information &<br />
Management, Vol 40, Issue 1, pp 11-24.<br />
Jones, P., Beynon-Davies, P., and Muir, E. (2003) “E-business barriers to growth within the SME sector”, Journal of<br />
Systems and Information Technology, Vol 7, Issue 1/2, pp 1 – 25.<br />
Kollberg, M., and Dreyer, H. (2006) “Exploring the impact of ICT on integration in supply chain control: a research<br />
model”, Department of Production and Quality Engineering, Norwegian University of Science and Technology,<br />
Norway.<br />
Lawal, A. A. (2002) Entrepreneurship Development in Small and Medium Enterprises: Key Success Factor. Lagos<br />
Journal of Business, Vol 2, No 1, pp 35-46.<br />
Lawrence, J. E. (2008) “The challenges and utilization of E-commerce: use of Internet by Small to Medium Sized<br />
Enterprises in the United Kingdom”, Information, Society & Justice, Vol 1, No 2, pp 99-113.<br />
Leenders, M. A. A. M. and Wierenga, B. (2002) “The Effectiveness of Different Mechanisms for Integrating Marketing<br />
and R&D. Journal of Product Innovation Management, Vol 19, No 4, pp 305-317.<br />
Love, P. E. D., Irani, Z. and Edwards, D. J. (2004) “Industry-Centric benchmarking of information technology benefits,<br />
costs and risks for small and medium sized enterprises in construction”, Automation in Construction, Vol 13, No<br />
4, pp 507-524.<br />
Maldeni, H. M. C. M., and Jayasena, S. (2009) “Information and Communication Technology Usage and Bank<br />
Branch Performance”, The International Journal on Advances in ICT for Emerging Regions (ICTer), Vol 2, No 2,<br />
pp 29 – 37.<br />
Ojukwu, D. (2006) “Achieving sustainable growth through the adoption of integrated business and information<br />
solutions: a case study of Nigeria small and medium sized enterprises”, Journal of Information Technology<br />
Impact, Vol 6, No 1, pp 47-60.<br />
Ongori, H. and Migiro, S. O. (2010) “Information and Communication technology adoption: a literature review”,<br />
Journal of Chinese Entrepreneurship, Vol 2, No 1, pp 93-104.<br />
Orlikowski, W.J. and Baroudi, J. J. (1991) “Studying Information Technology in organisations: Research Approaches<br />
and Assumptions”, Information Systems Research, Vol 2, pp 1-28.<br />
490
Idisemi Apulu and Ann Latham<br />
Parida, V., Westerberg, M., and Yinenpaa, H. (2009) “How do small firms use ICT for business purpose? A study of<br />
Swedish technology-based firms”, International Journal of Electronic Business, Vol 7, No 5, pp 536-551.<br />
Pavic, S., Koh, S. C. L. Simpson., M. and Padmore, J. (2007) “Could e-business create a competitive advantage in<br />
UK SMEs?” Benchmarking: An International Journal, Vol14, No 3, pp 320-351.<br />
Pokharel, S. (2005) “Perceptions on information and communication technology perspectives in logistics: A study of<br />
transportation and warehouses sections in Singapore”, Journal of Enterprise Information Management, Vol 18,<br />
Issue 1/2, pp 136-149.<br />
Porter, M. E. and Millar, V.E. (1985) “How information gives you competitive advantage”, Harvard Business Review<br />
(July-August), Vol 63, Issue 4, pp 149-174.<br />
Remenyi, D., and Williams, B. (1996) “The Nature of Research: Qualitative or Quantitative, Narrative or<br />
Paradigmatic?”, Information Systems Journal, Vol 6, pp 131-146.<br />
Sarosa, S. and Underwood, J. (2005) “Factors affecting ICT adoption within Indonesian SMEs: Manager’s<br />
perspective”, PACIS 2005 proceedings, paper 6.<br />
Sauer, C. (1993) Why Information Systems Fail: A case Study Approach. Henley-on-Thames: Alfred Wailer.<br />
Siagala, M. (2003) “The information and communication technologies productivity impact on the UK hotel sector”,<br />
International Journal of Operations & Production Management, Vol 23, No 10, pp 1224-1245.<br />
Small and Medium Enterprises Agency of Nigeria (SMEDAN) (2005) “SME Success Digest”, Volume 3, No 1,<br />
January-June, 2005.<br />
Swift, M. (2009) “ICTs for SMEs – How ICT can be a force of viability”, ICT for SMEs Seminar, The Cascadia Hotel &<br />
<strong>Conference</strong> Centre Port of Spain, Trinidad & Tobago Wednesday, 25th March, 2009. [online]<br />
http://www.iberpymeonline.org/Trinidad0309/MichaelSwift.pdf [Accessed 10th June 2009].<br />
Van Akkeren, J and Cavaye, L. M. A. (1999) “Factors affecting entry-level internet technologies adoption by small<br />
business in Australian: an empirical study”, 10 th Australasian <strong>Conference</strong> on Information Systems, Vol 199, pp<br />
1071-1083.<br />
Wielicki, T. and Arendt, L (2010) “A knowledge-driven shift in perception of ICT implementation barriers: comparative<br />
study of US and <strong>European</strong> SMEs”, Journal of Information Science, Vol 36, No 2, pp 162-124.<br />
Yin, R. K (2003) Case Study Research: Design and Methods. 3rd Edition, Applied Social Science Method Series,<br />
Vol. 5, Sage Publications, Thousand Oaks, CA.<br />
Zhouying, S., Weidong, L., and Turong, M. (2009) “The sustainable development of ICTs in China: An empirical<br />
study of small and medium enterprise”, Chinese Journal of Population, Resources and Environment, Vol 7, No<br />
3, pp 20-29.<br />
491
Establishing the Suitability of Dynamic Web Applications Development<br />
Platforms for Learning web Techniques in Tertiary<br />
Institutions<br />
Johnson Dehinbo<br />
Tshwane University of Technology, Pretoria, South Africa<br />
DehinboOJ@tut.ac.za<br />
Abstract: The availability of many platforms for developing dynamic Web applications suggests the existence of the<br />
problem of choosing the most suitable platform for learning the concepts of web applications development by undergraduate<br />
students in tertiary institutions. Students may not perform at their best capacity level if the platform chosen<br />
by the institution is not very suitable for learning the relevant concepts. As part of the framework to determine the<br />
most suitable platform for leaning web applications development in tertiary institutions, this study establishes a set of<br />
criteria for evaluating the suitability for learning the concepts of Web techniques in Web application development<br />
such as Web page serving, HyperText Transmission Protocol (HTTP) basics, Server information, Form validation<br />
and processing, as well as maintenance of Session and Application states. These criteria were applied by evaluating<br />
four platforms namely Java Servlets, Java Server Pages, Active Server Pages and PHP using various research<br />
methods including descriptive inquiry, document analysis, observations and programming tests. While ASP was<br />
found to be more suitable on applying the criteria, the significance of the study lies in the establishment of a comprehensive<br />
but specific set of criteria that can be used as a scientific basis for selection.<br />
Keywords: web applications development platforms, programming languages, web techniques, comparison frameworks<br />
1. Background of the study<br />
This study establishes part of a framework containing various criteria that can be used to evaluate dynamic<br />
Web platforms to determine a suitable platform for learning Web applications development in tertiary<br />
institutions. This part of the framework is to determine platform that will be suitable for learning the<br />
concepts of Web techniques in Web applications development.<br />
There are currently many platforms for implementing dynamic application programs on the World Wide<br />
Web and for learning dynamic web applications development in tertiary institutions today. Lim (2002)<br />
states that information systems departments need to reexamine their curricula in order to prepare students<br />
to face the challenge of being productive in a computing world swamped with web technologies.<br />
The use of various programming and scripting languages could lead to elements of repetition and confusion.<br />
Yet, comprehension is very important in undergraduate studies (Wiedenbeck 1999:5).<br />
The comprehension of relevant concepts by students may be affected by the choice of platform because<br />
according to Sebesta (1996:2-3), the language in which programmers develop software places various<br />
limits on the development efforts. Thus, the choice of dynamic web application development platform<br />
should involve critical evaluation in line with the statement below:<br />
Web development tools need to be analyzed in terms of its purpose (what it is designed to<br />
do), technology (ease of use, robustness, scalability, security, performance, etc.), support<br />
(portability, cost, ISP support), and how well it works in the real world (Ashenfelter 1999:<br />
105).<br />
With regard to the purpose-based analysis of the platforms, appropriate Web techniques are critical for<br />
developing dynamic Web applications. The research question then is: How do we determine the dynamic<br />
Web platform that will be suitable for learning the concepts of Web techniques in Web applications development<br />
by undergraduate students?<br />
The objectives of the study are therefore given as follows:<br />
1.) To identify the concepts of Web techniques that needs to be learnt by undergraduate students of<br />
Web applications development in tertiary institutions.<br />
2.) To establish the criteria to determine suitability for learning the concepts.<br />
3.) To apply the criteria by using them to evaluate the suitability of four specified platforms.<br />
492
2. Literature review<br />
Johnson Dehinbo<br />
Evaluating programming languages, development platforms and tools is very important for understanding<br />
the effects on novice programmers, but is difficult to carry out (Wiedenbeck et al. 1999). This could be the<br />
reasons why Apte et al. (2003) note that a study of existing literature showed varying conclusions about<br />
the superiority of one dynamic Web platform over another. Prechelt (2000) also indicates that programmers<br />
usually hold varied strong opinions on development platforms. These are illustrated in the various<br />
approaches towards the comparisons of platforms in the studies reviewed below.<br />
2.1 Previous studies comparing platforms<br />
In a survey of middleware platforms, Cooper (2001) concluded that ColdFusion is fast to learn and fast to<br />
use, Common Gateway Interface (CGI) is also fast to learn and that Servlets are hard to learn and use,<br />
even by someone who already knows Java. Bishop and Hurter (1999) examined the Scripting languages;<br />
Tcl/Tk, Perl and Python and found out that Python incorporated the features of Modula-3 into its scripting<br />
syntax thus making it suitable for "programming in the large". Programs in Python were also found to be<br />
typically much shorter than equivalent in C or Java.<br />
Prechelt (2000) observes that designing and writing programs in the scripting languages, namely Perl,<br />
Python, Rexx, or Tcl takes no more than half as much time as writing it in C, C++, or Java. Moreover, the<br />
resulting program is only half as long. Kruse (2003) illustrates the differences in the strengths and weaknesses<br />
of Personal Home Page (PHP) and Java.<br />
However, most of these studies did not use any criterion as a basis for the comparisons. The comparisons<br />
in these studies seem to be based on intuition rather than scientific facts. Comparison needs to be<br />
based on a variety of factors supported by scientific facts and results.<br />
A common criterion used in other studies is performance. Renaud et al. (2003) indicate various metrics<br />
can be used to measure performance. These include response or waiting time, synch delay, number of<br />
messages exchanged, throughput, communication delay, node fairness, Processor and memory usage.<br />
Marshak and Levy (2003) evaluated platforms only in terms of user-perceived latency.<br />
Interestingly, Vinoski (2003) pointed out that performance has somehow been overemphasized as people<br />
check only those qualities that are easily measurable, such as performance. This has unintentionally led<br />
many programming language users to presume that “high performance” is the same as “high quality”.<br />
Such presumptions could be entirely meaningless, depending on the nature of one’s application. So,<br />
comparisons should involve other relevant factors as criteria.<br />
2.2 Studies using various criteria in their comparisons<br />
Cecchet et al (2003) evaluate three specific mechanisms namely PHP, Java Servlets, and Enterprise<br />
Java Beans (EJB) with respect to performance and ease of development. The study attributes PHP’s better<br />
performance to the fact that it executes as a module in the Web server, sharing the same process<br />
(address space), thereby minimizing communication overhead. This is unlike Java Servlets which run in a<br />
Java Virtual Machine (JVM) as a separate process from the Web server (Cecchet et al 2003). In terms of<br />
ease of development, Cecchet et al (2003) explain that PHP scripts are easy to write because they embed<br />
code directly into an HTML page.<br />
In examining ASP, PHP and ASP.NET, Hartman (2001) mentioned two factors that complicate choosing<br />
a scripting environment. First, there is the issue of culture among developers which has a lot to do with<br />
the ideological camps to which they belong. If they love to tinker with source code to develop solutions<br />
more efficient than off-the-shelf products, and embellished their cubicles with defaced portraits of Bill<br />
Gates, they would probably prefer to use PHP. Convenience lovers would probably prefer to use ASP.<br />
Yet, very few developers are equally willing to use both, or talk about "the other" technology without a<br />
trace of disdain. Hartman’s second factor is the website's future scalability and functional requirements<br />
which might restrict which servers and platforms the site could run on or impact the feasibility of developing<br />
future features, such as database-linked connectivity (Hartman 2001).<br />
However, the most important criterion is the suitability for doing the job for which a tool is needed. This<br />
view is presented next.<br />
493
Johnson Dehinbo<br />
2.3 Studies emphasizing suitability for achieving the purpose of systems<br />
Web development tools need to be analyzed in terms of its purpose (Ashenfelter 1999:105). Similarly, for<br />
choosing a language or platform that is suitable for learning introductory programming, Holt et al. (1997)<br />
use the criteria that it should be appropriate for introducing programming concepts used in the real world<br />
such as in business, science and government; and it should encourage systematic problem solving. Similarly,<br />
Kolling and Rosenberg (1996) suggest clean concepts directly reflecting the theoretical model and<br />
avoidance of conceptual redundancy as achieving the same thing in a variety of ways can mean flexibility<br />
to the expert, but usually confuses beginners.<br />
Moreover, Hadjerrouit (1998) evaluates the suitability of Java as a first programming language using the<br />
criteria like: programming concepts to be taught including problem solving skills, algorithmic thinking and<br />
structured programming; use in subsequent courses and programming paradigm support.<br />
These studies emphasize the need for the suitability for doing the job for which a tool is needed as Web<br />
techniques is required in Web applications development.<br />
2.4 Summary of the literature review<br />
Table 1 summarizes the literature review. A key point observed is that choosing a suitable platform<br />
should involve exhaustive evaluations of various options based on various relevant criteria that are<br />
backed by scientific facts and results. This study fills this gap in the body of knowledge by being unique in<br />
the following ways: The advantages and strengths of platforms should be examined and ranked in the<br />
light of certain desired qualities relevant to specific use such as the suitability for learning Web techniques,<br />
just as for example, the ease of learning increases productivity in institutions than flexibility which<br />
could be preferable in industries. This informs the identification of desirable concepts of Web techniques<br />
before establishing criteria for evaluations.<br />
Table 1: Scoring for the platforms based on the criteria on web techniques.<br />
Sources Tasks / Conclusions Limitations / Points noted<br />
Cooper(2001) Surveyed some middleware platforms and concludes<br />
that Coldfussion is fast to learn and use while Servlet is<br />
hard to learn and use.<br />
Bishop & Hurter<br />
(1999)<br />
Examined scripting languages Tcl/Tk, Perl and Python<br />
and concludes that Python programs are typically<br />
shorter.<br />
Prechelt(2000) Observes that designing and writing programs in the<br />
scripting languages, namely Perl, Python, Rexx, or Tcl<br />
takes no more than half as much time as writing it in C,<br />
C++, or Java. Moreover, the resulting program is only<br />
half as long<br />
Kruse (2003) Illustrates the differences in the strengths and weaknesses<br />
of Personal Home Page (PHP) and Java<br />
Renaud et al.<br />
(2003)<br />
Marshak and<br />
Levy (2003)<br />
Indicate various metrics can be used to measure performance.<br />
These include response or waiting time,<br />
synch delay, number of messages exchanged,<br />
throughput, communication delay, node fairness, Proc-<br />
essor and memory usage<br />
Also evaluated platforms only in terms of userperceived<br />
latency.<br />
Vinoski (2003) Pointed out that performance has somehow been<br />
overemphasized as people check only those qualities<br />
that are easily measurable, such as performance<br />
494<br />
Systematic approach not<br />
used in evaluating ease<br />
of learning and ease of<br />
use<br />
Comparison not based<br />
on specific use<br />
Did not use any specific<br />
criterion as a basis for<br />
the comparisons.<br />
The comparisons seem<br />
to be based on intuition<br />
rather than scientific<br />
facts<br />
Comparison concentrates<br />
on performance<br />
which Vinoski states has<br />
been over-emphasized<br />
Comparison based on<br />
latency form of perform-<br />
ance as above<br />
Performance may not be<br />
the ultimate depending<br />
on the nature of one’s<br />
application
Cecchet et al<br />
(2003)<br />
Johnson Dehinbo<br />
Evaluate PHP, Java Servlets, and Enterprise Java<br />
Beans (EJB) with respect to performance and ease of<br />
development. The study attributes PHP’s better performance<br />
to the fact that it executes as a module in the<br />
Web server, sharing the same process (address<br />
space), thereby minimizing communication overhead.<br />
Serve as example towards<br />
multiple factors<br />
used in the comparisons<br />
Sources Tasks / Conclusions Limitations / Points noted<br />
Hartman (2001) Examined ASP, PHP and ASP.NET, mentioned two<br />
factors that complicate choosing a scripting environment.<br />
These are the issue of culture among developers<br />
dealing with the ideological camps to which they belong<br />
and the website's future scalability with functional<br />
Ashenfelter<br />
(1999: 105)<br />
requirements.<br />
Established that “Web development tools need to be<br />
analyzed in terms of its purpose (what it is designed to<br />
do), technology (ease of use, robustness, scalability,<br />
security, performance, etc.), support (portability, cost,<br />
ISP support), and how well it works in the real world”<br />
Holt et al. (1997) Use the criteria that it should be appropriate for introducing<br />
programming concepts used in the real world<br />
such as in business, science and government; and it<br />
Kolling and<br />
Rosenberg<br />
(1996)<br />
Hadjerrouit<br />
(1998)<br />
should encourage systematic problem solving<br />
Suggest clean concepts directly reflecting the theoretical<br />
model and avoidance of conceptual redundancy as<br />
achieving the same thing in a variety of ways can<br />
mean flexibility to the expert, but usually confuses be-<br />
ginners<br />
Evaluates the suitability of Java as a first programming<br />
language using the criteria like: programming concepts<br />
to be taught including problem solving skills, algorithmic<br />
thinking and structured programming; use in<br />
subsequent courses and programming paradigm support.<br />
3. Research design and methodology<br />
Also serve as example<br />
towards multiple factors<br />
used in the comparisons<br />
This serves as for this<br />
study’s evaluation based<br />
on ‘purpose’ which is the<br />
learning of Web tech-<br />
niques<br />
This serves as a pointer<br />
into the formulation of<br />
criteria to be established<br />
Also serves as a pointer<br />
into the formulation of<br />
criteria to be established<br />
Emphasize the need for<br />
the suitability for doing<br />
the job for which a tool is<br />
needed.<br />
An interpretive research design is used involving elements of descriptive, analytical and comparative<br />
studies. McMillan and Schumacher (2001:33) state that while a descriptive study describes a system with<br />
the aim of characterizing it as it is, by using numbers, comparative study investigates the differences,<br />
thereby taking descriptive study a step further. Therefore, a descriptive approach is used to characterize<br />
desirable features of the platforms towards establishing a set of criteria. In applying the established criteria,<br />
an analytical approach is used to analyze and evaluate platforms to determine the level of satisfaction<br />
of the criteria for learning Web techniques.<br />
3.1 Research methods adopted to obtain the results<br />
The descriptive method used involves document review and study of various manuals and textbooks for<br />
different platforms from various sources and established body of knowledge, including those written by<br />
the designers and originators of the platforms as well as various websites for the applicable web servers<br />
such as IIS and tomcat, to identify features that could enhance the learning of Web techniques in Web<br />
applications development. These features were characterized to establish criteria for evaluating the suitability<br />
of platforms for learning Web techniques. The established criteria serve as a model. Bowling<br />
(2002:141) describes models as abstract representations of the essential characteristics of phenomena<br />
of interest that make explicit, the relationships and comparison between the characteristics. The model is<br />
then applied for analyzing, evaluating and comparing specific platforms.<br />
The analytical method is aimed at evaluating specific platforms to ascertain their level of satisfaction of<br />
the established criteria. Therefore, documents were reviewed for the specific platforms to be evaluated<br />
and compared, and answers were sought to the questions and the availability of features that serve as<br />
the criteria.<br />
495
Johnson Dehinbo<br />
Moreover, experience was used to physically observe the support for various Web techniques by the platforms.<br />
Bowling (2002:358) describes observation as a research method in which the investigator systematically<br />
watches, listens to and records the phenomenon of interest. Scores were then assigned to each<br />
platform based on the availability of desired features, the level of support enabled on specific tasks or the<br />
inherent characteristics of the platform.<br />
3.2 Measuring scale used<br />
Using close-ended "Yes/No" questions, the measuring tool has values on a scale of 1 to 3, where:<br />
3 = "Yes",<br />
2 = "Not quite or with some workaround", and<br />
1= "No".<br />
A scale of 1 to 3 avoids subjective situations where it could be difficult to distinguish between, for example,<br />
a score of 3 or 4 in a scale of 1 to 5.<br />
3.3 Establishing reliability and validity of the descriptive inquiry<br />
It is important that a measuring scale or instrument be consistent and reliable. It should produce more or<br />
less the same accurate results every time it is applied (Coertze & Heath 1997: 78). Also, Coertze and<br />
Heath (1997:79) indicate that validity is concerned with soundness or the effectiveness of the measuring<br />
instrument. As a way of increasing validity, answers to the criteria questions were sought from established<br />
and recognized sources. The accompanying references are provided for verification or for more<br />
information. This is supplemented with practical experiences and program tests confirming the satisfaction<br />
of some of the criteria established.<br />
Also, to increase reliability, the quantitative characterization and evaluation using numbers would enhance<br />
clarity in the choice of platform with the highest score. This is unlike just using qualitative sentences<br />
to evaluate the platforms, at the end of which it could be difficult to say which platform is really<br />
more suitable. Furthermore, reliability is enhanced with the scale of 1 and 3 thereby avoiding subjective<br />
situations of distinguish between, say, a score of 3 or 4 in a scale of 1 to 5.<br />
3.4 Data analysis<br />
The data analysis was done using simple statistical parametric analysis, such as sums and means. The<br />
scores for all the criteria were summed up for each web based dynamic platform to obtain a total score<br />
from which the platform with the highest overall score can be identified.<br />
3.5 Limitations<br />
As an interpretive study, a particular limitation is that of the subjective view of the researcher. However,<br />
users can add/remove certain criteria or change their weights.<br />
4. Development of the criteria for the evaluations<br />
It is important that a suitable platform satisfies the purpose for which it is needed which in this case is the<br />
learning of Web techniques. Therefore, important Web techniques are first identified before establishing<br />
desired qualities and questions that serve as criteria for evaluating the suitability for implementing and<br />
learning these concepts.<br />
4.1 Important web techniques<br />
Web applications development is primarily about creating dynamic Web applications. However, dynamic<br />
contents are usually surrounded by static contents. This necessitates learning the generation of both<br />
static and dynamic contents. This involves techniques such as Web page serving, HTTP Basics, Server<br />
information, Form processing, as well as maintenance of Session and Application states.<br />
Web page serving<br />
This involves the creation, loading and redirection of Web pages. Students need to learn:<br />
496
Johnson Dehinbo<br />
How to install “normal HTML codes” on the Web server? Which directory the “HTML files” will be<br />
stored in order for the Web page to be accessible with appropriate URL addresses?<br />
How to access Web pages from the browsers? How the appropriate folders/directories are mapped<br />
relative to the default Web root directory. How appropriate URLs are mapped to the installed application/HTML<br />
files in relative folders?<br />
How to redirect Web pages to other pages? How pages are linked in HTML and how Web forms call<br />
the necessary programs to preprocess the form data?<br />
HTTP basics<br />
The Web runs on HTTP which governs how Web browsers request files from Web servers and how the<br />
servers send the files back. A basic understanding of HTTP is necessary to understand the various Web<br />
techniques. This includes knowledge of the HTTP request messages; its content, which includes some<br />
header information, and optional body; the Web server’s response message including some header information<br />
and usually a body (Lerdorf & Tatroe 2002:158).<br />
Server information<br />
Much information about the Web server and the client will usually be made available to the Web applications<br />
for effective Web management. Although some server information are obtainable from the response<br />
header, students need to learn how further information can be obtained from the server. Such further information,<br />
according to Lerdorf and Tatroe (2002:160-161) includes the hostname or IP address, the<br />
server port number, the fetch method (GET/POST), the path information (e.g.<br />
“…/students/science_faculty”), the script/application/program name useful for self-referencing scripts, the<br />
query string in the URL (e.g., “name=Johnson+age=38”), the client PC’s name/IP_address, the authentication<br />
method used, etc.<br />
Form processing<br />
Using form data moves data from a Web page to a server-side program (Hall, 1999). Students need to<br />
learn many tricks and techniques for working with forms:<br />
Attaching information as a query string at the end of the URL after a question mark, for GET requests,<br />
or sending it on a separate line, for POST requests (Sun Microsystems 1997).<br />
The significance of using POST rather than GET, with respect to the security of the transmitted data.<br />
Techniques of implementing self-processing pages that both generate the form and process it.<br />
Methods of extracting the sent information from form data which differs for various Web applications<br />
platforms, along with the ways of converting one type of data to another.<br />
File Uploading and how to handle the uploaded files?<br />
Form Validation<br />
It is necessary to validate entered data before using or storing it for later use. Several strategies are<br />
available for validating data. One involves using JavaScript or VBScript on the client side. However, since<br />
the user can choose to turn JavaScript or VBScript off, or may even be using a browser that does not<br />
support it (Lerdorf & Tatroe 2002:173), a more secure choice is to use the dynamic Web platform to do<br />
the validation.<br />
Students therefore need to learn how to:<br />
Check that a value was supplied: i.e. form variables are non-empty.<br />
Check that an email address is valid by requiring an “@” sign, checking for the presence of period “.”<br />
in the domain name, checking the last two or three digits of the domain name against the country<br />
codes and organization domains respectively.<br />
Check that a supplied filename is local and exists. This can be ensured by allowing the “browsing” of<br />
the filename on the required media.<br />
Validate a field to ensure that it contains a nonnegative integer.<br />
Ensure correctness of the comparison of a form value with other values by trimming leading/trailing<br />
spaces, converting to lowercase/uppercase etc.<br />
497
Maintaining Session and Application states<br />
Johnson Dehinbo<br />
HTTP is a stateless protocol (Lerdorf & Tatroe 2002:173; Sun Microsystems, 2002, Module 8: 3), which<br />
means that once a Web server completes a client’s request for a Web page, the connection between the<br />
two ceases to exist. Thus, there is no way for a server to recognize that a sequence of requests all originate<br />
from the same client.<br />
State is useful though. To build a shopping cart application for example, it is necessary to keep track of a<br />
sequence of requests from a single user (Lerdorf & Tatroe 2002:178). According to Hall (1999), maintaining<br />
states can help the server to provide visitors with a number of conveniences: Firstly, it enables identifying<br />
a user during an e-commerce session. The user selects and adds items to the shopping cart, and<br />
then continues shopping. Since the HTTP connection is closed after each page is sent, how does the<br />
store know that the user is the same user who put the previous item in his cart? Maintaining states using<br />
various techniques is a way of accomplishing this.<br />
Secondly, it is inconvenient to remember the username and password for many sites. An alternative is for<br />
the server to give a unique user-ID, and when the client reconnects at a later date, the user-ID is returned,<br />
which is verified by the server. Thirdly, it allows customizing a site and lastly, it enables the focusing<br />
of advertisement for displaying "directed" advertisements than "random" ones. This is possible if the<br />
server “remembers” clients’ interests.<br />
Techniques are available to get around the Web’s lack of state, and to keep track of state information<br />
between requests (i.e. session tracking). These include using Hidden form fields, Cookies, and URL rewriting.<br />
Hidden form fields can be used to pass around information using unique identifier assigned. With<br />
URL rewriting, where every local URL on which the user might click is dynamically modified to include<br />
extra information (Lerdorf & Tatroe 2002:178) such as a user-identification or session ID. A cookie is a bit<br />
of information that the server can give to the client. On every subsequent request, the client will give that<br />
information back to the server, thus identifying the client (Hall 1999: Lerdorf & Tatroe 2002:179). In using<br />
cookies, students will need to learn: How to retrieve the Session object from the request object passed?;<br />
How to store the Session attributes?; Accessing the Session attributes?; Destroying the Session; Tedious<br />
details, problems and limitations of Session management.<br />
Session and Application object store state information for use by script files and can also be used to<br />
share information among all users of a given application. One can therefore use Application variables to<br />
store global information, such that different users can modify Application variables simultaneously (MSDN<br />
2002: Module 5: 18). Students will need to learn how to create, lock, store and unlock the Application objects.<br />
4.2 Criteria for evaluating the suitability for teaching Web techniques<br />
From the concepts above, this section now establishes the criteria to ensure that Web applications development<br />
platforms could handle Web techniques with the following features:<br />
Facility for static Web page serving by the associated Web server: this is necessary to illustrate the<br />
use of the platform for static Web contents.<br />
Facilities for extracting information from the HTTP request header: this indicates the use of the request<br />
header as a medium for conveying useful information.<br />
Facilities to embed or set information into the HTTP response header: this illustrates the use of the<br />
response header to convey useful information.<br />
Provision of other relevant information by the Web server: this illustrates the wealth of information<br />
available through the server.<br />
Availability of default folders for placing program codes: these will be easier for beginners to use before<br />
they have mastered the art of mapping to various directories.<br />
Possibilities of sending information via the service method that could implement both GET and POST<br />
methods: this will ensure that the rest of the codes are not affected by changes between GET and<br />
POST methods.<br />
Facilities for form validation, e.g. checking for non empty fields, range of values, content of fields, and<br />
using hidden variables to check if form has been filled once, etc.; this will illustrate the importance of<br />
form validations.<br />
498
Johnson Dehinbo<br />
Facilities for identifying users in a session, e.g. using cookies, hidden variables or URL rewriting to<br />
illustrate the techniques used to overcome HTTP’s statelessness.<br />
Possibility of setting “age” for cookies thereby making cookies last longer or alternatively disallowing<br />
cache of pages.<br />
Facilities for maintaining Application objects: this illustrates the techniques for maintaining global Application<br />
states.<br />
5. Summary of the results of applying the criteria<br />
Application of the established criteria involves evaluating the suitability of four platforms namely Java<br />
Servlets, JSP, ASP and PHP for learning the Web techniques. Table 2 below gives the scoring for the<br />
platforms based on the criteria. From the table, it is obvious that most of the Web techniques features are<br />
supported by all the platforms, possibly because that is the major goal of all platforms. In order to serve<br />
static information that accompanies dynamic contents, the associated Web servers all have facilities for<br />
static Web page serving.<br />
The HTTP headers serve to pass valuable information, including error status, and as such all the platforms<br />
have facilities to extract information from the HTTP request-headers, and to embed information into<br />
the HTTP response-headers (Lerdorf & Tatroe 2002:175-176). All the platforms have methods of obtaining<br />
relevant information about client and server.<br />
Table 2: Scoring for the platforms based on the criteria on Web techniques.<br />
Criteria questions Servlet JSP ASP PHP<br />
1 Are there facilities for static Web page serving by the associated<br />
server?<br />
3 3 3 3<br />
2 Are there facilities to extract information from the HTTP request<br />
headers?<br />
3 3 3 3<br />
3 Are there facilities to embed or set information into the HTTP re- 3 3 3 3<br />
sponse headers?<br />
4 Are there ways of obtaining relevant information from the server? 3 3 3 3<br />
5 Is there availability of default folders for placing program codes? 3 3 3 3<br />
6 Is there the possibility of sending information via a service method<br />
that could implement both GET and POST methods?<br />
7 Do numeric values accepted from a Web form retain their numeric<br />
type instead of being in string format which still has to be recon-<br />
verted ?<br />
8 Are there simple facilities or functions for form validation, e.g.<br />
checking non-empty fields, range of values, field contents etc.?<br />
9 Are there facilities for identifying users in a session, e.g. using<br />
cookies, hidden variables or URL rewriting?<br />
10 Is there the possibility of setting the age of cookies thereby making<br />
cookies last longer?<br />
11 Are there facilities for maintaining application objects so as to maintain<br />
global application states?<br />
12 Pedagogically, is the platform good for teaching and learning Web<br />
techniques?<br />
TOTAL<br />
Scale: 3 = "Yes", 2 = "Not quite or with some workaround", and 1= "No".<br />
3<br />
1<br />
2<br />
3<br />
1<br />
2<br />
3<br />
3<br />
3<br />
3<br />
3<br />
2<br />
3 3 3 3<br />
3 3 3 3<br />
Novice Web developers would prefer putting their codes in default folders rather than creating and mapping<br />
new folders. All the platforms have default folders for this purpose.<br />
Experience shows that passing information via GET methods embeds the values along with the URL and<br />
as such is useful for debugging. But for good security, POST does not show the information on the URL.<br />
All the platforms allow for the sending of information via a service method that could implement both GET<br />
and POST methods.<br />
499<br />
2<br />
2<br />
31<br />
2<br />
2<br />
31<br />
3<br />
3<br />
36<br />
2<br />
3<br />
34
Johnson Dehinbo<br />
When numeric values are accepted from a Web form, it is better that they retain their numeric type instead<br />
of being in String format which still has to be reconverted. This is true in all the platforms except<br />
Java-based ones where form values that are not Strings must be parsed because they are all passed as<br />
Strings (Sebesta 1996:435). One has to use methods such as parseInt() to convert back to integer (Wigglesworth<br />
2000:821). However, experience has shown that in the case of ASP, when the operator used<br />
on the variables is an addition operator “+”, it always assumes the String concatenation operator “+”,<br />
thereby giving results such as “2+5 = 25”. ASP similarly provides the conversion functions such as CInt()<br />
to explicitly convert to integer (Deitel et al. 2001: 741) in such situations.<br />
The fact that the need for the explicit conversion explained above arising due to confusion between addition<br />
and String concatenation is not necessarily applicable in the case of other arithmetic operators such<br />
as subtraction “-”, multiplication “*” and division “/” makes such problems difficult to identify. If experienced<br />
programmers have difficulties in identifying such problems, one can imagine how long it will take<br />
students to do the same.<br />
Validation of values accepted from Web forms is important to ensure the integrity of the processing. In<br />
the Java-based platforms, one can use the length() method to determine whether the field is empty and<br />
one can “catch the NumberFormatException” if the entered value is non-numeric (Wigglesworth 2000:<br />
821). ASP has simple functions such as IsDate(), IsNumeric(), and IsEmpty() to assist in the validation<br />
(Deitel et al. 2001: 735). Also, there are facilities or functions for form validation such as checking nonempty<br />
fields, range of values, content of fields etc. in the PHP platform (Lerdorf & Tatroe 2002:173), but<br />
not as simple as in ASP.<br />
To remedy HTTP’s statelessness, all the platforms have facilities for identifying users in a session via<br />
cookies, hidden variables or URL rewriting, and there is the possibility for setting the age of cookies to<br />
make them last longer (Lerdorf & Tatroe 2002:178-179). In addition, all the platforms have ways of maintaining<br />
global Application states. Pedagogically, ASP and PHP, from experience, seem to be more suitable<br />
for teaching and learning Web techniques due to the simplicity of their commands and functions. In<br />
summary, applying the established criteria reveals that ASP has the highest score followed by PHP and<br />
then the Java-based platforms. It should however be noted that the scores are subjectively based on author’s<br />
knowledge and experience as well as the current design of the platforms. They are thus subject to<br />
change. Therefore, the emphasis of this study is on the establishment of the criteria that enables the selection<br />
among the platforms. Users can thus adapt these criteria to their own taste and can also test or<br />
apply the criteria on other platforms.<br />
6. Conclusions<br />
It is important to ensure that a platform selected for learning Web applications development is suitable for<br />
implementing and learning Web techniques. This is because Web development platforms differ from the<br />
general programming languages in their support for Web techniques used for generating both static and<br />
dynamic Web contents. By studying the features of many platforms, features that are desirable for supporting<br />
the Web techniques in Web applications are identified. These enabled the establishment of criteria<br />
to determine the suitability of the platforms for implementing and learning Web techniques. While ASP<br />
was found to be slightly more suitable on applying the criteria, the significance of the study lies in the establishment<br />
of a comprehensive but specific set of criteria that can be used as a scientific basis for selection.<br />
References<br />
Apte, V., Hansen, T. and Reeser, P. (2003) “Performance comparison of dynamic Web platforms.” Computer Communications.<br />
26 (8) pp. 888 – 898.<br />
Ashenfelter, J.P. 1999. Choosing a Database for Your Web Site, USA : Wiley Computer Publishing.<br />
Bishop, J. and Hurter, R. 1999. Competitors to Java: Scripting languages. In: Proceedings of the South African<br />
Computer Lecturers Association (SACLA) conference. June 1999. Golden Gate, South Africa. [Online]. Available<br />
WWW: http://www.cs.up.ac.za (Accessed 12/03/2009).<br />
Bowling, A. 2002. Research methods in health. 2 nd edition. UK: Open University Press.<br />
Cecchet, E., Chanda, A., Elnikety, S., Marguerite, J. & Zwaenepoel, W. 2003. Performance Comparison of Middleware<br />
Architectures for Generating Dynamic Web Content. Lecture Notes in Computer Science: Middleware.<br />
2672 (2003): 242-261.<br />
Coertze, D. and Heath, R. 1997. Research Methodology for Technikon Students: A practical Approach. South Africa :<br />
Technikon Natal publishing.<br />
Cooper, R. 2001. Software for managing Web sites. In: Proceedings of the South African Institute of Computer Scientists<br />
and Information Technologists (SAICSIT) annual conference. September 2001. Pretoria: South Africa.<br />
500
Johnson Dehinbo<br />
Deitel, H.M., Deitel, P.J. and Nieto, T.R. (2001) e-Business & e-Commerce: how to program. Prentice Hall, New Jersey.<br />
Hall, M. 1999. Tutorial on Servlet and JSP [Online]. Available from: http://www.apl.jhu.edu/~hall/java/Servlet-Tutorial/<br />
[Accessed: 5/06/2009].<br />
Hartman, H. 2001. Tools for dynamic Web sites: ASP vs PHP vs ASP.NET, Seybold Report Analysing Publishing<br />
Technologies, 15339211. 1 (12).<br />
Holt, R., Wortman, D., Barnard, D. and Cordy, J.R. (1977) “SP/k: a system for teaching computer programming.”<br />
Communications of the ACM. 20 (5) pp. 301-309.<br />
Kolling, M. and Rosenberg, A. (1996) “Blue: A language for teaching object-oriented programming.” Proceedings of<br />
the 27th SIGCSE Technical Symposium on Computer Science Education, Philadelphia. New York. pp.190-194.<br />
Kruse, W. (2003) “A comparison of PHP and J2EE.” Technologiae. 1 (2003) pp. 110-117.<br />
Lerdorf, R. and Tatroe, K. 2002. Programming PHP: Creating Dynamic Web Pages. USA : O’Reilly & Associates Inc.<br />
Lim, B.L. 2002. Teaching web development technologies: Past, present, and (near) future. Journal of Information<br />
Systems Education. 13 (2): 117-123.<br />
Marshak, M. and Levy, H. 2003. Evaluating web user perceived latency using server side measurements. Computer<br />
Communications. 26 (8): 872-887.<br />
McMillan, J.H. and Schumacher, S. 2001. Research in education. 5 th edition. USA: Addison Wesley Longman, Inc.<br />
MSDN. 2000. Introduction to Web Development Technologies. Course number 1912 Workbook. Part Number: X05-<br />
91011.<br />
Prechelt, L. 2000. An Empirical Comparison of Seven Programming Languages. Computer. 33 (10): 23-29.<br />
Renaud, K., Lo, J., Bishop, J., Van Zyl, P and Worrall, B. 1999. Algon: A framework for supporting comparison of<br />
distributed algorithm performance, In: Proceedings for PNDP ‘03 conference, February 2003. Genoa: Italy.<br />
Sebesta, R.W. 1996. Concepts of Programming Languages, 3 rd edition. USA: Addison-Wesley Publishing Company.<br />
Sun Educational Services. (2002) “Web Component Development With Java Technology.” SL-314 Student Guide.<br />
Revision A.1, Sun Microsystems, New York.<br />
Sun Microsystems. 1997. JDBC Guide: Getting Started [Online]. Available from:<br />
http://www.pearson.com.mx/component/component_java_21_days/jdk/1.2/doc/html/guide/jdbc/getstart/introTO<br />
C.doc.html [Accessed: 11/08/2008].<br />
Vinoski, S. (2003). “The performance presumption.” IEEE Internet Computing. 7 (2) pp. 88-90.<br />
Wiedenbeck, S., Ramalingam, V., Sarasamma, S. and Corritore, C.L. 1999. A comparison of the comprehension of<br />
object-oriented and procedural programs by novice programmers. Interacting with Computers. 11 (3): 252-282.<br />
Wigglesworth, J. and Lumby, P. 2000. Java Programming Advanced Topics. U.K.: Thompson Learning Course<br />
Technology.<br />
501
Human Resources Transformation Beyond Boundaries in<br />
Outsourcing Business Models - Expatriate Benchmarking<br />
Swathi Duppada 1 and Rama Chandra Aryasri 2<br />
1<br />
Satyam Computer Services Limited (Mahindra Satyam), Canada<br />
2<br />
School of Management Studies, JNTU, Hyderabad, India<br />
gsrkmedu@gmail.com,<br />
aryasri@yahoo.com<br />
Abstract: Human Resource (HR) divisions of multi national companies (MNCs) are under tremendous pressure<br />
globally with the challenges and opportunities with the outsourcing business models to maintain competitive position<br />
in the marketplace. Attracting the mobile talent with multi dimensional skill set to address effective, efficient and<br />
controllable business needs is becoming complex, hence expatriate management and training has gained much<br />
attention. Successful Expatriate assignments drive revenue, value and growth to the organization. The expatriation<br />
process requires huge amount of effort for analysis, planning, selection, training before departure of the associate to<br />
host country. The authors would like to bring the practical approaches that need to be considered in global business<br />
outsourcing model considering 3 dimensions – Associate delight, Customer delight and Investor delight with<br />
expatriate benchmarking. The research study also brings the expatriate management strategies in 3 categories –<br />
Onsite (foreign location / host country), Offshore (home country) and Near-shore (country close to host country, but<br />
with lesser delivery cost). To substantiate the research, the data is collected from several Human resource leaders<br />
and managers at various levels – HR Executives, HR Analysts, HR Managers, Senior Managers from IT<br />
organizations in different geographies through interviews and web based surveys. Statistical analysis are conducted<br />
on the data collected and the data analysis first revealed that 1) the organizations with good global management<br />
strategy had larger number of associates with better expatriate experience, steadier focus on leadership and resulted<br />
in better financials 2) The training is more focused on technology and job related skill set, but often ignored the level<br />
of depth in imparting the behavioral and cultural skills. The main contributions of the paper are expatriate training<br />
business drivers in the context of IT global outsourcing, proposing and benchmarking of Expatriate Canonical model<br />
framework, recommendations on delivery models with expatriate management with regard to IT outsourcing<br />
companies.<br />
Keywords: benchmarking, expatriate management, expatriate training, ROI, onsite, offshore, near-shore<br />
1. Introduction<br />
Technological advancement and innovation in the global business landscape is resulting in lot of<br />
challenges and opportunities for the IT industry. Driven by these challenges and opportunities,<br />
organizations are moving across national borders to align the business need to be in competitive market<br />
place to capitalize the wonderful opportunities and establishing foreign-owned subsidiaries, entering into<br />
joint ventures and strategic alliances. In today’s agile market conditions, the C-level executives are<br />
challenged to grab the opportunities, making referenceable customers to sustain in the business with<br />
human talent. To accomplish these goals, the executives are looking for new ways and models to exploit<br />
the current and next generation growth opportunities. National culture has received so much attention<br />
because it is hypothesized to be a constraint on management practice and organization culture (Barry,<br />
2008).To achieve global delivery excellence, organizations are reshaping the traditional business models,<br />
adopting the outsourcing models, deploying and utilizing its global human talent through expatriate<br />
assignments as platform to manage and win the business overseas. International business assignments<br />
are important to the success of the MNCs because they help to build the level of global competence and<br />
global integration within the organization [Paula MaryAnne M and Aparna, 1998). Due to expatriate<br />
assignments by multinational companies, language and mobility parameters have significantly impacted<br />
across the globe. As the world continues to globalize, firms are required to manage an increasingly<br />
diverse workforce with expatriation being just a subset of this challenge.<br />
Global expansion strategies are essential for small, mid and large sized IT organizations to expand and<br />
sustain in the business by recruiting and retaining globally competent people and utilize their managerial<br />
and technical capabilities for a competitive advantage. Expatriate management nowadays is regarded as<br />
one of the tools international organizations use to enhance global integration (Evans, Pucik, and Barsoux,<br />
2002). To lead these international operations successfully, flexible and highly skilled workforce with<br />
global mindset and strategic vision is needed and as such the expatriate Managers are on boarded.<br />
Successful Expatriate assignments drive revenue, value and growth to the organization and have always<br />
been a driving and determining factor for the success of the organization’s overseas operations. These<br />
expatriate managers become the cornerstones on which the international operations are built as they<br />
502
Swathi Duppada and Rama Chandra Aryasri<br />
help in informal control mechanism, for knowledge transfer, and for international team development.<br />
Hence, an emerging view of foreign assignments is that it may create a unique competitive advantage<br />
which is difficult to imitate by competitors.<br />
The research study revealed that the multinational organizations are structured and governed based on<br />
their working model and relationships. There are several types of global assignments that exist based on<br />
the business need and trends, hence the expatriates will be sent to foreign locations with different<br />
duration types – Short term, mid term and long term. Based on the survey analysis, most of the<br />
multinational companies classify the short term assignments duration may include 1-3 months, midterm<br />
assignments duration may include 3-6 months and long term duration may include more than 6 months.<br />
In this paper, the research is focused on outsourcing business model in IT organizations and expatriates<br />
transformation.<br />
2. Literature survey<br />
According to CIPD/KPMG labor market outlook report on skills, migration and off-shoring, almost half<br />
(45%) of the 600 employers surveyed report vacancies that are hard to fill, with 21% saying they are<br />
recruiting migrant workers for engineering vacancies, and 18% for both IT and accountancy/finance<br />
positions. As a result, almost one-fifth (17%) intend to recruit migrant workers in the third quarter of 2010.<br />
In the past three months, one in five (21%) of employers surveyed recruited migrant workers with over a<br />
third (37%) of these workers being recruited from outside the <strong>European</strong> Economic Area (EEA). More than<br />
half of migrant workers hired by the financial sector come from outside the EEA. The findings also point<br />
to greater off shoring activity. Nearly one in ten (9%) private sector companies plan to offshore jobs in the<br />
12 months to June 2011. Of those planning to offshore UK jobs, two-thirds (65%) intend to offshore to<br />
India, a third to China (36%) and three in ten to eastern Europe (29%). The most common functions<br />
outsourced by employers include call centres (55%), IT (51%), and finance (49%) (KPMG & CIPD, 2010).<br />
A fundamental challenge faced by multinational companies today is how to ensure that managers<br />
develop not only an overview of the organization in its entirety, but also a feel for international business<br />
(Gooderham & Nordhaug, 2003). A lot of research still continues to view the globalization topic is a work<br />
in progress research. Only a few firms are considered to have developed an effective capability to locate,<br />
source, manage human resources any where in the world (Lewin and Volberda, 2003), and multi national<br />
enterprises continue to have assets, sales, ownership of work forces and control concentrated in home<br />
countries or regions (Rugman & Verbeke, 2004),). The study of global staffing has traditionally<br />
concentrated on resourcing key positions within Multinational enterprises and top management top key<br />
positions at Head Quarters and subsidiary locations, generally idiosyncratic mixes of strategy rather than<br />
any logical progression of focus related to process of globalization Harzing, 2001, Harzing & Van<br />
Ruysseveldt, 2004) . The literature gives most attention to recruitment and selection of expatriates and<br />
international managers, talent management at HQ or local level, and more flexible forms of international<br />
business travelers, virtual teams and impatriates (Scullion & Collings, 2006). Briscoe and Schuler noted<br />
that the definition of international employee inside organizations continues to expand: “… the tradition of<br />
referring to all international employees as expatriates – or even international assignees – fall short of the<br />
need for international HR practitioners to understand the options available…and fit them to evolving<br />
international business strategies [(Briscoe and Schuler, 2004)”. In the context of international resourcing,<br />
this now includes a fragmentary group of individuals, ranging from: contract expatriates (Baruch and<br />
Altman, 2002), assignees on short term or intermediate term foreign postings (Morley and Hearty, 2004,<br />
Mayerhofer, Hartmann, and Herbert, 2004) ; international transferees (moving from one subsidiary to<br />
another) (Millar and Salt, 2006); virtual international employees active in cross-border project teams<br />
(Janssens and Brett , 2006) ; self-initiated movers who live in a third country but are willing to work for a<br />
multinational (Tharenou, 2003); immigrants actively and passively attracted to a national labor market<br />
and domestically based employees in a service centre but dealing with overseas customers, suppliers<br />
and partners on a regular basis . Millar and Salt draw attention to a number of factors that have increased<br />
demand for new forms of international mobility (Millar and Salt, 2006) : the need for skilled expatriates to<br />
help build new international markets (Findlay, Jowett and Skeldon, 2000) ; temporary and short term<br />
access to specialized talent in sending countries to assist the execution of overseas projects (Minbaeva &<br />
Michailova, 2004 and Hocking, Brown and Harzing , 2004), ; and the need for highly mobile elites of<br />
management to perform boundary-spanning roles to help build social networks and facilitate the<br />
exchange of knowledge (Tushman & Scanlan, 2005).<br />
3. Outsourcing delivery patterns in it organizations<br />
According to David J. Bryce and Michael Useem, outsourcing is defined as an outside company’s<br />
provision of the products or services associated with a major function or activity of a user organization<br />
503
Swathi Duppada and Rama Chandra Aryasri<br />
(David & Michael, 1998) . The definition of outsourcing is not yet standardized; however, the outsourcing<br />
term is much widely used in information technology business. According to Wikipedia, outsourcing is<br />
often viewed as involving the contracting out of a business function - commonly one previously performed<br />
in-house - to an external provider (Overby, 2007). In February 2004, the members of the President<br />
Bush’s council of economic advisers stated: outsourcing of professional services is a prominent example<br />
of a new type of trade” (Gregory, Kristin, Harvey, 2004). The following are the successful outsourcing<br />
implementation patterns due to reduced cost, access to most talented resources at cheater cost, 24 X 7<br />
productivity,<br />
3.1 Onsite delivery model / staff augmentation model<br />
In onsite delivery model, the customer/client would request for skilled expatriates from offshore to work at<br />
their location. In nutshell, the customer work will be done by expatriates at client location with lesser cost.<br />
Normally this model will be adopted when the continuous involvement of client is required at every stage<br />
and security of the data and information is required. Also this is best fitted when the requirements are not<br />
very clear. This model is suitable for short term engagements like strategy consulting, due diligence,<br />
architecture consulting, Roadmap Initiatives and will increase the top line revenues of the offshoring firm<br />
providing services to client.<br />
Figure 1: One view of delivery models<br />
3.2 Offsite delivery model / near shore model<br />
When the client needs more number of expatriates to deliver and if the client cannot extend the facilities<br />
to the expatriates, then the client would prefer the offsite delivery model as the expatriates can use the<br />
foreign/subsidiary office to deliver the projects. The only difference from Onsite delivery models is that the<br />
majority of the expatriates will work from the same country as the client location /city/near city. This model<br />
will enable expatriates to come and conduct face to face discussions and is suitable for medium sized<br />
project.<br />
504
3.3 Offshore delivery model<br />
Swathi Duppada and Rama Chandra Aryasri<br />
When the projects need to be done at cheaper / affordable cost, this model is ideally suitable, however<br />
this model requires clear planning and should possess clear objectives. In this model, the project delivery<br />
is accomplished at offshore (which is in different country). This type delivery model requires frequent<br />
communications with telephone calls and emails as there is no face to face communication.<br />
3.4 Onsite-offshore delivery model<br />
In onsite-offshore delivery model, the work is distributed between Onsite (Client location/host country)<br />
and offshore (home country), hence the client/customer gets most benefitted by taking the advantage of<br />
talent and time from offshore and expatriates from the same offshoring partner will be available at onsite.<br />
The normal distribution ratio of offshore and onsite is 70% and 30%. This model is most successful model<br />
adopted by the Multi National companies to deliver projects better and cheaper. This model is preferred<br />
by the customer as they see significant benefits - 24 X 7 productivity, Most talented offshore resources at<br />
cheaper cost, Access to well trained expatriated to communicate on day to day basis and less burden<br />
with the administration issues.<br />
4. Expatriate transformation canonical framework and benchmarking<br />
The suggested Expatriate transformation canonical framework is very useful to the IT organizations in<br />
transforming the offshore talent for global needs. The purpose of this framework is minimizing the<br />
expatriate dependencies across different business units (verticals) and competency units (technology<br />
units) in different geographic locations. The following diagram depicts the expatriate movement between<br />
locations based on the project needs.<br />
Figure 2: Expatriate transformation canonical framework<br />
Notation: Req - Expatriate Request Res Expatriate Response, MNC – Multi-National Company,<br />
APAC – Asia Pacific, T&M – Time & Material<br />
In order to excel in providing best-in-class expatriates, the canonical framework should be used with the<br />
suitable adapters for the transformation of the expatriate request and expatriate response. Most IT<br />
505
Swathi Duppada and Rama Chandra Aryasri<br />
organizations traditionally follow Hub and Spoke resource model for resources. The business units would<br />
communicate the resource requirements to corporate resource pool through the business unit talent<br />
management team. Hence the first transformation takes place from business unit expatriate resource<br />
request message to pool to transform organization protocol / common organization language. The<br />
request has several input parameters like duration, type, skill, role, etc. These parameters will be used in<br />
determining the right expatriate for the assignment request. This framework has the flexibility to add new<br />
types of expatriate requests.<br />
Figure 3: Expatriate movement beyond boundaries<br />
Typically expatriate management/ training and other related activities will have their own business<br />
process and each of these processes will have process steps. Most likely each process step will deliver<br />
some kind of services. These services may internally use all sorts of formats; hence there is a need to<br />
transform the messages to a common message format.<br />
5. Expatriate process benchmarking, tool and analysis<br />
According to Gartner survey report in December 2010, India is on the top list for offshore services<br />
(Gartner, 2010). To investigate the Expatriate management and training practices for HR excellence to<br />
meet the on demand needs, the authors have selected the MNCs having operations across the globe.<br />
The extensive research on the organizations is conducted to study the patterns and practices of<br />
expatriation. The following are samples of it some of the organizations reveal the need to study, analyze,<br />
benchmark the expatriate practices based on the annual reports.<br />
100.00%<br />
90.00%<br />
80.00%<br />
70.00%<br />
60.00%<br />
50.00%<br />
40.00%<br />
30.00%<br />
20.00%<br />
10.00%<br />
0.00%<br />
Cognizant Revenu Chart for past 4 years<br />
79.10% 79.10%<br />
18.50%<br />
19.20%<br />
82.82%<br />
16.06%<br />
86.19%<br />
12.91%<br />
2.40% 1.70% 1.12% 0.90%<br />
2009 2008 2007 2006<br />
Year<br />
Figure 4: Cognizant revenue contributions (Cognizant, 2009)<br />
506<br />
NA<br />
EUROPE<br />
APAC
% of Revenue<br />
70.00%<br />
60.00%<br />
50.00%<br />
40.00%<br />
30.00%<br />
20.00%<br />
10.00%<br />
0.00%<br />
Swathi Duppada and Rama Chandra Aryasri<br />
65.80%<br />
63.20%<br />
Infosys Revenue chart for past 2 years<br />
26.40%<br />
23.00%<br />
1.20% 1.30%<br />
10.00% 9.10%<br />
North America Europe India ROW<br />
Region<br />
Figure 5: Infosys revenue contributions (Infosys, 2009<br />
60<br />
30<br />
20<br />
10<br />
0<br />
52.8<br />
51.38<br />
50<br />
% of Revenue<br />
40<br />
16.18<br />
TCS % of Revunue for the Y 2009-10 and 2009-08<br />
18.99<br />
10.49 10.53<br />
8.65 7.85<br />
5.24 4.75 4.72 4.71<br />
1.92 1.79<br />
North America UK Europe India APAC Iberoamerica Middle Ease<br />
and Africa<br />
Region<br />
Figure 6: TCS revenue contributions (TCS, 2009) Figure 7: Wipro revenue contributions (Wipro, 2009)<br />
Figure 8: Expatriate process benchmarking<br />
2009-10<br />
2008-09<br />
507<br />
% of Revenue COntribution<br />
70%<br />
60%<br />
50%<br />
40%<br />
30%<br />
20%<br />
10%<br />
0%<br />
2009-10<br />
2008-09<br />
Wipro Revenue chart for past 2 years<br />
2009-10<br />
2008-09<br />
North America Europe Japan ROW<br />
Region and Year
Swathi Duppada and Rama Chandra Aryasri<br />
For these interviews, companies that previously had addressed or involved in addressing the<br />
transformation of their HR functions were selected. A 3-step process is used to conduct the study on the<br />
expatriate processes with multi channel methods. As a first step, the details of the existing Expatriate<br />
Management and training practices and models of the organizations are obtained. The data collection<br />
analysis activity began by identifying the benchmarking partner followed by making available the online<br />
survey weighted questionnaire about their expatriate life cycle process at their organization.<br />
The questionnaire includes – pre-assignment (to capture the drivers of assignment from the perspective<br />
of assignment key stakeholders), on-assignment (to monitor the deliverables and the progress of the<br />
assignment) and post assignment (to help assignment success and synergy to the organization). The<br />
criterion that was set for organizations to be a benchmarking partner is that the organization should have<br />
at least global presence, specifically in North America (NA), Europe and Asia-Pacific which includes<br />
India, in conducting IT business.<br />
Table 1: Normalized scores of the detailed survey<br />
BENCHMARKING PARTNER<br />
Benchmarking Framework Attribute 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17<br />
Structure of the organization 0.87 0.82 0.81 0.86 0.86 0.9 0.83 0.86 0.82 0.89 0.82 0.89 0.84 0.87 0.86 0.82 0.86<br />
Expatriate Request transformation 0.84 0.85 0.79 0.85 0.86 0.9 0.84 0.84 0.84 0.87 0.82 0.84 0.85 0.86 0.85 0.84 0.86<br />
Expat Planning Automation 0.83 0.82 0.79 0.82 0.84 0.87 0.82 0.82 0.82 0.84 0.79 0.83 0.84 0.84 0.84 0.81 0.83<br />
Expat Screening Automation 0.9 0.69 0.67 0.75 0.69 0.72 0.68 0.76 0.68 0.74 0.68 0.85 0.69 0.79 0.69 0.68 0.77<br />
Expat Interviewing and Automation 0.87 0.73 0.71 0.76 0.74 0.69 0.74 0.76 0.73 0.73 0.73 0.87 0.72 0.74 0.74 0.73 0.71<br />
Expatriate Selection Automation 0.81 0.87 0.84 0.86 0.89 0.91 0.87 0.84 0.87 0.92 0.85 0.81 0.89 0.89 0.83 0.87 0.85<br />
Immigration Process Automation 0.74 0.84 0.82 0.91 0.85 0.89 0.84 0.88 0.83 0.89 0.83 0.79 0.85 0.88 0.85 0.83 0.81<br />
Training Automation 0.89 0.82 0.82 0.81 0.87 0.93 0.83 0.85 0.82 0.91 0.82 0.85 0.84 0.87 0.84 0.82 0.84<br />
Depature/Deputation/Onsite transfer automation 0.81 0.81 0.81 0.79 0.81 0.88 0.83 0.83 0.81 0.85 0.81 0.81 0.81 0.86 0.81 0.81 0.79<br />
Arrival Support Automation 0.83 0.78 0.76 0.82 0.79 0.92 0.79 0.85 0.76 0.91 0.76 0.83 0.79 0.89 0.79 0.76 0.75<br />
Strategic Partnering 0.82 0.75 0.69 0.78 0.75 0.95 0.75 0.79 0.74 0.82 0.71 0.82 0.74 0.81 0.74 0.71 0.76<br />
Expatriate Service Development 0.78 0.68 0.52 0.75 0.79 0.7 0.69 0.78 0.67 0.81 0.62 0.74 0.71 0.79 0.71 0.62 0.77<br />
Sum of the Normalized Scores 9.99 9.46 9.03 9.76 9.74 10.26 9.51 9.86 9.39 10.18 9.24 9.93 9.57 10.09 9.55 9.3 9.6<br />
Score Ranking 4 13 17 7 8 1 12 6 14 2 16 5 10 3 11 15 9<br />
The above graphs clearly show that the organizations with good recruitment practices stand in top<br />
position; thereby the authors have arrived at the best practices in the next section.<br />
In order to find out the results and observations, the original sample included 100 associates from each<br />
firm were identified. With 69% of participants, the data is collected on technology, managerial and<br />
behavioral skills<br />
Normalized Score<br />
10.4<br />
10.2<br />
10<br />
9.8<br />
9.6<br />
9.4<br />
9.2<br />
9<br />
8.8<br />
8.6<br />
8.4<br />
Sum of the Normalized Scores V/S Benchmarking Partner<br />
B1 B2 B3 B4 B5 B6 B7 B8 B9 B10 B11 B12 B13 B14 B15 B16 B17<br />
Benchmarking Partner<br />
Sum of the Normalized Scores<br />
Normalized Values<br />
18<br />
16<br />
14<br />
12<br />
10<br />
8<br />
6<br />
4<br />
2<br />
0<br />
BP Partner Ranking with Normalized Values<br />
Sum of the Normalized Scores Score Ranking<br />
B1 B2 B3 B4 B5 B6 B7 B8 B9 B10 B11 B12 B13 B14 B15 B16 B17<br />
Benchmarking Partner<br />
Figure 9: Normalized scores with survey results Figure 10: Ranking with survey results<br />
Statistical analysis are performed to obtain the top most recruitment metrics – Full productivity, Quality of<br />
Service, Quality of Promotion and Knowledge Management of expatriates for mid to long term duration.<br />
The results are captured after the research, analysis and screening using Expatriate transformation<br />
framework.<br />
Quality of promotion is an important metric for branding. This parameter will have increased influence of<br />
people, processes and opportunities of the expatriate representing organization.<br />
6. Future trends<br />
In the days to come expatriate processes would be completely modernized and automated. The use of<br />
technology in the sector continues to increase at a remarkably faster pace with technologies hence;<br />
organizations have to be on top of technology and competencies along with the deep domain knowledge<br />
to address the future expatriate resource fulfillment needs.<br />
508
Full Productivity<br />
8<br />
7<br />
6<br />
5<br />
4<br />
3<br />
2<br />
1<br />
0<br />
-1<br />
-2<br />
-3<br />
Average Full Productivity w ith/out ETCFABM<br />
Swathi Duppada and Rama Chandra Aryasri<br />
With ETCFABM Adoption Without ETCFABM Adoption<br />
Intial Offshore EBMF Adoption Onsite Log term Onsite<br />
Time<br />
Quality of Service<br />
9<br />
8<br />
7<br />
6<br />
5<br />
4<br />
3<br />
2<br />
1<br />
0<br />
Average Quality of Service w ith/out ETCFABM<br />
With ETCFABM Adoption Without ETCFABM Adoption<br />
Intial Offshore EBMF<br />
Adoption<br />
Onsite Log term<br />
Onsite<br />
Time<br />
8 - Superior 6- Meritorious 4-Acheivable 2- Reasonable 0-Nornal<br />
Figure 11: Average full productivity Figure 12: Average quality of service<br />
Quality of Promotion<br />
4.5<br />
4<br />
3.5<br />
3<br />
2.5<br />
2<br />
1.5<br />
1<br />
0.5<br />
Quality of Promotion with/out ETCFABM<br />
With ETCFABM Adoption Without ETCFABM Adoption<br />
0<br />
Intial Offshore<br />
Time<br />
EBMF Adoption Onsite<br />
5 - Excellent 4- Very Good 3-Good 2-Reasonale 1-Nornal<br />
Knowledge Management<br />
6<br />
5<br />
4<br />
3<br />
2<br />
1<br />
0<br />
Knowledge Management with/out ETCFABM<br />
With ETCFABM Adoption Without ETCFABM Adoption<br />
Intial Offshore EBMF Adoption Onsite<br />
Time / Locale<br />
5 - Excellent 4- Very Good 3-Good 2-Reasonale 1-Nornal<br />
Figure 13: Average quality of promotion Figure 14: Average knowledge management<br />
Expatriate processes would become more efficient and robust and fully integrated with the overall<br />
organizational objectives. Firms will have to pay high attention to ensure that the expatriate solutions<br />
address the employer branding and regulatory compliances issues. The focus on building passive<br />
candidate pool and diverse recruitment would increase tremendously. High attention will be paid towards<br />
achieving collaboration between the hiring Manager and the recruiter.<br />
References<br />
Barry Gerhart, (2008) Cross Cultural management Research: Assumptions, Evidence and Suggested Directions,<br />
International Journal of Cross Cultural Management.<br />
Baruch, Y. and Altman, Y. (2002) Expatriation and Repatriation in MNCs : a Taxonomy, Journal Human Resource<br />
Management, Pg 239-259<br />
Briscoe, D. and Schuler, R.S. (2004) International Human Resource Management”, Second Edition, New York:<br />
Routledge.<br />
Cognizant Report. (2009) Cognizant Annual Report, http://investors.cognizant.com/index.php?s=52<br />
David J.Bryce and Michael Useem.(1998) The impact of corporate outsourcing on company value, <strong>European</strong><br />
Management Journal, Vol. 16, No.6, pp.635-643.<br />
Evans, P, Pucik, V and Barsoux, J.L, (2002) The Global Challenge: International Human Resource Management,<br />
NewYork: McGraw-Hill/Irwin.<br />
Findlay, A.M., Li, F.L.N., Jowett, A.J. and Skeldon, R. (2000) Skilled international migration and the global city: a<br />
study of expatriates in Hong Kong”. Applied Geography, Pg: 277- 304.<br />
Gartner Report. (2010) Gartner's 30 Leading Locations for Offshore Services, 2010-2011,<br />
http://www.gartner.com/it/page.jsp?id=1502714.<br />
Gooderham, P.N and Nordhaug, O (2003) International Management: Cross-boundary challenge, Oxford, Blackwell<br />
Publishing.<br />
Harzing, A.W.K (2001) Who’s in charge? An empirical study of executive staffing practices in Foreign Multinationals,<br />
Journal of Human Resource Management, 40 139-58.<br />
509
Swathi Duppada and Rama Chandra Aryasri<br />
Harzing, A.W.K and J.Van Ruysseveldt. (2004) Composing and International Staff, International Human Resource<br />
Management, Second Ed.<br />
Hocking, J.B., Brown, M. and Harzing, A.-W. (2004),A knowledge transfer perspective of strategic assignment<br />
purposes and their path-dependent outcomes, International Journal of Human Resource Management, 15 (3):<br />
565-586.<br />
Infosys Report. (2009) Infosys Annual Report, http://www.infosys.com/investors/pages/index.aspx<br />
Janssens, M. and Brett, J.M. (2006), Cultural intelligence in global teams: a fusion model of collaboration, Group and<br />
Organization Management, 31 (1): 124-153.<br />
KPMG & CIPD Quarterly Survey Report, (2010)Labor Market Outlook,<br />
http://www.cipd.co.uk/subjects/hrpract/hrtrends/_qtrends.htm.<br />
Lewin, A.Y. and Volberda, H.W (2003) Beyond Adaptation – Selection Research: Organizing Seld-Renewal in Co-<br />
Mayerhofer, H., Hartmann, L.C. and Herbert, A. (2004) Career management issues for flexpatriate international<br />
staff”, Thunderbird International Business Review, Pg: 647- 666.<br />
Millar, J. and Salt, J. (2006) The mobility of expertise in transnational corporations, Migration Research Unit. Working<br />
Paper. University College London.<br />
Minbaeva, D.B. and Michailova, S. (2004) Knowledge transfer and expatriation in multinational corporations: the role<br />
of disseminative capacity, Employee Relations, 26 (6): 663-679<br />
Morley, M. and Hearty, N. (2004) International Assignment and global careers, Thunderbird International Business<br />
Review, Pg: 633-646.<br />
N. Gregory Mankiw, Kristin J. Forbes, Harvey S. Rosen. (2004) The Economic Report of the President”, Washington,<br />
D.C.<br />
Overby, S. (2007) ABC: An Introduction to Outsourcing, www. CIO.com.<br />
Paula M.Caligiuri, MaryAnne M and Aparna Joshi (1998) Families on Global Assignments: Applying work/family<br />
theories abroad, Current Topics in Management, Volume 3, pages 313-328.<br />
Rugman, A. and Verbeke, A (2004) A perspective on regional and global strategies of multinational enterprises,<br />
Journal of International Business Studies, 35: 3-18.<br />
Scullion, H and Collings, D.G. (2006), “Global Stafing”, London: Routledge.<br />
TCS Report. (2009) TCS Annual Report,<br />
http://www.tcs.com/investors/Documents/Annual%20Reports/TCS_Annual_Report_2009-2010.pdf<br />
Tharenou, P. (2003) The initial development of receptivity to working abroad: self-initiated international work<br />
opportunities in young graduate employees, Journal of Occupational and Organizational Psychology, 76: 489-<br />
515.<br />
Tushman, M.I. and Scanlan, T.J. (2005) Boundary spanning individuals: their role in information transfer and their<br />
antecedents , Academy of Management Journal, 24 (2): 289-305.<br />
Wipro Report. (2009) Wipro Annual Report http://www.wipro.com/corporate/investors/pdf-files/annualreport/wipro_annual_report.pdf.<br />
510
The Diffusion of Mobile Phones for Business and Information<br />
Management in Kenya<br />
Wakari Gikenye<br />
University of Zululand, South Africa<br />
wagikenye@yahoo.com<br />
Abstract: Diffusion of innovation is the process by which an innovation is adopted and gains acceptance by<br />
members of a certain community, innovations that are perceived by individuals as having greater relative advantage,<br />
compatibility and less complexity may be adopted more rapidly than others (Rogers 2003:5 :16). Information and<br />
communication technologies (ICTs) like the computer and the internet have been widely adopted by people in<br />
western countries, while in developing countries they have taken much longer to be adopted. Mobile phones ICT on<br />
the other hand have been widely adopted by both the western and developing countries in the last decade. In Kenya,<br />
for instance, the number of subscribers had reached 20 million by March 2010, and still growing (Communication<br />
commission of Kenya 2010). The rapid spread and use of mobile phone technology has in the last three years been<br />
extended to the use of innovative mobile money transfer services both of which have increased the speed of<br />
economic activities for Kenyans, including the low income people who were not served by financial institutions like<br />
banks and the post office network. Money transfer services were first introduced in Kenya in 2007 in the form of ‘Mpesa’,<br />
a financial transaction medium, which has had phenomenal growth since its introduction and has greatly<br />
impacted economic activities. The M-pesa mobile money transfer service has been described as a revolutionary way<br />
of making cell phone technology work to serve a basic need by providing a fast, safe, cheaper and easy way to<br />
transfer money (Vodacom 2010). This paper examines the rapid diffusion of mobile phones and related mobile<br />
money transfer services in Kenya. It is based on an on-going research study on “The diffusion of ICTs in the informal<br />
sector in Kenya”.<br />
Keywords: diffusion of ICTs, mobile phones adoption, money transfer services, Kenya<br />
1. Introduction and background<br />
Diffusion of innovation is the process in which an innovation is adopted and gains acceptance by<br />
members of a certain community, innovations that are perceived by individuals as having greater relative<br />
advantage, compatibility and less complexity may be adopted more rapidly than others ((Rogers<br />
2003:5:16). Information and communication technologies (ICTs) like the computer and the internet have<br />
been widely adopted by people in the western countries, while in developing countries they are taking<br />
much longer to be adopted. Mobile phones ICT on the other hand have been widely adopted by both the<br />
western and developing countries during the last ten years. In Kenya, by March 2010 mobile phone<br />
subscription had reached 20 million and still growing (Communication commission of Kenya 2010). The<br />
rapid spread of mobile phone technology in Kenya has been extended to the use of innovative money<br />
transfer services. This has impacted the speed at which economic activities are carried out especially for<br />
those people who were not served by financial institutions like the banks.<br />
Money transfer services were first introduced in Kenya in 2007 in the form of ‘M-pesa’, it is a non-bank<br />
service where customers exchange cash at M-pesa agents in return for an e-money account, the virtual<br />
account can then be used for making payments, storing funds for future use and transferring funds to<br />
other phone users (Standard Chartered 2009/2010). The M-pesa money transfer service has been<br />
described as a revolutionary way of making cell phone technology work to serve a basic need by<br />
providing a fast, safe, cheaper and easy way to transfer money (Vodacom 2010). It was initially started as<br />
a mobile money transfer service to improve the efficiency of microfinance by using mobile technology to<br />
make transactions faster, cheaper and more secure (Geach, 2007:4) Other mobile money transfer<br />
services introduced later are Zap by Zain which is Safaricom’s main competitor in Kenya, ‘Yu Cash’ for<br />
YU mobile phone operator and orange money from Orange mobile phone operator, these services are<br />
heavily used for financial transactions and have also created jobs for agents in the many outlets across<br />
the country. As Kalba (2008:632) has rightly observed, mobile phones are “… spreading ubiquitously<br />
across the planet and are the latest phase of globalization’’ having out-diffused virtually every prior<br />
technology, whether TV sets, radios, wrist watches, fixed lines computers, Internet etc.<br />
This paper examines the rapid diffusion of mobile phones and mobile money transfer services and their<br />
effect on the information and business management in Kenya. It is based on an on-going research study<br />
“The diffusion of ICTs in the informal sector in Kenya”, the purpose of the study is establish the status of<br />
ICT penetration (or lack thereof) in the micro and small enterprise (MSE) sector in Kenya, the terms<br />
informal sector and MSEs sector are used interchangeably in the study. The study attempts to answer<br />
511
Wakari Gikenye<br />
the following research questions: What is the status of ICT penetration/access, use and impact (or lack<br />
thereof) in the MSE sector in Kenya?, What is the rate and level of ICT diffusion in the informal sector in<br />
Kenya?, What types of ICTs are being used by the MSE traders in Kenya, and what impact if any do they<br />
have on the businesses?, What challenges if any hinder the awareness and use of ICTs in the informal<br />
sector in Kenya?, What is the government involvement in putting the required infrastructure for ICT use in<br />
the informal sector Kenya.<br />
2. Review of relevant literature<br />
The rapid diffusion of the mobile phone in the last ten years can be explained by; among other things, the<br />
drop in the price of mobile handsets to within reach of those with low incomes, the drop in mobile tariffs<br />
as a result of stiff competition between mobile phone operators, mobile phones also require only basic<br />
literacy to use and this makes them accessible to a larger proportion of the population, they are easy to<br />
use and adaptable and for those without electricity phone–charging kiosks have quickly come up in small<br />
towns and shopping centers while solar-powered phones have also been introduced into the market<br />
(CCK 2010). The low cost of prepaid calling cards has also contributed to the rapid diffusion (Standard<br />
Chartered 2009/2010:61; Communication Commission of Kenya, 2010).<br />
Mobile ICT technology has made it possible to extend financial transfer services to millions of people<br />
including the low income at relatively low cost (Kimenyi and Ndungu 2009:2). Mas and Radcliffe (2010:2)<br />
have observed that, there nearly five times the number of M-pesa outlets than the total number of all the<br />
other financial services i.e. Post Bank branches, post offices, bank branches and automated teller<br />
machines all put together in the country. This is because unlike conventional banking branches that are<br />
only located in major urban centers due to high running expenses, M-pesa agents are the ones who<br />
decide on the most strategic points to locate their service outlets since they are easy to put up and to<br />
maintain, this has dramatically reduced the cost of delivering financial services, with the benefits being<br />
transferred to the users (Njenga nd :5).<br />
Mpesa has shown impressive growth in the utilization of mobile payments within a short period of time,<br />
with the number of monthly transactions having increased from 354,298 in July 2007 four months after it<br />
was introduced to 16,747,419 in July 2009, an increase of 4627 percent, the value of monthly<br />
transactions also rose from Kenya shillings 1.065 billion (Us $ 14.2 million) in July 2007 to 40.176 billion<br />
(Us $535.6 million) in July 2009, a growth of 3671 percent, these financial transactions generated<br />
substantial returns, and created job opportunities with about 12,300 agents in the country (Kimenyi and<br />
Ndungu 2009:3).<br />
Factors that have been responsible for the highly successful money transfer services in Kenya are; the<br />
fast adoption of mobile phones, the unmet need to access financial services as well as a strong demand<br />
for domestic remittances, such needs used to be met through informal and risky methods like sending<br />
money through relatives and friends, or matatus (shared taxi), buses and couriers by those working in the<br />
urban areas to their relatives in the rural areas, (Omwansa 2009).<br />
M-pesa therefore may have been introduced at the right time, to meet the needs of accessing financial<br />
services especially by those who were not served by the formal banking institutions. The cost of mobile<br />
money service is relatively low compared to the earlier methods, it is also faster more convenient and<br />
safer than cash. The adoption of mobile-based money transfer service has speeded up access to<br />
financial services not just for the unbanked but also for the banked as a preferred avenue of transactions<br />
(Daily Nation September 29 2010). The biggest beneficiaries, however, are the MSE traders who have<br />
discovered a new time-saving service leaving them with more time to work on their businesses<br />
(Omwansa 2009:111). The mobile money transfer have services also offered protection and safety by<br />
enabling instant and on-demand payments (Mas and Radcliffe 2010:5). Other innovative successful<br />
money transfer services in Africa are Wizzit in South Africa and Glofirst card in Nigeria which are used to<br />
draw cash, transfer money, make purchases, and check bank statements using the mobile phone<br />
(Standard Chartered 2009/2010:61).<br />
2.1 Methodology: How do MSEs in Kenya get information for their business<br />
management?<br />
According to Rogers’ theory of diffusion (2003:1), getting a new innovation adopted even when it has<br />
obvious advantages is difficult and many innovations require a period of many years to gain wide<br />
adoption. The slow adoption of computer and the internet for information provision in developing<br />
512
Wakari Gikenye<br />
countries provide examples while adoption and use for the same has been much faster in Western<br />
countries. Technology acceptance model (TAM) developed by Davis (1989) stresses perceived<br />
usefulness (PU) and perceived ease of use (PEOU) of an innovation (Khasawney and Ibrahim 2008:137).<br />
A combination of social, economic and psychological factors might explain how different communities<br />
become aware of and adopt particular innovations while they take much longer to perceive the<br />
usefulness and ease of use to others.<br />
A survey was carried out in two provinces in Kenya where a total of 377 MSEs were selected using a<br />
combination of multistage sampling technique consisting of purposive, cluster and random sampling. The<br />
two provinces selected are Nairobi and Central provinces representing an urban and a rural area<br />
respectively. The MSEs consisted of retail shops (mainly dealing with clothing and footwear), phone<br />
shops, auto-spare shops, furniture and hardware shops, and curio and horticultural traders. The clothing<br />
and footwear retail shops, phone shops and electronic shops were more concentrated in the city center<br />
while hardware, furniture, hair salons and construction material stores were found in the market centers.<br />
Curio traders were found in the up-market Village Market in Nairobi province while horticultural traders<br />
were based next to Jomo Kenyatta International Airport.<br />
Data was collected using comprehensive structured and non-structured questionnaires, to gather<br />
information from MSE owners, managers and employees on their use of ICTs for business management.<br />
The questions were based on their use, attitudes and impact of ICTs. Due to the informal nature of the<br />
business environment in the MSE sector, researcher-administered questionnaires were used.<br />
3. Results<br />
The results are presented in sections 3.1. to 3.7<br />
3.1 Gender composition, educational level and age of respondents<br />
The sampled MSE traders were 49 percent men and 51 percent women, making the gender balance<br />
quite close. As shown in table 1 the educational level of the traders varied from primary to degree level,<br />
with more than 27% (104 respondents) of the respondents having gone up to degree/diploma level and<br />
48% (182) up to secondary school level. 16% (60 respondents) had undergone various types of post<br />
secondary training, while over 6% (25 respondents) were primary school leavers. Majority of respondents<br />
had attained secondary school education and above.<br />
Table 1: Educational level of respondents<br />
Frequency Percent<br />
Primary Level 25 6.6<br />
Secondary Level 182 48.3<br />
Post Secondary Training 60 15.9<br />
Diploma/degree/Masters 104 27.6<br />
No answer 6 1.6<br />
377 100<br />
Table 2 shows the majority of the respondents are relatively young people between the age of 20 to 45,<br />
with 75 percent being under 35 years, these are young people trying to earn a living from MSE trade,<br />
most of whom are school leavers who failed to get jobs in the formal sector.<br />
Table 2: Age of respondents<br />
Frequency Percent<br />
Below 25 years 100 27<br />
25-35 years 180 48<br />
36-45 years 93 25<br />
Over 45 years 1 .3<br />
No answer 3 .8<br />
Total 377 377<br />
513
3.2 Size of MSEs in the study<br />
Wakari Gikenye<br />
The MSEs in the study can be divided into three categories based on their size and the premises they<br />
occupy. The first category operates from permanent buildings and enjoys infrastructural facilities such as<br />
water, electricity, they also pay relatively higher rents. They enjoy relative permanency and their capital<br />
outlay is also relatively higher. Examples of these are horticultural exporters and electronic and autospare<br />
shops. Most MSEs in this category use ICTs such as the computer, Internet and email for their<br />
businesses.<br />
The second category of MSEs operates micro-sized businesses in shared premises. Despite being<br />
housed in permanent premises and sheltered from the vagaries of the weather their businesses and<br />
operations are very small, these are the micro retail stalls found in the main streets of Nairobi.<br />
The third category operates from temporary sheds or stalls or in the open (this is the origin of the term<br />
“Jua Kali’’, Kiswahili for ‘’in the hot sun’’ used generally to refer to informal sector enterprises in Kenya).<br />
Their premises generally lack infrastructural facilities such as running water and electricity. They do not<br />
own or use computers and also lack awareness of ICT developments, examples are curio traders,<br />
grocery stores, food kiosk operators etc.<br />
Over seventy per cent (270) of the MSEs studied had between one and five employees, with 35 percent<br />
(133 respondents) having only one employee, only 7 percent had more than five employees.<br />
3.3 Diffusion and use of ICTs among the MSEs<br />
With the exception of mobile phones, the majority of MSE traders do not use ICTs such as computers for<br />
their businesses. The majority carry on as if such technological developments have nothing to do with<br />
them.<br />
As table three shows twenty six percent (96 respondents) owned a computer, while over 94% (356<br />
respondents) (shown in table 4) reported owning mobile phones. There was also a relatively high number<br />
of respondents with computer skills; 42 percent (158) but who were not using them because they did not<br />
think their businesses need the use of computers and said they could not afford them.<br />
Table 3: Ownership of computers in the MSEs<br />
Frequency Percent<br />
Yes 96 25.5<br />
No 281 74.5<br />
Total 377 100.0<br />
3.4 Diffusion and use of mobile phones and money transfer services for business<br />
management<br />
Findings shown in table 4 show that the mobile phone is quite popular among the MSE traders, with over<br />
94% reporting that they owned and used a mobile phone to conduct their businesses. 95 percent (357<br />
respondents) also reported that they were registered M-Pesa users, M-pesa was the pioneer for money<br />
transfer services and has been synonymous with mobile money transfer services in Kenya.<br />
Table 4: Ownership of mobile phone by MSE owners<br />
Frequency Percent<br />
Yes 356 94.4<br />
No 21 5.6<br />
Total 377 100<br />
As shown in table 5, 28 percent (108 respondents) reported using M-Pesa daily, 29% (111 respondents)<br />
use it once or twice a week; 21 % (80 respondents) said they used it whenever the need arose.<br />
Sixty three percent (237 respondents) of the M-pesa users said they used it to send and receive money<br />
from relatives while 79% (296 respondents) said they used it for business transactions. For most<br />
respondents, the mobile phone was their first contact with a phone, having never used the landline. Given<br />
that even the low prices of handsets and cost of airtime is a significant proportion of their earnings, the<br />
514
Wakari Gikenye<br />
high percentage of ownership and use of mobile phones means they value them highly and they would<br />
rather forgo other things to make sure they maintain the use of the mobile phone, which they reported<br />
using heavily for their work. In addition to being useful in business activities mobile phones are also seen<br />
as a status symbol.<br />
Table 5: Frequency on the use of M-pesa (mobile money transfer service)<br />
Frequency Percent<br />
Everyday 108 28<br />
Twice a week 20 5<br />
Once a week 91 24<br />
Once a month 17 5<br />
Other 36 10<br />
When need arises 80 21<br />
No answer 25 7<br />
Total 377 100<br />
Almost 60 per percent (222 respondents) reported that they used the mobile phone for their businesses<br />
transactions, 73% (275 respondents) said they used it for ordering goods, 79% (299 respondents) for<br />
contacting customers, and 50.4% (190) said they had used it to improve their businesses. 72 per cent<br />
(271) said they used the mobile phone for social communication, e.g. contacting friends and relatives.<br />
3.5 What difference has the mobile phone made for MSE traders in Kenya<br />
Fifty one percent (192 respondents) said the use of mobile phones made it easier and faster to<br />
communicate and carry out business transactions at any time as well as communicating with people in far<br />
off places. Fifteen percent (57 respondents) said the mobile phone helped them to get goods delivered<br />
more easily, 31 percent (115 respondents) said the use of mobile phones made it easier to contact and<br />
bring more customers leading to increased revenue.<br />
A small number of respondents (2%; 6 respondents) said that they were what they were because of the<br />
use of mobile phones, i.e. they had been able to start a business and maintain it due to the availability<br />
and use of the mobile phone. They reported that, by using the mobile phone, they were able to know<br />
which market to go to after comparing prices in different markets, before they left home. They were also<br />
able to change plans if the information received on the prices of a particular commodity would not bring<br />
them profits.<br />
Most respondents reported doing businesses they had been doing more easily and conveniently with the<br />
use of the mobile phone, 16%(62 respondents) said they were impressed by the genuineness, speed,<br />
convenience and reliability that the mobile phone had brought and which had resulted in making their<br />
work much easier and convenient, 4 percent (16) went as far as saying they could not do without it as it<br />
had ‘’become everything in their business’’.<br />
Only 2 per cent (6) said they relied on the money transfer service when a customer did not have money<br />
and could send it later, most traders said they prefer face-to-face transactions rather than encouraging<br />
credit services. 4 percent (14), said that M-Pesa is good for security and for carrying the money around<br />
instead of cash, because even if the phone is stolen the money is still safe. 3 per cent (10) said it is good<br />
for emergencies including medical ones due to its speed and convenience. 19 percent (72) said the<br />
money transfer services are conveniently available for more hours unlike banks, and that it is easier to<br />
save small amounts due to instant access, and one need not queue at the bank.<br />
Other uses the mobile phone is used for by the MSE traders are; telling the time (73 percent; 276<br />
respondents) thus replacing the wrist watch, 47 percent (178) use it as a camera, 46 percent (173) used<br />
it to access the Internet, 49 percent (185) use it as a radio.<br />
3.5.1 Problems of using mobile phones by MSE traders in Kenya<br />
Thirty seven percent (140 respondents) said they had problems buying credit for their phones but<br />
nonetheless indicated that it was a necessary cost for doing business, 54 percent per cent (207)<br />
complained about network congestion, a common problem experienced by Safaricom customers the<br />
main telephone service provider in Kenya and M-pesa money transfer service users. The mobile phone<br />
was also reported (46 per cent) to be an easy target by thieves and pickpockets. Only 1 percent (4<br />
515
Wakari Gikenye<br />
respondents) reported having problems with charging their phones, this is significant given the fact that<br />
there are many people in Kenya without access to electricity in their homes.<br />
3.5.2 Information sources for MSE business management<br />
Apart from getting information from mobile phone contacts and fellow traders, 34% (128 respondents) of<br />
MSE traders also reported getting their business information through acquaintances, friends and<br />
relatives, through word of mouth while only 2% reported getting information through government officers.<br />
23 percent said they got information through the mass media while 15% got their information through the<br />
internet.<br />
3.6 Challenges that hinder the awareness and use of ICTs by MSEs in Kenya<br />
As the findings show in 3.2 above the majority of MSEs are small in size with the majority having less<br />
than five employees, most of them are survival outfits, and their size and earnings do not enable the<br />
ownership of ICTs like the computer. As shown in table 6, they are started with minimum preparation as<br />
shown by the amount of starting capital, making the acquisition of expensive ICTs a remote possibility.<br />
Forty eight percent reported starting their businesses with less than 1000,000 Kenya shillings ($1250)<br />
with some having been started with as little as less than 10,000 Kenya shillings ($125).<br />
Table 6: Capital outlay at the beginning of business<br />
Frequency Percent<br />
Less than Kshs 10,000 7 2<br />
Kshs 10,000-20,000 49 13<br />
Kshs 20,000-40,000 41 11<br />
Kshs 40,000-60,000 41 11<br />
Kshs 60,000-100,000 40 11<br />
Over Kshs 100,000 51 14<br />
Do not know 55 15<br />
No answer 93 25<br />
Total 377 100<br />
3.7 Government Involvement in putting the required infrastructure for ICT use<br />
In spite of the efforts the government of Kenya has made to bring the internet closer to Kenyans by<br />
enabling three under sea cables to improve internet connectivity and the reduction on taxes on<br />
computers, the MSE traders do not think the government has done much to assist them in getting the<br />
necessary information for their businesses, they reported that they could not afford computers for their<br />
businesses even for those who have acquired computer skills (3.3 above). Asked why they did not<br />
recognize or appreciate the government efforts to help them get more information through ICTs they said<br />
that that their first priority was to get affordable and more permanent premises where they would<br />
establish their businesses, that the current status of MSE business premises in Kenya is unpredictable<br />
with the City Council’s occasional raids on semi-permanent premises. The MSE traders also felt the<br />
government was too keen to collect taxes and license fees but less eager to give services corresponding<br />
services to enable them establish more productive businesses.<br />
4. Discussion<br />
This study has found that, gender composition for MSE traders studied is the same for men and women,<br />
this is unlike other studies’ findings that there are more women than men involved in MSEs (Okello-Obura<br />
et al 2007:13).<br />
The small size of the MSEs has not endeared them to the use of ICTs like the computer, rather the<br />
majority of the MSEs traders have adopted mobile phones and mobile money transfer services for<br />
management of their businesses, the mobile phone technology has also been well adapted to the local<br />
economic environment for use by both the unbanked and the banked Kenyans.<br />
Due to the convenience resulting from the use of mobile phones and money transfer services and saved<br />
time and money in saved journeys respondents reported that business had become much easier and<br />
faster to run, and was more profitable. Thirty five per cent (131 respondents) said mobile money transfer<br />
services simplified business and social financial transactions, saves time and money, by cutting costs of<br />
516
Wakari Gikenye<br />
traveling and increases the efficiency of doing business. Money transfer services are more flexible and<br />
can be used for saving and withdrawing small amounts of money, they operate for more hours, unlike the<br />
banks, they are cheaper than bank accounts and require no minimum balance.<br />
The mobile phone was reported to have changed business practices for the better as it made it easier to<br />
order, supply and make payments for goods and services using the mobile phone. As summed up in a<br />
Standard Chartered bank report (2009/2010:60) “the mobile phone is having a dramatic effect …and is<br />
proving to be a life-transforming device”, mobile phones have become a key economic driver and are<br />
used in many different ways like marketing of goods and obtaining information, they have reduced travel<br />
time to distant bank branches and have enabled access to banking. According to Kimenyi and Ndungu<br />
(2009:2) Kenya’s mobile phone transfer service is a good example of how low cost approaches using<br />
modern technology can effectively expand the financial services frontier.<br />
Mas and Morawczynski (2009:89) are optimistic that there is more potential use of mobile transfer<br />
services that can be extended to payment of salaries and wages, payment of bills, school fees etc, and<br />
that, although not originally conceived to work as a saving service it has been used as such by the<br />
unbanked even with no interest. Emerging partnerships of various banks with mobile phone providers<br />
taking place in Kenya will make the realization of mobile banking a reality, as well as collectively bring<br />
together the aggregate savings from the wide network of money transfer services’ customers.<br />
While achievements of the mobile phone in the MSE sector are from private sector efforts the<br />
government has played a significant part in providing an enabling environment. This has not been felt in<br />
the adoption and acquisition of other ICTs like the computer and the internet, and the challenge remains<br />
for the government to do more so that the MSE traders can realize the benefits of the other ICTs.<br />
5. Conclusion<br />
This paper has found that, the status of ICT adoption especially that of computers is still low and out of<br />
reach of the MSE traders whose operations continue to remain small. The affordable and easy-to-use<br />
mobile phone technology however has been quickly and extensively adopted and is being used to<br />
provide easier communication of needed information by MSEs in Kenya. This has made businesses more<br />
efficient, less costly and more profitable in the MSE sector. The mobile phone technology has also been<br />
innovatively used to offer financial transfer services whose fast adoption is proof that it has been found<br />
useful and handy not just for the unbanked but also as a service to supplement the banked.<br />
The rapid diffusion of the mobile phone is not only enabling faster and more efficient communication but<br />
is also empowering people including the low income with information and services that were not available<br />
to them before. This will have a positive impact on the economy and with the telephone companies<br />
increasing their activities in the market and the growing partnerships with the banks the Kenyan economy<br />
is bound to grow as a result.<br />
The study also found out that for ICT adoption to be a reality for the MSE traders and indeed for all the<br />
other Kenyans, there is need for concerted efforts by the government, the internet service providers and<br />
for the people themselves to update themselves with what is happening around them and the rest of the<br />
world if they are to catch up and keep abreast with the rest of the world. A lot remains to be done to<br />
enable the realization of ICT benefits so as to maximize information acquisition for business<br />
management. While the mobile phone technology has been innovatively used to provide the MSE traders<br />
and other Kenyans with easier communication and money transfer services it has limitations on the<br />
impact that it can make on the economy of the country even at its best. The government needs to<br />
facilitate the creation of more jobs so as to have a vibrant economy which can increase the demand for<br />
goods and services, including those being produced by MSEs. Lack of jobs is the genesis for the flocking<br />
of the unemployed into the informal sector, where the participants try to eke out a living. Even with the<br />
government efforts to make internet connectivity easier to access and the zero-rating of taxes on<br />
computers and accessories people need to satisfy the basics before they can acquire ICTs.<br />
Appendix 1<br />
Questionnaire<br />
Section A: Personal Information<br />
1. .Name: ________________________________________ (Optional)<br />
517
Wakari Gikenye<br />
2. Age: (1) Below 25 years (2) 25-35 years (3) 36-45 years (3) Over 45<br />
3. Sex: (1) Male (2) Female<br />
4. Marital Status: (1) Single (2) Married (3) Other<br />
5. Educational Level: (1) Primary Education Level (2) Secondary Education Level (3)<br />
Post Secondary Training (specify)_________________(4) Diploma/degree Holder.<br />
Section B: Information about the business enterprise<br />
6. What is the name of your business ________________________________________<br />
7. Are you the (1) Owner (2) Part Owner (3) Manager (4) Employee<br />
8. What are the goals of your business enterprise? _______________________________<br />
9. How many permanent employees does the business have? (1) One only (2) 2-5<br />
(3) 6-10 (4) Over 10 specify how many____________________________________<br />
10. Do you have employees that are not permanent? (1) Yes (2) No If yes how many?<br />
11.Is the business service-oriented or goods- production enterprise? Service-<br />
oriented (1) goods-production enterprise (2)<br />
12. How many years has it been in existence? Less than 2 years (1) 2-5 years<br />
(2) 5-10 years ((3) Over 10 years (4)<br />
13. Are there many other such business enterprises in the neighbourhood? Yes (1) No (2)<br />
14. If yes how many? (1) 1-5 (2) 6-10 (3) Over 10.<br />
15. What facilities do they share if any?__________________________________<br />
Section C: Use of ICT (Information and Communication Technologies like the phone, computer<br />
and internet) for the business<br />
16. Does your business use ICTs? (1) Yes (2) No. If Yes, for what are the<br />
ICTs used for (1) To get information on raw materials/ordering goods (2) To get<br />
information on Customers, markets for goods/services (3) To contact friends and<br />
relatives (4). Any other need, specify _________________________ _<br />
17. How do you get information for your business? (1) Personal knowledge (2) Fellow<br />
Traders/producers (3) Mass media (radio, TV, newspapers) (4) internet (5) mobile<br />
phone contacts (4) Government officers (5) Acquaintances, friends and relatives,<br />
Other sources specify__________________________________________________<br />
18. How do you get information about customers for your work? Customers<br />
come on their own (1) Word of mouth (2) Advertising in the mass media (radio, TV,<br />
newspapers) (3) Through the internet and email contacts (4) Mobile phone contacts<br />
19. What else and by whom do you think needs to be done to make your access to<br />
information through ICTs better? (1) government (2) Phone companies (3) Other,<br />
specify ______________________________________________________________<br />
Section D :Acquisition and Use of (ICTs)<br />
518
Wakari Gikenye<br />
20. Does your business own a computer\mobile phone /or land line? Or other ICT<br />
resource, Specify_______________________________________________________<br />
21. How is/are the ICT resources used in the enterprise? (1) Typing<br />
(2) Communication (3) Internet use. (4) scanning (5) fax services (6) Sending money<br />
22. Do you think ICT resources are useful to the enterprise? Yes (1) No (2) Give<br />
Reasons _____________________________________________________________<br />
____________________________________________________________________<br />
23. Are there any people with computer skills in your organisation? (1) Yes (2) No<br />
24. If yes how many? (1) only one (2) all employees<br />
25. What kind of skill do they possess? (1) typing skills (2) skill for email and internet<br />
use (3) technical skills<br />
26. Are these skills used for the organisation? (1) Yes (2) No<br />
27. If yes what are the skills used for? (1) Typing (2) emails for keeping in touch with<br />
friends and relatives (3) business communication.<br />
28. What benefits does the use of ICTs bring to the business enterprise? ______________<br />
_____________________________________________________________________<br />
29. Do you know what was the approximate capital outlay at the beginning of this<br />
business? (1) Kshs 10,000-20, 000 (2) 21,000-40,000 (3) 41,000-60,000 (4) 61,000-<br />
100,000 (5) Over 100,000, Specify (6) Do not know.<br />
30. What is your approximate monthly turn-over (1) Kshs 10,000-20,000 (2) 21,000-<br />
40,000 (3) 41,000-60,000 (4) 61,000-100,000 (5) Other specify (6) Do not know.<br />
Section E: Access, Use and Impact of Mobile Phones<br />
31.How long have you had your mobile phone? (1) 1-2 years (2) 2-5 years (3) 5-10 years<br />
32. What do you use your mobile phone for? (1) For my work (2) For ordering goods (3)<br />
For contacting customers (4) For improving my business (5) For social<br />
communication (6) Other, specify ________________________________________<br />
33. How can you compare the time you did not have a mobile phone and now ?<br />
34. What other uses do you use your mobile phone for (1) For telling time (2) as a camera<br />
(3) for internet (4) For Mpesa (5) As a radio<br />
35. Has the mobile phone helped you to improve your business? (1) Yes (2) No<br />
36. What are the challenges\problems of using the mobile phone? (1) expensive to buy<br />
(2) air time (3) network congestion (4) easy target for thieves and pickpockets.<br />
37. Do you make/receive more business /more social calls on you mobile phone?<br />
(1) More Business-related calls (2) More social calls<br />
38. How many of the calls that you have made for the last one week are (1) work related<br />
(2) For social communication.<br />
39. Those which are work-related what were they used for? (1) To contact customers<br />
(2) To order goods\supplies (3) Other, Specify_____________________________<br />
40. Are you a registered Mpesa user? (1) Yes (2) No<br />
519
Wakari Gikenye<br />
41. How often do you use Mpesa (1) Everyday (2) Once a week (3) Other specify______<br />
42. What do you use Mpesa for? (1) T o send\receive money to\from relatives (2) For<br />
Business transactions.<br />
References<br />
Camner G. et al (2009). What makes a Successful Mobile Money Implementation? M-pesa in Kenya and Tanzania.<br />
Financial sector Deepening Kenya<br />
Communication Commision of Kenya (CCK) Quarterly sector statistics Report 2 nd quarter October 2009 –March2010<br />
http://wwwcck.go.ke./news/2010/news-30mar10.html<br />
Daily Nation, Wednesday September 29 2010 Mobile Banking boosts access to cash services. Nation Media group<br />
Limited.Nairobi.<br />
Geach N (2007) The Digital Divide, Financial exclusion and Mobile Phone Technology:Two Problems One Solution?<br />
British and Irish Low, Education and Technology Association 2007 Annual <strong>Conference</strong>. Hertfordshire16-<br />
17April. International Telecommunications Union (ITU) 2009. Measuring the Information society The ICT<br />
DevelopmentIndex.<br />
Kalba K. (2008) The adoption of mobile phones in emerging markets. Global diffusion and the rural challenge.<br />
International journal of Communication 2 (2008), P631-661.<br />
Khasawneh M. and Ibrahim H. (2008) Toward an information and communication technology development in<br />
developing countries. Communications of the International Business Information Management Association<br />
(IBIMA) Vol. 4 2008 p.135-140.<br />
Kimenyi M. and Ndungu N. (2009). Expanding the financial services frontier : Lessons from mobile phone banking in<br />
Kenya. Brookings Washington D.C.<br />
Mas I.and Morawczynski (2009) Designing mobile money services: Lessons from M-PESA. Innovations, a quarterly<br />
journal published by MIT Press vol.4, issue 2 2009 p. 77-91.<br />
Mas I and Radcliffe D. (2010). Mobile payments go viral : M-pesa in Kenya. Bill and Melinda Gates Foundation.<br />
Njenga A. n.d. Mobile phone banking: Usage experiences in Kenya. Catholic University of Eastern Africa.<br />
Okello-Obura, C. et al (2007). Assessment of business information access problems in Uganda. Partnership: The<br />
Canadian journal of library and information practice and research, vol.2 no.2 2007 pp.1-34.<br />
Omwansa T. (2009). M-pesa : Progress and prospects. Innovations/ Mobile World Congress 2009 p. 107-123<br />
Rogers E. (2003) Diffusion of innovations (5 th ed) New York Free Press.<br />
Standard Chartered, Asia, Africa and the Middle East : The guide to working capital Management 2009/2010.Banking<br />
the unbanked : Going mobile in Africa.<br />
M-pesa launched in South Africa (Vodacom) http:www.howwemadeitinafrica.com/M-pesa-launched-in-southafrica/3611/<br />
accessed on 22 nd September 2010.<br />
520
Querying DTI Analysis Results Within Deep Perisylvian Area<br />
Sarmad Istephan and Mohammad Siadat<br />
Oakland University, Rochester, USA<br />
sistephan@gmail.com<br />
siadat@oakland.edu<br />
Abstract: This paper is an advancement in the field of Data Driven Medicine. There has been a tremendous amount<br />
of work done in integrating technology into the medical field. Part of that work is in the storage and querying of<br />
medical images which is the focus of this paper. The high level motivation of this paper is to design a database<br />
driven system that empowers radiologists to validate their hypotheses against a set of medical images. More<br />
specifically, a system has been developed that answers queries pertaining to the average Fractional Anisotropy (FA)<br />
within a given brain structure. The proposed system enables querying of Diffusion Tensor Imaging (DTI) images to<br />
evaluate hypotheses that correlate changes in DTI analysis results (e.g., FA map) with brain epileptogenicity. The<br />
system makes the assumption that the following data are available: a) high-resolution and high-contrast image data<br />
such as T1-weighted MRI, b) models of brain structures of interest segmented in the high-resolution and highcontrast<br />
image space, c) DTI analysis maps such as FA map, d) registration Information that transfers the brain<br />
structures’ models from their native image space (T1) to target image space (FA map). A data model to store this<br />
information has been designed. The data model is implemented in a database that uses SQL Server 2008<br />
technology. This database stores patients’ data, various image modalities, segmented models, and registration<br />
information. Part of importing the images into the database includes reading the dicom image tags and feeding the<br />
available tags into the database. After all the required pieces of data are in the database, the calculation of the<br />
average FA within a given model is done as follows. First, using the Visualization Tool Kit (VTK) libraries and the<br />
registration information the segmented model is transferred from its native space to the target image space (e.g.,<br />
from T1-weighted MRI to FA map). Next, a bonding box is created around the model and each voxel within the<br />
bonding box is examined to see if it is inside the segmented model. At this point all the voxels of interest (inside the<br />
model) are identified and the average FA map would be calculated using the Image Tool Kit (ITK) libraries. The<br />
primary focus of this work was to calculate the average FA but the system can be extended to include many other<br />
features of interest such as fiber density, curvature measures, etc. In conclusion, this system creates a link between<br />
quantitative measurements of several modalities of data e.g. DTI and qualitative patients’ situations in clinical<br />
practices, which is an important element in the field of Data Driven Medicine.<br />
Keywords: medical image querying, medical image database, DTI/MRI querying, medical image data model,<br />
DICOM, medical imaging<br />
1. Introduction<br />
Data Driven Medicine is an initiative to improve health care in the United States by leveraging technology.<br />
The United States Department of Health and Human Services (HHS) has recently received a $3.6 billion<br />
budget to improve Information Technology (IT) within health care. A part of that budget goes to Health IT<br />
which is part of HHS. Health IT replaces traditional paper methods for storing patient information with<br />
Electronic Health Records (EHRs). The usage of EHRs enables the storage and secure dissemination of<br />
comprehensive patients’ health history. The availability of EHRs provides doctors with a mechanism to<br />
reach an accurate diagnosis sooner than traditional paper methods would (HHS Press Office 2010; The<br />
Office of the National Coordinator for Health Information Technology 2010).<br />
Patients’ medical image scans of various modalities are a subset of EHRs. These modalities include<br />
Magnetic Resonance Images (MRI), Diffusion Tensor Images (DTI), Computed Tomography (CT) and<br />
Ultra Sound (US). There are many medical image modalities created daily. In 2002, on a daily basis, at<br />
least 12,000 radiological images were created at the University Hospital of Geneva. In one year this<br />
amounts to 1 terabyte of image data (Muller et al 2004). Another study showed that tens of terabytes of<br />
image data is produced by an average size radiology department which means pedabytes of image data<br />
per country per year (Glatard et al Magnin 2004). This humongous amount of data can be used to<br />
complement traditional medicine in many ways. It is impossible to ask radiologists to analyze this data<br />
because their time is expensive and analyzing this amount of data manually would take a very long time.<br />
Therefore, a system must be designed that can automatically query the data.<br />
This paper focuses on developing a system with a database to store medical image modalities and<br />
develops a mechanism to query the image data automatically. Since this is an automatic system, then it<br />
addresses subjective analysis that may occur if a radiologist was performing the analysis. Also, the<br />
system will be able to analyze many more aspects of the images much faster than a radiologist can.<br />
521
Sarmad Istephan and Mohammad Siadat<br />
A data driven system has potential to be used in many areas. One usage of such a system is in medical<br />
school education where students can leverage the system to construct and test various hypotheses that<br />
they develop and verify them across many images. A second use of this system is to leverage the image<br />
data to determine an accurate quantitative outcome prediction. A quantitative outcome prediction would<br />
provide the patient with information using their image modalities in conjunction with those from previous<br />
patients with a similar disease to come up with conclusions. An example of a conclusion would be that<br />
90% of patients going through treatment plan X had side effect Y and it is up to the patient whether to<br />
undertake such a treatment that only 10% of patients did not have the side effect of. A third use of such a<br />
system is to come up with patient centric treatment plans. This means that treatment plan X can be<br />
tailored to a specific patient by leveraging information from the image data. A fourth use of such a system<br />
is as a decision support system. A decision support system will use the pre-operation and post-operation<br />
image data from previous patients during a new patient’s invasive surgery to decide on where to cut in<br />
order to minimize/eliminate tissue damage. Finally, the image data can be used as a mechanism to<br />
prevent the need for invasive techniques all together because it may unlock other treatment options that<br />
would not have been discovered by only using patient provided information (Muller et al 2004).<br />
The remainder of the paper has the following structure. In Section 2, a brief examination of previous work<br />
is provided. In Section 3, the method this paper uses is explained. In section 4, the technology used for is<br />
provided. In section 5, simulated data and results are provided. Finally, in section 6, the paper is<br />
concluded.<br />
2. Previous work<br />
There are three approaches to query medical data. One approach is to use the structured data.<br />
Structured data includes demographic data of a patient such as age, gender, height, weight, etc. Another<br />
approach is to use unstructured data. Unstructured data includes the image data of patients. The third<br />
approach is to use a hybrid of structured and unstructured data (Siadat et al 2009). This paper focuses<br />
on developing hypotheses and validating them using the unstructured image data.<br />
There are two methods used to query unstructured image data. One method is to create structured data<br />
(annotations) that represents the unstructured data. This approach is accomplished manually by a<br />
human. Sine transcribing unstructured data into structured data is manual then it is prone to error in the<br />
sense that the human can make mistakes while dictating the data. Also, it is irreproducible in that two<br />
humans may come up with different annotations or even the same human might come up with different<br />
annotations at different times for the same image data. In addition, this manual process is time<br />
consuming because only one volumetric image can be operated on at a given time. Finally, it is neither<br />
feasible nor possible to have a complete set of structured data to describe all the features of an<br />
unstructured image (Chbeir et al 2002) (Glatard et al 2004).<br />
The other method to query unstructured data is through the use of image segmentation, feature<br />
extraction and other imaging techniques that operate directly on the image itself. This approach<br />
addresses all the concerns of using structured data to represent unstructured data and hence is the<br />
preferable approach (Chbeir et al 2002). This paper queries the unstructured data directly to answer<br />
queries.<br />
The hybrid system approach to querying medical data is to use the unstructured image data as a<br />
complement to traditional textual queries. Textual queries are those pertaining to structured data. The<br />
querying of structured data will always be needed as a way to focus the study on certain demographics.<br />
However, leveraging unstructured data as a complement to the structured data provides the ability to<br />
answer many questions that are difficult to answer otherwise (Muller et al 2004). An example of such<br />
questions is, “Is there an intensity difference in the medial side of the perisylvian area of the left<br />
hemisphere of the brain compared to the medial side of the perisylvian area in the right hemisphere for<br />
patients within the ages of 20 and 25?” The age part of the question uses structured data and everything<br />
before that uses the unstructured data. (Chbeir 2002).<br />
There have been many good systems designed so far that use unstructured data. Most of these systems<br />
are designed using the Content Based Image Retrieval (CBIR) technique. In this technique, a system is<br />
given an image and asked to find all similar images in the database. The process of finding the similar<br />
images in the database is then refined by selecting an image from the result set and asking the system to<br />
again search. This iterative process is repeated over and over until accurate results are reached. The<br />
Spine and Pathology and Image Retrieval System (SPIRS) achieved a 20% increase after three iterations<br />
522
Sarmad Istephan and Mohammad Siadat<br />
of similarity based retrieval (Hsu et al 2008). Another system that uses CBIR is the Medical Image<br />
Management System (MIMS) which combines the use of structured and unstructured data to answer<br />
queries (Chbeir 2002). Another system that uses CBIR by exposing various methods as Web Services is<br />
MedImGrid (Jin 2007). Another system that performs Gabor Filter feature extraction and CBIR is done by<br />
Tristan Glatard (Glatard 2004). There are many other papers that also use CBIR as a mechanism to<br />
query unstructured data and are all reviewed in Henning Muller’s review of CBIR systems (Muller 2004).<br />
The theme amongst all these systems is the requirement of an initial image to initiate a query. This paper<br />
develops a system that does not require an initial image. Furthermore, these systems compare against<br />
one image modality. However, the system in this paper is able to do registration between different<br />
modalities of images thus allowing registering anatomical information to functional information.<br />
The research in medical data storage and retrieval has resulted in many systems that are widely used in<br />
hospitals. The overall system is the Hospital Information Systems (HIS). The HIS system is a system that<br />
consists of many subsystems used in the capturing, storage and retrieval of everything from patient<br />
appointments to image modalities acquisition to billing (Biohealthmatics 06). One subsystem of HIS is the<br />
Picture Archiving and communication Systems (PACS). PACS allows for the capturing, storing and<br />
retrieval of medical image modalities without the use of traditional image films (Becker and Arenson 94).<br />
A complement to PACS is the Radiological Information Systems (RIS). RIS enables the storage of patient<br />
data such as patients’ history from admission to discharge, patients’ appointments, patient centric reports<br />
and traditional film tracking (Biohealthmatics 06). The future of the system designed in this paper is a<br />
complement to all of these systems with the primary focus on the validation of hypotheses by querying<br />
the unstructured image data.<br />
3. Method<br />
In this paper, a new system was developed that enables the storage of unstructured medical images and<br />
hypotheses validation through querying the images directly. The goal of the system is to extend the<br />
power of Structured Query Language (SQL) through the use of the Common Library Runtime (CLR) to<br />
query the unstructured data. The general form of the query the system supports is,<br />
SELECT [Field Name]<br />
FROM [Database.Tables]<br />
WHERE [FeatureX] [Spatial Relation to Structure Name] [Structure Name] of [Image Map] [Comparison<br />
Operator] [Comparison Value]<br />
An example of this query is:<br />
SELECT [Field Name = Patient Name]<br />
FROM [Database.Tables]<br />
WHERE [FeatureX = AverageFA] [Spatial Relation to Structure Name = Adjacent to Medial Side] of<br />
[Structure Name = Perisylvian Area] of [Image Map = DTI] [Comparison Operator = >] [Comparison Value<br />
= 50]<br />
This general query has been encapsulated into a function called QueryImage. QueryImage function is a<br />
SQL CLR function that can be called directly from a query’s where clause. This function was created as<br />
an easy mechanism for the communication between the application and the database.<br />
QueryImage takes the following parameters,<br />
[FeatureX]: Type of module to be executed (e.g., Average FA).<br />
[Spatial Relation to Structure Name]: Geometric relationship to be used (e.g., within).<br />
[Structure Name]: The name of the structured to be queried (e.g., Perisylvian area).<br />
[Image Map]: The type of image modality to be used (e.g., MRI).<br />
[Comparison Operator]: The type of comparison to make (e.g., >).<br />
[Comparison Value]: The value to compare against (e.g. 50)<br />
523
Sarmad Istephan and Mohammad Siadat<br />
The power of this design is that new modules can be added to the system at any time without affecting<br />
the rest of the system. Using the AverageFA module, a radiologist can form the hypothesis, “the Average<br />
FA goes down next to an epileptogenic area.” This hypothesis can be answered using the query,<br />
SELECT PatientName<br />
FROM [Database.Tables]<br />
WHERE QueryImage(“AverageFA”, “ Adjacent to Medial Side”, “Perisylvian Area”, “DTI”, “50). The system then fetches the next patient and the cycle repeats itself.<br />
524
No<br />
No<br />
No<br />
Execute Query<br />
Get Next Patient<br />
Is there more<br />
patients<br />
Yes<br />
Is FeatureX<br />
Available?<br />
No<br />
Is Image Map<br />
available?<br />
Yes<br />
Is Model<br />
available?<br />
Yes<br />
Is Registration<br />
Information<br />
available?<br />
Sarmad Istephan and Mohammad Siadat<br />
Yes<br />
Yes<br />
Figure 2: System’s query execution architecture<br />
No<br />
Database<br />
Return Result Set<br />
Generate<br />
SpacialRelationToStructure<br />
Name<br />
Transfer<br />
SpatialRelationOfInterest<br />
model to ImageMap<br />
Calculate FeatureX<br />
Save feature value<br />
Is feature true for<br />
ComaprisonValue Using<br />
ComaprisonOperator?<br />
Yes<br />
Accumulate result<br />
into Result Set<br />
The steps to generate spatial relation to the structure name, the transfer of the relation and the<br />
calculation of feature are depicted in figure 3. This is done by the application first receiving the query,<br />
getting the image map, model and registration information from the database and calculating featurex<br />
using VTK libraries and Image Tool Kit (ITK) libraries. More specifically, VTK library is used for the<br />
following steps. First, load the model data from the database. Once the VTK polydata model is loaded<br />
then transfer the model from the model space to the image space using the registration information. The<br />
transfer of the model to the image space is illustrated in figure 4. Next, a bounding box is drawn around<br />
the model. Then an array containing a 1 for inside and 0 for outside for each voxel from the bounding box<br />
is generated. Figure 5 shows two voxels, once outside the model and one inside the model. Then, ITK<br />
libraries are used to load the image space volume. Finally, the sum of the intensity values for all inside<br />
voxels from the image space are all summed and divided by the total number of inside voxels. This sum<br />
represents the average FA.<br />
525<br />
No
Sarmad Istephan and Mohammad Siadat<br />
Figure 3: Details of the steps generation of the spatial relation to the structure name, the transfer of the<br />
relation and the calculation of feature from Figure 2<br />
Figure 4: The model space transformed to Figure 5: Model showing voxel outside and inside<br />
the image space<br />
4. Technologies used<br />
The data model is implemented in a database that uses SQL Server 2008 technology. This database<br />
stores patient data, various image modalities, registration information and segmented model information.<br />
An ASP.NET web application has been built on top of this database leveraging the .NET Framework 4.0<br />
technology. This web application contains forms in which the available images are imported into the<br />
database. The backend for the forms is written using the C# programming language. Furthermore, the<br />
backend code uses ClearCanvas to read the dicom image tags. In addition, C# VTK library DLLs and C#<br />
ITK library DLLs are used form image and model manipulation. It should be noted that both VTK library<br />
and the ITK library are written in C++ and therefore, CMAKE was used to create the C# DLLs from the<br />
native C++ code. Finally, Matlab was used to generate simulated image volumes and 3D Slicer was used<br />
to generate VTK models for the image data.<br />
5. Simulated data and result<br />
In this system, simulated data was generated and used to test the system because the outcome is<br />
known. It would be difficult to use real MRI/DTI data to verify the system because the outcome is not<br />
known. The simulated data includes sets consisting of a DICOM volume (see figure 8 for an example), a<br />
surface model (see figure 9 for an example) and registration information. The system was verified that it<br />
is able to query the simulated images using the QueryImage function.<br />
The process of generating the simulated data is depicted in figure 6. First, Matlab was used to randomly<br />
generate the model’s parameters. R1, R2, Zmin, Zmax, Xmin, Xmax, Ymin and Ymax. The R1 represents<br />
the radius from the center of the image to the outer edge of the structure. R2 represents the radius from<br />
the center of the image to the inner edge of the structure. In other words, R1 and R2 are used to control<br />
the width of the generated structure. Zmin and Zmax is the minimum and maximum number of slices the<br />
system is to generate (see figure 7). Xmin, Xmax and Ymin, Ymax are the dimensions of the slices. Using<br />
526
Sarmad Istephan and Mohammad Siadat<br />
the values of R1, R2, Zmin, Zmax, Xmin, Xmax, Ymin and Ymax, a model is generated and then saved<br />
as a DICOM image (see figure 8). If the system is to generate more images then another set of<br />
parameters are evaluated creating another DICOM image.<br />
The models are the imported into 3D slicer to generate a geometrical surface model which is then saved<br />
in VTK format. Then, the model is loaded using VTK libraries and a transformation (translation and<br />
rotation) is applied to it. Finally, all this information is saved into the database using the web forms. At this<br />
point the system is ready to be queried.<br />
Figure 6: Simulated data generation<br />
Figure 7: Axial, Coronal and Sagittal views of a slice within the 3 volume<br />
Figure 8: Model created out of the 3D volume<br />
6. Conclusion<br />
This paper created a system that is used to store and query medical images which serves as an<br />
extension to the field of Data Driven Medicine. This system can be used in academia, hospital research<br />
and hospital clinical diagnosis and prognosis. This paper shows the feasibility of such a system and will<br />
be extended into a complete system in future work by adding more module types such as fiber density<br />
measures, curvature measures and standard deviation of the intensity modules.<br />
References<br />
Becker, S. and Arenson Ronald (1994) “Costs and Benefits of Picture Archiving and Communication Systems”,<br />
Journal of the American Medical Informatics Association, vol. 1:4, November, pp.361 – 371.<br />
527
Sarmad Istephan and Mohammad Siadat<br />
Biohealthmatics (2006), Healthcare Technologies [Online], Available:<br />
http://www.biohealthmatics.com/technologies/healthtech.aspx [3 Dec 2010].<br />
Chbeir, R., Atnafu, S. and Brunie, L. (2002) “Image Data Model for an Efficient Multi-Criteria Query: A Case in<br />
Medical Databases”, Proceedings of the 14th International <strong>Conference</strong> on Scientific and Statistical Database<br />
Management IEEE. pp. 165-174.<br />
Glatard, T., Montagnat, J. and Magnin, I. (2004) “Texture Based Medical Image Indexing and Retrieval: Application to<br />
Cardiac Imaging”, Proceedings of the <strong>6th</strong> ACM SIGMM international workshop on Multimedia information<br />
retrieval.<br />
HHS Press Office. (2010) HHS Budget Makes Smart Investments, Protects the Health and Safety of America’s<br />
Families, [Online], Available: http://www.hhs.gov/news/press/2010pres/02/20100201a.html [29 Nov 2010].<br />
Hsu, W., Antani, S., Long, R. Neve, L. and Thoma, G. (2009) “SPIRS: A Web-based image retrieval system for large<br />
biomedical databases”, International journal of medical informatics, vol 78, pp. 13-24.<br />
Jin, H., Sun, A., Zheng, R. He, R. Zhang, Q., Shi, Y., Yang, W. (2007) “Content and Semantic Context Based Image<br />
Retrieval for Medical Image Grid”, GRID '07 Proceedings of the 8th IEEE/ACM International <strong>Conference</strong> on Grid<br />
Computing.<br />
Muller, H., Michoux, N., Bandon D. and Geissbuhler, A. (2004) “A review of content-based image retrieval systems in<br />
medical applications - clinical benefits and future directions”, International Journal of Medical Informatics, vol.<br />
73, pp 1-23.<br />
Siadat, M., Hammad, R., Shetty, A., Soltanian-Zadeh, H., Sethi, I., Eetemadi, A., Elisvich, K. (2009) “DTI data<br />
modeling for unlimited query support”, SPIE, Vol. 7264.<br />
The Office of the National Coordinator for Health Information Technology (2010), Why Health IT? [Online], Available:<br />
http://healthit.hhs.gov/portal/server.pt/community/healthit_hhs_gov__home/1204 [29 Nov 2010].<br />
528
Towards an Evaluation Method for Information Quality<br />
Management of Health Information Systems<br />
Siti Asma Mohammed and Maryati Mohd Yusof<br />
National University of Malaysia, Bangi, Selangor<br />
asmadixie@gmail.com<br />
mmy@ftsm.ukm.my<br />
Abstract: Information quality has been regarded by most researchers as one of the critical success factors of Health<br />
Information Systems (HIS) adoption. However, there is increasing evidence of poor quality information found in HIS.<br />
Poor information quality increases medical errors and adverse events. The occurrence of poor information quality<br />
reveals the need to evaluate information quality management (IQM) practices in healthcare organizations. It is crucial<br />
to assess the effectiveness of IQM to ensure the quality of information produced by HIS. A review of current<br />
evaluation methods for IQM in healthcare has been carried out to identify any existing gaps for further improvement.<br />
The literature review was conducted from the health informatics and information systems (IS) disciplines. It was<br />
found that the lack of information quality was due to a combination of human, organizational and technological<br />
factors. In the literature, guidelines and best practices are suggested in order to manage information quality in<br />
healthcare. However, there is inadequate literature that offers specific approaches and criteria for assessing quality<br />
information from the perspective of humans, organizations and technology. Therefore, this paper proposes a<br />
comprehensive and systematic evaluation method for HIS information quality management in improving the quality of<br />
information in HIS. The proposed method is developed based on PMBOK (Project Management Body of Knowledge)<br />
in order to provide a systematic evaluation process. With regards to evaluation criteria, this research focuses<br />
primarily on the criteria identified from a combination of quality management, IQM and IS/HIS evaluation findings<br />
from the literature. This paper contributes to the information quality body of knowledge within the information systems<br />
area. This evaluation method can be applied as a tool to evaluate the IQM practice in healthcare organizations and<br />
assist HIS users in improving the quality of information.<br />
Keywords: health information systems, information quality management, evaluation, project management, human<br />
factors, organizational factors<br />
1. Introduction<br />
Most researchers consider information quality (IQ) as one of the critical factors for the success of HIS<br />
(DeLone and McLean, 1992; Heeks, 2006; Yusof et al, 2008). Unfortunately, increasing evidence shows<br />
that HIS has not been producing the intended quality of information (Gaikwad et al, 2007; Miettinen and<br />
Korhonen, 2008; Agnew-Blais et al, 2009). Poor IQ increases medical errors and affects the quality of<br />
patient care (Institute of Medicine, 2000). Human, organizational and technological factors are among the<br />
elements that influence poor IQ in HIS (Caccia-Bava, Guimaraes and Harrington, 2006; Kaplan and<br />
Harris-Salamone, 2009). For example, organizational studies show that users are unable to align<br />
information usage with management and clinical performance (Garrib et al, 2008). Further, poor IQ<br />
occurs because of unclear responsibility for IQ and lack of clarity in IQ expectations among users (Kerr,<br />
2006). Lack of quality requirement specification is the reason why users have little understanding of the<br />
purpose of collecting and utilizing information to support decision making (Sutherland and Steinum,<br />
2009). Consequently, users consider less to produce quality information, which later becomes the main<br />
hindrance to adopting HIS (Orfanidis et al, 2004).<br />
2. Theoretical background<br />
Having Information quality management (IQM) practices is necessary to achieve optimal quality of<br />
information. IQM is a set of activities where an organization establishes, controls, conducts quality<br />
strategies so that the process of producing information satisfies user requirements (Caballero et al,<br />
2008). These activities include establishment of roles, responsibilities, IQ requirements and procedures.<br />
IQM provides guidance on the meaning of IQ to the organization (Price and Shanks, 2005). Information is<br />
of high quality if it is fit for use where the information is able to satisfy the actual use of the information<br />
(Tayi and Ballou, 1998). IQ and data quality (DQ) are always used interchangeably in the literature by<br />
most researchers. Therefore, this paper will also use the term IQ interchangeably with DQ unless<br />
specified. There are guidelines and best practices being developed for having strategic IQM in healthcare<br />
organizations (AHIMA, 1998; Arts et al, 2002; Kerr, 2006; Aqil, Lippeveld and Hozumi, 2009). However,<br />
the most critical factors for IQM and the criteria are still unclear to benchmark the quality of HIS<br />
information (Ammenwerth, Breu and Paech, 2010). In addition, more research is needed to study the<br />
practicalities of information use such as the relationships between users and the HIS in disseminating<br />
information, the dynamic changes of clinical relationship in addressing information discrepancy and other<br />
529
Siti Asma Mohammed and Maryati Mohd Yusof<br />
stakeholders’ influences (Greenhalgh et al, 2009). Evaluation can be undertaken to investigate the IQM<br />
practice in healthcare organizations for effective HIS. A more holistic evaluation approach is needed to<br />
understand the production of IQ by HIS given human and organizational interventions. To date, there is<br />
inadequate literature that offers a comprehensive and systematic evaluation approach together with<br />
specific criteria for having quality information. Hence, more work is required to investigate this issue<br />
further.<br />
2.1 Previous works on HIS information quality management<br />
Table 1 below shows the review of best practices that have been developed to strategically manage<br />
information production in HIS. This review attempts to identify the gap in HIS IQM.<br />
Table 1: Selected Works on HIS IQM<br />
Author Objective Information<br />
Life Cycle<br />
Factors and Criteria<br />
Method<br />
P A M A H O T<br />
Data Quality<br />
Assurance in<br />
Medical<br />
Registries<br />
(Arts et al,<br />
2002 )<br />
AHIMA Data<br />
Quality<br />
Management<br />
Model<br />
(AHIMA,<br />
1998)<br />
Data Quality<br />
Improvement<br />
Strategy<br />
(Kerr, 2006)<br />
PRISM<br />
(Aqil,<br />
Lippeveld<br />
and Hozumi,<br />
2009)<br />
1) To identify<br />
the type<br />
and causes of<br />
poor data<br />
quality<br />
2) To develop<br />
data quality<br />
assurance<br />
procedures for<br />
medical<br />
registries<br />
To ensure data<br />
quality<br />
and continuous<br />
quality<br />
improvement<br />
To<br />
institutionalize<br />
and establish<br />
strategic<br />
management of<br />
data quality in<br />
the New<br />
Zealand health<br />
sector<br />
To evaluate the<br />
performance of<br />
HIS<br />
from the<br />
perspective of<br />
information<br />
quality handling<br />
X X X Training<br />
Motivation<br />
Quality follow<br />
up<br />
X X X Data collection<br />
purpose<br />
Education<br />
Training<br />
Communicate<br />
changes and<br />
updates<br />
Accountability<br />
Quality data<br />
sources<br />
X X X Quality data<br />
source<br />
Training<br />
Quality<br />
awareness<br />
Peer review<br />
Frequent<br />
checks,<br />
analysis and<br />
profiling of<br />
data<br />
Data quality<br />
mapping<br />
(business<br />
rules and data<br />
quality<br />
metrics)<br />
X X X Checking skill<br />
HIS tasks<br />
competency<br />
Motivation<br />
Confidence for<br />
HIS tasks<br />
530<br />
Data<br />
requirements<br />
Quality<br />
assurance plan<br />
Ethics<br />
Scope<br />
Information<br />
requirements<br />
Processes<br />
Documentation<br />
Monitoring<br />
Regulation<br />
Data<br />
stewardship<br />
Governance<br />
Planning<br />
Resources<br />
Training<br />
Supervision<br />
Information<br />
quality<br />
culture<br />
promotion<br />
Database<br />
design and<br />
function<br />
Technical<br />
support<br />
Interface<br />
usability<br />
Data<br />
ownership<br />
Accessibility<br />
Correct<br />
algorithms<br />
and formulas<br />
Data<br />
currency<br />
Database<br />
normalization<br />
Database<br />
maintenance<br />
and support<br />
Database<br />
design and<br />
normalization<br />
Interface<br />
System<br />
maintenance<br />
Interface<br />
complexity<br />
HIS design<br />
IT complexity<br />
Formative<br />
Formative<br />
Formative<br />
Formative
Siti Asma Mohammed and Maryati Mohd Yusof<br />
Note: Information Life Cycle: P – Plan, A – Acquisition, M – Maintenance, A - Apply<br />
Factors: H – Human, O – Organization, T – Technology<br />
Generally, all the four methods mentioned in Table 1 provide clear guidance for having quality information<br />
in HIS. For example, Arts, Keizer and Scheffer developed a generic framework that outlines the<br />
procedures for data quality assurance for medical registry system. The procedures are divided into three<br />
categories; system registry setup and organisation (plan), detection during data collection (acquisition)<br />
and actions to be taken to improve data quality (maintain).<br />
As for American Health Information Management Association (AHIMA), an IQM model is developed to<br />
continuously improve IQ. The model includes data application (purpose for which data are collected),<br />
data collection (process by which data elements are accumulated), warehousing (processes and systems<br />
used to archive data) and analysis (process of translating data).<br />
Another IQM initiative, which is a DQ improvement strategy, has been developed by Kerr to promote DQ<br />
in New Zealand. This strategy was designed at the Ministry of Health New Zealand with involvement from<br />
healthcare stakeholders in New Zealand. This program is meant to provide DQ standard operating<br />
procedures for healthcare organizations in New Zealand.<br />
Aqil, Lippeveld and Hozumi developed a PRISM model that details the determinants influencing HIS<br />
performance from the perspective of IQ handling. The behavioral factors, which explain the human<br />
behavior in handling data, discussed briefly on the use of information technology. For example, HIS tasks<br />
competency and the confidence level of using HIS.<br />
The similarity between these studies is the emphasis on IQ behavior or IQ handling, which comes under<br />
the human factor. Human factors such as training, motivation, education and quality awareness are<br />
among the important determinants that will influence IQ. DQ requirements, clinical process, management<br />
supervision, IQ culture and IQ governance are the examples for organization factors. For the technical<br />
aspects, while AHIMA focusing on the stability of database design to achieve IQ, the other three methods<br />
include the importance of HIS design such as interface usability. On the other hand, Aqil, Lippeveld and<br />
Hozumi depicts that information technology usage will also affect IQ.<br />
It can be seen that all methods share similar opinion on the importance of quality awareness, DQ<br />
requirements and IT technical support in ensuring IQ. Further, the methods are applied across the<br />
information life cycle (ILC) through a formative evaluation. This research believes more can be added to<br />
the evaluation criteria proposed within the human, organizational and technological factors. There should<br />
be more explicit evaluation categories and evaluation criteria that can be assigned to each ILC. Thus, a<br />
holistic evaluation method that includes criteria from human, organization and technology factors<br />
throughout the ILC would offer a better solution.<br />
2.2 Project Management Body of Knowledge (PMBOK)<br />
In designing a systematic evaluation approach, this research recognizes the importance of project<br />
management discipline. We believe project management is important when conducting an evaluation.<br />
This is because project management is based on planning, organizing and controlling each activity so<br />
that evaluation objective is achieved (Irani, 2010). As such, a project management discipline, Project<br />
Management Body of Knowledge (PMBOK), becomes the basis for developing the evaluation method<br />
(PMI, 2008). PMBOK provides common operational steps with clear inputs and outputs. PMBOK is<br />
recognized as a good practice and contributes to project success and organization value (Zwikael, 2009).<br />
PMBOK defines project management as ‘the application of skills, tools and techniques to project activities<br />
to meet project requirements’ (PMI, 2008:6). PMBOK includes a collection of processes to manage all<br />
types of projects. These process groups are performed sequentially and iteratively prior to completing the<br />
project. Each process group has its individual processes that are linked by their inputs and outputs. The<br />
output of one process becomes the input of another. Tools and techniques are the mechanisms applied<br />
to create the output.<br />
In addition, most studies in IS evaluation have failed to focus on understanding the ‘how’ aspect or the<br />
evaluation process itself (Tuten, 2009; Irani, 2010). This may explain why IS evaluation in healthcare<br />
531
Siti Asma Mohammed and Maryati Mohd Yusof<br />
organization is unclear and does not provide a strong basis for a good evaluation theory and an effective<br />
evaluation method (Ammenwerth et al, 2004).<br />
Thus, this research believes that PMBOK complements IS evaluation in the sense that PMBOK provides<br />
more guidance and details on how to plan, execute, manage and control the evaluation method. This<br />
research uses PMBOK to develop the evaluation method to support users in improving the quality of<br />
information in HIS continuously.<br />
3. Proposed evaluation method for HIS information quality management<br />
The proposed evaluation model (Figure 1) is developed based on PMBOK (PMI, 2008:39). It describes<br />
the process for conducting the evaluation. This is done by phases throughout the ILC. In order to find the<br />
specific criteria that determine the production of IQ in HIS, this research makes use most of the criteria<br />
identified from a combination of quality management, IQ and IS/HIS evaluation findings. Three major<br />
factors that influence the production of quality information are identified; human, organization and<br />
technology. Even though those criteria do not explicitly focus on IQ, they are all relevant and applicable in<br />
evaluating HIS IQM. Some criteria used are also based on IQM best practices given in the literature such<br />
as the information behavior criteria. Important criteria pertaining to the use of HIS and user satisfaction<br />
towards the system in use are also included after reviewing the literature from health informatics. The<br />
way a user behaves towards the information he possesses, the way he uses the system and his<br />
satisfaction towards the system used will determine the quality of information produced. This research<br />
makes use of the HOT-fit (Human, Organization and Technology-fit) framework in categorizing the factors<br />
and dimensions (Yusof et al, 2008). This is because HOT-fit gives explicit evaluation categories, and<br />
indicates clear relationships and definitions between the evaluation factors and dimensions. However, the<br />
detail criteria are the findings from literature reviews pertaining to effective IQM. Therefore, criteria used<br />
for this study are categorized under different suitable factors and dimensions.<br />
Figure 1: Evaluation method for HIS information quality management<br />
The model above describes the evaluation phases, key activities, ILC involved under each evaluation<br />
phase, input (evaluation resource), evaluation output (deliverables), criteria for each ILC involved and the<br />
methods used to conduct the evaluation. Each evaluation phase has its own activities, inputs/outputs,<br />
criteria and methods. The evaluation criteria are distributed according to different ILC being evaluated.<br />
Some criteria are repeatedly included in different ILCs depending on its relevance. Below are the<br />
definitions for ILC stages for this research.<br />
Table 3: ILC stages definitions (adapted from English, 1999)<br />
ILC Stage Definition<br />
Plan Plan for resources including information requirements, IQ definitions and metrics, process design,<br />
information modeling, database architecture, information resources, user roles, system usage<br />
preparation and method for collecting the information.<br />
Acquisition Process of collecting, verifying and storing the data/information into HIS.<br />
Maintenance Ensure that all resources continue to operate properly to make the information assessable and<br />
usable. These include updating, changing, cleansing and validating data and system maintenance.<br />
Apply Information is used to accomplish objectives.<br />
532
3.1 Evaluation criteria<br />
3.1.1 Human<br />
Siti Asma Mohammed and Maryati Mohd Yusof<br />
System Use refers to the frequency and breadth of HIS use (Yusof et al, 2008). System Use is<br />
associated with higher performance of quality of care (Poon et al, 2010). System Use includes the<br />
training provided for the HIS users in using the system, the level of system acceptance, the level of<br />
motivation to use the system and their knowledge of the system that is beyond their current job<br />
assignments (Yusof et al, 2008). User role is another criterion for System Use. Clear user roles when<br />
using the system determines the quality of information in HIS. User roles refer to data collectors, data<br />
custodians and data consumers (Kerr, 2006; Wang et al, 1998). Information Behavior is the HIS user’s<br />
effort in ensuring quality in collecting, creating and using the information. Information Behavior criteria<br />
include education and training on medical information, and quality awareness on the information<br />
possessed. The information must be patient-centred and should satisfy the actual use of data consumers<br />
(Macinati, 2008; Ahire, 1996).<br />
3.1.2 Organization<br />
The organization Structure of healthcare organization has a major influence on the production of HIS IQ.<br />
Structure includes leadership, physician involvement, IQ governance, IQ culture, IQ team and IQ<br />
requirements. Leadership indicates the role and commitment of the top management in motivating quality<br />
culture among the HIS users (Wang and Strong, 1996; English, 1999). IQ culture is a very important<br />
criterion that is emphasized by English (English, 1999). It refers to organization culture transformation<br />
that regards IQM activities as a habit and mind-set by applying continuous quality improvement<br />
principles. IQ team is the main entity that initiates and encourages IQM activities. IQ requirements are<br />
part of the IQM activities where IQ standards and definitions are developed during the ILC planning<br />
stage.<br />
3.1.3 Technology<br />
System Quality affects the production of IQ. Database design, ease of use, availability, usefulness and<br />
data security and are among the important criteria for System Quality (Ammenwerth, Breu and Paech,<br />
2010). Database design refers to the data inside the database that is efficiently organized for easy<br />
retrieval and reduces anomalies. Ease of use includes the convenient of using HIS such as the interface<br />
usability. Availability refers to the run time of the HIS for a desirable or expected length of time. Service<br />
Quality is the overall support received from the IT team within the healthcare organization and from the<br />
HIS vendor. Criteria for Service Quality are responsiveness, follow-up, assurance and empathy (Yusof et<br />
al, 2008)<br />
3.2 Evaluation phase<br />
3.2.1 Initiation<br />
Evaluation initiation is the process of initiating a new evaluation project (PMI, 2008:44). During the<br />
initiation phase, the stakeholders with the most knowledge of the clinical process, medical information<br />
being collected, the HIS in use and those responsible for maintaining the quality of information, are<br />
identified. The stakeholders include the quality team, medical officers, physicians, nurses and HIS users.<br />
The stakeholders must be those who interact and influence the overall outcome of the evaluation project.<br />
These people will become the input for this evaluation phase and throughout the evaluation project. An<br />
analysis of the stakeholders, to understand the degree of influential and impact to the evaluation, will be<br />
done using Social Network Analysis (SNA). During this phase, the evaluation goal and the project scope<br />
are explained. The evaluator informs the stakeholders of the goal, the reasons for conducting the<br />
evaluation and the significance of the evaluation that triggered the evaluation. All parties are expected to<br />
agree to the evaluation before the scope is further defined. A decision is made; whether to continue,<br />
delay or discontinue the evaluation. The evaluation scope defines the evaluation boundaries, which<br />
include the parts of the process and system that are to be studied. The time and resources needed to<br />
conduct the evaluation are also negotiated during this initiation phase. The deliverables for this phase are<br />
the Evaluation Charter and Stakeholder Analysis. The Evaluation Charter formally authorizes the<br />
evaluation project and documents initial requirements that satisfy the stakeholder’s needs and<br />
expectations. Table 4 explains the parameters for evaluation initiation phase.<br />
533
Table 4: Parameters for evaluation initiation phase<br />
3.2.2 Planning<br />
Siti Asma Mohammed and Maryati Mohd Yusof<br />
Key Activity Identify stakeholders<br />
Sub-activities 1. Identify stakeholders<br />
2. Brief evaluation goal<br />
3. Discuss and agree on evaluation scope<br />
ILC -<br />
Input Stakeholders<br />
Evaluation Criteria -<br />
Output Evaluation charter, Stakeholder analysis<br />
Method Interviews, SNA<br />
Evaluation planning is the process of designing the evaluation (PMI, 2008:46). During this phase, the<br />
evaluator attempts to learn about the current system, the clinical process, the environment supporting the<br />
clinical process and the IQ requirements. The evaluator will try to understand the problem domain in<br />
which the IQ problems exist and assesses the status of quality management effort concerning the<br />
planning and strategy for having quality information in HIS. The quality management team is expected to<br />
describe and guide the evaluator on the initiatives about IQ. The aim of the appraisal is to understand the<br />
IQM effort at the planning stage of the ILC. Further, the evaluator will identify HIS user roles involved in<br />
the different stages of ILC and their degree of dependencies using SNA. The key deliverables for this<br />
phase will be Evaluation Plan, Workflow Analysis and SNA. The Evaluation Plan defines how to execute<br />
the evaluation. Table 5 describes the parameters for evaluation planning phase.<br />
Table 5: Parameters for evaluation planning phase<br />
Key Activity Quality management environment analysis<br />
Sub-activities 1. Review of strategic and operative environment<br />
2. Understand and assess the status of quality management effort<br />
ILC Plan<br />
Input Evaluation Charter, stakeholders<br />
Evaluation Criteria<br />
Human Organization Technology<br />
System Use<br />
Training<br />
User role<br />
Information Behavior<br />
Education and training<br />
Structure<br />
Management leadership<br />
Physician involvement<br />
IQ governance<br />
IQ culture<br />
IQ team<br />
IQ requirements<br />
Output Evaluation plan, Workflow Analysis, SNA, IQ standards<br />
Method Interviews , Document Analysis, Observations<br />
System Quality<br />
Database design<br />
Ease of use<br />
Availability<br />
Usefulness<br />
Data security<br />
3.2.3 Execution<br />
Evaluation execution is the process of performing the evaluation at the operational level where medical<br />
information is collected, stored, maintained and applied (PMI, 2008:55). The aim of this assessment is to<br />
get an overview of information production during the delivery of care. The criteria for this phase will be<br />
used to guide the evaluation. During the assessment, any possible issues are monitored and captured.<br />
The problems underlying these issues are analyzed. Table 6, Table 7 and Table 8 describe the<br />
parameters for execution phase according to three different stages of ILC, which are acquisition,<br />
maintenance and apply.<br />
534
Siti Asma Mohammed and Maryati Mohd Yusof<br />
Table 6: Parameters for evaluation execution phase (ILC - acquisition)<br />
Key Activity Information production process and business impact assessment<br />
Sub-activities 1. Records quality activities<br />
2. Monitor and capture various possible issues<br />
3. Analyze problems<br />
4. Identify, update and refine variances from the evaluation project<br />
ILC Acquisition<br />
Input Evaluation plan, Workflow Analysis, SNA, Quality team, HIS users<br />
Evaluation Criteria<br />
Human Organization Technology<br />
System Use<br />
Acceptance<br />
System motivation<br />
Knowledge/skills<br />
Information Behavior<br />
Quality awareness<br />
Patient centered<br />
Structure<br />
Physician involvement<br />
IQ culture<br />
Output Evaluation results, cause-and-effect analysis<br />
System Quality<br />
Database design<br />
Ease of use<br />
Availability<br />
Usefulness<br />
Data security<br />
Service Quality<br />
Responsiveness<br />
Follow up<br />
Assurance<br />
Empathy<br />
Method Interviews, Observations, Trend analysis, Expert and user judgments<br />
Table 7: Parameters for evaluation execution phase (ILC - maintenance)<br />
Key Activity System maintenance assessment<br />
Sub-activities 1. Investigate maintenance efforts<br />
2. Monitor and capture various possible issues<br />
3. Analyze problems<br />
4. Identify, update and refine variances from the evaluation project<br />
ILC Maintenance<br />
Input Evaluation plan, Workflow Analysis, Quality team, HIS users<br />
Evaluation Criteria<br />
Human Technology<br />
Information Behavior<br />
Quality awareness<br />
Motivation<br />
Patient centered<br />
Output Evaluation results, cause-and-effect analysis<br />
System Quality<br />
Database design<br />
Ease of use<br />
Availability<br />
Usefulness<br />
Data security<br />
Service Quality<br />
Responsiveness<br />
Follow-up<br />
Assurance<br />
Empathy team<br />
Vendor<br />
Method Interviews, Observations, Trend analysis, Expert and user judgments<br />
535
Siti Asma Mohammed and Maryati Mohd Yusof<br />
Table 8: Parameters for evaluation execution phase (ILC - apply)<br />
Key Activity IQ and business impact assessment<br />
Sub-activities 1. Assess and map the IQ usefulness based on users’ predefined IQ standards at planning<br />
stage.<br />
2. Assess the information satisfaction<br />
3. Identify, update and refine variances from the evaluation project<br />
ILC Apply<br />
Input Evaluation project plan, Workflow Analysis, Quality team, HIS users<br />
Evaluation<br />
Human<br />
Criteria Information Behavior<br />
Information satisfaction<br />
Output Evaluation results, cause-and-effect analysis<br />
Method Interviews, Observations, Trend analysis, Expert and user judgments<br />
3.2.4 Monitor and control<br />
Monitoring and controlling the evaluation occurs at every evaluation phase especially during the biggest<br />
part of evaluation which is the execution phase (PMI, 2008:59). Monitoring is to track the evaluation<br />
progress and to identify any variances or issues during evaluation. If necessary, the evaluator may<br />
control the issues by adjusting the scope, plan or resources. A progress report is necessary to gain an<br />
insight into the evaluation progress and to identify areas that need attention.<br />
3.2.5 Closure and actions<br />
The closure phase consists of activities to finalize and complete the evaluation project (PMI, 2008:64).<br />
During this phase, the evaluator communicates the evaluation findings and recommendations to the<br />
stakeholders. Experiences and lessons learned during evaluation are shared with the stakeholders.<br />
Based on the findings and lessons learned, this phase should be a continuous process of improvement.<br />
Future actions must be considered and implemented. An evaluation plan must be included to assess<br />
these future actions in order to continuously improve IQM practices in healthcare organization.<br />
4. Conclusions<br />
This paper has highlighted the issues of poor IQ in HIS, reviewed the existing works on IQM in healthcare<br />
organizations and proposed an evaluation method for IQM evaluation. From the literature, we found that<br />
poor IQ in HIS is influenced by three factors; human, organizational and technological factors. Review of<br />
current IQM practices in healthcare organization suggests that a comprehensive criteria for having quality<br />
information in HIS is needed. In addition, there is a need for a systematic and effective evaluation<br />
method. Thus, an evaluation method for IQM of HIS that is based on project management discipline is<br />
proposed. The evaluation method is designed based on PMBOK, to complements the method for<br />
conducting IS evaluation. In this paper, we have justified the need for project management knowledge in<br />
IS evaluation. The evaluation method also includes the evaluation criteria based on human,<br />
organizational and technological factors, which are assigned to each ILC according to its relevance. The<br />
evaluation method is meant to evaluate the IQM practices for HIS at the operational level.<br />
We believe this research contributes to IS discipline especially related to IQ and HIS evaluation. First, this<br />
research extends the knowledge that IQ is also influenced by the user’s behavior towards the system in<br />
use. Second, this research takes a holistic approach to examine the production of IQ in healthcare<br />
organization, which goes beyond usual studies that evaluate IQ to validate the HIS efficiency in producing<br />
quality information without taking into consideration the factors that affect quality. This research<br />
addressed the need for a better understanding of the HIS evaluation process by adapting project<br />
management discipline. Lastly, this research proposes a systematic evaluation method for HIS<br />
information quality management to support the users in improving continuously the quality of information<br />
in the HIS. For future work, opinions from experts are required to verify the criteria for having quality<br />
information and the evaluation method. Afterward, the evaluation method will be validated in healthcare<br />
organization for further refinement. This research will use qualitative study to gain an in-depth knowledge<br />
about HIS IQM in healthcare organization, with an intention to develop directions or predictions about the<br />
practice of HIS users in producing quality information.<br />
536
References<br />
Siti Asma Mohammed and Maryati Mohd Yusof<br />
Agnew-Blais, J. C., Coblyn, J. S., Katz, J. N., Anderson, R. J., Mehta, J. & Solomon, D. H. (2009) "Measuring Quality<br />
of care for rheumatic diseases using an electronic medical record". Annals Rheumatic Diseases, vol. 68, no. 5,<br />
May, pp. 680-684.<br />
AHIMA (1998) "Data Quality Management Model". Journal of AHIMA, vol 69, no. 6, June, pp 2-7.<br />
Ahire, S. L., Golhar, D. Y. & Waller, M. A. (1996) "Development and validation of TQM implementation constructs".<br />
Decision Sciences, vol. 27, no. 1, March, pp. 23-56.<br />
Ammenwerth, E., Brender, J., Nykanen, P., Prokosch, H.-U., Rigby, M. & Talmon, J. (2004) "Visions and strategies to<br />
improve evaluation of health information systems Reflections and lessons based on the HIS-EVAL workshop in<br />
Innsbruck". International Journal of Medical Informatics, vol. 73, no. 6, June, pp. 479-491.<br />
Ammenwerth, E., Breu, R. & Paech, B. (2010) User-Oriented Quality Assessment of IT-Supported Healthcare<br />
Processes - A Position Paper. In Rinderle-Ma, S., Sadiq, S. & Leymann, F. (Eds.). Business Process<br />
Management Workshop 2009. Springer Berlin Heidelberg.<br />
Ammenwerth, E., Ehlers, F., Hirsch, B. & Gratl, G. (2007) "HIS-Monitor: An approach to assess the quality of<br />
information processing in hospitals". International Journal of Medical Informatics, vol. 76, no. 2-3, February, pp<br />
216-225.<br />
Aqil, A., Lippeveld, T. & Hozumi, D. (2009) "PRISM framework: a paradigm shift for designing, strengthening and<br />
evaluating routine health information systems ". Health Policy and Planning, vol. 24, March, pp 217-228.<br />
Arts, D. G. T., Keizer, N. F. d. & Scheffer, G.-J. (2002) "Defining and Improving Data Quality in Medical Registries: A<br />
Literature Review, Case Study, and Generic Framework". Journal of the<br />
American Medical Informatics Association, vol. 9, no.6, Nov/Dec, pp. 600-611.<br />
Caballerro, I., Caro, A., Calero, C. & Piattini, M. (2008) "IQM3: Information Quality Management Maturity Model".<br />
Journal of Universal Computer Science, vol. 14, no. 2, December, pp 3658-3685.<br />
Caccia-Bava, M. d. C., Guimaraes, T. & Harrington, S. J. (2006) "Hospital organization culture, capacity to innovate<br />
and success in technology adoption". Journal of Health Organization and Management, vol. 20, no. 3, pp 194-<br />
217.<br />
DeLone, W. H. & McLean, E. R. (1992) "Information Systems Success: The Quest for Dependent Variable".<br />
Information Systems Research, vol. 3, no. 1, pp. 60-95.<br />
English, L. P. (1999) Improving Data Warehouse and Business Information Quality, New York, John Wiley & Sons,<br />
Inc.<br />
Gaikwad, R., Sketris, I., Shepherd, M. & Duffy, J. (2007) "Evaluation of accuracy of drug interaction alerts triggered<br />
by two electronic medical record systems in primary healthcare". Health Informatics Journal, vol. 13, no. 3,<br />
September, pp. 163-177.<br />
Garrib, A., Stoops, N., McKenzie, A., Dlamini, L., Govender, T., Rhode, J. & Herbst, K. (2008) "An evaluation of the<br />
District Health Information System in rural South Africa". S Afr Med, vol. 98, no. 7, July, pp. 549-522.<br />
Greenhalgh, T., Potts, H. W. W., Wong, G., Bark, P. & Swinglehurst, D. (2009) "Tensions and Paradoxes in<br />
Electronic Patient Record Research: A Systematic Literature Review Using the Meta-narrative Method". The<br />
Milbank Quarterly, vol. 87, no. 4, pp 729-788.<br />
Heeks, R. (2006) "Health information systems: Failure, success and improvisation". International Journal of Medical<br />
Informatics, vol. 75, no. 2, pp 125-137.<br />
Institute of Medicine (2000) To Err is Human: Building a Safer Health System, Washington D.C., National Academy<br />
Press.<br />
Irani, Z. (2010) "Investment evaluation within project management: an information systems perspective". Journal of<br />
the Operational Research Society, vol. 61, no. 6, June, pp. 917-928.<br />
Kaplan, B. & Harris-Salamone, K. D. (2009) "Health IT Success and Failure: Recommendations from Literature and<br />
AMIA Workshop". Journal of the American Medical Informatics Association, vol. 16, no. 3, May-June, pp. 291-<br />
299.<br />
Kerr, K. (2006) The institutionalisation of data quality in the New Zealand health sector. PhD Thesis University of<br />
Auckland.<br />
Macinati, M. S. (2008) "The relationship between quality management systems and organizational performance in<br />
the Italian National Health Service". Health Policy, vol. 85, no. 2, February, pp. 228-241.<br />
Miettinen, M. & Korhonen, M. (2008) "Information Quality in Healthcare: Coherence of Data Compared Between<br />
Organization's Electronic Patient Records". CBMS '08 Proceedings of the 2008 21st IEEE International<br />
Symposium on Computer-Based Medical Systems, pp. 488 - 493<br />
Orfanidis, L., Bamidis, P. D. & Eaglestone, B. (2004) "Data quality Issues in electronic health records: An adaptation<br />
framework for the Greek health system". Health Informatics Journal, vol. 10, no. 1, March, pp. 23-26.PMI (2008)<br />
PMBOK: A Guide to the Project Management Body of Knowledge 4th Edition. Project Management Institute<br />
Communications.<br />
Poon, E. G., Wright, A., Simon, S. R., Jenter, C. A., Kaushal, R., Vol, L. A., Cleary, P. D., Singer, J. A., Tumolo, A. Z.<br />
& Bates, D. W. (2010) "Relationship Between Use of Electronic Health Record Features and Health Care<br />
Quality". Med Care, vol 48, no. 3, March, pp 203-209.<br />
Price, R. J. & Shanks, G. (2005) "Empirical Refinement of a Semiotic Information Quality Framework". Proceedings<br />
of the 38th Hawaii International <strong>Conference</strong> on System Sciences, pp. 1-10.<br />
Sutherland, J. M. & Steinum, O. (2009) "Hospital factors associated with clinical data quality". Health Policy, vol. 91,<br />
no. 3, August, pp. 321-326.<br />
537
Siti Asma Mohammed and Maryati Mohd Yusof<br />
Tayi, G. K. & Ballou, D. P. (1998) "Examining Data Quality". Communications of the ACM, vol. 41, no. 2, February,<br />
pp. 54-57.<br />
Tuten, P. M. (2009) A Model for the Evaluation of IS/IT Investments. PhD Thesis Graduate School of Computer and<br />
Information Sciences. Nova Southeastern University.<br />
Wand, Y. & Wang, R. Y. (1996) "Anchoring Data Quality Dimensions in Ontological Foundations". Communications<br />
of the ACM, vol. 39, no. 11, November, pp. 86-95.<br />
Wang, R. Y. & Strong, D. M. (1996) "Beyond Accuracy: What Data Quality Means to Data Consumers". Journal of<br />
Management Information Systems, vol. 12, no. 4, pp. 5-34.<br />
Yusof, M. M., Kuljis, J., Papazafeiropoulou, A. & Stergioulas, L. K. (2008) "An evaluation framework for Health<br />
Information Systems: human, organization and technology-fit factors (HOT-fit)". International Journal of Medical<br />
Informatics, vol. 77, no. 6, pp. 386-398.<br />
Zwikael, O. (2009) "The Relative Importance of the PMBOK Guide's Nine Knowledge Areas During Project<br />
Planning". Project Management Journal, vol. 40, no. 4, December, pp. 94-103.<br />
538
Developing a Change Management Model for Iranian<br />
<strong>Academic</strong> Libraries: A Delphi Study<br />
Maryam Nakhoda 1 and Sirous Alidousti 2<br />
1<br />
University of Tehran, Tehran, Iran<br />
2<br />
Iranian Research Institute for Information Science and Technology (IranDoc),<br />
Tehran, Iran<br />
mnakhoda@ut.ac.ir<br />
alidousti@irandoc.ac.ir<br />
Abstract: Purpose- This paper aims to develop a change management model for Iranian academic libraries. Change<br />
is inevitable in organizations where it is now experienced as a daily phenomenon. <strong>Academic</strong> libraries perform in the<br />
rapidly growing environment of education and research and they need to be active in facing change in order to<br />
realize their mission. To facilitate the planning and implementation of change, several models have been designed in<br />
the past, but these are either too general or have been developed in specific contexts and cannot easily be adopted<br />
in other contexts. In order to overcome the challenges of change and to improve change management, Iranian<br />
academic libraries need their own change model. Design/methodology/approach-The model was designed based on<br />
a qualitative approach and a Delphi study. A panel of twenty-two experts in Management and Library Science and<br />
Information Science contributed to the design and validated the model in three rounds of Delphi. The input of the<br />
Delphi process was the “Information Services Framework for Managing Change” by Penfold and 54 predefined<br />
actions for change extracted from change models related to libraries. The experts were asked to allocate each action<br />
to one (or more) of the stages of the basic model. The panel members could also add their own preferred actions to<br />
the list. The consensus was measured by percentage agreement (60% or greater) for each item. Findings- Thirteen<br />
actions were added by the experts to the initial list. From the resulting 67 actions, only one of them was rejected by<br />
the panel, leaving 66 actions, which were consolidated to 58. These actions were then divided into five stages:<br />
reviewing the current status, analysing and defining library users, planning, implementing the change, and<br />
maintaining the change. After defining the problem, defining change management and describing the respective<br />
model, the methodology of the study and the findings were discussed and finally suggestions for future work were<br />
made.<br />
Keywords: change; change management; change models; academic libraries; Iran<br />
1. Introduction<br />
Modern organizations encounter more extensive changes. Dealing with these changes effectively<br />
requires planning. Change management and its models discuss how to guide an organization during the<br />
change process. Due to their pivotal role in education, libraries and information centres, especially<br />
academic libraries, need strategies for efficient management of change. This paper presents the research<br />
findings to design a change management model for academic libraries, especially in Iran.<br />
2. Statement of the problem<br />
Studies have shown that more than 70% of all efforts for organizational change have been unsuccessful<br />
(By 2005 and Goodrich 2008) mostly because of unfamiliarity with change management (Cicmil 1999 and<br />
Smith 2006). In addition, libraries encounter changes caused by internal and external stimulants (Mark<br />
1997and Pugh 2007). Hirshon (1999) referred to the importance of change management for the<br />
academic libraries worldwide.<br />
Iran is a developing country with an increasing academic growth. Advances in information technologies<br />
and the rise in the number of universities, students, and researchers (Iran’s statistical yearbook 2008)<br />
pave the way of academic improvements in Iran. Such factors create challenges for Iranian academic<br />
libraries which have to be dealt with in order to fulfil their roles. Therefore the need for change in the<br />
academic libraries’ services, spaces, and materials’ formats would appear. According to the importance<br />
of context in developing the managerial processes, the change model of the Iranian academic libraries<br />
must be devised according to their special status and the environment in which they perform their<br />
activities.<br />
The objective of this study is presenting a change model in order to improve the implementation of<br />
changes in Iranian academic libraries as well as increasing the efficiency by reducing the costs and<br />
redundancy.<br />
539
Maryam Nakhoda and Sirous Alidousti<br />
This study contributes to providing a perfect change model for academic libraries, especially academic<br />
libraries of Iran. Also it is of value to academic libraries’ directors to aid them in managing changes in<br />
their libraries.<br />
3. Change management and change models<br />
Change can be defined as kind of state conversion with its tendency being from a relatively undesirable<br />
state towards a more desirable one. Change management and change models are branches of<br />
management studies. Paton, Beranek, and Smith (2008) mentioned the functions of a change<br />
management plan. In change management at least three fields are studied: the current state of the<br />
organization, the state the organization should reach in the future, and finally how to guide this<br />
conversion from the current state to the desired state.<br />
Several studies have been done on change management in academic libraries (Clarke and Morris 1998;<br />
Konings and Dekker 2005; and Ferguson 2007). However, there is little evidence of efforts to apply<br />
existing change models in academic libraries (Sidorko 2008). The first change models followed a threestage<br />
process, including preparation of the organization for change, change implementation and<br />
consolidation of new procedures in the organizational culture. For instance, Lewin’s change model (1974)<br />
consisted of three stages, namely “unfreezing”, “moving”, and “refreezing” (Gilley, Godek, and Gilley<br />
2009). Herein, the term “model” denotes a systematic description of the stages and activities necessary<br />
to carry out a change. The literature review reveals common characteristics among a relatively large<br />
number of change models. Burke (2002) refers to two common functions of change models, namely,<br />
“conversion of organization’s diverse features to simple and manageable parts” and “helping to focus on<br />
more significant parts” in the change process.<br />
4. Research methodology<br />
In the present study, a qualitative approach and the Delphi method in three rounds were applied. The<br />
Delphi method was firstly employed in 1950s in defence sciences research by the RAND Company and<br />
has been used in other fields as well (Green 2000; Ludwig and Starr 2005). Reaching a judgment based<br />
on consensus could be considered as the outcome of a Delphi study (Linstone and Turoff 1975). The<br />
three major characteristics of the Delphi method are “anonymity” of members to each other, “feedback”<br />
and “iteration” (Ambrosiadou and Goulis 1999). The researcher collects the results of the first round of<br />
the study and sends them back again to the panel of experts. After providing the panel of experts with the<br />
results of the first round, each of them expresses his/her opinion once more. This process may continue<br />
for several rounds until consensus is finally reached (Bradley and Stewart 2002).<br />
Among the applications of the Delphi method, which are close to the aim of the present study, design and<br />
validation of models in other fields can be mentioned (Sharma et al. 2008; Minkman et al. 2009; and<br />
Minkman, Ahaus and Huijsman 2009).<br />
4.1 Selection of Delphi panel members<br />
In Delphi studies, the “panel of informed individuals” or “experts” possess knowledge of the subject under<br />
study. In the determination of these experts, purposive sampling is often used (Hasson, Keeney, and<br />
McKenna 2000), in which participants are selected to solve a specific problem using their knowledge.<br />
Concerning the size of the sample, Okoli and Pawlowski (2004) considered ten to twenty qualified<br />
experts as a valid sample.<br />
In the current research, members of the Delphi panel were selected by purposive sampling. A list of<br />
twenty-nine individuals appropriate for participating in this study was initially prepared. Twenty-two of<br />
them expressed their willingness to participate in the first round. The characteristics of the participants<br />
were as follows:<br />
Current director of a central academic library: eight individuals, considering that in Iran most of the<br />
academic (university) libraries are considered “central” academic libraries (Haghighi 2006);<br />
Ex-director of a central academic library: six individuals;<br />
Current assistant of a central academic library: four individuals;<br />
Faculty member of Librarianship and Information Science familiar with the management of academic<br />
libraries or with a background as a manager/consultant/executive responsible for projects for change<br />
in different organizations (such as libraries): one individual;<br />
540
Maryam Nakhoda and Sirous Alidousti<br />
Faculty member of management familiar with the management of academic libraries or with a<br />
background as a manager/consultant/executive responsible for projects for change in different<br />
organizations (such as libraries): three individuals.<br />
Among the current directors of central academic libraries in the panel, seven of them have a PhD degree<br />
in Librarianship and Information Science. One of them has a PhD degree in Mechanical Engineering. The<br />
average duration of their experience as the director of an academic library is 6.5 years. Of the members<br />
with a background as the director of a central academic library, four of them have PhD degrees in<br />
Librarianship and Information Science and two of them have a PhD degree in Management in the Wood<br />
and Paper Sciences industry. The average duration of their experience as the director of a central<br />
academic library is 5.6 years. Of the assistants of central academic libraries who were members of the<br />
panel, three have M.A. degrees in Librarianship and Information Science and one has a M.A. degree in<br />
Management. The average duration of their experience as an assistant in a central academic library is<br />
7.5 years. The faculty member of Librarianship and Information Science, who was member of the panel,<br />
has twenty-five years of experience, and the faculty members of Management in the panel have 19.3<br />
years of experience in this position.<br />
4.2 Preparation of Delphi questionnaire<br />
The design of the Delphi questionnaire was based upon a model that was earlier selected to include two<br />
stages. Of the twenty-nine change models found in the literature for the first stage, three suitable models<br />
for libraries were selected. In the second stage, using consensus method and analysis techniques related<br />
to Multiple Attribute Decision Making (MADM), Penfold’s model (1999), called “information services<br />
framework for managing change” was chosen as the basic model for the Delphi process (Nakhoda,<br />
Fadaie, and Alidousti 2010).<br />
To prepare the Delphi questionnaire, the essential activities for change management in academic<br />
libraries were determined by reviewing appropriate change models to libraries. So, Kotter’s change model<br />
(Kotter 1998), Penfold’s change model (Penfold 1999), and Curzon’s change model known as the “cycle<br />
of change model” (Curzon 2005) were studied to extract the necessary activities to bring about change.<br />
Besides these three models, Lewin’s change model called the “action research model” was also analysed<br />
for the extraction of activities. In comparison with current approaches towards change, this model is<br />
fundamental and among the oldest of the change models (Cummings and Worley 2001). Each of the four<br />
change models mentioned is composed of a series of stages. In turn, each stage of these models<br />
consists of several actions. The resulting collection and summarization of 54 actions from these models<br />
were included in the first round of the Delphi questionnaire as suggested actions for change.<br />
Hence, the main two elements of the first round of the Delphi questionnaire were the five stages of the<br />
basic model (Penfold’s model) and proposed actions for implementing change in academic libraries. The<br />
development and confirmation of the model, based upon conditions in Iranian academic libraries, were<br />
accomplished by members of Delphi panel. The consensus level in this study had a numerical value of<br />
60% of all respondents (Williams and Webb 1994). In other words, in the case that at least 60% of<br />
respondents in second and third rounds were to select a specific stage for a suggested action, this stage<br />
would be considered as one of consensus.<br />
5. Findings and conclusion<br />
In the first round questionnaire, a matrix was provided, composed of five stages of the basic model (in the<br />
first row) and 54 suggested actions for implementing change in academic libraries (in the rightmost<br />
column). To determine its necessity to the change model, each action was first specified by two options,<br />
i.e., yes or no. In the case of a positive response, the respondent was required to select the stage of the<br />
basic model that they considered the most suitable for that action. At the end of questionnaire, the<br />
respondent could write down those suggested actions for implementing change in academic libraries,<br />
which he/she thought, should be added to the list. Among the twenty-two questionnaires sent in the first<br />
round, three of them were not completed, so, nineteen questionnaires (equal to 86.3%) were submitted.<br />
Some Delphi panel members selected more than one stage in the allocation of actions to the Penfold'<br />
model. Furthermore, some members considered some actions not suitable for any of the stages. Table 1<br />
provides the actions added by panel members in the first round.<br />
541
Maryam Nakhoda and Sirous Alidousti<br />
Table 1: Actions added by panel members in the first round<br />
No. Description of actions<br />
1 Preparing an evaluation report of changes<br />
2 Standardizing the services<br />
3 Standardizing the necessary space in the library<br />
4 Reinforcing the expert human resource in the library<br />
5 Convincing top managers of the necessity of change<br />
6 Convincing top managers of the library regarding change and its probable costs<br />
7 Convincing top managers of the library regarding advantages of change<br />
8 Institutionalizing the need for change in the library<br />
9 Forming the culture required for each change<br />
10 Separating plans related to software and hardware changes in the library<br />
Handing out forms to users concerning whether or not to implement a change, accompanied by expressing<br />
11<br />
the positive and negative aspects of changes<br />
Forming a planning group whose members are selected from authorities of the parent organization, library<br />
12<br />
managers, users and staff for comprehensive decision-making<br />
13 Approving the main changes in the university to assure that they will be followed by future managers<br />
In the second round, the responses of each member in first round as well as the frequency of responses<br />
of all Delphi panel members were provided to panel members and they were again asked to express their<br />
opinions. It should be noted that, in this round, the number of suggested actions for implementation of<br />
change was again 54, and an opportunity was again provided to panel members for further judgment,<br />
considering the opinions of the group. The respondents were then asked to review the actions added by<br />
panel members to the list of 54 actions in the first round, and to allocate those actions that they find<br />
necessary to the most suitable stage(s) of Penfold’s model. All respondents in this round had also<br />
participated in the previous round. After collecting the data related to the second round of the Delphi<br />
method and calculating the frequencies of responses, a consensus criterion was applied to the data. The<br />
consensus level was considered as at least 60% of respondents (10 or more members).Therefore,<br />
thirteen actions remained below consensus level.<br />
In third round questionnaire, the responses of each expert in second round and the respective stages<br />
with the highest frequencies were provided. In this stage, frequencies of less than six were not<br />
mentioned. The respondents were asked to first identify the necessity of each of the thirteen actions, and<br />
then allocate each of them to the most appropriate stage(s).Out of the seventeen questionnaires that<br />
were sent in this stage, sixteen of them were filled out and returned. All respondents in this round had<br />
also participated in the previous round. After collecting the data related to the third round of the Delphi<br />
method, the consensus criterion was once again applied to data. The consensus level was set to at least<br />
60% of respondents in this round (nine or more members). Among the thirteen actions that were judged<br />
again, five of them were added to the list of actions by the panel members from the first round. Only the<br />
action, “terminating the contract with an organizational development consultant”, was not identified as<br />
being necessary in a change model for Iranian academic libraries, according to the votes of the majority<br />
of respondents (ten members). In total, 66 actions were allocated to the different stages of the basic<br />
model of the study (Penfold’s change model).<br />
During revision of the Delphi results, it was observed that some actions were covered by some others<br />
and had similar content. Hence, combining these actions into a single action would simplify the model. It<br />
was revealed that eight actions out of the 66 actions had such characteristics. These actions were added<br />
by the panel members in the first round and passed the consensus criterion in the second round.<br />
Ultimately, 58 actions were yielded and each of them belonged to a stage of the basic model. Table 2<br />
provides the final Delphi model, i.e., the stages and actions in the change management model for Iranian<br />
academic libraries. The number of actions does not denote the order of implementing the actions.<br />
Table 2: Change model for Iranian academic libraries<br />
Actions in stage 1: Reviewing the current state<br />
1. Recognising the problem<br />
2. Understanding the nature of change (positive/negative or permanent/temporary)<br />
3. Carrying out SWOT (analysis of strengths, weaknesses, opportunities and threatens) for the library<br />
4. Carrying out PEST (political, environmental, social and technological analysis) for the library<br />
5. Analysing the library’s current plans and strategies<br />
6. Identifying the key factors for success (activities, scope and key conditions) to reach the library’s goals<br />
542
Maryam Nakhoda and Sirous Alidousti<br />
Actions in stage 2: Analysing and identifying the users<br />
1. Evaluating users’ needs and making practical suggestions<br />
2. Identifying library users<br />
3. Handing out forms to users concerning whether or not to implement a change, accompanied by an opportunity<br />
to express positive and negative aspects of changes<br />
Actions in stage 3: Planning<br />
1. Consulting with a behavioural sciences expert regarding the management of the staff’s reactions towards<br />
change<br />
2. Contracting with an organizational development consultant from outside of the library<br />
3. Studying how to obtain and keep the existing information in the library and its costs<br />
4. Studying the vision and mission of the library<br />
5. Studying the general and specific objectives<br />
6. Determining the library services, users of services and requirements for providing services<br />
7. Acquiring information concerning the problem and initial recognition of possible solutions<br />
8. Comparing the library services and processes with similar libraries and the libraries in other sections<br />
9. Providing the acquired information to users or key groups which will be influenced by the change<br />
10. Discussing the pressures and problems with the staff in order to prepare the library for change<br />
11. Holding meetings and discussion with the library staff concerning change<br />
12. Making the decision about whether or not planning is needed for a specific change (evaluating the<br />
advantages, disadvantages and probable impacts of change)<br />
13. Making an agreement on the problem-solving method(s) among the group responsible for change<br />
14. Forming a powerful executive group for planning the change<br />
15. Describing the vision of the change, including its goal and direction<br />
16. Making strategies for reaching the change vision<br />
17. Discussing the change vision with the library staff<br />
18. Searching and finding organizational development resources and support of other sections for implementing<br />
the change<br />
19. Assigning some budget for unpredictable situations in the change process<br />
20. Searching for the source of the library staff’s resistance to change<br />
21. Making the plan for the implementation of change<br />
22. Participating in the strategic planning of the parent organization and developing it<br />
23. Analysing costs and proving the value of the library to the parent organization<br />
24. Separating the plans related to software and hardware changes in library<br />
25. Standardizing the services provided in an academic library<br />
26. Standardizing the necessary space in the academic library<br />
Actions in stage 4: Implementing the change<br />
1. Training in new behaviours which are in accordance with change practically, by the group guiding the change<br />
2. Alleviating barriers to change (personal, structural or those related to systems or skills) in the library<br />
3. Piloting the change in a limited part of library<br />
4. Converting the state of the library from its current state into the desired state<br />
5. Encouraging team work and communication among the library staff<br />
6. Keeping open the communication channels between the director of the library and staff during change<br />
7. Applying the required modifications in the process of execution of the change<br />
8. Analysing of the order of the actions carried out for change<br />
9. Reinforcing the expert human resources in the library<br />
Actions in stage 5: Maintaining change achievements<br />
1. Highlighting advances in the library’s performance (as a result of change)and informing the staff about them<br />
2. Recognizing and appreciating the staff who had a role in the improvements<br />
3. Supervising and evaluating the activities related to change periodically in order to measure their level of<br />
progress<br />
4. Determining the time for evaluating the changes made<br />
5. Determining persons for evaluating the changes made<br />
6. Determining the method(s) for evaluating the changes made<br />
7. Acquiring new information after performing actions, in order to evaluate and determine the impacts of the<br />
actions carried out<br />
8. Disseminating the results of change evaluation among library staff<br />
9. Growing the staff who can assist the realization of the change vision<br />
10. Reinforcing new behaviours caused by change<br />
11. Explaining to the staff the relationship between the new behaviours caused by change and library’s success<br />
12. Making sure that the next top manager will follow the performed change<br />
13. Measuring the satisfaction of the library users<br />
14. Disseminating information about the library’s new services<br />
543
6. Discussion<br />
Maryam Nakhoda and Sirous Alidousti<br />
Approved actions in the model have been confirmed in previous studies too. Recognising the problem<br />
has been considered as an action to recognize issues creating the need for change (Harigopal 2001;<br />
Cummings and Worley 2001). Other actions in stage one will clarify different aspects of change (Penfold<br />
1999). The second stage contains actions to know library users, their needs, and points of view much<br />
better. Adcroft, Willis, and Hurst (2008) pointed to the importance of this knowledge. Considering stage<br />
three, it has been suggested to use the help of consultants to manage psychological and organizational<br />
dimensions in the change process (Lippitt 1972; Harigopal 2001; and Cummings and Worley 2001).<br />
According to Burnes (1992), change could be carried out after the diagnosis of the problem by gathering<br />
relative information. There are actions in planning stage to discuss the change with staff (Leavitt 1965;<br />
Porras and Silvers 1991; and Burke and Litwin 1992). Forming the planning group for change and<br />
assigning budget have been confirmed respectively by Kotter (1998) and Vollmann (1996), Austin and<br />
Bartunek (2006), and Nilakant and Ramnarayan (2006). Software and hardware changes in the library<br />
could be separated in planning stage due to their specifications. In moving from the current stage to the<br />
desired one in stage four (Burnes 1992), developing the library staff, providing them with necessary skills,<br />
and alleviating barriers to change are significant actions (Kotter 1998; Curzon 2005; and Sidorko 2008).<br />
Encouraging team work and communications were confirmed by Kotter (1998), Penfold (1999), and<br />
Curzon (2005) as necessary actions for change. Cummings and Worley (2001) pointed to the evaluation<br />
of change process during the implementation. In stage five, highlighting the successful results of change<br />
and appreciating the helpful staff would be effective in maintaining the change (Kotter 1998). Evaluating<br />
the activities related to change, disseminating the results and information about new services, and<br />
measuring the users’ satisfaction could help the library director in making decision to continue or give up<br />
the change (Penfold 1999). Also in stage five new behaviours caused by change would be reinforced and<br />
the staff would know the relation between their role in change and the library success. The commitment<br />
of next top manager to follow change could help the effective changes to be more permanent (Kotter<br />
1998).<br />
7. Limitations and suggestions<br />
The first limitation of this research is originated form the nature of Delphi method, as the results of Delphi<br />
study could be altered by using a different panel of experts. In addition; the final model of this study<br />
needs to be validated in action to gain the empirical validity too.<br />
In future studies, the design of instruments for the model’s actions can be studied. Regarding the actions<br />
in the proposed change model, another study can be devoted to the order of these actions in the model.<br />
The time needed for performing each action is another issue for future research<br />
The directors of Iranian academic libraries can apply the suggested model to improve the process of<br />
change management in these libraries. The knowledge and skills for managing change could be included<br />
in the lessons of Library and Information science. This would help the students to apply necessary<br />
techniques in their working environments.<br />
References<br />
Adcroft, A., Willis, R., and Hurst, J. (2008) “A new model for managing change: the holistic view”, The Journal of<br />
Business Strategy, Vol. 29, No.1, pp. 40 -57.<br />
Ambrosiadou B. V. and Goulis D. G. (1999) “The DELPHI method as a consensus and knowledge acquisition tool for<br />
the evaluation of the DIABETES system for insulin administration” Medical Informatics, Vol. 24, No. 4, pp. 257-<br />
268.<br />
Austin, J. R, and Bartunek, J. M. (2006) “Theories and Practices of Organizational Development”, Organization<br />
Development: A Jossey-Bass Reader. Editor Gallos, J. V. San Francisco: JOSSEY-BASS, John Wiley & Sons,<br />
Inc.<br />
Bradley, L. and Stewart, K. (2002) “A Delphi study of the drivers and inhibitors of Internet banking”, The International<br />
Journal of Bank Marketing, Vol. 20, No. 6, pp. 250-260.<br />
Burke, W. W. (2002) Organization Change: Theory and Practice, Sage publications, London.<br />
Burke, W.W and Litwin, G. H. (1992) “A Causal Model of Organizational Performance and Change”, Journal of<br />
Management, Vol. 18, No.3, pp. 523-545.<br />
Burnes, B. (1992) Managing Change: a strategic approach to organizational development and renewal, Pitman<br />
Publishing, London.<br />
By, R. T. (2005) “Organizational Change Management: A Critical Review”, Journal of Change Management, Vol. 5,<br />
No. 4, pp. 369-380.<br />
Cicmil, S. (1999) “An insight into management of organizational change projects”, Journal of Workplace Learning,<br />
Vol. 11, No. 1, pp. 5-15.<br />
544
Maryam Nakhoda and Sirous Alidousti<br />
Clarke, L. and Morris, A. (1998) “Library system migration: a case study of change management at Oxford<br />
University”, Libri: International Journal of Libraries & Information Services, Vol.3, No. 48, pp. 153-162.<br />
Cummings, T. G. and Worley, C. G. (2001) Essentials of Organization Development and Change, South-Western<br />
College Publishing, Ohio.<br />
Curzon, S. C. (2005) Managing Change: A How-To-Do-It Manual for Librarians, Revised Edition, Neal-Schuman<br />
Publishers, New York.<br />
Ferguson, A. W. (2007) “Managing change in the Hong Kong library environment”, Library Management, Vol. 28, No.<br />
8 / 9, pp. 524-530.<br />
Gilley, A., Godek, M., and Gilley, J. W. (2009) “Change, Resistance, and the Organizational Immune System”,S.A.M.<br />
Advanced Management Journal, Vol. 74, No. 4, pp. 4-10.<br />
Goodrich, D. F. (2008) “The Relationship of the Theory of Constraints Implementation to Change Management<br />
Integration in Professional Service Organizations” Doctor of Business Administration Diss, Nova Southeastern<br />
University.<br />
Green, J. W. (2000) “Delphi Method in Web Site Selection: Using the Experts”, The Reference Librarian, No. 69/70,<br />
pp. 299-310.<br />
Haghighi, Mahmud(2006) “<strong>Academic</strong> Libraries”, Encyclopaedia of Library and Information Science, By F.<br />
Khosravi.Senior editor: E. Afshar, National Library of Iran, Tehran, pp. 1462-1481.<br />
Harigopal, K. (2001) “Management of Organizational Change: Leveraging Transformation”, Response Books, New<br />
Delhi.<br />
Hasson, F., Keeney, S., and McKenna, H. (2000) “Research Guidelines for the Delphi Survey Technique”, Journal of<br />
Advanced Nursing, Vol. 32, No. 4, pp. 1008-1015.<br />
Hirshon, A. (1999) “Libraries, Consortia, and Change Management”, The Journal of <strong>Academic</strong> Librarianship, Vol. 25,<br />
No. 2, pp. 124-126.<br />
Iran’s Statistical Yearbook of 2007 (2008) [Online].http://salnameh.amar-kr.ir:7001<br />
Konings, k. and Dekker, R. (2005) “Strategy and Change management in Delft University of Technology Library”,<br />
New Review of Information Networking, Vol. 11, No. 1, pp. 109- 121.<br />
Kotter, J. P. (1998) “Leading change: Why transformation efforts fail”. Harvard Business Review on Change, Harvard<br />
Business School Publishing, Boston, pp. 1- 20.<br />
Leavitt, H. J. (1965) “Applying organizational change in industry: Structural, technological, and humanistic<br />
approaches”, Handbook of Organizations, edited by J. M. March, Rand McNally, Chicago, pp. 1144-1170<br />
Linstone, H. A. and Turoff M. (1975) The Delphi Method: Techniques and Applications. Reading, MA: Columbia<br />
University Press.<br />
Lippitt, G. (1972) “Criteria for Selecting, Evaluating and Developing Consultants”, Training and Development Journal,<br />
Vol. 28, pp. 10-15.<br />
Ludwig, L. and Starr, S. (2005) “Library as place: results of a Delphi study”, Journal of Medical Library Association,<br />
Vol. 93, No. 3, pp. 315-326.<br />
Mark, G. (1997) “Reengineering library services for the digital age”, APLIS, Vol. 10, No. 4, pp 22-32.<br />
Minkman, M. et al (2009) “A quality management model for integrated care: results of a Delphi and Concept Mapping<br />
study”, International Journal for Quality in Health Care, Vol. 21, No. 1, pp 66–75.<br />
Minkman, M. MN, Ahaus, K. TB., and Huijsman R. (2009) “A four phase development model for integrated care<br />
services in the Netherlands”, BMC Health Services Research, Vol. 9, No. 42, pp 1-<br />
11.[Online]http://www.biomedcentral.com/1472-6963/9/42<br />
Nakhoda, M., Fadaei, G. R., and Alidousti, S. (2010) “Application of Two Decision Making Methods in Analyzing<br />
Experts’ Ranking of Change Models”,[Online] http://jist.irandoc.ac.ir/browse.php?a_code=A-10-1-<br />
132&sid=1&slc_lang=fa<br />
Nilakant, V. and Ramnarayan, S. (2006) Change Management: Altering Mindsets in a Global Context, Response<br />
Books, New Delhi.<br />
Okoli, C., and Pawlowski, S. D. (2004) “The Delphi method as a research tool: An example, design considerations<br />
and applications”, Information& Management, Vol.42, No. 1, pp. 15-30.<br />
Paton, B., Beranek, L., and Smith, I. (2008) “The transit lounge: a view of organizational change from a point in the<br />
journey”, Library Management, Vol. 29, No. 1 / 2, pp. 87-103.<br />
Penfold, S. (1999) Change Management for Information Services, Bowker-Saur, London.<br />
Porras, J. I. and Silvers, R. C. (1991) “Organization Development and Transformation”, Annual Review of<br />
Psychology, Vol. 42, pp. 51-78.<br />
Pugh, L. (2007) Change Management in Information Services, Ashgate Publishing, Burlington.<br />
Sharma, R. V. et al (2008) “Beyond the digital divide: a conceptual framework for analysing knowledge societies”,<br />
Journal of Knowledge Management, Vol. 12, No. 5, pp. 151-164.<br />
Sidorko, P. E. (2008) “Transforming library and higher education support services: can change models help?”Library<br />
Management, Vol. 29, No. 4 / 5, pp. 307-318.<br />
Smith, I. (2006) “Continuing professional development and workplace learning_15: Achieving successful<br />
organizational change_ do’s and don’ts of change management”, Library Management, Vol. 27, No. 4 / 5, pp.<br />
300- 306.<br />
Vollmann, T. E. (1996) the Transformation Imperative: Achieving Market Dominance through Radical Change,<br />
Harvard Business School Press, Boston.<br />
Williams P. L. and C. Webb (1994) “The Delphi technique: a methodological discussion”, Journal of Advanced<br />
Nursing, No.19, pp. 180-186.<br />
545
Adding Action to the Information Audit<br />
Huan Vo-Tran<br />
RMIT University, Melbourne, Australia<br />
huan.vo-tran@rmit.edu.au<br />
Abstract: The Information Audit (IA) has long been seen as an important tool within the Information Management<br />
field, with its origins stemming from financial audits. It is used extensively in libraries as an improvement tool and,<br />
although many have tried to define it, such as Guy St. Clair (1997), Orna (1999) and Henczel (2001a), there is still no<br />
general consensus on a definition, or the steps taken to achieve it. Whatever form it may take, it is agreed that to<br />
undertake such a task requires a structured approach. The following study will propose a hybrid approach in which<br />
Henczel’s seven-stage Information Audit model will be coupled with the Action Research (AR) methodology in order<br />
to assist a mid-sized architectural practice to manage their information throughout the architectural design process,<br />
and, in particular, as they attempt to design a new academic building for a prominent Australian university.<br />
Keywords: information audit, information management, architectural design process, action research<br />
1. Introduction<br />
As organisations begin to create and accumulate information, they tend to lose sight of how to mange it,<br />
and what information they actually possess. We are currently seeing this information being created at an<br />
astonishing rate, and manifesting itself in many forms. Organisations such as architectural firms are not<br />
only dealing with traditional paper-based documents, but are now also dealing with an abundance of<br />
unstructured data such as emails, pictures from building sites, technical drawings, formal documents<br />
from governing organisations, handwritten notes taken during interviews with clients and even transcripts<br />
collected from focus group activities.<br />
In order for an organisation to achieve high performance it is stated that, “effective information<br />
management is the key” (Roglaski 2006), but the sheer amount of unstructured data being created by<br />
these organisations present it with many challenges. Rogalski continues by stating that, “finding the right<br />
information is difficult, information is not well leveraged among partners and it is not coming together in<br />
ways that will yield useful new insights” (Roglaski 2006).<br />
However, before an organisation can effectively and efficiently manage the information that they create,<br />
possess and disseminate, they must first complete an audit of their current information practices. An<br />
Information Audit (IA) can be seen as a tool that can be used to assist them with this process. Botha and<br />
Boon (2003) suggest that an Information Audit can be defined as:<br />
The systematic examination of the information resources, information use, information flows<br />
and the management of these in an organisation. It involves the identification of users’<br />
information needs and how effectively (or not) these are being met.<br />
The study will attempt to audit the information created by a mid-sized architectural practice as they cycle<br />
through the architectural design process in order to design a new academic building for a prominent<br />
Australian university. This will be done through the use of a hybrid methodological approach in which Sue<br />
Henczel’s seven-stage Information Audit model will be combined with Kemmis and McTaggart’s Action<br />
Research (AR) model.<br />
2. Literature review<br />
2.1 Information Audit (IA)<br />
Over the last three decades there have been many attempts to define what an IA is, and what it should<br />
encompass. Yet to date, there is still no universal consensus. Early IA definitions (Reynolds 1980; Burk &<br />
Horton 1988) tended to focus on more formal information sources with a strong emphasis on document<br />
management, while recent approaches (Buchanan & Gibb 1998; Henczel 2001; Orna 1999) have moved<br />
away from this narrow approach to recognise and incorporate the importance of organisational<br />
approaches and the broad range of information resources.<br />
In order to demonstrate the differences in approaches and definitions of Information Audits a brief<br />
examination of some of the definitions is provided :<br />
546
Huan Vo-Tran<br />
LaRosa (1991)<br />
… A systematic method of exploring and analysing where a library’s various publics are<br />
going strategically, and determining the challenges and obstacles facing those publics. The<br />
audit, which raises questions about where and when these publics find and use information,<br />
gives the library a better understanding of the present and future needs of its constituents<br />
which in turn helps the library determine its most appropriate strategic direction.<br />
St. Clair (1997)<br />
A process that examines how well the organisation’s information needs and deliverables<br />
connects with the organisational missions, needs and goals and objectives.<br />
Buchanan and Gibb (1998)<br />
Discovering, monitoring and evaluating an organisation’s information resources in order to<br />
implement, maintain, or improve the organisation’s management of information.<br />
Orna (1999)<br />
A systematic evaluation of information use, resources and flows, with verification by<br />
reference to both people and existing documents, in order to establish the extent to which<br />
they are contributing to an organisation’s objectives.<br />
Henczel (2001b)<br />
Is a process that will effectively determine the current information environment by identifying<br />
what information is required to meet the needs of the organisation. It establishes what<br />
information is currently supplied, and allows matching of the two to identify gaps,<br />
inconsistencies and duplications. The process will also facilitate the mapping of information<br />
flows throughout the organisation and between the organisation and its external environment<br />
to enable the identification of bottlenecks and inefficiencies.<br />
As we can see from the definitions listed above, a majority focus on the organisations and the<br />
management of their information. It is only LaRosa who makes reference to the end users of this<br />
information, and only a few make reference to the use of Information Technology (IT) in their primary<br />
definition of IA, which can now be seen as an essential part of every organisation and the way they<br />
manage their information.<br />
2.2 Approaches to information audits<br />
Buchanan and Gibb (2007) describe the IA as being, “central to the effective organisational management<br />
of information, however there is evidence from the field that IA is neither fully accepted nor commonly<br />
practiced” (Buchanan & Gibb 2007). However, when an IA is executed by an information practitioner it<br />
can be seen as a very costly exercise for an organisation due to the time and resources that must be<br />
allocated for such an undertaking.<br />
Currently there is no standard or agreed methodological approach within the field, and it is generally left<br />
to the practitioner to sort through a myriad of academic and proprietary methods – some more<br />
comprehensive than others. Once an appropriate methodology has been selected, the practitioner is<br />
required to identify “the numerous tools and technique(s) required to support the methodological process"<br />
(Buchanan & Gibb 2007).<br />
Many may also argue that a standard for IA is not required, as each organisation needs to be treated as<br />
a separate entity and requires a different approach. Buchanan and Gibb (2007) also suggest that there is<br />
a, “lack of an agreed methodological approach” which in turn makes the selection of the methodology<br />
somewhat challenging. To add to the complexity of methodology selection, there has also been limited<br />
empirical evidence regarding the usability of the existing approaches.<br />
In support of Buchanan and Gibb's (2007) argument, Botha and Boon (2003) concluded that, “more<br />
research is required on the topic of information and more of the methodologies need to be tested in<br />
practice” (Botha & Boon 2003), which in turn will allow both practitioners and academics to develop more<br />
reliable IA methodologies that can be confidently used and re-used.<br />
547
2.3 Information audit methodologies<br />
Huan Vo-Tran<br />
As stated in section 2.2 there has been a myriad of methodologies created and adopted. The table below<br />
provides an outline of some of the major methodologies, in chronological order.<br />
Table 1: Outline of some major information audit methodologies in chronological order<br />
Year Author Brief description<br />
1987 Worlock Worlock discusses a framework of headings for the auditing process after testing it<br />
out in various environments and suggests that the judgement of these headings<br />
rests with the person undertaking the audit. In total, there are five headings, each of<br />
which should not be seen as being mutually exclusive. The five headings are:<br />
Utility analysis.<br />
Quality values.<br />
Productivity factors.<br />
Implementation criteria.<br />
Strategic impact statements.<br />
1988 Burk and Horton Burk and Horton were the first to develop InfoMap; it was seen as the first IA<br />
methodology developed for widespread use in the industry. Its focus was to evaluate<br />
the information resources using a four-stage process.<br />
Survey staff using questionnaires or surveys.<br />
Measure the information resources against cost/value.<br />
Analyse resources.<br />
Synthesise the findings and map the strengths and weaknesses of the information<br />
1993 Booth and<br />
Haines<br />
resources against the objectives of the organisation.<br />
Booth and Haines made use of the IA for organisational change and for the<br />
development of a new information policy for a regional health authority in the UK.<br />
Their strategy involved five components, which were:<br />
Identify and review the corporate objective.<br />
Decide what information is needed to meet the corporate requirements.<br />
Conduct an IA through the use of questionnaires and interviews to determine if the<br />
current required information exists within the organisation and if so, how it is<br />
currently being utilised.<br />
Address the identified information gaps and problems where possible.<br />
Develop a comprehensive information management policy for the organisation.<br />
Year Author Brief description<br />
1993 Ellis, Barker,<br />
Potter and<br />
Pridegeon<br />
1994 Webb<br />
(cited in Botha<br />
and Boon 2003)<br />
Ellis, Barker, Potter and Pridegeon acknowledge that there are many different<br />
approaches to information auditing; however, they suggest that, to fulfil its function,<br />
an IA must encompass the following:<br />
Establish what the major goals of the organisation/operation are and what kind of<br />
organizational constraints act upon the operational information systems.<br />
Determine the needs of the users.<br />
Inventory the resources available.<br />
Build up a coherent picture of how the system functions from the information<br />
gathered in the first three stages.<br />
Webb describes the IA according to three distinct stages:<br />
Initial audit (inventory).<br />
Collecting the data.<br />
Data analysis.<br />
It can be seen as an operational advisory audit as it looks at how the IA can be used<br />
to audit the current system and how effectively and efficiently the resources are<br />
being used.<br />
1997 St. Clair St. Clair states that the information audit can be grouped into five main areas which<br />
are:<br />
Getting the ball rolling<br />
Conducting interviews<br />
Organising and conducting interviews<br />
Follow up<br />
1997<br />
1998<br />
2007<br />
2008<br />
Buchanan and<br />
Gibb<br />
Typical pitfalls and mistakes.<br />
Buchanan and Gibb studied a number of IA case studies and developed what they<br />
described as a “universal model” for conducting IAs. Their approach was to create<br />
an IA model that could be used in a number of different environments and for the<br />
purposes of developing an effective information strategy for organisations. The<br />
“universal model” proposed by Buchanan and Gibb consists of five phases, these<br />
being:<br />
Promote<br />
548
Huan Vo-Tran<br />
Identify<br />
Analyse<br />
Account<br />
Synthesise.<br />
2001 Henczel Henczel's work in 2001 leveraged off the strengths of Orna and Buchanan and Gibb<br />
to produce a seven-stage auditing process. Henczel also suggests that the use of<br />
the IA should focus more on the strategic direction of the organisation and that it is<br />
the first step in the development of a knowledge audit or knowledge management<br />
strategy. Henczel's seven stages are:<br />
Planning<br />
Data collection<br />
Data analysis<br />
Data evaluation<br />
Communicating recommendations<br />
Implementing recommendations<br />
The information audit as a continuum.<br />
Year Author Brief description<br />
1990<br />
1999<br />
2004<br />
Orna Orna makes a metaphorical reference to the financial audit in her description of the<br />
IA as an, “authoritative examination of accounts with verification by reference to<br />
witnesses and documents” (Orna, 1990). Orna discusses the scope of the IA in<br />
terms of seven phases which are:<br />
Plan.<br />
Investigate the information available in the organisation.<br />
Identify the resources that are available for making information accessible.<br />
Determine information used to further the purposes of the organisation.<br />
Identify those that are responsible for managing and processing the information,<br />
respectively.<br />
Identify and evaluate the information technology that is used to manage information<br />
resources.<br />
Calculate the cost and determine the value of organisational information resources.<br />
Since then, Orna (1999) has developed an alternative auditing process comprising<br />
ten steps:<br />
Conduct a preliminary review to confirm operational/strategic direction<br />
Gain support/resources from management<br />
Gain commitment from the other stakeholders (staff)<br />
Plan, including the project, team, tools and techniques<br />
Identify the IR, information flow and produce a cost/value assessment<br />
Interpret findings based upon current versus desired state<br />
Produce a report to present findings<br />
Implement recommendations<br />
Monitor effects of change<br />
Repeat the IA.<br />
From the analysis of Table 1, we can clearly see that there are many different approaches to conducting<br />
an IA, ranging from Webb’s (1994) three-step process all the way up to Orna's (1999) ten-step process. It<br />
is also of interest to note that not all IAs focus on the same aspects and contain the same level of<br />
structure and detail.<br />
3. Methodology<br />
As stated in the introduction section, the study will attempt to audit the information created, processed<br />
and disseminated by a mid-sized architectural practice as they cycle through the architectural design<br />
process in order to design a new academic building for a prominent Australian university. Although there<br />
are many different methodologies that could have been selected as outlined in section 2.3, the<br />
researcher has opted for a hybrid approach in which two methodologies will be combined. This particular<br />
study will attempt to combine Henczel’s (2001) seven-stage IA model with Kemmis and McTaggart’s AR<br />
spiral.<br />
3.1 Justification<br />
After careful consideration of the methodologies on offer, the researcher has decided to select Henczel’s<br />
seven-stage IA model over the other competing methodologies as it:<br />
Leverages off the strengths from previous work by Orna and Buchanan and Gibb to produce a<br />
seven-stage auditing process.<br />
549
Huan Vo-Tran<br />
Focuses on a strategic direction for the organisation.<br />
Is similar to the AR cycle – Henczel’s model advocates the IA as a continuum where there is an<br />
establishment of a cyclical process.<br />
Allows for the creation of a knowledge audit or knowledge management strategy.<br />
Is a structured, descriptive and methodical process as opposed to that of Worlock (1987).<br />
Has incorporated IT into its overall process.<br />
Has considered management and operational aspects with the submission of a business case in the<br />
planning stage before proceeding.<br />
In addition to Henczel’s seven-stage IA model, the researcher has also incorporated Kemmis and<br />
McTaggart’s AR spiral. This will be utilised in place of stage six – implementing recommendations. The<br />
reason for this selection is that the researcher believes that, although Henczel’s stage six fits well into the<br />
overall IA process, the use of the Kemmis and McTaggart’s AR spiral will serve as a better diagnostic<br />
and implementation tool as it can be seen as a more strategic and structured approach for implementing<br />
change. The use of an AR methodology should also, “assist in solving an organisational problem, or in<br />
some instances takes a step forward in deepening an organisation’s understanding of themselves”<br />
(Emerald, 2008).<br />
3.2 Participants<br />
The study will make use of a mid-sized architectural and urban design practice as a case study. It will<br />
attempt to audit the information created by a mid-sized architectural practice as they cycle through the<br />
architectural design process in order to design a new academic building for a prominent Australian<br />
university. The practice was established in 1996 and is based in the Central Business District (CBD) of<br />
Melbourne, Australia. It is currently being led by a team of five directors and also encompasses ninety<br />
professionals, who include: architects, technical staff, designers and construction staff.<br />
The practice has been involved with, and completed numerous large-scale commissions for both private<br />
and government sectors, with some of their most recent projects including major commercial buildings,<br />
university education and training facilities, and government and defence projects.<br />
3.3 Henczel’s seven-stage IA model<br />
Similar to Orna’s (1990) IA model, Henczel’s IA model also consists of seven main stages. As<br />
demonstrated in Figure 1 each stage is clearlydefined and each stage needs to be complete before the<br />
next one can commence. Listed below is a short description of what is involved in each stage.<br />
Stage one: Planning<br />
Involves planning and the submission of a business case for approval by the organisation before<br />
proceeding on to the data collection.<br />
Stage two: Data Collection<br />
Involves the collection and development of an Information Resource (IR) database and its population<br />
through survey techniques such as questionnaires, personal interviews or focus groups.<br />
Stage three: Data Analysis<br />
Once the data has been collected, it must be organised in a way that allows it to be analysed. It is up to<br />
the person leading the IA to select the most appropriate method for the analysis of the data. Henczel<br />
suggests that, as specialist skills are required to accomplish this, it might be worthwhile to contract<br />
experts outside your organisation.<br />
Stage four: Data Evaluation<br />
At the data evaluation stage, the data begins to show the person conducting the audit a “snapshot” of the<br />
organisation’s information environment. This will facilitate the interpretation and formulation of the<br />
recommendations<br />
550
Huan Vo-Tran<br />
Figure 1: Henczel’s seven-stage information audit model<br />
Stage five: Communicating Recommendations<br />
Henczel suggests that there are many ways of doing this, depending on the size of the organisation and<br />
the scope of the recommendations. However, the key people who need to be kept informed through this<br />
process may include:<br />
Anyone who has championed the audit<br />
Any sponsors<br />
People who directly participated<br />
Those who will be affected by the recommended changes. (Henczel 2001).<br />
The communication of these recommendations may be delivered in many forms and could include:<br />
reports, website or intranet, verbal presentations, seminars and personal feedback to participants.<br />
Stage six: Implementing Recommendations<br />
This stage can be seen as the second-last stage in Henczel’s seven-stage IA model. It involves the<br />
development of a plan for when and how the recommendations outlined in the previous stage will be<br />
implemented. Henczel suggests that the organisation, together with the nature of the recommendations,<br />
will influence how the implementation occurs.<br />
Stage seven: The Information Audit as a Continuum<br />
Henczel suggests that when you reach the end of the audit, it is really only the beginning and that<br />
organisations need to be thinking about the IA as a continuous process to ensure that data gathered<br />
initially can be re-assessed and updated.<br />
3.4 Kemmis and McTaggart’s Action Research spiral<br />
McKay and Marshall (2001) describe how Action Research (AR) in its simplest form involves both action<br />
and participation within a particular field. Its focus is to problem-solve in order to improve the way<br />
processes are performed and services are delivered.<br />
551
Huan Vo-Tran<br />
AR typically makes use of four main phases of continuous change, which are: Plan, Act, Observe and<br />
Reflect<br />
For the purpose of this study, Kemmis and McTaggart’s AR Spiral was selected and will be embedded<br />
within the implementation recommendations stage (stage six) of Henczel’s seven-stage IA model. The<br />
spiral consists of the four phases of AR and is to be carried out collaboratively with the organisation that<br />
is being studied.<br />
Figure 2 represents Kemmis and McTaggart’s AR Spiral (1988). It includes the four main phases of<br />
continuous change (plan, act, observe and reflect). Provided below is a brief description of what happens<br />
within each phase.<br />
Plan:<br />
Develop a plan of critically informed action to improve what is already happening within the organisation<br />
that is being studied.<br />
Act:<br />
Based on the plan formulated, act upon the plan and implement it.<br />
Observe:<br />
Observe the effects of the critically informed action in the context in which it occurs.<br />
Reflect:<br />
Reflect on these effects as the basis for future planning, subsequent critically informed action and so on,<br />
through a cycle of succession cycles. (Kemmis & McTaggart, 1988)<br />
Figure 2: Kemmis and McTaggart’s (1998) action research spiral<br />
3.5 Combination of the two methods<br />
As stated in section 3.1, the study will take on a hybrid approach where Henczel’s seven-stage IA model<br />
and Kemmis and McTaggart’s AR Spiral will be used in conjunction with one another. The study will cycle<br />
through each one of the seven stages of Henczel’s IA model; however, there will be a modification made<br />
552
Huan Vo-Tran<br />
to the sixth stage (implementing recommendations) of the audit instead of making use of Henczel’s sixth<br />
stage, where the outputs should include:<br />
An information policy<br />
Post-implementation strategy<br />
Revised business plan and information management process<br />
An updated information resources database.<br />
With the addition of Kemmis and McTaggart’s AR Spiral, the findings of the previous stages of Henczel’s<br />
seven-stage IA model will inform the planning phase and, in the last component of reflection, will lead into<br />
the final stage, which is the information audit as a continuum. Figure 3 demonstrates how this will be<br />
accomplished.<br />
To what extent is the current<br />
process working and how<br />
can it be streamlined?<br />
Record the process taken<br />
as new information comes<br />
in. Also record the interview<br />
responses from the<br />
architects to see how they<br />
manage and disseminate<br />
this data.<br />
Has this process been<br />
streamlined, if so how? How<br />
important is the collected<br />
data for the next phase in<br />
the architectural design<br />
process?<br />
Follow up on the<br />
observations by interviewing<br />
the architects to see if the<br />
suggested changes have<br />
been beneficial.<br />
In the feasibility stage of the<br />
architectural design<br />
process, information is<br />
collected from a wide range<br />
of sources including focus<br />
group interviews and<br />
emails. This is then placed<br />
in a folder called ‘briefing<br />
folder’. From an information<br />
management point of view,<br />
how can this process be<br />
improved? Is teaching<br />
generic information<br />
management skills to<br />
architects going to improve<br />
the process?<br />
Shift the way architects<br />
think about information<br />
management by suggesting<br />
‘best practice’ tools and<br />
techniques then allowing<br />
them to explore these.<br />
Try questioning the<br />
architects so that the can<br />
critically reflect on their<br />
information management<br />
approaches and see if they<br />
can come up with other<br />
solutions.<br />
Continue general aim but try<br />
and emphasize ‘best<br />
practices’ in information<br />
management.<br />
Record any changes in the<br />
architects’ approach to<br />
information management<br />
after suggestions have been<br />
made.<br />
Figure 3: The incorporation of Kemmis and McTaggart’s action research spiral within stage six of<br />
Henczel’s seven-stage information audit model. (Vo-Tran 2010)<br />
In terms of the techniques that will be used within the combined methodologies, the researcher has<br />
decided to select two commonly used techniques, these being observations and interviews. The use of<br />
observations for this study will allow the researcher to understand, without interfering with the current<br />
practices within the organisation during their the planning phases. Observations can be used again to<br />
see if the changes that have been made during the implementation stages have been beneficial. In<br />
addition to the observations, interviews will be conducted to explore the attitudes and responses to the<br />
changes that have been made. This will also allow the researcher to gather qualitative information in the<br />
form of narratives. It is anticipated that the interviews will take place at the organisation and take on a<br />
planned but unstructured approach.<br />
553
4. Preliminary findings<br />
Huan Vo-Tran<br />
As the study is currently in progress, this section will outline the major findings of the first two stages to<br />
date. The findings will be broken down and reported in terms of Henczel’s seven-stage IA model.<br />
Stage one: planning<br />
A formal letter was submitted to the directors of the architectural practice seeking their permission to<br />
participate in the study. Written permission was granted by the practice and the researcher then sought<br />
and received ethics clearance from their home university.<br />
Adhering to Henczel’s seven-stage IA model, the following tasks and activities were considered and<br />
implemented:<br />
Table 2: Outline of the tasks and activities that were considered and implemented<br />
Task Activity<br />
Developing clear objectives An understanding of what the IA is trying to achieve for both the researcher and<br />
architectural practice<br />
Understanding the organisation and its core business<br />
Identification of the key stakeholders<br />
Defining and scoping the<br />
resource allocation<br />
Physical and information scope<br />
Human, financial, technical and physical resource allocation<br />
Insource, outsource options<br />
Selection of a methodology Data collection, analysis and evaluation<br />
The presentation of the findings and recommendations<br />
Action plan for implementing the recommendations<br />
Development of a<br />
communication strategy<br />
(Henczel 2001a)<br />
Before the audit<br />
During the audit<br />
After the audit<br />
From the tasks and activities listed in the table above, a business case was formulated and presented to<br />
the directors of the architectural practice; this has subsequently been approved and signed off.<br />
Stage two: data collection<br />
Upon completion of the business case, data needed to be collected. This was undertaken using a<br />
combination of techniques, such as onsite interviews and observations. From this, an Information<br />
Resources (IR) database was established and greater understanding of the architectural practice was<br />
achieved. Listed below are some of the findings that resulted from the interviews and observations.<br />
Projects are won through the tendering process, in which the architectural practice presents their<br />
ideas to prospective clients.<br />
Staff at the architectural practice work around project-based teams. They can be working on multiple<br />
projects at any given time depending on workload and expertise.<br />
Each project has one of the architectural practice’s five directors in charge, and these directors may<br />
be spread across no more than three projects.<br />
All project teams are multidiscipline, and may include architects, draftsmen, consultants, engineers<br />
and designers.<br />
Projects may last anywhere between six months and four years, depending on the size and<br />
complexity.<br />
In addition to the practices and processes listed above, the data collection also identified some<br />
challenges in information management the architectural practice is facing:<br />
Storage of documents and images. Documents and images are currently being duplicated in both<br />
physical and electronic formats. These may not be identical as it might take time to update the<br />
documents to reflect either format. In addition to this, electronic copies of documents are stored<br />
within folders according to the file type, e.g. all the PDF files in one location while all the word<br />
documents are in another.<br />
554
Huan Vo-Tran<br />
Recording decisions. The architects have stated that this is one of their biggest problems; any<br />
changes that have been made to a building go straight onto the drawing and there is no record to<br />
why this decision has been made. This causes problems later on when they have no idea why these<br />
changes have been made.<br />
Re-use of information. Although no two projects will be ever the same, there are some elements or<br />
aspects that could be reused. Instead, the architects and draughtsmen spend lengthy amounts of<br />
time trying to re-create almost identical drawings when they could be spending the time working on<br />
other elements of the building.<br />
Transfer of project details to new staff members. As the practices’ staff members move from one<br />
project to another, it takes them time to read through the brief, and attempt to get up to speed as<br />
soon as possible. This is done through reading both the physical and electronic documents, which<br />
may not always contain the most recent changes. This in turn, means that they will have to spend<br />
more time with another staff member going through the finer details.<br />
5. Directions for future research<br />
As the research is currently in progress, it is anticipated that over the next six months the study will move<br />
from the data collection stage into the data analysis and evaluation stages. The data collected so far will<br />
now inform what needs to be done in terms of implementing the recommendations and structuring the<br />
action research component of the hybrid methodology. It is also anticipated that once the IA has been<br />
completed, it will add to the body of knowledge about which Botha and Boon (2003) concluded in a<br />
paper, “more research is required on the topic of information and more of the methodologies need to be<br />
tested in practice” (Botha & Boon 2003).<br />
References<br />
Booth, A. and Haines, M. (1993) "Information audit: whose line is it anyway?" Health Libraries Review, Vol 10, pp<br />
224–232.<br />
Botha, H. and Boon, J. A. (2003) "The Information Audit: Principles and Guidelines", Libri, Vol 53, pp 23–38.<br />
Buchanan, S. and F. Gibb (1998). "The Information Audit: An Integrated Strategic Approach." International Journal of<br />
Information Management, Vol 18, No. 1, pp 29 - 47.<br />
Buchanan, S. and F. Gibb (2007). "The Information Audit: Role and Scope." International Journal of Information<br />
Management, Vol 27, No. 3, pp 159 - 172.<br />
Buchanan, S. and F. Gibb (2008). "The information audit: Methodology selection." International Journal of<br />
Information Management, Vol 28, No. 1, pp 3 - 11.<br />
Burk, C. F. and F. W. Horton (1988). InfoMap: A complete guide to discovering corporate information resources.<br />
Englewood Cliffs, NJ, Prentice-Hall.<br />
Ellis, D., R. Barker, et al. (1993). "Information Audits, Communication Audits and Information Mapping: A Review and<br />
Survey." International Journal of Information Management, Vol 13, No. 2, pp 134 - 151.<br />
Emerald. (2008). "How to... carry out action research." [Online], Emerald,<br />
http://www.emeraldinsight.com/research/guides/methods/action_research.htm.<br />
Orna, E. (1990). Practical information policies: how to manage information flow in organizations. Aldershot, Gower<br />
Publishing Limited.<br />
Orna, E. (1999). Practical Information Policies. Hampshire, Aldershot.<br />
Orna, E. (2004). Information strategy in practice. Aldershot, Gower Publishing Limited.<br />
Dubois, C. P. R. (1995). "The information audit: its contribution to decision making." Library Management, Vol 16, No.<br />
7, pp 20 - 24.<br />
Henczel, S. (2001a). The Information Audit: A Practical Guide. München, K.G.Saur.<br />
Henczel, S. (2001b). "The Information Audit as a First Step Towards Effective Knowledge Management." Information<br />
Outlook, Vol 5, No. 6, pp 13.<br />
Kemmis, S. and R. McTaggart (1988). The Action research planner. Waurn Ponds, Deakin University Press.<br />
McKay, J. and P. Marshall (2001). "The dual imperatives of action research." Information Technology & People, Vol<br />
14, No. 1, pp 46 - 59.<br />
LaRosa, S. (1991). "The corporate information audit." Library Management Quarterly, Vol 14, No. 2, pp 7-9.<br />
Roglaski, S. (2006). "The Rising Importance of Enterprise Content Management." DM Review, Vol 16, No. 11, pp 36.<br />
St. Clair, G. (1997). "The information audit I: defining the process." InfoManage, Vol 4, No. 6, pp 5-6.<br />
Swash, G. D. (1997). "The information audit." Journal of Managerial Psychology, Vol 12, No. 5, pp 312-318.<br />
Vo-Tran, H. (2010). "Auditing the architectural design process", Paper read at 1 st International <strong>Conference</strong> on<br />
Information Management and Evaluation. Cape Town, South Africa.<br />
Worlock, D. (1987). Implementing the information audit. 59th Aslib Annual <strong>Conference</strong>, 'The information Manager'.<br />
University of Sussex, Electronic Publishing Services Ltd. 39: 255-260.<br />
555
Business Intelligence Best Practices for Success<br />
Joseph Woodside<br />
Cleveland State University, USA<br />
joseph.m.woodside@gmail.com<br />
Abstract: Business Intelligence (BI) is in a high adoption and high growth area, as users quickly value the<br />
capabilities and increasingly demand more features to compete in today’s economic climate. However, from a return<br />
on investment (ROI) standpoint, BI is similar to ERP and CRM, in that it has a poor risk/reward profile, as it regularly<br />
runs into cost overruns, due to scope creep and limitless requests for support from end-users (Bernard 2009). Unlike<br />
operational systems which often have specific requirements and implementation completion timelines, BI<br />
environments are constantly evolving to meet business and information requirements (Moss 2007). Given the<br />
complexity of most system implementations, no single measure exists for Business Intelligence success. In order to<br />
effectively evaluate BI success, measures are developed to identify critical implementation factors based on the<br />
research objectives and investigation (Wixom 2001). As an organization progresses in BI maturity, the value of its<br />
activities expands. Successful organizations increasingly utilize analytical approaches to identify and enact modest<br />
improvements that increase profitability and return on business intelligence investments. This paper presents several<br />
key findings, lessons learned, success evaluation methods, and best practices as identified through prior literature<br />
review and a formal empirical study, which extends and enhances prior literature and understanding of BI.<br />
Keywords: BI, business intelligence, critical success factors, implementation, success<br />
1. Road to Business Intelligence<br />
Beginning in the late 1960s, experiments began with Decision Support Systems (DSS), utilizing<br />
computers to analyze data and offer decision-making support. DSS were typically used for narrowly<br />
focused activities such as production planning, investment management, and transportation applications,<br />
and several inputs are required to prepare the analysis (Leidner 1993). With the introduction of software<br />
applications such as SAS and SPSS in the 1970s, statistical software became more available and<br />
accessible to end users. Despite this introduction, DSS did not prosper and evolved into Executive<br />
Support Systems (ESS). (Davenport 2007; Ranjan 2008).<br />
ESS were utilized by executives for viewing firm performance and focused less on decision making<br />
support. The feature found in most ESS was single database access with current organizational<br />
information, in an easy to access manner. ESS usage was also found to be positively related to problem<br />
identification, decision making, and analysis. Other ESS features included non-keyboard interface,<br />
organizational database, drill-down capabilities, trend analysis, exception reports, graphics, and critical<br />
information monitoring. The focus of an ESS was on the organizations day-to-day activities as well as<br />
marketplace indicators. A DSS by contrast was intended to allow on demand decisions and routine<br />
analysis (Leidner 1993). ESS were referred to as high-risk/high-return systems, as the systems serve<br />
executives whose information needs are complex, but also have greater influence. ESS provided<br />
executives easy to use information that supports their critical success objectives. ESS also failed to enjoy<br />
widespread usage due to resistance by executives to hands-on usage (Rainer 1995).<br />
Firms have since made major investments in such systems as enterprise resource planning (ERP),<br />
supply chain management (SCM) and customer relationship management (CRM), yet struggle to achieve<br />
competitive advantage. Firms need streamlined access and analysis of the underlying information in<br />
order to make operational decisions. Strategic organizations sought to improve efficiency through faster<br />
and better-informed decision making, and looked to technology to enhance strategic and tactical results<br />
to improve time to market, connectivity, integration, and visibility into their business. Unbeknownst to<br />
users at the time, this data from these systems was a significant organizational asset, which would later<br />
be leveraged for success and competitive advantage. In order to realize these benefits however, the data<br />
must be developed into an enterprise wide unified view. Construction and integration of knowledge is a<br />
key to succeeding in the competitive global market. Information Technology moved to support day-to-day<br />
operations and all aspects of decision making, with differentiation through technology becoming<br />
increasingly important. New generations of technology savvy users and executives were finding ways to<br />
utilize previously untapped information. The entire field emerged as Business Intelligence (BI) and<br />
includes collection, management, and reporting of decision making data and information. BI capabilities<br />
have consistently been identified as the number one technology priority for organizations according to<br />
current industry surveys (Davenport 2007; Ranjan 2008).<br />
556
2. Business Intelligence<br />
Joseph Woodside<br />
BI is not a single product, application, program, user, area, or system, rather an architecture of integrated<br />
systems that provide users with easy access to and storage of information for decision making and<br />
learning. Competitive pressures cause organizations to continually improve and adapt in order to be<br />
successful in the ever changing business environment, and information is required by employees<br />
throughout all levels of the organization for ongoing decision making (Ranjan 2008; Saha 2007).<br />
Business Intelligence refers to applications and technologies used to gather, capture, access,<br />
consolidate, and analyze information to improve decision making by various horizontal and vertical levels<br />
of stakeholders. These systems capture important metrics on business operations, as well as providing a<br />
mechanism for improved decision making. At the various levels these information items may include<br />
documents, calendars, wikis, links, reports, dashboards, scorecards, search, databases, lists, user<br />
knowledge, and much more. For example, these technologies can help coordinate projects, calendars,<br />
schedules, discuss ideas, review documents, share information, keep in touch with others, utilize Key<br />
Performance Indicators (KPI) to gauge operational status, and generate reporting information ondemand.<br />
The BI process is one that allows large amounts of disparate data to come together into a single<br />
repository and turn that data into meaningful information for decision support processes. BI can include<br />
various forms of analysis, data mining, scorecards, dashboards, metrics, reporting, portals, data<br />
warehouse, OLAP, decision support, knowledge management, etc. This information is available to all<br />
levels of the organization and associated stakeholders, on-demand, and in an easy-to-use fashion (Moss<br />
2007; Ranjan 2008; Saha 2007).<br />
3. Research model<br />
During interviews with executives, managers, and professional staff, before, during, and after BI<br />
implementation and from review of prior literature, a path-analytic BI implementation success model was<br />
proposed as shown in Figure 1. Literature review from information systems was utilized to develop the<br />
research model and relationships between those areas. Prior literature has studied information success<br />
using multiple methods. Researchers are instructed to utilize appropriate success measures based on<br />
the objectives of the study and investigation (Wixom 2001).<br />
Figure 1: BI implementation success model<br />
557
3.1 Collaborative culture<br />
Joseph Woodside<br />
Team work is an important aspect throughout any implementation project. This requires close<br />
cooperation between all department areas and business and technical teams, including top management,<br />
consultants, end-users, and vendors (Bhatti 2005). Aligning resources with business strategies is also<br />
important, as typically there is limited alignment between BI strategy and business strategy (Williams<br />
2007). Organizational learning through shared visions and commitment to learning is an important aspect<br />
to collaborative culture and can be achieved through team work and alignment between units in support<br />
of a common objective and set of goals.<br />
H1: There is a positive relationship between collaborative culture and implementation success<br />
3.2 Customization<br />
Business Process Re-engineering (BPR) involves rethinking and redesigning business processes to<br />
improve key performance measures such as cost and quality of service. Most organizations are required<br />
to modify their existing business processes to fit the application software as a way to limit customizations<br />
(Bhatti 2005). Arnott describes this similarly as the degree of fit between the organization and the<br />
software and hardware (Arnott 2008). Many times an implementation may not meet expectations, due to<br />
an underestimation of change management complexities and encountering resistance to change (Bhatti<br />
2005). In many cases, end users must be trained in the new paradigm after and during BI<br />
implementation. Users must also be trained in understanding and adjusting to changes in business<br />
processes (Williams 2007). BI platforms should be selected which allow for customizations to occur<br />
dynamically through the user interface, this resolves prior system issues which occurred at time of<br />
upgrade of customized applications or modules. The flexibility of most BI platforms allow for<br />
customizations which previously were unavailable in the scope of an enterprise application.<br />
H2: There is a positive relationship between customization and implementation success<br />
3.3 Communication<br />
Wide information sharing and understanding must occur by all stakeholders throughout the<br />
implementation stages and beyond. Communication should start as early as possible to gain<br />
organizational understanding and acceptance (Bhatti 2005). Beyond development of BI and user training,<br />
the vision must be marketed and communicated. The BI applications must be viewed as mission critical<br />
and all users share that vision (Williams 2007). Communication should be started early in the form of<br />
several announcements and organizational newsletters, including email, meeting, and Intranet<br />
announcements. Kick-off meetings with key personnel and staff resources should also take place, along<br />
with regularly recurring meetings.<br />
H3: There is a positive relationship between communication and implementation success<br />
3.4 Project management<br />
Organizations should use structured and formal approach for BI projects. Many projects fail to adequately<br />
account for organizational requirements, resources, and funding necessary to support a successful BI<br />
implementation (Williams 2007). PM includes coordinating, scheduling, scope, and monitoring activities<br />
and resources in line with the project objectives. PM is also responsible for the overall implementation<br />
process and developing organizational support. The DW/BI systems should be developed iteratively<br />
building to a complete application set. (Arnott 2008; Bhatti 2005). Agile methodology was adopted, having<br />
a formal project methodology and a formal project management office for oversight and project tracking is<br />
critical to the project’s success. It is important to establish critical success indicators and metrics from<br />
project inception, to ensure expectations are met and exceeded. To allow improved deployment speed,<br />
an iterative methodology with rapid prototyping should be employed. Parallel user sub-groups should be<br />
established to allow continuous feedback from rapid prototyping and to reduce periods of inactivity.<br />
H4: There is a positive relationship between project management and implementation success<br />
3.5 Resources<br />
Resources can include financial, people, hardware, software and time for project completion. It is also<br />
important to fund new activities required as a result of BI implementation such as meta data<br />
558
Joseph Woodside<br />
management. Resource issues often have a negative impact on implementation success (Arnott 2008;<br />
Williams 2007; Wixom 2001). Consultants are often required due to a knowledge gap and complexity of<br />
new systems. User involvement can occur through requirements gathering, implementation participation,<br />
and use after go-live (Bhatti 2005). Dedicated resources should be assigned to avoid inevitable<br />
competing projects and priorities. Dedicated consultant resources should be allocated to improve timeline<br />
and knowledge transfer for new technology areas. Though consultants should only be utilized as a<br />
temporary solution, as knowledge-loss occurs with continued usage. Dedicated department or area<br />
based resources should be assigned for local subject matter experts and knowledge diffusion.<br />
H5: There is a positive relationship between resources and implementation success<br />
3.6 Top management support<br />
Top management provides the required resources in a direct or indirect manner through financing, as<br />
well as the power and support. Top management is also responsible for setting a clear direction, overall<br />
project objectives, project guidance, representation, and establishing these throughout the organization<br />
(Arnott 2008; Bhatti 2005). Sponsorship across the entire management team, allows others in the<br />
organization to support the project through reducing political resistance, and facilitate participation. Top<br />
management support must include top management champions and are viewed similarly (Wixom 2001).<br />
All top management should be advised of the project by the sponsors, and any issues or concerns<br />
addressed initially. It is vital for the sponsors to continually update top management early in the project,<br />
and as components are released to end-users.<br />
H6: There is a positive relationship between top management support and implementation success<br />
3.7 Training<br />
Training end-users is important to improve knowledge and appropriate use of the system. General BI<br />
concepts, components, demonstration, and use are key training areas. Training should also include<br />
process changes and overall flow of information and integration (Bhatti 2005). Training also includes the<br />
standards and policies that must be followed for the new BI applications, to optimize use of BI by endusers<br />
(Williams 2007). Training modules and materials should be developed prior to the initial go-live,<br />
along governance plans, such as best practices, content and technical standards, and policies and<br />
procedures should be developed for training purposes. In addition, proactive monitoring of training issues<br />
should be immediately addressed to avoid long-term paradigm creation in early stages. New users<br />
should be required to take established training and existing users on an annual basis and directed to the<br />
training materials as questions arise.<br />
H7: There is a positive relationship between training and implementation success<br />
3.8 Vertical integration<br />
In prior BI implementations multiple vendor solutions required purchase, as one vendor did not provide a<br />
fully integrated solution, also other companies chose a best of breed approach based on vendor<br />
offerings. Today, several vendors offer completely integrated solutions with equitable offerings across<br />
services. In one survey organizations utilized an average of 3.2 vendors with 8-13 tools (Howson 2008).<br />
For the small-medium business scenario, utilizing a vertical architecture with a single vendor was<br />
identified as a critical success factor. Due to resource and funding requirements, the ability to match<br />
expert skillsets with a best of breed approach is not possible. The use of a single vendor also improves<br />
delivery time through ease of installation and avoids integration issues that commonly arise when utilizing<br />
multiple vendor solutions vendors. Here, a vertical architecture is defined as having a single vendor<br />
designed and developed BI platform, which includes Knowledge Management, Content Management,<br />
Performance Management, End-User Tools, Querying / Reporting, Analysis, and Database Management<br />
System.<br />
H8: There is a positive relationship between vertical architecture and implementation success<br />
3.9 Success factors<br />
BI project implementation success is measured through perceived success, whether the project<br />
completed timely, completed on budget, and overall satisfaction with the BI (Howson 2008; Wixom 2001).<br />
Table 1 displays the summary of supporting works for construct and hypothesis development.<br />
559
Table 1: Summary of supporting works<br />
Wixom and<br />
Watson (2005)<br />
Joseph Woodside<br />
Bhatti<br />
(2005)<br />
Williams and<br />
Williams (2007)<br />
Howson<br />
(2008)<br />
Implementation Factors<br />
Collaborative Culture X X<br />
Customization X X X<br />
Communication X X<br />
Project Management X X X X<br />
Resources X X X X<br />
Top Management Support X X X<br />
Training X X<br />
Vertical Integration X<br />
Success Factors<br />
Perceived Success X X<br />
Timely Implementation X X<br />
Satisfaction X X<br />
4. Research methodology<br />
Arnott<br />
(2008)<br />
A survey was development followed Moore and Benbasat's (1991) identified stages of item creation,<br />
scale development, and testing: Item creation, in which existing items were utilized from prior literature,<br />
then additional items added to those components which fit the definitions. Scale development, where<br />
similar categories of items were created and refined as needed. Testing, in which sample surveys were<br />
conducted, and then was followed by revisions and larger distribution (Moore 1991). The final survey was<br />
randomized to reduce order effects. The survey uses a seven point likert scale, and measures the level of<br />
agreement with each statement, with 1 being strong disagreement and 7 strong agreement. The survey<br />
was administered to a national healthcare organization who had recently completed a BI implementation.<br />
The survey was reviewed by subject matter experts in and across organizational levels as part of a pretest<br />
to resolve any concerns or address any identified discrepancies, as well as re-word identified items,<br />
or remove non-weighting or duplicative manifest variables. A total of 141 responses were received with<br />
105 (75%) non-managerial users, and 36 (25%) managerial users. Multiple geographic locations were<br />
included, along with multiple staff functions and department areas. Overall 148 responses were received,<br />
with 7 responses removed that were incomplete past general user information, for a total of 141 usable<br />
responses. 106 respondents were non-managerial users, with the remaining 35 respondents holding a<br />
supervisory position through executive position. Users from multiple geographic office locations and<br />
functional areas were represented within the study group, as well as users with short and long-term<br />
tenure at the firm.<br />
5. Data analysis<br />
A path analysis is employed using SmartPLS 2.0 software to analyze the results and determine model fit<br />
(Ringle 2005). A model with significant loadings is developed. Significance of relationships was<br />
determined between implementation factors and success factors. The final model specification includes<br />
the supported paths, all paths were supported for significance at p = 0.01. Results showed that 73.8% of<br />
the variability in implementation success is explained from the model.<br />
Table 2 displays the quality of the model. The average variance extracted is indicated by the column<br />
AVE, and is the average communality for the latent factors in the model. AVE is utilized for convergent<br />
validity, and should be greater than or equal to 0.5, which the latent factors exceed. Composite reliability<br />
is also utilized as Cronbach’s alpha commonly underestimates or overestimates reliability. Composite<br />
reliability follows similarly to Cronbach’s alpha, with 0.80 to be considered good, 0.70 to be considered<br />
acceptable, and 0.60 to be considered for exploratory requirements. Composite reliability exceed 0.90 for<br />
this model. R-Square displays the effect size measure, and is not shown for exogenous constructs. An R-<br />
Square of 0.67 is considered substantial, 0.33 considered moderate, and 0.19 considered weak. The R-<br />
Square for this model is 0.73. Cronbach’s alpha should be equal to at least 0.80 to be considered good,<br />
0.70 to be considered acceptable, and 0.60 to be considered for exploratory requirements. For short<br />
scales, Cronbach’s alpha may be biased, this study utilized 7 point scales for all questions measured,<br />
and Cronbach's alpha exceed 0.80 (Garson 2010; Ringle 2005). Table 3 displays the hypothesis results.<br />
560
Table 2: Measurement model quality<br />
Joseph Woodside<br />
AVE Composite Reliability R Square Cronbachs Alpha<br />
Implementation Factors 0.57986 0.90581 0.878317<br />
Success Factors 0.66762 0.922954 0.738381 0.899059<br />
Table 3: Hypothesis results<br />
Hypothesis Result<br />
H1: There is a positive relationship between collaborative culture and implementation success Supported<br />
H2: There is a positive relationship between customization and implementation success Supported<br />
H3: There is a positive relationship between communication and implementation success Supported<br />
H4: There is a positive relationship between project management and implementation success Supported<br />
H5: There is a positive relationship between resources and implementation success Supported<br />
H6: There is a positive relationship between top management support and implementation<br />
success Supported<br />
H7: There is a positive relationship between training and implementation success Supported<br />
H8: There is a positive relationship between vertical architecture and implementation success Supported<br />
6. Conclusion and future directions<br />
This paper presents several key findings as identified through a formal study, and improves the power of<br />
the explanatory success model, which extends and enhances prior literature and understanding of BI.<br />
The first key finding is the identified implementation construct addition of a vertical architecture,<br />
particularly for the small-medium business (SMB) scenario. A vertically-integrated architecture improves<br />
the implementation timeline and required resource base for implementation success. The second key<br />
finding is around establishing a collaborative culture to promote organizational learning capabilities. This<br />
extends previous notions of team work and business-IT alignment, and is necessary to support adoption<br />
and use of BI. Third involves implementation success outcomes when democratization or universal user<br />
adoption of BI has been achieved. In past studies measuring BI success, only a small portion of users<br />
had access to BI capabilities, successful outcomes can be realized while extending BI benefits to all<br />
users.<br />
Limitations include the use of single organization study; additional organizations should be reviewed to<br />
increase sample size in and across various industries and firm sizes. There is an importance to<br />
identifying universally applying critical success factors, and the ability for tailoring those factors to an<br />
individual organization or implementation. However it is the adaptability of the BI capabilities and the<br />
overall project that will ensure successful completion. Other identified areas of study and importance<br />
beyond implementation include establishing a competency center to ensure continue usage of business<br />
intelligence, stakeholder satisfaction, and decreased costs. Another area is establishing an architecture<br />
roadmap for future iterations and system updates, these include enhancements to key capabilities and<br />
features, bug fixes, security improvements, and ensured vendor support. A governance plan is also<br />
important to establish technical roles, support service level agreements, back and recovery, database<br />
and data standards, metadata standards, content branding, life cycle policies, and training.<br />
References<br />
Arnott, D. "Success Factors for Data Warehouse and Business Intelligence Systems," Australasian <strong>Conference</strong> on<br />
Information Systems) 2008, pp 55-65.<br />
Bernard, A. "Four Technology Best (and not so best) Bets for 2009," in: CIO Update, 2009.<br />
Bhatti, T.R. "Critical Success Factors for the Implementation of Enterprise Resource Planning (ERP): Empirical<br />
Validation," The Second International <strong>Conference</strong> on Innovation in Information Technology) 2005.<br />
Davenport, T.H., Harris, Jeanne G. Competing on Analytics Harvard Business School Publishing Corporation,<br />
Boston, 2007.<br />
Garson, G.D. "Partial Least Squares Regression (PLS)," N.C.S. University (ed.), 2010.<br />
Howson, C. Successful Business Intelligence: Secrets to Making BI a Killer App, 2008.<br />
Leidner, D., Elam, Joyce "Executive Information Systems: Their Impact on Executive Decision Making," Journal of<br />
Management Information Systems (101:3) 1993, pp 139-155.<br />
Moore, G.C., Benbasat, Izak "Development of an Instrument to Measure the Perceptions of Adopting an Information<br />
Technology Innovation," Information Systems Research (2:3) 1991, pp 192-222.<br />
Moss, L.T., Atre, Shaku Business Intelligence Roadmap, Boston, 2007.<br />
Rainer, R.K., Watson, Hugh "The Keys to Executive Information System Success," Journal of Management<br />
Information Systems (12:2) 1995, pp 83-98.<br />
561
Joseph Woodside<br />
Ranjan, J. "Business Justification with Business Intelligence," The Journal of Information and Knowledge<br />
Management Systems (38:4) 2008, pp 461-475.<br />
Ringle, C.M., Wende, S., Will, S. "SmartPLS 2.0 (M3) Beta," Hamburg, 2005.<br />
Saha, G.K. "Business Intelligence Computing Issues," ACM Ubiquity (8:25) 2007.<br />
Williams, S., Williams, Nancy The Profit Impact of Business Intelligence, 2007.<br />
Wixom, B.H., Watson, Hugh J. "An Empirical Investigation of the Factors Affecting Data Warehouseing Success,"<br />
MIS Quarterly (25:1) 2001, pp 17-41.<br />
562
Work<br />
in<br />
Progress<br />
Papers<br />
563
564
The Evolution of IT Governance in Multiple Industry Units<br />
and the Business Case for an Outside Partner<br />
Ganeshprasad Chandrasekaran 1, 2 , Chandramohan Annavarapu 2 and Balaji<br />
Rathinasamy 2<br />
1<br />
Tata Consultancy Service Limited, Chennai, India<br />
2<br />
SRM School of Management, SRM University, Chennai, India<br />
cgprasad@gmail.com<br />
profacmohan@gmail.com<br />
balaji99@gmail.com<br />
Abstract: IT Governance has become the integral part of the strategic thinking process for the growing number of<br />
companies across different industries. Companies are increasingly evaluating what is core to their business and<br />
weighing the benefits of outsourcing critical non-core functions such as IT (Information Technology). IT Governance<br />
exists within the context of corporate governance and the principles are especially the same. IT governance is an<br />
accountability framework and management process that helps to define and communicate what must be done and<br />
provides the rigorous oversight to ensure that it is. The accountability framework is typically planned with well defined<br />
roles and responsibilities reflecting decision rights among the stakeholders in the IT Management process and is<br />
reinforced with effective reporting. This paper reports on an investigation which focused on the role of IT governance<br />
and the outsourcing partner as mechanism for the effective and efficient management of IT. Results are presented<br />
from a survey in which respondents consisted of IT executives with responsibilities for IT governance and<br />
Outsourcing from different parts of the globe, who represent multiple industries. The outcome of the survey indicates<br />
a strong correlation between the overall use of IT governance practices and overall governance performance.<br />
Keywords: IT governance, outsourcing, corporate governance, business<br />
1. Introduction<br />
The advent of the 21 st century has brought about new challenges for the corporate world as it continues<br />
to evolve in adapting to ever-changing operating environments. However, business expectations for IT<br />
are rising. The vast majority of business processes are enabled by computers and organizations have no<br />
fallback paper processes. IT-enabled products and services have become a competitive differentiator in<br />
almost every industry. To help meet the strategic needs of the business, many companies are<br />
implementing IT governance practices to enhance IT and business alignment. IT governance is<br />
recognized as an extension of corporate governance. The IT Governance Institute (ITGI, 2008) indicates<br />
that corporate governance is the “Methodology by which a corporation is directed, administered and<br />
controlled, whereas IT governance supports achieving corporate objectives, strategy, direction,<br />
administration and control, using appropriate IT investment and resource management.” This paper<br />
reports on investigated results about “IT governance practices being used to balance the competing IT<br />
priorities like control, agility, cost, customer, business needs, revenue growth and profitability”.<br />
2. Literature review<br />
IT governance is the responsibility of executives and the board of directors. It consists of the leadership,<br />
organizational structures and processes that ensure that the enterprise’s IT sustains and extends the<br />
organization’s strategies and objectives (ITGI, 2008). Leveraging IT successfully to transform the<br />
enterprise and create value added products and services has become a universal business competency.<br />
IT is fundamental for managing enterprise resources, dealing with suppliers and customers, and enabling<br />
increasingly global transactions (Eric and William, 2008).<br />
According to the ITGI, “Boards and executive management generally expect their enterprise’s IT to<br />
deliver business value, i.e., provide fast, secured, high-quality solutions and services; generate<br />
reasonable return on investment; and move from efficiency and productivity gains toward value creation<br />
and business effectiveness (Alan, 2009). The governance of outsourcing is the function described as “it is<br />
no longer a company’s ownership of capabilities that matters but rather its ability to control and make the<br />
most of critical capabilities, whether or not they reside on the company’s balance sheet” (Lucin and<br />
Assaf, 2009). Outsourcing in simple terms is “the delegation of few services or IT-intensive business<br />
processes to an external provider for increasing the availability of service and the overall profitability”<br />
(Spremic, et al., 2008, p.252).<br />
565
3. Research findings<br />
Ganeshprasad Chandrasekaran et al.<br />
In addition to the literature review, the research study was supported by a survey and an interview. The<br />
survey was conducted with twenty one IT professionals and interview with three IT professionals. Both<br />
the survey and the interview were with IT professionals who fall under the category of senior executives<br />
and work for different CMM (Capability Maturity Model) organizations. During the interview the same<br />
survey questions were used, however had in detail discussion around every question. The literature<br />
review was instrumental in designing the survey questions and the survey has targeted use of the IT<br />
governance practices and its performance. The below points were used to frame the survey and they are<br />
(a).Are IT Governance practices able to balance the internally focused cost and control objectives?<br />
(b).Are IT Governance practices able to balance the externally focused customer and business<br />
needs?<br />
(c).Is IT Governance helps to meet the internal and external business objectives?<br />
(d).Can IT Governance practices ensures to prioritize the business demands?<br />
(e).Is there any real growth in revenue and profitability later to the implementation of IT governance?<br />
(f).What is the proportional focus of IT governance and an Overall Business Objective?<br />
(g).In General, are IT Governance initiatives working?<br />
The overall use of IT governance practices was measured by the average combined score and the<br />
results of the survey respondents indicates a strong correlation between the overall use of IT governance<br />
practices and overall governance performance. For the open-ended questions raised on IT governance,<br />
answers indicate that most common IT governance objectives are focused on cost containment (includes<br />
efficiency, standardization and automation) and risk reduction (includes compliance, security, and public<br />
scrutiny of IT failure). This is perhaps not surprising, given the recent economic climate and global surge<br />
in regulatory requirements imposed on IT organizations.<br />
Survey Response<br />
35<br />
30<br />
25<br />
20<br />
15<br />
10<br />
5<br />
0<br />
IT Goverance - by Maturity<br />
Cost Complaince Cusotomer Decision<br />
Category<br />
Figure 1: IT Governance objectives by maturity group<br />
High<br />
Medium<br />
Maturity is determined by a combination of use of governance practices and performance. The study<br />
revealed (as shown in Figure 1) the following about the different maturity groups.<br />
(i). Low maturity organizations (those with lower usage of IT governance practices and lower<br />
performance scores) have IT governance objectives that are more focused on cost and risk reduction.<br />
The high-impact IT governance practices that are shown to optimize performance at the initial level of<br />
maturity, however move beyond cost and risk to also focus on agility as well as project and service value<br />
that may not be cost oriented.<br />
(ii). Medium maturity organizations (those with moderate IT governance practice usage and medium<br />
performance scores) also focus on cost objectives, but they have a greater focus on customer and<br />
business related objectives. The high-impact IT governance practices at this maturity level build on the<br />
foundational practices to include supporting a process culture, asset management, tactical demand<br />
management and improving the transparency of project assumptions.<br />
566<br />
Low
Ganeshprasad Chandrasekaran et al.<br />
(iii). High maturity organizations (those with broadest use of IT governance practices, and highest<br />
performance scores) have the highest focus on customer and business related objectives and the lowest<br />
level of focus on cost containment. Mature organizations have learned to use IT auditing as a tool to<br />
achieve their operating objectives as well as addressing compliance mandates.<br />
4. Conclusion<br />
Although best practices are openly available and clearly described in literature, they are not necessarily<br />
widely adopted. This implies that in many organizations, there is significant room for improvement in the<br />
area of IT governance specifically in the low and medium matured organizations. As the IT organization<br />
makes the transition from utility service provider to a more strategic revenue enabler, IT governance can<br />
be a powerful tool for balancing competing IT priorities. However, this study suggests that in many<br />
organizations, IT governance initiatives are more focused on cost containment and risk reduction<br />
objectives. Focus on customer differentiation and enabling business needs has taken a lower priority than<br />
more foundational cost and risk reduction objectives. The results, however also indicate a positive<br />
element, IT governance initiatives are working and the further research will continue to understand the<br />
methods adopted by organizations to utilize the IT governance principles to the fullest extend.<br />
References<br />
Alan, C. (2009) IT Governance: Implementing Frameworks and Standards for the corporate Governance of IT, ITGI<br />
Publications, Illinois.<br />
Andreas, W.M. (2008) “Modes of Governance in Business Process Outsourcing: Equity Control or Tradability of<br />
Services, an Executive versus Market's Perspective”, Proceedings of the XLIst Annual Hawaii International<br />
<strong>Conference</strong> on System Sciences, Hawaii, USA.<br />
Brown, A.E. and Grant, G.G. (2005) “Framing the Frameworks: A review of IT Governance Research“,<br />
Communications of the Association for Information Systems, Vol. 15, pp 696-712.<br />
David, N. (2005) IT Governance: Managing Information Technology for Business, Thorogood Publications, London.<br />
Eric, J.B and William, A.Y (2009) The effective CIO: How to achieve outstanding success through strategic<br />
Alignment, Financial Management and IT Governance, Auerbach Publications, Florida.<br />
IT Governance Institute. (2008) Aligning COBIT®4.1, ITIL® V3 and ISO/IEC 27002 for Business Benefit. [online],<br />
Available at: < http://www.isaca.org/Knowledge-Center/Research/ResearchDeliverables/Pages/Aligning-COBIT-<br />
4-1-ITIL-V3-and-ISO-IEC-27002-for-BusinessBenefit.aspx > [Accessed on 10 November 2010].<br />
Lawrence, J.A, Susan, P., Gary, F.P and Dasaratha V.R (2007) “Corporate Governance, Audit Quality and the<br />
Sarbanes-Oxley Act: Evidence from Internal Audit Outsourcing“, The Journal of Accounting Review, Vol. 82, pp<br />
22-31.<br />
Lucian, A.B. and Assaf, H. (2009)”The Elusive Quest for Global Governance Standards“, University of Pennsylvania<br />
Law Review, Vol. 157, pp 1263-1317.<br />
Patel, N.V. (2002) “Global Ebusiness IT Governance: Radical Re-Directions”, Proceedings of the XXXVth Hawaii<br />
International conference on System Sciences, IEEE Computer Society, Hawaii, USA.<br />
Short, J. and Gerrard, M. (2009) “IT Governance Must be Driven by Corporate Governance” Gartner Research,<br />
Stamford, USA, November.<br />
Spremic, M., Zmirak, Z. and Kraljevic, K. (2008) “Evolving IT Governance Model – Research Study on Croatian<br />
Large Companies”, WSEAS Transactions on Business and Economics, Issue 5, Vol. 5, May, pp 250-259.<br />
Weill, P. and Ross, J.W. (2004) “IT Governance on One Page”, Center for Information Systems Research,<br />
Massachusetts Institute of Technology, Cambridge, USA.<br />
567
The Relationship Between Quality Management and<br />
Knowledge Management in the Service Industries<br />
Amir Honarpour and Ahmad Jusoh<br />
Management and Human Resource Development, University Technology<br />
Malaysia, Johor, Malaysia<br />
Amir_honarpour@yahoo.com<br />
ahmadj@utm.my<br />
Abstract: Regarding the knowledge based economy master plan (2002), Malaysia is transforming itself towards a<br />
knowledge-based economy, which concentrates on knowledge, people and virtual networks. Researchers argue that<br />
achieving such an objective and gaining value from knowledge causes some problems including changes in<br />
customer demands, developing new skills, knowledge creation, and modifying the economical structure to a<br />
knowledge-based economy etc. (Plump and Zamfir, 2009; Suete, 2001;Nonaka and Teece, 2001). These challenges<br />
affect the way organizations collect, use and disseminate information. Knowledge management can lead to a<br />
sustainable competitive advantage. On the one hand, it is supposed that knowledge is an intangible and unique<br />
asset, which provides an organization with competitive advantage, while on the other, QM contributes significantly to<br />
the performance of organizations, mostly in the manufacturing sector. Considering the organizational life and<br />
structure, both QM and KM are professional communities that are positioned at different points in their maturity<br />
lifecycle, however, they have similar aims and positions in regard to management. It appears that both quality<br />
management and knowledge management are complementary, if not compatible, as the strategies of both are long<br />
term for the intention of gaining competitive advantage and overall improved performance. QM aims to help the<br />
value-creators of organizations in improving their performance as well as KM. They hold somesimilar basic<br />
assumptions, e.g. the importance of cultural changes and process improvement. There are valuable studies on the<br />
description of the interaction between QM and KM (see Jaime et al., 2006; Dvir, 2004). Regarding this point of view,<br />
this research is going to consider implementing KM in an organization with regard to QM practice and show how they<br />
interact and how they improve their effectiveness.<br />
Keywords: knowledge, knowledge management, quality management<br />
1. Introduction<br />
In the early 1990s, Malaysiaset out a 30-year plan as vision 2020 to become a developed country in<br />
terms of economic performance and technological capability. Transferring from a manufacturing industrial<br />
economy to a knowledge-based economy is a part of this plan (Ramlee and Abu, 2005; Fleming and<br />
Soborg, 2010). A knowledge-based economy (k-economy) focuses on knowledge, people and virtual<br />
networks, and requires the adoption and application of new management and organizational systems<br />
along with knowledgeable, skilled, dynamic, creative and innovative human resources (Plumb and Zamfir,<br />
2009; Ramlee and Abu, 2005; Fleming and Soborg, 2010). Moving towards a knowledge-based economy<br />
needs intensive knowledge investment such as research & development (R&D), software, information<br />
and communication technology (ICT), education and training (Suete, 1999). To succeed in this way,<br />
knowledge management has an essential role. As Chong et al. (2010) suggested, knowledge<br />
management (KM) has a strong effect on innovation adoption. In addition, to support scientific activities,<br />
especially in R&D, uses of quality management complemented by knowledge management will improve<br />
the quality of research (Jaime, 2006; Holt and Dilani 2009). Finally, the object of education and training<br />
should be to make quality human resources and knowledge workers the backbone of the k-economy.<br />
Such a labour force is able to create and affect knowledge to produce useful actors with strong and<br />
analytical skills (Ramlee and Abu, 2005). According to Abadesco (2004), a knowledge worker is ‘anyone<br />
who makes a living out of creating, manipulating or disseminating knowledge’. Therefore, KM is the<br />
infrastructure for knowledge workers' jobs, and also has an important role in supporting activities required<br />
for achieving a K-economy. However, the point is that recent studies have revealed that a large amount<br />
of KM projects (more than 50%) fail or do not achieve their goals (Prusak and Weiss, 2007;Madanmohan,<br />
2005). This may be caused by many reasons such as the lack of top management commitment or failing<br />
to locate, capture, organize and disseminate the tacit knowledge (see Akhavanet al., 2005).<br />
Researchers have linked Quality Management (QM) with organizational learning in an attempt to show<br />
how Quality Management leads to continuous improvement (Choo et al., 2007; Konidari and Abernot,<br />
2006). Linderman et al. (2004) stated that knowledge is created by quality management practices as<br />
shown in the Nonaka's Knowledge Creation Process. Molina et al. (2007) advocated that QM practices<br />
have an important role in the transferring of knowledge. In addition, the knowledge-based economy is<br />
replacing the meaning of service quality management in a way that it becomes knowledge-driven and<br />
568
Amir Honarpour and Ahmad Jusoh<br />
relies on people’s continuous growth and sharing ideas and knowledge (Plumb and Zamfir, 2009). It<br />
seems that the notion of knowledge and quality are strongly associated, as Jaime(2006) said "as<br />
Knowledge Management reaches its maturity in terms of acceptance as an important part of doing<br />
business in the modern world, quality will again become the mantra of successful companies".<br />
Based on the literature, quality management is a good basis and facilitator for implementing knowledge<br />
management, however, empirical research on this subject is insufficient. In Malaysia 2,383 organizations<br />
have received ISO 9001:2000 and other firms are mostly ready to execute IMS in their organization<br />
(Kadir et al., 2009). As QM practices and systems are widely used in the country, this research can be<br />
helpful concerning Vision 2020.<br />
2. Knowledge management and quality management<br />
QM has contributed significantly to the performance of organizations, mostly in the manufacturing sector.<br />
It played a major role in the rise of the Japanese industry in the 60's and 70's (James, 1997), and the<br />
recovery of some sectors of the American industry in the 80's. In the 80's and early 90's there was<br />
growing management attention on QM, which was accompanied by a considerable growth in the<br />
community. There are different QM models and techniques such as ISO 9000, Deming Prize, the<br />
American Malcolm Baldrige National Quality Award, and the <strong>European</strong> Quality Award. Many researchers<br />
from the mid 1990s have studiedthe field of knowledge management and have contributed to relevant<br />
theories. In this study, the systematic framework of organizational knowledge management processes<br />
introduced by Alavi and Leidner (2001) is applied. This framework is based on the sociology of<br />
knowledge and considers the cognitive and social nature of organizational knowledge. It also represents<br />
an individual's cognition and practices as well as the organizational ones (Alavi and Leidner, 2001). The<br />
knowledge management processes include knowledge creation, knowledge storage/retrieval, knowledge<br />
transfer, and knowledge application. According to Pentland (in Alavi and Leidner, 2001), knowledge<br />
creation refers to developing new knowledge or replacing existing content within the organization's tacit<br />
and explicit knowledge. Knowledge storage or retrieval refers to the knowledge that resides in different<br />
component forms, such as written documentation or structured information stored in electronic<br />
databases. Knowledge transfer refers to transferring knowledge between individuals or from individuals to<br />
organizational memory or across groups. Knowledge application is defined as the integration of<br />
knowledge for creating organizational capability (Alavi and Leidner, 2001).<br />
Regarding to previous studies the conceptual model has been developed. This model is introduced in<br />
figure 1. Based on this model some propositions has been presented to investigate the relationship<br />
between QM and KM.<br />
Figure 1: Conceptual framework of the study<br />
569
2.1 QM and knowledge creation<br />
Amir Honarpour and Ahmad Jusoh<br />
The application of information technology and systems has recently attracted traditional service providers<br />
such as airlines and travel agencies into service-value networks (Plumb and Zamfir, 2009). These<br />
changes are towards the satisfaction of the changing needs of customers. Customer satisfaction also<br />
means considering internal customers (Yusof and Eng, 2003). This relationship can be provide by<br />
listening mechanisms, QFD, customer stories etc. Linderman (2004) claimed that, "Quality management<br />
practices that promote monitoring and feedback of customer information allow knowledge to be created".<br />
Practices such as customer relationship and workforce management. Therefore, we propose the<br />
following proposition:<br />
Proposition1: Quality management practices that are related to customer satisfaction can affect the<br />
knowledge creation process.<br />
2.2 QM and knowledge storage/retrieval<br />
One of the quality management values is management by facts. This means that all the decisions should<br />
be based on documented facts (Lagrosen and Lagrosen, 2006). This concept relies on practices and<br />
tools that are used in different QM practices and techniques such as quality information, the idea bank,<br />
flowcharts, and decision making tools. This value is a base for documenting culture and is considered as<br />
a facilitator for the use of organizational memory, such as databases and repositories. Hence, the<br />
following proposition is formulated:<br />
Proposition2: Quality management practices affect the knowledge storage/retrieval process.<br />
2.3 QM and knowledge transfer<br />
One of the principals of quality management is establishing teams (Molina et al., 2006). According to<br />
Linderman et al. (2004), teams begin to develop ideas to resolve quality problems. These team meetings<br />
are a place in which knowledge is transferred. There are also other practices such as process<br />
management and product/service design that use some specific tools like SPC, cause and affect analysis<br />
and statistical charts, which lead to transfer knowledge from a separate part of the organization to other<br />
parts. Training and education are also important factors of QM. Such practices and tools transfer<br />
knowledge from individuals to individuals (e.g. team meetings in process management practice) or from<br />
individuals to group (e.g. educations in workforce management). Thus, the following proposition could be<br />
formulated:<br />
Proposition3: Quality management practices and tools affect the knowledge transfer process<br />
2.4 QM and knowledge application<br />
To apply knowledge management, Demsetz (in Alavi and Leidner, 2001) identified three steps of the<br />
knowledge integration process: first, directives or a set of rules, standards, and procedures; second,<br />
organizational routines or interaction protocols, and process specification; and third, self-contained task<br />
teams or teams of individuals with prerequisite knowledge for problem solving (Alavi and Leidner, 2001).<br />
As mentioned earlier, teams facilitate developing new ideas to resolve quality problems (see Linderman<br />
et al., 2004). System flowcharts and process flowcharts are QM tools, which identify what scope of<br />
information is required to do work or what codes and standards govern the work (Shearer, 1994). Also<br />
the main aim of process management practice is to reduce process variation to reduce reworks and<br />
make stable process. Therefore, this study proposes that:<br />
Proposition4: Quality management practices and tools affect the knowledge application process<br />
2.5 Senior management commitment<br />
Senior management commitment is a necessary condition for the implementation of quality management<br />
(Gitlow, 2001). This commitment is one of the values of quality management implementation and, as<br />
Lagrosen and Lagrosen (2006) said, "leaders can excel if only their hearts are into their quality journey<br />
and they have a determination to carry it through". Quality management is a long-term strategy for<br />
gaining competitive advantage and improving overall performance. It aims to improve quality and<br />
productivity. Therefore, senior managers who are committed to improve quality and productivity through<br />
QM are more likely to invest in knowledge management. As a result:<br />
570
Amir Honarpour and Ahmad Jusoh<br />
Proposition 5: Top management commitment to QM will affect the acceptance of implementing KM<br />
3. Conclusion<br />
In conclusion, the implementation of knowledge management can be conceptualized as quality<br />
management practices. The conceptual model proposes that the commitment of senior managers, quality<br />
management practices which promote monitoring and feedback of customer information, management by<br />
facts, teamwork, problem solving tools and process orientation work can improve the success of<br />
knowledge management implementation. The next step in this investigation is to develop a survey to<br />
examine the model and its proposed relationships.<br />
References<br />
Akhavan, P., Jafari, M. and Fathian, M. (2005) Exploring Failure-Factors Of Implementing Knowledge Management<br />
Systems In Organizations, Journal of Knowledge Management Practice<br />
Alavi, M. Leidner, D. (2001) Knowledge Management and Knowledge Management System: Conceptual Foundations<br />
and Research Issues, MIS Quarterly, Vol. 25, No. 1, pp: 107-136<br />
Abadesco, V. (2004) Training knowledge worker,Asian Productivity Organization, Tokyo<br />
Chong, A. Y., Ooib, K., Linc, B. and Tehb, P. (2010) TQM, knowledge management and collaborative commerce<br />
adoption: A literature review and research framework, Total Quality Management, Vol 21, No. 5, pp: 457–473<br />
Choo, A. S., Linderman, K. W. and Schroeder, R. G. (2007) Method and context perspectives on learning and<br />
knowledge creation in quality management, Journal of Operations Management, Vol 25, pp: 918–931<br />
Dvir, R. (2000) Quality By Knowledge: Collective Creation of A New Concept, Quality Progress, August 2000<br />
Eng, Q, E., Yusof, S, M. (2003) A survey of TQM practices in the Malaysian electrical and electronic industry, Total<br />
Quality Management, Vol 14, No 1, pp: 63-67<br />
Fleming, D. and Soborg, H. (2010) Malaysia’s Human Resource Strategies for a Knowledge-Based Economy –<br />
Comparing the Influence of Different Labour Market Relations, <strong>European</strong> Journal of Social Sciences, Vo 16,<br />
Number 2, pp:286-306<br />
Gitlow, H. S. (2001) Quality management systems: a practical guide, St. Lucie Press<br />
Holt, R. and Dilani, J. (2009) Knowledge and quality management: An R&D perspective, Technovation, Vol29, pp:<br />
775–785<br />
Jaime, A., Gardoni, M., Mosca, J. and Vink, D. (2006) From Quality Management to Knowledge Managementin<br />
Research Organisations, IJIM, Vol10, No. 2, pp:197-215<br />
James, D. (1997). TQM Seeks a New Relevancy in the Emerging 'Knowledge Economy', Business Review Weekly<br />
(Australia),September22, p 199<br />
Kadir, A. et al (2009) Implementation of Integrated Management System in Malaysia: The Level of Organization’s<br />
Understanding and Awareness, <strong>European</strong> Journal of Scientific Research, vol 31, No. 2, pp.188-195<br />
Konidari, V. Abernot, Y (2007) From TQM to learning organization: Another way for quality management in<br />
educational institutions, International Journal of Quality & Reliability Management, Vol 23, No. 1, pp: 8-26<br />
Lagrosen, S. and Lagrosen, Y. (2006) A dive into the depths of quality management, <strong>European</strong> Business Review,<br />
V18, N.2, pp: 84-96<br />
Linderman, K., Schroeder, R. G., Zaheer, S., Liedtke, C. and Choo, A. S. (2007) Integrating quality management<br />
practices with knowledge creation processes, Journal of Operations Management, Vol22, pp: 589–607<br />
Madanmohan, R. (2005) Knowledge management tools and techniques: practitioners and experts, Elsevier, Oxford<br />
Molina, L. M., Lloréns-Montes, J. and Ruiz-Moreno, A. (2007) Relationship between quality management practices<br />
and knowledge transfer, Journal of Operations Management, Vol25, pp: 682–701<br />
Nonaka, I. and Teece, D. (2001). Managing Industrial Knowledge: Creation, Transfer and Utilization, SAGE<br />
publications, London<br />
Ramlee, M. and Abu, A. (2005) Malaysia transitions toward a knowledge-basedEconomy, The Journal of Technology<br />
Studies, 51-61<br />
Plumb, I. and Zamfir, A. (2009) Managing service quality within the knowledge based economy: opportunities and<br />
challenges, Quality Management in Services, VolXI, No 26<br />
Prusak, L. and Weiss, L. (2007) Knowledge Creation and Management: New Challenges for Managers, Oxford<br />
University Press<br />
Shearer, C. (1995) Practical Continuous Improvement for Professional Services, American Society for Industrial<br />
Security<br />
571
Construction and Validation of eSchool Success Model<br />
Hesbon Nyagowa 1 , Dennis Ocholla 1 and Stephen Mutula 2<br />
1 University of Zululand, South Africa<br />
2 University of Botswana, Botswana<br />
nyagowa_hesbon@yahoo.co.uk<br />
docholla@pan.uzulu.ac.za<br />
MUTULASM@mopipi.ub.bw<br />
Abstract: DeLone and McLean (1992) IS Success Model has been widely adapted in evaluating success of many<br />
different information systems including e-commerce, e-learning and knowledge management systems. Adapting the<br />
D&M IS success model requires re-specification to suit specific systems. The re-specified models tend to introduce<br />
new dimensions and therefore new measuring instruments that call for empirical validation. In the current study<br />
evaluating NEPAD eSchool pilot programme in Kenya, a distinct new dimension was added to the D&M IS success<br />
model and the net-benefit dimension measure was re-defined to capture the objectives of NEPAD eSchool<br />
programme. The study specified seven dimensions to eSchool system: infrastructure quality, information content and<br />
communication quality, technical personnel service quality, training effectiveness, extent and exploitative use, user<br />
satisfaction and net-benefits. The eSchool success model was conceptualised theoretically and subsequently<br />
validated empirically using data collected from 776 students and teachers in six schools which have implemented<br />
NEPAD eSchool system. The data was split into two to facilitate implementation of both exploratory factor analysis<br />
(EFA) and confirmatory factor analysis (CFA). The first half of the data set was used in performing principal<br />
component analysis on one new dimension (training effectiveness) and the re-specified dimension (net-benefit). The<br />
second half of the data set was used in confirmatory factor analysis to test the goodness-of-fit of the eSchool<br />
success model in evaluating eSchool success. This paper discusses the eSchool success model construction, its<br />
validation outcome and the contribution of the study to IS evaluation. This is part of a study in progress.<br />
Keywords: eSchool, eSchool success model, validation of eSchool system, validation, IS evaluation<br />
1. Introduction<br />
1.1 eSchool as an Information System (IS)<br />
In its original patent, the eSchool teaching system has four sub-systems; students’ sub-system, teachers’<br />
sub-system, library sub-system and administration sub-systems (Nobles et al., 1989) all interfaced such<br />
that user groups are restricted to modules and content relevant for their roles. Information system<br />
consists of computers, communication networks, instructions (application modules), stored files, people<br />
and procedures. ISs are characterized in organizational work system (Alter, 2008) and serve as<br />
information management tools. They therefore enable user groups to engage in retrieval of data, sharing<br />
of data, manipulation of information, and transmission of information. ESchool system enables students<br />
and teachers to share information through communication network inbuilt in the computer systems,<br />
facilitate storage, sharing and retrieval of data by users and user groups.<br />
Specifically, eSchool systems are designed to avail content relevant to curriculum, sharing of class work<br />
materials between teachers and students, students & students, and teachers & teachers’, retrieval of<br />
curriculum support materials from other data bases, communication between users, and management of<br />
student data within the school network. Teaching, learning and education management is therefore<br />
accorded an alternative platform in eSchool system. These characteristics of eSchool systems qualify<br />
them as information system as defined in Alter (2008).<br />
1.2 Justification of need for eSchool evaluation<br />
Implementing eSchool system has three major cost components; infrastructure development, content<br />
development and upgrading of ICT expertise of students and teachers. In addition to ICT expertise<br />
upgrading, teachers require retraining on pedagogical approaches that reduce their direct involvement in<br />
the learning process (Hennessy, Ruthven & Brindley, 2005) as sage-on-the-stage to facilitators. ESchool<br />
infrastructure, content and human skills development require high capital outlay that needs<br />
rationalization. To ascertain sustainability and success of an information system, evaluation mechanism<br />
should be put in place to give indicators of its success and re-strategizing the implementation to improve<br />
the success rate.<br />
572
Hesbon Nyagowa et al.<br />
Traditionally, IS programmes have had evaluation inbuilt as part of IS development cycle (Gremberen,<br />
2001). The theories and practices behind the traditional evaluations have tended to be specific to the<br />
programme and they deploy varied non-standardized tools which deny IS community opportunity to<br />
compare evaluation outcomes (DeLone & McLean, 1992; Petter, DeLone & McLean, 2008). Studies<br />
which have used similar evaluation models such as DeLone and McLean (1992) IS success model have<br />
concentrated on one or two dimensions of the model. Absence of methodological comparability limits the<br />
trend in knowledge growth particularly in extension and generalization of knowledge (Petter, DeLone and<br />
McLean, 2008) in new environment. It is in this respect that Petter, DeLone and McLean (2008) urged IS<br />
success evaluation researchers to use validated instruments in measuring success of any IS.<br />
Proliferation of eSchool systems has been experienced in many countries in the recent past (Simpson,<br />
Payne & Condie, 2005; Condie & Livingston, 2007; Ya’Acob, Nor & Azman, 2005; UNICT Task Force,<br />
2005). The force behind this trend is the high expectations organizations have placed on the potentials of<br />
eSchool system; studies converge on the possible benefits of eSchools which include productivity<br />
improvement in the education sector, imparting lifelong learning skills to students and creating team work<br />
skills amongst students through collaboration between different learning communities (EL-Halawany &<br />
Huwail, 2008). To safeguard against moribund projects, eSchool researchers should seek dependable<br />
success evaluation model and instruments.<br />
The current study conceptualized an eSchool success evaluation model based on DeLone and McLean<br />
(1992) IS success model. The theoretically conceptualized model was empirically validated to establish<br />
consistency and stability using factor analysis. This poster presents the conceptualized model and the<br />
results of its validation.<br />
2. Methods<br />
2.1 Conceptualization of eSchool success model<br />
Following analysis of several IS success evaluation studies, DeLone and McLean (1992) developed an IS<br />
success model which has seven dimensions namely; system quality, information quality, intention to use,<br />
use, user satisfaction, individual impact and organization impact (DeLone & McLean, 1992). The model<br />
has been validated (ICIS, 2007; Wu & Wang, 2006; Bokhari, 2005; Seddon & Kiew, 1994; Roldan & Leal,<br />
2003) and augmented (Pitt, Watson & Kavan, 1995; DeLone & McLean, 2003; Adeyinka, 2009; DeLone<br />
& McLean, 2008). In its present form D&M IS success model’s seven dimensions are schematically<br />
represented in figure 1.<br />
Information<br />
Quality<br />
System Quality<br />
Service Quality<br />
Intention to Use<br />
Use<br />
User Satisfaction<br />
Net Benefit<br />
Figure 1: D&M IS success model’s seven dimensions (source: DeLone and McLean (2003))<br />
573
Hesbon Nyagowa et al.<br />
DeLone and McLean IS success model (2003) has widely been accepted in many evaluation studies<br />
(Molla & Licker, 2001; Ong, Chorng-Shyong & Lai, Jung-Yu, 2004; Adeyinka, 2009) but each study has<br />
had to re-specify the model to suit the system being studied. Underpinning on extant literature, the<br />
current study conceptualized a seven dimensional eSchool success model. ESchool users (students and<br />
teachers) are introduced to eSchool use when they have very limited exposure to ICT literacy. Teachers<br />
in particular, were previously trained on pedagogies which had not anticipated integration of ICT in<br />
teaching and learning. Invariably, retraining of teachers on pedagogies consistent with eSchool<br />
requirements and imparting to both teachers and students ICT skills give them positive attitude to<br />
eSchool and may play a big role in the success of eSchool. The current study therefore conceptualized<br />
training effectiveness as a necessary dimension to eSchool success model. As training impacts on<br />
attitude, and intention to use IS is an attitudinal construct, the current study dropped the dimension<br />
‘intention to use’ in the D&M IS success model. Further, in eSchool, availability of content relevant to<br />
curriculum will to a great extent influence its adoption and use in teaching and learning. The current study<br />
therefore conceptualized that content should be given prominence more so for the fact that sustainability<br />
of eSchool partly rely on availability of content. Lastly, the study argues that accruing measurable<br />
benefits from eSchool may on its own drive more investment in eSchool thus causal relationship between<br />
‘net-benefit and eSchool infrastructure quality. The conceptualized eSchool success model is<br />
represented in figure 2.<br />
E-school<br />
infrastructure Quality<br />
Content and<br />
Communication<br />
Quality<br />
Service Quality of<br />
Technical<br />
Department<br />
Training for Use<br />
Effectiveness<br />
Figure 2: Modified D and M model for eSchools<br />
2.2 Validation design<br />
Use<br />
Teacher and Learners<br />
Satisfaction<br />
Ha<br />
Net Benefit<br />
The current study was implemented using NEPAD eSchool project pilot in Kenya as a case. NEPAD e-<br />
Africa Commission is piloting eSchool system in 17 African countries including Kenya. The eSchools<br />
have computers with local area network (LAN) and wide area network (WAN) established through V-SAT<br />
connection to the internet. The project has training component for users. The main objectives of NEPAD<br />
eSchool include improving the quality of graduating students. This would be evident in the extent to which<br />
students are better prepared to fit well in the modern work life. The modern work life is intensive in<br />
computer use and therefore students should be enabled to acquire computer skills, develop team work<br />
skills through collaborations and lifelong learning skills. Developing lifelong learning skills require learners<br />
to engage in self directed learning which is achieved through practicing acquisition of learning materials<br />
independently.<br />
574
Hesbon Nyagowa et al.<br />
The populations for the study were the six secondary schools:Chevakali High school in Western<br />
Province; Maranda High School in Nyanza Province; Menengai Mixed Secondary School in the Rift<br />
Valley Province; Isiolo Girls Secondary School in Eastern region; Mumbi Girls Secondary School in<br />
Central Province; and Wajir Girls Secondary School in North Eastern region. The other part of the<br />
population was users of the eSchool systems in the six schools. The six schools altogether had a<br />
population of 4,936 (1,887 F and 3,049 M) students with 250 teachers. A sample of 1,508 users was<br />
identified using stratified random techniques. Stratification was based on school population, class size in<br />
the different levels (years 9, 10, 11 and 12 of education) and gender (for co-education schools). A<br />
response rate of 51.4% was realized.<br />
Data collected was cleaned and entered into SPSS programme. Reliability and validity of the dater was<br />
assessed. Reliability was assessed using Cronbach Alpha while validity was tested using factor analysis<br />
applying principal component analysis.<br />
3. Results<br />
3.1 Reliability<br />
Test alpha obtained from first half of the data was found to be very closely associated with confirmatory<br />
alpha obtained from second half of the data as shown in table 1 and figure 3.<br />
Table 1: Test and confirmatory alpha<br />
Dimension Test Alpha Confirmatory Alpha<br />
Infrastructure Quality 0.6592 0.704<br />
Information Content and Communication Quality 0.7448 0.7416<br />
Service Quality 0.7182 0.754<br />
Training Effectiveness 0.3962 0.4818<br />
Extent and Exploitative Use 0.7893 0.7766<br />
User Satisfaction 0.916 0.9071<br />
Figure 3: Scatter plots of test and confirmatory alpha<br />
Net Benefits 0.8519 0.822<br />
575
3.2 Validity<br />
Hesbon Nyagowa et al.<br />
Scree plot revealed an elbow at a point lying between ordinate 5 and 9 as in figure 4. This confirmed<br />
existence of at least seven components to the eSchool success model as conceptualized.<br />
Eigenvalue<br />
Figure 4: Scree plot<br />
16<br />
14<br />
12<br />
10<br />
8<br />
6<br />
4<br />
2<br />
0<br />
1<br />
Scree Plot<br />
5<br />
9<br />
13<br />
17<br />
21<br />
Component Number<br />
25<br />
29<br />
3.3 Construct and factor loading<br />
Extraction table was drawn to identify the measurement items which fall under same factor. The mean<br />
factor loading was found to range between 0.4786 and 0.6625 as shown in table 2 and which is<br />
considered acceptable for exploratory study.<br />
Table 2: Mean factor loading<br />
Factor Mean Alpha<br />
Infrastructure Quality 0.5586<br />
Information Content and Communication Quality 0.6176<br />
Service Quality 0.5687<br />
Training Effectiveness 0.4787<br />
Extent and Exploitative Use 0.5534<br />
User Satisfaction 0.6835<br />
Net Benefits 0.6625<br />
4. Conclusion and recommendation<br />
4.1 Conclusion<br />
Confirmatory alpha for the data was generally high above 0.65 with the exception of ‘training<br />
effectiveness which is a new dimension in IS success under exploration. Scree plot confirmed existence<br />
of at least seven dimensions to the eSchool success model conceptualized. The item factor loading to the<br />
dimensions had a mean above 0.4 which is considered good construct validity for sample size of 200<br />
(Hair et al., 1998). Agreement between theoretical conceptualization and empirical validation is therefore<br />
strong enough to support suitability of the eSchool success model for evaluating other eSchool systems.<br />
576<br />
33<br />
37<br />
41<br />
45<br />
49<br />
53<br />
57<br />
61
4.2 Recommendation<br />
Hesbon Nyagowa et al.<br />
Of the seven dimensions, the training effectiveness achieved lowest item mean factor loading of 0.4787.<br />
Some items loading to this dimension were much lower. Since training effectiveness measurement<br />
instrument was exploratory, it would benefit the IS evaluation community to contribute towards refinement<br />
of the construct and its measurement. The study could not collect data from students who have<br />
completed schooling in eSchools and therefore could not measure the all the anticipated benefits of<br />
eSchool. It is recommended that future studied include graduates of eSchools to shed more lights on the<br />
benefits of eSchool. Lastly, with the good agreement between conceptual model and empirical testing,<br />
more studies should consider adopting the eSchool success model to determine its validity.<br />
References<br />
Adeyinka, T. (2009). An evaluation of WEBCT course content management system at the University of Botswana<br />
(Unpublished thesis).<br />
Condie, R. and Livingston, K. (2007). Blending online learning with traditional approaches: Changing practices,<br />
British Journal of Educational Technology, Vol. 38, No. 2, pp. 337-348.<br />
DeLone, H.W. and McLean, R.E. (2002). Information System Success Revisited, Proceedings of the 35 th Hawaii<br />
International <strong>Conference</strong> on System Science.<br />
Delone, W.H. & McLean, E.R. (1992) “Information systems success: The quest for the dependent variable”.<br />
Information Systems Research. 3(1), 60-95.<br />
EL-Halawany, H. and Huwail, E. I. (2008). Malaysian Smart Schools: A Fruitful Case Study for Analysis to Synopsize<br />
Lessons Applicable to the Egyptian Context, International Journal of Education Using ICT, Vol. 4, No. 2.<br />
Grembergen, W.V. (2001). Information technology evaluation methods & management, IDEA, Group Publishing,<br />
London, UK.<br />
Hair, J.F. jr., Anderson, R.E>, Tatham, R.C., and Black, W.C. (1998). Multivariate data analysis, 6 ed. Englewood<br />
Cliffs: Prentice Hall.<br />
International <strong>Conference</strong> on information System (2007). Using Market Exchange Theory to Conceptualize IS Quality<br />
and Re-Conceptualize the IS Success Model, Montreal.<br />
Molla, A. and Licker, S.P. (2001). E-commerce system success: An attempt to extend and respecify the DeLone and<br />
McLean model of IS success, Journal of Electronic Commerce Research, Vol. 2, No. 4, pp. 131-141.<br />
Nobles, A.A.et al. (1989). Electronic school teaching system, United States Patent Number 4820167.<br />
Ong, Chorng-Shyong and Lai, Jung-Yu (2004). Developing an instrument for measuring user satisfaction with<br />
knowledge management systems, proceeding of the 37 th Hawaii International <strong>Conference</strong> on System Science<br />
Petter, S., DeLone, W. and McLean, E. (2008). Measuring Information Systems Success: Models, Dimensions,<br />
Measures, and Interrelationships, <strong>European</strong> Journal of Information Systems, Volume 17, July 2008, pp. 236-<br />
263.<br />
Pitt, F.L., Watson, T.R., and Kavan, B.C. (1995). Service Quality: A measure of Information System Effectiveness,<br />
MIS Quarterly, Vol. 19, No. 2, pp. 173-187.<br />
Pitt, F.L., Watson, T.R., and Kavan, B.C. (1995). Service Quality: A measure of Information System Effectiveness,<br />
MIS Quarterly, Vol. 19, No. 2, pp. 173-187.<br />
Roldan, J.L., Leal, A. (2003). A Validation Test of an Adaptation of the DeLone and McLean’s Model in the Spanish<br />
EIS Field, Idea Group Publishing.<br />
Seddon, B.P. and Kiew, M. (1994). A partial test and development of DeLone and McLean’s Model of IS, Australian<br />
Journal of Information System, Vol.4, No. 1.<br />
Simpson, M., Payne, F. and Condie, R. (2005). Introducing ICT in Secondary Schools: A Context for Reflection on<br />
Management and Professional Norms, Education Management Leadership, 33; 331<br />
UNICT Taskforce (2005). Harnessing the Potential of ICT for Education: A Multi Stakeholder Approach, proceedings<br />
from the Dublin Global Forum for the United Nations ICT Task Force<br />
Wu, J-H. and Wang, Y-M. (2006). Measuring KMS Success: A respecification of DeLone and McLean’s Model,<br />
Information and Management, 43, 728-739.<br />
Ya’Acob, A., Nor, M.F.N and Azman H. (2005). Implementation of Malaysian Smart schools: An investigation of<br />
teaching-learning practices and teacher-student readiness, Internet Journal of e-Language Learning and<br />
Teaching, Vol. 2, No. 2, pp. 16-26.<br />
577