Health Information Management: Integrating Information Technology ...
Health Information Management: Integrating Information Technology ...
Health Information Management: Integrating Information Technology ...
Create successful ePaper yourself
Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.
202 EVALUATION OF PCISS<br />
evaluations, integrating quantitative and qualitative research methods in this way<br />
requires that these quantitative designs should be quite ‘modest’ in nature: before<br />
—after designs are useful, but more strictly experimental designs (like RCTs)<br />
exceed their aim. Their design is too rigorous to be of help, often directly hinders<br />
the organizational change process that is occurring, and they yield very few<br />
results in the light of their costs.<br />
In the next paragraph we will elaborate on the different steps that come to play<br />
when designing and conducting an evaluation.<br />
STEPS IN DESIGNING AN EVALUATION<br />
In this paragraph we will describe the design of an evaluation as following<br />
several, successive steps (as shown in Figure 11.1) (Heathfield et al. 2001). We<br />
will focus primarily, but not exclusively, on formative evaluation. In reality, the<br />
individual steps are often difficult to separate, and usually impossible to manage<br />
as a simple and steady sequence. The process of designing an evaluation is one<br />
of adjustment and compromise, where choices concerning study design,<br />
evaluation criteria and data collection methods must be offset against the<br />
constraints of conflicting stake-holders’ aims, deadlines, resources available, the<br />
intrusiveness of the evaluation and ethical considerations. The design of an<br />
evaluation is as much a social and political process as that of IT procurement and<br />
implementation itself. By using case material (from the UK and the<br />
Netherlands), we will show that taking these steps can be quite complicated.<br />
Every next step is determined by the situation at hand. Though the cases differ in<br />
technology, set up and success, both can be seen as a realistic example of the<br />
implementation of a PCIS in health care. The more general ‘Issues’ and ‘Lessons<br />
learned’ at the end of each step are based on material from these two sites as<br />
well as other case material.<br />
CASE STUDY<br />
The first case is based upon a study in which the UK NHS Executive,<br />
through its <strong>Information</strong> <strong>Management</strong> Group, initiated two programmes of<br />
work which sought to improve patient records using information<br />
technology: the Electronic Patient Record (EPR) project and The Clinical<br />
Workstation (ICWS) project. The EPR project was a three-year research<br />
and development programme to assist clinicians in acute hospitals to<br />
provide better care to patients through the use of electronic patient record<br />
systems. The ICWS project was a two-year project concerned<br />
with providing interface facilities to an EPR. A number of sites with<br />
established Hospital <strong>Information</strong> Systems and sites in the early stages of<br />
procurement and implementation but with an organizational commitment<br />
to an EPR were selected. External evaluation was seen as essential to both<br />
the EPR and ICWS projects and the evaluation covered all five sites.