2005 SAGES Abstracts
2005 SAGES Abstracts
2005 SAGES Abstracts
Create successful ePaper yourself
Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.
POSTER ABSTRACTS<br />
urement tools. Rating scales used during live observation of<br />
operative performance may be divided into global, procedurespecific<br />
and checklist-based. These scales may also be used<br />
for retrospective video-based observation with a resultant<br />
increase in objectivity of the assessment, though it is unclear<br />
which type of scale would be the most reliable for this purpose.<br />
The aim of this study is to define the optimal scale for<br />
video-based assessment of technical skill in laparoscopic surgery.<br />
Methods: Twenty-eight laparoscopic cholecystectomies (LC)<br />
performed by a total of 17 surgeons were recorded, following<br />
the confirmation of informed consent from all patients and<br />
surgeons involved in the cases. Each LC was assigned a code,<br />
and rated by three other experienced laparoscopic surgeons in<br />
a blinded manner on each of four rating scales. These were<br />
the OSATS global rating scale of technical skill, a modified<br />
OSATS global rating scale specifically adapted for video-based<br />
assessment of laparoscopic surgery, a procedure-specific rating<br />
scale, and finally a procedural checklist. Scores for each<br />
procedure were collated, and analysed using Cronbach?s<br />
alpha test for inter-rater reliability.<br />
Results: The inter-rater reliabilities for each of the four rating<br />
scales were as follows: OSATS (0.79); modified OSATS for<br />
video-based observation (0.75); procedure-specific rating scale<br />
(0.82); and procedural checklist (0.48).<br />
Conclusions: OSATS, modified OSATS and the procedure-specific<br />
scale all possess high inter-rater reliabilities, even though<br />
the assessment is based upon retrospective, blinded, videobased<br />
observation of performance. Such a method can ensure<br />
increased objectivity and may be performed after the event,<br />
making the assessment more feasible by not requiring the<br />
presence of expert faculty to be available at the time of the<br />
procedure. Finally, this study did not find the checklist to be a<br />
reliable method of assessment, replicating the results of previous<br />
work in this field.<br />
P337–Minimally Invasive Other<br />
THE OPTIMAL METHOD OF TRAINING ON A VIRTUAL REALI-<br />
TY LAPAROSCOPIC SIMULATOR, R Aggarwal MD, G Dew,J<br />
Hance MD,N Selvapatt,A Darzi MD, Department of Surgical<br />
Oncology & Technology, Imperial College London, UK.<br />
Introduction: Virtual reality (VR) laparoscopic simulators have<br />
been shown to teach the skills required for laparoscopic surgery.<br />
However, with increased fidelity and on-screen feedback,<br />
the question should be asked whether it is necessary to have<br />
an expert tutor available during every training session? It is<br />
the aim of this study to determine the optimal method of feedback<br />
to teach the skills required for laparoscopic suturing.<br />
Methods: Forty laparoscopic novices trained on laparoscopic<br />
VR simulators in a stepwise approach, commencing with a<br />
previously validated basic skills curriculum on the MIST-VR<br />
simulator. Subjects then completed five half-hour sessions on<br />
the laparoscopic suturing module of the LapSim VR simulator.<br />
Prior to their first LapSim session, subjects were randomly<br />
allocated to one of four groups, each group receiving a different<br />
type of feedback during their training sessions. The groups<br />
received either expert tutor feedback, checklist feedback, VR<br />
simulator feedback, or no feedback. Assessments of laparoscopic<br />
suturing skill were carried out at the beginning and end<br />
of each subject?s training period, using a synthetic bowel<br />
model placed in a video trainer. Each subject?s performance<br />
was scored objectively using a validated motion analysis system,<br />
together with blinded checklist scoring of videos of each<br />
procedure. Data analysis used non-parametric tests, p