13.07.2015 Views

Evaluating non-randomised intervention studies - NIHR Health ...

Evaluating non-randomised intervention studies - NIHR Health ...

Evaluating non-randomised intervention studies - NIHR Health ...

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

<strong>Health</strong> Technology Assessment 2003; Vol. 7: No. 27Abstract<strong>Evaluating</strong> <strong>non</strong>-<strong>randomised</strong> <strong>intervention</strong> <strong>studies</strong>JJ Deeks, 1* J Dinnes, 2 R D’Amico, 1 AJ Sowden, 3 C Sakarovitch, 1 F Song, 4M Petticrew 5 and DG Altman 1In collaboration with the International Stroke Trial and the European Carotid Surgery Trial Collaborative Groups1 Centre for Statistics in Medicine, Institute of <strong>Health</strong> Sciences, Oxford, UK2 Southampton <strong>Health</strong> Technology Assessments Centre, University of Southampton, UK3 NHS Centre for Reviews and Dissemination, University of York, UK4 Department of Public <strong>Health</strong> and Epidemiology, University of Birmingham, UK5 MRC Social and Public <strong>Health</strong> Sciences Unit, University of Glasgow, UK* Corresponding authorObjectives: To consider methods and related evidencefor evaluating bias in <strong>non</strong>-<strong>randomised</strong> <strong>intervention</strong><strong>studies</strong>.Data sources: Systematic reviews and methodologicalpapers were identified from a search of electronicdatabases; handsearches of key medical journals andcontact with experts working in the field. Newempirical <strong>studies</strong> were conducted using data from twolarge <strong>randomised</strong> clinical trials.Methods: Three systematic reviews and new empiricalinvestigations were conducted. The reviews considered,in regard to <strong>non</strong>-<strong>randomised</strong> <strong>studies</strong>, (1) the existingevidence of bias, (2) the content of quality assessmenttools, (3) the ways that study quality has been assessedand addressed. (4) The empirical investigations wereconducted generating <strong>non</strong>-<strong>randomised</strong> <strong>studies</strong> fromtwo large, multicentre <strong>randomised</strong> controlled trials(RCTs) and selectively resampling trial participantsaccording to allocated treatment, centre andperiod.Results: In the systematic reviews, eight <strong>studies</strong>compared results of <strong>randomised</strong> and <strong>non</strong>-<strong>randomised</strong><strong>studies</strong> across multiple <strong>intervention</strong>s using metaepidemiologicaltechniques. A total of 194 tools wereidentified that could be or had been used to assess<strong>non</strong>-<strong>randomised</strong> <strong>studies</strong>. Sixty tools covered at leastfive of six pre-specified internal validity domains.Fourteen tools covered three of four core items ofparticular importance for <strong>non</strong>-<strong>randomised</strong> <strong>studies</strong>. Sixtools were thought suitable for use in systematicreviews. Of 511 systematic reviews that included <strong>non</strong><strong>randomised</strong><strong>studies</strong>, only 169 (33%) assessed studyquality. Sixty-nine reviews investigated the impact ofquality on study results in a quantitative manner. Thenew empirical <strong>studies</strong> estimated the bias associatedwith <strong>non</strong>-random allocation and found that the biascould lead to consistent over- or underestimations oftreatment effects, also the bias increased variation inresults for both historical and concurrent controls,owing to haphazard differences in case-mix betweengroups. The biases were large enough to lead <strong>studies</strong>falsely to conclude significant findings of benefit orharm. Four strategies for case-mix adjustment wereevaluated: <strong>non</strong>e adequately adjusted for bias inhistorically and concurrently controlled <strong>studies</strong>. Logisticregression on average increased bias. Propensity scoremethods performed better, but were not satisfactory inmost situations. Detailed investigation revealed thatadequate adjustment can only be achieved in theunrealistic situation when selection depends on a singlefactor.Conclusions: Results of <strong>non</strong>-<strong>randomised</strong> <strong>studies</strong>sometimes, but not always, differ from results of<strong>randomised</strong> <strong>studies</strong> of the same <strong>intervention</strong>. Non<strong>randomised</strong><strong>studies</strong> may still give seriously misleadingresults when treated and control groups appear similarin key prognostic factors. Standard methods of case-mixadjustment do not guarantee removal of bias. Residualconfounding may be high even when good prognosticdata are available, and in some situations adjustedresults may appear more biased than unadjusted results.Although many quality assessment tools exist and havebeen used for appraising <strong>non</strong>-<strong>randomised</strong> <strong>studies</strong>, mostomit key quality domains. <strong>Health</strong>care policies basedupon <strong>non</strong>-<strong>randomised</strong> <strong>studies</strong> or systematic reviews of<strong>non</strong>-<strong>randomised</strong> <strong>studies</strong> may need re-evaluation if theuncertainty in the true evidence base was not fullyappreciated when policies were made. The inability ofiii© Queen’s Printer and Controller of HMSO 2003. All rights reserved.

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!