13.07.2015 Views

Evaluating non-randomised intervention studies - NIHR Health ...

Evaluating non-randomised intervention studies - NIHR Health ...

Evaluating non-randomised intervention studies - NIHR Health ...

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

<strong>Health</strong> Technology Assessment 2003; Vol. 7: No. 27Unpredictability in biasWhen bias acts unpredictably, it will sometimeslead to an overestimation and sometimes tounderestimation of an effect. Although thesebiases may on average ‘cancel out’ across a set of<strong>studies</strong> such that no difference is observed inaverage ORs, the biases will still affect the resultsof individual <strong>studies</strong>. The presence of systematicbias may therefore be missed if the comparison ofresults is restricted to a comparison of averagevalues, as was done in five of the eight previousreviews summarised in Chapter 3. 25–28,32Unpredictable over- and underestimation willincrease the variability (or heterogeneity) of theresults of a set of <strong>studies</strong>. In the concurrentcomparisons such an increase in variability(measured by the standard deviation) wasobserved for the IST (Table 15), even though theaverage treatment effects in the concurrentlycontrolled and <strong>randomised</strong> <strong>studies</strong> were thesame. A similar pattern was observed forhistorically controlled <strong>studies</strong> generated from theIST when the haphazard within-region timetrends were aggregated in the overall analysis(Table 16).How do these biases occur, and how do they differfrom the variability seen between RCTs? Variabilityalways occurs between the results of multipleRCTs. The principal reason is the ‘play of chance’or sampling variation. A treatment effect observedin a particular RCT is unlikely to be the preciseeffect of the <strong>intervention</strong>. For example, randomlydividing the study sample into two does notguarantee that the groups are identical in allrespects, and the differences that do exist in casemixwill lead to either under- or overestimates ofthe treatment effect in an individual trial. We donot normally talk about these differences asbiases, but rather as uncertainties. We know thedistribution with which under-and overestimatesarise in RCTs, enabling us to draw correctinferences within specified degrees of certainty. Wecannot identify whether a particular trial isaffected by such bias, but we can calculate boundswithin which we are reasonably sure possible biasis encompassed, which we term confidenceintervals. Importantly, we know that the possibledifferences between the groups due to samplingvariation (and hence confidence intervals) reducewith increasing sample size.The extra variability we see in the <strong>non</strong><strong>randomised</strong><strong>studies</strong> arises in a similar but moretroubling manner. Rather than randomly dividinga single group of individuals, we start with twodifferent groups of individuals. We therefore startwith differences between the groups in measurableand unmeasurable factors. These potentiallyinclude differences in case-mix, additionaltreatments and methods of assessment ofoutcome. Importantly, in addition to not beingable to identify all these differences, we may notknow in which way many of the factors act, so thatthere is overall uncertainty as to whether they willcause under- or overestimates of the treatmenteffect. Sampling from these populationsintroduces the same sampling variation as in theRCT. While we can estimate the impact of thesampling variation (and calculate standardconfidence intervals), there is no mathematicalway of knowing how pre-existing differencesbetween the groups behave. It is therefore notpossible to include an allowance in the confidenceinterval for a single study that accounts for theextra uncertainty introduced through unsystematicbias. As we cannot mathematically allow for thisvariation when drawing conclusions, it isappropriate to call such extra variation ‘bias’ eventhough it is ‘uncertain’. In contrast to samplingvariation, the extra uncertainty is independent ofsample size as it is a feature of the pre-existingdifferences between the two populations fromwhich the samples were drawn.Our resampling <strong>studies</strong> provide a uniqueopportunity to calculate the distribution of thisextra uncertainty for the specific situations studiedin the IST and ECST by calculating the increase invariance seen with <strong>non</strong>-<strong>randomised</strong> concurrentlycontrolled <strong>studies</strong> compared with RCTs. Thiscomputation is possible as we ensured that foreach study the RCTs are the same size as theconcurrent comparisons, such that the differencesin variability cannot be explained by differences insampling variability. The results of thesecomputations are given in Table 19. The extravariance in log OR was 0.61 for regional ISTcomparisons, 0.57 for UK city IST comparisonsand 0.01 for regional ECST comparisons. Giventhese estimates, it is possible to calculate newadapted confidence intervals for these <strong>studies</strong> thatallow for these potential uncertain biases inaddition to sampling variation. They areexpressed in Table 19 as multiplicative increases inthe width of the standard confidence intervals. Assampling variability decreases with increasingsample size but the unsystematic bias remainsconstant, the ratio of the extra allowance in thewidth of the confidence interval due tounsystematic bias increases with sample size. Theratios presented in Table 19 reveal that standardconfidence intervals for many <strong>non</strong>-<strong>randomised</strong>59© Queen’s Printer and Controller of HMSO 2003. All rights reserved.

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!