Network Meta-Analysis for Decision-Making
By Sofia Dias, A. E. Ades, Nicky J. Welton and
()
About this ebook
A practical guide to network meta-analysis with examples and code
In the evaluation of healthcare, rigorous methods of quantitative assessment are necessary to establish which interventions are effective and cost-effective. Often a single study will not provide the answers and it is desirable to synthesise evidence from multiple sources, usually randomised controlled trials. This book takes an approach to evidence synthesis that is specifically intended for decision making when there are two or more treatment alternatives being evaluated, and assumes that the purpose of every synthesis is to answer the question "for this pre-identified population of patients, which treatment is 'best'?"
A comprehensive, coherent framework for network meta-analysis (mixed treatment comparisons) is adopted and estimated using Bayesian Markov Chain Monte Carlo methods implemented in the freely available software WinBUGS. Each chapter contains worked examples, exercises, solutions and code that may be adapted by readers to apply to their own analyses.
This book can be used as an introduction to evidence synthesis and network meta-analysis, its key properties and policy implications. Examples and advanced methods are also presented for the more experienced reader.
- Methods used throughout this book can be applied consistently: model critique and checking for evidence consistency are emphasised.
- Methods are based on technical support documents produced for NICE Decision Support Unit, which support the NICE Methods of Technology Appraisal.
- Code presented is also the basis for the code used by the ISPOR Task Force on Indirect Comparisons.
- Includes extensive carefully worked examples, with thorough explanations of how to set out data for use in WinBUGS and how to interpret the output.
Network Meta-Analysis for Decision Making will be of interest to decision makers, medical statisticians, health economists, and anyone involved in Health Technology Assessment including the pharmaceutical industry.
Related to Network Meta-Analysis for Decision-Making
Titles in the series (57)
Using Statistical Methods for Water Quality Management: Issues, Problems and Solutions Rating: 0 out of 5 stars0 ratingsMissing Data in Clinical Studies Rating: 0 out of 5 stars0 ratingsCompeting Risks: A Practical Perspective Rating: 0 out of 5 stars0 ratingsStatistical Methods in e-Commerce Research Rating: 5 out of 5 stars5/5Statistical Methods for Groundwater Monitoring Rating: 0 out of 5 stars0 ratingsIntroduction to Distribution Logistics Rating: 0 out of 5 stars0 ratingsStatistical Monitoring of Complex Multivatiate Processes: With Applications in Industrial Process Control Rating: 0 out of 5 stars0 ratingsBioequivalence Studies in Drug Development: Methods and Applications Rating: 0 out of 5 stars0 ratingsStatistical Practice in Business and Industry Rating: 0 out of 5 stars0 ratingsQuality of Life Outcomes in Clinical Trials and Health-Care Evaluation: A Practical Guide to Analysis and Interpretation Rating: 0 out of 5 stars0 ratingsUncertainty Modeling in Dose Response: Bench Testing Environmental Toxicity Rating: 0 out of 5 stars0 ratingsAn Introduction to Optimal Designs for Social and Biomedical Research Rating: 0 out of 5 stars0 ratingsData Analysis in Forensic Science: A Bayesian Decision Perspective Rating: 0 out of 5 stars0 ratingsModeling Online Auctions Rating: 0 out of 5 stars0 ratingsStatistical Analysis of Cost-Effectiveness Data Rating: 3 out of 5 stars3/5Statistical Framework for Recreational Water Quality Criteria and Monitoring Rating: 0 out of 5 stars0 ratingsBayesian Networks: A Practical Guide to Applications Rating: 3 out of 5 stars3/5Selection Bias and Covariate Imbalances in Randomized Clinical Trials Rating: 0 out of 5 stars0 ratingsEvidence Synthesis for Decision Making in Healthcare Rating: 0 out of 5 stars0 ratingsFinancial Surveillance Rating: 0 out of 5 stars0 ratingsStatistical Analysis and Modelling of Spatial Point Patterns Rating: 0 out of 5 stars0 ratingsStatistical Issues in Drug Development Rating: 0 out of 5 stars0 ratingsSpatio-temporal Design: Advances in Efficient Data Acquisition Rating: 0 out of 5 stars0 ratingsBinary Data Analysis of Randomized Clinical Trials with Noncompliance Rating: 0 out of 5 stars0 ratingsModern Analysis of Customer Surveys: with Applications using R Rating: 3 out of 5 stars3/5Multiple Imputation and its Application Rating: 0 out of 5 stars0 ratingsSimplicity, Complexity and Modelling Rating: 0 out of 5 stars0 ratingsUnderstanding Biostatistics Rating: 0 out of 5 stars0 ratingsComparing Clinical Measurement Methods: A Practical Guide Rating: 0 out of 5 stars0 ratingsA Practical Guide to Cluster Randomised Trials in Health Services Research Rating: 0 out of 5 stars0 ratings
Related ebooks
A Concise Guide to Observational Studies in Healthcare Rating: 0 out of 5 stars0 ratingsCrossover Designs: Testing, Estimation, and Sample Size Rating: 0 out of 5 stars0 ratingsCausality in a Social World: Moderation, Mediation and Spill-over Rating: 0 out of 5 stars0 ratingsStatistical Thinking for Non-Statisticians in Drug Regulation Rating: 0 out of 5 stars0 ratingsDecision Analytics and Optimization in Disease Prevention and Treatment Rating: 0 out of 5 stars0 ratingsIntroduction to Population Pharmacokinetic / Pharmacodynamic Analysis with Nonlinear Mixed Effects Models Rating: 0 out of 5 stars0 ratingsSample Sizes for Clinical, Laboratory and Epidemiology Studies Rating: 0 out of 5 stars0 ratingsStatistical Analysis of Ecotoxicity Studies Rating: 0 out of 5 stars0 ratingsModelling Optimization and Control of Biomedical Systems Rating: 0 out of 5 stars0 ratingsAdvanced Analysis of Variance Rating: 0 out of 5 stars0 ratingsHandbook for Clinical Trials of Imaging and Image-Guided Interventions Rating: 0 out of 5 stars0 ratingsStatistical Methodologies with Medical Applications Rating: 0 out of 5 stars0 ratingsStatistical Methods for Hospital Monitoring with R Rating: 0 out of 5 stars0 ratingsIntegration of Omics Approaches and Systems Biology for Clinical Applications Rating: 0 out of 5 stars0 ratingsQuantile Regression: Estimation and Simulation Rating: 4 out of 5 stars4/5How to Read a Paper: The Basics of Evidence-based Medicine and Healthcare Rating: 4 out of 5 stars4/5Biomolecular and Bioanalytical Techniques: Theory, Methodology and Applications Rating: 0 out of 5 stars0 ratingsNeuroendocrinology of Stress Rating: 0 out of 5 stars0 ratingsUnderstanding Biostatistics Rating: 0 out of 5 stars0 ratingsThe Diagnostic Process: Graphic Approach to Probability and Inference in Clinical Medicine Rating: 0 out of 5 stars0 ratingsHow to Design, Analyse and Report Cluster Randomised Trials in Medicine and Health Related Research Rating: 0 out of 5 stars0 ratingsMultiple Imputation and its Application Rating: 0 out of 5 stars0 ratingsClinical Trials in Older Adults Rating: 0 out of 5 stars0 ratingsSmart Business Problems and Analytical Hints in Cancer Research Rating: 0 out of 5 stars0 ratingsEXPRESS STATISTICS "Hassle Free" ® For Public Administrators, Educators, Students, and Research Practitioners Rating: 0 out of 5 stars0 ratingsStatistics at Square Two: Understanding Modern Statistical Applications in Medicine Rating: 0 out of 5 stars0 ratingsComputational and Statistical Methods for Protein Quantification by Mass Spectrometry Rating: 0 out of 5 stars0 ratingsStatistical Methods in Medical Research Rating: 0 out of 5 stars0 ratings
Mathematics For You
What If?: Serious Scientific Answers to Absurd Hypothetical Questions Rating: 5 out of 5 stars5/5Calculus Made Easy Rating: 4 out of 5 stars4/5Quantum Physics for Beginners Rating: 4 out of 5 stars4/5Algebra - The Very Basics Rating: 5 out of 5 stars5/5Basic Math & Pre-Algebra For Dummies Rating: 4 out of 5 stars4/5My Best Mathematical and Logic Puzzles Rating: 4 out of 5 stars4/5The Golden Ratio: The Divine Beauty of Mathematics Rating: 5 out of 5 stars5/5Pre-Calculus For Dummies Rating: 5 out of 5 stars5/5Algebra II For Dummies Rating: 3 out of 5 stars3/5Game Theory: A Simple Introduction Rating: 4 out of 5 stars4/5Math Magic: How To Master Everyday Math Problems Rating: 3 out of 5 stars3/5Mental Math: Tricks To Become A Human Calculator Rating: 5 out of 5 stars5/5Real Estate by the Numbers: A Complete Reference Guide to Deal Analysis Rating: 0 out of 5 stars0 ratingsThe Little Book of Mathematical Principles, Theories & Things Rating: 3 out of 5 stars3/5Mental Math Secrets - How To Be a Human Calculator Rating: 5 out of 5 stars5/5Alan Turing: The Enigma: The Book That Inspired the Film The Imitation Game - Updated Edition Rating: 4 out of 5 stars4/5Standard Deviations: Flawed Assumptions, Tortured Data, and Other Ways to Lie with Statistics Rating: 4 out of 5 stars4/5Algebra I Workbook For Dummies Rating: 3 out of 5 stars3/5Limitless Mind: Learn, Lead, and Live Without Barriers Rating: 4 out of 5 stars4/5How to Solve It: A New Aspect of Mathematical Method Rating: 4 out of 5 stars4/5Sneaky Math: A Graphic Primer with Projects Rating: 0 out of 5 stars0 ratingsInfinite Powers: How Calculus Reveals the Secrets of the Universe Rating: 4 out of 5 stars4/5Basic Math & Pre-Algebra Workbook For Dummies with Online Practice Rating: 4 out of 5 stars4/5How Math Explains the World: A Guide to the Power of Numbers, from Car Repair to Modern Physics Rating: 3 out of 5 stars3/5Summary of The Black Swan: by Nassim Nicholas Taleb | Includes Analysis Rating: 5 out of 5 stars5/5Logicomix: An epic search for truth Rating: 4 out of 5 stars4/5Calculus Essentials For Dummies Rating: 5 out of 5 stars5/5Introducing Game Theory: A Graphic Guide Rating: 4 out of 5 stars4/5
Reviews for Network Meta-Analysis for Decision-Making
0 ratings0 reviews
Book preview
Network Meta-Analysis for Decision-Making - Sofia Dias
1
Introduction to Evidence Synthesis
1.1 Introduction
This chapter gives a broad overview of indirect comparisons and network meta-analysis. Using a worked example, we look at what these methods are for, what they can do, the core assumptions, the kinds of output they can produce and some of the implications for decision-making. These topics are picked up and covered in more detail in later chapters.
We assume that the evidence that has been assembled for meta-analysis has been identified by a protocol-driven systematic review. Although the methods of systematic review are not covered in this book, it will become clear that the underlying assumptions of meta-analysis put certain constraints on the conduct of the systematic review. This will be discussed further, particularly in Chapter 12 on the validity of network meta-analysis. We also assume that only randomised controlled trials (RCTs) will be included in the synthesis, although we include some comments on how observational studies might be included in Chapter 9 on bias models.
We begin by discussing the purpose of indirect comparisons and network meta-analysis, which is to form the basis for coherent, evidence-based treatment decisions. We then look at some simple methods for networks involving three treatments to illustrate the concepts. Next a worked example of a larger network meta-analysis is presented to show the main properties of the method. After mentioning the assumptions made by network meta-analysis, we briefly look at the question of which trials should be included in a network meta-analysis. This turns into a discussion of the participants, intervention, comparators, outcomes (PICO), the standard ‘script’ that defines inclusion criteria in systematic reviews (Sackett et al., 2000; Higgins and Green, 2008). We briefly consider whether PICO definitions should be reconsidered for the purposes of network meta-analysis.
1.2 Why Indirect Comparisons and Network Meta-Analysis?
While pairwise meta-analysis seeks to combine evidence from trials comparing two treatments, A and B, indirect comparisons involve more than two treatments. A variety of structures can be found, some of which are shown in Figure 1.1. More complex structures are also common and will be presented in later chapters. The first use of indirect comparisons was by Bucher et al. (1997a, 1997b). Their objective was to draw inferences about the relative treatment effects of treatments B and C, using data from B versus A (AB) and C versus A (AC) trials. An indirect estimate of dBC, the effect of C relative to B, is formed by subtracting the direct estimate of dAB from the direct estimate of dAC:
(1.1)
Image described by caption and surrounding text.Figure 1.1 Some examples of connected networks: (a) a pairwise comparison, (b) an indirect comparison via reference treatment A, (c) a ‘snake’ indirect comparison structure, (d) another indirect comparison structure, (e) a simple triangle network and (f) a network of evidence on thrombolytics drugs for acute myocardial infarction (Boland et al., 2003).
While it is true that the indirect comparison provides an estimate of dBC, in practice this may not be the main reason for looking at the AC and AB trials. More often than not, the objective is to consider which is the best of the three treatments. To make such a comparison, based on one AB trial and one AC trial, we have to make the assumption that, although the event rates in the A arm may have been different in the AB and AC trials, the treatment effects, that is, the differences between arms, that are seen in one trial would – subject to sampling error – also be seen in the other. It is for this reason that some authors have referred to this as an ‘adjusted’ indirect comparison (Glenny et al., 2005), because it takes account of the fact that the AB and AC trials will not necessarily have the same absolute event rate on the A arm. But there is probably no need for this terminology, as the idea of an ‘unadjusted’ indirect comparison, based on the absolute event rates, has never been seriously considered in the (pairwise) meta-analysis literature as it breaks the fundamental property of randomisation. (A form of unadjusted indirect comparison has, however, been proposed for use in network meta-analysis under the heading ‘arm-based models’ (Hong et al., 2013, 2015, 2016). These models are reviewed in Chapter 5.)
The assumption that treatment differences on a suitable scale are relatively stable across trials, while absolute event rates can vary, is a fundamental assumption in clinical trials and meta-analysis.
The introduction of ‘loops’ of evidence into a network represents an additional level of complexity. Examples in Figure 1.1e and f include at least one closed loop of evidence. In the triangle structure of Figure 1.1e, for example, we have two sources of evidence on dBC, direct evidence from the BC trials and indirect evidence from the AB and AC trials. Loops have a special significance as they mean that we are in a position to combine the direct and indirect evidence on dBC and form a single pooled estimate. We are also in a position to assess the degree of consistency (or agreement) between the direct and indirect estimates.
Although we continue to use the terminology ‘direct’ and ‘indirect’ throughout the book, these words only have meaning when referring to a specific treatment contrast (we use the term ‘contrast’ to refer to a pairwise comparison between two treatments). As the network grows in complexity, networks involving over 20 treatments are not uncommon (Corbett et al., 2013; Kriston et al., 2014; Mayo-Wilson et al., 2014), and particularly in a decision-making context, the terms ‘direct’ and ‘indirect’ lose all meaning, as evidence that is indirect for one contrast is direct for another, and – depending on network structure – evidence on a single contrast may impact on estimates throughout the whole network. The challenge facing the investigator is first to assemble all the trials comparing any of the treatments of interest in the target population and second to use the totality of this trial evidence to determine an internally consistent set of estimated relative treatment effects between all treatments. Thus, every trial in the network can be considered as contributing ‘direct’ evidence towards the decision problem. Furthermore, if we set aside for the moment the results of the trial, its size, and its position in the network structure, we can say that every trial is equally relevant and has an equal chance of influencing the decision.
The primary objective of network meta-analysis is, therefore, simply to combine all the available data in a coherent and internally consistent way. To see what this means, consider the triangle structure of Figure 1.1e. One way to analyse these data would be to carry out three separate meta-analyses on the AB, AC and BC sets of trials to generate three estimates: . However, these estimates cannot in themselves form a coherent basis for deciding which treatment is the best. Suppose we had three boys: John, Paul and George. John is 2 inches taller than Paul, and Paul is 3 inches taller than George. From this we know that John is 5 inches taller than George. If we were not able to make this inference, if John could be 6 or 2 inches taller than George, then we would not be able to decide who was tallest.
What is needed, therefore, is not the original, unrelated estimates , but three coherent estimates that have the property: . With John = A, Paul = B and George = C, we can deduce that the AC difference is 5, the sum of the AB and BC differences. Network meta-analysis is no more than a technical way of processing the trial-specific treatment effects to arrive at the coherent estimates that are required for coherent decisions about which treatment is best while correctly reflecting parameter uncertainty. However, it has to be appreciated that the coherent estimates represent a view of the data that makes some assumptions. These are set out in Section 1.5.
A network meta-analysis can be carried out whenever the included trials and treatments form a connected network, that is, a network where there is a path linking each treatment to every other (Figure 1.1). In that sense, pairwise meta-analysis and indirect comparisons are special cases of simple network meta-analysis. See Chapter 2 for more details on this.
1.3 Some Simple Methods
The first explicit use of indirect comparisons was due to Bucher et al. (1997a, 1997b). An indirect estimate for the treatment effect of C relative to B is formed by subtracting the estimate of B relative to A from the estimate of C relative to A, as in equation (1.1). Before going further, note that we keep to a convention that dXY is the additional effect of Y relative to X and that we always keep treatment indices in numerical or alphabetical order, dXY or d35; is equivalent to dXY, but we never refer to the former to avoid confusion. A helpful way of remembering how the notation works is to draw the line:
Schematic displaying a horizontal line with short vertical lines at both ends and at the middle labeled A and C and B, respectively.Now consider the relative effect of B compared to A as the ‘distance’ in relative effect terms between points A and B on the line (hence the notation dAB). This immediately clarifies why the indirect estimate of dBC is calculated as in equation (1.1), and not . The variance of the indirect estimate is the sum of the variances of the two direct estimates, as each is derived from different trials including different patients and therefore independent:
(1.2)
See Exercise 1.1. Notice that equation (1.2) tells us that indirect evidence will always have a higher variance than any of its component parts. As the number of links increases, as in Figure 1.1c, the variance of the indirect estimate will become very large (see Exercise 1.3).
We can extend this by pooling the indirect and direct estimates into a single combined estimate. In the triangle structure of Figure 1.1e, we have three sets of trials on AB, AC and BC, and we now conduct three separate pairwise meta-analyses to generate the three pairwise pooled estimates and their variances . We therefore have two independent sources of evidence on dBC, one direct and the other indirect. This invites us to pool the two estimates into a single combined estimate, for example, using inverse-variance weighting:
(1.3)
See Exercise 1.2. Before accepting this combined estimate, it will be important to check that the direct and indirect estimates of dBC are consistent with each other. This is explored in Chapter 7.
The aforementioned procedure gives us a pooled estimate of dBC, but it does not immediately yield the set of coherent estimates of all three parameters. We can generate these by simply repeating the process for the other edges. The pooled estimates , and will have the required property of coherence, but we can arrive at these estimates more efficiently in a single step using suitable computer programs. Indeed, the simple approach to indirect and mixed treatment comparisons in equations (1.2) and (1.3) is extremely limited: as we go from 3 to 4, 5, 10, 20 or 30 treatments, the number of pairwise comparisons that can be made increases from 3 to 6 and then 15, 45, 190 and finally 445. (For S treatments, there are S(S − 1)/2 possible contrasts.) It is evident that distinguishing between direct and indirect evidence in the context of a decision on which of several treatments is best is not only a pointless exercise but also a very tedious one. What is needed is a method that can put all this evidence together to produce the coherent estimates required and some methods for assessing the consistency of evidence from different sources.
1.4 An Example of a Network Meta-Analysis
We now develop a worked example of a network meta-analysis. This is taken from a health technology assessment report (Boland et al., 2003) that was published in 2003, based on an analysis of thrombolytic treatments carried out for NICE, but before the use of network meta-analysis had become a routine feature of NICE technology appraisals. The evidence structure (Figure 1.1f) comprises six active treatments: streptokinase (SK), tissue-plasminogen activator (t-PA), accelerated tissue-plasminogen activator (Acc t-PA), SK + t-PA, tenecteplase (TNK) and reteplase (r-PA). Table 1.1 is a logically ordered layout of the evidence structure that makes it easy to see that the majority of the evidence is on trials that compare SK with t-PA. In the original report the authors presented their findings as shown in Table 1.2, which shows the results of four separate pairwise meta-analyses, each undertaken independently, and therefore not having the required property of coherence.
Table 1.1 The thrombolytics dataset, 14 trials, six treatments (data from Boland et al., 2003): streptokinase (SK), tissue-plasminogen activator (t-PA), accelerated tissue-plasminogen activator (Acc t-PA), tenecteplase (TNK), and reteplase (r-PA) 14 trials.
Fifteen possible pairwise comparisons (a tick represents comparisons made in RCTs). The network is shown in Figure 1.1f.
Table 1.2 Findings from the HTA report on thrombolytics drugs (Boland et al., 2003).
In the absence of a single coherent analysis, the authors were obliged to put forward their conclusions in terms of the pairwise analyses:
… streptokinase is as effective as non-accelerated alteplase … tenecteplase is as effective as accelerated alteplase … reteplase is at least as effective as streptokinase … [is] streptokinase as effective as, or inferior to accelerated alteplase … [is] reteplase as effective as accelerated alteplase or not …
(Boland et al., 2003).
It is difficult to conclude from statements of this sort which treatment performs best. Indeed, given that each estimate is subject to sampling error, under some circumstances, it could prove impossible to identify the best treatment using this approach.
Table 1.3 sets out the full set of estimated median odds ratio (OR) estimates from a fixed effects network meta-analysis in the lower left triangle and the median ORs from the unrelated pairwise analyses in the upper right triangle, along with their 95% credible intervals. Details on how the network estimates were obtained are given in Chapter 2. The network meta-analysis fills in each of the 15 possible pairwise contrasts. See Exercise 1.4.
Table 1.3 Thrombolytics example, fixed effect analysis: odds ratios (posterior medians and 95% CrI).
Upper right triangle gives the ORs from pairwise comparisons (column treatment relative to row), lower left triangle ORs from the network meta-analysis (row treatment relative to column).
It is always informative to look at the relationship between the pairwise estimates and the network estimates. One critical question is whether the direct evidence is consistent with the indirect evidence on each contrast; this cannot be answered from this summary table, but formal methods to detect inconsistency are introduced in Chapter 7. However, there are a number of useful ‘reality’ checks that can be carried out. In some parts of the network, for example, the network meta-analysis does not appear to have had much effect: the OR of t-PA relative to SK and its credible interval are entirely unchanged. Examination of the network structure (Figure 1.1f) immediately explains why: the (SK, t-PA) edge is an isolated spur on the network and the SK versus t-PA trials are the sole source of data on this contrast. On the other hand, the network estimate of the (SK, r-PA) comparison has a tighter credible interval than the pairwise estimate, and this is due to the additional indirect evidence contributed by the (SK, Acc t-PA) trials. However, we do not see the same effect with the (SK, SK + t-PA) comparison because this contrast is informed only by a single three-arm trial so that the ‘loop’ formed by the (SK + t-PA, Acc t-PA) and (SK, Acc t-PA) edges does not constitute additional information. The way that data in one part of the network affects estimates in other parts is determined by the network structure and the quantity of evidence on different edges. Investigators need to be able to ‘read’ the table in this way to gain an understanding of the main ‘drivers’ of the results.
The best treatment is Acc t-PA (OR relative to SK 0.86), with TNK a close second (OR 0.87). Following the accepted principles of pairwise meta-analysis, we distinguish between the relative treatment effects on mortality, which are pooled in evidence synthesis and are expressed here as ORs, and the absolute mortality on specific treatments. Our view is that absolute effects for given treatments must be considered quite separately, because although trials are essential to inform relative effects, they are not necessarily the best source of information on absolute effects of treatments. This is dealt with in detail in Chapter 5. However, if we can assume a distribution for the absolute 35-day mortality on reference treatment SK, which is appropriate for our target populations, we can use the relative effect estimates from the network meta-analysis to infer the absolute mortality on all the treatments. This is shown in Table 1.4. Given a mortality on SK of 8.33%, we expect 7.29% mortality on Acc t-PA and 7.34% on TNK.
Table 1.4 Thrombolytics example, fixed effect analysis: posterior summaries.
Odds ratios, treatment rankings with credible intervals, probability ‘best’, and absolute mortality. The odds ratios are taken from Table 1.3.
Also shown in Table 1.4 are the posterior median ranks of each treatment (rank ‘1’ indicates lowest mortality) and their credible intervals. The posterior distribution of the rankings and the probability that each treatment is ‘best’, which is derived from the rankings, are evidently quite unstable. Indeed, the reason for looking at these statistics is only to make oneself aware of the statistical uncertainty in the treatment differences. Decisions should never be based on these probability statements. For example, if a decision was to be made strictly on the grounds of the effect of treatment on a 35-day survival, regardless of cost or side effects, the treatment chosen should be the one that was associated with the lowest expected mortality, in this case Acc t-PA, as shown by the ranking of the posterior expected effects in Table 1.4. This is not necessarily the same as the treatment that is most likely to be associated with the lowest mortality, which in this case is TNK (see Exercise 1.5). Similarly, decisions based on net benefit would be based on the treatment with the highest expected net benefit, not on the treatment that is most likely to have the highest net benefit (see Chapter 5). These probability statements can be seen as a way of avoiding the complexity of 15 pairwise significance tests, but should be interpreted with caution. In this case, under the fixed effects model, the data appear to rule out SK, t-PA and SK + t-PA, but do not distinguish between Acc t-PA, TNK, and ‘3rd place’ r-PA.
1.5 Assumptions Made by Indirect Comparisons and Network Meta-Analysis
The assumptions made by network meta-analysis can be stated in several ways. First, in a fixed effects model, one is assuming that the dAB effect that is estimated in the AB trials is the same as the dAB effect that would be estimated in the AC, AD, BC and CD trials if these trials had included treatments A and B. With a random effects model, one is assuming that the trial-specific effect δi,AB, which is estimated in trial i comparing AB, is a sample from a random effects distribution with mean dAB and variance , that all the other AB trials are estimating effects from this same distribution and that if A and B were included in the AC, AD, BC and CD trials, they too would be estimating parameters that are random samples from the same random effects distribution (see Chapter 2 for further details).
Another way to express this is to imagine that all the trials were in fact multi-arm trials of treatments A, B, C and D, but only a subset of the treatments are reported and the remaining treatments are missing at random (Little and Rubin, 2002). This does not mean that every treatment is equally likely to be entered into trials, nor even that missingness is unrelated to efficacy. What is required is that missingness mechanism operates with no regard to the true relative efficacy of the treatments (see Chapter 12).
These assumptions are often called the ‘consistency assumptions’. Contrary to what is often stated (Song et al., 2009), the consistency assumptions are not additional assumptions made by network meta-analysis, but they follow from more basic assumptions underlying all meta-analyses (Lu and Ades, 2009), as explained in Chapter 2 and explored further in Chapter 12. Even so, all the doubts that have been expressed about indirect comparisons and network meta-analysis can be seen as doubts about this assumption. For example:
Between-trial comparisons [indirect comparisons] are unreliable. Patient populations may differ in their responsiveness to treatment…. Thus an apparently more effective treatment may … have been tested in a more responsive population
(Cranney et al., 2002).
Placebo controlled trials lacking an active control give little useful information about comparative effectiveness…. Such information cannot reliably be obtained from cross-study [indirect] comparisons, as the conditions of the studies may have been quite different
(ICH Expert Working Group, 2000).
The warnings ‘patient populations may differ’ or ‘the conditions of the study may have been quite different’ constitute a general warning that there may be unrecognised effect modifiers, present in the AB trials but not in the AC trials, so that a BC comparison based on this information may be confounded. This is unquestionably correct. But it is not clear that indirect comparisons and network meta-analysis are really different in this aspect from pairwise meta-analysis, where unrecognised effect modifiers are also present, as proven by the frequent need to fit random effects models (Engels et al., 2000), often with extremely high levels of between-study variation (Turner et al., 2012).
Putting this in another way, while there is no doubt that indirect comparisons may be subject to biases arising from failures in internal or external validity, equation (1.1) tells us that indirect evidence (on the left hand side) can only be biased if direct evidence (on the right hand side) is biased. We can unpick this paradox only by reminding ourselves that the concept of ‘bias’ is relative to a target parameter. For example, suppose our objective is to estimate the effect of C relative to B in patients who are naïve on both B and C, say, . We need to be sure that the included AB and AC trials are based on patients who were indeed naïve to both these treatments. If we are satisfied that and are unbiased estimates of and relating to patients who are naïve on both B and C, we can be satisfied that is also unbiased for this population. If, however, the AC trials are performed on patients who have failed on B, we may think that the direct estimate, , is ‘biased’ with respect to the target population, in this case the indirect estimate will inherit this bias.
This, of course, highlights the importance of paying careful attention to which evidence is included in a network meta-analysis (see Chapter 12), or, alternatively, it invites us to consider some form of covariate adjustment (Chapter 8) or bias modelling (Chapter 9). Although we return to this theme in the later chapters, readers may find it useful to see the direction of travel in advance.
The trial data that are available for evidence synthesis are likely, in many instances, to suffer from a range of imperfections. Trials may be only partially relevant (external bias) because the patient population is mixed, or the clinical settings are different. Further, the conduct of the trial may have characteristics, such as lack of blinding or allocation concealment, that have been shown to be associated with compromised internal validity (internal bias) (Schulz et al., 1995). This has given rise to a range of methods centred on trial ‘quality’. One approach has been the Cochrane Risk of Bias tool (Lundh and Gotzsche, 2008). Another is the Grading of Recommendations Assessment, Development and Evaluation (GRADE) system of assessing the quality of pairwise meta-analyses on a set of pre-determined criteria (Guyatt et al., 2008), extended to deliver quality ratings on network meta-analyses (Puhan et al., 2014; Salanti et al., 2014). These approaches are discussed in more detail in Chapter 12.
Our approach will be somewhat different. Rather than carrying out a network meta-analysis on trial results that we suspect are biased and rating our results from such analysis as unreliable, we would prefer to address any problems with the evidence in advance so as to produce the best possible estimates, whose reliability is appropriately reflected in their credible intervals.
Our strategy to achieve this has several planks. First, knowing that network meta-analysis makes certain assumptions, the most important step is to do everything possible to ensure that the data that are assembled to answer the decision question actually meet those assumptions (see Chapter 12). Second, if there is variation in effect modifiers across the trials in the network, we can use meta-regression (see Chapter 8) to carry out covariate adjustment. Third, if there are potential biases associated with explicit markers of trial ‘quality’, bias adjustment methods can be applied (see Chapter 9). These can take the form of a regression adjustment, or use can be made of a growing literature that quantifies the extent of potential bias in trials (Wood et al., 2008; Savovic et al., 2012a, 2012b).
1.6 Which Trials to Include in a Network
The existence of methods that can compare multiple treatments, based on all the RCT evidence that compares any of them with any of the others, raises a series of further, more strategic issues. When evaluating whether a new product should be adopted by a health service or whether it should be reimbursed by insurance, one approach is to compare the new therapy to one or more standard therapies and issue a recommendation based on whether it is clinically more effective, or more cost-effective, than standard treatments. Another option is to evaluate the new product against all the standard comparators and against all the new competitor products, and then choose the best.
These two procedures appear to be very different, and they clearly have different implications for manufacturers, as the ‘choose the best’ strategy has the effect of removing all but one of a new set of treatments from the field, and yet both decisions could be based on the same set of trials, and indeed on the same network meta-analysis. This is because the ‘evidence space’ does not have to coincide with the ‘decision space’. The procedure of ranking and choosing the best treatment does not have to be applied to all the treatments; it can apply only to the ones that are relevant for the decision, and it is simpler and probably fairer to all concerned, for both decisions to have the same evidential base.
1.6.1 The Need for a Unique Set of Trials
A similar principle lay behind the decision in the methods guide by the National Institute for Health and Clinical Excellence (NICE, 2008b) to permit for the first time the inclusion in evidence networks of treatments that were not of intrinsic interest themselves, but which formed a link between otherwise disconnected networks (Figure 1.2a). If there was no treatment G that had been trialled with at least one of the treatments A, B and C and one of the treatments X and Y, then there would be no way of using network meta-analysis to compare X or Y with A, B or C. However, once a treatment G that links the two networks has been identified (Figure 1.2b), are there other treatments that could have made the required connection? If there are, these should also be included, to prevent any suggestions of ‘gaming’ with the network structure (choosing comparisons that flattered one particular product over the others). The addition of treatment H (Figure 1.2c) illustrates this situation. Finally, if the comparator set is A, B, C, D, G, H, X and Y, we must ask: have all the trials comparing any two or more of these treatments been included? These are no less ‘relevant’ to the decision problem than the others and should therefore also be included (Figure 1.2d).
Image described by caption and surrounding text.Figure 1.2 (a) An unconnected network; (b) use of an intermediate treatment to link a network (dashed lines); (c) other intermediate treatments that could also be used as links; and (d) incorporation of all trials comparing the enlarged set of treatments.
To ensure a transparent process that removes any temptation to manipulate trial inclusion/exclusion criteria, a procedure is required, which identifies a unique set of trials. Following the illustration in Figure 1.2, we suggest the following algorithm:
Identify all the comparators of interest.
Identify all the trials that compare two or more of the comparators in the population of interest.
Remove trial arms that are not comparators of interest from trials with more than two arms. If this forms a connected network, stop.
If this is not a connected network,
Identify further treatments that have been compared with any treatments in both sub-networks (including any additional arms that may have been removed in step 3).
Add these to the comparators of interest.
Carry out steps (2) and (3) with the larger set of comparators.
This procedure identifies a minimum set of trials that form a connected network while still being unique. Of course, this does not necessarily guarantee that a connected network can be obtained (see Section 1.7). In addition, a degree of subjective judgement remains as there can always be debate about the relation between a trial population, often poorly described, and the target ‘population of interest’, as in any synthesis.
1.7 The Definition of Treatments and Outcomes: Network Connectivity
This raises the question: supposing no connected network of trial evidence can be formed, can one use one-arm studies? Can one use observational studies? Rather than raise false hopes in the reader’s mind, we make it clear from the outset that evidence synthesis on treatment effects, as conceived in this book, is based exclusively on randomised trial evidence. Nevertheless, the methods discussed in Chapter 9 on bias models can be applied to observational data on treatment efficacy. These methods force the user to be absolutely explicit about the distribution of potential bias. Similarly, the entire book follows the tradition of meta-analytic methods that concerns the statistical combination of relative treatment effects, which rules one-arm studies out of bounds – except when they are used to inform a natural history model. These points are taken up in Chapter 5.
With these provisos, the issue of network connectedness cannot be separated from how treatments and outcomes are defined. In traditional pairwise meta-analysis, this has been determined by the PICO script (Sackett et al., 2000). In the era of network meta-analysis, we suggest in Sections 1.7.1–1.7.3 that there is a need to liberalise the I, C and O constraints on study inclusion while tightening up quite considerably on the P.
1.7.1 Lumping and Splitting
With pairwise meta-analyses it remains common to ‘lump’ treatments together. Sometimes this is done so there will be enough trials to justify a quantitative synthesis (Gotzsche, 2000). One example of this is a comparison of thrombolytics treatments for acute myocardial infarction against the surgical procedure percutaneous transluminal coronary angioplasty (PTCA) (Keeley et al., 2003b). This analysis, which lumps the thrombolytics treatments (Figure 1.1f), found that PTCA was (statistically) significantly superior to thrombolytics taken as a class. However, in subsequent published correspondence, a ‘splitter’ objected that this was not the relevant comparison; instead one should compare PTCA with the best of the thrombolytics treatments, which was Acc t-PA (Fresco et al., 2003). On this comparison – based on very few trials – PTCA was superior, but this was no longer statistically significant. Revealingly, the original author, evidently a committed ‘lumper’, declared herself ‘mystified’ by the criticism of ‘lumping trials together’ because ‘meta-analyses, by definition, pool data’ (Keeley et al., 2003a).
In a network meta-analysis one can accept that different thrombolytics may not all be equally effective while at the same time retaining statistical power by combining direct and indirect evidence. A network meta-analysis of the same dataset demonstrated a statistically significant advantage of PTCA over Acc t-PA (Caldwell et al., 2005). See Chapters 2, 3 and 7 for more details on this example.
The default network meta-analysis, therefore, treats every intervention, every dose and every treatment combination a separate ‘treatment’. On the other hand, the finer the distinction between treatments, the greater the risk of disconnected networks. There are three ways in which this can be ameliorated: dose response models, treatment combination models or class effects models. The effect of these techniques is to reduce – sometimes dramatically – the number of effective parameters that need to be estimated, and this impacts powerfully on the connectivity of the network. These models are described further in Chapter 8.
1.7.2 Relationships Between Multiple Outcomes
Another way to create more connections is to include more than one ‘outcome’ in the same synthesis. Trials very frequently report different outcomes at different follow-up times, and some report more than one outcome at more than one time. However, if we restrict ourselves to one particular outcome, or one particular time, we will produce evidence networks that are sparse, poorly connected, or even disconnected. If, on the other hand, the different outcomes can be included in the same analysis, a more connected network is available. What is required is an explicit model for how the treatment effects at different times or on different outcomes are related. However, this is not a trivial undertaking. Chapter 11 gives a number of examples showing how clinical and logical relationships between outcomes can be exploited to provide a single, coherent analysis of all the treatments on all the outcomes and at the same time creating connected networks from unconnected components.
1.7.3 How Large Should a Network Be?
There is no theoretical limit to how large a network can be, and a question that is often asked is ‘how large should a network be?’ In theory, the larger the network, the more robust the conclusions will be, in the sense that the conclusions will become increasingly insensitive to the inclusion or exclusion of any particular trial. However, there is little doubt that the larger the network becomes, the greater the risk of clinical heterogeneity in the trial populations, particularly as one reaches further back to include older trials. This is because new treatments tend to be introduced for more severely ill patients. As time goes on more data on side effects accumulates, and eventually clinicians start to feel it might benefit less ill patients who are then included in later trials.
For this reason it may often turn out that larger networks will fail to produce more precise estimates: greater clinical heterogeneity may lead to greater statistical heterogeneity in treatment effects, larger between-trial variances in random effects models and therefore lower precision in estimates of mean treatment effects (Cooper et al., 2011). It may also increase the risk of inconsistency, as older treatments will have been compared on more severely ill populations than newer treatments. Thus different sources of evidence (direct and indirect) may conflict.
1.8 Summary
In this introductory chapter we have looked at why indirect comparisons and network meta-analysis are needed. We have emphasised that their key property is that they produce an internally coherent set of estimates, which is essential for rational decision-making, whether based on efficacy of cost-effectiveness. We have observed that the simpler methods, which can be applied to triangular networks, have limited scope, and they become completely unfeasible as the networks grow in complexity. We have set out results from a worked example and shown how to interpret them and how comparison with pairwise meta-analysis results can help us understand the main ‘drivers’ of the results. Finally we have set out the key assumptions in an informal way and looked at some of the strategic policy issues that the existence of network meta-analysis immediately raises: what treatments should be compared, which trials should be included, and how large should a network of comparisons be.
As we work through the book, many of these themes will be picked up again and covered in more detail.
1.9 Exercises
1.1 This table gives results, in the form of log hazard ratios from a study of virologic suppression following highly active antiretroviral therapy for HIV (Chou et al., 2006). Obtain an indirect estimate for .
1.2 Pool the direct and indirect estimates on BC from the previous exercise. Comment on the validity of the pooled estimate.
1.3 Find an indirect estimate for and its standard error using the data below. (Hint: draw the treatment network first.)
1.4 The table below shows posterior mean and median odds ratio from two sets of analyses on a triangle network of three treatments A, B and C. One of these analyses represents pairwise summaries from unrelated meta-analyses, while the other comes from a network meta-analysis. Which is which?
1.5 *In Section 1.4 it was noted that the treatment with the highest expected treatment effect may not always be the same as the treatment that is most likely to have the highest treatment effect. Under what circumstances can this happen?
2
The Core Model
2.1 Bayesian Meta-Analysis
There is a substantial literature on statistical methods for meta-analysis, going back to methods for combination of results from two-by-two tables and the introduction of random effects meta-analysis (DerSimonian and Laird, 1986), an important benchmark in the development of the field. Over the years methodological and software advances have contributed to the widespread use of meta-analytic techniques. A series of instructional texts and reviews have appeared (Cooper and Hedges, 1994; Smith et al., 1995; Egger et al., 2001; Sutton and Abrams, 2001; Higgins and Green, 2008; Sutton and Higgins, 2008), but there have been only a few attempts to produce a comprehensive guide to the statistical theory behind meta-analysis (Whitehead and Whitehead, 1991; Whitehead, 2002).
We present a single unified framework for evidence synthesis of aggregate data from RCTs that delivers an internally consistent set of estimates while respecting the randomisation in the evidence (Glenny et al., 2005). The core models presented in this chapter can synthesise data from pairwise meta-analysis, indirect and mixed treatment comparisons, that is, network meta-analysis (NMA) with or without multi-arm trials (i.e. trials with more than two arms), without distinction. Indeed, pairwise meta-analysis and indirect comparisons are special cases of NMA, and the general NMA model and WinBUGS code presented here instantiate that.
We take a Bayesian approach to synthesis using Markov chain Monte Carlo (MCMC) implemented in WinBUGS (Lunn et al., 2013) mainly due to the flexibility of the Bayesian approach, ease of implementation of models and parameter estimation in WinBUGS and its natural extension to decision modelling (see Preface). While we do not describe Bayesian methods, MCMC or the basics of running WinBUGS in this book, basic knowledge of all these concepts is required to fully understand the code presented. However, the underlying principles and concepts described can still be understood without detailed knowledge of Bayesian methods, MCMC or WinBUGS. Further references and a guide on how to read the book are given in the Preface.
We begin by presenting the core Bayesian models for meta-analysis of aggregate data presented as the number of events observed in a fixed and known number of individuals – commonly referred to as binary, dichotomous or binomial data. We describe fixed and random treatment effects models and show how the core models, based on the model proposed by Smith et al. (1995) for pairwise meta-analysis, are immediately applicable to indirect comparisons and NMA with multi-arm trials without the need for any further extension. This general framework is then applied to other types of data in Chapter 4.
2.2 Development of the Core Models
Consider a set of M trials comparing two treatments, 1 and 2, in a pre-specified target patient population, which are to be synthesised in a meta-analysis. A fixed effects analysis would assume that each study i generates an estimate of the same parameter d12, the relative effect of treatment 2 compared with 1 on some scale, subject to sampling error. In a random effects model, each study i provides an estimate of the study-specific treatment effects δi,12, the relative effect of treatment 2 compared with 1 on some scale in trial i, which are assumed not to be equal but rather exchangeable. This means that all δi,12 are ‘similar’ in a way that assumes that the trial labels, i, attached to the treatment effects δi,12 are irrelevant. In other words, the information that the trials provide is independent of the order in which they were carried out, over the population of interest (Bernardo and Smith, 1994). The exchangeability assumption is equivalent to saying that the trial-specific treatment effects come from a common distribution, and this is often termed the ‘random effects distribution’. In a Bayesian framework, the study-specific treatment effects, δi, are estimated for each study and are often termed the ‘shrunken’ estimates. They estimate the ‘true’ treatment effect for each trial and therefore ‘shrink’ the observed trial effects towards the random effects mean. See Chapters 3 and 8 for further details. The normal distribution is often chosen, so