Home | About | Journals | Submit | Contact Us | Français |

**|**BMC Med Res Methodol**|**v.10; 2010**|**PMC2906500

Formats

Article sections

- Abstract
- Background
- Methods
- Results
- Discussion
- Conclusions
- Competing interests
- Authors' contributions
- Appendix - WinBUGS code
- Pre-publication history
- References

Authors

Related links

BMC Med Res Methodol. 2010; 10: 54.

Published online 2010 June 10. doi: 10.1186/1471-2288-10-54

PMCID: PMC2906500

Beth S Woods: moc.semoctuodrofxo@sdoow.hteb; Neil Hawkins: moc.semoctuodrofxo@snikwah.lien; David A Scott: moc.semoctuodrofxo@ttocs.divad

Received 2009 May 9; Accepted 2010 June 10.

Copyright ©2010 Woods et al; licensee BioMed Central Ltd.

This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

This article has been cited by other articles in PMC.

Data on survival endpoints are usually summarised using either hazard ratio, cumulative number of events, or median survival statistics. Network meta-analysis, an extension of traditional pairwise meta-analysis, is typically based on a single statistic. In this case, studies which do not report the chosen statistic are excluded from the analysis which may introduce bias.

In this paper we present a tutorial illustrating how network meta-analyses of survival endpoints can combine count and hazard ratio statistics in a single analysis on the hazard ratio scale. We also describe methods for accounting for the correlations in relative treatment effects (such as hazard ratios) that arise in trials with more than two arms. Combination of count and hazard ratio data in a single analysis is achieved by estimating the cumulative hazard for each trial arm reporting count data. Correlation in relative treatment effects in multi-arm trials is preserved by converting the relative treatment effect estimates (the hazard ratios) to arm-specific outcomes (hazards).

A worked example of an analysis of mortality data in chronic obstructive pulmonary disease (COPD) is used to illustrate the methods. The data set and WinBUGS code for fixed and random effects models are provided.

By incorporating all data presentations in a single analysis, we avoid the potential selection bias associated with conducting an analysis for a single statistic and the potential difficulties of interpretation, misleading results and loss of available treatment comparisons associated with conducting separate analyses for different summary statistics.

Network meta-analyses enable us to combine trials that compare different sets of treatments, and form a network of evidence, within a single analysis [1] and to use all available direct and indirect evidence to inform a given comparison between treatments. Network meta-analysis is based on the assumption that, on a suitable scale, we can add and subtract within-trial estimates of relative treatment effects i.e. the difference in effect between treatments A & B (*d*_{AB}) is equal to the difference in effects between treatments A & C and B & C (*d*_{AB }= *d*_{AC }- *d*_{BC}) [1-3].

In this paper we show how network meta-analyses of survival endpoints can be conducted on the hazard ratio scale when some or all trials report cumulative count data. We also describe how trials with more than two arms reporting relative treatment effects (such as hazard ratios) should be included.

A survival endpoint is one where, over time, an increasing number of patients experience an event. Although death is the ultimate survival endpoint, many other endpoints may also be considered as survival endpoints. For example, in the study of epilepsy, seizure freedom may be regarded as a survival endpoint as over time the number of patients experiencing one or more seizures can only increase and the probability of being seizure free decreases. In contrast, the number of patients with a greater than 50% reduction in seizure frequency relative to baseline is not a survival endpoint as the number of patients achieving this response can increase or decrease over time.

Data on survival endpoints may be expressed either in the form of hazard ratio statistics derived from parametric or non-parametric methods of survival analysis [4,5] or as cumulative count statistics (the total number of subjects who have experienced an event at a specific time point). Unlike cumulative count statistics, hazard ratio statistics from survival analysis account for censoring, incorporate time to event information, and may be adjusted for co-variables. Where both hazard ratio and cumulative count statistics are available for a given study, use of the hazard ratio data may therefore be considered preferable.

The methods described in this paper allow hazard ratio and cumulative count survival statistics to be combined within a single network meta-analysis on the log-hazard scale; this might be termed a multi-statistic evidence synthesis. Treatment effects can then be estimated based on an inclusive set of data, and separate analyses for hazard ratio and count statistics are avoided [6-9].

Network meta-analyses should account for the correlations in relative treatment effect estimates that arise from trials with more than two treatment arms (multi-arm trials) [10]. These correlations are accounted for 'by default' when count statistics for individual trial arms are included in a network meta-analysis [10]. When hazard ratio statistics are used, we show how these correlations can be accounted for by deriving estimates of the mean log hazards (and their variances) for individual trial arms.

We use an example data set describing mortality in randomised controlled trials of treatments for chronic obstructive pulmonary disease (COPD). The data set consists of a subset of trials from Baker et al [11]. We chose five trials providing mortality data for the following comparators: salmeterol, fluticasone, salmeterol fluticasone combination (SFC) or placebo. Three of the trials report count data [12-14], one reports hazard ratio data from a two arm trial [15] and one reports hazard ratio data from a multi (four)-arm trial [16]. Hazard ratio data were obtained from the trial publications and an existing meta-analysis of hazard ratio data [17].

The count statistics used are presented as Table Table11 and the hazard ratio statistics as Table Table2.2. The derived estimates of the mean log hazard ratio and it's standard error, required for the analysis, are also presented in Table Table2,2, these were estimated using formulae (1) and (2):

(1)

(2)

The results of this example analysis are purely illustrative of the methodology and do not provide any indication of the comparative effectiveness of treatments for decision-making purposes, as the data set omits relevant direct and indirect data.

Using the example data set we show how to:

(a) Perform a meta-analysis of count statistics on the log hazard ratio scale;

(b) Reflect correlation in relative treatment effect estimates from multi-arm trials;

(c) Combine count and hazard ratio statistics in a single analysis on the log-hazard ratio scale; and

(d) Include a random effect in to the analysis, whilst preserving the correlation in relative treatment effects for multi-arm trials.

We also discuss how other possible presentations of survival data, not available from our motivating example, could be incorporated into an analysis on the log-hazard ratio scale.

All analyses described were conducted using WinBUGS [18]. The WinBUGS code for the fixed and random effects analyses are presented as an Appendix.

(a) Meta-analysing count statistics on the log hazard ratio scale

The count data are incorporated in the network meta-analysis model using a binomial likelihood:

(3)

where *r*_{s,k }is the cumulative count of subjects who have experienced an event in arm *k *of study *s*; *n*_{s,k }is the total number of subjects in arm *k *of study *s*; and *F*_{s,k }is the cumulative probability of a subject having experienced an event (or 'failure').

A log cumulative hazard for each trial arm ln(*H*_{s,k}) is then derived from *F*_{s,k}.

(4)

The log cumulative hazard estimates are then included in a treatment effect model with a linear regression structure. The log cumulative hazard is estimated as the sum of a study specific 'baseline' term *α*_{s }and a treatment effect coefficient *β*_{k}:

(5)

where *β*_{1 }= 0 for the reference treatment (placebo in our example) and *β*_{b }represents the treatment effect for the baseline treatment in study *s*. The fixed study level 'baseline' term is a nuisance parameter, included to ensure that the treatment effect estimates are informed by within trial differences between treatment arms and not by differences in baseline event rates across trials.

Under an assumption of proportional hazards, the *β*_{k }coefficient is equal to both the log cumulative hazard ratio and the log hazard ratio:

(6)

where *h*_{s,b }represents the hazard for the baseline treatment in study *s*. This identity allows us to combine the count statistics analysed on the log cumulative hazard scale with the hazard ratio data analysed on the log hazard scale. It also demonstrates that analysis of count statistics on the log hazard scale does not require a stronger assumption than proportional hazards.

(b) Reflecting correlations in relative treatment effects from multi-arm trials

Estimates of relative treatment effects from trials with more than two treatment arms will be correlated [10]. In our example the TORCH trial [16] is the only multi-arm trial reporting hazard ratio data. Estimated treatment effects from TORCH will be correlated, for example, the hazard ratio comparing SFC to placebo and the hazard ratio comparing salmeterol to placebo will be correlated due to their joint dependence on the time to event data in the placebo arm.

If a network meta-analysis is based on estimates of treatment effect in individual trial arms rather than estimates of relative treatment effect between arms ("contrast" statistics), this correlation will automatically be captured in the analysis [10]. This is the case when a network meta-analysis is based on count statistics. However, if the network meta-analysis is conducted based on estimates of relative treatment effect, this correlation between arms will not automatically be captured [10].

For multi-arm trials reporting hazard ratio statistics, this problem can be addressed by converting the log hazard ratios (contrast statistics) to log hazards (arm-specific statistics). Log hazards for individual trial arms are derived by nominally setting the log hazard for the baseline treatment *b*for the trial to zero. The mean log hazards for the other treatments are then equal to the log hazard ratios compared to baseline treatment.

The variance for a log hazard ratio is the sum of the variances for the individual log hazards. Standard errors of the log hazards for each trial arm can therefore be estimated by solving simultaneous equations based on the standard errors for the set of log-hazard ratios. For example:

(7)

Where *se*^{2}_{i,j }is the variance of the log hazard ratio comparing arm *i *to arm *j *and *se*_{i }is the standard error of the log hazard for arm *i*.

The standard errors of the log hazards for the other treatment arms are then estimated as:

(8)

Mean log hazards and associated standard errors for each treatment in TORCH [16] are reported in Table Table3.3. These are derived from the hazard ratio data presented in Table Table22.

In order to estimate standard errors of the log hazards for each treatment, we required estimates of the uncertainty associated with four treatment contrasts. In some cases this data may not be available and thus the methods presented in equations 7 and 8 may not be feasible. For example, hazard ratios and associated measures of uncertainty may only be available for each active treatment relative to a single common comparator (e.g. placebo) as is commonly reported in the published literature.

In this situation, we can approximate the standard error for the comparison between active treatments by assuming the standard error is proportional to . For example

(9)

(c) Combining count and hazard ratio statistics in a network meta-analysis

The log hazard ratio statistics from two arm trials comparing treatments *k* to *b* are incorporated in the network meta-analysis model using a normal likelihood:

(10)

where is the log hazard ratio estimate for study s comparing treatments *k *to *b *and is the corresponding variance.

The log hazard ratio estimates are then included in a treatment effect model with a linear regression structure, with the predicted log hazard ratio for a study *s *comparing treatments *k *and *b *equal to the difference between the two treatment coefficients:

(11)

where *β*_{1 }= 0 for the reference treatment (in our example placebo) and *β*_{b }represents the treatment effect for the baseline treatment in study *s*. As in equation 5, the *β*_{k }coefficient is equal to the log hazard ratio for treatment *k *compared to the reference treatment.

The log hazard statistics from a multi-arm trial are incorporated in the analysis using the following likelihood functions. For the baseline treatment, *b*:

(12)

For the other treatments:

(13)

where is the log hazard for treatment arm *k *from study *s *and is the associated variance.

The log hazard estimates are then included in a treatment effect model with a linear regression structure. The log hazard is estimated as the sum of a study specific 'baseline' term *α*_{s }and a treatment effect coefficient *β*_{k}:

(14)

where *β*_{1 }= 0 for the reference treatment (placebo in our example) and *β*_{b }represents the treatment effect for the baseline treatment in study *s*. The fixed study level 'baseline' term is a nuisance parameter, included to ensure that the treatment effect estimates are informed by within trial differences between treatment arms and not by differences in baseline event rates across trials. As the *β*_{k }coefficient is equal to the log hazard ratio for the cumulative count data, the log hazard ratio data and the log hazard data, they can be combined within a single analysis. Where an individual study reports both cumulative count and hazard ratio data, only one set of data should be included in the analysis to avoid double counting.

(d) Incorporating a random effect

In a random effects analysis of a network containing multi-arm trial contrast data, the correlation in the random effects must also be taken in to account. Again this is due to the joint dependence of the multiple contrast estimates on common trial arms.

This correlation is reflected in the model by separating the random effect deviation for each contrast in to the contributions to the random effect deviation of the two treatments that form the contrast. This is achieved by modifying the linear predictor component of the model for the cumulative count, log hazard ratio and log hazard data:

(15)

(16)

(17)

where *re*_{s,k }is the random effect deviation for arm *k *of study *s *and is assumed to be normally distributed with zero mean and variance *σ*^{2 }/2 where *σ*^{2 }is the random effect variance for a treatment contrast:

(18)

This approach assumes that *σ*^{2 }is the same for all treatments and consequently that the random effect variance will be the same for all treatment contrasts. The assumption of a common random effect variance across treatment contrasts implies that the covariance for any pair of treatment contrasts from the same study will equal half the treatment contrast random effect variance [10].

A vague prior for the study specific baseline *α*_{s}~*N*(0,10^{6}) is used to ensure estimates of treatment effect are informed by within trial differences between treatment arms, and not by differences in absolute response between trials. A vague prior is also used for the treatment effect coefficients with *β*_{k }~ *N*(0,10^{6}) and *β*_{1 }= 0 (representing placebo).

Each model was run for 40,000 burn-in simulations and 200,000 runs which were then thinned every 20^{th }simulation to reduce autocorrelation.

Two sets of initial values were used and convergence was assessed by examining caterpillar plots and Brooks Gelman-Rubin (BGR) statistics. The deviance information criteria (DIC) was used to compare the fit of the fixed and random effects models [19].

Results of the fixed and random effects models are presented as Table Table4.4. The results provide no evidence that the random effects model is preferred; the DIC for the random effects model is marginally higher (lower DIC values are preferred, with differences of 2-5 considered important [19,20]) and the high level of uncertainty around the random effects standard deviation estimate indicates that there is little information to inform the random effect parameter.

Figure Figure11 provides a presentation of the uncertainty in the analysis, showing the probability that each treatment takes each possible ranking (1st best, 2nd best, etc). For example, the figure tells us that there is a very low probability that fluticasone is the first or second most efficacious treatment in this analysis and that there is a 56% probability that is the third best and a 42% probability that it is the worst treatment.

Again it must be noted that this example data set is not an appropriate basis for answering the clinical question of which is the most efficacious treatment with respect to the mortality endpoint.

In this tutorial we have described how network meta-analysis can be conducted on the log-hazard ratio scale when the data on the survival endpoint available from the network of studies takes varying forms: count statistics and hazard ratio statistics from two and multi-arm trials.

In highlighting the importance of network meta-analysis as an extension to conventional pairwise meta-analysis Ades et al point out that "...to ignore indirect evidence either makes the unwarranted claim that it is irrelevant, or breaks the established precept of systematic review that synthesis should embrace all available evidence" [6]. A similar criticism can be levelled at meta-analyses that omit data points on the basis of the summary measure reported. This may result in intentional or unintentional selection bias. Furthermore, even if separate analyses are conducted for each summary measure available, this may have undesirable consequences by: producing conflicting results; producing coherent results that would not be supported by an 'all embracing' analysis; and unlinking some comparators from the network entirely.

Other authors have presented methods for combining arm-based mean change from baseline data with contrast mean difference data [21] and for combining mean time, median time and count data within a parametric survival model [22]. Further research is required in this area, for example in to methods for combining incidence rate data with count data, as these summary measures are often reported interchangeably across trials.

Other approaches to combining alternative reporting of time to event data have been discussed [8,23-26], recommended [27] and implemented [28,29]. These approaches involve the approximation of the log hazard ratio and its variance using available count statistics. The methods used in this tutorial avoid the use of approximations, instead utilising cumulative count statistics directly in the analysis.

A further common statistic reported for survival endpoints is arm-specific median survival time. This data can be incorporated using a similar approach as for count data. A binomial likelihood is used to incorporate the number of subjects experiencing an event (half the number of patients) and the number of patients analysed in to the model:

(19)

Note that *r*_{s,k }should be calculated outside the model to ensure it is a whole number. A log hazard for each trial arm ln (*h*_{s,k}) is then derived from the cumulative failure probability *F*_{s,k }and median survival time *T*_{s,k}:

(20)

The arm-specific log hazard estimates are then included in a treatment effect model that takes the same form as that for hazard ratio reporting multi-arm trials (see equation 14). It should be noted that analysis of median survival times requires the strong assumption of a constant hazard in each trial arm. The code for incorporating median survival time data is also provided in the Appendix.

Where multiple statistics are reported for a trial, hazard ratio statistics should be used in preference to median survival time and count data as hazard ratio statistics incorporate information about time to event and censoring, and analysis of hazard ratio statistics does not require assumptions stronger than proportional hazards. If both median survival time and count data are reported, a judgement is required to weigh up the relative merits of each presentation, as median survival time incorporates information about censoring but its analysis requires the stronger assumption of constant hazards.

Finally, this tutorial discusses methods for running analyses on the log hazard ratio scale. However, the question of which is the most appropriate scale for a given analysis (the scale on which transitivity and exchangeability are most likely to hold) - is an empirical question. Further work is required to develop methods for selecting the most appropriate scale for a given data set.

Meta-analysis of summary statistics continues to play an important role in medical research [30,31]. Where only summary statistics are available, the analyst performing pairwise or network meta-analysis may be faced with multiple summary statistics for a given endpoint. We present methods for meta-analysing different statistics summarising survival data on the hazard ratio scale. By incorporating all data presentations in a single analysis, we avoid the potential selection bias associated with conducting an analysis for a single statistic only, and the potential difficulties of interpretation, misleading results and loss of available treatment comparisons associated with conducting separate analyses for different summary measures. The methods described also allow use of the most informative statistic available from each study.

No specific funding was provided for the preparation of this manuscript. However, some of the methods presented were developed as part of an empirical project funded by GlaxoSmithKline.

BW conducted the statistical analysis and drafted the manuscript. NH conceived the study and developed the winBUGS code. DAS contributed to the design of the study and drafting of the manuscript and validated the statistical analysis. All authors reviewed and approved the manuscript.

*A) Fixed effects analysis*

model{

#Define Prior Distributions

#On tx effect mean

beta[1] < -0

for (tt in 2:nTx){

beta[tt]~dnorm(0,1.0E-6)

}

#On individual study baseline effect

for(ss in 1:nStudies){

alpha[ss] ~ dnorm(0,1.0E-6)

}

#Fit data

#For hazard ratio reporting studies

for(ii in 1:LnObs ){

Lmu[ii] < - alpha[Lstudy[ii]]*multi[ii] + beta[Ltx[ii]] - beta[Lbase[ii]]

Lprec[ii] < - 1/pow(Lse[ii],2)

Lmean[ii] ~ dnorm(Lmu[ii],Lprec[ii])

}

#For binary data reporting studies

for(ss in 1:BnObs){

logCumHaz[ss] < - alpha[Bstudy[ss]] + beta[Btx[ss]] - beta[Bbase[ss]]

cumFail[ss] < - 1-exp(-1*exp(logCumHaz[ss]))

Br[ss] ~ dbin(cumFail[ss], Bn[ss])

}

# Calculate HRs

for (hh in 1:nTx) {

hr[hh] < -exp(beta[hh])

}

# Ranking plot

for (ll in 1:nTx) {

for (mm in 1:nTx) {

rk[ll,mm] < - equals(ranked(beta[],mm),beta[ll])

}

}

}

# Data

# Data set descriptors

list(LnObs = 5, BnObs = 8, nTx = 4, nStudies = 5)

# Log hazard ratio and log hazard data

Lstudy[] Ltx[] Lbase[] Lmean[] Lse[] multi[]

1 1 1 0 0.066 1

1 2 1 0.055 0.063 1

1 3 1 -0.154 0.070 1

1 4 1 -0.209 0.072 1

2 2 1 -0.276 0.203 0

END

# Binary data

Bstudy[] Btx[] Bbase[] Br[] Bn[]

3 3 1 1 229

3 1 1 1 227

4 2 1 4 374

4 3 1 3 372

4 4 1 2 358

4 1 1 7 361

5 3 1 1 554

5 1 1 2 270

END

# Initial values

list(alpha = c(-0.50,-0.50,-0.50,-0.50,-0.50), beta = c(NA,-0.5,-0.5,-0.5))

list(alpha = c(0.50,0.50,0.50,0.50,0.50), beta = c(NA,0.5,0.5,0.5))

*B)Random effects analysis **(changes required to incorporate random effect in bold)*

model{

#Define Prior Distributions

#on random tx effect variance

**sd~dunif(0,5)**

**reTau < - 2/pow(sd,2)**

#On tx effect mean

beta[1] < -0

for (tt in 2:nTx){

beta[tt]~dnorm(0,1.0E-6)

}

#On individual study baseline effect

for(ss in 1:nStudies){

alpha[ss] ~ dnorm(0,1.0E-6)

}

#Define random effect

**for (ss in 1:nStudies){**

**for(tt in 1:nTx){**

**re[ss,tt]~dnorm(0,reTau)**

**}**

**}**

#Fit data

#For hazard ratio reporting studies

for(ii in 1:LnObs ){

Lmu[ii] < - alpha[Lstudy[ii]]*multi[ii] **+ re[Lstudy[ii],Ltx[ii]] -**

**re[Lstudy[ii],Lbase[ii]] **+ beta[Ltx[ii]] - beta[Lbase[ii]]

Lprec[ii] < - 1/pow(Lse[ii],2)

Lmean[ii] ~ dnorm(Lmu[ii],Lprec[ii])

}

#For binary data reporting studies

for(ss in 1:BnObs){

logCumHaz[ss] < - alpha[Bstudy[ss]] **+ re[Bstudy[ss],Btx[ss]] -**

**re[Bstudy[ss],Bbase[ss]] **+ beta[Btx[ss]] - beta[Bbase[ss]]

cumFail[ss] < - 1-exp(-1*exp(logCumHaz[ss]))

Br[ss] ~ dbin(cumFail[ss], Bn[ss])

}

# Calculate HRs

for (hh in 2:nTx) {

hr[hh] < -exp(beta[hh])

}

# Ranking plot

for (ll in 1:nTx) {

for (mm in 1:nTx) {

rk[ll,mm] < - equals(ranked(beta[],mm),beta[ll])

}

}

}

# Data as for fixed effects analysis

############################

# Initial values

list(alpha = c(-0.50,-0.50,-0.50,-0.50,-0.50), beta = c(NA,-0.5,-0.5,-0.5),**sd = 0.1**)

list(alpha = c(0.50,0.50,0.50,0.50,0.50), beta = c(NA,0.5,0.5,0.5),**sd = 1**)

*C) Additional code required for data reported as median survival times*

for (ii in 1:medianNObs ){

medianMu[ii] < - alpha[medianStudy[ii]] + beta[medianTx[ii]] -

beta[medianBase[ii]]

prob[ii] < - exp(-median[ii]*exp(medianMu[ii]))

medianR[ii] ~ dbin(prob[ii],medianN[ii])

}

The pre-publication history for this paper can be accessed here:

James Roger and Oliver Keene are acknowledged for their valuable comments on the methods presented in this paper.

- Caldwell DM, Ades AE, Higgins JPT. Simultaneous comparison of multiple treatments: combining direct and indirect evidence. BMJ. 2005;331:897–900. doi: 10.1136/bmj.331.7521.897. [PMC free article] [PubMed] [Cross Ref]
- Ades AE. A chain of evidence with mixed comparisons: model for multi-parameter synthesis and consistency of evidence. Stat Med. 2003;22(19):2995–3016. doi: 10.1002/sim.1566. [PubMed] [Cross Ref]
- Hasselblad V. Meta-analysis of multitreatment studies. Med Decis Making. 1998;18:37–43. doi: 10.1177/0272989X9801800110. [PubMed] [Cross Ref]
- Collett D. Modelling survival data in medical research. Second. 2003. Chapman & Hall/CRC.
- Kleinbaum DG, Klein M. Survival analysis: a self-learning text. Second. Springer, U.S; 2005.
- Ades A, Sculpher M, Sutton AJ, Cooper NJ, Abrams KR, Welton N, Lu G. Bayesian Methods for Evidence Synthesis in Cost-Effectiveness Analysis. Pharmacoeconomics. 2006;24:1–19. doi: 10.2165/00019053-200624010-00001. [PubMed] [Cross Ref]
- Chan AW, Hróbjartsson A, Haahr MT, Gøtzsche PC, Altman DG. Empirical evidence for selective reporting of outcomes in randomized trials: comparison of protocols to published articles. JAMA. 2004;291(20):2457–65. doi: 10.1001/jama.291.20.2457. [PubMed] [Cross Ref]
- Parmar MKB, Torri V, Stewart L. Extracting summary statistics to perform meta-analyses of the published literature for survival endpoints. Stat Med. 1998;17:2815–34. doi: 10.1002/(SICI)1097-0258(19981230)17:24<2815::AID-SIM110>3.0.CO;2-8. [PubMed] [Cross Ref]
- Whitehead A. Meta-Analysis of Controlled Clinical Trials. John Wiley and Sons: Chichester; 2002.
- Salanti G, Higgins JP, Ades AE, Ioannidis JP. Evaluation of networks of randomised trials. Stat Methods Med Res. 2008;17(3):279–301. doi: 10.1177/0962280207080643. [PubMed] [Cross Ref]
- Baker WL, Baker EL, Coleman C. Pharmacologic treatments for chronic obstructive pulmonary disease: a mixed-treatment comparison meta-analysis. Pharmacotherapy. 2009;29(8):891–905. doi: 10.1592/phco.29.8.891. [PubMed] [Cross Ref]
- Boyd G, Morice AH, Pounsford JC, Siebert M, Peslis N, Crawford C. An evaluation of salmeterol in the treatment of chronic obstructive pulmonary disease (COPD) Eur Respir J. 1997;10:815–21. [PubMed]
- Calverley P, Pauwels R, Vestbo J, Jones P, Pride N, Gulsvik A, Anderson J, Maden C. Trial of Inhaled Steroids and Long-Acting β2-Agonists (TRISTAN) Study Group. Combined salmeterol and fluticasone in the treatment of chronic obstructive pulmonary disease: a randomized controlled trial. Lancet. 2003;361:449–56. doi: 10.1016/S0140-6736(03)12459-2. [PubMed] [Cross Ref]
- Celli B, Halpin D, Hepburn R, Byrne N, Keating ET, Goldman M. Symptoms are an important outcome in chronic obstructive pulmonary disease clinical trials: results of a 3-month comparative study using the breathlessness, cough, and sputum scale (BCSS) Resp Med. 2003;97(suppl A):S35–43. [PubMed]
- Burge PS, Calverley PMA, Spencer S, Anderson JA, Maslen TK. on behalf of the ISOLDE Study Investigators. Randomised, double-blind, placebo-controlled study of fluticasone propionate in patients with moderate to severe chronic obstructive pulmonary disease: the ISOLDE trial. BMJ. 2000;320:1297–303. doi: 10.1136/bmj.320.7245.1297. [PMC free article] [PubMed] [Cross Ref]
- Calverley PMA, Anderson JA, Celli B, Ferguson GT, Jenkins C, Jones PW, Yates JC, Vestbo J. for the TORCH Investigators. Salmeterol and fluticasone propionate and survival in chronic obstructive pulmonary disease. N Engl J Med. 2007;356:775–89. doi: 10.1056/NEJMoa063070. [PubMed] [Cross Ref]
- Sin DD, Wu L, Anderson JA. Inhaled corticosteroids and mortality in chronic obstructive pulmonary disease. Thorax. 2005;60:992–7. doi: 10.1136/thx.2005.045385. [PMC free article] [PubMed] [Cross Ref]
- Lunn DJ, Thomas A, Best N, Spiegelhalter D. WinBUGS - a Bayesian modelling framework: concepts, structure, and extensibility. Stat Comput. 2000;10:325–337. doi: 10.1023/A:1008929526011. [Cross Ref]
- Spiegelhalter DJ, Best NJ, Carlin BP, Van der Linde A. Bayesian measures of model complexity and fit (with discussion) J R Stat Soc. 2002;64:1–34. doi: 10.1111/1467-9868.00353. [Cross Ref]
- The BUGS project. DIC: Deviance Information Criterion. http://www.mrc-bsu.cam.ac.uk/bugs/winbugs/dicpage.shtml#q9 Accessed 10/02/2010.
- Welton NJ. Mixed treatment comparisons: outcome measures. Lecture notes from Indirect and Multiple Comparisons course. Leicester University; 2008.
- Welton NJ, Cooper NJ, Ades AE, Lu G, Sutton AJ. Mixed treatment comparison with multiple outcomes reported inconsistently across trials: Evaluation of antivirals for treatment of influenza A and B. Stat Med. 2008;27:5620–39. doi: 10.1002/sim.3377. [PubMed] [Cross Ref]
- Ma J, Liu W, Hunter A, Zhang W. Performing meta-analysis with incomplete statistical information in clinical trials. BMC Med Res Methodol. 2008;8(1):56. doi: 10.1186/1471-2288-8-56. [PMC free article] [PubMed] [Cross Ref]
- Michiels S, Piedbois P, Burdett S, Syz N, Stewart L, Pignon JP. Meta-analysis when only the median survival times are known: a comparison with individual patient data results. Int J Technol Assess. 2005;21(1):119–25. [PubMed]
- Tierney JF, Stewart LA, Ghersi D, Burdett S, Sydes MR. Practical methods for incorporating summary time-to-event data into meta-analysis. Trials. 2007;8:16. doi: 10.1186/1745-6215-8-16. [PMC free article] [PubMed] [Cross Ref]
- Williamson PR, Tudur Smith C, Hutton JL, Marson AG. Aggregate data meta-analysis with time-to-event outcomes. Stat Med. 2002;21:3337–51. doi: 10.1002/sim.1303. [PubMed] [Cross Ref]
- (editors) Chapter 7: Selecting studies and collecting data. In: Higgins JPT, Green S, editor. Cochrane Handbook for Systematic Reviews of Interventions Version 5.0.0 (updated February 2008). The Cochrane Collaboration. 2008. http://www.cochrane-handbook.org
- Green JA, Kirwan JM, Tierney JF, Symonds P, Fresco L, Collingwood M, Williams CJ. Survival and recurrence after concomitant chemotherapy and radiotherapy for cancer of the uterine cervix: a systematic review and meta-analysis. Lancet. 2001;358(8):781–86. doi: 10.1016/S0140-6736(01)05965-7. [PubMed] [Cross Ref]
- Langendijk JA, Leemans CR, Buter J, Berkhof J, Slotman BJ. The Additional Value of Chemotherapy to Radiotherapy in Locally Advanced Nasopharyngeal Carcinoma: A Meta-Analysis of the Published Literature. J Clin Oncol. 2004;22(22):4604–12. doi: 10.1200/JCO.2004.10.074. [PubMed] [Cross Ref]
- Lyman GH, Kuderer NM. The strengths and limitations of meta-analyses based on aggregate data. BMC Med Res Methodol. 2005;5(1):14. doi: 10.1186/1471-2288-5-14. [PMC free article] [PubMed] [Cross Ref]
- Riley RD, Simmonds MC, Look MP. Evidence synthesis combining individual patient data and aggregate data: a systematic review identified current practice and possible methods. J Clin Epidemiol. 2007;60(5):431–9. doi: 10.1016/j.jclinepi.2006.09.009. [PubMed] [Cross Ref]

Articles from BMC Medical Research Methodology are provided here courtesy of **BioMed Central**

PubMed Central Canada is a service of the Canadian Institutes of Health Research (CIHR) working in partnership with the National Research Council's national science library in cooperation with the National Center for Biotechnology Information at the U.S. National Library of Medicine(NCBI/NLM). It includes content provided to the PubMed Central International archive by participating publishers. |