U.S. flag

An official website of the United States government

NCBI Bookshelf. A service of the National Library of Medicine, National Institutes of Health.

Trikalinos TA, Hoaglin DC, Small KM, et al. Evaluating Practices and Developing Tools for Comparative Effectiveness Reviews of Diagnostic Test Accuracy: Methods for the Joint Meta-Analysis of Multiple Tests [Internet]. Rockville (MD): Agency for Healthcare Research and Quality (US); 2013 Jan.

Cover of Evaluating Practices and Developing Tools for Comparative Effectiveness Reviews of Diagnostic Test Accuracy

Evaluating Practices and Developing Tools for Comparative Effectiveness Reviews of Diagnostic Test Accuracy: Methods for the Joint Meta-Analysis of Multiple Tests [Internet].

Show details

Estimation and Inference

Separate (one test at a time) and joint meta-analysis models using the normal approximation can be fit using (restricted) maximum likelihood.

Separate meta-analyses models that use the binomial distribution can be fit in the generalized linear mixed models framework using routines readily available in general statistical packages such as xtmelogit in Stata or lmer in R. However, the joint meta-analysis models using the multinomial likelihood cannot be fit in these general routines. The available generalized linear mixed model (GLMM) packages in R, Stata and SAS do not allow the user to specify the random effects distribution in (24), where the random effects pertain to sums of the probabilities in Table 5. Optimizing the likelihood for joint meta-analysis using the multinomial likelihood outside a GLMM package is nontrivial, because it involves calculating complicated integrals numerically. Thus we did not develop routines for fitting this model. Instead we fitted the model using Markov Chain Monte Carlo (MCMC) methods in the Bayesian framework, as described later in this section.

Maximum Likelihood Estimation (Model Using the Normal Approximation)

To fit the normal approximation model, optimize the log likelihood

LogL=12k=1K(log(|Wk|)-DkWkDk),
35

where Wk = (k + T)−1 and Dk=(η^k-Hξ^k-Ξ); |Wk| denotes the determinant of Wk. The parameters to be estimated are the summary effects H and Ξ and the elements of the between-study covariance matrix T. Alternatively, one can optimize the restricted likelihood, which was the approach we used in the applied example:

LogL*=12k=1K(log(|Wk|)-DkWkDk)+12log(|k=1KWk|)
36

As mentioned previously, it is typical meta-analytic practice to consider the elements of k known, but calculate them from the data. Appendix A provides formulas for these calculations. The matrix equations for the log likelihood remain the same for bivariate meta-analysis of one test and for the joint meta-analysis of two or more tests.

By optimizing (35) or (36) we obtain the (restricted) maximum likelihood estimators Ĥ, Ξ̂ and T̂. We also obtain the (2M+1 − 2)×(2M+1 − 2) estimated covariance matrix C = (cij) of the (Ĥ, Ξ̂′)′ as the inverse Hessian matrix.

Confidence Intervals

Confidence Intervals for the Summary Estimates Hm and Ξm

Confidence intervals for summary estimates are obtained in a similar manner for bivariate analyses of one test and for joint meta-analyses of two or more tests. Therefore, the formulas below are for M tests.

The 100(1−α)% simultaneous confidence interval (usually a 95% confidence interval) for Hm (the summary logit-TPR in test m) is given by:

(H^m-qαcmm,H^m+qαcmm),
37

where cmm is the variance of Ĥm, and qα is the square root of the 100(1−α) percentile of the chi-squared distribution with 2M+1 − 2 degrees of freedom. This simultaneous confidence interval is a special case of Scheffé’s F-projections for multiple comparisons; it controls type I error for the family of all possible linear combinations of the estimated parameters.47 The simultaneous confidence interval for Ξm (the summary logit-FPR in test m) is given by:

(Ξ^m-qαcm+2M-1,m+2M-1,Ξ^m+qαcm+2M-1,m+2M-1),
38

where cm+2M−1,m+2M−1 is the variance of Ξ̂m.

Confidence Intervals for Differences HiHj and ΞiΞj Between Summary Estimates of Two Tests

For two tests that have been applied to the same patients, one can either perform a meta-analysis for Test 1 and a separate one for Test 2, or a joint meta-analysis for the two tests. In either case, one can compare the diagnostic accuracy of the tests by calculating the difference between the logit-TPRs H1H2 and the difference between the logit-FPRs Ξ1Ξ2. The confidence intervals for such differences are calculated in different ways for separate versus joint meta-analyses of the two tests.

Confidence Intervals for Differences Based on Separate Meta-Analyses Per Test

Separate bivariate meta-analyses of the two tests ignore within-study correlations and treat the summary estimates of the two tests as independent. The resulting asymptotic confidence interval for the difference in logit TPR of tests i and j is

(H^i-H^j-zα/2var(H^i)+var(H^j),H^i-H^j+zα/2var(H^i)+var(H^j)),
39

where zα/2 is the upper α/2 percentile of the standard normal distribution. Because the confidence intervals in (39) ignore within-study correlations, their coverage differs from the nominal 100(1−α)%. Bonferroni’s inequality offers a simple adjustment to control the type I error. One substitutes zα/(2f) for zα/2 in (39), where f is the number of comparisons of interest. It may be reasonable to consider f = 2M + M(M−1), which equals the number of estimated mean logit-TPRs and mean logit-FPRs plus the total number of pairwise differences among the mean logit-TPRs and plus the total number of pairwise differences among the mean logit-FPRs. (The above considers all other modeled quantities, such as the logit-JTPR and the logit-JFPR, as nuisance parameters that are not of interest.)

Confidence Intervals for Differences Based on Joint Meta-Analyses of All Tests

For joint multivariate meta-analyses of all tests, differences and simultaneous confidence intervals are obtained as follows. For convenience, write β=(HΞ); that is, arrange the true summary logit-transformed quantities in a column vector. For a vector a = (a1, …, a2(2M−1))′ let L(a, β) = aβ be a linear combination of the true summaries, and L(a,β^)=a(H^Ξ^) its estimate. Then 100(1−α)% simultaneous confidence intervals for all possible linear combinations are given by

(L(a,β^)-qα(aCa),L(a,β^)+qα(aCa)).
40

In particular, to estimate differences between the summary logit-TPRs of tests i and j, set ai = 1, aj = −1, and all other elements of a to 0. Then L(a, β̂) = ĤiĤj, and the confidence interval in (40) becomes

(H^i-H^j-qαcii+cjj-2cij,H^i-H^j+qαcii+cjj-2cij).
41

In an analogous manner, to estimate differences between summary logit-FPRs for tests i and j, set ai+2M−1 = 1, aj+2M−1 = 1, and all other elements of a to 0, and proceed as in (40) to obtain

(Ξ^i-Ξ^j-qαci+2M-1,i+2M-1+cj+2M-1,j+2M-1-2ci+2M-1,j+2M-1,Ξ^i-Ξ^j+qαci+2M-1,i+2M-1+cj+2M-1,j+2M-1-2ci+2M-1,j+2M-1).
42

MCMC Estimation and Credible Intervals for Models Using Discrete Likelihoods

We fit models using the binomial and multinomial distributions at the within-study level with MCMC methods. To this end, and in addition to equations in the Models and Estimation chapter, we specified vague prior distributions for the following modeled parameters.

The true means were assigned independent vague normal priors:

(HΞ)~N(0,106·I6),

where I6 is the 6 × 6 identity matrix.

To assign priors for the covariance matrix T we use the factorization T = diag(τ)R diag(τ) where diag(τ) is the diagonal matrix whose diagonal elements are the square roots of the variances of the ηk and ξk and R is the correlation matrix corresponding to the covariances of the ηk and ξk. We assign independent uniform priors to the elements of τ, i.e., the standard deviations of the random effects):

τm~U(10-4,5)

The priors for R must guarantee that the matrix is positive definite with elements between −1 and 1. We follow Lu and Ades48 in factorizing R using the Cholesky decomposition for square symmetric matrices R = LL′, and in assigning specially constructed priors to the elements of the lower triangular matrix L (this is the spherical parameterization of Pinheiro and Bates49):

L=[100000cos(ϕ21)sin(ϕ21)0000cos(ϕ31)cos(ϕ32)·sin(ϕ31)sin(ϕ32)·sin(ϕ31)000cos(ϕ41)cos(ϕ42)·sin(ϕ41)cos(ϕ43)·sin(ϕ42)·sin(ϕ41)sin(ϕ43)sin(ϕ42)·sin(ϕ41)00cos(ϕ51)cos(ϕ52)·sin(ϕ51)cos(ϕ53)·sin(ϕ52)·sin(ϕ51)cos(ϕ54)sin(ϕ53)·sin(ϕ52)sin(ϕ41)sin(ϕ54)sin(ϕ53)·sin(ϕ52)sin(ϕ41)0cos(ϕ61)cos(ϕ62)·sin(ϕ61)cos(ϕ63)·sin(ϕ62)·sin(ϕ61)cos(ϕ61)sin(ϕ63)·sin(ϕ62)sin(ϕ64)cos(ϕ65)sin(ϕ64)·sin(ϕ63)sin(ϕ62)·sin(ϕ61)sin(ϕ65)sin(ϕ64)·sin(ϕ63)sin(ϕ62)·sin(ϕ61)]

Setting uniform independent priors for ϕ’s in the interval 0 to π = 3.14159... yields a prior for R in which all elements are between −1 and 1 and positive definiteness is guaranteed

ϕij~U(0,π).

See Lu and Ades for a short discussion on the density of the elements of R using the priors above.48 See Pinheiro and Bates for a discussion of additional parameterizations.49

95% Credible Intervals

With MCMC it is straighfrorward to obtain credible intervals for any quantity or any function of quantities explicitly, by simulation. In particular, we used 95% central credible intervals as the 2.5 and 97.5 percentile of the MCMC simulations.

Software and Computation

For the normal approximation models, the log likelihood in (35) and (36) for the unstructured variant of the T matrix can be optimized using routines such as mvmeta in Stata. We have developed our own Stata routines to optimize both the structured and the unstructured variant of T. (mvmeta uses a simple imputation of zero point estimates and large variances or covariances to simplify programming when handling studies with missing data; our routines do not need such imputations.) For convergence, starting values from fixed effect meta-analysis estimates appear to suffice. Note however, that the routine for the structured covariance matrix is not as robust: it failed to converge in the dataset used in this example (but does converge in other datasets). The optimization uses a modified Newton-Raphson algorithm. The routines are available from the authors upon request (see also www.cebm.brown.edu).

We ran MCMC analyses using JAGS version 3.1.0 through the R package rjags. We used three chains with a burn-in of at least 100,000 iterations and between 100,000 and 800,000 iterations for recording results. We monitored convergence with the Gelman-Rubin diagnostic for stochastic nodes corresponding to the meta-analysis means and the elements of their between-study covariance matrices. We declared convergence when the 97.5 percentile of the diagnostic was 1.10 or less for all monitored stochastic nodes, and provided that on visual inspection the traceplots of the MCMC chains were suggestive of good mixing.

Views

  • PubReader
  • Print View
  • Cite this Page
  • PDF version of this title (3.4M)

Recent Activity

Your browsing activity is empty.

Activity recording is turned off.

Turn recording back on

See more...