|Home | About | Journals | Submit | Contact Us | Français|
In diagnostic medicine, estimating the diagnostic accuracy of a group of raters or medical tests relative to the gold standard is often the primary goal. When a gold standard is absent, latent class models where the unknown gold standard test is treated as a latent variable are often used. However, these models have been criticized in the literature from both a conceptual and a robustness perspective. As an alternative, we propose an approach where we exploit an imperfect reference standard with unknown diagnostic accuracy and conduct sensitivity analysis by varying this accuracy over scientifically reasonable ranges. In this article, a latent class model with crossed random effects is proposed for estimating the diagnostic accuracy of regional obstetrics and gynaecological (OB/GYN) physicians in diagnosing endometriosis. To avoid the pitfalls of models without a gold standard, we exploit the diagnostic results of a group of OB/GYN physicians with an international reputation for the diagnosis of endometriosis. We construct an ordinal reference standard based on the discordance among these international experts and propose a mechanism for conducting sensitivity analysis relative to the unknown diagnostic accuracy among them. A Monte-Carlo EM algorithm is proposed for parameter estimation and a BIC-type model selection procedure is presented. Through simulations and data analysis we show that this new approach provides a useful alternative to traditional latent class modeling approaches used in this setting.
The motivation for this statistical research comes from the Physician Reliability Study (PRS) (Schliep et al., 2012) that investigated the diagnosis of endometriosis for various types of physicians by using different combinations of clinical information. Endometriosis is a gynecological medical condition in which cells from the lining of the uterus appear and flourish outside the uterine cavity, most commonly on the ovaries. The diagnosis of endometriosis can be complicated and there is no consensus in the field on what constitutes the gold standard (e.g., Brosens and Brosens, 2000). Of interest in the PRS is estimating the diagnostic accuracy of a group of regional obstetrics and gynecological (OB/GYN) physicians (R-OB/GYNs) in terms of their endometriosis diagnosis. The R-OB/GYNs were presented with digital images of the uterus and adnexal structure of participants that were taken during laparoscopies and were asked to make a diagnosis of endometriosis. Since these R-OB/GYNs see patients daily and are affiliated with the same medical center, an assessment of their diagnostic accuracy can help the medical center in designing specific training programs to improve their diagnosis.
Diagnostic accuracy of raters or medical tests are of considerable interest in many public health and biomedical fields (Zou, McClish, and Obuchowski, 2002; Pepe, 2003; Hui and Walter, 1980). The estimation of diagnostic accuracy is straightforward when the true disease status is known. In many cases such as for endometriosis, however, a gold standard is not available. Methods have been proposed to estimate diagnostic accuracy without a gold standard using latent class models for which the true disease status is considered to be a latent variable (Hui and Walter, 1980; Hui and Zhou, 1998). Qu, Tan, and Kutner (1996) proposed a random-effects latent class model that introduces conditional dependence between tests with normally distributed random effects. Albert et al. (2001) proposed a latent class model with a finite mixture structure to account for dependence between tests. More recently, it has been shown that latent class models for estimating diagnostic accuracy may be problematic in many practical situations (Albert and Dodd, 2004; Pepe and Janes, 2006). Specifically, Albert and Dodd (2004) showed that with a small number of binary tests, estimates of diagnostic accuracy are biased under a misspecified dependence structure; yet in many situations it is nearly impossible to distinguish between models with different dependence structures from observed data.
Between the two extremes of no gold standard on anyone and a gold standard on all individuals, there are situations where a gold standard does not exist but some imperfect information is available. When there is no gold standard, the best available reference tests can be employed to help the estimation of diagnostic accuracy of new tests. Those best available reference tests may themselves be subject to small error, and therefore are called imperfect reference standard. In the motivating PRS example, there are diagnostic results on the same subjects from a group of international expert (IE) OB/GYN physicians. These IEs all had directed specialized training in laparoscopic surgery, accrued extensive clinical and research experience in diagnosing and treating endometriosis, and have international reputations in the field. A scatterplot of the correlation between the IE and R-OB/GYN ratings is included in Web Appendix A of the web-based supplementary materials. In this paper, we propose new methodology to estimate the R-OB/GYNs’ diagnostic accuracy by exploiting the IEs’ diagnostic results in the PRS.
Valenstein (1990), Begg (1987), and Qu and Hadgu (1998) have discussed the bias in estimating diagnostic accuracy using an imperfect reference standard. Using both analytical and simulation techniques, Albert (2009) showed that, with the aid of an imperfect reference standard with high sensitivity and specificity, inferences on diagnostic accuracy are robust to misspecification of the conditional dependence between tests. However, this approach assumes that the diagnostic accuracy of the imperfect reference standard is known or can be estimated from other data sources. In some cases, no gold standard exists and it is impossible to obtain estimates of the diagnostic accuracy of the imperfect reference standard relative to the gold standard from other studies. In this situation, we show how multiple expert raters or more definitive tests can be used along with a sensitivity analysis to estimate the diagnostic accuracy of other raters or tests.
Our proposed approach for estimating the average diagnostic accuracy among R-OB/GYNs makes use of the latent class model as in the aforementioned literature for models without a gold standard. Since each physician examines each subject in the PRS, we develop an approach where R-OB/GYNs are random and crossed with subjects. To exploit the IEs’ diagnostic results, we construct an ordinal composite imperfect reference standard from the individual diagnostic results of the IEs. We first assume that we know the diagnostic accuracy of the imperfect reference standard and proceed with estimating the diagnostic accuracy of the R-OB/GYNs. In this step, the value of the diagnostic accuracy of the imperfect reference standard is chosen such that the corresponding posterior probability of the latent disease status given the observed ordinal reference standard is reasonable. We then vary this choice widely within a scientifically sensible range of the posterior probabilities in order to assess the robustness of the estimated diagnostic accuracy of the R-OB/GYNs. To handle the computational challenge arising from the crossed random effects, we develop a Monte-Carlo EM algorithm for parameter estimation. We investigate the robustness of the proposed latent class model with respect to the misspecification of the dependence structure between tests. We show that, without the imperfect reference standard (i.e., IE reviewers), estimates of diagnostic parameters are biased under a misspecified dependence structure. Moreover, the model selection criterion (Ibrahim, Zhu, and Tang, 2008) has difficulty distinguishing the various competing models. However, with the aid of the imperfect ordinal standard, (i) estimates of diagnostic accuracy are nearly unbiased, even when the dependence structure between the R-OB/GYN tests is misspecified or when the assumed diagnostic accuracy of the imperfect reference standard deviates from the truth in a reasonable way and (ii), we are able to distinguish between competing models for the dependence between R-OB/GYNs.
In section 2, we propose a latent class modeling approach for estimating the diagnostic accuracy of the R-OB/GYNs that exploits the IEs by constructing an ordinal reference standard. In Section 3, we investigate the bias from misspecifying the random effects structure with and without the use of the imperfect reference standard. In Section 4, we apply the proposed model to data from the PRS in the diagnosis of endometriosis. A discussion follows in Section 5.
Let Yij denote the binary diagnostic result of endometriosis (Yij = 1(0) for having (not having) endometriosis) for the ith subject from the jth R-OB/GYN, i = 1, 2, · · ·, I and j = 1, 2, · · ·, J. In the PRS, we consider 79 subjects who had complete diagnoses from 8 R-OB/GYNs and 4 IEs; thus, I = 79 and J = 8. We denote Di as the true disease status of the ith subject. Due to the lack of a gold or imperfect reference standard for endometriosis, we model Di as a binary latent variable. In the PRS, the eight R-OB/GYNs are chosen from a group of regional physicians, and interest is on estimating the average sensitivity and specificity across the population of these physicians, rather than the eight physician-specific sensitivities and specificities themselves. Therefore, we consider the following model with two crossed random effects:
where bi is the subject-specific random effect with probability density distribution (p.d.f.) g1(x), cj is the rater-specific random effect with p.d.f. g2(x), and the three unobserved latent variables Di, bi, and cj are assumed to be independent of each other. Let πdi = P(Di = di) and note π1 is the prevalence of endometriosis in the population.
Contrast to the latent class models with a Gaussian random effect model in Qu, Tan and Kutner (1996) and Albert (2001), where the rater sensitivity and specificity were treated as fixed effects, (1) can be used to estimate the average sensitivity and specificity across the population of regional physicians as follows: and under the normality assumption of bi and cj. These expressions are obtained by integrating out both subject- and rater-level random effects. For notational brevity, we will simply call them sensitivity and specificity in the remainder of the paper. In addition to the normality assumption on the random effects, we consider the case where g1(x) and g2(x) are two-group mixture normal distribution with p.d.f.:
where λm is the probability in the first group, 0 ≤ λm ≤ 1, and (x; μ, ν2) is the p.d.f. of a normal distribution with mean μ and variance ν2 (Albert et al., 2001). For model identification, we assume λmμ1m + (1 − λm)μ2m = 0, and μ1m < μ2m. Similar to the Gaussian random effects model with crossed random effects, the model-based estimates of sensitivity and specificity can be obtained by marginalizing over the two-group mixture models.
The likelihood of (1) is complicated by the two crossed random effects bi and cj. Let Yi = (Yi1, Yi2, · · ·, YiJ )′ be the J dichotomous rating results on the ith subject. Also, let , D = (D1, D2, · · ·, DI) and θ be the vector of unknown parameters βdi, σdi, τdi, πdi and the unknown parameters in gi(x). Then, the likelihood of the proposed model (1) is given by
The likelihood (3) involves high-dimensional integration and summation, which is difficult to evaluate by numerical approximation. As a consequence, a Monte-Carlo EM algorithm is presented in Web Appendix B of the web-based supplementary materials to obtain the maximum-likelihood estimation of (3).
In addition to the eight R-OB/GYNs, four IEs provided diagnoses of endometriosis for each subject in the PRS. These IEs are well-known OB/GYN physicians in the field and are expected to have better diagnostic accuracy than other physicians. Although a gold standard does not exist for endometriosis, the diagnostic results from these IEs can be used to construct an imperfect reference standard to improve the estimation of average diagnostic accuracy of the R-OB/GYNs. Specifically, an ordinal imperfect reference standard can be constructed based on the multiple IE binary ratings. Generally, suppose there are L binary IE ratings for the ith subject, l = 1, · · ·, L. We propose to use the sum of those binary ratings as the imperfect reference standard , where Ti takes values 0, 1, · · ·, L. In the PRS, L is equal to 4. Letting T = (T1, T2, · · ·, TI), the likelihood of the observed rating results Y and imperfect reference standard T is
where characterizes the diagnostic accuracy of the imperfect reference standard relative to the true disease status. In this section, we will assume is known. However, for endometriosis there is no established gold standard available for estimating these quantities. When no gold standard exists, we propose a methodological approach for conducting a sensitivity analysis in section 2.3.
To incorporate the information of the imperfect reference standard, we consider the model
where we make the assumption that the observed ratings do not depend on the imperfect reference standard if the true disease status is available. This is a natural assumption which is usually true in practice. The assumption can be relaxed by allowing β to depend on Ti (i.e., replacing βdi with βdiTi in (5)) or by introducing dependence between the Yij’s and Ti’s through a shared random effect (see Web Appendix F). The corresponding likelihood is given by
A Monte-Carlo EM algorithm is used to obtain maximum likelihood estimates of βdi, σdi and τdi, di = 0 or 1, by maximizing (6).
There are special cases of the proposed methodology. When , it reduces to the case in which the true disease status is observed. For diagnosing endometriosis in the PRS, it corresponds to the scenario where all four IEs report positive results if the subject has endometriosis and they all report negative results if the subject does not (i.e., IEs have perfect classifications). When for ti = 0, 1, · · ·, L, the imperfect reference standard (IEs) adds no additional information, and the approach reduces to a latent class model without a gold standard for the R-OB/GYNs alone.
As stated in Section 2.2, we construct the imperfect reference standard T by using the diagnostic results from the four IEs in the PRS. In this application, we do not know the diagnostic accuracy of the IEs and, consequently, do not know the diagnostic accuracy of the constructed imperfect reference standard. Hence, we discuss how we conduct sensitivity analysis for diagnostic accuracy estimation of the R-OB/GYNs by varying the diagnostic accuracy of the imperfect reference standard over a wide range of reasonable values. In particular, we assume the following polychotomous logit model,
and a symmetrically defined . Note here that . Denoting , then by Bayes rule,
which characterizes the posterior probability of disease given the observed imperfect reference standard. We focus on the sum of the IE ratings, rather than their individual values, since this provides a simple and intuitive approach for sensitivity analysis that does not require the specification of the dependence structure between the multiple IE ratings. With regard to choosing parameter of (7), we suggest that parameters be chosen so that and are both close to zero (i.e., the probability that a woman is truly negative (positive) for endometriosis but all IEs independently diagnose her as positive (negative) is reasonably assumed to be zero). Figure 1 shows posterior probability of having disease for different values of γ2 in (8) where γ0 = −4.5, γ1 = 0.1 and π1 = 0.70. The impact of the change of γ2 on the estimation of the diagnostic accuracy of the regional physicians is examined in both the simulation and application sections.
In the simulation and discussion sections as well as the web-based supplementary materials, we show that the proposed approach for exploiting the IEs and conducting sensitivity analysis is robust to (i) the assumed dependence structure among the R-OB/GYN ratings and (ii) the exchangeability assumption implicit in using the sum of the IE ratings as an imperfect reference standard.
The simulation datasets were generated from the latent class models with random effects that follow mixture normal distribution (“true model”) and were fit to models with either normal or mixture normal random effects (“working model”). All simulations were conducted with 500 replications, each with 100 subjects (I = 100). The average sensitivity and specificity and the disease prevalence over 500 replications and their standard deviations are shown with five (J = 5) or ten (J = 10) raters in Table 1.
Table 1(A) presents the results when there is no gold or imperfect reference standard. The results show that, with no help of a gold or imperfect reference standard, there is sizable bias in estimation when the random effects (both between subjects and between raters) in the true model follow mixture normal distributions and in the working model follow normal distributions. The averages of estimated mean sensitivity and specificity are both 0.81 for 5 raters, and 0.82 and 0.83 for 10 raters, respectively, compared to the true sensitivity of 0.88 and true specificity of 0.87. Moreover, we are unable to distinguish between the true and the misspecified models by using model selection criterion ICH(0),Q (Ibrahim, Zhu, and Tang, 2008); the percentages that ICH(0),Q selected the true random effects structure are just sightly above 50% (55% for 5 raters; 52% for 10 raters). These results are consistent with Albert and Dodd (2004) who showed that for fixed rater-specific diagnostic accuracy, estimation is sensitive to misspecification of the random effects distribution, yet it is difficult to distinguish between random effects distributions using the observed data.
In the case when an ordinal imperfect reference standard is used and the diagnostic accuracy of the imperfect reference standard is known, we assume that the imperfect reference standard Ti is ordinal taking values 0, 1, · · ·, 4, and that the true disease status Di remained binary taking 0 (no disease) or 1 (disease). The imperfect reference standard was generated under (7) with γ0 = −4.5, γ1 = 0.1, and γ2 = 0.2. The corresponding posterior disease probabilities given ti = 0, 1, 2, 3, 4, when π1 = 0.70, were 0.02, 0.28, 0.70, 0.93, and 1.00, respectively. Table 1(B) shows that, with the aid of the imperfect reference standard that has known diagnostic accuracy, the proposed latent class model is robust in estimating sensitivity, specificity, and disease prevalence when the random effect distributions are mis-specified. More specifically, when the true model random effects distributions are mixture normal distributions and the working model assumes normal random effects distributions, the estimates are nearly unbiased with the average estimated sensitivities at 0.87 for 5 raters and 0.88 for 10 raters, and the average estimated specificities both at 0.87. In addition, we are able to distinguish between the true and misspecified models by using model selection criterion ICH(0),Q; the percentages that ICH(0),Q selected the true random effects structure are about 90% or above in both the 5- and 10-rater cases. With the imperfect reference standard, there is also no observed efficient loss under misspecified random effect distributions relative to the correct distributions.
The results in Table 1(B) show how incorporating an imperfect reference standard improves the robustness of the estimation, assuming that we know the diagnostic accuracy of it. However, in most practice situations, we do not know the diagnostic accuracy of the imperfect reference standard. In the PRS, we construct the imperfect reference standard from four IEs, but have no information about the diagnostic accuracy of this imperfect reference standard. Thus, it is of interest to investigate the robustness of the proposed latent class model to reasonable misspecification of the diagnostic accuracy of the imperfect reference standard. For this reason, we repeated the simulation study in Table 1(B), but with a misspecified diagnostic accuracy (γ0 = −4.5, γ1 = 0.1, and γ2 = 0.1 in (7)) for the imperfect reference standard. The resulting posterior disease probabilities given ti = 0, 1, 2, 3, 4 were 0.02, 0.46, 0.70, 0.86, and 1.00, respectively. Table 1(C) shows that even with the misspecified diagnostic accuracy of the imperfect reference standard, the estimates of sensitivity and specificity are still nearly unbiased. Further, we are still able to distinguish between the true and misspecified models by using model selection criterion ICH(0),Q: the percentages ICH(0),Q selected the true random effects structure are around 90%.
Although the robustness of the proposed latent class model in estimating diagnostic accuracy is shown in Table 1 with the imperfect reference standard, we also investigate the sensitivity of the proposed model to different values of the diagnostic accuracy of the imperfect reference standard. Specifically, we estimate sensitivity, specificity, and prevalence for various values of the diagnostic accuracy of the imperfect reference standard by varying γ2 in the parameterization given in (7). The simulations were also conducted with 500 replications for 100 individuals (I = 100). Table 2 shows results of the average estimated sensitivity and specificity in the population with five (J = 5) or ten (J = 10) raters under the scenarios that the true imperfect reference standard was generated from (7) with γ0 = −4.5, γ1 = 0.1, and γ2 = 0.2 and the misspecified imperfect reference standards with γ0 = −4.5, γ1 = 0.1, but γ2 = −0.1, 0, 0.1, 0.15, 0.25, and 0.3 were used. Severe departure was not considered here, because it is not likely for investigators to use imperfect reference standards that have unreasonable poor quality. The simulation datasets were generated from the latent class models with random effects that follow mixture normal distribution and were fit to the models with normal random effects, but with the aid of the imperfect reference standard with an incorrectly specified diagnostic accuracy. As shown in Table 2, the robustness of the latent class model remains in all scenarios when the diagnostic accuracy of the imperfect reference standard is misspecified. Thus, the simulations demonstrate that exploiting the ordinal imperfect reference standard (e.g., a group of expert raters) and performing a sensitivity analysis provides a more robust solution than latent class models without a gold or imperfect reference standard for estimating diagnostic accuracy. Further, it is simpler to distinguish between competing models for the dependence between the experimental raters (e.g. ROB/GYNs) when we exploit the ordinal imperfect reference standard.
One alternative of using the IE ratings is to incorporate both R-OB/GYNs and IEs into the latent class approach and fit a conventional model without a gold standard. We show the inherent problem of this alternative approach by considering the following model that was modified from Equation (1):
Equation (9) is (1) with the addition of Ej as an indicator variable for IEs where both types of ratings are included in the latent class model. To investigate the performance of (9) in estimating the sensitivity and specificity of the R-OB/GYNs, we set α1 = −α0 = 0.5 so that the IEs have higher sensitivity and specificity than the R-OB/GYNs. The simulation study in Table 1(A) was repeated under (9). Table S.1 in Web Appendix C of the web-based supplementary materials shows the results from the simulation study. In summary, the estimates of sensitivity and specificity of the R-OB/GYNs are biased when the dependence structure is misspecified. Further, the results shows that it is difficult to distinguish between the different models with this approach. In conclusion, the lack of robustness and difficulty in choosing between models makes the latent class approach on both R-OB/GYNs and IEs less attractive for incorporating the imperfect reference standard (IEs) than the proposed approach with a sensitivity analysis.
The proposed methodology is applied to data from the PRS in the diagnosis of endometriosis. In this study, eight R-OB/GYNs diagnosed 79 subjects for endometriosis based on digital images from laparoscopies. Our interest is to obtain average sensitivity and specificity of the R-OB/GYNs in the diagnosis of endometriosis.
We estimated the diagnostic accuracy of R-OB/GYNs using the proposed latent class model with crossed random effects, under two random effects distributions for the between-subject and between-rater variation: normal distribution and mixture normal distribution. Table 3 shows the overall estimates of prevalence, sensitivity, and specificity for the diagnosis of endometriosis, as well as the ICH(0),Q values of the fitted models. Bootstrap standard errors based 500 replications are also presented under each model. When no imperfect reference standard information is incorporated (Table 3(A)), estimates of diagnostic accuracy are close across models with the different random effects structures. For example, the sensitivity is 0.96 under the normal random effects distribution and is 0.93 under the mixture normal random effects distribution. The ICH(0),Q values are also close for the two random effects specifications, suggesting that it is difficult to distinguish between the two models.
Due to the lack of a gold standard diagnosing endometriosis in the PRS, the ratings from the four IEs are used to construct an imperfect reference standard as stated in Section 2.2. This imperfect reference standard is ordinal, with Ti = 0 representing all four IEs’ diagnosing no disease for the ith subject and Ti = 4 representing all four IEs diagnosing the ith subject having endometriosis. Since the diagnostic accuracy of the imperfect reference standard is unknown, we first assumed that the diagnostic accuracy of the IEs was governed by setting γ0 = −4.5, γ1 = 0.1, and γ2 = 0.2 in (7). The resulting posterior disease probabilities given ti = 0, 1, 2, 3, 4 are 0.02, 0.46, 0.70, 0.86, and 1.00, respectively. The estimation results are reported in Table 3(B). Consistent with the simulation studies, the estimates of sensitivity and specificity from the models with normal and mixture normal random effects structures are very close. The model selection criterion ICH(0),Q can distinguish the models and identifies the mixture normal model as the better model.
We further examined the robustness of the estimates of the proposed latent class model with respect to the different diagnostic accuracy of the imperfect reference standard information using (7). Figure 2 shows estimates of sensitivity and specificity, along with disease prevalence, for the latent class model that has normal random effects and incorporates the different imperfect reference standard information. The diagnostic accuracy of the imperfect reference standard was generated from (7) with γ0 = −4.5, γ1 = 0.1 and γ2 changing from 0 to 0.5 by 0.01. Varying values of γ2 corresponds to the probabilities as shown in Figure 1. The overall estimates of sensitivity and specificity for the R-OB/GYN physicians are nearly identical across γ2. Thus, estimation of the average sensitivity and specificity forthe R-OB/GYNs are insensitive over a wide range of scientifically reasonable values of the diagnostic accuracy of the imperfect reference standard (derived from the IEs’ diagnoses).
Estimating diagnostic accuracy without a gold standard is a challenging problem that has received substantial recent attention. Most of these methods involve latent class models where the true disease state is considered latent. However, these approaches have received criticism from a conceptual (Pepe and Janes, 2007) and robustness (Albert and Dodd, 2004) prospective.
In practical situations, we are still left with the problem of whether it is possible to estimate diagnostic accuracy when there is no gold standard. This was the motivation in the PRS where an important goal was estimating the sensitivity and specificity of diagnosing endometriosis from viewing intra-uterus digital pictures taken during laparoscopic surgeries for the R-OB/GYNs at a Utah site. Fortunately, we have additional information on a set of four IEs which we can exploit to estimate diagnostic accuracy of endometriosis among typical obstetrics and gynecology physicians. The methodology assumes that we know the diagnostic accuracy of the IEs, and we perform sensitivity analysis by varying this diagnostic accuracy in scientifically sensible ways. For this application, we are confident that when all raters agree there is only a negligible probability of misclassification. Hence, we primarily focus on varying the diagnostic accuracy for situations where the IEs differ. Through simulations and data analysis we show that the diagnostic accuracy estimation is remarkably robust to reasonable misspecification of assumed diagnostic accuracy of the IEs as well as to the distributional assumptions on the random effects for modeling the dependence among ROB/GYNs.
The robustness for estimating the diagnostic accuracy of the R-OB/GYNs to the assumed accuracy for the IEs is in part due to the low frequency of discordance in IEs’ ratings, and the reasonable assumption that there is no misclassification when all IEs agree. We expect that estimation will be more sensitive to variation in the diagnostic accuracy of the imperfect standard when there is less consensus among the IEs. Our approach will still be useful in this case since we can report ranges of diagnostic accuracy estimates of R-OB/GYNs corresponding to ranges in the assumed diagnostic accuracy of the IEs.
This work has important implications for diagnostic accuracy studies, suggesting that exploiting expert raters or multiple high quality tests may provide a good approach for estimating diagnostic accuracy of other raters or tests. Especially, we showed that these types of experts raters or high quality tests can greatly improve the robustness of the estimation with respect to the misspecification of random effect distributions in the latent class models. Thus, this approach provides a robust alternative to traditional latent class models for estimating diagnostic accuracy without a gold standard.
The use of IE data is one of the key aspects in the application of the proposed methodology. First, the fact that the inference are not sensitive to parameters of the polychotomous logit model is in part due to and for all parameters considered. This is sensible since it would seem very unlikely that one would have a positive (negative) gold standard when all the IEs were negative (positive). One would need a large enough group of IEs to have this confidence, with this number being application dependent. However, for a general rule, we recommend a minimum of four expert ratings. Second, we conducted simulation studies to investigate the performance of the proposed method when IEs only examine a subset of the patients. Estimates of the sensitivity and specificity of the R-OB/GYNs are nearly unbiased under a misspecified dependence structure when the IEs examine 80% and 50% of the patients. When the proportion of examined patients decreased to 20%, the estimates have substantially more bias. Third, we conducted simulation studies to investigate the performance of the proposed method when the patients are not examined by all IEs. The simulations indicate that the estimates of the sensitivity and specificity of the R-OB/GYNs are robust to dependence misspecification. Therefore, we suggest that, when the patients are not examined by all IEs, the proposed method can still function very well with the appropriate imputation for the missingness. Please see Web Appendix D in the web-based supplementary materials for simulation results.
The use of the proposed polychotomous logit model for the imperfect reference standard provides nearly unbiased estimation of the diagnostic accuracy and disease prevalence, regardless of whether or not the IEs are exchangeable (exhangeablilty of the four IEs means, for any combination of e1, e2, e3, and e4, , where di = 0 or 1). Also, in (5), we assume independence between the random effect bi and the imperfect reference standard Ti. In the circumstance where these assumptions are violated, the proposed methodology is still able to provide robust estimates for the diagnostic accuracy of R-OB/GYNs and disease prevalence (please refer to Web Appendices E and F in the web-based supplementary materials for more detailed discussions of these two points).
The focus of this work was on estimating diagnostic accuracy measures such as sensitivity, specificity, and prevalence. Other work in the area of diagnostic testing have criticized these measures in favor of positive and negative predictive value (Moons et al., 1997). These alternative measures can easily be estimated from estimated sensitivity, specificity, and prevalence discussed in this paper. The analysis of the PRS data with the new methodology show that the average R-OB/GYN’s sensitivity is high (≈0.93) with the average specificity being rather low (≈0.77). This is important new information since it suggests that R-OB/GYNs are overly diagnosing endometriosis in regular clinical practice.
We thank Editor, Associate Editor and the anonymous reviewer for their thoughtful and constructive comments, which have led to an improved article. This research was supported by the Intramural Research Program of the National Institutes of Health, Eunice Kennedy Shriver National Institute of Child Health and Human Development. We thank the Center for Information Technology, the National Institutes of Health, for providing access to the high performance computational capabilities of the Biowulf Linux cluster.