 Original research
 Open Access
 Published:
Computeraided diagnosis of renal obstruction: utility of loglinear modeling versus standard ROC and kappa analysis
EJNMMI Research volume 1, Article number: 5 (2011)
Abstract
Background
The accuracy of computeraided diagnosis (CAD) software is best evaluated by comparison to a gold standard which represents the true status of disease. In many settings, however, knowledge of the true status of disease is not possible and accuracy is evaluated against the interpretations of an expert panel. Common statistical approaches to evaluate accuracy include receiver operating characteristic (ROC) and kappa analysis but both of these methods have significant limitations and cannot answer the question of equivalence: Is the CAD performance equivalent to that of an expert? The goal of this study is to show the strength of loglinear analysis over standard ROC and kappa statistics in evaluating the accuracy of computeraided diagnosis of renal obstruction compared to the diagnosis provided by expert readers.
Methods
Loglinear modeling was utilized to analyze a previously published database that used ROC and kappa statistics to compare diuresis renography scan interpretations (nonobstructed, equivocal, or obstructed) generated by a renal expert system (RENEX) in 185 kidneys (95 patients) with the independent and consensus scan interpretations of three experts who were blinded to clinical information and prospectively and independently graded each kidney as obstructed, equivocal, or nonobstructed.
Results
Loglinear modeling showed that RENEX and the expert consensus had beyondchance agreement in both nonobstructed and obstructed readings (both p < 0.0001). Moreover, pairwise agreement between experts and pairwise agreement between each expert and RENEX were not significantly different (p = 0.41, 0.95, 0.81 for the nonobstructed, equivocal, and obstructed categories, respectively). Similarly, the threeway agreement of the three experts and threeway agreement of two experts and RENEX was not significantly different for nonobstructed (p = 0.79) and obstructed (p = 0.49) categories.
Conclusion
Loglinear modeling showed that RENEX was equivalent to any expert in rating kidneys, particularly in the obstructed and nonobstructed categories. This conclusion, which could not be derived from the original ROC and kappa analysis, emphasizes and illustrates the role and importance of loglinear modeling in the absence of a gold standard. The loglinear analysis also provides additional evidence that RENEX has the potential to assist in the interpretation of diuresis renography studies.
Background
The increase in the number and complexity of diagnostic studies, subjectivity in image interpretation, physician time constraints, and high error rates have stimulated the development of computeraided diagnostic (CAD) tools to help nuclear medicine physicians and radiologists interpret studies at faster rate and with higher accuracy [1–5]. The introduction of new decision support tools, however, has raised a critical question: What is the best way to evaluate the performance of these new diagnostic tools? Ideally, the accuracy of a new diagnostic tool should be measured against a gold standard which represents the true status of the disease, i.e., disease present or disease absent. Unfortunately, in many circumstances, a gold standard is not available due to the fact that the gold standard is unacceptably invasive, prohibitively expensive, or simply nonexistent [6–9]. A common approach to this problem is to compare the diagnosis of a new CAD tool with those of expert readers. However, since experts do not always agree, the CAD diagnosis is often compared to a consensus diagnosis of experts. The best standard, however, is not how well the new diagnostic tool performs compared to a consensus interpretation of experts but to determine if its performance is equivalent to the diagnostic performance of any expert. When the performance of the new CAD tool is equivalent to any expert, the new computeraided tool can be considered to be sufficient to assist in scan interpretation.
Receiver operating characteristic (ROC) and kappa methodologies have been and continue to be popular methods to assess the reliability of computeraided diagnosis tools [7, 8, 10–12], but both of these common approaches have significant limitations. ROC analysis requires an independent measure of truth, and it requires the measure to be dichotomized (e.g., disease present or absent). In practice, image interpretation may not be definitive and the report may be qualified by terms like "indeterminate," "possible" or "questionable." In contrast, kappa statistics [13, 14] measure the degree of agreement beyond that expected by chance alone. For example, when there are three categories such as "normal", "equivocal", and "obstruction" in rating of kidney images, the kappa statistic provides a number between 0 and 1, indicating the strength of agreement beyond chance across all categories. A major disadvantage of kappa is that, by construction, it provides an overall summary of beyondchance agreement across all categories and there is a loss of information [14] in summarizing the data and it does not specifically address how two raters agree on a certain category. Moreover, kappatype statistics [9, 15, 16] can be heavily influenced by the distribution of disease in the population as well as by differences or similarities among raters [17]. It is also difficult to interpret the magnitude of the kappa statistic, particularly the degree to which a change can be considered to be an improvement. For example, is kappa = 0.7 clinically superior to kappa = 0.6 in terms of agreement? In fact, common statistical methods such as kappa and ROC are not designed to determine if a computeraided diagnosis tool provides interpretations that are "equivalent" to expert interpretations and a new framework is needed for addressing these questions.
In this manuscript, we present a statistical modeling approach [16] called loglinear modeling which is more informative and useful for evaluating a new computeraided diagnostic tool against experts than ROC curves and kappa statistics. This approach can fully characterize the accuracy of computeraided diagnosis tool against experts by evaluating the pattern of agreement across rating categories. Moreover, it can quantify the magnitude of the agreement and assess its statistical significance. In particular, the modeling approach can address the critical question: Is the performance of a new diagnostic tool equivalent to performance of an expert? To illustrate the added value of loglinear analysis over ROC and kappa, we compared the accuracy of a new CAD approach for the diagnosis of renal obstruction (RENEX) to the diagnosis provided by three experts where RENEX and the experts rated each kidney in a series of diuresis renography studies as nonobstructed, equivocal, and obstructed.
Methods
Institutional review board approval was obtained for this HIPAAcompliant study; the requirement for informed patient consent was waived. RENEX is a renal expert system for detecting renal obstruction using pre and postfurosemide Tc99 m mercaptoacetyltriglycine (MAG3) renal scans [18]. RENEX consists of: (1) a parameter knowledge library with the list of the boundary conditions necessary for transforming the values of each quantitative parameter such as time to peak height of the renogram curve or time to half maximum counts (T 1/2) to a certainty factor describing the degree of abnormality or normality of that parameter, (2) a knowledge base of heuristic rules that uses certainty factors describing the degree of normality or abnormality of specific parameters to generate new certainty factors specifying the likelihood of obstruction,(3) and an inference engine to combine the certainty factors of the rules and parameters to reach a final certainty factor (conclusion) in regard to obstruction [19]. Detailed description of the architecture of RENEX is presented in a separate publication [18]. RENEX was optimized using pilot data [18] and prospectively validated [10].
This study analyzed a previously published data base that compared diuresis renography scan interpretation generated by RENEX with the consensus and individual scan interpretation of three experts using ROC and kappa analysis [10]. The database consisted of 95 patient studies (55 males and 40 females with a mean age of 58.6 years, SD = 16.5) and contained 185 kidneys classified by RENEX and three experts as obstructed, equivocal, or nonobstructed. Readers were defined as "expert" on the basis of the fact that each reader had > 20years experience in fulltime academic nuclear medicine, had multiple publications in renal nuclear medicine and have been invited to give renal nuclear medicine educational session as national radiology and nuclear medicine meetings. The experts were blinded to clinical information and had prospectively and independently rated each kidney as obstructed, equivocal, or nonobstructed; a consensus reading was subsequently obtained by resolving the differences of expert readings. RENEX analyzed the 95 patient studies based on quantitative parameters automatically extracted from baseline and furosemide acquisitions [20, 21] and used clinically validated optimal cutoff points to classify a kidney as obstructed, equivocal, or nonobstructed.
The diuresis renography protocol was a twostage acquisition based on a minor modification of the consensus recommendations [22]. A 24min baseline Tc99 m MAG3 scan was first obtained. If there was prompt bilateral drainage, obstruction was excluded and furosemide was not administered; if there was delayed drainage in one or both kidneys, the patient received furosemide and an additional 20min scan was obtained. Exclusion of clearly nonobstructed patients (those with a normal baseline acquisition who, consequently, did not receive furosemide) weighted the study population toward a higher percentage of patients with an indeterminate or obstructed kidney.
Statistical modeling
Our primary interest in using a loglinear modeling approach was to characterize the overall structure of agreement present in the data. Carefully considering possible reasons why agreement is present, our modeling procedure can quantify the pattern and magnitude of agreement. For example, we can address the question as to whether agreement in the data is due to chance or due to actual agreement among the raters. Moreover, the actual rater agreement can be further divided into categoryspecific agreement components (obstructed, equivocal, nonobstructed) because it is possible to have different agreement patterns in each of the different response categories. These various components were incorporated by specifying a series of statistical models, starting with the independence model. Goodnessoffit tests were conducted to select the best model that characterizes the structure of agreement present in the data. For model selection, significance level α = 0.05 was used. Once a model was selected, we used the more conservative significance level of α = 0.01 due to multiple comparisons.
We employed an appropriate loglinear model to address the two questions:

1.
How does the rating of RENEX compare with the consensus interpretation of three experts?

2.
How does the rating of RENEX compare with the interpretations of the individual experts?
To address the first question, we treated consensus reading as the interpretation of one expert. To address the second question, we evaluated two agreement patterns: (1) pairwise agreement (i.e., two raters) and (2) threeway agreement (i.e., three raters). This evaluation allowed us to determine if the performance of RENEX was equivalent to the performance of expert readers.
Comparison of RENEX to the consensus interpretation
To investigate the agreement between RENEX and expert consensus, we developed a sequence of nested loglinear models starting with the simplest model, a baseline model for Table 1. This model is also called the independence model which assumes that the agreement between consensus and RENEX is due to chance alone. The cell count in the 3 × 3 table is modeled via two components: a separate effect due to RENEX and a separate effect due to consensus interpretation.
Experts are expected to agree among themselves more often than chance would allow. In this case, expert ratings will not be independent and an association in the 3 × 3 contingency table will exist. The resulting rating pattern may thus be described by a configuration with a larger number of counts on the main diagonal than would be expected under independence. If this pattern occurs, the independence model fits the data poorly. When the independence model is not adequate to explain the data, a component measuring the extra agreement present on the main diagonal is added to the model. This model is referred to as the homogeneous agreement model and assigns equal strength of agreement between RENEX and consensus readings across each category (nonobstructed, equivocal, and obstructed). The homogeneous agreement model thus has two components: the first representing chance, the second representing agreement. A significant positive agreement of the second component suggests positive agreement beyond that expected by chance.
When the homogeneous agreement model still cannot adequately capture the agreement information in the data, the homogeneous agreement term is replaced by three terms representing different agreement strengths in each reading category. This is called the nonhomogeneous agreement model. Note that if all the categories have a uniform level of agreement, we will have the homogeneous agreement model. The modeling procedure is described in detail in the appendix.
The independence model, homogeneous agreement model, and nonhomogeneous agreement model form a nested sequence of models. As such, a likelihoodratio test can used to examine the improvement in fit. Regression coefficients associated with the agreement terms are calculated under the bestfitting model.
Comparison of RENEX to expert raters
The goal of this comparison is to determine if the performance of the RENEX is actually equivalent to that of expert readers. There were three experts and RENEX; hence, the data could be considered as having four raters, each evaluating kidneys into three categories. To address the question of equivalence of RENEX with respect to individual expert readers, we compared the agreement within experts to the agreement between RENEX and experts. Because we had three experts, it was natural to consider two agreement patterns: (1) pairwise agreement and (2) threeway agreement. We thus examined the agreement between RENEX and individual experts by taking two or three raters at a time.
As before, we first started with the independence model which included effects due to all three experts and RENEX. Next, a model allowing pairwise agreement was considered. This was done by adding terms to the previous model which are effects due to pairwise agreement among experts and effects due to pairwise agreement between RENEX and an expert. This is the homogenous pairwise agreement model. The third model extended this homogeneous model by expanding the terms described in the homogeneous pairwise agreement model to reflect different strengths of pairwise agreement according to response categories. Finally, we considered the threeway agreement model by including effect due to threeway agreement among experts and threeway agreement among expert and RENEX. The modeling procedure is detailed in the appendix.
Results
Agreement between RENEX and the Consensus Interpretations
The agreement between RENEX and the consensus readings for 185 kidneys is shown in Table 1. The expert system agreed with the consensus reading in 84% (101/120) of nonobstructed kidneys, in 92% (33/36) of obstructed kidneys, and in 45% (13/29) of equivocal kidneys.
To determine the best model for the agreement between RENEX and expert consensus, a series of models were examined. The likelihoodratio statistics (G^{2}) for both the independence model, G^{2} = 138.55 [df (degrees of freedom) = 4, p < 0.001] and homogeneous agreement model G^{2} = 21.38 (df = 3, p < 0.001) indicated that neither of these models was adequate to describe the data. (When performing the likelihoodratio test in loglinear analysis, a model is considered adequate if its p value is at or above 0.05). For the nonhomogeneous agreement model, G^{2} = 0.16 (df = 1, p = 0.69) showing the adequacy of the model in describing the pattern of agreement; the agreement pattern in the data favors assigning different strengths of agreement to the three response categories.
The results based on the nonhomogeneous agreement model are displayed in Table 2. Of the 185 kidneys, expert consensus classified 65%, 16%, and 19%, respectively, as nonobstructed, equivocal, and obstructed. On the other hand, RENEX classified 59%, 16%, and 25%, respectively, as nonobstructed, equivocal, and obstructed. The regression coefficients for the nonobstructed, equivocal, and obstructed categories were, respectively, 1.57 (p < 0.0001), 0.28 (p = 0.37), and 1.82 (p < 0.0001). These coefficients show that there was a significant positive agreement between RENEX and expert consensus in the nonobstructed and obstructed categories.
Both kappa [10] and loglinear analysis showed that consensus and RENEX interpretations agreed beyond chance; however the loglinear modeling approach further suggested that the agreement pattern among the three response categories was not uniform. In particular, RENEX and expert consensus rated the renal scans with high agreement in the nonobstructed and obstructed categories while they did not seem to agree well in the equivocal category.
Agreement between RENEX and the Individual Experts
To address the question of whether RENEX is equivalent to an expert, we examined the pattern of agreement between experts and RENEX by considering pairwise agreement and threeway agreement.

1.
Pairwise agreement within experts and between experts and RENEX. Based on likelihoodratio tests, the nonhomogeneous model (G ^{2} = 58.69, df = 54, p = 0.31) was preferred over the independence model (G ^{2} = 530.70, df = 72, p < 0.0001) and the homogeneous model (G ^{2} = 102.61, df = 66, p < 0.01). The results based on the nonhomogeneous agreement model are displayed in Table 3. Although the pattern of pairwise agreement is not apparent across all raters, coefficients in the nonobstructed category seem to indicate positive significant agreement. A hypothesistesting approach provides more insight into the pattern of agreement, which is described next.
Tests of hypothesis in Table 3 indicated that for all response categories, the pairwise agreement among experts was the same (p = 0.15, 0.08, 0.11, respectively, for the nonobstructed, equivocal, and obstructed categories). Moreover, the pairwise agreement between RENEX and an expert are also the same (p = 0.47, 0.34, 0.58, respectively, for the nonobstructed, equivocal and obstructed categories). Finally, the overall pairwise agreement between two experts and the overall pairwise agreement between RENEX and an expert are the same (p = 0.41, 0.95, 0.81, respectively). Hence, pairwise agreement appeared to remain the same when an expert was replaced by RENEX. In terms of pairwise comparisons, the performance of RENEX is equivalent to that of an expert.

2.
Threeway agreement within experts and between experts and RENEX. Based on likelihoodratio tests, the nonhomogeneous agreement model (G ^{2} = 61.63, df = 60, p = 0.42) was preferred over the homogeneous agreement model (G ^{2} = 113.60, df = 68, p < 0.001) showing that the pattern of threeway agreement in the data is different for the three response categories. The results based on the nonhomogeneous agreement model are displayed in Table 4. Coefficients in the nonobstructed and obstructed categories indicate significant positive agreement. Tests of hypothesis suggest that agreement among three experts is the same as agreement among two experts and RENEX for the nonobstructed (p = 0.79), obstructed (p = 0.49) and equivocal categories (p = 0.03). Since none of these values reached the level of significance (p ≤ 0.01, Table 4), RENEX appears to be equivalent to an expert in all three categories.
Discussion
One goal of this manuscript was to show the advantages of loglinear regression analysis when a gold standard is absent by analyzing a previously published database that assessed the accuracy of computeraided diagnosis of renal obstruction against the diagnosis provided by expert readers using kappa and ROC methods [10]. In the ROC analysis, the expert consensus was used as the gold standard but this approach is problematic because ROC analysis should have a gold standard independent of the test under evaluation. Unfortunately, this problem occurs whenever an expert panel is used as the gold standard. Secondly, ROC analysis requires just two categories, disease present or disease absent. To apply ROC analysis, equivocal interpretations have to be placed into the disease present or disease absent category [10]. This requirement may obscure critical information and fails to represent the clinical setting where some interpretations are, in fact, equivocal.
An alternative to ROC analysis is kappa analysis. The weighted kappa statistic between RENEX and expert consensus readings was 0.72 which indicated good agreement between RENEX and experts [10]. The weighted kappa coefficients between each pair of experts and between RENEX and each expert also ranged from 0.61 to 0.73 [10]; this close agreement of kappa coefficients suggested that RENEX was performing similarly to an expert. However, kappa analysis does not provide a framework for evaluating the pattern of agreement across different categories. In our analysis, we found that RENEX has better agreement with consensus and experts in obstructed and nonobstructed categories, but not in the equivocal category.
Loglinear models [16] can establish the general pattern and magnitude of agreement which provides valuable information for improving the reliability of a computeraided diagnosis system. In a loglinear framework, agreement is specified by two components: one represents the effect of chance and the other represents the effect of rater agreement beyond chance. Compared to a summary statistic like kappa, loglinear models provide a straightforward test of the magnitude of the difference in agreement and also provide more information about agreement such as the structure and pattern of the agreement across categories.
For example, a kappa of 0.72 only gives us a sense that RENEX agrees well with consensus, but it is hard to say whether the agreement is high in all three reading categories (obstructed, equivocal or nonobstructed) or exists only in some categories. Loglinear modeling shows us that the overall agreement between RENEX and consensus across the reading categories exists beyond chance (p < 0.001). Furthermore, the significance tests under a nonhomogeneous agreement model suggest that the beyondchance agreement mainly comes from the ratings in nonobstructed and obstructed categories. To determine if a new diagnostic tool is "equivalent to" an expert, the advantage of loglinear models becomes more apparent since the agreement among various combinations of raters can be specified in one single model and can be tested directly in this context (Table 4). The results of significance tests led us to conclude that RENEX behaves equally to an expert in all reading categories.
Conclusions
Loglinear modeling (1) provided more insight into the pattern and magnitude of interrater agreement than ROC and kappa analysis, (2) showed that RENEX performed as well as any expert reader particularly rating in obstructed and nonobstructed categories, and (3) should be considered when a gold standard is absent. This analysis provides additional evidence that the renal expert system (RENEX) interprets diuresis renography studies as well as human experts and has the potential to assist in the interpretation of diuresis renography studies.
Authors' information
AM is Professor of Biostatistics and Bioinformatics, Rollins School of Public Health, Emory University; JNGB is Research Associate Professor in the Department of Biostatistics and Bioinformatics, Rollins School of Public Health, Emory University; ATT is Professor of Radiology, Emory University School of Medicine, Atanta, GA, USA.
Appendix
Loglinear modeling
Comparison of RENEX vs. consensus
Let u _{ ij } be the expected cell count in a 3 × 3 contingency table where the consensus reading falls in the i th category and the RENEX reading falls in the j th category (see Table 1). u represents the overall effect, represents the effect due to the i th level of consensus, and represents the effect due to the j th level of RENEX. Similar to the regression framework, a series of models is fitted to determine the pattern of agreement. The simplest model to start with is the independence model, which can be written as
.
The next model to be considered is the homogeneous agreement model which assumes the same strength of agreement across the three reading categories. That is,
where I = 1 if RENEX agrees with CONSENSUS and 0 otherwise. The parameter δ indicates the beyondchance homogeneous agreement. When δ is zero, the homogeneous agreement model reduces to the independence model.
The third model to be considered is the nonhomogeneous agreement model which assumes a different strength of agreement for different reading categories. That is,
where I _{1} = 1 RENEX agrees with CONSENSUS on the first category and 0 otherwise; I _{2} = 1 RENEX agrees with CONSENSUS on the second category and 0 otherwise, and so on.
If the strength in agreement for different reading categories is the same, then δ_{1}=δ_{2}=δ_{3}, and the nonhomogeneous agreement model reduces to the homogeneous agreement model.
For our data, we fitted these three models using Proc CATMOD in SAS software [23]. For our data, the best model was the nonhomogeneous model. Table 4 shows the regression coefficients, standard errors, and p values.
Comparison of RENEX vs. individual expert readers
In the pairwise homogeneous agreement model, the agreement component assumes the equal strength of agreement between any pair of raters:
The superscripts E _{1}, E _{2}, E _{3}, and R refer to the main effect of expert 1, expert 2, expert 3, and RENEX, respectively. I _{1} = 1 if experts 1 and 2 agree, 0 otherwise; I _{2} = 1 if experts 1 and 3 agree, 0 otherwise; and so on.
When δ _{1} = δ _{2} = δ _{3} = θ _{1} = θ _{2} = θ _{3} = 0, the homogeneous pairwise agreement reduces to the independence model. To determine whether RENEX is behaving similar to experts, we tested the null hypothesis: H_{0}: δ _{1} + δ _{2} + δ _{3} = θ _{1} + θ _{2} + θ _{3}.
The next model to be considered was the nonhomogeneous pairwise agreement model which permitted a different strength of agreement for different reading categories. That is,
I _{11} = 1 if experts 1 and 2 agree on the first category, 0 otherwise; I _{12} = 1 if experts 1 and 3 agree on the first category, 0 otherwise; and so on. Here, we were interested in testing the null hypotheses for the mth category: H _{0}: δ _{m1} + δ _{m2} + δ _{m3} = θ _{m1} + θ _{m2} + θ _{m3}.
The threeway agreement can be modeled in a similar way by appropriately changing the definitions of δ s and θ s to accommodate threeway agreement.
As before, we fit these models using Proc CATMOD in SAS software [23]. For our data the best models were the nonhomogeneous pairwise agreement model and the nonhomogeneous threeway agreement model. Table 4 shows the regression coefficients, standard errors, and p values.
Abbreviations
 ROC:

receiver operating characteristic
 CAD:

computeraided diagnosis
 df:

degrees of freedom
 SD:

standard deviation.
References
 1.
Li F, Engleman R, Metz CE, Doi K, MacMahon H: Lung cancers missed on chest radiographs: Results obtained with a commercial computeraided detection program. Radiology 2008, 246: 273–280.
 2.
Taylor SA, Charmin SC, Lefere P, McFarland EG, Paulson EK, Yee J, Aslam R, Barlow JM, Gupta A, Kim DH, Miller CM, Halligan S: CT Colonography: Investigation of the optimum reader paradigm by using computeraided detection software. Radiology 2008, 246: 463–471.
 3.
Iglehart J: The new era of medical imagingprogress and pitfalls. N Eng J Med 2006, 354: 2822–2828. 10.1056/NEJMhpr061219
 4.
IMV Medical information division: 2003 nuclear medicine census market summary report. Volume IV. IMV Limited, Des Plaines, IL; 2003:7–11.
 5.
Hunsche A: A value of quantitative data in the interpretation of diuresis renography for suspected urinary tract obstruction. In Ph D thesis. Federal University of Rio Grande o Sul, Porto Alegre, Rio Grande o Sul; 2006.
 6.
Kupinski MA, Hoppin JW, Clarkson E, Barrett HH, Kastis GA: Estimation in medical imaging without a gold standard. Academic Radiology 2002, 9: 290–297. 10.1016/S10766332(03)803720
 7.
Kundel HL, Polansky M: Mixture distribution and receiver operating characteristic analysis of bedside chest imaging with screenfilm and computed radiology. Acad Radiol 1997, 4: 1–7. 10.1016/S10766332(97)801523
 8.
Kung JW, Matsumoto S, Hasegawa I, Nguyen B, Toto LC, Kundel H, Hatabu H: Mixture distribution analysis of a computer assisted diagnostic method for the evaluation of pulmonary nodules on computed tomography scan. Acad Radiol 2004, 11: 281–285. 10.1016/S10766332(03)007177
 9.
Nelson JC and Pepe MS: Statistical description of interrater variability in ordinal ratings. Statistical Methods in Medical Research 2000,9(5):475–496. 10.1191/096228000701555262
 10.
Taylor A Jr, Garcia EV, Binongo J, Manatunga A, Folks RD, Dubovsky E: Diagnostic performance of an expert system for the interpretation of Tc99 m MAG3 scans to detect renal obstruction. J Nucl Med 2008, 49: 216–224. 10.2967/jnumed.107.045484
 11.
Chan HP, Sahiner B, Helvie MA, Petrick N, Roubidoux MA, Wilson TE, Adler DD, Paramagul C, Newman JS, and SanjayGopal S: Improvement of radiologists' characterization of mammographic masses by using computeraided diagnosis: an ROC study. Radiology 1999, 212: 817.
 12.
Chakraborty DP, Breatnach ES, Yester MV, Soto B, Barnes GT, and Fraser RG: Digital and conventional chest imaging: a modified ROC study of observer performance using simulated nodules. Radiology 1986, 158: 35–39.
 13.
Cohen J: A coefficient of agreement for nominal tables. Educational and Psychological measurement 1960, 20: 37–46. 10.1177/001316446002000104
 14.
Agresti A: A model for agreement between ratings on an ordinal scale. Biometrics 1988, 44: 539–548. 10.2307/2531866
 15.
Light RJ: Measures of response agreement for qualitative data: some generalizations and alternatives. Psychological Bulletin 1971, 5: 365–377.
 16.
Tanner MA, Young MA: Modeling agreement among raters. JASA 1985, 80: 175–180.
 17.
Kraemer HC: Ramifications of a population model for kappa as a coefficient of reliability. Psychometrika 1979, 44: 461–472. 10.1007/BF02296208
 18.
Garcia EV, Taylor A, Halkar R et al: RENEX: An expert system for the interpretation of Tc99 m MAG3 scans to detect renal obstruction. J Nucl Med 2006, 47: 320–329.
 19.
Taylor A, Manatunga A, Garcia EV: Decision support systems in diuresis renography. Semin Nucl Med 2008, 38: 67–81. 10.1053/j.semnuclmed.2007.09.006
 20.
Taylor A Jr, Corrigan PL, Galt J, et al.: Measuring technetium99 mMAG3 clearance with an improved camerabased method. J Nucl Med 1995, 36: 1689–1695.
 21.
Taylor A Jr, Manatunga A, Morton K, et al.: Multicenter trial validation of a camerabased method to measure Tc99 m mercaptoacetyltriglycine, or Tc99 m MAG3, clearance. Radiology 1997, 204: 47–54.
 22.
O'Reilly P, Aurell M, Britton K, et al.: Consensus on diuresis renography for investigating the dilated upper urinary tract. J Nucl Med 1996, 37: 1872–1876.
 23.
SAS/STAT^{®} 9.2 User's Guide. Chapter 28: The CATMOD Procedure Cary, NC: SAS Institute 1998, 1092–1127.
Acknowledgements
This work was supported by a National Institute of Health grant, RO1EB008838, funded from the National Institute of Biomedical Imaging and BioEngineering and from the National Institute of Diabetes and Digestive and Kidney Diseases and a URC grant from Emory University. We would also like to thank Eva V. Dubovsky, MD, PhD and Raghuveer Halkar, MD for serving as expert readers and Russell Folks, CNMT for his assistance in the acquisition and organization of the data.
Author information
Affiliations
Corresponding author
Additional information
Competing interests
ATT receives royalties from the sale of the application software QuantEM™ related to the research described in this article. The terms of this arrangement have been reviewed and approved by Emory University in accordance with its conflictofinterest practice. ATT used to consult for Mallinckrodt, Inc (now Covidien); he has not consulted for the past 3 years.
Authors' contributions
AM and JNGB designed the loglinear model, conducted the statistical analysis, wrote the appendix, and assisted in writing the manuscript. ATT designed the project, wrote the first draft of the manuscript, critically reviewed the data and completed the final version of the manuscript following input from AM and JNGB.
Rights and permissions
Open Access This article is distributed under the terms of the Creative Commons Attribution 2.0 International License ( https://creativecommons.org/licenses/by/2.0 ), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
About this article
Cite this article
Manatunga, A.K., G Binongo, J.N. & Taylor, A.T. Computeraided diagnosis of renal obstruction: utility of loglinear modeling versus standard ROC and kappa analysis. EJNMMI Res 1, 5 (2011). https://doi.org/10.1186/2191219X15
Received:
Accepted:
Published:
Keywords
 Loglinear modeling
 Renal obstruction
 Diuresis renography