Skip to main content

Main menu

  • Home
  • Content
    • Current
    • Ahead of print
    • Past Issues
    • JNM Supplement
    • SNMMI Annual Meeting Abstracts
    • Continuing Education
    • JNM Podcasts
  • Subscriptions
    • Subscribers
    • Institutional and Non-member
    • Rates
    • Journal Claims
    • Corporate & Special Sales
  • Authors
    • Submit to JNM
    • Information for Authors
    • Assignment of Copyright
    • AQARA requirements
  • Info
    • Reviewers
    • Permissions
    • Advertisers
  • About
    • About Us
    • Editorial Board
    • Contact Information
  • More
    • Alerts
    • Feedback
    • Help
    • SNMMI Journals
  • SNMMI
    • JNM
    • JNMT
    • SNMMI Journals
    • SNMMI

User menu

  • Subscribe
  • My alerts
  • Log in
  • My Cart

Search

  • Advanced search
Journal of Nuclear Medicine
  • SNMMI
    • JNM
    • JNMT
    • SNMMI Journals
    • SNMMI
  • Subscribe
  • My alerts
  • Log in
  • My Cart
Journal of Nuclear Medicine

Advanced Search

  • Home
  • Content
    • Current
    • Ahead of print
    • Past Issues
    • JNM Supplement
    • SNMMI Annual Meeting Abstracts
    • Continuing Education
    • JNM Podcasts
  • Subscriptions
    • Subscribers
    • Institutional and Non-member
    • Rates
    • Journal Claims
    • Corporate & Special Sales
  • Authors
    • Submit to JNM
    • Information for Authors
    • Assignment of Copyright
    • AQARA requirements
  • Info
    • Reviewers
    • Permissions
    • Advertisers
  • About
    • About Us
    • Editorial Board
    • Contact Information
  • More
    • Alerts
    • Feedback
    • Help
    • SNMMI Journals
  • View or Listen to JNM Podcast
  • Visit JNM on Facebook
  • Join JNM on LinkedIn
  • Follow JNM on Twitter
  • Subscribe to our RSS feeds
LetterLetters to the Editor

A Common Mistake in Assessing the Diagnostic Value of a Test: Failure to Account for Statistical and Methodologic Issues

Siamak Sabour
Journal of Nuclear Medicine July 2017, 58 (7) 1182-1183; DOI: https://doi.org/10.2967/jnumed.115.156745
Siamak Sabour
Shahid Beheshti University of Medical Sciences Tehran 1413663633, Islamic Republic of Iran E-mail:
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • For correspondence: s.sabour@sbmu.ac.ir
  • Article
  • Info & Metrics
  • PDF
Loading

TO THE EDITOR: I was interested to read the paper by Anand et al. in the December 2016 edition of The Journal of Nuclear Medicine (1). The purpose of the authors was to assess the impact of variability in scanning speed and in vendor-specific γ-camera settings on the reproducibility and accuracy of the automated bone scan index (BSI) (1). They measured reproducibility as the absolute difference between repeated BSI values, and they measured accuracy as the absolute difference between observed BSI values and phantom BSI values. Descriptive statistics were used to compare the generated data.

Reproducibility (reliability) and accuracy (validity), as two completely different methodologic issues, should be assessed using appropriate tests. It is crucial to be aware that, regarding reliability, one should use the intraclass correlation coefficient for quantitative variables and the weighted κ-test for qualitative variables. However, regarding validity, one should use the interclass correlation coefficient (Pearson r) for quantitative variables whereas the most appropriate tests for qualitative variables may include sensitivity, specificity, positive and negative predictive value, positive and negative likelihood ratio, diagnostic accuracy, and odds ratio. Moreover, in analyzing reliability, one should apply an individual-based approach using single-measure intraclass correlation coefficient agreement because applying a global-average approach (absolute difference) can be misleading. A test may indicate high validity, yet there may be no reliability at all (2–8).

Anand et al. enrolled 25 patients in each of 3 groups and observed a significantly lower reproducibility for group 2 (mean ± SD, 0.35 ± 0.59) than for group 1 (0.10 ± 0.13; P < 0.0001) or group 3 (0.09 ± 0.10; P < 0.0001). However, no significant difference in reproducibility was observed between group 3 and group 1 (P = 0.388) (1). Statistical significance and clinical importance are two completely different issues, and in clinical research—especially reliability analysis—we should not put the emphasis on significance level (P value) (2–8).

They concluded that the accuracy and reproducibility of automated BSI were dependent on scanning speed but not on vendor-specific γ-camera settings. Such a conclusion should be supported by the above-mentioned statistical and methodologic issues. Otherwise, in clinical practice, misdiagnosis and patient mismanagement may occur.

Footnotes

  • Published online Jan. 12, 2017.

  • © 2017 by the Society of Nuclear Medicine and Molecular Imaging.

REFERENCES

  1. 1.↵
    1. Anand A,
    2. Morris MJ,
    3. Kaboteh R,
    4. et al
    . A preanalytic validation study of automated bone scan index: effect on accuracy and reproducibility due to the procedural variabilities in bone scan image acquisition. J Nucl Med. 2016;57:1865–1871.
    OpenUrlAbstract/FREE Full Text
  2. 2.↵
    1. Szklo M,
    2. Nieto FJ
    . Epidemiology Beyond the Basics. 2nd ed. Manhattan, NY: Jones and Bartlett; 2007.
  3. 3.
    1. Sabour S
    . Myocardial blood flow quantification by Rb-82 cardiac PET/CT: methodological issues on reproducibility study. J Nucl Cardiol. September 6, 2016 [Epub ahead of print].
  4. 4.
    1. Sabour S
    . Reproducibility of semi-automatic coronary plaque quantification in coronary CT angiography with sub-mSv radiation dose: common mistakes. J Cardiovasc Comput Tomogr. 2016;10:e21–e22.
    OpenUrl
  5. 5.
    1. Sabour S
    . Reliability of a new modified tear breakup time method: methodological and statistical issues. Graefes Arch Clin Exp Ophthalmol. 2016;254:595–596.
    OpenUrl
  6. 6.
    1. Sabour S,
    2. Farzaneh F,
    3. Peymani P
    . Evaluation of the sensitivity and reliability of primary rainbow trout hepatocyte vitellogenin expression as a screening assay for estrogen mimics: methodological issues. Aquat Toxicol. 2015;164:175–176.
    OpenUrl
  7. 7.
    1. Sabour S
    . Re: does the experience level of the radiologist, assessment in consensus, or the addition of the abduction and external rotation view improve the diagnostic reproducibility and accuracy of MRA of the shoulder? [comment]. Clin Radiol. 2015;70:333–334.
    OpenUrl
  8. 8.↵
    1. Sabour S
    . The reliability of routine clinical post-processing software in assessing potential diffusion-weighted MRI “biomarkers” in brain metastases, common mistake [comment]. Magn Reson Imaging. 2014;32:1162.
    OpenUrl
PreviousNext
Back to top

In this issue

Journal of Nuclear Medicine: 58 (7)
Journal of Nuclear Medicine
Vol. 58, Issue 7
July 1, 2017
  • Table of Contents
  • Table of Contents (PDF)
  • About the Cover
  • Index by author
Print
Download PDF
Article Alerts
Sign In to Email Alerts with your Email Address
Email Article

Thank you for your interest in spreading the word on Journal of Nuclear Medicine.

NOTE: We only request your email address so that the person you are recommending the page to knows that you wanted them to see it, and that it is not junk mail. We do not capture any email address.

Enter multiple addresses on separate lines or separate them with commas.
A Common Mistake in Assessing the Diagnostic Value of a Test: Failure to Account for Statistical and Methodologic Issues
(Your Name) has sent you a message from Journal of Nuclear Medicine
(Your Name) thought you would like to see the Journal of Nuclear Medicine web site.
Citation Tools
A Common Mistake in Assessing the Diagnostic Value of a Test: Failure to Account for Statistical and Methodologic Issues
Siamak Sabour
Journal of Nuclear Medicine Jul 2017, 58 (7) 1182-1183; DOI: 10.2967/jnumed.115.156745

Citation Manager Formats

  • BibTeX
  • Bookends
  • EasyBib
  • EndNote (tagged)
  • EndNote 8 (xml)
  • Medlars
  • Mendeley
  • Papers
  • RefWorks Tagged
  • Ref Manager
  • RIS
  • Zotero
Share
A Common Mistake in Assessing the Diagnostic Value of a Test: Failure to Account for Statistical and Methodologic Issues
Siamak Sabour
Journal of Nuclear Medicine Jul 2017, 58 (7) 1182-1183; DOI: 10.2967/jnumed.115.156745
Twitter logo Facebook logo LinkedIn logo Mendeley logo
  • Tweet Widget
  • Facebook Like
  • Google Plus One
Bookmark this article

Jump to section

  • Article
    • Footnotes
    • REFERENCES
  • Info & Metrics
  • PDF

Related Articles

  • No related articles found.
  • PubMed
  • Google Scholar

Cited By...

  • Diagnostic accuracy of novel ultrasonographic halo score for giant cell arteritis: methodological issues
  • Google Scholar

More in this TOC Section

  • Business Model Beats Science and Logic: Dosimetry and Paucity of Its Use
  • Determining PSMA-617 Mass and Molar Activity in Pluvicto Doses
  • The Value of Functional PET in Quantifying Neurotransmitter Dynamics
Show more Letters to the Editor

Similar Articles

SNMMI

© 2025 SNMMI

Powered by HighWire