Skip to main content

Main menu

  • Home
  • Content
    • Current
    • Ahead of print
    • Past Issues
    • JNM Supplement
    • SNMMI Annual Meeting Abstracts
    • Continuing Education
    • JNM Podcasts
  • Subscriptions
    • Subscribers
    • Institutional and Non-member
    • Rates
    • Journal Claims
    • Corporate & Special Sales
  • Authors
    • Submit to JNM
    • Information for Authors
    • Assignment of Copyright
    • AQARA requirements
  • Info
    • Reviewers
    • Permissions
    • Advertisers
  • About
    • About Us
    • Editorial Board
    • Contact Information
  • More
    • Alerts
    • Feedback
    • Help
    • SNMMI Journals
  • SNMMI
    • JNM
    • JNMT
    • SNMMI Journals
    • SNMMI

User menu

  • Subscribe
  • My alerts
  • Log in
  • Log out
  • My Cart

Search

  • Advanced search
Journal of Nuclear Medicine
  • SNMMI
    • JNM
    • JNMT
    • SNMMI Journals
    • SNMMI
  • Subscribe
  • My alerts
  • Log in
  • Log out
  • My Cart
Journal of Nuclear Medicine

Advanced Search

  • Home
  • Content
    • Current
    • Ahead of print
    • Past Issues
    • JNM Supplement
    • SNMMI Annual Meeting Abstracts
    • Continuing Education
    • JNM Podcasts
  • Subscriptions
    • Subscribers
    • Institutional and Non-member
    • Rates
    • Journal Claims
    • Corporate & Special Sales
  • Authors
    • Submit to JNM
    • Information for Authors
    • Assignment of Copyright
    • AQARA requirements
  • Info
    • Reviewers
    • Permissions
    • Advertisers
  • About
    • About Us
    • Editorial Board
    • Contact Information
  • More
    • Alerts
    • Feedback
    • Help
    • SNMMI Journals
  • View or Listen to JNM Podcast
  • Visit JNM on Facebook
  • Join JNM on LinkedIn
  • Follow JNM on Twitter
  • Subscribe to our RSS feeds
EditorialEditorial

An Opinion on ChatGPT in Health Care—Written by Humans Only

Jens Kleesiek, Yonghui Wu, Gregor Stiglic, Jan Egger and Jiang Bian
Journal of Nuclear Medicine May 2023, 64 (5) 701-703; DOI: https://doi.org/10.2967/jnumed.123.265687
Jens Kleesiek
1Institute for AI in Medicine, University Medicine Essen, Essen, Germany;
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Yonghui Wu
2Department of Health Outcomes and Biomedical Informatics, College of Medicine, University of Florida, Gainesville, Florida; and
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Gregor Stiglic
3Faculty of Health Sciences, University of Maribor, Maribor, Slovenia
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Jan Egger
1Institute for AI in Medicine, University Medicine Essen, Essen, Germany;
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Jiang Bian
2Department of Health Outcomes and Biomedical Informatics, College of Medicine, University of Florida, Gainesville, Florida; and
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • Article
  • Figures & Data
  • Info & Metrics
  • PDF
Loading

ChatGPT, created by OpenAI, has taken the world by storm, and its user base is growing even faster than the previous record held by TikTok, reaching 100 million users in just 2 mo after it launched. Textual context, presentations, and even source code are already being generated using ChatGPT. Many publications have been issued, and meanwhile, ChatGPT has been banned as an author by many publishing companies for several different reasons, such as plagiarism, incorrect information, or inaccurate information (1,2), whereas others argue its benefits, such as the ability to write more coherent sentences than nonnative speakers (3). But that does not stop people from all walks of health care from using it.

ChatGPT is powered by a generative pretrained transformer (GPT-3.5), which is a large language model (LLM) trained with 175 billion parameters (4). LLMs originate in natural language processing to formulate the probability distribution of a sequence of words or the next word in a sequence. Recent studies report that LLMs are foundation models in which a single model can be adapted to solve a wide range of different natural language-processing tasks because of few-shot learning, zero-shot learning, and transfer learning ability (5). The conversational artificial intelligence (AI) ability is achieved using LLM-based prompt learning (6). To alleviate the toxic responses and integrate human ethics, ChatGPT applied a strategy of reinforcement learning from human feedback to align LLMs to follow human instructions (7). These breakthroughs in natural language processing empower ChatGPT with conversational AI ability so good it has surprised the world. Even within OpenAI, ChatGPT has been a surprise. AI chatbots are not a new thing, but many previous attempts have not achieved the sensation that ChatGPT achieved. Meta’s BlenderBot was a disappointment. What may be different for ChatGPT, beyond the unknown technologies, is OpenAI’s goal of creating artificial general intelligence to match human-level intellect (8). ChatGPT certainly is not an artificial general intelligence, but it sure looks like one because of the breadth and depth of the knowledge it demonstrates through conversations.

Even though many are excited by its first use, disillusionment often sets in over time, for several reasons. On the one hand, ChatGPT gives wrong answers and is prone to confabulation (“a memory error defined as the production of fabricated, distorted, or misinterpreted memories about oneself or the world” (9)). This is exacerbated by the fact that we set different standards for communication among humans and between humans and computers. The belief is that a computer will not make mistakes. Moreover, many users’ expectations are wrong, especially for medical interactions. The program was trained and designed for conversation, not diagnostic support or treatment recommendations. Yet, questions arise as to whether ChatGPT is a medical product and who is liable, even though ChatGPT always generates a disclaimer that it is not a health-care professional licensed to give medical advice. This is a typical case of intended use versus actual use as described in the medical device regulation. We argue that there is a difference between general-purpose conversational AI—in which the focus is the conversational ability such as readability—and medical AI—in which the focus is the health facts about flesh-and-blood humans. Speaking a fake fact using elegant words is amusing (that is why many ChatGPT users are tricking this conversational AI), but providing a wrong fact in medical AI is dangerous—indeed, making ChatGPT a medical device if it should turn out that doctors are actually using it to diagnose and treat their patients. Nevertheless, philosophically, asking ChatGPT for health-related information (to inform health decision making) is not much different from asking Dr. Google, which has long been criticized for not just giving but spreading medical misinformation (10). Nevertheless, this is again not only the gap between intended use versus actual use but also the consistent push and pull between the expectations of the developers versus the end users.

As always with any potentially disruptive technologies, such use can be seen as either a threat or an opportunity. Many articles are optimistic, pointing to the potential symbiosis, the modern centaur, a combination of humans and computers leading to a beneficial augmentation of our capabilities. But pessimistic views also need to be discussed. Take the global positioning system, for example. Because of this technology, many young people are no longer able to navigate with a compass and map. Of course, one could argue that use of a map is not required as a basic skill anymore. But that is certainly not the case with language. If we as humans lose the ability to communicate, debate, and think critically, then we are taking a step backward, leading to devolution.

The question remains: what is the actual use of ChatGPT, despite all the hype during the last few months? Of course, it can be used to generate simple text and to produce code snippets (but often with errors). It can even quickly analyze a research topic and generate an academic paper—again, with frequent errors that may go unnoticed even by reviewers and editors of scientific journals (11). This application may be helpful for student assignments but will not be of much use for learning, in which an individual must come up with a solution through a step-by-step thought process. And when the teacher turns to an oral exam at the end of the semester, irresponsible students who have used the ChatGPT approach will most likely fail, as they might later in their actual work life.

In a health-care setting, one cannot afford to stay on the surface. ChatGPT produces false information that requires checking and correcting of every sentence. In addition to the significant time investment, which calls into question the efficiency of this approach, the real danger comes from false information that goes undetected by the human corrector.

We argue that this technology is worth building toward a clinical knowledge system that can provide health and clinical decision support and enable better self-care and patient care in an era of skill shortages. Hundreds of medical articles are published every day, and it is impossible for humans to cope with this flood of information. In particular, we must keep in mind that information is not the same as knowledge. Filtering information and extracting knowledge from it have enormous potential. Nevertheless, from a health behavior perspective, knowledge is only a small component of healthful behavior and decision making (12). Other aspects, such as beliefs, feelings, norms, and the importance of healthful behavior, are equally important. In our view, developing ChatGPT into a medical product such as a clinical decision support system needs to be considered in a broader context with a wider range of other aspects (e.g., reliability, ethics, and fairness) than just model performance, and like any other AI system, humans must be in the loop (13).

LLMs have much potential in health care. For example, text-to-text generation may help autocomplete the sentences and paragraphs of a clinical document (e.g., a progress report) based on short phrases provided by a human clinician, thus reducing the documentation burden (14). When used for the generation of clinical documents, LLMs also have the potential to integrate the observations of clinicians and knowledge about clinical guidelines, thus reflecting real-world diagnosis and treatment patterns and subsequently being helpful for compiling a differential diagnosis and composing treatment plans. But much more research and development are needed to achieve this goal. We recently developed the first (to our knowledge) clinical LLM, GatorTron (8.9 billion parameters) using over 90 billion words of text (including 82 billion words of clinical text) and demonstrated its power in clinical natural language processing (15). We also examined the text generation ability of SynGatorTron (5 billion and 20 billion parameters), a generative clinical LLM based on the GPT-3 architecture (16). There is ongoing research on LLMs addressing fundamental issues, such as incorporating chains of reasoning through selection-inference and chain-of-thought prompting (17–19). Once current limitations have been addressed, many applications are conceivable using ChatGPT and the next generation of LLMs (Table 1).

View this table:
  • View inline
  • View popup
TABLE 1.

Potential Applications and Areas of Health-Care Research for ChatGPT and Similar LLMs

More than a decade ago, self-driving cars were heralded as a disruptive technology. As with many technologies, the last 20% of development takes 80% of the total time. The same may be happening with ChatGPT, and additional development time will be needed for productive use in health care. Like the Human Genome Project, in which mapping out the base pairs in the human DNA is not the end but only the start of the genetic revolution, Chat GPT’s eventual possibilities are certainly more exciting than the current hype. Whether this technology is a revolution or just an evolution remains to be seen. What is certain in any case is that there will be no more stepping back. As a society, we have the responsibility to shape its future development.

DISCLOSURE

No potential conflict of interest relevant to this article was reported.

Footnotes

  • Published online Apr. 13, 2023.

  • © 2023 by the Society of Nuclear Medicine and Molecular Imaging.

REFERENCES

  1. 1.↵
    1. Shen Y,
    2. Heacock L,
    3. Elias J,
    4. et al
    . ChatGPT and other large language models are double-edged swords. Radiology. January 26, 2023 [Epub ahead of print].
  2. 2.↵
    Introducing ChatGPT. OpenAI website. https://openai.com/blog/chatgpt. Published November 30, 2022. Accessed March 20, 2023.
  3. 3.↵
    1. Ma Y,
    2. Liu J,
    3. Yi F,
    4. et al
    . AI vs. human: differentiation analysis of scientific content generation. arXiv website. https://arxiv.org/abs/2301.10416. Published January 24, 2023. Revised February 12, 2023. Accessed March 20, 2023.
  4. 4.↵
    1. Floridi L,
    2. Chiriatti M
    . GPT-3: its nature, scope, limits, and consequences. Minds Machines. 2020;30:681–694.
    OpenUrl
  5. 5.↵
    1. Brown T,
    2. Mann B,
    3. Ryder N,
    4. et al
    . Language models are few-shot learners. In: Advances in Neural Information Processing Systems. Vol 33. Curran Associates, Inc.; 2020:1877–1901.
    OpenUrl
  6. 6.↵
    1. Gao T,
    2. Fisch A,
    3. Chen D
    . Making pre-trained language models better few-shot learners. ACL Anthology website. https://aclanthology.org/2021.acl-long.295/. Published August 2021. Accessed March 20, 2023.
  7. 7.↵
    Aligning language models to follow instructions. OpenAI website. https://openai.com/research/instruction-following. Published January 27, 2022. Accessed March 20, 2023.
  8. 8.↵
    1. Roose K
    . How ChatGPT kicked off an A.I. arms race. The New York Times website. https://www.nytimes.com/2023/02/03/technology/chatgpt-openai-artificial-intelligence.html. Published February 3, 2023. Accessed March 6, 2023.
  9. 9.↵
    Confabulation. Wikipedia website. https://en.wikipedia.org/wiki/Confabulation. Updated March 16, 2023. Accessed April 4, 2023.
  10. 10.↵
    1. Granter SR,
    2. Papke DJ
    . Medical misinformation in the era of Google: computational approaches to a pervasive problem. Proc Natl Acad Sci USA. 2018;115:6318–6321.
    OpenUrlFREE Full Text
  11. 11.↵
    1. Anderson N,
    2. Belavy DL,
    3. Perle SM,
    4. et al
    . AI did not write this manuscript, or did it? Can we trick the AI text detector into generated texts? The potential future of ChatGPT and AI in sports & exercise medicine manuscript generation. BMJ Open Sport Exerc Med. 2023;9:e001568.
    OpenUrlFREE Full Text
  12. 12.↵
    Integrated behavior model. Perelman School of Medicine website. https://www.med.upenn.edu/hbhe4/part2-ch4-integrated-behavior-model.shtml. Accessed March 20, 2023.
  13. 13.↵
    1. Patel BN,
    2. Rosenberg L,
    3. Willcox G,
    4. et al
    . Human–machine partnership with artificial intelligence for chest radiograph diagnosis. NPJ Digit Med. 2019;2:111.
    OpenUrl
  14. 14.↵
    1. Patel SB,
    2. Lam K
    . ChatGPT: the future of discharge summaries? Lancet Dig Health. 2023;5:E107–E108.
    OpenUrl
  15. 15.↵
    1. Yang X,
    2. Chen A,
    3. PourNejatian N,
    4. et al
    . A large language model for electronic health records. NPJ Digit Med. 2022;5:194.
    OpenUrl
  16. 16.↵
    SynGatorTron: a large clinical natural language generation model for synthetic data generation and zero-shot tasks. NVIDIA website. https://www.nvidia.com/en-us/on-demand/session/gtcspring22-s41638/. Published March 2022. Accessed March 20, 2023.
  17. 17.↵
    1. Creswell A,
    2. Shanahan M,
    3. Higgins I
    . Selection-inference: exploiting large language models for interpretable logical reasoning. arXiv website. https://arxiv.org/abs/2205.09712. Published May 19, 2022. Accessed March 20, 2023.
  18. 18.
    1. Tafjord O,
    2. Mishra BD,
    3. Clark P
    . Entailer: answering questions with faithful and truthful chains of reasoning. arXiv website. https://arxiv.org/abs/2210.12217. Published October 21, 2022. Accessed March 20, 2023.
  19. 19.↵
    1. Kazemi SM,
    2. Kim N,
    3. Bhatia D,
    4. Xu X,
    5. Ramachandran D
    . LAMBADA: backward chaining for automated reasoning in natural language. arXiv website. https://arxiv.org/abs/2212.13894. Published December 20, 2022. Accessed March 20, 2023.
  20. 20.
    1. Fink MA,
    2. Kades K,
    3. Bischoff A,
    4. et al
    . Deep learning–based assessment of oncologic outcomes from natural language processing of structured radiology reports. Radiol Artif Intell. 2022;4:e220055.
    OpenUrl
  21. 21.
    1. Biswas S
    . ChatGPT and the future of medical writing. Radiology. February 2, 2023 [Epub ahead of print].
  • Received for publication March 7, 2023.
  • Revision received March 14, 2023.
PreviousNext
Back to top

In this issue

Journal of Nuclear Medicine: 64 (5)
Journal of Nuclear Medicine
Vol. 64, Issue 5
May 1, 2023
  • Table of Contents
  • Table of Contents (PDF)
  • About the Cover
  • Index by author
  • Complete Issue (PDF)
Print
Download PDF
Article Alerts
Sign In to Email Alerts with your Email Address
Email Article

Thank you for your interest in spreading the word on Journal of Nuclear Medicine.

NOTE: We only request your email address so that the person you are recommending the page to knows that you wanted them to see it, and that it is not junk mail. We do not capture any email address.

Enter multiple addresses on separate lines or separate them with commas.
An Opinion on ChatGPT in Health Care—Written by Humans Only
(Your Name) has sent you a message from Journal of Nuclear Medicine
(Your Name) thought you would like to see the Journal of Nuclear Medicine web site.
Citation Tools
An Opinion on ChatGPT in Health Care—Written by Humans Only
Jens Kleesiek, Yonghui Wu, Gregor Stiglic, Jan Egger, Jiang Bian
Journal of Nuclear Medicine May 2023, 64 (5) 701-703; DOI: 10.2967/jnumed.123.265687

Citation Manager Formats

  • BibTeX
  • Bookends
  • EasyBib
  • EndNote (tagged)
  • EndNote 8 (xml)
  • Medlars
  • Mendeley
  • Papers
  • RefWorks Tagged
  • Ref Manager
  • RIS
  • Zotero
Share
An Opinion on ChatGPT in Health Care—Written by Humans Only
Jens Kleesiek, Yonghui Wu, Gregor Stiglic, Jan Egger, Jiang Bian
Journal of Nuclear Medicine May 2023, 64 (5) 701-703; DOI: 10.2967/jnumed.123.265687
Twitter logo Facebook logo LinkedIn logo Mendeley logo
  • Tweet Widget
  • Facebook Like
  • Google Plus One
Bookmark this article

Jump to section

  • Article
    • DISCLOSURE
    • Footnotes
    • REFERENCES
  • Figures & Data
  • Info & Metrics
  • PDF

Related Articles

  • No related articles found.
  • PubMed
  • Google Scholar

Cited By...

  • Comparative efficacy of ChatGPT 3.5, ChatGPT 4, and other large language models in gynecology and infertility research
  • Google Scholar

More in this TOC Section

  • The Australian National Total-Body PET Facility—A Shared Resource and Risk Model for Implementing Total-Body PET
  • Funding Reductions Threaten the Future of Medical Innovation
  • A Brief Report on the Results of the 2024 National Survey of Nuclear Medicine Conducted by the Chinese Society of Nuclear Medicine
Show more Editorial

Similar Articles

SNMMI

© 2025 SNMMI

Powered by HighWire