Received 2023-06-29

Revised 2023-08-22

Accepted 2023-09-16

Application of ChatGPT in Medical Content

Development: Challenges and Hope

Akhilesh Vikram Singh 1 , Anudwipa Singh 2

1 Department of Biotechnology, Graphic Era Deemed to be University, Uttarakhand, India

2 Sri Sai Jyothi College of Pharmacy, Hyderabad, Telangana, India

Dear Editor,

Recently, large language models (LLM), specifically “Chat Generative Pre-Trained Transformer” (ChatGPT), has been evidence of artificial intelligence (AI) tool that extensively perform work similar to humans through training on vast amounts of data interlinking with multi-layer recurrent neural networks [1]. As ChatGPT coding language has the potential to produce write-ups like stories and so on, it has capabilities of drafting simulations of scientific abstract writing and medical literature of free grammatical errors, unlike human researchers [2].

ChatGPT has beneficial aspects in several fields, especially academic writing and medical content development. For academic purposes, features of ChatGPT efficiently clear the queries of students; it has a narration of content for providing the prompt and conveyable meaning of text information; through generating codes, it can summarize and provide meaningful essays as well written by students, quality content for research papers, research abstracts created by ChatGPT were indistinguishable by scientists, hence, some writers may seek this tool for a logical application to get massive outputs, unlike traditional laborious manual methods to sort and analyze large volumes of text.  It has efficiently pulled the author’s details, research findings, and publication date details. It saves the time of academicians by providing required paper titles and avoids tedious article searches. Also, it is a response giver to the audible conversation to appropriate questions and provides answers; another side, it even rejects incorrect and inappropriate requests [3, 4].

ChatGPT is not a setback in medical writing; it has generated information specific to medical knowledge but also to radiological reports, patient educational information, scientific articles and regulatory documents. It drafts the essential information for clinical trial protocols, study reports and translation of medical information into various languages for comfortable reading and understanding.

This has gained significance in diagnosing critical diseases and can write a short case report of a brain tumor l extra-ventricular neurocytoma (EVN), a rare central nervous system tumor. Translational medicine is another medical field that can fill the lag between basic research and clinical practice through the application of ChatGPT by translating the clinical observations of patients to practical amendment, which improves patient medical care for better output.

Therefore, AI suits this task well as it can analyze vast data accurately within no time to identify patterns and trends that might not be apparent to humans [5, 6]. ChatGPT-AI technology has made man’s work easy but brought other ethical and unresolved concerns that a human can’t eliminate. However, users face some issues related to violation of the copyright of the content, transparency in AI-generated content used and medico-legal complications. The content credibility, inaccurate results and plagiarism are severe issues found with the medical content developed with the help of LLM tools. The results generated give biased and harmful outputs, leaving us confused about distinguishing between reliable and unreliable sources of information. As a result, the study is left as an irrelevant reference for others [7]. A recent study by Gao et al. evaluated the scientific abstracts written with ChatGPT and original abstracts with plagiarized detector websites and found that original abstracts scored higher with a median ‘plagiarized’ score of 62.5% (IQR 43.25%, 84.75%) compared to with generated abstracts with median ‘plagiarized’ score of 0% (IQR 0, 0)[8]. Similarly, comparing case reports written with ChatGPT and a human author (physician) showed that ChatGPT’s case report lacked the patient’s medical history [9].

Sam Altman, a Chief Executive Officer concerned with Open AI’s application, has highlighted that there would be a prominent threat associated with cyber-attacks and disinformation [10]. This would be observed more often at greater significance in the medical field. The training data embedded in AI systems have depicted discriminatory information and the potentiality of the existing stereotypes. By including the generative language models (GLMs) in medical education, a necessary demand is sought to address the existing potential biases. However, several past incidents tagged with Microsoft’s Tay chatbot tweeting racist and sexist content and racial biases in facial recognition technology demonstrate the need for vigilance. Further intellectual property issues, data privacy, and transparency must be eagle-eyed to ensure these tools are used responsibly. Manipulating AI-generated content may mislead medical information or, in some instances, approve illegal or unethical treatments, which would risk the life of medical students and the patient’s critical condition [11].

A two-fold risk is involved when AI-generated information is distributed unauthorizedly, and significant legal and ethical issues may be faced. On the one hand, sharing of AI-simulation models used for medical demonstrations for education purposes and also using patient information without maintaining confidentiality requires necessary copyright permissions, and without getting, it is an illegal issue and mainly causes serious problems as these actions violate privacy laws and copyright regulations; and leaves a potential concern for AI-generated medical content that was unintentionally caused a threat to ethical issues which must be followed during its training phase. These scenarios emphasize keen observation of the data governance protocols and clear usage policies when incorporating AI into medical content development [11].

Further, in the future, the implementation of AI will become essential in medical education and clinical applications, as it will attain a significant hold in clinical support tools for healthcare professionals. [GMJ.2023;12:e3106] DOI:3106

Conflict of Interest

The authors certify no conflict of interest with any financial organization regarding the material discussed in the manuscript.

Keywords

ChatGPT; Medical Content Development; AI Tool; Clinical Application; Medical Education

GMJ

Copyright© 2023, Galen Medical Journal.

This is an open-access article distributed

under the terms of the Creative Commons

Attribution 4.0 International License

(http://creativecommons.org/licenses/by/4.0/)

Email:info@gmj.ir

Correspondence to:

Akhilesh Vikram Singh, Department of Biotechnology, Graphic Era, Dehradun, 248002, Uttarakhand, India.

Telephone Number: +91-7718057352

Email Address: akhileshvikram@outlook.com

GMJ.2023;12:e3106

www.gmj.ir

Singh A, et al.

ChatGPT in Medical Content Development

2

GMJ.2023;12:e3106

www.gmj.ir

References

  1. Looi MK. Sixty seconds on ChatGPT. BMJ. 2023; 380:205.
  2. Mann DL. Artificial Intelligence Discusses the Role of Artificial Intelligence in Translational Medicine: A JACC: Basic to Translational Science Interview With ChatGPT. JACC Basic Transl Sci. 2023; 8(2):221-223.
  3. Bhatia P. ChatGPT for academic writing: A game changer or a disruptive tool? J Anaesthesiol Clin Pharmacol. 2023; 39(1):1-2.
  4. Dergaa I, Chamari K, Zmijewski P, Ben Saad H. From human writing to artificial intelligence generated text: examining the prospects and potential threats of ChatGPT in academic writing. Biol Sport. 2023; 40(2):615-622.
  5. Kitamura FC. ChatGPT is shaping the future of medical writing but still requires human judgment. Radiology. 2023;307(2):e230171.
  6. Hegde A, Srinivasan S, Menon G. Extraventricular neurocytoma of the posterior fossa: A case report written by ChatGPT. Cureus. 2023;15(3):e35850.
  7. Dave T, Athaluri SA, Singh S. ChatGPT in medicine: an overview of its applications, advantages, limitations, future prospects, and ethical considerations. Front Artif Intell. 2023;6:1169595.
  8. Gao CA, Howard FM, Markov NS, Dyer EC, Ramesh S, Luo Y, et al. Comparing scientific abstracts generated by ChatGPT to real abstracts with detectors and blinded human reviewers. NPJ Digit Med. 2023; 6(1): 75.
  9. Ye Y, Sarkar S, Bhaskar A, Tomlinson B, Monteiro O. Using ChatGPT in a clinical setting: A case report. MedComm–Future Medicine. 2023;2:e51.
  10. The Guardian. We are a little bit scared’: OpenAI CEO warns of risks of artificial intelligence. UK: The Guardian; Available from: https://www.theguardian.com/technology/2023/mar/17/openai-sam-altman-artificial-intelligence-warning-gpt4.
  11. Karabacak M, Ozkara BB, Margetis K, Wintermark M, Bisdas S. The Advent of Generative Language Models in Medical Education. JMIR Medical Education. 2023 Jun 6;9:e48163.

ChatGPT in Medical Content Development

Singh A, et al.

GMJ.2023;12:e3106

www.gmj.ir

3