Skip to main content
Opinion Papers

Leveraging Large Language Models to Streamline Clinical Trial Data Administration: Balancing Efficiency with Ethical Responsibility

Authors: Amrita Ghosh orcid logo (Stanford University) , Caroline J Huang orcid logo (Stanford University)

  • Leveraging Large Language Models to Streamline Clinical Trial Data Administration: Balancing Efficiency with Ethical Responsibility

    Opinion Papers

    Leveraging Large Language Models to Streamline Clinical Trial Data Administration: Balancing Efficiency with Ethical Responsibility

    Authors: ,

Abstract

Clinical trials are critical for advancing medical knowledge and developing new treatments, yet they often involve substantial administrative burdens that can impede progress and reduce efficiency. Effective clinical trial data management requires seamless participant engagement, accurate data collection, strict regulatory compliance, and coordinated efforts among multidisciplinary teams. These demands can strain resources and lead to workflow inefficiencies.

This research paper explores the potential of large language models (LLMs), such as ChatGPT, to enhance clinical coordination by optimizing administrative processes in clinical trials. We assess the model's applications in streamlining standard operating procedures (SOPs), clinical data management, automating documentation, and supporting regulatory compliance. To ensure responsible implementation, we also examine key challenges related to ethical considerations, data biases, and safety concerns, while proposing strategies for mitigating these risks.

Our findings indicate that integrating AI-driven solutions like ChatGPT can significantly improve operational efficiency, reduce administrative workloads, and allow clinical trial teams to dedicate more time to patient care and scientific inquiry. By leveraging AI responsibly, we can make clinical research more agile, adaptive, and focused on advancing medical innovation.

How to Cite:

Ghosh, A. & Huang, C. J., (2025) “Leveraging Large Language Models to Streamline Clinical Trial Data Administration: Balancing Efficiency with Ethical Responsibility”, Journal of the Society for Clinical Data Management 1. doi: https://doi.org/10.47912/jscdm.438

131 Views

19 Downloads

Published on
30 Oct 2025
Peer Reviewed

1. Introduction

Clinical trials are essential for advancing medical science, yet they impose considerable administrative burdens on clinical coordination efforts. Managing participant engagement, regulatory compliance, data collection, and documentation requires seamless collaboration among clinical research staff, investigators, and regulatory teams. These challenges often result in delays, increased errors, and operational inefficiencies, ultimately impacting trial timelines and data integrity.

The integration of artificial intelligence (AI) tools and large language models (LLMs) presents an opportunity to streamline administrative tasks, optimize workflows, and enhance overall clinical coordination. Approximately 900 articles were found on PubMed mentioning “ChatGPT,” with most of the articles being published in the past two years, and the scientific community has expressed great interest in using LLMs for research purposes.1,2 It is apparent that LLMs are being vastly utilized in the scientific community. In this paper, we examine methods of using LLMs to assist in drafting study documents, automating data management, and ensuring regulatory compliance, thereby allowing clinical teams to focus on patient safety, trial oversight, and scientific rigor.

2. Administrative Challenges in Clinical Trials

The complexity of modern clinical trials places significant pressure on clinical coordination research teams. A 2022 study reported that approximately 61% of clinical researchers experienced signs of burnout,3 which is defined by the World Health Organization as exhaustion, increased mental distance from one’s job, and reduced professional efficacy.4 66% of clinical researchers believed that investing in technology could mitigate administrative burdens.3 Administrative duties, which include drafting informed consent forms, managing regulatory documentation, developing training materials, and preparing grant proposals, require significant time and effort.

In 2023, a systematic review examined ChatGPT as an example of an LLM in the context of healthcare education and research.5 This review included a total of 60 records, in which 85% of the records cited the benefits of using ChatGPT in healthcare and research by improving scientific writing, enhancing research equity, highlighting ChatGPT’s utility in healthcare research, and providing benefits in healthcare education.5 When these responsibilities are inefficiently managed, it may compromise trial completion, data quality, and participant retention. Effective clinical coordination is critical to ensuring seamless communication among stakeholders, compliance with ethical and regulatory requirements, and efficient data processing. Given these challenges, leveraging LLMs to streamline administrative processes is a logical step in improving the efficiency of clinical trials.

3. Applications of LLMs in Clinical Trials

LLMs are trained using reinforcement learning from human feedback, enabling them to generate structured, context-specific, and user-friendly responses.6 Unlike traditional search engines that deliver vast amounts of information based on keyword searches, LLMs provide tailored answers in the desired format, making them particularly useful for managing time-intensive administrative tasks in clinical trials. Traditional clinical workflow tools such as Electronic Data Capture (EDC) platforms and Clinical Trial Management Systems (CTMS) efficiently manage various aspects of clinical trials but focus on oversight and data collection, while LLMs have capabilities that extend beyond these functions and reduce the administrative burden.

LLMs are not substitutes for professional expertise and must be verified against authoritative sources. They serve as a valuable assistant in clinical coordination by reducing administrative workload and improving efficiency in key trial processes. Table 1 compares LLM features that are useful in reducing the researchers’ workload in comparison to the traditional clinical workflow tools.

Table 1

LLMs vs. Traditional Workflow Solutions: Reducing Clinical Coordination Burden.

Feature LLMs Traditional Workflow Solutions
Electronic Data Capture (EDC) platforms Clinical Trial Management Systems (CTMS)
Real-time natural language interaction
  • Human-like responses, customizable for different workflows

  • No conversational capability

  • No conversational capability

Document drafting
  • Fast, accurate, context-aware drafting

  • Not a core feature

  • Mostly requires manual data entry

  • Requires manual data entry

Data validation
  • Requires prompt customization for structured validation

  • Needs to be validated by a human

  • Built-in validation rules and logic

  • May have advanced data validation workflows

Plain language integration (conversion of medical jargon)
  • Converts medical jargon into patient-friendly language when paired with the right prompt

  • Typically, heavy with medical terms

  • Limited plain language outputs

Set-up and training
  • No setup and training needed*

  • Requires user onboarding and role-based training

  • Requires technical training and customization

Cost
  • Free version available

  • The institution should purchase and provide it

  • The institution should purchase and provide it

  • *Prompt training may be needed, but is optional.

3.1 Key Areas of Clinical Coordination Supported by LLMs

Documentation and Consent Forms

Documentation is the backbone of clinical trials, with consent forms being among the most crucial documents in the clinical research process. Researchers at LifeSpan Healthcare System (LHS) analyzed 798 federally funded clinical trial consent forms and found that dropout rates increased by almost 16% due to the high-grade reading level of the documents. To address this issue, the research team utilized ChatGPT-4 as an example of an LLM to transform a surgical consent form from a 12.6 Flesch-Kincaid reading level to a 6.7 reading level.7 We have seen how the ability of LLMs to simplify complex medical terminology and generate clear, patient-friendly consent forms makes clinical information more accessible, promoting informed decision-making.

Beyond consent forms, LLMs can support various documentation tasks, including summarizing data from research articles, clinical records, and patient records. For instance, a cardiovascular study demonstrated the potential of LLMs to generate structured data from unstructured text, showcasing their capability to enhance data management processes.8 By reframing technical jargon and making medical information more understandable, LLMs help reduce barriers and facilitate better patient engagement.

Regulatory Compliance

LLMs can significantly support regulatory compliance by assisting in drafting and reviewing documents, as well as summarizing guidelines from governmental agencies such as the US Food and Drug Administration (FDA). In this context, LLMs can generate detailed descriptions of medical devices, risk management plans, general safety and performance requirements checklists, and technical reports.9

One of the strengths of LLMs lies in generating standardized templates for Institutional Review Board (IRB) submissions, minimizing redundant efforts, and saving time. Furthermore, with constantly evolving regulations, LLMs can help keep documentation up to date by scanning files for compliance and indicating specific regulatory guidelines present. For instance, when drafting sections on drug risks and side effects, specific LLM prompts can identify areas that may require clearer explanations or more prominent disclosures, thereby promoting patient safety.10

Recruitment Strategies

Recruitment tasks in clinical trials often involve crafting targeted materials that resonate with specific demographics while being culturally sensitive and engaging. LLMs can help reduce the recruitment burden by generating such materials efficiently.

A notable example is the National Institutes of Health’s development of TrialGPT, an AI algorithm designed to match potential volunteers to clinical trials by identifying relevant trials for which individuals are eligible. This technology enhances outreach efforts and has the potential to improve participant enrollment rates by making recruitment strategies more targeted and accessible.11

Grant Writing and Funding Applications

Grant writing is a critical yet time-consuming aspect of clinical research. LLMs can expedite the development of grant proposals by structuring content, aligning it with funding agency requirements, and refining scientific arguments. This allows researchers to concentrate on core study design aspects.

LLMs can also assist in conducting literature reviews by automating the extraction and summarization of relevant information from extensive collections of scientific articles and publications.12 Additionally, when developing grant budgets, LLMs can help identify potential expenses, distinguish between direct and indirect costs, and suggest appropriate buffers for unforeseen expenditures.13 By supporting researchers in articulating evaluation methods and sustainability strategies, LLMs can make grant proposals more robust and comprehensive.

Training and Onboarding Support

Training and onboarding clinical staff efficiently are essential for maintaining high standards in clinical trials. LLMs can generate training materials, including FAQs, compliance checklists, and interactive learning tools. Studies have shown that using LLMs as a learning assistant can enhance the acquisition of medical terminology, and can provide adaptive, personalized training experiences.14

By generating diverse scenarios quickly and cost-effectively, LLMs accommodate specific learning objectives and skill development needs. These training approaches are particularly suitable for individual, group, or remote training sessions. Furthermore, standardized training resources help maintain consistency across study sites, while LLM-based simulations provide a more affordable and scalable alternative to traditional training methods.15

3.2 Prompt Generation

The effectiveness of LLMs largely depends on how prompts are written and structured. Prompts can be questions, statements, or combinations of both that are designed to generate information, reframe data, or analyze content. LLM responses are based on language patterns learned during pre-training and user interactions.

However, LLMs are sensitive to subtle changes in prompt phrasing, which can result in different outputs.6 To generate accurate and useful responses—especially for tasks like regulatory guidance, protocol summaries, or document templates—prompts must be clear and precise. Table 2 provides examples of optimized prompts tailored for clinical coordination tasks.

Table 2

LLM Prompt Generation Guide for Clinical Studies.

Prompt Requirement Prompt Example
Define the research question and objective Example: “Explain how a randomized controlled trial (RCT) can be used to compare two COVID-19 vaccines. Focus on safety outcomes like adverse events and include trial design considerations.”
Define the area and agencies the information focuses on Example: “What FDA regulatory guidelines should be followed when designing a Phase III clinical trial for a new antiviral drug?”
Define boundaries Example: “Summarize the safety data from COVID-19 vaccine clinical trials conducted between January 2020 and December 2021, focusing on Phase III trials in adults aged 18–65. Limit the summary to studies conducted in North America and Europe.”
Define writing style Example: “Explain the safety profile of the Pfizer COVID-19 vaccine to a non-medical audience in bullet points.”
Do not enter patient health information (PHI) Example: “Provide an example of a treatment plan for an adult male in his 40s diagnosed with Type 2 diabetes in 2015, considering general treatment guidelines.”
Keep it focused but flexible Example: “Discuss the differences in adverse event profiles between mRNA and adenovirus-based COVID-19 vaccines, focusing on the general adult population.”

By incorporating LLMs into these processes, clinical coordination teams can improve workflow efficiency, reduce administrative errors, and allocate more time to patient care and trial oversight.

4. Ethical Challenges of LLMs

While LLMs offer promising benefits for clinical coordination, their implementation raises several ethical and practical concerns. Ensuring data integrity, regulatory compliance, and ethical AI use is critical for maintaining trust among trial stakeholders.

4.1 Key Challenges and Mitigation Strategies for Ethical Considerations of LLMs

Accuracy and Reliability

LLMs generate responses based on their training data, which can occasionally lead to “hallucinations,” which are factually incorrect statements.16 Hallucinations cause LLMs to frequently generate non-existent or incorrect academic references, which is a particularly concerning issue in scientific and medical contexts.16 For example, evaluations of LLMs in a study examining urinary tract infection diagnosis and management found that it fabricated 12 of 24 provided citations, creating entirely fictional medical references that appeared credible.17

Other studies evaluating LLMs’ responses to common medical questions found numerous instances where the AI provided incorrect drug information or failed to mention important contraindications.18 Users should be wary of information that contradicts common knowledge, is unsupported by citations, or source material.19 Thus, human oversight is essential to validate AI-generated regulatory guidance and clinical documentation.

Data Privacy and Security

LLMs are not inherently compliant with privacy regulations such as the Health Insurance Portability and Accountability Act (HIPAA) or General Data Protection Regulation (GDPR). LLMs are susceptible to jailbreaking, which is caused when the model’s built-in safeguards are bypassed.20 In medical research, jailbreaking causes a confidentiality breach.20 Organizations must establish protocols to anonymize sensitive data before using AI tools in clinical workflows.

Bias in AI Output

Since LLMs learn from a diverse range of data sources, inherent biases may be present. As LLMs are trained from generalized sources, they may not provide accurate region and population-based information. Moreover, there is a black box issue associated with LLMs like ChatGPT, in which the decision-making processes are opaque, and the logic that produces the output is hidden.21 The combination of sampling and exclusion biases in LLMs has significant implications for clinical administration, potentially impacting resource allocation, clinical guideline development, and healthcare access.22 Periodic audits and validation processes can help identify and mitigate these biases.6

Consistency in AI Responses

LLMs’ output can vary depending on prompt phrasing. Different prompts cause different outputs. Erroneous outputs may contain incorrect usage of medical terms, errors in diagnosis/treatment/management, and provide irrelevant information. For example, although there is evidence that the ChatGPT 4.0 model version shows better information processing capabilities than the 3.5 model version, variability in outputs can lead to errors in clinical contexts, risking patient safety and causing disruptions in clinical tasks due to unreliable recommendations.23 Standardizing prompts and evaluating response consistency across different iterations can improve reliability.

Ethical Use and Oversight

AI tools should complement, not replace, human expertise. Their roles should be limited to providing support for routine tasks, with all critical decisions remaining the responsibility of qualified professionals. Establishing Standard Operating Procedures (SOPs) for AI use in clinical trials can help define appropriate applications and boundaries while ensuring alignment with ethical and regulatory standards.

Organizations integrating LLMs into clinical coordination processes must implement risk assessment frameworks to monitor AI use and mitigate potential issues. Figure 1 outlines the recommended steps for risk assessment, mitigation, and monitoring.

Figure 1
Figure 1

Risk Analysis Framework for LLMS in Clinical Trials.

5. Risk Analysis of LLMs

The integration of LLMs into clinical study tasks presents both opportunities and challenges. While the potential for enhancing efficiency and reducing administrative burdens is promising, the use of AI-driven tools also carries inherent risks that must be carefully managed. This section outlines a comprehensive risk analysis framework, addressing key considerations such as human oversight, data integrity, regulatory compliance, and continuous monitoring to safeguard patient welfare and clinical accuracy (Figure 1).

Risk Assessment

  • Nature of the risk:

    Analyze the nature of each identified risk and assess how frequently they are likely to occur when using LLMs for different tasks.

    • Example: When generating a report or summary from medical data, the primary risks may involve inconsistencies and hallucinations. Similarly, when using LLMs to suggest patient communication, the risk lies in using technical jargon or an inappropriate tone. Adjusting the prompt can help mitigate these risks.24

  • Impact of the risk:

    Evaluate the potential consequences if these risks materialize and establish procedures to address such a situation if it arises.

    • Example: The impact of a data privacy breach can be severe, resulting in legal repercussions and reputational damage.

  • Risk-sensitive prompts:

    Identify the most sensitive prompts within clinical studies.

    • Example: Interpreting study results requires precision, as errors can have significant consequences. In these cases, physician oversight is essential.

Risk Mitigation

  • Human oversight:

    LLMs should be considered a tool to assist clinical research coordinators, not a replacement for medical professionals. Qualified professionals must review any content generated by ChatGPT, as the success of a clinical study depends on clinical decision-making and accurate data interpretation.

    • Context understanding: AI tools like LLMs may not fully comprehend the scientific context or the intricate medical complexities involved.24 Human oversight ensures that the content addresses the intended audience accurately.

    • Accuracy assurance: As noted in the challenges section of this paper, ChatGPT can produce inaccurate or nonsensical data due to hallucinations.19 Human reviewers are needed to validate data credibility and relevance.

    • Structural integrity: While prompt engineering can control language, tone, and text format to an extent, human oversight ensures the response meets regulatory standards and specific requirements.

  • Data anonymization:

    Before inputting data into LLMs, remove all patient-identifiable information to comply with privacy regulations. The responsibility for this lies with the individual structuring and inputting the prompt.11

  • Validation:

    Implement mechanisms to validate the accuracy of LLM outputs by comparing them to established clinical guidelines, literature, or expert opinions. This task should be performed by data management personnel or clinical managers.

    • Prompt evaluation: Develop a scoring guideline from test prompts to categorize responses as pass, fail, inappropriate, or need human assistance.

    • Bias audits: Regularly audit the model’s outputs to check for biases, especially when working with diverse patient populations or varied study data. 20

  • User access controls:

    Restrict access to LLMs by assigning specific roles to users and monitoring usage to prevent unauthorized data input or sharing.20

  • Adherence to regulations:

    Ensure that LLM usage aligns with clinical research regulatory frameworks. Engage with regulatory bodies such as the FDA and IRB to clarify acceptable AI usage in clinical trials.9

  • Training and guidelines:

    Provide clear guidelines and SOPs for clinical coordinators on the appropriate use of LLMs. Clearly specify which tasks it can assist with and which require human intervention.

Risk Monitoring

  • Incident reporting:

    Establish a clear Corrective and Preventive Action (CAPA) process for reporting any errors, inaccuracies, or data breaches related to LLM involvement in clinical study tasks.20

  • Audits:

    Conduct periodic audits performed by data management or clinical research managers to ensure LLM usage complies with both internal policies and external regulations.

  • Feedback loop:

    Collect feedback from clinical research coordinators on their experiences using LLMs, including any errors or concerns. Use this feedback to update SOPs and improve guidelines regularly.

6. An Organizational Framework for Implementation of LLMs

To maximize the benefits of LLMs while ensuring responsible use, organizations should adopt a structured approach:

Develop SOPs

SOPs should define appropriate AI applications and establish validation protocols for maintaining data privacy and regulatory compliance. Organizations should develop policies to ensure patient privacy when using LLMs, which may require going beyond existing regulatory requirements, such as HIPAA.

Train Clinical Staff

Clinical teams should be trained to utilize AI-based tools and to identify tasks that require human oversight. Organizations should engage healthcare educators and professionals in the development and review process to enhance the quality and accuracy of AI-generated training content.15

Implement Continuous Monitoring and Audits

Regular performance audits and feedback loops should be established to refine the integration of LLMs or AI tools into clinical workflows.

Engage with Regulatory Bodies

Researchers must acknowledge any use of AI-assisted technologies, including LLMs, in research and manuscript preparation, typically in the Methods or Acknowledgments section.12, 25 This acknowledgment will enable agencies like the FDA and NIH to clarify best practices for AI use in clinical research and ensure compliance with industry regulations. For example, scientific journals such as Nature have added LLM-use rules in their author guide, including such directions as no LLMs will be accepted as a credited author on a research paper, and researchers using LLMs should document this in the methods or acknowledgments section.26

7. The Future of AI in Clinical Research

The adoption of LLMs in clinical trials reflects a broader trend toward integrating AI into healthcare. Numerous studies have demonstrated that models like ChatGPT-4 exhibit remarkable proficiency in medical knowledge tasks.27 These models have shown varying levels of accuracy across diverse topics, suggesting potential applications in specific administrative domains where their knowledge base is most robust.27 Additionally, LLMs have proven highly effective as a tool for academic research, prompting institutions such as the University of Michigan, Harvard University, Washington University, the University of California-Irvine, New York University, and Stanford to develop their own versions of LLMs.28 This trend is likely to continue as AI technologies advance.

However, in recent years, safety concerns regarding LLMs have emerged, including significant legal challenges when ChatGPT was banned in Italy. In 2023, the Italian Data Protection Authority accused OpenAI of violating EU data protection rules.29 Despite these challenges, the future points toward the global inclusion of ChatGPT and other LLMs within regulatory frameworks. As AI continues to evolve in healthcare settings, the development of robust frameworks for responsible use will become increasingly vital across administrative functions.

Several regulations have been established to address the rapid growth of AI while ensuring stringent controls on its use, including the European Union’s GDPR,30 California’s Health Care Services: Artificial Intelligence Act,31 and China’s Personal Information Protection Law.32 These frameworks aim to strike a balance between innovation and safeguarding data privacy. As Robert M. Califf, former Commissioner of the US FDA, emphasized, regulating large language models is crucial to harnessing their potential while mitigating risks.33 Furthermore, the European Medicines Regulatory Network envisions AI systems functioning as personal assistants to support users with daily workplace tasks, significantly enhancing productivity while maintaining compliance with data protection legislation.34

The ability of AI-based tools, such as LLMs, to improve clinical coordination marks a critical step toward more advanced applications, including AI-assisted patient monitoring and personalized medicine. As the field progresses, responsible innovation and regulatory vigilance will be paramount to realizing the full potential of AI in healthcare.

8. Conclusion

LLMs present a promising solution for addressing administrative burdens in clinical coordination, including streamlining documentation, regulatory compliance, and patient engagement. An example of this is an exploratory case study that assessed the quality of radiology reports simplified using ChatGPT.35 In the study, ChatGPT was prompted with the instruction: “Explain this medical report to a child using simple language.” The simplified radiology reports were subsequently evaluated by 15 radiologists, who rated their quality based on factual correctness, completeness, and the potential for patient harm. The results indicated that most reports were complete and factually accurate, with no identified potential for patient harm. However, there were instances in which potential harm could be caused by the reports’ inclusion of incorrect data or the omission of pertinent medical information.35

These findings underscore the need for careful planning, robust oversight, and strict adherence to ethical standards when integrating LLMs into clinical administration. Healthcare organizations must strike a balance between leveraging the capabilities of AI and implementing appropriate safeguards to prevent biases from undermining administrative decision-making, compromising patient care, or exacerbating health disparities.36

By implementing AI responsibly, clinical teams can enhance efficiency, reduce errors, and improve trial outcomes. As AI technologies continue to evolve, LLMs will undoubtedly play an increasingly pivotal role in shaping the future of clinical research and reducing administrative burdens.

Acknowledgements

The authors acknowledge the use of LLMs, such as ChatGPT, to assist in refining research prompts. All content was subsequently reviewed, verified, and edited by the authors for accuracy and appropriateness.

Competing Interests

The authors have no competing interests to declare.

Author contributions

The authors contributed equally as co-first authors.

References

1. Grillo R. The rising tide of artificial intelligence in scientific journals: A profound shift in research landscape. European Journal of Therapeutics. 2023; 29(3):686–688. DOI:  http://doi.org/10.58600/eurjther1735

2. Khalifa AA, Ibrahim M. Artificial intelligence (AI) and ChatGPT involvement in scientific and medical writing: A new concern for researchers. A scoping review. Arab Gulf Journal of Scientific Research. 2024; 42(4):1770–1787. DOI:  http://doi.org/10.1108/AGJSR-09-2023-0423

3. OpenClinica. Clinical trial researchers are burned out, too. OpenClinica Blog; 2022, April 28. https://www.openclinica.com/blog/clinical-trial-researchers-are-burned-out-too/ Accessed September 26, 2024.

4. World Health Organization. Burn-out an “occupational phenomenon”: International Classification of Diseases. WHO Newsroom; 2019, May 28. https://www.who.int/news/item/28-05-2019-burn-out-an-occupational-phenomenon-international-classification-of-diseases Accessed September 26, 2024.

5. Sallam M. ChatGPT utility in healthcare education, research, and practice: Systematic review on the promising perspectives and valid concerns. Healthcare. 2023; 11(6):887. DOI:  http://doi.org/10.3390/healthcare11060887

6. OpenAI. ChatGPT. https://openai.com/index/chatgpt/ Accessed September 26, 2024.

7. Mirza FN, Tang OY, Connolly ID, et al. Using ChatGPT to facilitate truly informed medical consent. NEJM AI. 2024; 1(2). DOI:  http://doi.org/10.1056/AIcs2300145

8. Moons P, Van Bulck L. ChatGPT: Can artificial intelligence language models be of value for cardiovascular nurses and allied health professionals. European Journal of Cardiovascular Nursing. 2023; 22(7):e55–e59. DOI:  http://doi.org/10.1093/eurjcn/zvad022

9. Di Bello F, Russo E, Sartori M. Enhancing regulatory affairs in the market placing of new medical devices: How LLMs like ChatGPT may support and simplify processes. AboutOpen. 2024; 11(1):77–88. DOI:  http://doi.org/10.33393/ao.2024.3302

10. GlobalVision. 10 ChatGPT prompts to enhance pharmaceutical proofreading. GlobalVision Blog; 2025, March 25. https://www.globalvision.co/blog/10-chatgpt-prompts-to-enhance-pharmaceutical-proofreading Accessed March 30, 2025.

11. Jin Q, Wang Z, Floudas CS, et al. Matching patients to clinical trials with large language models. Nature Communications. 2024; 15:9074. DOI:  http://doi.org/10.1038/s41467-024-53081-z

12. Lee PY, Salim H, Abdullah A, Teo CH. Use of ChatGPT in medical research and scientific writing. Malaysian Family Physician. 2023; 18:58. DOI:  http://doi.org/10.51866/cm0006

13. Morrow SP. 3 examples of grant budgets that will win over funders (with examples). Instrumentl Blog; 2024. https://www.instrumentl.com/blog/grant-budget-examples Accessed March 30, 2025.

14. Hsu M-H. Mastering medical terminology with ChatGPT and Termbot. Health Education Journal. 2024; 83(4):352–358. DOI:  http://doi.org/10.1177/00178969231197371

15. Tully MJ. Revolutionizing healthcare simulation: ChatGPT’s impact and potential. Healthy Simulation; 2023. https://www.healthysimulation.com/healthcare-simulation-chatgpt/ Accessed March 15, 2025.

16. Masters K. Medical Teacher’s first ChatGPT’s referencing hallucinations: Lessons for editors, reviewers, and teachers. Medical Teacher. 2023; 45(7):673–675. DOI:  http://doi.org/10.1080/0142159X.2023.2208731

17. Gupta K, O’Brien W, Strymish J. 104. Accuracy of ChatGPT for UTI diagnosis and management questions. Open Forum Infectious Diseases. 2023; 10(Suppl 2):ofad500.020. DOI:  http://doi.org/10.1093/ofid/ofad500.020

18. Gravina AG, Pellegrino R, Cipullo M, et al. May ChatGPT be a tool producing medical information for common inflammatory bowel disease patients’ questions? An evidence-controlled analysis. World Journal of Gastroenterology. 2024; 30(1):17–33. DOI:  http://doi.org/10.3748/wjg.v30.i1.17

19. Alkaissi H, McFarlane SI. Artificial hallucinations in ChatGPT: Implications in scientific writing. Cureus. 2023; 15(2):e35179. DOI:  http://doi.org/10.7759/cureus.35179

20. Zhang H, Lou Q, Wang Y. Towards safe AI clinicians: A comprehensive study on large language model jailbreaking in healthcare. arXiv Preprint; 2025. https://arxiv.org/abs/2501.18632 Accessed April 20, 2025.

21. Schopow N, Osterhoff G, Baur D. Applications of the natural language processing tool ChatGPT in clinical practice: Comparative study and augmented systematic review. JMIR Medical Informatics. 2023; 11: e48933. DOI:  http://doi.org/10.2196/48933

22. Adam H, Balagopalan A, Alsentzer E, et al. Mitigating the impact of biased artificial intelligence in emergency decision-making. Communications Medicine. 2022; 2:149. DOI:  http://doi.org/10.1038/s43856-022-00214-4

23. Mu L J, Wang TT, Miao YD. Advancements in AI-driven oncology: Assessing ChatGPT’s impact from GPT-3.5 to GPT-4o. International Journal of Surgery. 2025; 111(1):1669–1670. DOI:  http://doi.org/10.1097/JS9.0000000000001989

24. Biassoni F, Gnerre M. Exploring ChatGPT’s communication behaviour in healthcare interactions: A psycholinguistic perspective. Patient Education and Counseling. 2025; 134:108663. DOI:  http://doi.org/10.1016/j.pec.2025.108663

25. Ganjavi C, Eppler MB, Pekcan A, Biedermann B, Abreu A, Collins GS, Gill IS, Cacciamani GE. Publishers’ and journals’ instructions to authors on use of generative artificial intelligence in academic and scientific publishing: Bibliometric analysis. BMJ. 2024; 384:e077192. DOI:  http://doi.org/10.1136/bmj-2023-077192

26. Nature Editorial. Tools such as ChatGPT threaten transparent science; Here are our ground rules for their use. Nature. 2023; 613(7945):612. DOI:  http://doi.org/10.1038/d41586-023-00191-1

27. Angel MC, Rinehart JB, Cannesson MP, Baldi P. Clinical knowledge and reasoning abilities of AI large language models in anesthesiology: A comparative study on the American Board of Anesthesiology examination. Anesthesia & Analgesia. 2024; 139(2):349–356. DOI:  http://doi.org/10.1213/ANE.0000000000006892

28. Coffey L. Universities build their own ChatGPT-like AI tools. Inside Higher Ed; 2024, March 21. https://www.insidehighered.com/news/tech-innovation/artificial-intelligence/2024/03/21/universities-build-their-own-chatgpt-ai Accessed April 20, 2025.

29. Vincent J. OpenAI’s regulatory troubles are just beginning. The Verge; 2023, May 5. https://www.theverge.com/2023/5/5/23709833/openai-chatgpt-gdpr-ai-regulation-europe-eu-italy Accessed April 20, 2025.

30. European Union. General Data Protection Regulation (GDPR). EUR-Lex. https://eur-lex.europa.eu/EN/legal-content/summary/general-data-protection-regulation-gdpr.html Accessed April 20, 2025.

31. California Legislative Information. AB-3030 Health care services: Artificial intelligence. https://leginfo.legislature.ca.gov/ Accessed April 20, 2025.

32. Personal Information Protection Law. Personal Information Protection Law (PIPL). https://personalinformationprotectionlaw.com/ Accessed April 20, 2025.

33. U.S. Food and Drug Administration. Speech by Robert M. Califf, M.D. at the National Health Council’s 2023 Science of Patient Engagement Symposium: The patient at the center; 2023, May 9. https://www.fda.gov/news-events/speeches-fda-officials/speech-robert-m-califf-md-national-health-councils-2023-science-patient-engagement-symposium-patient Accessed September 26, 2024.

34. European Medicines Agency. Artificial intelligence: Workplan to guide use of AI in medicines regulation. EMA News; 2023, December 18. https://www.ema.europa.eu/en/news/artificial-intelligence-workplan-guide-use-ai-medicines-regulation Accessed April 20, 2025.

35. Jeblick K, Schachtner B, Dexl J, et al. ChatGPT makes medicine easy to swallow: An exploratory case study on simplified radiology reports. Eur Radiol. 2024; 34:2817–2825. DOI:  http://doi.org/10.1007/s00330-023-10213-1

36. Nair A, Dsouza KD, Reshmi B. Artificial intelligence & sustainable development goals in health care. Indian Journal of Community Medicine. 2024; 49(Suppl 1):S121. DOI:  http://doi.org/10.4103/ijcm.ijcm_abstract422