Open access peer-reviewed chapter - ONLINE FIRST

Artificial Intelligence in Healthcare: Considerations for Adoption and Adaptation in Academic Medical Settings

Written By

Jacob A. Gould, Stanislaw P. Stawicki, Ryan Yimeng Lee and Anna Ng-Pellegrino

Submitted: 05 June 2024 Reviewed: 08 August 2024 Published: 16 September 2024

DOI: 10.5772/intechopen.115397

Artificial Intelligence in Medicine and Surgery - An Exploration of Current Trends, Potential Opportunities, and Evolving Threats - Volume 2 IntechOpen
Artificial Intelligence in Medicine and Surgery - An Exploration ... Edited by Stanislaw P. Stawicki

From the Edited Volume

Artificial Intelligence in Medicine and Surgery - An Exploration of Current Trends, Potential Opportunities, and Evolving Threats - Volume 2 [Working Title]

Dr. Stanislaw P. Stawicki

Chapter metrics overview

10 Chapter Downloads

View Full Metrics

Abstract

Discussions around artificial intelligence (AI) and machine learning (ML) and their applicability within academic medicine have become prominent over the past several years. Various end-user-focused AI/ML tools have emerged, offering opportunities to enhance efficiency and improve outcomes in biomedical research and medical education. While AI holds the promise of revolutionizing many aspects of academic medicine, the gravitas of the medical field necessitates scrupulous consideration and forward planning when implementing AI/ML in medical settings. Consequently, frameworks to guide AI/ML implementation discussions within academic medicine are crucial for mitigating the inherent pitfalls of such technology. This chapter proposes a framework to assist decision-makers in the academic medicine ecosystem with AI/ML implementation decisions. The framework emphasizes [A] understanding the functionality of different types of AI (Large Language Models, Computer Vision, and Omics Learning Models) to identify inherent use cases and limitations; [B] considering regulatory constraints and ethical principles specific to the implementation context; and [C] evaluating the overall costs and benefits of AI/ML implementation. Proactively balancing innovation with human oversight is essential to leveraging AI’s benefits while mitigating risks. As AI in healthcare evolves, ongoing research, collaboration, and regulations will be vital to ensure AI is aligned with the goal of advancing healthcare responsibly.

Keywords

  • academic medicine
  • artificial intelligence
  • machine learning
  • large language model
  • computer vision model
  • OMNI specific model
  • healthcare
  • medical education
  • medical AI implementation framework
  • pitfalls of medical AI

1. Introduction

The public unveiling of end-user-focused applications of artificial intelligence (AI) such as ChatGPT (generative pre-training transformer) has been followed by significant discussion around AI technologies and their potential role and impact [1]. Potential benefits of AI applications have now been extensively explored, from redefining the dynamic of educational classrooms to facilitating individualized gene sequencing [2, 3]. At the same time, the rapid and somewhat unchecked advancement of AI has been met with calls for regulatory framework creation to help address potential pitfalls and disruptive socio-economic externalities related to the widespread adoption of this technology [4, 5]. Potential implementations of AI in healthcare must satisfy a value proposition that addresses relevant challenges facing the field today, ranging from clinical practice to academic research to medical education [6, 7]. The sensitive social foundation and gravitas that underpin health professions present a need for heightened considerations around the adoption of AI in healthcare. Aforesaid, there is a need for an overview of the appropriate and inappropriate use cases of AI in healthcare in order to optimize the benefits of this technology in a manner that is safe, ethical, and responsible [8, 9]. Here, we provide a general overview of the current use cases of AI in clinical, academic, and educational settings within the healthcare industry alongside a framework to facilitate the assessment of pitfalls and benefits of AI implementations.

1.1 The economic value proposition of artificial intelligence in healthcare

Before analyzing the potential use cases and pitfalls of AI in academic medicine, it is important to first understand the underlying factors driving AI adoption in healthcare. The US healthcare market presents a unique opportunity for AI integration as a means of value creation through increased efficiency, safety, and productivity [6, 1011]. According to the Centers for Medicare and Medicaid Services (CMS), National Healthcare Expenditure (NHE) in the US totaled nearly $4.3 trillion in 2021, accounting for approximately 18.3% of US gross domestic product (GDP). Looking ahead, NHE is projected to grow annually at 5.6%, outpacing anticipated 4.6% annual GDP growth over the period of 2022–2031 [12]. Moreover, the market for healthcare products and services faces many challenges and inefficiencies, including, but not limited to a geographically imbalanced distribution of medical facilities and providers, workforce shortages and associated provider burnout, and increasing costs of medical education [13, 14]. In this context, the functional capability of AI to perform tasks that are perceived as menial and/or repetitive tasks may help address many of the aforementioned challenges [3]. In fact, widespread adoption of AI is estimated to be associated with NHE savings of 5-10% annually [6]. Consequently, significant opportunities exist to realize untapped value within the existing healthcare paradigm, with AI being among the most promising and paradigm-changing developments [15].

The overall value proposition presented by the integration of AI into academic healthcare has drawn an influx of dedicated resources, spurring significant new development and investment [16, 17]. This led to the creation of AI platforms that are academic-centric, with emphasis on streamlining processes and improving efficiencies related to academic medicine environments [7, 18, 19, 20]. Specific domains included in this innovation push include medical education, clinical productivity, scholarly publishing, and other related platforms [18, 19, 20, 21, 22]. The corresponding amount of private investment into AI has been especially generous in the healthcare sector when compared to other industry verticals [Figure 1] [16]. Still, important questions persist, including the question of AI’s ability to provide levels of transparency inherently required across our healthcare settings [23].

Figure 1.

Global private investment in artificial intelligence by industry vertical [16]. Note: A comparison of total global private investment spent on artificial intelligence across sectors demonstrates that healthcare-focused artificial intelligence applications received the most total private investment in 2022 among various sectors at $6.1 billion.

Advertisement

2. Basic considerations and definitions

To individuals not well-versed with the technology, AI nomenclature may seem both overwhelming and somewhat confusing. While the terms “artificial intelligence,” “machine learning,” and “large language models” are often used interchangeably in colloquial speech, there are precise intersections and distinctions between these forms of technology. Understanding the relationships between these terms can be useful in avoiding conflation and misnomers.

2.1 Artificial intelligence, machine learning, and large language models: Intersections and distinctions

Since the public launch of Open AI’s ChatGPT (San Francisco, California, USA) in late 2022, and the widespread public exposure to AI, Google ™ (Alphabet, Inc., Mountain View, California, USA) searches for the term “artificial intelligence” increased by nearly 250% from ChatGPT’s launch to April 16, 2023 [24]. Likewise, the term “large language model” increased by nearly 592% during the same time period [25]. Thus, the introduction of the end-user-oriented large language models (LLMs) coincided with notable increases in public interest around AI topics. In parallel, other important actors entered the market, including open-source platforms such as Ollama (Toronto, Ontario, Canada) and H2O AI (Mountain View, California, USA). These platforms offer subscription-free and local LLM functionalities to end-users, thus enabling truly autonomous and private use of their models for the general public free of charge.

The term “artificial intelligence” was first coined in 1956 by computer scientist John McCarthy [26]. Artificial intelligence broadly refers to the ability of a computer to perform tasks typically associated with intelligent beings [27]. As per its definition, AI encompasses any non-naturally occurring system that displays capabilities typically associated with intelligence: communication, self-awareness, world knowledge, intentionality, creativity, and learning [28]. Artificial intelligence can be represented through a variety of end-user technological manifestations and implementations, as outlined in previous sections of this chapter.

Machine learning represents a distinct facet of the larger AI umbrella. Machine learning (ML) is a branch of computer science that allows computers to learn from, and derive outputs using data-driven relationships autonomously [29]. Given its functionality, ML is often leveraged as a prediction and classification tool with applications including natural language processing, computer vision, and omics learning models [30, 31, 32]. Provided with a large set of data, ML algorithms attempt to identify relationships between variables within a dataset, and the machine can then make predictions about the values of an outcome variable based on a given input or set of inputs. For example, ML may be able to train itself on the relationship between body mass index (BMI) and the probability that a patient has type 2 diabetes mellitus [33]. When provided with sufficient data, ML algorithms will be more likely to identify a meaningful relationship between input variables [34]. The algorithm may then be used to predict, for example, the probability that a future patient has type 2 diabetes mellitus given information about his or her BMI [29]. Ultimately, ML is a subset of AI that centers on a computer’s ability to autonomously learn associations between pieces of data.

Large language models (LLMs), also often referred to as natural language processors (NLPs), are text-specific applications of ML algorithms that have become a popular end-user product and consequently warrant dedicated discussion. Large language models, such as generative pre-trained transformers (GPTs) or Bidirectional Encoder Representations from Transformers (BERT), can generate coherent text in response to a given end-user prompt [35]. As a form of ML, large language models are trained on large amounts of text that are fed into algorithmic architecture to guide the machine in developing associations between words. By understanding the association between words, pertinent grammar and other language rules, the machine can provide a coherent response to a verbal prompt by predicting which words may be most appropriate in a response based on the LLM’s trained associations. All to say, large language models are a specific application of machine learning, which are trained specifically on text-based data (Figure 2).

Figure 2.

Relationships of the various domains and subdomains of artificial intelligence. Note: Understanding the intersections and distinctions of nomenclature related to artificial intelligence is foundational to higher-level consideration regarding AI implementation.

Understanding the connections between AI, ML, and LLMs is important to understanding the scope of applications for which each may be employed in the healthcare field, including medical academics and related domains. The applicability of AI technologies can vary widely across the spectrum of healthcare and academic medicine. Therefore, a framework may be useful to physician executives, hospital administrators, medical researchers, educators, and practitioners for guiding considerations around the implementation of AI in healthcare settings.

Advertisement

3. Discussion

In this section, we will provide an in-depth discussion of key considerations regarding new and existing implementations of artificial intelligence in healthcare, including strong focus on academics. We will formulate our discussion around the well-established “what? where? and why?” framework, with additional explorations of immediately adjacent areas and topics. Our discussion will then be followed by a dedicated section on the limitations of AI/ML in healthcare, and specifically in academics.

3.1 Considering applications and pitfalls of artificial intelligence in healthcare: What? Where? And Why? Framework

Artificial intelligence has potential applicability across many niche functions of healthcare delivery and medical academics [3, 8, 36]. Consequently, a framework that organizes the functionality of the AI technology, the context in which it is being used, and the end goal (or intended goal) of each use-case helps guide our considerations for AI implementation. In other words, when thinking about AI integration in healthcare and academic medicine, we propose a framework to consider what technology is being used, where it is being implemented, and why it is being used.

3.1.1 What technology is being used?

Artificial intelligence in healthcare and medical academics has many potential applications based on the functionality of AI technology being deployed. Particularly, ML technology in the forms of LLMs, computer vision models, and omics-specific learning models have demonstrated potential use cases in academic medicine (Figure 3) [3738]. These forms of ML technology function as pillars upon which many healthcare end-user AI applications are built, and therefore, understanding the functionality of each can be beneficial in identifying their potential use cases and inherent flaws [3].

Figure 3.

General classifications of “what” technology is being used. Note: LLMs, computer vision models, and omics learning models are three pillar types of AI/ML technology that have demonstrated use cases in academic medicine.

Fundamentally, LLMs are able to interact with written text including the ability to help edit sentences, aid in translation, and respond to prompts by a third-party user [39]. Large language models such as ChatGPT, LaMDA, and BERT produce text by predicting the appropriateness of every “next word” based on associations the model makes when trained on millions of lines of human-produced text [4041]. Functionally, this leads to the “context problem,” where the model does not actually “comprehend” the text in terms of its semantic context; it merely produces probabilistically “optimal responses” based upon the data on which it was trained [42]. Therefore, the architectural development of LLMs lends itself to a few generalized pitfalls of such models including reliability concerns (the propensity of the model to provide inaccurate or non-factual information), consistency issues (the ability of the model to produce different responses to the same prompt), prompt-interpretation complications (the sensitivity of the model to respond differently to minor variations in prompt phrasing), response bias (the tendency of a model to produce biased responses as a reflection of its training data), and black-box explainability barriers (the inability to concretely explain why the model responds the way it does) [42, 43, 44]. What follows is the realization that any implementations of LLMs in academic medicine, especially when it comes to the requirement for high-fidelity, quantifiable, reliable, and ethical data processing, may be inherently flawed without the a priori definition of strict curbs to be utilized in the academic research process [45, 46, 47]. After all, the LLM itself may not be “aware” of the boundary between “true” and “confabulated” reality [47, 48, 49].

Computer vision (CV) serves as a technological tool to many end-user AI applications in the medical field, especially within the academic, educational, and research domains [50]. Computer vision refers to the ability of a computer to process and identify images and video, much like humans do, and hence, CV can be leveraged for object detection, facial recognition, image classification, reverse image searches, and more [51, 52, 53]. Computer vision is an attempt to develop computer-based perception, which is relatively rigorous compared to the development of computer-based logic, as in the case of LLMs. In order to develop a CV model, computers must be trained on thousands of labeled images or videos, which allows the machine to develop an association between the image and the label [51]. This lends itself to a few potential pitfalls. First, the reliability and accuracy of a CV model are, in part, a function of the size of the training dataset. However, obtaining images, particularly those of human subjects, with appropriate consent can be both logistically difficult and costly. This presents a quandary (and an important trade-off) between improving model accuracy and developing a model in a timely, cost-effective, and ethical manner. All to say, CV models can be error-prone. Additionally, CV models are particularly vulnerable to identification bias. Even with sufficient training data, each image must receive a specific label. Incongruent with the adage that “a picture is worth a thousand words,” images in CV training receive a single label. Consequently, labels can be subject to the biased perspective of the labeler. Furthermore, the task of labeling, despite its notable downstream influence on the model’s functionality, is often delegated to third-party contributors outside of the primary development pipeline of the model [52]. Inaccurate or biased labeling may propagate and be reinforced through the use of the model, creating potential hazards.

Omics-targeted AI technologies aid in the characterization of biological products. The field of omics refers to an array of biological subfields that study the molecular processes occurring within the various and diverse constituents and components of living cells [54, 55]. Omics fields include genomics, proteomics, metabolomics, and more. Omics-targeted AI models have shown the ability to facilitate omics research by efficiently organizing, processing, and deriving relationships within large datasets necessary to analyze omics processes [56]. The primary challenges inherent to omics models are statistical in nature. This is because omics models are often built upon a large number of variables/parameters that must be inherently accounted for and accurately represented within the very complex and intertwined matrix of biomolecular processes. Given the large number of parameters that need to be addressed, developing models that are sufficiently robust and reproducible over multiple independent studies is difficult [57]. In statistical practice, training a model in which the number of samples (n) used in training is less than the number of parameters (p) in the model creates a problem with respect to the reliability of the model. This relationship between n and p should be considered critical for omics models, given the significant number of parameters used and the challenges associated with obtaining clinical and experimental data. Lastly, a key challenge inherent to omics learning models, given their size and complexity, lies within interpretability. The utility of a model can be hindered if it is difficult to understand and interpret the relevance of the results of the model [58, 59].

Ultimately, understanding what technology (LLM, CV, or omics model) is being used can allow individuals to assess generalized pitfalls and use cases for the technology, and is hence, a starting point for the overall framework.

3.1.2 Where is the technology being used?

The modern landscape of healthcare and academic medicine presents a unique environment for the introduction of AI technology. Settings for AI integration include clinical practice, medical research, and medical education. Factors ranging from ethical considerations around human subjects to provider-patient confidentiality present additional layers of context that must be considered when implementing AI into the healthcare field [60].

Within a clinical setting, perhaps the most notable concern around the deployment of AI rests in patient safety. Assurance in the reliability and validity of AI technology, in addition to transparency around its development, is required prior to any implementation into clinical practice [61]. However, the reliable functionality of AI technology is not unilaterally responsible for safe deployment of AI, as it also must be ensured that healthcare professionals are sufficiently trained to both use and understand the technology in the context of their practice [8]. Patient confidentiality also remains a key concern in clinical medicine; thus, the privacy of a patient’s health data should not be relinquished with the introduction of an AI technology. Furthermore, responsibility for potential liabilities associated with AI is often an open-ended question with respect to the deployment of AI into the practice of clinical medicine. One component of addressing AI-linked liability is the requirement, like with any other procedure, for informed patient consent for the use of such technology [61]. Lastly, key to the clinical relationship is a patient-provider relationship based on trust and compassion, which is critical when introducing new human-like technologies into such a dynamic [62].

Academic medical research is founded on the principle that contributions to the field are novel, yet based on well-tested, verifiable, and reproducible information [63, 64, 65]. Therefore, concerns around data authenticity, plagiarism, and justifiability of research insights are notable in the context of medical research [66, 67]. Additionally, traditional ethics surrounding clinical research, including privacy rights, informed consent, independent review, fair sample selection, methodological transparency, and favorable risk-reward ratio should all be considered without compromise when implementing AI-based technologies into academic research [68].

Medical education presents another focus area for AI application [69, 70]. A highly notable consideration underpinning medical education is the need for an up-to-date curriculum with experiences that enhance students’ didactic knowledge and relevant technical and interpersonal skills [71, 72]. Other general considerations around medical education include concerns of academic integrity, reliability of information, mechanisms for student evaluation, and opportunities for student feedback [44, 73, 74].

3.1.3 Why is the technology being used?

The anticipated benefits of AI in healthcare and academic medicine are varied. Such benefits range from reducing the workload of overburdened physicians to improving the medical education experience, as well as minimizing costs due to the optimization of diagnostic testing and improved diagnostic accuracy [75]. More generally, such benefits can be classified into “buckets” based on the ultimate end goal of implementing the technology. These “buckets” include, but are not limited to, improving outcomes, enhancing productivity, advancing the patient/provider experience, and reducing costs. The acronym Productivity, Experience, Costs, Outcomes (PECO) can be used to organize these benefits. Below, we outline each of the four components in more detail.

Productivity: Productivity aligns with the ability of a single working unit to provide more goods or services with the same amount of resources. Improvements in productivity may include the ability of a provider to see more patients due to a reduced administrative workload that is facilitated by natural language processing or the ability of medical researchers to spend less time on text editing as a consequence of GPTs.

Experience: Experience refers to the perceived convenience by which medical services, medical research, and education are administered and received. Natural language processors that simplify complex medical vocabulary in a patient’s health record may help patients feel more informed about their course of treatment. Likewise, LLMs that aid in the completion of administrative paperwork may minimize physician burden and reduce physician burnout.

Costs: Costs refer to the reduced net monetary expenditure associated with healthcare delivery, medical research, and medical education. In the context of AI, it is important to consider cost savings with respect to integration costs including all equipment, maintenance, subscription, and electricity costs [76]. An example of cost reduction may come from the decrease in superfluous medical testing as a result of improved AI-assisted diagnostic tools and better “medical data coordination” overall.

Outcomes: Outcomes refer to an increase in the quality of results. This can take on various forms, ranging from shortened diagnosis timelines to deeper medical research insights to improved knowledge retention among health science students.

It is important to emphasize that AI-based technologies may not fall exclusively into one “bucket” as there may be several (or even multiple) associated end goals and outcomes.

Advertisement

4. Artificial intelligence across various medical domains: limitations versus use cases

Combining elements of the What, Where, and Why framework may be useful in thinking about benefits and pitfalls of AI implementation in healthcare (Figure 4). Considerations of the functionality of the technology (LLM, CV, or omics learning models) being implemented will allow individuals to consider generalized pitfalls and applications of the AI technology. Considerations regarding where (clinical setting, research setting, or educational setting) the technology is being implemented is superimposed on context-specific considerations and then the corresponding generalized pitfalls and applications. Lastly, considering why (outcomes, productivity, experience, and costs) the technology is being implemented can serve as a basis for an in-depth pro-con analysis of the uses and pitfalls of the technology within a specific context.

Figure 4.

A framework for conceptualizing AI implementation in healthcare settings.

4.1 Overview of applications and pitfalls around artificial intelligence in clinical practice

The integration of AI into both clinical and academic domains of medicine has the potential to transform many aspects of clinical practice, education, and research, while simultaneously addressing some of the challenges in healthcare delivery. Applications vary widely in their utility and potential benefits, including clinical documentation optimization, improvements in educational experience, more streamlined and effective research, and augmented diagnostic imaging capabilities, among other areas [3]. However, regardless of the specific technology at hand, integrating AI into relevant settings is a major endeavor that comes with inherent risks that should be considered, managed, and mitigated to prioritize any potential downstream consequences, from patient safety issues to research biases [60].

Both LLMs and natural language processors (NLPs) have the potential to provide numerous administrative and clinical benefits to diverse groups of users, including medical students, medical practitioners, researchers, allied health professionals, and patients [77, 78, 79]. Given the importance of fluid communication between various stakeholders, LLMs may serve to bridge communication gaps. For example, LLM-based paradigm may help improve patient understanding of their medical record by simplifying complex medical language [62]. In addition, LLMs may assist healthcare providers in drafting clinical documentation, summarizing dense medical text, and aiding in diagnosis and triage [80, 81]. Additionally, large language models have the ability to facilitate administrative clinical tasks, such as the extraction of pertinent information for medical billing and insurance authorization approvals, as demonstrated in Figure 5 [82, 83]. Prior authorization and clinical documentation are major contributors to physician burnout, and automating parts of these processes may lead to a reduction in administrative burden while also reducing costs and improving the patient-provider experience [84].

Figure 5.

Example of autonomous identification of relevant ICD-10 Billing Codes from text [44]. Note: This figure demonstrates the capability of LLMs to facilitate processes related to text recognition.

However, it is important to understand the pitfalls associated with the implementation of large language models in clinical settings. First, the general ability of LLMs to “fabricate” information and promote trained biases presents a consequential concern that should be ultimately mitigated through expert oversight and hardwired quality control tools [76, 85]. Furthermore, while these tools may have the ability to augment patient-provider interactions, they cannot replace practitioners in building trust with patients [76]. Additionally, over-reliance on LLMs to guide diagnosis and medical triage may impair a provider’s critical thinking skills over time [86].

Computer vision technology within clinical medicine has demonstrated potential to aid in radiologic, pathologic, and dermatologic image recognition [3]. Such image recognition tools may help physicians detect abnormalities sooner or with more efficiency [87]. Through potential productivity enhancement, these tools may help expand care in the face of physician shortage across specialties including radiology and pathology [88]. However, the consequences of CV error and propagated bias are heightened in the clinical setting, particularly if such missteps compromise patient safety. In the event of misdiagnosis by a CV model, it is currently unclear to whom or what liability would fall [89]. Ultimately, potentiality for error, concerns of patient-provider trust, and regulatory hurdles, prevent a fully autonomous employment of diagnostic imaging tools in the present [90].

Additional uses of CV within clinical practice include the development of immersive experiences for patients and providers using augmented reality (AR) and virtual reality (VR). Augmented reality may be used to guide individuals through surgical procedures or emergency response techniques through the use of lenses that provide direction and feedback to the user [3]. Such technology has the potential to further advance opportunities in minimally invasive surgery. Minimally invasive surgery has become a focus of medical research as it presents an opportunity to mitigate operative trauma and expedite postoperative recovery. However, the consequence of achieving these improvement outcomes is a more difficult surgical task with limited vision and restricted vision. Augmented reality tools have the potential to relieve this tradeoff by increasing viewing area [91]. Augmented reality technologies have already been employed in certain intraoperative settings in the field of spine surgery and will most likely continue to expand initially in orthopedic and neurosurgery due to the immobility of the skeletal system and the spine [92]. However, with the rapid advancement and increased use of peri- and intraoperative imaging in thoracic and abdominal surgeries, AR could potentially impact a much broader range of surgeries [92, 93, 94]. Besides aiding physicians in providing clinical care, AR and VR tools can directly impact patient experiences. For instance, by providing an escape to virtual worlds and environments, VR immersive experiences have shown benefits in improving therapeutic outcomes in patients experiencing debilitating or painful illnesses such as chronic stroke. Furthermore, immersive AI technology used as an important adjunct has demonstrated partial relief to late-stage cancer patients (Table 1) [3].

Areas of AI integration in clinical medicineExamples of use cases
Communication
  • Translation of medical text into lay terms

Administrative and clerical duties
  • Automation of charting and billing

  • Automation of administrative duties (i.e., prior authorization)

Image recognition
  • Radiology and pathology image recognition

Augmented and virtual reality
  • Simulation of procedures

  • Improvement of minimally invasive procedures

  • Improvement of patient experiences

Precision medicine
  • Optimization of care based on omics AI models

Table 1.

Summary of uses of AI in clinical medicine.

Applicability of omics AI models within a clinical setting lies primarily within the field of precision medicine—the ability to tailor healthcare interventions to an individual based on genomic composition, age, gender, geography, race, family history, immunity profile, metabolic panel, microbiome, and environmental sensitivities [3]. While precision medicine may boast a number of envisioned benefits including improved drug response, minimized complications, and reduced costs, the ability to develop sufficiently reliable omics models under social and economic constraints remains difficult [95].

4.2 Overview of applications and pitfalls around artificial intelligence in modern medical research

Artificial intelligence can enhance the process of medical research by streamlining clerical tasks and presenting new tools for data analysis. Large language models in medical research can assist in many clerical tasks associated with the research process including enhancing vocabulary, rectifying grammar errors, annotating documents, and scraping text-based information [76, 96]. A particularly powerful use case of LLMs is the ability to summarize and condense medical text. This can be of benefit to researchers when developing annotated bibliographies, performing integrative reviews, polishing language/grammar, and more [76]. A notable concern in the use of LLMs in medical research is the ability of the model to “fabricate” or overgeneralize research insights, resulting in inappropriate conclusions and pointing to a need for continued critical review of the overall research process by humans. Perhaps transparent disclosure of the use of AI in the research process can aid in identifying potential instances of fabrication or bias [76]. The ability to track and correlate the potential effects and biases of AI-based implementations on research quality and fidelity will be an important first step in ensuring that robust error-reduction frameworks can be operationalized in the future (Table 2).

Areas of AI integration in medical researchExamples of use cases
Research interpretation
  • Editing and revising

  • Summarizing medical literature

Research study design
  • Clinical trial recruitment

  • Determining the novelty of research questions

Omics AI models
  • Integration of gene expression and clinical parameters

  • Drug discovery and development

Table 2.

Summary of uses of AI in medical research.

Besides aiding in streamlining various research tasks, AI may be able to improve research study design. For instance, patient selection is a vital component of clinical trials and presents one of the main bottlenecks when it comes to trial completion [97]. Various barriers exist, including overwhelming medical jargon and the non-centralized distribution of information on clinical trials [98]. To aid patients in finding and understanding available clinical trials, NLP programs can be leveraged to mine available clinical trial databases for eligible studies and to translate these studies to improve patient understanding [98]. Other potential uses for AI in research design include determining the novelty of research questions and assisting in statistical analysis.

Artificial intelligence can also directly influence medical research in potentially powerful ways. Large machine learning omics models can be leveraged in medical research to enhance the study of biomolecular processes. Given their functionality of identifying relationships between parameters, especially in multi-parametric and highly complex systems, ML omics algorithms can be leveraged to identify relationships between gene expression and other EMR data [3, 99]. Another notable area of applicability for omics ML models falls into drug discovery and development [100]. For instance, researchers may leverage such models to predict drug activity, target interaction, and other properties, which may aid in the molecular design process [3, 100]. The key challenge in this context consists of developing cost-effective models that are sufficiently reliable to meet safety requirements for human research subjects and, ultimately, patients receiving the resultant therapeutic agent [58].

4.3 Overview of utility and limitations of artificial intelligence applications in today’s medical education

There has been significant development in education focused AI technologies over the past two decades influencing the methodologies by which students learn and curricula are designed [101, 102]. The introduction of “teacher bots,” which are effectively education-focused chat bots that serve as individualized tutors, has provided a mechanism for students to address learning needs at their convenience [103]. This area of development is bound to be highly controversial for a number of well-founded reasons [104]. Education underpins the foundation of modern medicine, from undergraduate coursework to post-graduate studies and lifelong continuing medical education (CME). Given the educational nature of the field, medicine presents a unique opportunity to introduce educational AI tools across the entire learner continuum, and realization of this opportunity has been reflected by increased enthusiasm and research around the topic of AI in medical education [44, 105].

Artificial intelligence tools in education can take on a variety of forms including intelligent tutoring systems, integrative learning environments, assistant chatbots, grading and assessment tools, diagnostic assistants, analytics monitors, and medical simulations [106, 107, 108]. In addition, when properly and safely implemented, with appropriate checks and balances that focus on content validity and model fidelity, AI-based educational tools may allow entire regions of the globe (especially in low-resource environments) to make an “educational leap” not dissimilar to the leap made when highly restricted traditional telephony gave way to omnipresent cellular telephony across the world [109, 110, 111, 112].

Intelligent tutoring systems and integrated learning environments are learning tools that adaptively expose students to learning content based on the students’ level of mastery as perceived by the AI model overseeing the learner’s progress. Perhaps the most readily apparent use of artificial intelligence in medical education is the use of AI in generation of practice or assessment questions. Some studies have already investigated the use of generative AI in producing multiple-choice, board-style questions for undergraduate medical education, and although AI-created questions generally still required further human editing to reach satisfactory accuracy and comprehensiveness, this process was still more efficient than traditional question generation [113]. In the future, these use cases may be extended to include generation of open-ended questions. Consequently, these tools have the potential to aid in the delivery of lesson plans and provide the opportunity for a “flipped classroom” in which each student has the opportunity to learn at their own speed, thereby potentially enhancing student outcomes and improving the productivity of time in-class [74, 114]. However, notable pitfalls inherent to AI-based tutoring systems include the hindrance of social interaction, detriment to human communication skills, and inability to monitor students’ body language [114]. Finally, over-reliance on intelligent tutoring/learning systems may result in both quantitative and qualitative declines in the educational workforce. In this context, intelligent tutoring systems and integrated learning environments are best suited as supplementary educational tools, and appropriate mechanisms should be utilized to ensure that the AI educational revolution does not result in irreversible brain drain and talent erosion.

Educational chatbots can serve to field questions from educational stakeholders. Chatbots are end-user-focused LLMs that can aid medical students in the individual learning process. This on-demand access to receive answers to questions may improve the productivity and the experience of medical students, while reducing the burden on education providers; however, there are pitfalls to consider [74]. Leveraging, the What, Where, Why? framework, we must remember that currently, available LLMs have the ability to “fabricate” information, which can be further compounded by a lack of transparency regarding the generative process itself [115, 116, 117]. In addition, the constant and ever-accelerating growth of the existing medical information available to LLMs for processing and use, would require a great deal of oversight and real-time model updates, inclusive of purging of information that is no longer valid and/or known to be harmful [118, 119]. Further, concerns linked generally to ML algorithms may present students with a gap in clinical reasoning. In other words, it must be ensured that intelligent tutoring systems in medical education are capable of explaining the reasoning behind the answers they generate in order to provide students with proper cognitive intuition [44]. The potential for the provision of misinformation coupled with potential barriers thus requires another reference point for students to engage [115].

Another potential benefit of AI-based tools, particularly LLMs, in medical education is within the area of grading and assessment of students [120]. Artificial intelligence can provide a mechanism for grading assessments without the influence of a human grader for questions beyond multiple choice and other more traditional standardized testing. This, in theory, presents the opportunity for educators to assess students in a more open-ended fashion (i.e., free response questions and essays) without the concern for ease of grading and/or standardization of the assessment [74]. At the same time, given concerns inherent to LLMs, understanding why the grading tool assigns a specific grade to a student may be problematic, and therefore, all final grades may still need to be reviewed by a human grader [121]. More so, the nature of assessments in medical education is varied, and thus, AI-based grading systems may not be useful in all contexts of student evaluation, such as in the assessment of a student’s performance on a physical examination [44]. Nonetheless, such grading tools may be implemented to provide a first-pass review of a student’s work, thus improving the overall systemic efficiency.

Artificial intelligence also poses benefits to simulation-based medical education [122, 123]. Tools including augmented reality resuscitation simulators and AI-simulated patients, in theory, present the opportunity for medical and surgical trainees to practice skills related to clinical evaluation and surgical procedures in a seemingly realistic context without the potential harm to an actual patient [86]. In the future, surgical simulation software may also be able to track how well students are performing or making progress, highlighting areas for improvement and even predicting the number of repetitions needed to master certain skills [124]. These simulations could be further supplemented with CV algorithms to track learner movements, leading to the possibility of providing feedback on surgical dexterity and motion economy [125]. However, the key challenge in this dynamic is training future doctors, who will be responsible for the care of human beings, using a training platform or an educational entity that is intrinsically non-human. This may create both potential biases and “unintended standardizations” of future doctors with less exposure to the real-life diversity of physical and social differences among patients – something that healthcare providers are bound to experience throughout their careers. Despite the above risks, some researchers have proposed ways that AI could be effectively leveraged to enhance empathy and compassion among both trainees and future physicians, such as through immersive experiential simulations of conditions like old age or disabilities [126]. Still, these technologies must be further improved before widespread implementation in medical education. In aggregate, our current knowledge supports the use of simulated/robotic patients primarily as supplementary tools, but by no means a replacement for real-life bedside or clinic experience (Table 3) [127].

Areas of AI integration in medical educationExamples of use cases
Autonomous tutoring
  • Educational chatbots

Assessment generation
  • Generation of multiple-choice style questions

  • Generation of flipped classroom lesson plans

Assessment grading
  • Automation of assessment grading

Simulation-based learning
  • Simulation-based skills training

  • Automation of simulation training feedback

  • Immersive experiential simulations for development of compassion

Table 3.

Summary of uses of AI in medical education.

In summary, exposure to AI during medical education, if implemented with appropriate guardrails, may also serve to equip the next generation of practitioners with the technical and cognitive skills, and the understanding needed to use AI-based technologies in future practice. The overall comfort levels and understanding of AI technology by providers is key to integrating AI into clinical practice, and this is optimally established during medical training [128].

Advertisement

5. Synthesis and conclusion

Overall, the exploration of AI applications in medicine reveals a promising yet complex landscape marked by both substantial and transformational advancements and potential pitfalls. The integration of AI into healthcare and medical academics holds immense potential to revolutionize diagnostics, treatment, education, research, and healthcare delivery. However, it is imperative to proactively address any associated ethical considerations, data privacy concerns, and potential biases inherent in AI algorithms. Given the increasing interest around AI integration in healthcare, we aim to provide decision-makers in academic medical settings with a framework to guide discussions around AI implementation. To guide AI implementation that is capable of harnessing the full benefits of AI, while minimizing the associated risks, we propose that decision-makers consider deeply the functionality and limitations of the type of AI being implemented, the legal and ethical barriers associated with various implementation contexts, and the overall cost-benefit analyses to result from implementation. Striking a balance between technological innovation and human oversight is crucial to the future integration of AI into the healthcare field. As AI applications continue to evolve and use cases for AI in healthcare settings continue to expand, our proposed framework may need to be expanded and adapted. While our framework and discussion around current use cases of AI in academic medical contexts may be useful in some contexts, it is not completely exhaustive. Further research, case studies, and discussions around each of the AI use cases mentioned in this chapter would be beneficial in developing detailed and nuanced implementation guides that are specific to both a given technology and a given context. As the field continues to evolve, ongoing research, collaborative efforts, and stringent regulatory frameworks will be essential to navigate the applications and pitfalls of AI, ensuring that its integration into healthcare practices aligns with the overarching goal of improving patient outcomes and advancing the overall quality of medical care.

Advertisement

Acknowledgments

The authors would like to acknowledge the support of the following non-author contributors: Dr. Allison Walker (St. Luke’s University Health Network, Easton, PA) and Dr. Maria T. Martinez Baladejo (St. Luke’s University Health Network, Easton, PA). Their assistance in the areas of project logistics and general support has been valuable to both the overall team effort and the completion of this manuscript.

References

  1. 1. De Silva D et al. ChatGPT and generative AI guidelines for addressing academic integrity and augmenting pre-existing Chatbots. In: 2023 IEEE International Conference on Industrial Technology (ICIT). Piscataway, NJ: IEEE; 2023
  2. 2. Celik I et al. The promises and challenges of artificial intelligence for teachers: A systematic review of research. TechTrends. 2022;66(4):616-630
  3. 3. Bohr A, Memarzadeh K. The rise of artificial intelligence in healthcare applications. In: Artificial Intelligence in Healthcare. Cambridge, MA: Academic Press; 2020. pp. 25-60
  4. 4. Meskó B, Topol EJ. The imperative for regulatory oversight of large language models (or generative AI) in healthcare. npj Digital Medicine. 2023;6(1):120
  5. 5. Larsen BC. Governing artificial intelligence: Lessons from the United States and China [Phd]; Copenhagen Business School. 2022
  6. 6. Sahni N et al. The Potential Impact of Artificial Intelligence on Healthcare Spending. Cambridge, MA: National Bureau of Economic Research; 2023
  7. 7. Stawicki SP et al. Introductory chapter: Artificial intelligence in healthcare-Where do we go from here? In: Artificial Intelligence in Medicine and Surgery-An Exploration of Current Trends, Potential Opportunities, and Evolving Threats. Vol. 1. London, UK: IntechOpen; 2023
  8. 8. Sapci AH, Sapci HA. Artificial intelligence education and tools for medical and health informatics students: Systematic review. JMIR Medical Education. 2020;6(1):e19285
  9. 9. Luxton DD. Recommendations for the ethical use and design of artificial intelligent care providers. Artificial Intelligence in Medicine. 2014;62(1):1-10
  10. 10. Lee D, Yoon SN. Application of artificial intelligence-based technologies in the healthcare industry: Opportunities and challenges. International Journal of Environmental Research and Public Health. 2021;18(1):271
  11. 11. Dwivedi YK et al. Artificial intelligence (AI): Multidisciplinary perspectives on emerging challenges, opportunities, and agenda for research, practice and policy. International Journal of Information Management. 2021;57:101994
  12. 12. Services, C.f.M.a.M. NHE fact sheet. 2023. Available from: https://www.cms.gov/data-research/statistics-trends-and-reports/national-health-expenditure-data/nhe-fact-sheet [Accessed: Dec 15, 2023]
  13. 13. Executives, A.C.o.H. Survey: Personnel Shortages Cited above Financial Challenges by CEOs as Top Issue Confronting Hospitals in 2021. Chicago, IL: The Foundation of ACHE; 2022. p. 3
  14. 14. Summit, I.o.M.U.C.o.t.H.P.E, Greiner A, Knebel E. Challenges facing the health system and implications for educational reform. In: Health Professions Education: A Bridge to Quality. Vol. 192. United States: National Academies Press; 2003
  15. 15. Aminizadeh S et al. Opportunities and challenges of artificial intelligence and distributed systems to improve the quality of healthcare service. Artificial Intelligence in Medicine. 2024;149:102779
  16. 16. Maslej N. Artificial intelligence index report 2023. In: Maslej N, editor. AI Index Report. United States: Stanford University; 2023. p. 386
  17. 17. Frenk J et al. Health professionals for a new century: Transforming education to strengthen health systems in an interdependent world. The Lancet. 2010;376(9756):1923-1958
  18. 18. DeVilbiss MB, Roberts LW. Artificial intelligence tools in scholarly publishing: Guidance for academic medicine authors. Academic Medicine. 2023;98(8):865-866
  19. 19. Lee J et al. Artificial intelligence in undergraduate medical education: A scoping review. Academic Medicine. 2021;96(11S):S62-S70
  20. 20. Kim JK et al. ChatGPT and large language model (LLM) chatbots: The current state of acceptability and a proposal for guidelines on utilization in academic medicine. Journal of Pediatric Urology. 2023;19(5):598-604
  21. 21. Bajwa J et al. Artificial intelligence in healthcare: Transforming the practice of medicine. Future Healthcare Journal. 2021;8(2):e188-e194
  22. 22. Triola MM et al. Artificial intelligence screening of medical school applications: Development and validation of a machine-learning algorithm. Academic Medicine. 2023;98(9):1036-1043
  23. 23. Webster CS. Can artificial intelligence tools operate transparently in medicine? Academic Medicine. 2022;97(10):1431-1432
  24. 24. Trends G. Artificial intelligence. 2023. Available from: https://trends.google.com/trends/explore?date=2022-11-21%202023-12-13&geo=US&q=%2Fm%2F0mkz&hl=en
  25. 25. Trends G. Large Language Model. Mountain View, CA: Google Trends; 2023
  26. 26. Smith C et al. The History of Artificial Intelligence. United States: University of Washington; 2006. p. 27
  27. 27. Copeland BJ. Artificial intelligence. In: Encyclopedia Britannica. Chicago, IL: Encyclopedia Britannica; 2023
  28. 28. Schank RC. What is AI, anyway? In: AI Magazine. United States: AI Magazine; 1987. p. 7
  29. 29. Bi Q et al. What is machine learning? A primer for the epidemiologist. American Journal of Epidemiology. 2019;118(12):2222-2239
  30. 30. Hamamoto R. Application of artificial intelligence for medical research. Biomolecules. 2021;11(1):90
  31. 31. Li C et al. Data analysis methods for defining biomarkers from omics data. Analytical and Bioanalytical Chemistry. 2022;414(1):235-250
  32. 32. Reel PS et al. Using machine learning approaches for multi-omics data analysis: A review. Biotechnology Advances. 2021;49:107739
  33. 33. Joshi RD, Dhakal CK. Predicting type 2 diabetes using logistic regression and machine learning approaches. International Journal of Environmental Research and Public Health. 2021;18(14):7346
  34. 34. Angelov PP et al. Explainable artificial intelligence: An analytical review. Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery. 2021;11(5):e1424
  35. 35. Tamkin A et al. Understanding the Capabilities, Limitations, and Societal Impact of Large Language Models. Stanford, CA: Stanford Natural Language Processing Group; 2020. pp. 1-8
  36. 36. Kochanny SE, Pearson AT. Academics as leaders in the cancer artificial intelligence revolution. Cancer. 2021;127(5):664-671
  37. 37. Rajendran S et al. Learning across Diverse Biomedical Data Modalities and Cohorts: Challenges and Opportunities for Innovation. Cambridge, MA: Patterns; 2024
  38. 38. Rajendran S et al. Patchwork Learning: A Paradigm towards Integrative Analysis across Diverse Biomedical Data Sources. Appleton, WI: International Conference on Learning Representations; 2024. pp. 1-21
  39. 39. Jiang Z et al. How can we know what language models know? Transactions of the Association for Computational Linguistics. 2020;8:423-438
  40. 40. Yang J et al. Harnessing the power of llms in practice: A survey on chatgpt and beyond. ACM Transactions on Knowledge Discovery from Data. 2024;18(6):1-32
  41. 41. McTear M, Ashurkina M. Transforming Conversational AI: Exploring the Power of Large Language Models in Interactive Conversational Agents. New York, NY: Apress; 2024
  42. 42. Urs S. The power and pitfalls of large language models: A fireside chat with Ricardo Baeza-yates. Information Matters. 2022;2(5):1-5
  43. 43. Koga S. Exploring the Pitfalls of Large Language Models: Inconsistency and Inaccuracy in Answering Pathology Board Examination-Style Questions. Laurel Hollow, NY: Cold Spring Harbor Laboratory; 2023
  44. 44. Chan KS, Zary N. Applications and challenges of implementing artificial intelligence in medical education: Integrative review. JMIR Medical Education. 2019;5(1):e13930
  45. 45. Cardenuto JP et al. The age of synthetic realities: Challenges and opportunities. APSIPA Transactions on Signal and Information Processing. 2023;12(1):1-62
  46. 46. Lee P. Synthetic Data and the Future of AI. Ithaca, NY: Cornell Law Review; 2024
  47. 47. McGuire A. Leveraging ChatGPT for rethinking plagiarism, digital literacy, and the ethics of Co-authorship in higher education: A position paper and comparative critical reflection of composing processes. Irish Journal of Technology Enhanced Learning. 2023;7(2):21-31
  48. 48. Lloyd D. What is it like to be a bot?: The world according to GPT-4. Frontiers in Psychology (Lausanne, Switzerland). 2024;15
  49. 49. Culture, A.R.G.o.t.C.f.D. Encountering artificial intelligence: Ethical and anthropological investigations. Journal of Moral Theology. 2023;1(Theological Investigations of AI):i-262
  50. 50. Orhei C et al. End-to-end computer vision framework: An open-source platform for research and education. Sensors. 2021;21(11):3691
  51. 51. Szeliski R. Computer Vision: Algorithms and Applications. 2nd ed. Berlin, Germany: Springer Nature; 2022. p. 925
  52. 52. Birhane A. Automating Ambiguity: Challenges and Pitfalls of Artificial Intelligence, in School of Computer Science. Dublin, Ireland: University College Dublin; 2021. p. 189
  53. 53. Thyagharajan K, Kalaiarasi G. A review on near-duplicate detection of images using computer vision techniques. Archives of Computational Methods in Engineering. 2021;28(3):897-916
  54. 54. Irvine GW, Nguyen S. An overview of the “-omics” fields at the forefront of next-generation personalized medicine and fundamental systems biology studies. Biomedical Genetics and Genomics. 2019;4(2):3-5
  55. 55. Ali AM, Mohammed MA. A comprehensive review of artificial intelligence approaches in omics data processing: Evaluating Progress and challenges. International Journal of Mathematics, Statistics, and Computer Science. 2024;2:114-167
  56. 56. D’Adamo GL, Widdop JT, Giles EM. The future is now? Clinical and translational aspects of “omics” technologies. Immunology & Cell Biology. 2021;99(2):168-176
  57. 57. Grapov D et al. Rise of deep learning for genomic, proteomic, and metabolomic data integration in precision medicine. OMICS: A Journal of Integrative Biology. 2018;22(10):630-636
  58. 58. Zhang Z et al. Deep learning in omics: A survey and guideline. Briefings in Functional Genomics. 2019;18(1):41-57
  59. 59. Cannataro M et al. Artificial Intelligence in Bioinformatics: From Omics Analysis to Deep Learning and Network Mining. Amsterdam, Netherlands: Elsevier; 2022
  60. 60. Farhud D, Zokaei S. Ethical issues of artificial intelligence in medicine and healthcare. Iran Journal of Public Health. 2021;50(11):i-iv
  61. 61. Gerke S, Minssen T, Cohen G. Ethical and legal challenges of artificial intelligence-driven healthcare. In: Bohr KMA, editor. Artificial Intelligence in Healthcare. Amsterdam, Netherlands: Elsevier; 2020. pp. 295-336
  62. 62. Atallah SB et al. How large language models including generative pre-trained transformer (GPT) 3 and 4 will impact medicine and surgery. Techniques in Coloproctology. 2023;27(8):609-614
  63. 63. Stodden VC. Reproducible Research: Addressing the Need for Data and Code Sharing in Computational Science. New Haven, CT: Yale Law School Roundtable; 2010
  64. 64. Alger BE. Defense of the Scientific Hypothesis: From Reproducibility Crisis to Big Data. USA: Oxford University Press; 2019
  65. 65. Cimino JJ. Clinical research data: Characteristics, representation, storage, and retrieval. In: Principles and Practice of Clinical Research. Amsterdam, Netherlands: Elsevier; 2018. pp. 547-557
  66. 66. Lipworth W et al. Science at warp speed: Medical research, publication, and translation during the COVID-19 pandemic. Journal of Bioethical Inquiry. 2020;17(4):555-561
  67. 67. Tyson J. Shortcomings of ChatGPT. Journal of Chemical Education. 2023;100(8):3098-3101
  68. 68. Center, N.C. Ethics in Clinical Research. 2021. Available from: https://clinicalcenter.nih.gov/recruit/ethics.html#4 [Accessed: Jan 1, 2024]
  69. 69. Lillehaug S-I, Lajoie SP. AI in medical education—Another grand challenge for medical informatics. Artificial Intelligence in Medicine. 1998;12(3):197-225
  70. 70. Sousa MJ et al. The potential of AI in health higher education to increase the students’ learning outcomes. TEM Journal. 2021;10(2):488-497
  71. 71. Greenberg R, Loyd G, Wesley G. Integrated simulation experiences to enhance clinical education. Medical Education. 2002;36(11):1109-1110
  72. 72. Stawicki TT et al. From “Pearls” to “Tweets:” How Social Media and Web-Based Applications Are Revolutionizing Medical Education. Mumbai, India: Medknow; 2018. pp. 93-97
  73. 73. Cotton DRE, Cotton PA, Shipway JR. Chatting and cheating: Ensuring academic integrity in the era of ChatGPT. Innovations in Education and Teaching International. 2023;61(2):1-12
  74. 74. Remian D. Augmenting education: Ethical considerations for incorporating artificial intelligence in education. In: Instructional Design. Vol. 57. University of Massachusetts at Boston: United States; 2019
  75. 75. Amisha PM et al. Overview of artificial intelligence in medicine. Journal of Family Medicine and Primary Care. 2019;8(7):2328-2334
  76. 76. Stawicki S, Martinez-Baladejo M, Ng-Pellegrino A. What’s new in academic international medicine? Artificial intelligence and machine learning is here to stay, forcing rapid adoption and adaptation. International Journal of Academic Medicine. 2023;9(3):117-120
  77. 77. Sallam M. The utility of ChatGPT as an example of large language models in healthcare education, research and practice: Systematic review on the future perspectives and potential limitations. MedRxiv. 2023:1-34
  78. 78. Hobensack M et al. A rapid review on current and potential uses of large language models in nursing. International Journal of Nursing Studies. 2024;154:104753
  79. 79. Rane NL et al. Contribution and performance of ChatGPT and other large language models (LLM) for scientific and research advancements: A double-edged sword. International Research Journal of Modernization in Engineering Technology and Science. 2023;5(10):875-899
  80. 80. Cascella M et al. The breakthrough of large language models release for medical applications: 1-year timeline and perspectives. Journal of Medical Systems. 2024;48(1):1-11
  81. 81. Yu P et al. Leveraging generative AI and large language models: A comprehensive roadmap for healthcare integration. In: Healthcare. Basel, Switzerland: MDPI; 2023
  82. 82. Vani S et al. Natural language processing utilisation in healthcare. In: Machine Learning Approaches and Applications in Applied Intelligence for Healthcare Analytics. Boca Raton, FL: CRC Press; 2022
  83. 83. Abid R, Salzman G. Evaluating physician burnout and the need for organizational support. Missouri Medicine. 2021;118(3):185
  84. 84. Thirunavukarasu AJ et al. Large language models in medicine. Nature Medicine. 2023;29:1930-1940
  85. 85. Tang L et al. Evaluating large language models on medical evidence summarization. npj Digital Medicine. 2023;6(10):158
  86. 86. Komasawa N, Yokohira M. Simulation-Based Education in the Artificial Intelligence Era. San Francisco, CA: Cureus; 2023
  87. 87. Van Leeuwen KG et al. How does artificial intelligence in radiology improve efficiency and health outcomes? Pediatric Radiology. 2022;52(11):2087-2093
  88. 88. Analysis, N.C.f.H.W. Physician Workforce: Projections, 2020-2035. United States: National Center for Healthcare Workforce Analysis; 2022
  89. 89. Cestonaro C et al. Defining medical liability when artificial intelligence is applied on diagnostic algorithms: A systematic review. Frontiers in Medicine. 2023;10
  90. 90. Goisauf M, Abadia MC. Ethics of AI in radiology: A review of ethical and societal implications. Frontiers in Medicine. 2022:5
  91. 91. Paolis LTD, Aloisio G. Augmented Reality in Minimally Invasive Surgery, in Advances in Biomedical Sensing, Measurements, Instrumentation and Systems. Berlin: Springer; 2010
  92. 92. Peh S et al. Accuracy of augmented reality surgical navigation for minimally invasive pedicle screw insertion in the thoracic and lumbar spine with a new tracking device. The Spine Journal. 2020;20(4):629-637
  93. 93. Soler L et al. Virtual reality and augmented reality in digestive surgery. In: Third IEEE and ACM International Symposium on Mixed and Augmented Reality. Piscataway, NJ: IEEE; 2004
  94. 94. Arjomandi Rad A et al. Extended, virtual and augmented reality in thoracic surgery: A systematic review. Interactive Cardiovascular and Thoracic Surgery. 2022;34(2):201-211
  95. 95. Naithani N et al. Precision medicine: Uses and challenges. Medical Journal of Armed Forces India. 2021;77(3):258-265
  96. 96. Goel A et al. LLMs accelerate annotation for medical information extraction. In: Stefan H et al., editors. Proceedings of the 3rd Machine Learning for Health Symposium. Cambridge, MA: PMLR: Proceedings of Machine Learning Research; 2023. pp. 82-100
  97. 97. Nipp RD, Hong K, Paskett ED. Overcoming barriers to clinical trial enrollment. In: American Society of Clinical Oncology Educational Book. Annual Meeting. Alexandria, VA: American Society of Clinical Oncology; 2019
  98. 98. Harrer S et al. Artificial intelligence for clinical trial design. Trends in Pharmacological Sciences. 2019;40(8):577-591
  99. 99. Lipkova J et al. Artificial intelligence for multimodal data integration in oncology. Cancer Cell. 2022;40(10):1095-1110
  100. 100. Perakakis N et al. Omics, Big Data and Machine Learning as Tools to Propel Understanding of Biological Mechanisms and to Discover Novel Diagnostics and Therapeutics. Amsterdam, Netherlands: Elsevier; 2018. pp. A1-A9
  101. 101. Roll I, Wylie R. Evolution and revolution in artificial intelligence in education. International Journal of Artificial Intelligence in Education. 2016;26(2):582-599
  102. 102. Guan C, Mou J, Jiang Z. Artificial intelligence innovation in education: A twenty-year data-driven historical analysis. International Journal of Innovation Studies. 2020;4(4):134-147
  103. 103. Bayne S. Teacherbot: Interventions in automated teaching. Teaching in Higher Education. 2015;20(4):455-467
  104. 104. Rudolph J, Tan S, Tan S. ChatGPT: Bullshit spewer or the end of traditional assessments in higher education? Journal of applied learning and teaching. 2023;6(1):342-363
  105. 105. Sun L et al. Artificial intelligence for healthcare and medical education: A systematic review. American Journal of Translational Research. 2023;15(7):4820-4828
  106. 106. González-Calatayud V, Prendes-Espinosa P, Roig-Vila R. Artificial intelligence for student assessment: A systematic review. Applied Sciences. 2021;11(12):5467
  107. 107. Dai C-P, Ke F. Educational applications of artificial intelligence in simulation-based learning: A systematic mapping review. Computers and Education: Artificial Intelligence. 2022;3:100087
  108. 108. Pedro F et al. Artificial Intelligence in Education: Challenges and Opportunities for Sustainable Development. New York, NY: United Nations; 2019
  109. 109. Hart SL, Christensen CM. The great leap: Driving innovation from the base of the pyramid. MIT Sloan Management Review. 2002;44(1):51
  110. 110. Etzo S, Collender G. The mobile phone ‘revolution’ in Africa: Rhetoric or reality? African Affairs. 2010;109(437):659-668
  111. 111. Kendall J, Voorhies R. The mobile-finance revolution; how cell phones can spur development. Foreign Affairs. 2014;93:9
  112. 112. Kalil T. Harnessing the mobile revolution. Innovations: Technology Governance Globalization. 2009;4(1):9
  113. 113. Zuckerman M et al. ChatGPT for assessment writing. Medical Teacher. 2023;45(11):1224-1227
  114. 114. Akyuz Y. Effects of intelligent tutoring systems (ITS) on personalized learning (PL). Creative Education. 2020;11(06):953-978
  115. 115. Kooli C. Chatbots in education and research: A critical examination of ethical implications and solutions. Sustainability. 2023;15(7):5614
  116. 116. Chen Y, Esmaeilzadeh P. Generative AI in medical practice: in-depth exploration of privacy and security challenges. Journal of Medical Internet Research. 2024;26:e53008
  117. 117. Sison AJG et al. ChatGPT: More than a “weapon of mass deception” ethical challenges and responses from the human-centered artificial intelligence (HCAI) perspective. International Journal of Human–Computer Interaction. 2023;14(9):1-20
  118. 118. Zhou J et al. Synthetic lies: Understanding AI-generated misinformation and evaluating algorithmic and human solutions. In: Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems. Hamburg, Germany: CHI; 2023
  119. 119. Rahimzadeh V et al. Ethics education for healthcare professionals in the era of ChatGPT and other large language models: Do we still need it? The American Journal of Bioethics. 2023;23(10):17-27
  120. 120. Abd-Alrazaq A et al. Large language models in medical education: Opportunities, challenges, and future directions. JMIR Medical Education. 2023;9(1):e48291
  121. 121. Hall E, Seyam M, Dunlap D. Identifying usability challenges in AI-based essay grading tools. In: Artificial Intelligence in Education. Posters and Late Breaking Results, Workshops and Tutorials, Industry and Innovation Tracks, Practitioners, Doctoral Consortium and Blue Sky. Cham: Springer Nature Switzerland; 2023
  122. 122. Mir MM et al. Application of artificial intelligence in medical education: Current scenario and future perspectives. Journal of Advances in Medical Education & Professionalism. 2023;11(3):133
  123. 123. Park JJ, Tiefenbach J, Demetriades AK. The role of artificial intelligence in surgical simulation. Frontiers in Medical Technology. 2022;4:1076755
  124. 124. Gao Y et al. A machine learning approach to predict surgical learning curves. Surgery. 2020;167(2):321-327
  125. 125. Azari DP et al. Modeling surgical technical skill using expert assessment for automated computer rating. Annals of Surgery. 2019;269(3):574
  126. 126. Brydon M et al. Virtual reality as a tool for eliciting empathetic behaviour in carers: An integrative review. Journal of Medical Imaging and Radiation Sciences. 2021;52(3):466-477
  127. 127. Hayasaka Y, Fujikara T, Kashimura M. Expectations for the next generation of simulated patients born from thoughtful anticipation of artificial intelligence-equipped robot. Journal of Nippon Medical School. 2018;85(6):347-349
  128. 128. Paranjape K et al. Introducing artificial intelligence training in medical education. JMIR Medical Education. 2019;5(2):e16048

Written By

Jacob A. Gould, Stanislaw P. Stawicki, Ryan Yimeng Lee and Anna Ng-Pellegrino

Submitted: 05 June 2024 Reviewed: 08 August 2024 Published: 16 September 2024