Review
Abstract
Background: Artificial intelligence (AI) is rapidly advancing in health care, particularly in medical imaging, offering potential for improved efficiency and reduced workload. However, there is little systematic evidence on process factors for successful AI technology implementation into clinical workflows.
Objective: This study aimed to systematically assess and synthesize the facilitators and barriers to AI implementation reported in studies evaluating AI solutions in routine medical imaging.
Methods: We conducted a systematic review of 6 medical databases. Using a qualitative content analysis, we extracted the reported facilitators and barriers, outcomes, and moderators in the implementation process of AI. Two reviewers analyzed and categorized the data separately. We then used epistemic network analysis to explore their relationships across different stages of AI implementation.
Results: Our search yielded 13,756 records. After screening, we included 38 original studies in our final review. We identified 12 key dimensions and 37 subthemes that influence the implementation of AI in health care workflows. Key dimensions included evaluation of AI use and fit into workflow, with frequency depending considerably on the stage of the implementation process. In total, 20 themes were mentioned as both facilitators and barriers to AI implementation. Studies often focused predominantly on performance metrics over the experiences or outcomes of clinicians.
Conclusions: This systematic review provides a thorough synthesis of facilitators and barriers to successful AI implementation in medical imaging. Our study highlights the usefulness of AI technologies in clinical care and the fit of their integration into routine clinical workflows. Most studies did not directly report facilitators and barriers to AI implementation, underscoring the importance of comprehensive reporting to foster knowledge sharing. Our findings reveal a predominant focus on technological aspects of AI adoption in clinical work, highlighting the need for holistic, human-centric consideration to fully leverage the potential of AI in health care.
Trial Registration: PROSPERO CRD42022303439; https://www.crd.york.ac.uk/PROSPERO/view/CRD42022303439
International Registered Report Identifier (IRRID): RR2-10.2196/40485
doi:10.2196/63649
Keywords
Introduction
Background
Advancements in the development of artificial intelligence (AI) have increased the accessibility and awareness of AI solutions in health care [,]. AI in health care has numerous potential applications, which can be categorized into 4 areas of application: diagnostics, therapeutics, administration and regulation, and population health management []. AI is mostly applied to data-driven tasks due to its ability to adapt to input data. It can process and analyze large volumes of health care data more quickly [,].
In the United States and Europe, AI technologies in health care can be categorized as software as a medical device, referring to software designed for medical purposes without requiring hardware integration []. These purposes, as defined by the Food and Drug Administration, encompass treating, diagnosing, curing, mitigating, or preventing diseases or conditions []. The growing recognition of the potential of AI algorithms in health care is supported by the surge of Food and Drug Administration approvals since 2016 for AI-enabled devices []. Notably, >75% of approvals are related to radiology []. These numbers are consistent with reports that highlight image-based disciplines at the forefront of AI integration in clinical practice due to their data-driven nature and continuously increasing workload demands [,,].
Despite the increasing availability of AI algorithms, there remains a limited understanding of their integration into clinical practice. A critical gap persists between broad research on algorithm development and limited evaluation of their actual use in clinical practice [,]. Most AI solutions are tested under controlled experimental conditions, which may underestimate the real-world impact of contextual factors on their utility and are therefore not necessarily transferable to clinical applications []. Depending on the users, the implementation process, and the clinical setting, the usefulness of AI solutions can significantly differ from previous evaluations or applications [,].
Complex sociotechnical systems, such as health care, “can be characterised by high uncertainty, multiple interacting elements and dynamic change” []. According to the sociotechnical systems theory, a sociotechnical system refers to the integration of humans, machines, environments, and organizational processes working together toward a shared objective. It consists of 2 interconnected subsystems: the technology subsystem, which encompasses tools and work organization, and the social subsystem, which involves individuals, teams, and coordination needs [,]. Sociotechnical frameworks of real-world clinical care offer a valuable approach to scrutinizing implementation complexities as well as the multiple intricacies of technology adoption [,].
A framework based on the sociotechnical systems theory that captures these complex demands and relations in health care settings is the Systems Engineering Initiative for Patient Safety (SEIPS) model []. The SEIPS model—most recently refined as SEIPS 3.0 []—proposes that sociotechnical systems consist of 5 interacting components: people, tasks, tools and technologies, organization, and environment. When one of the components changes, it affects the other components of the work system and subsequently the outcomes, that is, for patients, health care professionals, or organizations []. The model emphasizes the human as the center of the work system, which should be designed to support human performance and minimize negative impacts resulting from the work setting [,]. The SEIPS model can be applied to identify barriers and facilitators, which result from 1 element or the interaction between elements []. Hoonakker et al [] introduced the concept of dimensions, which can function as either facilitators or barriers.
While the SEIPS model is useful for understanding work system dynamics, other frameworks also help analyze health care technology implementation. The Consolidated Framework for Implementation Research (CFIR) evaluates implementation processes in health services through 5 domains: intervention characteristics, outer setting, inner setting, individual characteristics, and the implementation process, overlapping with SEIPS in addressing the involved people and their environment [,]. The nonadoption, abandonment, scale-up, spread, and sustainability (NASSS) framework examines factors influencing each of these outcomes and is specifically designed for technology implementation, while SEIPS covers broader work system design [,]. The integrate, design, assess, and share (IDEAS) framework, focusing on the full development cycle, is more suited for creating health technology solutions but less relevant to our study, which focuses on evaluating already implemented AI solutions []. The key distinction of SEIPS 3.0 is its human-centered approach, placing patients, clinicians, and caregivers at the core of the work system and emphasizing human-technology interaction and alignment in real-world clinical environments [].
A thorough understanding of how professionals in real-world clinical settings use AI technologies and how these tools can support their performance seems imperative, given the increasing availability of AI in health care []. While current literature extensively addresses the potential of AI in overviews and opinion articles, limited empirical evidence stems from actual clinical care [,-]. This leads to a critical lack of comprehensive understanding of AI implementation challenges and processes, potentially limiting the future development of evidence-based recommendations for successful AI technology implementation in clinical practice.
Objectives
Given the growing number of AI solutions in imaging-based disciplines, we aimed to explore and synthesize the existing literature on facilitators and barriers to AI implementation in routine medical imaging. We explored the relationships among AI implementation factors by drawing upon the SEIPS model. This approach allows for a concept-based and comprehensive synthesis of the available literature, generating a nuanced understanding of key process facilitators and barriers and their interactions in the implementation of AI technology into sociotechnical work systems in health care. Moreover, it contributes to a holistic picture of AI implementation in clinical work with consideration of important outcomes and moderating factors.
Methods
Registration and Protocol
Before starting, we registered our systematic literature review, which included qualitative analysis and synthesis, in the PROSPERO database (CRD42022303439) and published the review protocol (RR2-10.2196/40485) [].
The primary aim of this study was to assess and synthesize facilitators and barriers to AI workflow integration in medical imaging. This study was part of a larger review project on the impact of AI solutions on workflow efficiency in medical imaging, with a separate publication on the effect of AI on efficiency outcomes []. Our report follows the PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) reporting guidelines ().
Eligibility Criteria
We analyzed original clinical imaging studies in German or English published in peer-reviewed journals from January 2000 onward. Eligible studies implemented AI into real-world clinical workflows; therefore, we included observational and interventional studies (eg, randomized controlled trials) conducted in health care facilities using medical imaging. We focused on AI tools interpreting image data for disease diagnosis and screening.
We excluded dissertations, conference proceedings, and gray literature. In addition, due to our focus on real-world implementation of AI, we excluded studies conducted in experimental or laboratory settings.
Search Strategy
We searched the following electronic databases: MEDLINE (PubMed), Embase, PsycINFO, Web of Science, IEEE Xplore, and Cochrane CENTRAL. The databases were selected to reflect the interdisciplinary research on AI implementation in health care by including sources from medicine, psychology, and IT. Databases such as Cochrane, which only list systematic reviews or meta-analyses, were excluded in accordance with our eligibility criteria.
The detailed search strategy followed the PICO (population, intervention, comparison, and outcome) framework and can be found in the study by Wenderott et al []. The searches were performed on July 21, 2022, and on May 19, 2023. In a backward search, we identified additional relevant studies through screening the references of the included studies from the database search. Due to the time-consuming process of a systematic review with the in-depth qualitative analysis of the included studies, we performed an additional search on November 28, 2024, to identify relevant, recently published studies on facilitators and barriers to AI implementation in medical imaging []. This additional step ensured an update as well as the incorporation of interim published evidence on the topic. Further details are provided in [,-].
Screening and Selection Procedure
All gathered articles were imported into the Rayyan tool (Rayyan) [] for initial title and abstract screening. Two study team members (KW plus JK, MW, or Nikoloz Gambashidze), trained beforehand, individually assessed the titles and abstracts and reviewed their decisions in a consensus-oriented discussion. Subsequently, KW and JK screened the full texts of all eligible publications. Any disagreements regarding article inclusion were resolved through discussions with a third team member (MW). Exclusion reasons were documented and presented a flow diagram [].
Data Extraction
For qualitative data extraction, full texts of all eligible articles were imported into MAXQDA 22 (VERBI Software GmbH) []. This program allows users to mark text segments with different semantic codes, in this case the key characteristics, and automatically creates Excel (Microsoft Corporation) files of all the marked segments. Two researchers (JK and Fiona Zaruchas) extracted key study characteristics, including country, sample size, and any reported conflicts of interest (for more details, refer to the study protocol []). Countries and authors were imported into RStudio (2025.05.1+513; Posit PBC) to create a map of the geographical distribution [].
Regarding the reported stage and status of AI tool implementation in clinical practice, we used the studies by Bertram et al [] and Pane and Sarno [] to develop our classification of “level of implementation.” We defined 3 distinct levels: external validation, initial implementation, and full implementation (). We categorized all the included studies accordingly.
External validation
- Evaluation of the AI solution using real-world data
- Participants (ie, clinicians) recruited for the study
- Participants potentially blinded to other patient data
- Approximate simulation of the routine workflow
Initial implementation
- Partial implementation into the usual workflow
- Participants recruited in their usual work
- Different study groups possible
Full implementation
- Used for all eligible patients
- Implemented into the routine workflow of clinicians
Data Analysis
We applied a multistep procedure for data analysis. We first used a structured qualitative content analysis in a stepwise process []. In the initial phase, JK and KW independently classified the following key content categories of AI technology process factors in all the retrieved study texts:
- Facilitators, defined as “any factor that promotes or expands the integration or use of the AI system in the workflow” [].
- Barriers, defined as “any factor that limits or restricts the integration or use of the AI system” [].
- Outcomes of AI use, defined as the impact the AI use has on clinicians, patients, organizations, or the workflow.
- Moderators, defined as external factors, independent of the AI tool, that influence its use, for example, the setting or user [].
Subsequently, JK and KW engaged in a consensus-oriented discussion to reconcile all coded text segments [,]. In the following step, we defined subcategories following an inductive process. We noted a thematic overlap between topics being reported as a facilitator or barrier, depending on the study. Therefore, we decided to code categories that encompass facilitators as well as barriers, noting their valence (ie, positive or negative) separately. We organized the categories in a comprehensive codebook with corresponding definitions []. To establish consistency between raters throughout the coding process, the codebook underwent testing across 5 publications, where we discussed any coding issues and adjusted definitions as needed. Moving forward, both researchers (KW and JK) independently coded segments and subsequently discussed their codes to establish a consensus. Two researchers (KW and ARW) independently identified the proximally involved work system elements of the dimensions and then met to discuss their categorization and reached a consensus [,]. Using an inductive methodology, individual statements per dimension were clustered into themes that were mentioned frequently.
Epistemic Network Analysis
Epistemic network analysis (ENA) examines relationships between codes by modeling how frequently they co-occur in datasets. ENA was developed, validated, and widely applied in engineering education studies and has subsequently been used in research focused on human factors in health care [-]. ENA quantifies qualitative data by applying mathematics similar to social network analysis and principal component analysis to generate a weighted network of co-occurrences of codes. The matrix is then depicted graphically for each unit within the dataset. In each graph, the node size represents how frequently a code occurred in that unit; the thickness of the edges between the nodes corresponds to the weight, or frequency, at which a pair of codes co-occurred. The placement of each node is based on plotting vectors from the weighted co-occurrence matrix in a high-dimensional space, normalizing the vectors, reducing the dimensions using singular value decomposition (similar to principal component analysis), and then performing a rigid body rotation to preserve meaning. The x-axis is the dimension that accounts for the highest variation in the dataset, and the y-axis is a dimension orthogonal to the first that explains the next highest percentage of variance. Due to the preservation of meaning, these dimensions can be interpreted conceptually based on the qualitative data analysis. The fit of the resulting model can be evaluated both with Spearman and Pearson correlation coefficients. Importantly, ENA evaluates all networks concurrently, yielding a collection of networks that can be compared both visually and statistically. For more details on the method, including the mathematics and validation, please refer to the studies by Andrist et al [], Bowman et al [], Shaffer [], Shaffer et al [], and Shaffer and Ruis [].
ENA serves as a valuable method to analyze and visualize the findings of our qualitative content analysis, that is, the co-occurrence of the dimensions of facilitators or barriers in the included studies [,-]. In this study, we used the ENA web tool (version 1.7.0) []. The data were uploaded to the ENA web tool in a .csv file, with each row representing a barrier or facilitator identified through qualitative analysis; the columns included metadata such as the study, type of implementation, if that row contained a barrier or a facilitator, the dimension that specific barrier or facilitator was categorized as, and the coded excerpt from the study. ENA was used to generate 6 network graphs that depict the relationships between barriers or facilitators reported in each study, separated by the level of implementation. Thus, in each graph, the node size corresponds to the frequency that a barrier or facilitator occurred across all studies in that type of implementation; the thickness of the edges between nodes indicates how often a pair of barriers or facilitators co-occurred within the same study.
Results
Study Selection
We identified 22,684 records in the databases and an additional 295 articles through a backward search. After the removal of duplicates, 13,756 remaining records were included in the title and abstract screening. Afterward, 207 full texts were screened, of which 169 were excluded primarily because they did not meet the inclusion criteria, that is, experimental studies or studies not focusing on AI tools for interpreting imaging data (for more details, refer to the study by Wenderott et al []). A total of 10 studies were excluded because they did not describe any facilitator or barrier in the course of clinical implementation. Finally, 38 studies were included in the review and data extraction. A PRISMA flowchart is presented in .

Study Characteristics
Of the 38 included studies, 24 (63%) were performed in a single institution and 14 (37%) were multicenter studies. Only 5% (2/38) of the studies were published before 2012, whereas all others (36/38, 95%) were published from 2018 onward. The geographical distribution of the studies is depicted in . On the basis of the heterogeneity in the regulatory frameworks of AI in health care, we included a comparison across dimensions between the 2 main geographical clusters, the European Union and the United States ( [-]). Most studies (25/38, 66%) were conducted in radiology, followed by gastroenterology (5/38, 13%; ). A total of 47% (18/38) of the studies reported a potentially relevant conflict of interest. For the risk of bias assessment, we used the Risk of Bias in Nonrandomized Studies of Interventions tool and the Cochrane Risk of Bias version 2 tool for the 1 included randomized study [,]. From the included 37 nonrandomized studies, only 1 (3%) study was classified as having a low risk of bias. In total, 11% (4/37) of the studies were rated as having a moderate risk, 65% (24/37) of the studies had a serious risk, and 22% (8/37) of the studies were assessed as having a critical risk of bias. The included randomized study was determined to have a high overall risk of bias. For a detailed risk of bias and quality of reporting assessment, refer to the supplementary material of the study by Wenderott et al [].

| Study | Data collection | Source of data | Professionals, n | Cases, patients, or scans, n | Level of implementation |
| Arbabshirani et al [] | Prospective | No information | Radiologists (not specified) | 347 patients | Full |
| Batra et al [] | Retrospective | Time stamps | 32 radiologists | 2501 examinations of 2197 patients | Full |
| Carlile et al [] | Prospective | Survey | 112 EDa physicians | 1855 scans and a survey on 202 scans | Initial |
| Cha et al [] | Prospective | Survey | 18 physicians | 173 patients | Full |
| Cheikh et al [] | Retrospective | Performance metrics and survey | 79 radiologists | 7323 examinations | Initial |
| Chen et al [] | Retrospective | Performance metrics and time measurement | 4 radiologists | 85 patients | External |
| Conant et al [] | Retrospective | Performance metrics and time measurement | 24 radiologists (including 13 breast subspecialists) | 260 cases | External |
| Davis et al [] | Prospective | Time stamps | Radiologists (not specified) | 50,654 cases | Full |
| Diao et al [] | Prospective | Time stamps and survey | 7 radiologists | 251 patients | Initial |
| Duron et al [] | Retrospective | Performance metrics and time stamps | 6 radiologists and 6 ED physicians | 600 cases | External |
| Elijovich et al [] | Retrospective | Chart review | Neurologists and neurointerventionalists (not specified) | 680 patients | Full |
| Ginat [] | Retrospective | Time stamps | 5 radiologists | 8723 scans | Initial |
| Hassan et al [] | Retrospective | Chart review | Technologists, radiologists, ED physicians, neurologists, and interventionalists (not specified) | 63 patients | Full |
| Jones et al [] | Prospective | Survey | 11 radiologists | 2972 scans of 2665 patients | Initial |
| Ladabaum et al [] | Retrospective | Chart review | 52 endoscopists | 2329 patients | Initial |
| Levy et al [] | Retrospective | Performance metrics and time stamps | 30 gastroenterologists | 4414 patients | Full |
| Marwaha et al [] | Retrospective | Survey | Genetic counselors and trainees (15 in total) | 72 patients | Initial |
| Mueller et al [] | Prospective | Observation, interview, and survey | 2 radiologists | 90 scans | Full |
| Nehme et al [] | Prospective | Performance metrics, time stamps, and surveys | Endoscopists and staff members (45 in total) | 1041 patients | Initial |
| Oppenheimer et al [] | Prospective | Performance metrics | 2 radiologists | 1163 examinations of 735 patients | Full |
| Pierce et al [] | Retrospective | Case review | Radiologists (not specified) | 30,847 examinations | Full |
| Potrezke et al [] | Prospective | Performance metrics | 49 radiologists and 12 medical image analysts | 170 cases of 161 patients | Initial |
| Quan et al [] | Prospective | Performance metrics and time measurement | 6 endoscopists | 600 patients | Full |
| Raya-Povedano et al [] | Retrospective | Performance metrics and workload | 5 breast radiologists | 15,986 patients | External |
| Ruamviboonsuk et al [] | Prospective | Performance metrics and surveys | Staff members and nurses (12 in total) | 7651 patients | Full |
| Sandbank et al [] | Prospective | Performance metrics | Pathologists (not specified) | 5954 cases | Full |
| Schmuelling et al [] | Retrospective | Performance metrics and time stamps | Radiologists (not specified) | 1808 scans of 1770 patients | Full |
| Seyam et al [] | Retrospective | Performance metrics and time stamps | Radiologists (not specified) | 4450 patients | Full |
| Tchou et al [] | Prospective | Observation | 5 radiologists | 267 cases | External |
| Tricarico et al [] | Prospective | Performance metrics | Radiologists (not specified) | 2942 scans | Initial |
| Vassallo et al [] | Retrospective | Observation and performance metrics | 3 radiologists | 225 patients | External |
| Wang et al [] | Prospective | Performance metrics and time measurement | 8 endoscopists | 1058 patients | External |
| Wang et al [] | Retrospective | Chart review | 2 radiologists | 2120 patients | External |
| Wittenberg et al [] | Retrospective | Performance metrics and time measurement | 6 radiologists | 209 patients | External |
| Wong et al [] | Prospective | Survey | Radiation therapists and oncologists (39 in total) | 174 cases | Full |
| Wong et al [] | Prospective | Performance metrics and survey | Radiologists and internists (17 in total) | 214 scans | Initial |
| Yang et al [] | Prospective | Performance metrics and time measurement | Ophthalmologists | 1001 patients | Initial |
| Zia et al [] | Prospective | Performance metrics, time stamps, and survey | 49 radiologists | 1446 scans | Initial |
aED: emergency department.
Regarding the level of AI implementation, we identified 24% (9/38) of the studies evaluating external validation, 34% (13/38) of the studies focusing on initial implementation, and 42% (16/38) of the studies focusing on an AI tool being fully integrated in the clinic. presents the key characteristics of all the included studies. There was a substantial variety of AI technologies, with 42% (16/38) of the studies using commercial AI solutions and 55% (21/38) of the studies evaluating self-developed tools (1 study did not specify the source of the AI solution []). More details about the AI tools are provided in [-]. The methods that were most frequently used were the analysis of performance metrics (21/38, 55%) or time stamps (10/38, 26%). In total, 29% (11/38) of the studies used some form of survey or questionnaire to gather the opinions and experiences of clinicians. Most commonly, they used self-reports on the impact of AI use on the diagnosis and efficiency, followed by their attitude toward AI, their satisfaction or usefulness, as well as the usability of the AI tool. Notably, only the study by Jones et al [] used an established tool, that is, the Systems Usability Scale. Further details on the surveys described in the studies are provided in [,,,,-,, ,,].
Facilitators and Barriers to AI Implementation
Identification and Classification of Process Factors (Qualitative Content Analysis Results)
Overview
Drawing upon the qualitative analyses of the included studies, we identified 180 statements from the included publications that described the factors influencing AI implementation in clinical practice. These statements were systematically categorized into 12 overarching dimensions, as described in detail in . Within each dimension, we clustered recurring themes. This resulted in a total of 37 themes; the details and example quotations from the studies are listed in [-]. Many themes were stated simultaneously as facilitators and barriers, mostly depending on the presence or absence of the mentioned theme in the study (). For example, the theme impact on decision-making was referenced positively in the study by Cheikh et al []:
Radiologists stressed the importance of AI to strengthen their conclusions, especially to confirm negative findings, or to ensure the absence of distal PE [pulmonary embolism] in poor-quality examinations.
In contrast, Oppenheimer et al [] stated the following:
In some edge cases, both residents reported feeling somewhat unsure of their diagnosis, in particular if they decided on a fracture and the AI result was negative.
With 64% (115/180) of the segments, we identified more facilitators in general than barriers (65/180, 36% segments). The dimensions attitudes and values and stakeholder engagement were mostly stated as facilitators, highlighting their positive impact on AI implementation. Medicolegal concerns was the only dimension that was exclusively mentioned as a barrier. In the subsequent sections, we describe the 3 dimensions with the most frequently coded segments in more detail.

| Dimensions | Definition | Codes, n | Work system elements | |||||
| People | Tasks | TTa | Organization | PEb | EEc | |||
| Evaluation of AI use | Clinicians’ or patients’ evaluation of the usefulness of the AI tool impacting its integration. | 37 | ✓ | ✓ | ✓ | |||
| Fit into the workflow | The AI is embedded into the workflow or processes of the local health care facility, including both clinical workflows and technical aspects such as data processing. | 29 | ✓ | ✓ | ✓ | |||
| Implementation procedure | The AI implementation follows an implementation protocol or a prespecified plan, including users receiving training on the AI tool. | 24 | ✓ | ✓ | ✓ | |||
| Explainability of AI | The capability of understanding and justifying the decisions made by the AI tool. | 13 | ✓ | |||||
| Attitudes and values | The beliefs, ethical principles, judgments, or priorities that might have been present before using AI influence clinicians’ acceptance, adoption, and use of AI. | 12 | ✓ | ✓ | ||||
| Interoperability | Ensures that AI can seamlessly communicate and share data with other technologies used. | 12 | ✓ | ✓ | ||||
| Stakeholder involvement | In the course of implementing or using AI, important stakeholders are included in the process. | 12 | ✓ | ✓ | ✓ | ✓ | ||
| Usability | Users can interact effectively and intuitively with the AI tool to accomplish their goals. | 12 | ✓ | ✓ | ||||
| Reliability | The reliability of the AI tool that impacts its use in the workflow. | 11 | ✓ | |||||
| Individual work organization | Fit of the AI tool with the individual preferences of the users’ work organization. | 7 | ✓ | ✓ | ✓ | |||
| Impact on the role of clinicians | AI use alters the role of clinicians, how they perceive autonomy, and whether they feel responsible for their diagnosis. | 6 | ✓ | ✓ | ✓ | |||
| Medicolegal concerns | Intersection of medical practice and legal regulations, mitigation of legal risks, and safeguarding of patients and their rights when using the AI tool. | 5 | ✓ | ✓ | ✓ | |||
aTT: tools and technologies.
bPE: physical environment.
cEE: external environment.
Evaluation of AI Use
The dimension evaluation of AI use reflected whether a positive or negative evaluation of the use of the AI solution aided the AI integration. This dimension was most frequently mentioned, reflecting the focus of the included studies on AI evaluation in clinical practice. We identified people, tasks, and tools and technologies as proximally involved work system elements. Two themes emerged in this dimension. Overall, the usefulness was the most frequently mentioned theme. This is supported by evidence that perceived usefulness or performance expectancy are strong determinants of the actual use of technologies [,], focusing on the behavior of users. The impact on decision-making emerged as a second theme in this dimension. Positively, clinicians valued the support provided by the AI tool, as AI use can increase the confidence of clinicians []. Negatively, the studies mentioned risks, such as alert fatigue [], over trust [,], or insecurities due to diverging diagnostic decisions [].
Fit Into the Workflow
The dimension fit into the workflow focused on how well AI technology fits into the workflow, which is an important factor to consider during the implementation of a novel technology [,]. The proximally involved work system elements were tasks, tools and technologies, and organization. In this dimension, 5 themes were identified. The most frequently and favorably mentioned theme was the accessibility of results, for example, by results being forwarded automatically to the clinicians [] or providing a notification platform []. This also applied to the theme of data processing, where automatic and fast processing was a facilitating factor [,,,,]. Regarding the themes distractions or disruptions due to AI, the facilitating factors were characterized by the absence of these, whereas the barriers reflected the negative influence of the AI tool on the workflow of the users, for example, through alarms that potentially distracted the clinicians. The theme additional work steps was only mentioned in the study by Batra et al [].
Implementation Procedure
The dimension implementation procedure focused on the descriptions of the implementation process to install the AI system in the clinical workflow. The related work system elements were people, tools and technologies, and organization. In this dimension, the themes internal testing of the AI tool; continuous maintenance, that is, the ongoing monitoring of the AI tool with adaptations if necessary; and the training of users were exclusively mentioned as facilitators. Of the 38 studies, only 3 (8%) described a deployment strategy [,,], with Ladabaum et al [] describing that their minimalist approach was not sufficient to successfully implement the AI tool. In total, 13% (5/38) of the studies discussed the strategies or preconditions to the technology readiness of the organization, which can be defined as the willingness to “embrace and use new technologies to accomplish goals.... It is a combination of positive and negative technology-related beliefs” []. In the study by Ruamviboonsuk et al [], the authors encountered the challenge that the hospital was still working with paper-based records, and the internet connectivity was slow, highlighting the role of the pre-existing digital infrastructure.
Comparison of Facilitators and Barriers Across the Levels of Implementation (Results of ENA)
We used ENA to model the differences in facilitators and barriers across the level of implementation, resulting in 6 distinct network graphs (). The axes identified in our ENA can be associated with work system elements of the SEIPS model []. The x-axis represents the work system element people in the negative direction, as indicated by the dimensions attitudes and values and stakeholder involvement being the farthest in this direction, and the work system element technology in the positive direction, which we concluded from the dimensions reliability, interoperability, and usability presented in this direction. For the x-axis and the y-axis, the coregistration correlations were 1 (both Pearson and Spearman), showing a strong goodness of fit []. The x-axis accounted for 37.2% of the variance. The y-axis accounted for 21% of the variance. The positive direction of the y-axis can be associated with the work system element tasks, with the ENA showing the dimension usability as the farthest node in this direction. In contrast, the negative side of the y-axis represents the work system element organization, which we inferred from the dimensions fit into the workflow and interoperability being the most distant nodes in this direction.
For the studies describing external validations of AI solutions, a total of 19 coded segments (segments per study: mean 2.11, SD 1.27; median 2, IQR 1-2) were included in the ENA. The resulting networks showed a small number of involved dimensions and connections, highlighting the dimensions evaluation of AI use and explainability of AI as facilitators and the dimension usability as a barrier (A and 4D).
For the initial implementation studies, we analyzed 85 coded segments (segments per study: mean 6.54, SD 4.74; median 5, IQR 3-9). The facilitators showed an accumulation in the quadrant of the work system elements tasks and people, with the dimensions implementation procedure and evaluation of AI use being the largest nodes. The strongest connection for the facilitators was between the dimensions evaluation of AI use and implementation procedure, whereas the strongest connection for the barriers was between the dimensions evaluation of AI use and attitudes and values, with the dimension implementation procedure being also mentioned frequently (B and 4E).
Regarding the publications reporting the full implementation of AI solutions, the network graphs were based on 76 coded segments (segments per study: mean 4.75, SD 4.11; median 3.5, IQR 2.5-7). The frequently mentioned facilitators were the dimensions fit into the workflow and evaluation of AI use, with a strong connection between these dimensions (C). The barriers centered on the dimension reliability, with a strong connection to the dimension fit into the workflow (F).

Reported Outcomes of AI Implementation
The included studies examined various outcomes stemming from the implementation of AI tools in medical imaging tasks. Of the 38 included studies, 31 (82%) reported efficiency outcomes, with 71% (22/31) of the studies showing enhanced efficiency, while 6% (2/31) of the studies reported a negative impact, and 23% (7/31) of the studies indicated no changes in efficiency. 13% (5/38) of the included studies assessed the impact of AI on workload or required work steps, with 80% (4/5) of the studies reporting reductions and 20% (1/5) of the studies indicating an increase. Of the 38 included studies, 16 (42%) reported on the performance of AI solutions in terms of changes in detection rates, need for human oversight, or quality of the AI-based results. In addition, 34% (13/38) discussed outcomes for patients, such as enhanced safety or quality control due to AI; a reduced time to diagnosis or treatment; prolonged stay in the emergency department; and increased detection rates, possibly leading to additional unnecessary treatments or increased workload []. The full details on the reported study outcomes are provided in [-,,, -].
Moderating Factors of AI Implementation
Of the 38 included studies, 18 (47%) identified moderators, which are defined as factors that influence AI use but are independent of the AI itself, such as the setting or the users. Details on the studies reporting moderators are provided in [,,,,,-,-,,, ,,,,].
The setting, precisely the shifts, times of the day, or whether it was a weekday or a weekend, was mentioned by 5% (2/38) of the studies [,]. Schmuelling et al [] and Wong et al [] also highlighted the significant influence of the clinical environment or pre-existing clinical workflows on AI implementation [,].
In addition, 21% (8/38) of the studies described that the implementation and use of AI are impacted by how health care professionals use the AI system, such as through personal preferences concerning their workflow or change in behaviors when they are not being observed. In total, 11% (4/38) examined the impact of human behavior on the evaluation of AI solutions in terms of interobserver variability or the missing reporting of errors.
In total, 26% (10/38) of the studies listed task-related factors, for example, differences due to input image quality, task type, or criticality of the findings. Moreover, 18% (7/38) of the studies noted that job experience or familiarity with AI has an impact on AI use.
Of the 38 included studies, 5 (13%) investigated physician performance when using AI regarding their job experience, with 20% (1/5) of the studies reporting no association []. Furthermore, 40% (2/5) of the studies reported a more positive AI use evaluation [,] or an enhanced detection rate [] for less experienced readers, while 20% (1/5) of the studies reported that “the time to review the CAD images increased with the experience of the reader” [].
Additional Search to Include Recent Evidence
We searched 6 databases (PubMed, Web of Science, Embase, CENTRAL, Cochrane, and IEEE Xplore) to further identify recently published, relevant evidence, including review articles, in contrast to our original review process. While we retrieved and screened 1016 records, we identified 9 studies investigating facilitators and barriers of AI implementation in medical imaging. Among the 9 studies, 5 (56%) were scoping reviews, with 40% (2/5) of them focusing on AI implementation in health care in general [,], 40% (2/5) of the reviews studying AI for breast imaging [,], and 20% (1/5) of the reviews focusing on AI in radiology []. Only Chomutare et al [] used a theoretical framework, the CFIR, to guide their analysis. All reviews provided a narrative synthesis of the results. In addition, of the 9 studies retrieved through the additional search, we identified 4 (44%) original studies, all using interviews as a qualitative methodology for studying facilitators and barriers of AI in medical imaging. Among those, 50% (2/4) of the studies did not study a specific AI implementation [,] and the other 50% (2/4) of the studies focused on specific AI solutions and were published after our second search [,]. Further details on these studies are provided in .
Discussion
Principal Findings
Our systematic review provides, to the best of our knowledge, the first qualitative and quantitative synthesis that analyzes facilitators and barriers reported in studies on AI implementation in real-world clinical practice. Using our differentiation between the 3 levels of implementation, we were able to delve into the complexities of transferring AI technologies from model development and testing into the actual clinical environment []. To strengthen our conclusions, we used the SEIPS model, which is a strong asset for the system-based analysis of health care work environments []. In our analysis, we found that the frequency of various facilitators and barriers differed significantly across the stages of implementation. However, a consistently wide range of factors was identified, emphasizing the complex interplay of various elements when integrating AI into routine care processes. Consequently, our study offers a consolidated list of key factors that should be considered during AI implementation.
Focusing on categories across the implementation levels and matching them to work system elements can guide future implementation processes. In the conducted ENAs, the work system elements tasks, tools and technology, organization, and people were associated with the different axes, which provided a visualization of the importance of interactions between the work system elements. Missing in this categorization was the work system element physical environment, likely due to the diverse study settings and minimal impact of AI on work environments in the included studies. All studies focused on software as a medical device solutions that mostly did not alter their physical environment, and only 2 studies [,] reported physical changes because the AI solution was displayed on separate monitors. Referring to our resulting network graphs (), it is noteworthy that the dimension implementation procedure was linked to work system elements tasks and people, while typically it is associated with organizational decisions [,]. Our classification showed that the included studies focused on evaluating AI on a microsystem level, that is, the individual health professionals and the tasks associated with AI use [,].
Studies describing external validations of AI solutions reported facilitators mostly related to the dimension evaluation of AI use, which was also the most prominent dimension overall. Barriers often stemmed from the AI technology itself, especially from the issues with usability. The focus of these networks highlights that external validation is still a part of the algorithm development process in which the clinical applicability of the AI solutions is being assessed. This is also supported by the outcomes reported in these studies, which were mostly time related, such as efficiency, treatment times, or workload. Moderating factors were not very prominent in these studies and were predominantly task related. These studies usually test the algorithm’s interaction with various work system elements for the first time under realistic conditions, which is often not done during the AI development phase before clinical validation [].
Studies focusing on the initial implementation tested how AI solutions can be fitted into the existing workflow, while not yet being applied to all patients or cases. Barriers and facilitators in these studies mainly focus on the work system elements people and tasks, with most connections in the ENA stemming from this quadrant. In addition, these studies presented a broader spectrum of outcomes, such as satisfaction or patient outcomes. Moderating factors to AI use in these studies were also diverse, including experience of clinicians and their behavior. This focus aligns with the SEIPS model, which prioritizes the people and a human-centered design []. This resonates well with the identified initial implementation studies that tested and studied AI integration into the work system, and determined the necessary optimizations. The rising recognition of the significance of human-centered design and stakeholder engagement in the adoption of AI in health care is supported by our findings [,,-].
In the network analysis of studies assessing AI solutions that have been fully integrated into routine care, the dimension fit into the workflow emerges as the largest node of facilitators, with also the most connections, supporting the literature that highlights the integration of AI into work processes as crucial for success [,,]. The themes we observed as being most important were accessibility of results and no disruptions due to AI, with the latter being mentioned positively by the absence of AI-related disruptions to the workflow. As workflow disruptions can increase the procedure duration, this is highly relevant in medical imaging, as radiologists and other physicians face increasing workloads and time pressures due to the large amount of medical imaging data to be interpreted [,]. Interestingly, barriers in these studies showed a strong connection between the dimensions reliability and fit into the workflow. This aligns with our recent findings that technical issues can largely impact the workflow, contrasting with the literature that often emphasizes ethical debates, medicolegal concerns, or AI explainability, which were less prominent in our analysis [,]. Nevertheless, most outcomes reported in these studies were positive, such as increased efficiency, improved detection rates, or reduced treatment times, potentially reflecting that only the AI solutions that have overcome most barriers manage the transfer from the initial development stage to full implementation [].
Comparison to Previous Work
Compared to previous research in the field, our results contribute important insights and show consistencies and discrepancies in AI implementation research. Few reviews have focused on the implementation of AI in clinical practice, and even fewer have specifically examined the facilitators and barriers to AI implementation. In our additional search, we only identified 5 scoping reviews targeting this topic in relation to AI for medical imaging. Hassan et al [] provided a recent review on the facilitators and barriers to AI adoption, noting that most of the included studies focused on radiology and oncology. The authors identified 18 categories of facilitators and barriers, and similar to our findings, they observed that the same factor can be described as both a facilitator and a barrier []. However, because Hassan et al [] do not offer a detailed overview of the included studies and only present a narrative synthesis, the comparison with our included studies, their settings, and designs is limited.
Lokaj et al [] reviewed AI development and implementation for breast imaging diagnosis, identifying clinical workflow as a key facilitator. However, they emphasized technical aspects and algorithm development, with barriers such as data, evaluation, and validation issues. They noted the inclusion of very few prospective studies. In contrast, our review focuses on AI solutions evaluated after the development phase, in real-world clinical settings; therefore, technical aspects do not play a significant role in our developed set of facilitators and barriers.
Chomutare et al [] also reviewed AI implementation in health care using the CFIR focusing on late-stage implementations. Despite including only 19 studies, they identified dimensions similar to ours, such as interoperability and transparency. Using ENAs based on implementation levels, our study provides a detailed overview of the facilitators and barriers at different implementation stages. Our findings further support the claim of Chomutare et al [] that limited knowledge exists about the clinicians working with AI. Our review found that 29% (11/38) of the included studies incorporated user feedback, revealing a significant research gap. This underscores the need for research to adopt human-centered design, defined by the International Organization for Standardization standard 9241-210:2019 as follows: “an approach to interactive systems development that aims to make systems usable and useful by focusing on the users, their needs and requirements, and by applying human factors/ergonomics, and usability knowledge and techniques. This approach enhances effectiveness and efficiency, improves human well-being, user satisfaction, accessibility and sustainability; and counteracts possible adverse effects of use on human health, safety and performance” []. Using human-centered design principles is crucial for developing AI systems that benefit clinicians and patients [,].
Factors influencing AI adoption in health care are similar to those for other health information technologies, for example, electronic health records or e-prescription systems [-]. Key success factors, such as stakeholder involvement and system usability, are comparable across these technologies [,]. Recommendations for AI implementation can be drawn from health information technology research, such as that by Yen et al [], who emphasize the importance of the sociotechnical context and longitudinal studies over cross-sectional outcomes. Although few of our included studies reported on the implementation process over time, our network analyses by implementation level can help identify the criteria that must be met in the course of AI tool transitions from research to clinical practice. AI introduces unique considerations to health care workflows, such as shared decision-making and human oversight [], and presents new challenges requiring a broader understanding of the technology [].
Clinicians need to understand the data used to train AI tools, as biases and limitations can arise, a point highlighted by Pierce et al [] through their educational campaign before AI implementation. As AI solutions present the possibility of algorithmic bias, which might not be detected by clinicians, it is noteworthy that we identified user training and transparency as facilitators of AI implementation. The diverse nature of algorithmic biases, for example, stemming from biased training data, data gaps on underrepresented groups, human bias of the developers, or a lack of data standards, is an important information to be considered by the users [-]. Algorithmic bias holds the potential for patient harm, especially for populations considered disadvantaged []. While we identified strategies that can limit the impact of bias, such as user training, continuous monitoring, or transparency, most of the included studies did not explicitly mention bias, as described in by Wenderott et al []. Beyond algorithmic bias, it is also essential to address the legal and ethical challenges surrounding AI-supported decisions in health care []. Although these topics are widely discussed in research and politics, only 13% (5/38) of the studies we reviewed discussed medicolegal concerns in terms of data privacy concerns and legal implications. Thus, although AI solutions have been successfully implemented into routine medical care, issues of liability remain unresolved [,]. As AI continues to evolve and becomes more integrated into clinical practice, it is crucial to carefully consider these factors to ensure its safe, effective, and responsible use in health care settings.
Limitations
Our study has a few limitations worth noting. First, we focused exclusively on AI tools in medical imaging, aiming to ensure the comparability of our findings. However, we encountered significant diversity in study settings, AI solutions, and purposes for decision-making or diagnostics. Because we only reviewed peer-reviewed original studies, some evaluations of AI implementation in health care might have been missed. Second, our findings showed more facilitators than barriers, which could be associated with a potential publication bias toward a more positive reporting of AI implementation, especially in combination with the high number of studies that reported a conflict of interest. In addition, we only searched for peer-reviewed literature, possibly missing reports on AI implementation from gray literature. AI implementation might also occur in clinical practice without scientific evaluation or reporting of results, which could also contribute to a publication bias. Third, the rapidly evolving nature of AI research indicates that certain processes or issues discussed in the studies may already be outdated by the time of publication, a challenge particularly relevant to the time-consuming process of systematic reviews, which often face delays from the literature search to final publication []. Therefore, while our review provides the first comprehensive, thorough, and methodologically rigorous overview of the facilitators and barriers to AI implementation in medical imaging, we recommend that future studies consider adopting shorter review cycles to ensure more timely publication and greater relevance in light of ongoing technical advancements. Fourth, facilitators and barriers were mainly extracted from study discussions, with separate reporting being rare, possibly introducing bias. In general, we noted that the descriptions of the implementation procedure and setting were sparse. Future research should provide details on their implementation strategy, processes, and subsequent adjustments to best integrate technology into the unique workflow []. This would enable comparisons across studies and facilitate learning in the scientific community. In addition, our established dimensions were formed inductively, requiring further validation. Fifth, while we used the SEIPS model for our analysis, we acknowledge that other frameworks exist such as the CFIR; the IDEAS framework; or the NASSS framework [,,]. We planned to use the NASSS as specified in the review protocol but eventually chose the SEIPS model due to its human-centered and system-based approach []. Finally, our focus was on real-world investigations in clinical settings. Although our classification of “level of implementation” was useful for comparing different studies, its applicability to other clinical tasks, medical specialties, and work settings needs further examination. Furthermore, future studies should explore the impact of regulatory settings on research outcomes. While this was not feasible in our review due to the limited number of studies, the growing number of available AI algorithms and academic publications on AI in medicine will potentially provide sufficient data for these analyses [,].
Conclusions
In conclusion, the facilitators and barriers identified in medical imaging studies have produced a comprehensive list of dimensions and themes essential for AI implementation in clinical care. Our research underscores the pressing necessity for holistic investigations into AI implementation, encompassing not only the technical aspects but also their impact on users, teams, and work processes. Furthermore, our results corroborate the future need for transparent reporting of AI implementation procedures. This transparency fosters knowledge exchange within the scientific community, facilitating the translation of research findings into actionable strategies for clinical care. A deeper understanding of how AI solutions affect clinicians and their workflows can help reduce clinician workload and improve patient care.
Acknowledgments
This work was supported by a fellowship from the Deutscher Akademischer Austauschdienst (DAAD; German Academic Exchange Service) awarded to KW. The publication of this work was supported by the Open Access Publication Fund of the University of Bonn. The authors sincerely thank Dr Nikoloz Gambashidze and Fiona Zaruchas (Institute for Patient Safety, University Hospital Bonn) for helping with the title and abstract screening and data extraction. During the preparation of this paper, the authors used ChatGPT (version GPT-3.5, OpenAI) to optimize the readability and wording of the manuscript. This was done by asking ChatGPT for synonyms or the spelling of single words or for sentences using prompts such as “Can you check for spelling or grammar mistakes?” or “Can you enhance the readability of this sentence?” (). After using this tool, the authors reviewed and edited the content as required and take full responsibility for the content of the paper.
Data Availability
The datasets generated or analyzed during this study are available from the corresponding author on reasonable request.
Authors' Contributions
KW was responsible for conceptualization. data curation, formal analysis, investigation, methodology, project administration, software development, and visualization. KW also led the writing of the original draft and contributed to the preparation, review, and editing of the manuscript. JK contributed to data curation, investigation, visualization, and the review and editing of the manuscript. MW was involved in conceptualization, funding acquisition, supervision, validation, and manuscript review and editing. ARW contributed to methodology, software development, supervision, validation, and the review and editing of the manuscript.
Conflicts of Interest
None declared.
PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) checklist.
DOCX File , 53 KBAdditional search.
DOCX File , 78 KBGeographical comparison.
DOCX File , 19 KBArtificial intelligence solutions.
DOCX File , 60 KBOverview on surveys used in the included publications.
DOCX File , 32 KBDetails on the extracted themes.
DOCX File , 110 KBOutcomes extracted from the included publications.
DOCX File , 62 KBModerators extracted from the included publications.
DOCX File , 55 KBChatGPT transcript.
DOCX File , 32 KBReferences
- AlZaabi A, AlMaskari S, AalAbdulsalam A. Are physicians and medical students ready for artificial intelligence applications in healthcare? Digit Health. Jan 26, 2023;9:20552076231152167. [FREE Full text] [CrossRef] [Medline]
- Beets B, Newman TP, Howell EL, Bao L, Yang S. Surveying public perceptions of artificial intelligence in health care in the United States: systematic review. J Med Internet Res. Apr 04, 2023;25:e40337. [FREE Full text] [CrossRef] [Medline]
- He J, Baxter SL, Xu J, Xu J, Zhou X, Zhang K. The practical implementation of artificial intelligence technologies in medicine. Nat Med. Jan 2019;25(1):30-36. [FREE Full text] [CrossRef] [Medline]
- Artificial intelligence and machine learning in software as a medical device. U.S. Food & Drug Administration. URL: https://www.fda.gov/medical-devices/software-medical-device-samd/artificial-intelligence-and-machine-learning-software-medical-device [accessed 2022-04-29]
- Pesapane F, Codari M, Sardanelli F. Artificial intelligence in medical imaging: threat or opportunity? Radiologists again at the forefront of innovation in medicine. Eur Radiol Exp. Oct 24, 2018;2(1):35. [FREE Full text] [CrossRef] [Medline]
- Software as a Medical Device (SaMD): key definitions. International Medical Device Regulators Forum. Dec 9, 2013. URL: https://www.imdrf.org/sites/default/files/docs/imdrf/final/technical/imdrf-tech-131209-samd-key-definitions-140901.pdf [accessed 2025-06-15]
- Proposed regulatory framework for modifications to artificial intelligence/machine learning (AI/ML)-based software as a medical device (SaMD) - discussion paper and request for feedback. U.S. Food & Drug Administration. 2019. URL: https://www.fda.gov/media/122535/download [accessed 2025-06-16]
- Joshi G, Jain A, Araveeti SR, Adhikari S, Garg H, Bhandari M. FDA-approved artificial intelligence and machine learning (AI/ML)-enabled medical devices: an updated landscape. Electronics. Jan 24, 2024;13(3):498. [CrossRef]
- Ahmad OF, Mori Y, Misawa M, Kudo SE, Anderson JT, Bernal J, et al. Establishing key research questions for the implementation of artificial intelligence in colonoscopy: a modified Delphi method. Endoscopy. Sep 2021;53(9):893-901. [FREE Full text] [CrossRef] [Medline]
- Wolff J, Pauling J, Keck A, Baumbach J. Success factors of artificial intelligence implementation in healthcare. Front Digit Health. Jun 16, 2021;3:594971. [FREE Full text] [CrossRef] [Medline]
- Yin J, Ngiam KY, Teo HH. Role of artificial intelligence applications in real-life clinical practice: systematic review. J Med Internet Res. Apr 22, 2021;23(4):e25759. [FREE Full text] [CrossRef] [Medline]
- Wenderott K, Krups J, Luetkens JA, Gambashidze N, Weigl M. Prospective effects of an artificial intelligence-based computer-aided detection system for prostate imaging on routine workflow and radiologists' outcomes. Eur J Radiol. Jan 2024;170:111252. [FREE Full text] [CrossRef] [Medline]
- Asan O, Bayrak AE, Choudhury A. Artificial intelligence and human trust in healthcare: focus on clinicians. J Med Internet Res. Jun 19, 2020;22(6):e15154. [FREE Full text] [CrossRef] [Medline]
- Felmingham CM, Adler NR, Ge Z, Morton RL, Janda M, Mar VJ. The importance of incorporating human factors in the design and implementation of artificial intelligence for skin cancer diagnosis in the real world. Am J Clin Dermatol. Mar 22, 2021;22(2):233-242. [CrossRef] [Medline]
- Carayon P, Hancock P, Leveson N, Noy I, Sznelwar L, van Hootegem G. Advancing a sociotechnical systems approach to workplace safety--developing the conceptual framework. Ergonomics. Apr 02, 2015;58(4):548-564. [FREE Full text] [CrossRef] [Medline]
- Mumford E. The story of socio‐technical design: reflections on its successes, failures and potential. Inf Syst J. Sep 04, 2006;16(4):317-342. [CrossRef]
- Carayon P, Schoofs Hundt A, Karsh BT, Gurses AP, Alvarado CJ, Smith M, et al. Work system design for patient safety: the SEIPS model. Qual Saf Health Care. Dec 2006;15 Suppl 1(Suppl 1):i50-i58. [FREE Full text] [CrossRef] [Medline]
- Hettinger LJ, Kirlik A, Goh YM, Buckle P. Modelling and simulation of complex sociotechnical systems: envisioning and analysing work environments. Ergonomics. Mar 11, 2015;58(4):600-614. [FREE Full text] [CrossRef] [Medline]
- Carayon P, Wooldridge A, Hoonakker P, Hundt AS, Kelly MM. SEIPS 3.0: human-centered design of the patient journey for patient safety. Appl Ergon. Apr 2020;84:103033. [FREE Full text] [CrossRef] [Medline]
- Wooldridge AR, Carayon P, Hoonakker P, Hose BZ, Eithun B, Brazelton T3, et al. Work system barriers and facilitators in inpatient care transitions of pediatric trauma patients. Appl Ergon. May 2020;85:103059. [FREE Full text] [CrossRef] [Medline]
- Hoonakker PL, Carayon P, Cartmill RS. The impact of secure messaging on workflow in primary care: results of a multiple-case, multiple-method study. Int J Med Inform. Apr 2017;100:63-76. [FREE Full text] [CrossRef] [Medline]
- Damschroder LJ, Aron DC, Keith RE, Kirsh SR, Alexander JA, Lowery JC. Fostering implementation of health services research findings into practice: a consolidated framework for advancing implementation science. Implement Sci. 2009;4:50. [FREE Full text] [CrossRef] [Medline]
- Damschroder LJ, Reardon CM, Widerquist MA, Lowery J. The updated Consolidated Framework for Implementation Research based on user feedback. Implement Sci. Oct 29, 2022;17(1):75. [FREE Full text] [CrossRef] [Medline]
- Greenhalgh T, Wherton J, Papoutsi C, Lynch J, Hughes G, A'Court C, et al. Beyond adoption: a new framework for theorizing and evaluating nonadoption, abandonment, and challenges to the scale-up, spread, and sustainability of health and care technologies. J Med Internet Res. Nov 01, 2017;19(11):e367. [FREE Full text] [CrossRef] [Medline]
- Abell B, Naicker S, Rodwell D, Donovan T, Tariq A, Baysari M, et al. Identifying barriers and facilitators to successful implementation of computerized clinical decision support systems in hospitals: a NASSS framework-informed scoping review. Implement Sci. Jul 26, 2023;18(1):32. [FREE Full text] [CrossRef] [Medline]
- Mummah SA, Robinson TN, King AC, Gardner CD, Sutton S. IDEAS (integrate, design, assess, and share): a framework and toolkit of strategies for the development of more effective digital interventions to change health behavior. J Med Internet Res. Dec 16, 2016;18(12):e317. [FREE Full text] [CrossRef] [Medline]
- Andersen TO, Nunes F, Wilcox L, Coiera E, Rogers Y. Introduction to the special issue on human-centred AI in healthcare: challenges appearing in the wild. ACM Trans Comput Hum Interact. Jun 30, 2023;30(2):1-12. [CrossRef]
- Wenderott K, Gambashidze N, Weigl M. Integration of artificial intelligence into sociotechnical work systems-effects of artificial intelligence solutions in medical imaging on clinical efficiency: protocol for a systematic literature review. JMIR Res Protoc. Dec 01, 2022;11(12):e40485. [FREE Full text] [CrossRef] [Medline]
- Chomutare T, Tejedor M, Svenning TO, Marco-Ruiz L, Tayefi M, Lind K, et al. Artificial intelligence implementation in healthcare: a theory-based scoping review of barriers and facilitators. Int J Environ Res Public Health. Dec 06, 2022;19(23):16359. [FREE Full text] [CrossRef] [Medline]
- Han R, Acosta JN, Shakeri Z, Ioannidis JP, Topol EJ, Rajpurkar P. Randomised controlled trials evaluating artificial intelligence in clinical practice: a scoping review. Lancet Digit Health. May 2024;6(5):e367-e373. [CrossRef]
- Wenderott K, Krups J, Zaruchas F, Weigl M. Effects of artificial intelligence implementation on efficiency in medical imaging-a systematic literature review and meta-analysis. NPJ Digit Med. Sep 30, 2024;7(1):265. [FREE Full text] [CrossRef] [Medline]
- Beller EM, Chen JK, Wang UL, Glasziou PP. Are systematic reviews up-to-date at the time of publication? Syst Rev. May 28, 2013;2:36. [FREE Full text] [CrossRef] [Medline]
- Wenderott K, Krups J, Luetkens JA, Weigl M. Radiologists' perspectives on the workflow integration of an artificial intelligence-based computer-aided detection system: a qualitative study. Appl Ergon. May 2024;117:104243. [FREE Full text] [CrossRef] [Medline]
- Hassan M, Kushniruk A, Borycki E. Barriers to and facilitators of artificial intelligence adoption in health care: scoping review. JMIR Hum Factors. Aug 29, 2024;11:e48633. [FREE Full text] [CrossRef] [Medline]
- Lokaj B, Pugliese MT, Kinkel K, Lovis C, Schmid J. Barriers and facilitators of artificial intelligence conception and implementation for breast imaging diagnosis in clinical practice: a scoping review. Eur Radiol. Mar 02, 2024;34(3):2096-2109. [FREE Full text] [CrossRef] [Medline]
- Masud R, Al-Rei M, Lokker C. Computer-aided detection for breast cancer screening in clinical settings: scoping review. JMIR Med Inform. Jul 18, 2019;7(3):e12660. [FREE Full text] [CrossRef] [Medline]
- Eltawil FA, Atalla M, Boulos E, Amirabadi A, Tyrrell PN. Analyzing barriers and enablers for the acceptance of artificial intelligence innovations into radiology practice: a scoping review. Tomography. Jul 28, 2023;9(4):1443-1455. [CrossRef] [Medline]
- Swillens JE, Nagtegaal ID, Engels S, Lugli A, Hermens RP, van der Laak JA. Pathologists' first opinions on barriers and facilitators of computational pathology adoption in oncological pathology: an international study. Oncogene. Sep 16, 2023;42(38):2816-2827. [FREE Full text] [CrossRef] [Medline]
- Strohm L, Hehakaya C, Ranschaert ER, Boon WP, Moors EH. Implementation of artificial intelligence (AI) applications in radiology: hindering and facilitating factors. Eur Radiol. Oct 26, 2020;30(10):5525-5532. [FREE Full text] [CrossRef] [Medline]
- Liao X, Yao C, Jin F, Zhang J, Liu L. Barriers and facilitators to implementing imaging-based diagnostic artificial intelligence-assisted decision-making software in hospitals in China: a qualitative study using the updated Consolidated Framework for Implementation Research. BMJ Open. Sep 10, 2024;14(9):e084398. [FREE Full text] [CrossRef] [Medline]
- Ouzzani M, Hammady H, Fedorowicz Z, Elmagarmid A. Rayyan-a web and mobile app for systematic reviews. Syst Rev. Dec 05, 2016;5(1):210. [FREE Full text] [CrossRef] [Medline]
- Page MJ, McKenzie JE, Bossuyt PM, Boutron I, Hoffmann TC, Mulrow CD, et al. The PRISMA 2020 statement: an updated guideline for reporting systematic reviews. BMJ. Mar 29, 2021;372:n71. [FREE Full text] [CrossRef] [Medline]
- VERBI software. MAXQDA. URL: https://www.maxqda.com/about [accessed 2023-04-21]
- RStudio. 2015. URL: http://www.rstudio.com/ [accessed 2025-07-03]
- Bertram RM, Blase KA, Fixsen DL. Improving programs and outcomes: implementation frameworks and organization change. Res Soc Work Pract. Jun 08, 2014;25(4):477-487. [CrossRef]
- Pane ES, Sarno R. Capability maturity model integration (CMMI) for optimizing object-oriented analysis and design (OOAD). Procedia Comput Sci. 2015;72:40-48. [CrossRef]
- Kuckartz U, Radiker S. Qualitative Content Analysis: Methods, Practice and Software [Qualitative Inhaltsanalyse: Methoden, Praxis, Computerunterstützung]. Weinheim, Germany. Beltz Juventa; 2022.
- Niezen MG, Mathijssen JJ. Reframing professional boundaries in healthcare: a systematic review of facilitators and barriers to task reallocation from the domain of medicine to the nursing domain. Health Policy. Aug 2014;117(2):151-169. [FREE Full text] [CrossRef] [Medline]
- Hopf C, Schmidt C. On the Relationship Between Intra-Familial Social Experiences, Personality Development and Political Orientations: Documentation and Discussion of the Methodological Procedure in a Study on This Topic. [Zum Verhältnis von innerfamilialen sozialen Erfahrungen, Persönlichkeitsentwicklung und politischen Orientierungen: Dokumentation und Erörterung des methodischen Vorgehens in einer Studie zu diesem Thema]. Hildesheim, Germany. Institut für Sozialwissenschaften der Universität Hildesheim; 1993.
- Wooldridge AR, Carayon P, Hundt AS, Hoonakker PL. SEIPS-based process modeling in primary care. Appl Ergon. Apr 2017;60:240-254. [FREE Full text] [CrossRef] [Medline]
- Wooldridge AR, Carayon P, Shaffer DW, Eagan B. Quantifying the qualitative with epistemic network analysis: a human factors case study of task-allocation communication in a primary care team. IISE Trans Healthc Syst Eng. 2018;8(1):72-82. [FREE Full text] [CrossRef] [Medline]
- Weiler DT, Lingg AJ, Eagan BR, Shaffer DW, Werner NE. Quantifying the qualitative: exploring epistemic network analysis as a method to study work system interactions. Ergonomics. Oct 2022;65(10):1434-1449. [FREE Full text] [CrossRef] [Medline]
- Shaffer DW, Hatfield D, Svarovsky GN, Nash P, Nulty A, Bagley E, et al. Epistemic network analysis: a prototype for 21st-century assessment of learning. Int J Learn Media. May 2009;1(2):33-53. [FREE Full text] [CrossRef]
- Arastoopour G, Chesler NC, Shaffer DW. Epistemic network analysis as a tool for engineering design assessment. In: Proceedings of the 122nd ASEE Annual Conference & Exposition. 2015. Presented at: ASEE 2015; June 14-17, 2015; Seattle, WA. URL: https://www.researchgate.net/publication/283441969_Epistemic_network_analysis_as_a_tool_for_engineering_design_assessment [CrossRef]
- D’Angelo CM, Clark DB, Shaffer DW. Epistemic network analysis: an alternative analysis technique for complex STEM thinking. In: Proceedings of the National Association of Research on Science Teaching Conference. 2012. Presented at: NARST 2012; March 25-28, 2012; Indianapolis, Indiana. [CrossRef]
- Shaffer DW, Collier W, Ruis AR. A tutorial on epistemic network analysis: analyzing the structure of connections in cognitive, social, and interaction data. J Learn Anal. Dec 19, 2016;3(3):9-45. [CrossRef]
- Andrist S, Collier W, Gleicher M, Mutlu B, Shaffer D. Look together: analyzing gaze coordination with epistemic network analysis. Front Psychol. Jul 21, 2015;6:1016. [FREE Full text] [CrossRef] [Medline]
- Bowman D, Swiecki Z, Cai Z, Wang Y, Eagan B, Linderoth J, et al. The mathematical foundations of epistemic network analysis. In: Proceedings of the Second International Conference on Advances in Quantitative Ethnography. 2021. Presented at: ICQE 2020; February 1-3, 2021; Malibu, CA. [CrossRef]
- Shaffer DW. Quantitative Ethnography. Charlottesville, VA. Cathcart Press; 2017.
- Shaffer DW, Ruis AR. Epistemic network analysis: a worked example of theory-based learning analytics. In: Lang C, Siemens G, Wise A, Grasevic D, editors. Handbook of Learning Analytics. Beaumont, AB. Society for Learning Analytics Research; 2017:175-187.
- Marquart CL, Hinojosa C, Swiecki Z, Shaffer DW. Epistemic network analysis. Epistemic Network. 2018. URL: https://app.epistemicnetwork.org/login.html [accessed 2025-07-03]
- Romagnoli A, Ferrara F, Langella R, Zovi A. Healthcare systems and artificial intelligence: focus on challenges and the international regulatory framework. Pharm Res. Apr 05, 2024;41(4):721-730. [CrossRef] [Medline]
- Muehlematter UJ, Daniore P, Vokinger KN. Approval of artificial intelligence and machine learning-based medical devices in the USA and Europe (2015–20): a comparative analysis. Lancet Digit Health. Mar 2021;3(3):e195-e203. [FREE Full text] [CrossRef] [Medline]
- Vokinger KN, Gasser U. Regulating AI in medicine in the United States and Europe. Nat Mach Intell. Sep 2021;3(9):738-739. [FREE Full text] [CrossRef] [Medline]
- Sterne JA, Hernán MA, Reeves BC, Savović J, Berkman ND, Viswanathan M, et al. ROBINS-I: a tool for assessing risk of bias in non-randomised studies of interventions. BMJ. Oct 12, 2016;355:i4919. [FREE Full text] [CrossRef] [Medline]
- Sterne JA, Savović J, Page MJ, Elbers RG, Blencowe NS, Boutron I, et al. RoB 2: a revised tool for assessing risk of bias in randomised trials. BMJ. Aug 28, 2019;366:l4898. [FREE Full text] [CrossRef] [Medline]
- Arbabshirani MR, Fornwalt BK, Mongelluzzo GJ, Suever JD, Geise BD, Patel AA, et al. Advanced machine learning in action: identification of intracranial hemorrhage on computed tomography scans of the head with clinical workflow integration. NPJ Digit Med. Apr 4, 2018;1(1):9. [FREE Full text] [CrossRef] [Medline]
- Batra K, Xi Y, Bhagwat S, Espino A, Peshock RM. Radiologist worklist reprioritization using artificial intelligence: impact on report turnaround times for CTPA examinations positive for acute pulmonary embolism. Am J Roentgenol. Sep 2023;221(3):324-333. [CrossRef]
- Carlile M, Hurt B, Hsiao A, Hogarth M, Longhurst CA, Dameff C. Deployment of artificial intelligence for radiographic diagnosis of COVID-19 pneumonia in the emergency department. J Am Coll Emerg Physicians Open. Dec 05, 2020;1(6):1459-1464. [FREE Full text] [CrossRef] [Medline]
- Cha E, Elguindi S, Onochie I, Gorovets D, Deasy JO, Zelefsky M, et al. Clinical implementation of deep learning contour autosegmentation for prostate radiotherapy. Radiother Oncol. Jun 2021;159:1-7. [FREE Full text] [CrossRef] [Medline]
- Cheikh AB, Gorincour G, Nivet H, May J, Seux M, Calame P, et al. How artificial intelligence improves radiological interpretation in suspected pulmonary embolism. Eur Radiol. Sep 22, 2022;32(9):5831-5842. [FREE Full text] [CrossRef] [Medline]
- Chen W, Wu J, Wei R, Wu S, Xia C, Wang D, et al. Improving the diagnosis of acute ischemic stroke on non-contrast CT using deep learning: a multicenter study. Insights Imaging. Dec 06, 2022;13(1):184. [FREE Full text] [CrossRef] [Medline]
- Conant EF, Toledano AY, Periaswamy S, Fotin SV, Go J, Boatsman JE, et al. Improving accuracy and efficiency with concurrent use of artificial intelligence for digital breast tomosynthesis. Radiol Artif Intell. Jul 31, 2019;1(4):e180096. [FREE Full text] [CrossRef] [Medline]
- Davis MA, Rao B, Cedeno PA, Saha A, Zohrabian VM. Machine learning and improved quality metrics in acute intracranial hemorrhage by noncontrast computed tomography. Curr Probl Diagn Radiol. Jul 2022;51(4):556-561. [CrossRef] [Medline]
- Diao K, Chen Y, Liu Y, Chen BJ, Li WJ, Zhang L, et al. Diagnostic study on clinical feasibility of an AI-based diagnostic system as a second reader on mobile CT images: a preliminary result. Ann Transl Med. Jun 2022;10(12):668. [FREE Full text] [CrossRef] [Medline]
- Duron L, Ducarouge A, Gillibert A, Lainé J, Allouche C, Cherel N, et al. Assessment of an AI aid in detection of adult appendicular skeletal fractures by emergency physicians and radiologists: a multicenter cross-sectional diagnostic study. Radiology. Jul 2021;300(1):120-129. [CrossRef] [Medline]
- Elijovich L, Dornbos Iii D, Nickele C, Alexandrov A, Inoa-Acosta V, Arthur AS, et al. Automated emergent large vessel occlusion detection by artificial intelligence improves stroke workflow in a hub and spoke stroke system of care. J Neurointerv Surg. Jul 20, 2022;14(7):704-708. [CrossRef] [Medline]
- Ginat D. Implementation of machine learning software on the radiology worklist decreases scan view delay for the detection of intracranial hemorrhage on CT. Brain Sci. Jun 23, 2021;11(7):832. [FREE Full text] [CrossRef] [Medline]
- Hassan AE, Ringheanu VM, Tekle WG. The implementation of artificial intelligence significantly reduces door-in-door-out times in a primary care center prior to transfer. Interv Neuroradiol. Dec 2023;29(6):631-636. [CrossRef] [Medline]
- Jones CM, Danaher L, Milne MR, Tang C, Seah J, Oakden-Rayner L, et al. Assessment of the effect of a comprehensive chest radiograph deep learning model on radiologist reports and patient outcomes: a real-world observational study. BMJ Open. Dec 20, 2021;11(12):e052902. [FREE Full text] [CrossRef] [Medline]
- Ladabaum U, Shepard J, Weng Y, Desai M, Singer S, Mannalithara A. Computer-aided detection of polyps does not improve colonoscopist performance in a pragmatic implementation trial. Gastroenterology. May 2023;164(6):S-152-S-153. [CrossRef]
- Levy I, Bruckmayer L, Klang E, Ben-Horin S, Kopylov U. Artificial intelligence-aided colonoscopy does not increase adenoma detection rate in routine clinical practice. Am J Gastroenterol. Nov 01, 2022;117(11):1871-1873. [CrossRef] [Medline]
- Marwaha A, Chitayat D, Meyn MS, Mendoza-Londono R, Chad L. The point-of-care use of a facial phenotyping tool in the genetics clinic: enhancing diagnosis and education with machine learning. Am J Med Genet A. Apr 08, 2021;185(4):1151-1158. [CrossRef] [Medline]
- Müller FC, Raaschou H, Akhtar N, Brejnebøl M, Collatz L, Andersen MB. Impact of concurrent use of artificial intelligence tools on radiologists reading time: a prospective feasibility study. Acad Radiol. Jul 2022;29(7):1085-1090. [FREE Full text] [CrossRef] [Medline]
- Nehme F, Coronel E, Barringer DA, Romero LG, Shafi MA, Ross WA, et al. Performance and attitudes toward real-time computer-aided polyp detection during colonoscopy in a large tertiary referral center in the United States. Gastrointest Endosc. Jul 2023;98(1):100-9.e6. [CrossRef] [Medline]
- Oppenheimer J, Lüken S, Hamm B, Niehues SM. A prospective approach to integration of AI fracture detection software in radiographs into clinical workflow. Life (Basel). Jan 13, 2023;13(1):223. [FREE Full text] [CrossRef] [Medline]
- Pierce JD, Rosipko B, Youngblood L, Gilkeson RC, Gupta A, Bittencourt LK. Seamless integration of artificial intelligence into the clinical environment: our experience with a novel pneumothorax detection artificial intelligence algorithm. J Am Coll Radiol. Nov 2021;18(11):1497-1505. [CrossRef] [Medline]
- Potretzke TA, Korfiatis P, Blezek DJ, Edwards ME, Klug JR, Cook CJ, et al. Clinical implementation of an artificial intelligence algorithm for magnetic resonance-derived measurement of total kidney volume. Mayo Clin Proc. May 2023;98(5):689-700. [FREE Full text] [CrossRef] [Medline]
- Quan SY, Wei MT, Lee J, Mohi-Ud-Din R, Mostaghim R, Sachdev R, et al. Clinical evaluation of a real-time artificial intelligence-based polyp detection system: a US multi-center pilot study. Sci Rep. Apr 21, 2022;12(1):6598. [FREE Full text] [CrossRef] [Medline]
- Raya-Povedano JL, Romero-Martín S, Elías-Cabot E, Gubern-Mérida A, Rodríguez-Ruiz A, Álvarez-Benito M. AI-based strategies to reduce workload in breast cancer screening with mammography and tomosynthesis: a retrospective evaluation. Radiology. Jul 2021;300(1):57-65. [FREE Full text] [CrossRef] [Medline]
- Ruamviboonsuk P, Tiwari R, Sayres R, Nganthavee V, Hemarat K, Kongprayoon A, et al. Real-time diabetic retinopathy screening by deep learning in a multisite national screening programme: a prospective interventional cohort study. Lancet Digit Health. Apr 2022;4(4):e235-e244. [FREE Full text] [CrossRef] [Medline]
- Sandbank J, Bataillon G, Nudelman A, Krasnitsky I, Mikulinsky R, Bien L, et al. Validation and real-world clinical application of an artificial intelligence algorithm for breast cancer detection in biopsies. NPJ Breast Cancer. Dec 06, 2022;8(1):129. [FREE Full text] [CrossRef] [Medline]
- Schmuelling L, Franzeck FC, Nickel CH, Mansella G, Bingisser R, Schmidt N, et al. Deep learning-based automated detection of pulmonary embolism on CT pulmonary angiograms: no significant effects on report communication times and patient turnaround in the emergency department nine months after technical implementation. Eur J Radiol. Aug 2021;141:109816. [FREE Full text] [CrossRef] [Medline]
- Seyam M, Weikert T, Sauter A, Brehm A, Psychogios MN, Blackham KA. Utilization of artificial intelligence-based intracranial hemorrhage detection on emergent noncontrast CT images in clinical workflow. Radiol Artif Intell. Mar 01, 2022;4(2):e210168. [FREE Full text] [CrossRef] [Medline]
- Tchou PM, Haygood TM, Atkinson EN, Stephens TW, Davis PL, Arribas EM, et al. Interpretation time of computer-aided detection at screening mammography. Radiology. Oct 2010;257(1):40-46. [CrossRef] [Medline]
- Tricarico D, Calandri M, Barba M, Piatti C, Geninatti C, Basile D, et al. Convolutional neural network-based automatic analysis of chest radiographs for the detection of COVID-19 pneumonia: a prioritizing tool in the emergency department, phase I study and preliminary "real life" results. Diagnostics (Basel). Feb 23, 2022;12(3):570. [FREE Full text] [CrossRef] [Medline]
- Vassallo L, Traverso A, Agnello M, Bracco C, Campanella D, Chiara G, et al. A cloud-based computer-aided detection system improves identification of lung nodules on computed tomography scans of patients with extra-thoracic malignancies. Eur Radiol. Jan 15, 2019;29(1):144-152. [CrossRef] [Medline]
- Wang P, Berzin TM, Glissen Brown JR, Bharadwaj S, Becq A, Xiao X, et al. Real-time automatic detection system increases colonoscopic polyp and adenoma detection rates: a prospective randomised controlled study. Gut. Oct 27, 2019;68(10):1813-1819. [FREE Full text] [CrossRef] [Medline]
- Wang M, Xia C, Huang L, Xu S, Qin C, Liu J, et al. Deep learning-based triage and analysis of lesion burden for COVID-19: a retrospective study with external validation. Lancet Digit Health. Oct 2020;2(10):e506-e515. [FREE Full text] [CrossRef] [Medline]
- Wittenberg R, Berger FH, Peters JF, Weber M, van Hoorn F, Beenen LF, et al. Acute pulmonary embolism: effect of a computer-assisted detection prototype on diagnosis--an observer study. Radiology. Jan 2012;262(1):305-313. [CrossRef] [Medline]
- Wong J, Huang V, Wells D, Giambattista J, Giambattista J, Kolbeck C, et al. Implementation of deep learning-based auto-segmentation for radiotherapy planning structures: a workflow study at two cancer centers. Radiat Oncol. Jun 08, 2021;16(1):101. [FREE Full text] [CrossRef] [Medline]
- Wong KP, Homer SY, Wei SH, Yaghmai N, Estrada Paz OA, Young TJ, et al. Integration and evaluation of chest X-ray artificial intelligence in clinical practice. J Med Imaging (Bellingham). Sep 2023;10(5):051805. [FREE Full text] [CrossRef] [Medline]
- Yang Y, Pan J, Yuan M, Lai K, Xie H, Ma L, et al. Performance of the AIDRScreening system in detecting diabetic retinopathy in the fundus photographs of Chinese patients: a prospective, multicenter, clinical study. Ann Transl Med. Oct 2022;10(20):1088. [FREE Full text] [CrossRef] [Medline]
- Zia A, Fletcher C, Bigwood S, Ratnakanthan P, Seah J, Lee R, et al. Retrospective analysis and prospective validation of an AI-based software for intracranial haemorrhage detection at a high-volume trauma centre. Sci Rep. Nov 18, 2022;12(1):19885. [FREE Full text] [CrossRef] [Medline]
- Holden RJ, Karsh BT. The technology acceptance model: its past and its future in health care. J Biomed Inform. Feb 2010;43(1):159-172. [FREE Full text] [CrossRef] [Medline]
- Venkatesh V, Morris MG, Davis GB, Davis FD. User acceptance of information technology: toward a unified view. MIS Q. 2003;27(3):425-478. [CrossRef]
- Chanda T, Hauser K, Hobelsberger S, Bucher TC, Garcia CN, Wies C, Reader Study Consortium, et al. Dermatologist-like explainable AI enhances trust and confidence in diagnosing melanoma. Nat Commun. Jan 15, 2024;15(1):524. [FREE Full text] [CrossRef] [Medline]
- Salwei ME, Carayon P, Hoonakker PL, Hundt AS, Wiegmann D, Pulia M, et al. Workflow integration analysis of a human factors-based clinical decision support in the emergency department. Appl Ergon. Nov 2021;97:103498. [FREE Full text] [CrossRef] [Medline]
- Salwei ME, Carayon P. A sociotechnical systems framework for the application of artificial intelligence in health care delivery. J Cogn Eng Decis Mak. Dec 11, 2022;16(4):194-206. [FREE Full text] [CrossRef] [Medline]
- Godoe P, Johansen TS. Understanding adoption of new technologies: technology readiness and technology acceptance as an integrated concept. J Eur Psychol Students. May 06, 2012;3:38. [CrossRef]
- Wooldridge AR, Morgan J, Ramadhani WA, Hanson K, Vazquez-Melendez E, Kendhari H, et al. Interactions in sociotechnical systems: achieving balance in the use of an augmented reality mobile application. Hum Factors. Mar 2024;66(3):658-682. [CrossRef] [Medline]
- Marco-Ruiz L, Hernández MÁ, Ngo PD, Makhlysheva A, Svenning TO, Dyb K, et al. A multinational study on artificial intelligence adoption: clinical implementers' perspectives. Int J Med Inform. Apr 2024;184:105377. [FREE Full text] [CrossRef] [Medline]
- Gunasekeran DV, Zheng F, Lim GY, Chong CC, Zhang S, Ng WY, et al. Acceptance and perception of artificial intelligence usability in eye care (APPRAISE) for ophthalmologists: a multinational perspective. Front Med (Lausanne). Oct 13, 2022;9:875242. [FREE Full text] [CrossRef] [Medline]
- Lennon MR, Bouamrane MM, Devlin AM, O'Connor S, O'Donnell C, Chetty U, et al. Readiness for delivering digital health at scale: lessons from a longitudinal qualitative evaluation of a national digital health innovation program in the United Kingdom. J Med Internet Res. Feb 16, 2017;19(2):e42. [FREE Full text] [CrossRef] [Medline]
- Widner K, Virmani S, Krause J, Nayar J, Tiwari R, Pedersen ER, et al. Lessons learned from translating AI from development to deployment in healthcare. Nat Med. Jun 29, 2023;29(6):1304-1306. [CrossRef] [Medline]
- Chen H, Gomez C, Huang CM, Unberath M. Explainable medical imaging AI needs human-centered design: guidelines and evidence from a systematic review. NPJ Digit Med. Oct 19, 2022;5(1):156. [FREE Full text] [CrossRef] [Medline]
- Herrmann T, Pfeiffer S. Keeping the organization in the loop: a socio-technical extension of human-centered artificial intelligence. AI Soc. Feb 18, 2022;38(4):1523-1542. [CrossRef]
- Chen Y, Clayton EW, Novak LL, Anders S, Malin B. Human-centered design to address biases in artificial intelligence. J Med Internet Res. Mar 24, 2023;25:e43251. [FREE Full text] [CrossRef] [Medline]
- Koch A, Burns J, Catchpole K, Weigl M. Associations of workflow disruptions in the operating room with surgical outcomes: a systematic review and narrative synthesis. BMJ Qual Saf. Dec 23, 2020;29(12):1033-1045. [CrossRef] [Medline]
- Gore JC. Artificial intelligence in medical imaging. Magn Reson Imaging. May 2020;68:A1-A4. [CrossRef] [Medline]
- Mennella C, Maniscalco U, De Pietro G, Esposito M. Ethical and regulatory challenges of AI technologies in healthcare: a narrative review. Heliyon. Feb 29, 2024;10(4):e26297. [FREE Full text] [CrossRef] [Medline]
- Ergonomics of human-system interaction: part 210: human-centred design for interactive systems. International Organization for Standardization. 2019. URL: https://www.iso.org/standard/77520.html [accessed 2025-01-14]
- Kruse CS, Kothman K, Anerobi K, Abanaka L. Adoption factors of the electronic health record: a systematic review. JMIR Med Inform. Jun 01, 2016;4(2):e19. [FREE Full text] [CrossRef] [Medline]
- Gagnon MP, Desmartis M, Labrecque M, Car J, Pagliari C, Pluye P, et al. Systematic review of factors influencing the adoption of information and communication technologies by healthcare professionals. J Med Syst. Feb 2012;36(1):241-277. [FREE Full text] [CrossRef] [Medline]
- Gagnon MP, Nsangou ER, Payne-Gagnon J, Grenier S, Sicotte C. Barriers and facilitators to implementing electronic prescription: a systematic review of user groups' perceptions. J Am Med Inform Assoc. 2014;21(3):535-541. [FREE Full text] [CrossRef] [Medline]
- Sidek YH, Martins JT. Perceived critical success factors of electronic health record system implementation in a dental clinic context: an organisational management perspective. Int J Med Inform. Nov 2017;107:88-100. [FREE Full text] [CrossRef] [Medline]
- Fragidis LL, Chatzoglou PD. Implementation of a nationwide electronic health record (EHR): the international experience in 13 countries. Int J Health Care Qual Assur. Mar 12, 2018;31(2):116-130. [FREE Full text] [CrossRef] [Medline]
- Yen PY, McAlearney AS, Sieck CJ, Hefner JL, Huerta TR. Health information technology (HIT) adaptation: refocusing on the journey to successful HIT implementation. JMIR Med Inform. Sep 07, 2017;5(3):e28. [FREE Full text] [CrossRef] [Medline]
- Gama F, Tyskbo D, Nygren J, Barlow J, Reed J, Svedberg P. Implementation frameworks for artificial intelligence translation into health care practice: scoping review. J Med Internet Res. Jan 27, 2022;24(1):e32215. [FREE Full text] [CrossRef] [Medline]
- Garvey KV, Thomas Craig KJ, Russell R, Novak LL, Moore D, Miller BM. Considering clinician competencies for the implementation of artificial intelligence-based tools in health care: findings from a scoping review. JMIR Med Inform. Nov 16, 2022;10(11):e37478. [FREE Full text] [CrossRef] [Medline]
- Norori N, Hu Q, Aellen FM, Faraci FD, Tzovara A. Addressing bias in big data and AI for health care: a call for open science. Patterns (N Y). Oct 08, 2021;2(10):100347. [FREE Full text] [CrossRef] [Medline]
- Mittermaier M, Raza MM, Kvedar JC. Bias in AI-based models for medical applications: challenges and mitigation strategies. NPJ Digit Med. Jun 14, 2023;6(1):113. [FREE Full text] [CrossRef] [Medline]
- Ratwani RM, Sutton K, Galarraga JE. Addressing AI algorithmic bias in health care. JAMA. Oct 01, 2024;332(13):1051-1052. [CrossRef] [Medline]
- Naik N, Hameed BM, Shetty DK, Swain D, Shah M, Paul R, et al. Legal and ethical consideration in artificial intelligence in healthcare: who takes responsibility? Front Surg. Mar 14, 2022;9:862322. [FREE Full text] [CrossRef] [Medline]
- Terranova C, Cestonaro C, Fava L, Cinquetti A. AI and professional liability assessment in healthcare. A revolution in legal medicine? Front Med (Lausanne). Jan 8, 2023;10:1337335. [FREE Full text] [CrossRef] [Medline]
- Eldakak A, Alremeithi A, Dahiyat E, El-Gheriani M, Mohamed H, Abdulrahim Abdulla MI. Civil liability for the actions of autonomous AI in healthcare: an invitation to further contemplation. Humanit Soc Sci Commun. Feb 23, 2024;11:305. [CrossRef]
Abbreviations
| AI: artificial intelligence |
| CFIR: Consolidated Framework for Implementation Research |
| ENA: epistemic network analysis |
| IDEAS: integrate, design, assess, and share |
| NASSS: nonadoption, abandonment, scale-up, spread, and sustainability |
| PICO: population, intervention, comparison, and outcome |
| PRISMA: Preferred Reporting Items for Systematic Reviews and Meta-Analyses |
| SEIPS: Systems Engineering Initiative for Patient Safety |
Edited by Y Li; submitted 25.06.24; peer-reviewed by S Antani, B Mesko, T Donovan; comments to author 26.11.24; revised version received 15.01.25; accepted 15.05.25; published 21.07.25.
Copyright©Katharina Wenderott, Jim Krups, Matthias Weigl, Abigail R Wooldridge. Originally published in the Journal of Medical Internet Research (https://www.jmir.org), 21.07.2025.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research (ISSN 1438-8871), is properly cited. The complete bibliographic information, a link to the original publication on https://www.jmir.org/, as well as this copyright and license information must be included.

