Published on in Vol 27 (2025)

Preprints (earlier versions) of this paper are available at https://preprints.jmir.org/preprint/71091, first published .
Imaging-Based AI for Predicting Lymphovascular Space Invasion in Cervical Cancer: Systematic Review and Meta-Analysis

Imaging-Based AI for Predicting Lymphovascular Space Invasion in Cervical Cancer: Systematic Review and Meta-Analysis

Imaging-Based AI for Predicting Lymphovascular Space Invasion in Cervical Cancer: Systematic Review and Meta-Analysis

Department of Radiation Oncology, The Second Hospital of Jilin University, 4026 Yatai Street, Changchun, Jilin Province, China

Corresponding Author:

Ling Qiu


Background: The role of artificial intelligence (AI) in enhancing the accuracy of lymphovascular space invasion (LVSI) detection in cervical cancer remains debated.

Objective: This meta-analysis aimed to evaluate the diagnostic accuracy of imaging-based AI for predicting LVSI in cervical cancer.

Methods: We conducted a comprehensive literature search across multiple databases, including PubMed, Embase, and Web of Science, identifying studies published up to November 9, 2024. Studies were included if they evaluated the diagnostic performance of imaging-based AI models in detecting LVSI in cervical cancer. We used a bivariate random-effects model to calculate pooled sensitivity and specificity with corresponding 95% confidence intervals. Study heterogeneity was assessed using the I2 statistic.

Results: Of 403 studies identified, 16 studies (2514 patients) were included. For the interval validation set, the pooled sensitivity, specificity, and area under the curve (AUC) for detecting LVSI were 0.84 (95% CI 0.79-0.87), 0.78 (95% CI 0.75-0.81), and 0.87 (95% CI 0.84-0.90). For the external validation set, the pooled sensitivity, specificity, and AUC for detecting LVSI were 0.79 (95% CI 0.70-0.86), 0.76 (95% CI 0.67-0.83), and 0.84 (95% CI 0.81-0.87). Using the likelihood ratio test for subgroup analysis, deep learning demonstrated significantly higher sensitivity compared to machine learning (P=.01). Moreover, AI models based on positron emission tomography/computed tomography exhibited superior sensitivity relative to those based on magnetic resonance imaging (P=.01).

Conclusions: Imaging-based AI, particularly deep learning algorithms, demonstrates promising diagnostic performance in predicting LVSI in cervical cancer. However, the limited external validation datasets and the retrospective nature of the research may introduce potential biases. These findings underscore AI’s potential as an auxiliary diagnostic tool, necessitating further large-scale prospective validation.

Trial Registration: PROSPERO CRD42024612008; https://www.crd.york.ac.uk/PROSPERO/view/CRD42024612008

J Med Internet Res 2025;27:e71091

doi:10.2196/71091

Keywords



Cervical cancer remains a predominant cause of cancer-related morbidity and mortality among women globally, with an estimated 660,000 new cases and 350,000 deaths reported in 2022 [1,2]. The prognosis of cervical cancer is determined by a variety of influencing factors, including tumor stage, histopathological type, size, lymphovascular space invasion (LVSI), and the presence of distant metastasis [3-5]. Among these, LVSI is a critical prognostic factor, as its presence is associated with an increased risk of lymph node metastasis, disease recurrence, and poorer overall survival [6]. It refers to the infiltration of cancer cells into the lymphatic system or blood vessels [6]. Researchers have demonstrated that LVSI is linked to worse survival outcomes. According to Marchiolé et al [7], the incidence of LVSI was 2-fold higher in patients with cervical cancer recurrence than in those without recurrence. Pol et al [8] demonstrated that patients without LVSI exhibited significantly superior outcomes in both disease-free survival and overall survival.

The presence of LVSI significantly impacts treatment decisions in cervical cancer [9]. Patients with LVSI-positive status often necessitate multimodal therapy, including radical surgery with lymphadenectomy combined with chemotherapy or radiotherapy, with the sequence determined by tumor stage and risk factors [10]. Current diagnostic methods for cervical cancer include imaging techniques such as computed tomography (CT), positron emission tomography/computed tomography (PET/CT), and magnetic resonance imaging (MRI), which are used to evaluate LVSI. However, these imaging modalities have limitations in accurately detecting LVSI. Based on findings from Holtz and Dunton [11,12] and Zhu et al [12,12], CT demonstrated inferior soft tissue visualization and diagnostic performance in assessing cervical cancer tumor invasion compared to MRI. Specifically, PET/CT exhibits inherent constraints in soft tissue resolution, limiting its diagnostic performance for detecting minute neoplastic lesions [13,14]. According to Park et al [15], among stage IB1 cervical cancer patients with no visible tumors on MRI, 4.7% still exhibited LVSI, underscoring the limitations of MRI in detecting microscopic features. Similarly, Woo et al [16] reported that MRI’s diagnostic rate for postconization residual tumors was 32.7%, significantly lower than the actual occurrence rate of 54.5%, highlighting its restricted sensitivity in detecting residual disease. These limitations demonstrate the need for advanced diagnostic tools, such as artificial intelligence (AI), which can enhance diagnostic accuracy by detecting subtle imaging patterns and providing quantitative, reproducible analyses that surpass human interpretation.

Consequently, there is growing interest in applying image-based AI to improve the accuracy of LVSI detection. AI-based diagnostic tools have demonstrated variable performance in predicting LVSI. Some studies show high performance with area under the curve (AUC) values of 0.94 and 0.923, respectively [17,18]. In contrast, other studies, such as Li et al [19] and Wang et al [20], observed considerably lower performance, with AUC values of 0.72 and 0.73, respectively. These discrepancies can be attributed to factors such as data quality, sample size, and model architecture. Low-quality datasets, such as retrospective studies or single-center studies, may introduce selection bias and limit the generalizability of models, thereby affecting the reliability of radiomics approaches in clinical practice [21]. Additionally, smaller or less diverse sample sizes also restrict model generalizability [22]. These discrepancies highlight the need for further research to assess the reliability and external validation of AI models in clinical practice.

Given the limitations of conventional imaging modalities and the inconsistent performance of AI tools, our systematic review and meta-analysis aim to comprehensively evaluate the diagnostic accuracy of imaging-based AI in predicting LVSI in cervical cancer, addressing critical gaps in current literature.


This meta-analysis was conducted in strict accordance with the Preferred Reporting Items for Systematic Reviews and Meta-Analyses of Diagnostic Test Accuracy (PRISMA-DTA) guidelines [23].

Search Strategy

A systematic literature search was conducted in PubMed, Embase, and Web of Science without date restrictions to identify relevant studies published up to November 9, 2024. The search strategy included 4 groups of terms: those related to AI (eg, “artificial intelligence,” “machine learning,” “deep learning”), cervical cancer (eg, “uterine cervical neoplasms,” “cervix,” “cervical”), lymph-vascular space invasion (eg, “LVSI,” “lymphatic vessel invasion,” “lymphatic permeation”) and imaging modalities (eg, “MRI,” “magnetic resonance imaging,” “PET/CT,” “positron emission tomography-computed tomography”). Both keywords and Medical Subject Headings terms were used, and complete strategy could be seen in Multimedia Appendix 1. We manually reviewed the reference lists of selected studies for additional articles and repeated the search on December 11, 2024, to ensure the inclusion of recent publications.

Inclusion and Exclusion Criteria

The selection of studies adhered to the PICOS (Population, Intervention, Comparison, Outcome, Study Design) framework to ensure rigorous assessment and relevance.

  • Population (P): Studies involved adult patients diagnosed with cervical cancer who underwent LVSI assessment.
  • Intervention (I): Imaging-based AI models, including CT, MRI, ultrasound, and PET/CT, were used for LVSI evaluation.
  • Comparison (C): No diagnostic tool was used as a comparator.
  • Outcome (O): Studies providing quantitative diagnostic performance metrics, including sensitivity, specificity, and AUC, with extractable true positive (TP), true negative (TN), false positive (FP), and false negative (FN) data.
  • Study design (S): Only studies published in English and focused on pretreatment prediction models were included.

Studies were excluded if they met any of the following conditions: (1) irrelevant titles or abstracts; (2) focused solely on lymph node metastasis rather than LVSI; (3) inappropriate article types, including reviews, conference abstracts, case reports, or meta-analyses; case reports were excluded due to their limited sample size, typically involving only a single patient, which makes them unsuitable for statistical analysis or model evaluation; (4) studies with incomplete or uninterpretable data, where TP, FP, FN, and TN values for internal and external validation sets could not be extracted; and (5) animal or in vitro studies. The screening process was conducted in duplicate by 2 independent reviewers (LS and YL), who first evaluated titles and abstracts for relevance. Full-text articles were then assessed against the inclusion and exclusion criteria. Duplicate articles were identified and removed using EndNote’s (Clarivate) duplicate detection tool, followed by manual verification. Discrepancies between reviewers during the screening process were resolved through discussion, and if consensus could not be reached, a third reviewer (HW) was consulted to make the final decision.

Quality Assessment

To comprehensively assess the quality of included studies, we adapted a recognized tool, the revised Quality Assessment of Diagnostic Accuracy Studies-2 (QUADAS-2) [24], by replacing certain criteria with more relevant ones from the Prediction Model Risk of Bias Assessment Tool (PROBAST) [25]. The rationale for this adaptation lies in the complementary strengths of the 2 tools. QUADAS-2 is widely recognized for evaluating the risk of bias in diagnostic accuracy studies, but it does not fully address methodological considerations specific to prediction models, such as variable selection, outcome definitions, or overfitting. PROBAST, on the other hand, is specifically designed to assess risk of bias in prediction model studies. By combining elements from both tools, we tailored the modified QUADAS-2 tool to account for the unique characteristics of studies exploring image-based AI models for LVSI detection.

Our modified QUADAS-2 tool evaluates 4 domains: participants, index test (AI algorithm), reference standard, and analysis. In addition to assessing the risk of bias across each domain, we also evaluated applicability concerns in the first 3 domains. Two independent reviewers applied the modified QUADAS-2 tool to assess the risk of bias in each study, resolving any disagreements through discussion.

Data Extraction

Two independent reviewers assessed study eligibility and performed data extraction, resolving any discrepancies through consensus, with a third reviewer serving as an adjudicator if needed. Data extraction encompassed the primary author’s name, publication year, study design, and country of origin. Additionally, key elements such as imaging modality, reference standard, and total and LVSI+ counts of patients, lesions, or images in training, internal validation, and external validation sets were recorded. Further details included age, AI method, AI model algorithms, and performance outcomes, specifically the TP, FP, TN, and FN counts for both internal and external validation sets. Internal validation refers to the evaluation of model performance using a subset of data from the same source as the training data. External validation, in contrast, involves using an independent dataset that originates from a different source than the training set.

Outcome Measures

The principal outcome measures used in this analysis were sensitivity, specificity, and AUC for both internal and external validation sets. Sensitivity was defined as the proportion of TP scans among the total number of positive cases (TP+FN) for patients, images, or lesions. Specificity was defined as the proportion of true negative (TN) scans among the total number of negative cases (TN+FP). The AUC, representing the area under the receiver operating characteristic curve, served as a summary measure of the model’s diagnostic ability. We extracted AI performance data from internal validation sets and external validation sets. For studies reporting multiple machine learning or deep learning algorithms, only the best diagnostic performance model or algorithm (with the highest AUC value) was included as a representative of the study in the meta-analysis.

Statistical Analysis

A bivariate random-effects analytical approach was used to generate pooled estimations of sensitivity and specificity outcomes for AI-based assessments of LVSI in both internal and external validation sets, accompanied by 95% CIs. We used a summary receiver operating characteristic model to generate the summary receiver operating characteristic curve and calculate the AUC. Additionally, Fagan plots were created to provide a visual representation of the clinical utility of the models.

To assess heterogeneity among studies, we calculated the I² statistic, interpreting values of 25%, 50%, and 75% as indicating low, moderate, and high heterogeneity, respectively. For internal validation sets, we performed subgroup analyses by the number of patients (>150 vs ≤150), region (single vs multiple centers), AI method (deep learning vs machine learning), AI algorithm (logistic regression vs support vector machine), and imaging modality (MRI vs PET/CT). Publication bias was evaluated using Deeks’ funnel plot. All statistical analyses were conducted in Stata 15.1, with significance set at P<.05. Risk of bias for study quality was assessed using RevMan 5.4 (The Cochrane Collaboration) for comprehensive risk of bias assessment.


Study Selection

A systematic literature retrieval and analysis was methodically executed across 3 authoritative databases: PubMed, Embase, and Web of Science, yielding 568 studies. After removing 165 duplicates, 383 studies were excluded based on initial screening criteria. Subsequently, 20 full-text articles were further assessed. In total, 4 studies were excluded due to unavailable data (TP, TN, FP, and FN, n=1), non-English language (n=1), or not focusing on LVSI (n=2). Ultimately, 16 studies met all inclusion criteria and were included in the final analysis [17-20,26-37]. The PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) flowchart detailing this selection process is shown in Figure 1.

Figure 1. Flow chart of the screening process for included studies. TP: true positive; FP: false positive; FN: false negative; TN: true negative; LVSI: lymphovascular space invasion.

Study Description and Quality Assessment

This meta-analysis included 16 studies, all of which included an internal validation set with a total of 1092 participants (ranging from 25 to 198 participants per study). In total, 4 studies included an external validation set [29,30,32,37], involving 203 participants (ranging from 26 to 102 participants). In total, 15 studies used MRI as the imaging modality, while 1 study used PET/CT [37]. Machine learning techniques were applied in 15 studies, with 1 study using deep learning [27]. Among the AI model algorithms, LR was the most common AI model algorithm, used in 9 studies [17,19,28,29,32-35,37], followed by support vector machine (SVM) in 5 studies [18,20,30,31,36], and decision tree and convolutional neural network algorithms, each used in 1 study [26,27]. The gold standard for diagnosis was all pathological examination. All studies were retrospective and conducted in China. Study and patient characteristics, as well as technical features, are summarized in Tables 1-3.

Quality assessment was performed using the QUADAS-2 revised tool, as shown in Figure 2 [9,17,18,20,26-37]. For patient selection, one study was classified as “high risk” due to inappropriate exclusion criteria that may omit specific populations, such as those with a history of radiotherapy or chemotherapy [29]. One study [19] was rated as “unclear” because it was uncertain whether the patients were included consecutively. Regarding the index test in risk of bias, one study was rated as “high risk” for inadequate reporting of model training and evaluation processes [32]. Overall, the quality of the included studies was deemed acceptable.

Table 1. Study and patient characteristics of the included studies.
AuthorYearCountryStudy designImaging modalityReference standard
Yu et al [26]2024ChinaRetroaMRIbPathology
Li et al [9]2019ChinaRetroMRIPathology
Jiang et al [27]2021ChinaRetroMRIPathology
Wu et al [28]2019ChinaRetroMRIPathology
Wu et al [29]2023ChinaRetroMRIPathology
Wang et al [30]2023ChinaRetroMRIPathology
Hua et al [31]2020ChinaRetroMRIPathology
Shi et al [32]2021ChinaRetroMRIPathology
Huang et al [17]2022ChinaRetroMRIPathology
Xiao et al [33]2022ChinaRetroMRIPathology
Cui et al [34]2022ChinaRetroMRIPathology
Ma et al [35]2024ChinaRetroMRIPathology
Wang et al [36]2024ChinaRetroMRIPathology
Li et al [37]2021ChinaRetroPET/CTcPathology
Du et al [18]2021ChinaRetroMRIPathology
Wang et al [20]2019ChinaRetroMRIPathology

aRetro: retrospective.

bMRI: magnetic resonance imaging.

cPET/CT: positron emission tomography-computed tomography.

Table 2. Patient characteristics of included studies.
AuthorYearPatients, lesions, or images per setAgeNumber of LVSIa+ patients, lesions, or images
TrainingInternal validationExternal validation
Yu et al [26]202412060NRb
  • Training: median 53 (range 47-57)
  • Internal validation: median 50 (range 46-54)
  • Training: 46
  • Internal validation: 27
Li et al [19]20197035NR
  • Training: LVSI+: mean 49.34 (SD 8.55); non-LVSI: mean 46.54 (SD 9.66)
  • Internal validation: LVSI+: mean 54.54 (SD 10.42); non-LVSI: mean 48.59 (SD 10.69)
  • Training: 29
  • Internal validation: 13
Jiang et al [27]202120562056NR
  • Median 50.42 (range 27-70)
  • Training: 862
  • Internal validation: 862
Wu et al [28]20195656NR
  • Median 50 (range 29‐67)
  • Training: 31
  • Internal validation: 31
Wu et al [29]202312912939
  • Training: LVSI+: mean 48.07 (SD 10.16); non-LVSI: mean 53.18 (SD 9.47)
  • Internal validation: LVSI+: mean 48.07 (SD 10.16); non-LVSI: mean 53.18 (SD 9.47)
  • External validation: LVSI+: mean 52.08 (SD 12.24); non-LVSI: mean 52.81 (SD 9.46)
  • Training: 46
  • Internal validation: 46
  • External validation: 12
Wang et al [30]2023198198102
  • Training: LVSI+: mean 51.35 (SD 10.49); non-LVSI: mean 51.63 (SD 10.84)
  • Internal validation: LVSI+: mean 51.35 (SD 10.49); non-LVSI: mean 51.63 (SD 10.84)
  • External validation: LVSI+: mean 49.06 (SD 10.52); non-LVSI: mean 51.81 (SD 8.73)
  • Training: 104
  • Internal validation: 104
  • External validation: 54
Hua et al [31]202011156NR
  • Training: LVSI+: mean 49.90 (SD 10.06); non-LVSI: mean 51.34 (SD 8.73)
  • Internal validation: LVSI+: mean 50.57 (SD 7.57); non-LVSI: mean 52.67 (SD 10.04)
  • Internal validation: 23
  • Training: 44
Shi et al [32]20211604436
  • Training: LVSI+: mean 49.97 (SD 9.78); non-LVSI: mean 52.68 (SD 8.72)
  • Internal validation: LVSI+: mean 54.62 (SD 9.07); non-LVSI: mean 49.53 (SD 9.85)
  • External validation: LVSI+: mean 50.21 (SD 12.67); non-LVSI: mean 55.36 (SD 10.36)
  • Training: 65
  • Internal validation: 16
  • External validation: 14
Huang et al [17]202210025NR
  • Total: mean 47.94 (SD 9.01);
  • LVSI+: mean 47.23 (SD 8.23); non-LVSI: mean 48.20 (SD 9.31)
  • Training: 29
  • Internal validation: 5
Xiao et al [33]202215479NR
  • Training: mean 50.0 (SD 9.3)
  • Internal validation: mean 49.9 (SD 10.8);
  • LVSI+: mean 49.5 (SD 10.1); non-LVSI: mean 50.9 (SD 9.2)
  • Training: 106
  • Internal validation: 45
Cui et al [34]202210855NR
  • Training: LVSI+: mean 49.42 (SD 9.16); non-LVSI: mean 51.11 (SD 9.43)
  • Internal validation: LVSI+: mean 52.23 (SD 9.75); non-LVSI: mean 52.94 (SD 9.03)
  • Training: 43
  • Internal validation: 22
Ma et al [35]20248638NR
  • Training: LVSI+: median 52 (range 47.2‐57.5);
  • Non-LVSI: median 57.5 (range 51.2‐64.8)
  • Internal validation: LVSI+: mean 53.2 (SD 6.8); non-LVSI: mean 53.4 (SD 12.7)
  • Training: 24
  • Internal validation: 8
Wang et al [36]20246140NR
  • LVSI+: mean 53.25 (SD 9.72);
  • Non-LVSI: mean 53.57 (SD 10.36)
  • Training: 34
  • Internal validation: 14
Li et al [37]2021612526
  • Training: median 50 (range 33-74)
  • Internal validation: median 51 (range 40-58)
  • External validation: median 52 (range 40-74)
  • Training: 30
  • Internal validation: 12
  • External validation: 15
Du et al [18]202110445NR
  • Training: LVSI+: median 45 (range 37-53); non-LVSI: median 48 (range 40-56)
  • Internal validation: LVSI+: median 44 (range 35-53); non-LVSI: median 46 (range 42-50)
  • Training: 45
  • Internal validation: 22
Wang et al [20]20198040NR
  • Training: median 49.20 (range 29-67)
  • Internal validation: mean 50.45 (SD 32-75)
  • Internal validation: 7
  • Training: 25

aLVSI: lymphovascular space invasion.

bNR not reported.

Table 3. Technical aspects of included studies.
AuthorYearAIa methodAI model algorithmsInterval validation setsExternal validation sets
TPbFPcFNdTNeTPFPFNTN
Yu et al [26]2024Machine learningDTf229524NRgNRNRNR
Li et al [9]2019Machine learningLRh95417NRNRNRNR
Jiang et al [27]2021Deep learningCNNi759296103898NRNRNRNR
Wu et al [28]2019Machine learningLR277418NRNRNRNR
Wu et al [29]2023Machine learningLR3471276105222
Wang et al [30]2023Machine learningSVMj8116237840121436
Hua et al [31]2020Machine learningSVM1711622NRNRNRNR
Shi et al [32]2022Machine learningLR148220137115
Huang et al [17]2022Machine learningLR56014NRNRNRNR
Xiao et al [33]2022Machine learningLR368926NRNRNRNR
Cui et al [34]2022Machine learningLR179524NRNRNRNR
Ma et al [35]2024Machine learningLR67223NRNRNRNR
Wang et al [36]2024Machine learningSVM139117NRNRNRNR
Li et al [37]2021Machine learningLR8041312239
Du et al [18]2021Machine learningSVM183420NRNRNRNR
Wang et al [20]2019Machine learningSVM612121NRNRNRNR

aAI: artificial intelligence.

bTP: true positive.

cFP: false positive.

dFN: false negative.

eTN: true negative.

fDT: decision tree.

gNR: not reported.

hLR: logistic regression.

iCNN: convolutional neural network.

jSVM: support vector machine.

Figure 2. Risk of bias and applicability assessment of included studies using the revised Quality Assessment of Diagnostic Accuracy Studies-2 (QUADAS-2) tool [9,17,18,20,26-37].

Diagnostic Performance of Internal Validation Sets and External Validation Sets for Artificial Intelligence in Predicting Lymph Vascular Space Invasion of Cervical Cancer

For internal validation sets, the sensitivity in detecting LVSI of cervical cancer was 0.84 (95% CI 0.79‐0.87), and the specificity was 0.79 (95% CI 0.75‐0.82; Figure 3 [9,17,18,20,26-37]), with an AUC of 0.88 (95% CI 0.84‐0.90; Figure 4A). Using a pretest probability of 20%, the Fagan nomogram demonstrates a positive likelihood ratio of 49% and a negative likelihood ratio of 5% (Figure 5A).

For external validation sets, the sensitivity in detecting LVSI was 0.79 (95% CI 0.70‐0.86), and the specificity was 0.76 (95% CI 0.67‐0.83; Figure 6 [29,30,32,37]), with an AUC of 0.84 (95% CI 0.81‐0.87; Figure 4B). Using a pretest probability of 20%, the Fagan nomogram demonstrates a positive likelihood ratio of 45% and a negative likelihood ratio of 6% (Figure 5B).

Figure 3. Forest plot of sensitivity and specificity for artificial intelligence–based lymphovascular space invasion diagnosis in cervical cancer: internal validation set [9,17,18,20,26-37].
Figure 4. Receiver operating characteristic (ROC) curves for artificial intelligence–based lymphovascular space invasion prediction in cervical cancer: (A) internal validation set and (B) external validation set. AUC: area under the curve; SENS: sensitivity; SPEC: specificity; SROC: summary receiver operating characteristic.
Figure 5. Fagan’s nomograms for artificial intelligence–based lymphovascular space invasion diagnostic performance in cervical cancer: (A) internal validation set and (B) external validation set. LR: logistic regression.
Figure 6. Forest plot of external validation set for sensitivity and specificity of artificial intelligence in diagnosing lymphovascular space invasion in cervical cancer [29,30,32,37].

Subgroup Analysis of Internal Validation Sets of Image-Based AI for Lymphovascular Space Invasion in Cervical Cancer

In the subgroup analysis, we evaluated diagnostic performance across the number of internal validation patients, region, AI methods, AI algorithms, imaging modalities, and data source (Table 4). Number of internal validation patients (>150 vs ≤150) showed no statistically significant differences in sensitivity (P=.35) and specificity (P=.08). Single-center studies demonstrated significantly different specificities compared to multicenter studies (P<.001). Deep learning (1 study) exhibited higher sensitivity compared to machine learning (15 studies, P=.01), with no significant differences in specificity (P=.29). Among AI algorithms, LR (9 studies) and SVM (5 studies) showed comparable diagnostic performance, with no significant differences in sensitivity (P=.77) or specificity (P=.36). Imaging modality analysis revealed a significant difference in sensitivity between MRI (15 studies) and PET/CT (1 study, P=.01), while specificity remained consistent (P=.29). The radiomic model showed a higher sensitivity, and the radiomic and clinical model showed higher specificity; the differences were statistically significant (P=.05 for both sensitivity and specificity comparisons).

Table 4. Subgroup analysis of imaging-based artificial intelligence performance in internal validation cohorts for lymphovascular space invasion detection in cervical cancer.
SubgroupStudies, nSensitivity (95% CI)Subgroup difference P valueSpecificity (95% CI)Subgroup difference P value
Number of internal validation patients.35.08
>15080.81 (0.75‐0.86)0.80 (0.75‐0.84)
≤15080.84 (0.78‐0.88)0.73 (0.65‐0.79)
Region.09<.001
Single center120.84 (0.79‐0.89)0.74 (0.70‐0.78)
Multiple centers40.77 (0.69‐0.84)0.85 (0.80‐0.89)
AIa method.01.29
Deep learning10.88 (0.86‐0.90)0.75 (0.73‐0.78)
Machine learning150.79 (0.75‐0.83)0.78 (0.74‐0.81)
AI algorithms.77.36
LRb90.79 (0.72‐0.84)0.79 (0.73‐0.84)
SVMc50.80 (0.73‐0.86)0.75 (0.66‐0.81)
Imaging-based.01.29
PET/CTd10.88 (0.86‐0.90)0.75 (0.73‐0.78)
MRIe150.79 (0.75‐0.83)0.78 (0.74‐0.81)
Data source.05.05
Radiomic70.86 (0.81‐0.90)0.75 (0.72‐0.78)
Radiomic and clinical90.78 (0.71‐0.83)0.81 (0.76‐0.85)

aAI: artificial intelligence.

bLR: logistic regression.

cSVM: support vector machine.

dPET/CT: positron emission tomography-computed tomography.

eMRI: magnetic resonance imaging.

Publication Bias

Deeks’ funnel plot asymmetry test showed that there is no significant publication bias both for the internal validation sets (P=.07; Figure 7A) and external validation sets (P=.09; Figure 7B) for AI.

Figure 7. Deeks’ funnel plots for publication bias assessment in artificial intelligence–based lymphovascular space invasion prediction: (A) internal validation set and (B) external validation set.

Principal Findings

Based on our comprehensive literature review, this is the first meta-analysis comprehensively evaluating imaging-based AI for detecting LVSI in cervical cancer. The pooled sensitivity, specificity, and AUC in our analysis for interval validation were 0.84, 0.79, and 0.88, respectively, while for external validation, these metrics were slightly lower at 0.79, 0.76, and 0.84. The slightly lower performance in external validation compared to internal validation is expected when applying AI models to independent datasets. This decline may be due to differences in patient populations, imaging protocols, data quality, and clinical practices across institutions. Despite this, the proposed AI models still show reasonable generalizability. They demonstrate potential for robust performance in diverse datasets and clinical settings.

Deep learning showed superior sensitivity (0.88) compared to machine learning (0.79), likely due to its advanced capabilities in processing complex imaging data. Deep learning models, particularly convolutional neural networks, excel at autonomously extracting hierarchical features from imaging datasets, revealing subtle patterns that manual feature extraction often misses [27]. And by integrating diverse imaging variables like tissue texture, morphological characteristics, and intensity variations, deep learning enables more precise LVSI-positive case differentiation [27,38,39]. However, given the limited number of deep learning studies, the observed higher sensitivity may be insufficiently generalizable. Larger-scale, multicenter studies are essential to validate the potential superiority of deep learning approaches in LVSI detection. Additionally, deep learning is challenged by high overfitting risk and significant computational demands, which can hinder its practical application in clinical settings [27,40].

Comparison to Prior Work

Comparative imaging analysis of LVSI detection revealed differential sensitivity across modalities. MRI-based AI demonstrated a pooled sensitivity of 0.79, whereas PET/CT-based AI reported a marginally superior sensitivity of 0.88. The potential enhanced diagnostic performance of PET/CT can be attributed to its sophisticated capacity for integrating functional and anatomical data, thereby facilitating more nuanced metabolic and structural characterization [41]. However, it is important to note that only 1 study used PET/CT in this analysis, and the limited sample size may have influenced the reported sensitivity. This warrants cautious interpretation of the results and highlights the need for further studies with larger PET/CT-based datasets to confirm these findings.

In 2024, Zhang et al [22] conducted a meta-analysis specifically examining MRI-based radiomics models for predicting LVSI in cervical cancer. Their findings reported a sensitivity of 0.79, specificity of 0.73, and an AUC of 0.83. In contrast, our study achieved superior results in the internal validation set, with a sensitivity of 0.84, specificity of 0.79, and an AUC of 0.88. The improved performance in our research could be attributed to the incorporation of more recent studies and a larger sample size, which likely enhanced the accuracy and robustness of the model. Similarly, a recent meta-analysis by Zhao et al [21] reported comparable diagnostic performance metrics, with a sensitivity of 0.83, specificity of 0.74, and an AUC of 0.86. Unlike their studies focusing only on MRI, our meta-analysis integrated multiple imaging modalities, including MRI and PET/CT, thus offering a more comprehensive evaluation of AI-based diagnostic approaches. Moreover, we pioneered the approach of stratifying the analysis into internal and external validation cohorts, enabling a more rigorous evaluation of the models’ diagnostic performance and generalizability. The slight decline in external validation cohorts indicates acceptable generalizability and provides clinicians with realistic expectations regarding model performance in real-world clinical scenarios.

Heterogeneity

Our meta-analysis demonstrated no significant heterogeneity in both internal and external validation datasets. However, the Deeks’ funnel plot revealed a borderline P value of approximately 0.07, suggesting the potential presence of publication bias. Therefore, we used a bivariate random-effects model for sensitivity and specificity pooling, acknowledging inherent clinical heterogeneity and potential bias. Subgroup analyses were strategically conducted based on the number of patients, region, AI method, AI algorithms, imaging modalities, and data source. The results indicated that both the region and the AI methods significantly influenced the pooled outcomes. However, several additional factors may also contribute to clinical heterogeneity, including variations in AI algorithm architectures, hyperparameter optimization strategies, image acquisition protocols, preprocessing techniques, feature selection methods, and discrepancies in LVSI assessment criteria among pathologists. As reported in Huang et al [17], variations in image acquisition protocols, such as differences in imaging resolution or contrast enhancement methods, may impact the diagnostic performance. Similarly, heterogeneity arising from AI algorithm design, such as the use of different convolutional neural network algorithms, has been shown to influence overall results [27].

Future Directions

Despite these challenges, our findings revealed promising diagnostic performance of AI-based approaches for LVSI detection in cervical cancer, suggesting their potential to streamline clinical workflows and enhance diagnostic accuracy. The implementation of AI-driven approaches in cervical cancer surveillance has increased over the years, yielding encouraging therapeutic prospects [42]. The implementation of imaging-based AI tools could particularly benefit primary health care systems, especially in resource-limited settings where specialist expertise may be scarce. However, challenges include limited computing infrastructure, data storage constraints, and the need for tailored training programs [43]. Additionally, AI systems could optimize screening efficiency and support treatment planning decisions [44]. There is the lack of direct comparisons between the performance of AI models and that of radiologists. Such comparisons are critical to understanding the relative strengths and weaknesses of AI in clinical practice. Future studies should aim to investigate head-to-head comparisons between AI algorithms and radiologists.

Limitations

There are several limitations in our meta-analysis. First, external validation was performed in only 4 of 16 studies, raising concerns about model generalizability. The limited external validation increases the risk of overfitting, where models may perform exceptionally well on training datasets but demonstrate reduced performance on independent datasets. Second, the inherently retrospective design of the included studies may introduce selection biases and limitations in data collection. Second, all studies were conducted in China, which raises concerns about the generalizability of our findings to other populations and health care settings. These factors underscore the critical need for future prospective investigations that not only address the biases associated with retrospective designs but also validate and strengthen the robustness of our findings across diverse demographics and clinical environments. Third, the absence of direct comparative analyses between AI and radiologist interpretations represents a significant research gap; further comparisons are needed.

In conclusion, imaging-based AI, particularly deep learning algorithms, demonstrates promising diagnostic performance in predicting LVSI in cervical cancer. However, the limited external validation datasets and the retrospective nature of the research may introduce potential biases. These findings underscore AI’s potential as an auxiliary diagnostic tool, necessitating further large-scale prospective validation.

Acknowledgments

During the preparation of this work, we used the generative artificial intelligence tool Sider in order to improve readability and language quality. After using this tool, the authors reviewed and edited the content as needed and take full responsibility for the content of the publication.

Data Availability

All primary research findings are presented within this manuscript. For additional information or clarification, please contact the corresponding investigators.

Authors' Contributions

LS and LQ established the investigative framework and research methodology. LS, YL, HW, JZ, YZ, and JC extracted and analyzed the data, while LS composing the initial manuscript draft. All contributing authors participated in the manuscript preparation and approved for the final version before submission.

Conflicts of Interest

None declared.

Multimedia Appendix 1

Additional materials.

DOCX File, 16 KB

Checklist 1

PRISMA (2020) checklist.

PDF File, 74 KB

  1. Singh D, Vignat J, Lorenzoni V, et al. Global estimates of incidence and mortality of cervical cancer in 2020: a baseline analysis of the WHO Global Cervical Cancer Elimination Initiative. Lancet Glob Health. Feb 2023;11(2):e197-e206. [CrossRef] [Medline]
  2. Bray F, Laversanne M, Sung H, et al. Global cancer statistics 2022: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA Cancer J Clin. May 2024;74(3):229-263. [CrossRef]
  3. Meng X, Jiang Y, Chang X, Zhang Y, Guo Y. Conditional survival analysis and real-time prognosis prediction for cervical cancer patients below the age of 65 years. Front Oncol. 2022;12:1049531. [CrossRef] [Medline]
  4. Xie L, Chu R, Wang K, et al. Prognostic assessment of cervical cancer patients by clinical staging and surgical-pathological factor: a support vector machine-based approach. Front Oncol. 2020;10:1353. [CrossRef] [Medline]
  5. Huang Y, Wen W, Li X, Xu D, Liu L. Prognostic value of lymphovascular space invasion in stage IA to IIB cervical cancer: a meta-analysis. Medicine (Baltimore). Apr 14, 2023;102(15):e33547. [CrossRef] [Medline]
  6. Weyl A, Illac C, Lusque A, et al. Prognostic value of lymphovascular space invasion in early-stage cervical cancer. Int J Gynecol Cancer. Oct 2020;30(10):1493-1499. [CrossRef] [Medline]
  7. Marchiolé P, Buénerd A, Benchaib M, Nezhat K, Dargent D, Mathevet P. Clinical significance of lympho vascular space involvement and lymph node micrometastases in early-stage cervical cancer: a retrospective case-control surgico-pathological study. Gynecol Oncol. Jun 2005;97(3):727-732. [CrossRef] [Medline]
  8. Pol FJM, Zusterzeel PLM, van Ham M, Kuijpers DAT, Bulten J, Massuger L. Satellite lymphovascular space invasion: an independent risk factor in early stage cervical cancer. Gynecol Oncol. Sep 2015;138(3):579-584. [CrossRef] [Medline]
  9. Li P, Liu P, Yang Y, et al. Hazard ratio analysis of laparoscopic radical hysterectomy for IA1 With LVSI-IIA2 cervical cancer: identifying the possible contraindications of laparoscopic surgery for cervical cancer. Front Oncol. 2020;10. [CrossRef]
  10. Abu-Rustum NR, Yashar CM, Arend R, et al. NCCN Guidelines® Insights: cervical cancer, version 1.2024. J Natl Compr Canc Netw. Dec 2023;21(12):1224-1233. [CrossRef] [Medline]
  11. Holtz DO, Dunton C. Traditional management of invasive cervical cancer. Obstet Gynecol Clin North Am. Dec 2002;29(4):645-657. [CrossRef] [Medline]
  12. Zhu Y, Shen B, Pei X, Liu H, Li G. CT, MRI, and PET imaging features in cervical cancer staging and lymph node metastasis. Am J Transl Res. 2021;13(9):10536-10544. [CrossRef] [Medline]
  13. Cabello J, Ziegler SI. Advances in PET/MR instrumentation and image reconstruction. Br J Radiol. Jan 2018;91(1081):20160363. [CrossRef] [Medline]
  14. Lucia F, Visvikis D, Desseroit MC, et al. Prediction of outcome using pretreatment 18F-FDG PET/CT and MRI radiomics in locally advanced cervical cancer treated with chemoradiotherapy. Eur J Nucl Med Mol Imaging. May 2018;45(5):768-786. [CrossRef]
  15. Park JY, Lee JW, Park BK, et al. Postoperative outcomes of MR-invisible stage IB1 cervical cancer. Am J Obstet Gynecol. Aug 2014;211(2):168. [CrossRef] [Medline]
  16. Woo S, Kim HS, Chung HH, Kim SY, Kim SH, Cho JY. Early stage cervical cancer: role of magnetic resonance imaging after conization in determining residual tumor. Acta Radiol. Oct 2016;57(10):1268-1276. [CrossRef]
  17. Huang G, Cui Y, Wang P, et al. Multi-parametric magnetic resonance imaging-based radiomics analysis of cervical cancer for preoperative prediction of lymphovascular space invasion. Front Oncol. 2021;11:663370. [CrossRef] [Medline]
  18. Du W, Wang Y, Li D, et al. Preoperative prediction of lymphovascular space invasion in cervical cancer with radiomics -based nomogram. Front Oncol. 2021;11:637794. [CrossRef] [Medline]
  19. Li Z, Li H, Wang S, et al. MR-based radiomics nomogram of cervical cancer in prediction of the lymph-vascular space invasion preoperatively. J Magn Reson Imaging. May 2019;49(5):1420-1426. [CrossRef] [Medline]
  20. Wang S, Chen X, Liu Z. Radiomics analysis on T2-MR image to predict lymphovascular space invasion in cervical cancer. Presented at: Medical Imaging 2019: Computer-Aided Diagnosis; 2019; San Diego, California, USA.
  21. Zhao M, Li Z, Gu X, et al. The role of radiomics for predicting of lymph-vascular space invasion in cervical cancer patients based on artificial intelligence: a systematic review and meta-analysis. J Gynecol Oncol. 2024;36(2). [CrossRef]
  22. Zhang H, Teng C, Yao Y, et al. MRI-based radiomics models for noninvasive evaluation of lymphovascular space invasion in cervical cancer: a systematic review and meta-analysis. Clin Radiol. Nov 2024;79(11):e1372-e1382. [CrossRef]
  23. McInnes MDF, Moher D, Thombs BD, et al. Preferred reporting items for a systematic review and meta-analysis of diagnostic test accuracy studies: The PRISMA-DTA statement. J Am Med Assoc. Jan 23, 2018;319(4):388-396. [CrossRef] [Medline]
  24. Whiting PF, Rutjes AWS, Westwood ME, et al. QUADAS-2: a revised tool for the quality assessment of diagnostic accuracy studies. Ann Intern Med. Oct 18, 2011;155(8):529-536. [CrossRef] [Medline]
  25. Wolff RF, Moons KGM, Riley RD, et al. PROBAST: a tool to assess the risk of bias and applicability of prediction model studies. Ann Intern Med. Jan 1, 2019;170(1):51-58. [CrossRef] [Medline]
  26. Yu Z, Zhihui Q, Linrui L, Long L, Qibing W. Machine learning-based models for assessing postoperative risk factors in patients with cervical cancer. Acad Radiol. Apr 2024;31(4):1410-1418. [CrossRef]
  27. Jiang X, Li J, Kan Y, et al. MRI based radiomics approach with deep learning for prediction of vessel invasion in early-stage cervical cancer. IEEE/ACM Trans Comput Biol Bioinform. 2021;18(3):995-1002. [CrossRef] [Medline]
  28. Wu Q, Shi D, Dou S, et al. Radiomics analysis of multiparametric MRI evaluates the pathological features of cervical squamous cell carcinoma. J Magn Reson Imaging. Apr 2019;49(4):1141-1148. [CrossRef] [Medline]
  29. Wu Y, Wang S, Chen Y, et al. A multicenter study on preoperative assessment of lymphovascular space invasion in early-stage cervical cancer based on multimodal MR radiomics. J Magn Reson Imaging. Nov 2023;58(5):1638-1648. [CrossRef] [Medline]
  30. Wang S, Liu X, Wu Y, et al. Habitat-based radiomics enhances the ability to predict lymphovascular space invasion in cervical cancer: a multi-center study. Front Oncol. 2023;13:1252074. [CrossRef] [Medline]
  31. Hua W, Xiao T, Jiang X, et al. Lymph-vascular space invasion prediction in cervical cancer: exploring radiomics and deep learning multilevel features of tumor and peritumor tissue on multiparametric MRI. Biomed Signal Process Control. Apr 2020;58:101869. [CrossRef]
  32. Shi J, Cui L, Wang H, et al. MRI-based intratumoral and peritumoral radiomics on prediction of lymph-vascular space invasion in cervical cancer: a multi-center study. Biomed Signal Process Control. Feb 2022;72:103373. [CrossRef]
  33. Xiao M, Li Y, Ma F, Zhang G, Qiang J. Multiparametric MRI radiomics nomogram for predicting lymph-vascular space invasion in early-stage cervical cancer. Br J Radiol. Jun 1, 2022;95(1134):20211076. [CrossRef] [Medline]
  34. Cui L, Yu T, Kan Y, Dong Y, Luo Y, Jiang X. Multi-parametric MRI-based peritumoral radiomics on prediction of lymph-vascular space invasion in early-stage cervical cancer. Diagn Interv Radiol. Jul 2022;28(4):312-321. [CrossRef] [Medline]
  35. Ma NN, Wang T, Lv YN, Li SD. An MRI radiomics-based model for the prediction of invasion of the lymphovascular space in patients with cervical cancer. Front Oncol. 2024;14. [CrossRef]
  36. Wang H, Meng J, Dong G, et al. Multi-parametric MRI combined with radiomics for the evaluation of lymphovascular space invasion in cervical cancer. Clin Exp Obstet Gynecol. 2024;51(4). [CrossRef]
  37. Li X, Xu C, Yu Y, Guo Y, Sun H. Prediction of lymphovascular space invasion using a combination of tenascin-C, cox-2, and PET/CT radiomics in patients with early-stage cervical squamous cell carcinoma. BMC Cancer. 2021. [CrossRef]
  38. Mahmood T, Li J, Pei Y, Akhtar F. An automated in-depth feature learning algorithm for breast abnormality prognosis and robust characterization from mammography images using deep transfer learning. Biology (Basel). Sep 2, 2021;10(9):859. [CrossRef] [Medline]
  39. Tran KA, Kondrashova O, Bradley A, Williams ED, Pearson JV, Waddell N. Deep learning in cancer diagnosis, prognosis and treatment selection. Genome Med. Sep 27, 2021;13(1):152. [CrossRef] [Medline]
  40. Bejani MM, Ghatee M. A systematic review on overfitting control in shallow and deep neural networks. Artif Intell Rev. Dec 2021;54(8):6391-6438. [CrossRef]
  41. Adam JA, van Diepen PR, Mom CH, Stoker J, van Eck-Smit BLF, Bipat S. [18F]FDG-PET or PET/CT in the evaluation of pelvic and para-aortic lymph nodes in patients with locally advanced cervical cancer: a systematic review of the literature. Gynecol Oncol. Nov 2020;159(2):588-596. [CrossRef] [Medline]
  42. Vargas-Cardona HD, Rodriguez-Lopez M, Arrivillaga M, et al. Artificial intelligence for cervical cancer screening: scoping review, 2009-2022. Int J Gynaecol Obstet. May 2024;165(2):566-578. [CrossRef] [Medline]
  43. Allahqoli L, Laganà AS, Mazidimoradi A, et al. Diagnosis of cervical cancer and pre-cancerous lesions by artificial intelligence: a systematic review. Diagnostics (Basel). 2022;12(11):2771. [CrossRef] [PMC]
  44. Esteva A, Robicquet A, Ramsundar B, et al. A guide to deep learning in healthcare. Nat Med. Jan 2019;25(1):24-29. [CrossRef] [Medline]


AI: artificial intelligence
AUC: area under the curve
CT: computed tomography
FN: false negative
FP: false positive
LVSI: lymphovascular space invasion
MRI: magnetic resonance imaging
PET/CT: positron emission tomography/computed tomography
PICOS: Population, Intervention, Comparison, Outcome, Study Design
PRISMA: Preferred Reporting Items for Systematic Reviews and Meta-Analyses
PRISMA-DTA: Preferred Reporting Items for Systematic Reviews and Meta-Analyses of Diagnostic Test Accuracy
PROBAST: Prediction Model Risk of Bias Assessment Tool
QUADAS-2: Quality Assessment of Diagnostic Accuracy Studies-2
SVM: support vector machine
TN: true negative
TP: true positive


Edited by Naomi Cahill; submitted 09.01.25; peer-reviewed by Amir Hassankhani, Mohammadhosein Akhlaghpasand; final revised version received 23.04.25; accepted 23.04.25; published 16.06.25.

Copyright

© Lizhen She, Yunfeng Li, Hongyong Wang, Jun Zhang, Yuechen Zhao, Jie Cui, Ling Qiu. Originally published in the Journal of Medical Internet Research (https://www.jmir.org), 16.6.2025.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research (ISSN 1438-8871), is properly cited. The complete bibliographic information, a link to the original publication on https://www.jmir.org/, as well as this copyright and license information must be included.