Published on in Vol 28 (2026)

Preprints (earlier versions) of this paper are available at https://preprints.jmir.org/preprint/78377, first published .
The Performance of Artificial Intelligence in Classifying Molecular Markers in Adult-Type Gliomas Using Histopathological Images: Systematic Review

The Performance of Artificial Intelligence in Classifying Molecular Markers in Adult-Type Gliomas Using Histopathological Images: Systematic Review

The Performance of Artificial Intelligence in Classifying Molecular Markers in Adult-Type Gliomas Using Histopathological Images: Systematic Review

1Faculty of Medicine, Yarmouk University, Irbid, Jordan

2Department of Computer Science and Engineering, College of Engineering, Qatar University, Doha, Qatar

3AI Center for Precision Health, Weill Cornell Medical College in Qatar, 103 Ezdan Oasis, Alwakrah, Doha, Qatar

4Department of Biomedical Sciences, College of Health Sciences, Qatar University, Doha, Qatar

Corresponding Author:

Alaa Abd-alrazaq, PhD


Background: Adult-type gliomas are among the most prevalent and lethal primary central nervous system tumors, where prompt and accurate diagnosis is essential for maximizing survival prospects. Molecular classification, particularly the detection of isocitrate dehydrogenase (IDH) mutations and 1p/19q codeletions, has become crucial for accurate diagnosis and prognosis. Artificial intelligence (AI) has emerged as a promising adjunct in enhancing diagnostic accuracy using histopathological images. Existing reviews mostly focused on radiology rather than histopathology, and no comprehensive systematic review has specifically evaluated AI performance exclusively from histopathological images for detecting these two molecular markers.

Objective: This study aims to systematically evaluate the performance of AI models in detecting and classifying IDH mutation status and 1p/19q gene codeletion in adult-type gliomas using histopathological images.

Methods: A systematic review was conducted in accordance with PRISMA-DTA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses–Extension for Diagnostic Test Accuracy) guidelines. Seven databases (MEDLINE, PsycINFO, Embase, IEEE Xplore, ACM Digital Library, Scopus, and Google Scholar) were searched for studies published between 2015 and 2025. Eligible studies used AI models on histopathological images for molecular classification of adult-type gliomas and reported performance metrics. Study selection, data extraction, and risk of bias assessment using a modified QUADAS-2 (Quality Assessment of Diagnostic Accuracy Studies 2) tool were conducted independently by two reviewers. Extracted data were synthesized narratively.

Results: A total of 2453 reports were identified, with 22 studies meeting the inclusion criteria. The pooled average accuracy, sensitivity, specificity, and area under the curve (AUC) across studies were 85.46%, 84.55%, 86.03%, and 86.53%, respectively. Hybrid models demonstrated the highest diagnostic performance (accuracy 92.80% and sensitivity 89.62%). In general, AI models that used multimodal data outperformed those that used unimodal data in terms of sensitivity (90.15% vs 84.31%) and AUC (88.93% vs 86.29%). Furthermore, models had a better overall performance in identifying IDH mutations than 1p/19q codeletions, with higher accuracy (86.13% vs 81.63%), specificity (86.61% vs 78.11%), and AUC (86.74% vs 85.15%). Unexpectedly, AI models designed for binary classification exhibited lower performance than those for multiclass classification in terms of both accuracy (91.98% vs 84.02%) and sensitivity (93.41% vs 80.18%). However, these differences should be interpreted as descriptive trends rather than statistically validated superiority, as formal between-group comparisons were not feasible.

Conclusions: AI models show strong potential as complementary tools for the molecular classification of adult-type gliomas using histopathology images, particularly for IDH mutation detection. However, these findings are constrained by the limited number of studies, the focus on adult-type gliomas, lack of meta-analysis, and restriction to English-language publications. While AI offers valuable diagnostic support, it must be integrated with expert clinical judgment. Future research should prioritize larger, more diverse datasets and multimodal AI frameworks and extend to other brain tumor types for broader applicability.

Trial Registration: PROSPERO CRD420250653668; https://www.crd.york.ac.uk/PROSPERO/view/CRD420250653668

J Med Internet Res 2026;28:e78377

doi:10.2196/78377

Keywords



Background

Gliomas are the most frequent primary tumors affecting the central nervous system (CNS) in adults, with a variable prognosis depending on the specific subtype and histological grade [1]. The global prevalence of CNS neoplasms in 5 years is approximately 771,110 cases. Primary brain tumors account for 1.7% of all cancers, with a global incidence of 3.9 cases per 100,000 people each year. Among these, gliomas are the most common malignant type, accounting for 75% of all malignant CNS tumors, with an incidence of 6 cases per 100,000 people annually [2]. Gliomas are among the tumors that are difficult to treat, with a 5-year overall survival of less than 35% [3]. Patients with brain tumors present with variable clinical symptoms according to the part of the brain affected. However, they usually share general symptoms that are nonspecific to anatomic location (eg, seizures, headaches).

Gliomas are neuroectodermal tumors arising from glial cells or their precursors and encompass astrocytomas, oligodendrogliomas, and ependymomas [3]. Mutations in isocitrate dehydrogenase 1 and 2 (IDH1 and IDH2) are considered early events in gliomagenesis and occur more frequently in lower-grade gliomas [3]. Diffuse gliomas harboring IDH1/2 mutations are associated with significantly improved prognosis compared with IDH–wild-type diffuse gliomas [3].

Codeletion of chromosomal arms 1p and 19q results from an unbalanced centromeric translocation, t(1;19)(q10;p10) [3]. In combination with IDH mutation, 1p/19q codeletion is a defining molecular feature required for the diagnosis of oligodendroglioma, IDH mutant, and 1p/19q codeleted. This molecular alteration is associated with a favorable prognosis among diffuse gliomas and predicts enhanced responsiveness to alkylating chemotherapy [3].

Conventional classification of gliomas relied on histologic features as a gold standard. Initially, pathologists utilized microscopic analysis of histochemically stained (eg, hematoxylin and eosin) sections for the diagnosis of gliomas [4]. However, ancillary immunohistochemistry has been increasingly used over the past few decades to enhance diagnostic accuracy [4]. With either method, pathologists followed a diagnostic algorithm to rule out nonneoplastic lesions (eg, infarcts) or other types of malignancies (eg, metastatic) based on histologic features. Clinical information (eg, age) and radiologic findings (eg, location) are also valuable clues in the diagnosis [4]. Nonetheless, the histologic-based classification system has a few limitations, particularly interobserver variability, in addition to insufficient or nonrepresentative tissue sampling [5]. Although histologic-based classification can be accurate for prototypic tumors, classification of tumors encountered in clinical practice is often more challenging due to a degree of mixed features in the same sample [5].

In order to address these limitations, molecular-based diagnosis has emerged as a possible solution. Recent advancements in genetics have helped to identify specific gene alterations involved in the development of different types of gliomas. Detection of these gene alterations is the base of molecular-based diagnosis. This method offers improved accuracy and made it possible to distinguish between primary and secondary gliomas, which was not possible with the previous histologic-based method [6]. Therefore, the World Health Organization (WHO) integrated molecular features into its most recent classification of CNS tumors, published in 2021, based on the presence or absence of IDH1 and IDH2 mutations, in addition to chromosome 1p/19q codeletion [7]. It categorized adult-type gliomas into three categories: astrocytoma with IDH mutation, oligodendroglioma with IDH mutation and 1p/19q codeletion, and glioblastoma with IDH wild-type [7]. Furthermore, a combined approach using molecular and histologic features is used for the grading of IDH-mutant astrocytomas as CNS WHO grade 2, 3, or 4. The new WHO classification also led to a more accurate prognosis of gliomas, as the absence of IDH mutation is often a sign of more aggressive tumors [7].

Neuroimaging provides noninvasive insight into the molecular landscape of diffuse gliomas. Advanced magnetic resonance imaging (MRI) features can suggest IDH mutation status, as IDH-mutant tumors typically exhibit lower perfusion and less aggressive radiological behavior than IDH–wild-type gliomas [5-7]. Magnetic resonance spectroscopy allows in vivo detection of the IDH-specific oncometabolite D-2-hydroxyglutarate, enabling preoperative identification of IDH-mutant tumors [7]. In addition, oligodendrogliomas defined by combined IDH mutation and 1p/19q codeletion display characteristic imaging features, including cortical localization, calcifications, and relatively increased perfusion, reflecting distinct tumor biology [4,7]. These radiogenomic associations support the integration of imaging into molecular stratification of adult-type diffuse gliomas [4-6].

As gliomas pose diagnostic challenges due to their complex histological and molecular features, artificial intelligence (AI) has emerged as a promising tool to enhance diagnostic accuracy and efficiency. AI models were able to offer rapid and accurate results with accuracy exceeding 90%, which can further improve the access and speed of molecular diagnosis [8]. In general, AI implementation in health care can lead to improved data synthesis (eg, patient data, medical literature) [9]. Additionally, AI can process multimodal data (two or more different data sources such as histopathological images, radiological data, clinical variables, genomic markers, and other relevant data types) to provide more accurate diagnosis and predictions. Moreover, it offers means for augmenting human performance, as human limitations can prevent processing large quantities of information and making the optimal judgment at all times. Thus, AI can improve care consistency, increase precision, accelerate discoveries, and minimize disparities [9].

Research Gaps and Aim

The use of AI in the diagnosis of brain tumors and gliomas has been widely studied, with many reviews summarizing the results of numerous studies. However, existing reviews exhibit several limitations. For instance, many reviews focused on AI implementation in other diagnostic methods of gliomas, such as radiology-based [10,11], surgery-based [12], and DNA-based diagnoses [13]. As for reviews focusing on AI implementation in the molecular diagnosis of gliomas, they fall into one of the following categories: not systematic [14], outdated [15], restricted by databases or search terms [11,16], lacking subtyping of gliomas [17], not focused on genetic markers (IDH mutations and 1p/19q codeletion) [11], or limited to a specific subfield of AI (eg, deep learning or machine learning) [16]. To address the limitations of the previous reviews, this study aims to systematically assess the performance of AI in detecting and classifying IDH mutation status and 1p/19q gene codeletion in adult-type gliomas using histopathological images.


This review was conducted in accordance with the PRISMA-DTA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses–Extension for Diagnostic Test Accuracy) [18] (see Checklist 1) and registered with the international Prospective Register of Systematic Reviews (PROSPERO) under the number CRD420250653668.

Search Strategy

The literature search was performed on November 15, 2024, using the following digital databases: MEDLINE, PsycINFO, Embase, IEEE Xplore, ACM Digital Library, Scopus, and Google Scholar. A biweekly automated search was set for 4 months, concluding on March 15, 2025. Given the large number of results from Google Scholar, which are ranked by relevance, we reviewed only the first 100 entries (10 pages). Our query utilized a combination of proper keywords and Boolean operators, including but not limited to ((“Artificial Intelligence” OR “Machine Learning” OR “Deep Learning”) AND (“Histopatholog*” OR “Whole-Slide Imag*”) AND (“Glioma*” OR “Astrocytoma*” OR “Oligodendroglioma*” OR “Glioblastoma*”)). The search was restricted to studies published between 2015 and 2025. The search query for each database is shown in Multimedia Appendix 1. To identify additional studies, we screened the reference lists of included studies (ie, backward reference list checking) and checked studies that cited the included studies (ie, forward reference list checking).

Study Selection

The selection process of relevant studies consisted of three steps. The first step involved removing duplicates using EndNote. In the second step, the studies were assigned to two authors, who screened the papers independently by title and abstract. The final step involved each author reading the consequent papers in full text independently. Discussions were held to resolve any differences or ambiguities in steps two and three.

Study Eligibility Criteria

Studies fulfilling the following criteria were included: (1) original articles, theses, dissertations, and conference papers written in English, (2) studies published in 2015 onward, (3) studies focused on patients with adult-type glioma with no restrictions on age, (4) studies that used histopathological images gathered from tissue biopsies, tumor excision, or digital histopathological images, (5) studies that assessed the performance of AI algorithms in the detection and classification of the molecular markers, specifically the IDH mutation status and the 1p/19q gene codeletion status using histopathology data, and (6) studies that reported results related to the performance of the developed AI models. No restriction was applied regarding study settings, study design, and country of publication. Studies were excluded if they (1) used data modalities other than histopathology images (eg, radiological images, stimulated Raman histology), (2) did not employ AI algorithms for molecular marker detection or classification, (3) were not human studies, or (4) were categorized as editorials, letters to editors, posters, conference abstracts, reviews, commentaries, short communications, and brief reports.

Data Extraction

Two authors independently extracted data from the eligible studies using predefined tables in Microsoft Excel sheet. Any discrepancies in the extracted data were resolved through discussion. The extracted data included four main categories: (1) study characteristics, including surname of first author, publication type, publication year, and country of publication; (2) participants and dataset characteristics, including number of participants, mean age, female percentage, dataset size, data source, magnification, slide image resolution, level of analysis, number of classes, and ground truth assessor; (3) AI model characteristics, covering problem-solving approaches, AI classifiers, type of classification, aim of classifier, data input to AI algorithms, features extraction methods, type of validation, and performance metrics; (4) results, including accuracy, sensitivity, specificity, and area under the curve (AUC). The data extraction form is shown in Multimedia Appendix 2.

Risk of Bias Assessment

To evaluate the quality and reliability of the included studies, two reviewers independently conducted a risk of bias assessment using an adapted version of the QUADAS-2 (Quality Assessment of Diagnostic Accuracy Studies 2) tool [19]. The original QUADAS-2 tool was modified to better suit the context of our review, which focuses on AI applications. Specifically, we incorporated relevant elements from the PROBAST (Prediction Model Risk of Bias Assessment Tool) to ensure that the assessment adequately addressed the methodological nuances of AI-based diagnostic models [20]. Each domain included four tailored signaling questions designed to determine the potential for bias and to evaluate applicability to the review’s research question. The risk of bias was rated separately for each domain, while concerns regarding applicability were assessed for the first three domains. For example, the “Participants” domain examined issues such as sample representativeness, exclusions, and subgroup balance. The “Index Test” domain evaluated the transparency and consistency of the AI model’s design and feature use. The “Reference Standard” focused on the validity and consistency of outcome definitions and assessor qualifications, while the “Analysis” domain addressed data inclusion, preprocessing, set division, and model performance evaluation. In addition to risk of bias, we also assessed concerns regarding applicability in the first three domains, evaluating whether the participants, AI model, and outcome definitions were aligned with the review’s objectives. To refine and validate our modified tool, we conducted a pilot assessment using five studies, which allowed us to fine-tune the criteria before full application. Subsequently, two reviewers independently assessed all included studies using the finalized tool (Multimedia Appendix 3). Any discrepancies in assessments were resolved through discussion and consensus.

Data Synthesis

A narrative approach was used to synthesize the data extracted from the included studies. Specifically, we used texts and tables to summarize and describe the characteristics of the included studies and results. We could not perform a meta-analysis because fewer than two studies reported the confusion matrices or other necessary details (eg, number of cases and controls in the test set) required to calculate the numerators and denominators for the same performance measures (accuracy, sensitivity, specificity, and AUC), AI algorithms, and measured outcomes (detection of IDH and 1p/19q). Instead of meta-analysis, we calculated the traditional mean for each metric (ie, accuracy, sensitivity, specificity, and AUC).


Search Results

As shown in Figure 1, the literature searches identified a total of 2453 reports from databases. After removing duplicate records (n=583), we screened the titles and abstracts of 1870 reports. This led to the exclusion of 1753 reports. Of the remaining 117 reports, we could not retrieve the full text for 5 reports. Subsequently, we checked the eligibility of the remaining 112 reports. This led to the exclusion of 91 reports for the following reasons: not related to glioma (n=7), not related to molecular classification (n=53), not pertaining to histopathology (n=11), inappropriate publication type (n=19), and not in the English language (n=1). We included an additional study by checking the reference list of the included studies. Ultimately, only 22 reports met the inclusion criteria and were included in the review [21-42].

Figure 1. Flowchart of the study selection process.

Characteristics of the Included Studies

The included studies were published between 2015 and 2024 (Table 1). The majority of studies were published in 2023 (9/22, 40.91%), followed by 2024 (5/22, 22.73%). Only 3 of the 22 studies (13.64%) are conference papers, whereas 16 (72.73%) are journal articles and 3 (13.64%) are preprints. Studies were carried out in different countries, with the United States and China being the most frequent contributors (6/22, 27.27%) and (5/22, 22.73%), respectively. The number of participants in the included studies ranges from 29 to 2845, with an average of 765.1 (SD 815.6). The mean of female participation rate is 42.45% with a range between 34.5% and 49.3%. The dataset size (ie, number of whole-slide image or patches) varied widely (47‐27,000), with an average of 2464 (SD 5851). It is worth mentioning that the unit of analysis used to compute performance metrics was predominantly whole-slide image level (21/22, 95.5%), while it was at the patch level in 1 (4.5%) study. The participants’ age ranged from 42.2 to 54.5 with an average of 48.03 (SD 4.2; Table 1). The characteristics of each included study are described in Multimedia Appendix 4.

Table 1. Characteristics of included studies.
FeaturesNumber of studiesReferences
Year of publication, n (%)
20245 (22.73)[24,28,35,37,42]
20239 (40.91)[21,22,25,27,30-33,40]
20222 (9.09)[26,34]
20213 (13.64)[29,38,41]
20202 (9.09)[23,36]
20191 (4.55)[39]
Publication type, n (%)
Journal article16 (72.73)[23,24,26,27,29-31,33-38,40-42]
Conference paper3 (13.64)[21,28,32]
Preprint3 (13.64)[22,25,39]
Country of publication, n (%)
United States6 (27.27)[28,29,32,34,38,39]
China5 (22.73)[23,25,35,40,42]
Germany2 (9.09)[21,27]
Australia2 (9.09)[36,41]
India1 (4.55)[22]
Luxembourg1 (4.55)[24]
Canada1 (4.55)[26]
Switzerland1 (4.55)[30]
South Korea1 (4.55)[31]
Saudi Arabia1 (4.55)[33]
Japan1 (4.55)[37]
Number of participantsa, mean (SD); range (%)765.1 (815.6); 29‐2845 (95.45)[21,22,24-42]
Female percentageb (%), mean (SD); range (%)42.45 (0.03); 34.5‐49.3 (59.09)[24,25,29,31,33-37,39-42]
Dataset sizec, mean (SD); range (%)2464 (5851); 47‐27,000 (95.45)[22-42]
Mean aged (year), mean (SD); range (%)48.03 (4.2); 42.2‐54.5 (59.09)[24,25,29,31,33-37,39-42]

aOne study did not report the mean age [23].

bStudies did not report the female percentage [21-23,26-28,30,32,38].

cOne study did not report the dataset size [21].

dStudies did not report the mean age [21-23,26-28,30,32,38].

Features of Histopathological Images

As presented in Table 2, the histopathological images were captured at various magnification levels, including 2.5×, 5×, 10×, 20×, 40×, and 100×. Among these, 20× magnification was the most commonly used (12/22, 54.55%). While 14 (63.64%) studies focused on architectural-level analysis, 13 (59.09%) studies focused on cellular-level analysis. Slide resolutions varied across studies, with 256×256 pixels being the most common (8/22, 36.36%). Datasets were sourced from either open-source databases (16/22, 72.73%) or closed-source datasets (13/22, 59.09%) (ie, data were collected by the study authors or obtained from previous studies). The open datasets used include The Cancer Genome Atlas (15/22, 68.18%) and the Digital Brain Tumor Atlas (1/22, 4.55%), with some studies combining open and closed sources. The histopathological images were labeled into 2 classes in about two-thirds of the studies (14/22, 63.64%). The features of histopathological images in each study are described in Multimedia Appendix 5.

Table 2. Features of histopathological imagesa.
FeaturesNumber of studies, n (%)References
Magnification
2.5×1 (4.55)[34]
2 (9.09)[32,34]
10×4 (18.18)[24,29,34,36]
20×b12 (54.55)[22,24-26,28,31,32,34,35,40-42]
40×5 (22.73)[24,30,31,35,41]
100×1 (4.55)[24]
Not reported7 (31.82)[21,23,27,33,37-39]
Level of analysis
Architecture level14 (63.64)[22,24-26,28,29,31,32,34-36,40-42]
Cell level13 (59.09)[22,24-26,28,30-32,34,35,40-42]
Not reported7 (31.82)[21,23,27,33,37-39]
Slide image resolution
224×2244 (18.18)[29,31,37,42]
256×2568 (36.36)[22,23,25,28,32,34,36,37]
512×5123 (13.64)[21,23,38]
1024×10245 (22.73)[23,33,35,36,40]
Other7 (31.82)[21,23,24,26,30,41,42]
Not reported2 (9.09)[27,39]
Data sources
Open source16 (72.73)[21-23,25,27-29,31-34,36-39,41]
Closed source13 (59.09)[24-27,30,31,33-36,40-42]
Open dataset name
TCGAc15 (68.18)[21,23,25,27-29,31-34,36-39,41]
DBTAd1 (4.55)[22]
Not applicable6 (27.27)[24,26,30,35,40,42]
Number of classes
Two14 (63.64)[22,23,28-37,41,42]
Three2 (9.09)[21,26]
Four4 (18.18)[25,27,38,39]
Five2 (9.09)[24,33]
Six1 (4.55)[40]

aStudies used both data sources, open and closed [25,31,33,34,36,41].

bValues in italics indicate highest score.

cTCGA: The Cancer Genome Atlas.

dDBTA: Digital Brain Tumor Atlas.

Features of AI

AI was primarily used for binary classification in 86.36% (19/22) of studies, where models differentiate between two outcomes (eg, IDH-mutant vs wild-type), whereas only 22.73% (5/22) employed multiclass classification, which predicts among 3 or more glioma subtypes (Table 3). The majority of studies (20/22, 90.91%) used AI for IDH subtyping (ie, distinguishing between IDH mutant and IDH wild-type), whereas only 6 (27.27%) studies applied AI for subtyping IDH mutant tumors in relation to 1p/19q gene codeletion. Convolutional neural networks (CNNs) were the most commonly used classifier and feature extraction method, appearing in 54.55% (12/22) and 90.91% (20/22) of studies. For specific classifiers, CNNs are the most utilized, appearing in 54.55% (12/22) of studies, followed by multiple instance learning (MIL) (which handles weakly labeled slide-level data by learning from sets of image patches without requiring pixel-level annotations) at 50% (11/22), transformers at 18.18% (4/22), and hybrid models (ie, single end-to-end architectures that integrate convolutional components with attention-based mechanisms [eg, CNN-transformer models] within a unified computational graph and are trained jointly) also at 18.18% (4/22). For specific feature extraction models, ResNet50 leads with 46.67% (10/22) usage, followed by densely connected convolutional network 121 (DenseNet121) and InceptionV3, both at 19.05% (4/22). Among the included studies, most models were developed using histopathological images alone (17/22, 77.3%), while a smaller number (5/22, 22.7%) incorporated additional modalities such as MRI scans (2/22, 9.1%), demographic information (3/22, 13.6%), and clinical variables (2/22, 9.1%). The most widely used method for model validation was k-fold cross-validation, used in 50% (11/22) of studies, followed by the train-test split approach in 45.45% (10/22) of studies. AI model performance was assessed using multiple metrics, with accuracy being the most frequently reported (19/22, 86.36%), followed by AUC and sensitivity, each used in 68.18% (15/22) of studies. A detailed summary of the AI models used in each study is provided in Multimedia Appendix 6.

Table 3. Features of artificial intelligence (AI).
FeatureNumber of studies, n (%)References
Type of classification
Binarya19 (86.36)[22,23,25,27-42]
 Multiclass5 (22.73)[21,24,26,33,40]
Aim of AI algorithmsb
Subtyping IDHc20 (90.91)[21-29,32-42]
 Subtyping 1p/19q codeletion6 (27.27)[25,27,30,31,38,39]
Classifier
Convolutional neural networks12 (54.55)[24,26,29,30,33-36,38-40,42]
 Multiple instance learning11 (50)[21-23,25,27,28,31-33,37,42]
 Transformers4 (18.18)[30,33,35,42]
 Hybrid models2 (9.09)[35,42]
 Ensemble model2 (9.09)[34,37]
 Logistic regression2 (9.09)[31,36]
 Random forest1 (4.55)[41]
 Multilayer perceptron1 (4.55)[21]
Specific classifier
AttMILd7 (31.82)[22,23,25,27,28,33,37]
ResNet505 (22.73)[24,35,36,40,42]
Traditional MILe4 (18.18)[21,31,33,42]
InceptionV34 (18.18)[24,33,35,36]
ViTf4 (18.18)[24,33,35,42]
DenseNet121g3 (13.64)[24,34,35]
VGG19h3 (13.64)[24,26,36]
Features extraction methods
Convolutional neural networks20 (90.91)[21-26,28-40,42]
 Transformers4 (18.18)[33,35,37,42]
 Hybrid models4 (18.18)[27,35,37,42]
 MIL1 (4.55)[42]
 Random forest1 (4.55)[41]
Specific features extraction model
ResNet5010 (45.45)[22,24,25,28,31,35-37,40,42]
 DenseNet1215 (22.73)[21,22,24,34,35]
 InceptionV34 (18.18)[24,33,36,37]
 ResNet183 (13.64)[24,29,32]
 VGG192 (9.09)[24,26,36]
 ViT2 (9.09)[33,42]
 Not reported2 (9.09)[23,30]
Data modality
 Unimodel data17 (77.3)[22-35,39,40,42]
 Multimodel data5 (22.7)[21,36-38,41]
Types of validation
k-fold cross-validation11 (50)[21,23,25,27,28,31,33,37-39,41]
 Training-test split10 (45.45)[22,24,26,29,30,32,34-36,40]
 External validation7 (31.82)[25,27,29,31,34,40,42]
Performance measures
Accuracy19 (86.36)[21-25,27,28,31-42]
 AUCi15 (68.18)[21-23,26,28,29,31,33-36,39-42]
 Sensitivity15 (68.18)[23-25,27,28,30-32,34-37,39,40,42]
F1-score9 (40.91)[21,24,25,31,32,35,37,40,42]
 Precision8 (36.36)[23,24,27,30-32,35,37]
 Specificity8 (36.36)[25,27,28,34,36,39,40,42]
 Others8 (36.36)[21,22,25,27-29,37,39]

aValues in italic indicate highest score.

bStudies reported results for both IDH mutation status and 1p/19q codeletion status [25,27,38,39].

cIDH: isocitrate dehydrogenase.

dAttMIL: attention-based multiple instance learning.

eMIL: multiple instance learning.

fViT: vision transformer.

gDenseNet121: densely connected convolutional network 121.

hVGG: visual geometry group.

iAUC: area under the curve.

Results of the Risk of Bias and Applicability

In the selection of participants domain, nearly a quarter of the included studies (5/22, 23%) provided sufficient information to confirm the use of an appropriate consecutive or random sampling of eligible participants. Nearly all studies (21/22, 95%) reported a sufficient sample size. Additionally, 8 (36%) studies avoided inappropriate exclusions, while over half (12/22, 55%) demonstrated balanced subgroup representation. Consequently, a small proportion of the studies (2/22, 9%) were assessed as having a low risk of bias, while a larger portion (8/22, 36%) were rated as having an unclear risk of bias in the “selection of participants” domain, as shown in Figure 2.

In terms of matching participants to the predefined requirements of the review question, a low level of concern was identified in 95% (21/22) of the included studies, as shown in Figure 3.

Figure 2. Results of the assessment of risk of bias in the included studies.
Figure 3. Results of the assessment of applicability concerns in the included studies.

A substantial majority of the included studies comprehensively detailed their AI models, with 21 of the 22 (95%) studies providing thorough descriptions. An overwhelming majority, all 22 studies (100%) clearly reported the features (predictors) used. Moreover, all 22 studies (100%) ensured that these features were sourced without prior knowledge of the outcome data. Consistency in feature assessment across participants was observed in 21 of the 22 (95%) studies. Consequently, the potential for bias in the “index test” domain was assessed as low in 20 of the 22 (91%) studies, as shown in Figure 2. In addition, all 22 (100%) studies were found to have minimal concerns regarding the alignment between the model’s predictors and the review question’s criteria, as illustrated in Figure 3.

In most of the included studies (20/22, 91%), the outcome of interest, histopathology image analysis, was conducted by an expert (eg, pathologist or neuropathologist), ensuring the necessary expertise to accurately classify the outcomes. Nearly all studies (22/22, 100%) applied consistent and uniform outcome definitions across all participants, with outcome classification (ie, image labeling) blinded to the AI model predictions. An overwhelming majority (22/22, 100%) determined the outcome without prior knowledge of the predictor information. Furthermore, in all 22 (100%) studies, the diagnostic process was carried out over an appropriate duration, using the same diagnostic criteria for all participants. As a result, the risk of bias in the “reference standard” domain was considered low in the majority of studies (20/22, 91%), as shown in Figure 2. Additionally, a similar proportion of studies (21/22, 95%) exhibited minimal concerns regarding discrepancies in outcome definition, timing, and determination methods, as shown in Figure 3.

Finally, a significant majority of the studies (22/22, 100%) ensured the inclusion of all enrolled participants in the data analysis. A substantial number of these studies (21/22, 95%) reported no missing values, or if missing values were present, they were handled appropriately (eg, through multiple imputation).

Similarly, a high proportion of studies (22/22, 100%) adopted suitable measures to evaluate the performance of their models. The confusion matrix was presented, or more than one evaluation measure was used, and the selected measures were deemed appropriate.

Nearly half of the studies (21/22, 95%) demonstrated an appropriate split among training, validation, and test sets. The chosen distribution aligned with best practices in the field (eg, 70%‐80% for training, 10%‐15% for validation, and 10%‐20% for testing).

Consequently, the conduct and interpretation of the analysis did not introduce bias in most studies, with 21/22 (95%) having a low risk of bias in the “analysis” domain, as shown in Figure 2. A detailed breakdown of the “risk of bias” and “applicability concerns” for each domain in every study is available in Multimedia Appendix 7.

Results of the Included Studies

As shown in Table 4, the accuracy of AI models in detecting or subtyping molecular markers in adult-type gliomas ranged from 58% to 100%, with a mean accuracy of 85.46%. Sensitivity ranged from 62% to 100%, averaging 84.55%, while specificity ranged from 46% to 100%, with a mean of 86.03%. The AUC values ranged from 46% to 99%, with an average of 86.53%. The hybrid model demonstrated the highest accuracy (92.80%) and sensitivity (89.62%), while the CNN model achieved the highest specificity (89.30%). Logistic regression outperformed other models in terms of AUC (93.05%). Among specific CNN architectures, DenseNet121 achieved the highest accuracy (90.82%), sensitivity (89.28%), and AUC (91.05%). Moreover, AI models that used multimodal data outperformed those that used unimodal data in terms of sensitivity (90.15% vs 84.31%) and AUC (88.93% vs 86.29%). In outcome prediction, AI models showed better performance in detecting IDH mutations compared to 1p/19q codeletions, with higher accuracy (86.13% vs 81.63%), specificity (86.61% vs 78.11%), and AUC (86.74% vs 85.15%). AI models also performed better in IDH subtyping for multiclass classification compared to binary classification in terms of accuracy (91.98% vs 84.02%) and sensitivity (93.41% vs 80.18%). Conversely, AI models for binary classification yielded a lower performance than multiclass classification in terms of accuracy (91.98% vs 84.02%) and sensitivity (93.41% vs 80.18%). However, these differences should be interpreted as descriptive trends rather than statistically validated superiority, as formal between-group comparisons were not feasible.

Table 4. Summary of overall performance of all models based on number of classes, type of data, and artificial intelligence (AI) algorithms in the target variablea.
GroupsAccuracySensitivitySpecificityAUCb
Number of studiesMean % (SD)RangeNumber of studiesMean % (SD)RangeNumber of studiesMean % (SD)RangeNumber of studiesMean % (SD)Range
Algorithms
CNNsc885.60 (0.09)0.72‐1.00884.09 (0.11)0.64‐1.00789.30d (0.09)0.75‐1.00886.54 (0.06)0.76‐0.99
MILe1080.83 (0.09)0.58‐0.91882.54 (0.13)0.62‐0.99368.69 (0.23)0.46‐0.941081.73 (0.16)0.46‐0.97
Transformer488.27 (0.03)0.83‐0.92381.62 (0.10)0.69‐0.92392.37 (0.04)0.89‐0.96
Hybrid model392.80 (0.05)0.85‐0.97389.62 (0.08)0.81‐0.97292.20 (0.04)0.89‐0.95
Ensemble model286.35 (0.12)0.78‐0.95289.60 (0.05)0.86‐0.93292.00 (0.09)0.85‐0.99
Logistic regression287.15 (0.01)0.86‐0.88286.00 (0.12)0.78‐0.94293.05 (0.00)0.93‐0.931
CNN models
ResNet (ResNet,18,50)787.04 (0.09)0.72‐1.00687.51 (0.11)0.69‐1.00591.48 (0.07)0.84‐1.00690.19 (0.06)0.81‐0.99
VGGf (VGG16,19)489.67 (0.09)0.79‐1.00487.03 (0.13)0.69‐1.00290.65 (0.13)0.81‐1.00281.55 (0.01)0.81‐0.82
DenseNet121g390.82 (0.10)0.79‐0.97389.28 (0.10)0.78‐0.97291.05 (0.08)0.86‐0.97
InceptionV3380.15 (0.05)0.77‐0.86375.82 (0.11)0.64‐0.86—-383.58 (0.03)0.80‐0.87
Type of data
Unimodel data2085.67 (0.09)0.58‐1.001684.31 (0.11)0.62‐1.00986.27 (0.15)0.46‐1.001686.29 (0.10)0.46‐0.99
Multimodel data583.64 (0.05)0.78‐0.90290.15 (0.06)0.86‐0.94488.93 (0.03)0.85‐0.93
Outcome prediction
1p/19q681.63 (0.12)0.58‐1.00687.64 (0.09)0.75‐1.00378.11 (0.28)0.46‐1.00585.15 (0.19)0.46‐0.99
IDHh1986.13 (0.08)0.72‐1.001483.96 (0.11)0.62‐1.00886.61 (0.12)0.52‐1.001686.74 (0.08)0.52‐0.99
Class prediction
Binary IDH1684.02 (0.07)0.72‐1.001280.18 (0.11)0.62‐1.00886.61 (0.12)0.52‐1.001586.87 (0.08)0.52‐0.99
Multiclass IDH391.98 (0.07)0.75‐1.00293.41 (0.05)0.85‐1.00
Overall2085.46 (0.08)0.58‐1.001684.55 (0.01)0.62‐1.00986.03 (0.15)0.46‐1.001786.53 (0.10)0.46‐0.99

aSpecificity not reported in all studies.

bAUC: area under the curve.

cCNNs: convolutional neural networks.

dValues in italics indicate highest score.

eMIL: multiple instance learning.

fVGG: visual geometry group.

gDenseNet121: densely connected convolutional network 121.

hIDH: isocitrate dehydrogenase.


Principal Findings

This review aimed to systematically evaluate the performance of AI models in detecting and classifying IDH mutation status and 1p/19q codeletion in adult-type gliomas using histopathological images. Our analysis indicates that these models demonstrate promising diagnostic capabilities, achieving pooled mean values of 85.46% for accuracy, 84.55% for sensitivity, 86.03% for specificity, and 86.53% for the area under the receiver operating characteristic curve. As noticed, the performance across all metrics was comparable, suggesting that AI models offer balanced diagnostic capabilities. Despite these promising results, they remain below the commonly cited thresholds for clinical adoption, which often exceed 90% for high-stakes diagnostic tasks [42]. As such, the current evidence is not sufficient to support stand-alone clinical decision-making. Therefore, the results of this meta-analysis should be interpreted as early indicators of feasibility rather than clinically actionable performance, underscoring the substantial methodological and validation work still required before these models can be safely integrated into clinical workflows.

These results are generally consistent with prior reviews. Specifically, a systematic review conducted by Farahani et al [43] reported a similar performance in IDH prediction using MRI, with pooled sensitivity of 84% and specificity of 87% and an AUC of 0.89. Further, Chen et al [44] found pooled sensitivity of 83%, specificity of 84%, and AUC of 0.90 for predicting IDH mutations using machine learning–based radiomics. Taken together, these comparisons suggest that our findings are in line with the broader body of research, reinforcing the reliability and clinical promise of AI-based histopathological approaches for glioma molecular classification.

Individual studies showed considerable variability across performance metrics; for instance, specificity values ranged from 46% to 100%. This variability likely stems from population differences, sample sizes, model architectures, and the inherent challenge of predicting rare molecular alterations (eg, telomerase reverse transcriptase mutations) from histopathology images, compounded by class imbalance, domain shift (ie, performance degradation when models are applied to data from different sources), and the absence of distinct histological correlates. While accuracy and sensitivity remain higher for well-characterized targets (eg, IDH, 1p/19q), specificity suffers when the model encounters rare or ambiguous cases in external validation.

Hybrid models achieved the highest mean accuracy (92.80%) and sensitivity (89.62%), while logistic regression models achieved the top mean AUC just slightly higher than hybrid models (93.05% vs 92.20%). These results suggest that hybrid models deliver the strongest overall performance. Hybrid models are defined as AI frameworks that integrate multiple deep learning architectures to leverage the complementary strengths of each component. In the context of glioma diagnosis, most recent reviews have highlighted the effectiveness of hybrid models combining CNNs with transformer-based architectures [45-47]. CNNs excel at extracting local spatial features from histopathological images, such as cellular morphology and texture patterns, while transformers are adept at capturing long-range dependencies and global contextual information through their self-attention mechanisms. By combining these capabilities, hybrid CNN-transformer models can achieve a more comprehensive understanding of complex tissue structures and molecular characteristics within gliomas, leading to improved predictive performance. Our findings align with a recent review by Redlich et al [45], which analyzed 83 studies on AI-based methods for whole-slide histopathology image analysis of gliomas. The review highlighted that hybrid models combining CNNs with other architectures, including transformers, consistently outperform single-architecture models in tasks such as subtyping, grading, and molecular marker prediction [45]

Among pure CNN backbones, DenseNet121 demonstrated the best performance, achieving an accuracy of 90.82%, sensitivity of 89.28%, and an AUC of 91.05%. DenseNet121’s superior performance may be due to its unique architectural features, notably dense connectivity, which enhance training dynamics, feature representation, and parameter efficiency.

In terms of data modalities employed by AI models, our analysis demonstrated that multimodal approaches exhibited superior performance in sensitivity (90.15% vs 84.31%) and AUC (88.93% vs 86.29%) compared to unimodal models. This suggests that multimodal models benefit from richer and more diverse feature representations, thereby enhancing their diagnostic capability. However, unimodal models achieved slightly higher accuracy (85.67% vs 83.64%), which may be attributed to the fact that none of the five multimodal studies included in the accuracy analysis employed hybrid modeling. Instead, they utilized simple model ensembles or comparisons involving architectures such as multilayer perceptron, logistic regression, random forests, and CNNs. Our findings are consistent with existing literature. For instance, a systematic review by Alleman et al [48] concluded that multimodal deep learning approaches enhance prognostic accuracy in glioma patients compared to unimodal models, particularly in predicting overall survival. Similarly, d’Este et al [49] demonstrated that combining multimodal imaging with AI techniques improves visualization of glioma infiltration, surpassing the capabilities of single-modality approaches. Furthermore, the GlioMT model, introduced by Byeon et al [50], which integrates imaging and clinical data, exhibited superior performance in predicting adult-type diffuse gliomas compared to conventional CNNs and visual transformers.

Regarding outcome prediction and AI model performance in detecting molecular markers, our analysis revealed that AI models demonstrated stronger performance in identifying IDH mutations compared to 1p/19q codeletions. Specifically, IDH detection yielded higher accuracy (86.13% vs 81.63%), specificity (86.61% vs 78.11%), and AUC (86.74% vs 85.15%). In contrast, AI models achieved slightly better sensitivity for 1p/19q codeletion detection (87.64% vs 83.96%). The discrepancy in performance between IDH mutation and 1p/19q codeletion detection may be attributed to several factors. IDH mutations often result in more distinct histopathological features, making them more amenable to detection by AI models. In contrast, 1p/19q codeletions may not produce as pronounced morphological changes, posing a greater challenge for AI-based detection. The findings of published studies align with our results. For example, the review by Farahani et al [43] showed higher performance in identifying IDH mutations compared to 1p/19q codeletions using MRI in terms of sensitivity (84% vs 76%) and specificity (87% vs 85%).

In our systematic review, CNNs and MIL algorithms were the most predominantly used in AI models, likely due to their suitability for histopathological image analysis. CNNs are designed to capture spatial hierarchies in image data through convolutional layers, making them effective at identifying morphological patterns such as cellular structures and tissue architecture that are critical in tumor classification. Histopathological slides of gliomas are typically of high resolution and contain complex visual features, which CNNs can learn to recognize with high accuracy. Meanwhile, MIL frameworks address the challenge of weakly labeled data, which is common in medical imaging, where only slide-level labels (eg, tumor type or mutation status) are available, rather than precise annotations at the pixel or region level. MIL allows models to learn from sets of image patches (instances) while associating predictions with the whole slide (bag), making it particularly useful in computational pathology where exhaustive labeling is impractical.

Practical and Research Implications

While our results are encouraging, they remain not optimal and arguably lower than that acceptable in clinical practice [42]. Thus, our findings should be interpreted with caution due to several limitations. First, 15 of the 22 studies included in the review utilized the same dataset, raising concerns regarding the generalizability of the results. Second, many of the studies were based on small sample sizes, limiting the robustness and reliability of the conclusions. Third, several key findings were derived from fewer than 5 studies, reducing the strength of the evidence. Lastly, the aggregation of results using simple arithmetic means rather than formal meta-analytic techniques may have limited the ability to draw definitive conclusions about the overall efficacy of AI models.

Given these limitations, AI should not be used as a stand-alone diagnostic tool for the detection or subtyping of adult-type gliomas but rather as a complementary approach alongside conventional methods. Future research should prioritize the development of AI models using larger and more diverse datasets to enhance performance and generalizability. Ensuring adequate sample sizes and improving the transparency of reporting, such as including confusion matrices, will also facilitate more rigorous meta-analyses. Notably, the majority of studies reviewed developed AI models exclusively from histopathological images. Developing multimodal AI models that incorporate additional data types (eg, radiological and clinical information) may further enhance diagnostic accuracy. Thus, continued efforts are needed to advance the integration of multimodal data in AI model development for the detection and subtyping of adult-type gliomas.

In this review, the majority of studies used AI algorithms that were either MIL or CNNs, indicating a research gap in exploring the performance and effectiveness of other types of AI algorithms (eg, transformers, hybrid models, ensemble models [ie, approaches that combine predictions from multiple independently trained models using averaging, voting, or weighted fusion strategies], or logistic regression). Additionally, we observed a lack of research assessing the performance of AI models that use multimodal data, which may have affected the accuracy of the results. Moreover, the number of studies evaluating AI models’ performance in detecting IDH mutations was noticeably greater than those that used 1p/19q as a predicted outcome. These studies predominantly employed binary IDH classification (ie, IDH mutation vs IDH wild-type), whereas studies using IDH multiclass classification (eg, astrocytoma, oligodendroglioma) were markedly fewer. Future research should address these gaps to produce more accurate and representative findings.

Our findings that AI hybrid models performed better in detecting adult-type gliomas than other models underscore the need for more advanced approaches to explore their effectiveness and their potential to achieve diagnostic outcomes comparable to those of histopathologists. However, their real-world applicability requires consideration of computational and operational demands. These models often require greater processing power (eg, GPU-enabled servers or high-performance cloud computing) and longer inference times compared to conventional CNNs. Such requirements may pose challenges for pathology laboratories lacking advanced digital infrastructure. Nonetheless, many modern clinical centers are increasingly equipped with GPU-accelerated digital pathology ecosystems, and cloud-based deployment can significantly reduce the need for on-site hardware. From a cost-benefit perspective, the additional computational resources required by hybrid models may be justified given their superior accuracy and robustness, particularly for molecular prediction tasks where diagnostic errors carry serious clinical implications. However, successful implementation in routine practice will depend on streamlined model optimization, efficient inference pipelines, and prospective validation of runtime performance to ensure that diagnostic gains meaningfully outweigh operational costs.

As our findings indicated superior performance of multimodal AI models in sensitivity and AUC, we suggest a complementary approach: multimodal AI models may serve as effective initial diagnostic tools due to their comprehensive data integration, while unimodal models can be employed subsequently for more accurate and specific outcome determination. Moreover, our results reflect a stronger performance of AI models in the detection of IDH mutations compared to 1p/19q codeletion in all metrics except sensitivity. Therefore, we suggest that while IDH mutation detection should remain a primary focus for model refinement and deployment due to its balanced high performance, models targeting 1p/19q codeletions may serve as effective initial diagnostic tools in adult-type gliomas due to their higher sensitivity, which is critical in early disease detection.

Given that the current review focused on IDH mutations and 1p/19q codeletions, we encourage researchers to undertake further reviews covering other glioma molecular markers, including MGMT methylation, ATRX, and telomerase reverse transcriptase mutations. Moreover, further research should extend to other types of brain tumors, such as meningiomas, pituitary adenomas, and schwannomas. These tumors differ significantly in terms of biological behavior, treatment options, and clinical outcomes, warranting a comprehensive evaluation of AI’s potential in diagnosing and subtyping each tumor type.

Limitations

This systematic review is subject to several limitations. First, the generalizability of the findings is limited, as approximately 68.18% of the included studies relied on the same dataset, and the number of included studies is relatively small (n=22). This raises concerns about the external validity of the results and their applicability to the broader population of adult-type glioma patients. To address the potential duplication of patients or slides across studies (ie, those relying on The Cancer Genome Atlas), we reviewed the data sources reported by each study to identify where overlap was likely. Because individual-level data were not available, we were unable to verify or remove duplicate cases. Instead, we treated studies drawing from the same dataset as nonindependent samples and interpreted pooled estimates with caution. As a result, our summary metrics should be viewed as descriptive indicators rather than independent, unbiased effect estimates. Second, the review focused exclusively on adult-type gliomas, thereby restricting the ability to draw conclusions about the performance of AI in detecting other adult brain tumors or brain tumors in the general population. Third, the analysis was confined to studies utilizing histopathological images, limiting the assessment of AI performance with alternative imaging modalities, such as MRI or computed tomography scans. Fourth, meta-analysis was not feasible due to incomplete reporting of key performance metrics across studies. Instead, mean values were calculated to provide a descriptive summary, though this approach lacks the statistical rigor and precision of formal meta-analysis. Finally, the inclusion of only English-language studies may have introduced language bias and led to the exclusion of relevant research published in other languages, potentially affecting the comprehensiveness of the review.

Conclusion

This systematic review demonstrates that AI models applied to histopathological images show strong potential for the molecular classification of adult-type gliomas, particularly in detecting IDH mutations and 1p/19q codeletions. Our findings indicate that while overall diagnostic performance is promising—especially for IDH mutations—variability in model performance and methodological limitations across studies temper the generalizability of results. Hybrid models and multimodal approaches emerged as particularly effective strategies, offering enhanced sensitivity and discriminative capability. However, the limited use of diverse datasets, the predominance of unimodal designs, and the lack of standardized reporting remain critical barriers to clinical translation. Given these constraints, AI models should currently serve as complementary tools rather than stand-alone diagnostic systems, functioning primarily to assist pathologists in pattern recognition and decision support while final diagnoses remain clinician-led. Future research should prioritize methodological standardization, broader data integration, and validation across larger, heterogeneous populations to support the clinical deployment of AI in glioma diagnostics.

Acknowledgments

The authors declare the use of generative AI in the research and writing process. According to the GAIDeT taxonomy (2025), the following tasks were delegated to generative artificial intelligence (GAI) tools under full human supervision: Proofreading and editing. The GAI tool used was ChatGPT-4. Responsibility for the final manuscript lies entirely with the authors. GAI tools are not listed as authors and do not bear responsibility for the final outcomes. All intellectual content, study design, data collection, analysis, and final interpretations are the sole responsibility of the authors.

Funding

The publication of this article was funded by the Weill Cornell Medicine–Qatar Health Sciences Library.

Data Availability

The datasets generated during and/or analyzed during the current study are available from the corresponding author upon reasonable request.

Authors' Contributions

Concept for the paper and experimental design: AA, RAD, and OA

Data collection and extraction: AA, RAD, and OA

Data analysis: AA and OA

Risk of bias assessment: AT

Writing original draft preparation: AO, RAD, and OA

Writing-review and editing: AA, AO, RAD, and OA

Supervision: AA

All authors agreed to submission of the manuscript.

Conflicts of Interest

AA is an associate editor of JMIR Nursing at the time of publication. All other authors declare no conflict of interest.

Multimedia Appendix 1

Search strategy.

DOCX File, 30 KB

Multimedia Appendix 2

Data extraction form.

DOCX File, 23 KB

Multimedia Appendix 3

Modified version of QUADAS-2.

DOCX File, 28 KB

Multimedia Appendix 4

Characteristics of each included study.

DOCX File, 43 KB

Multimedia Appendix 5

Features of histopathological images in each study.

DOCX File, 39 KB

Multimedia Appendix 6

Features of artificial intelligence in each study.

DOCX File, 44 KB

Multimedia Appendix 7

Reviewers’ judgments about each domain in “risk of bias” and "applicability concerns" for each included study.

DOCX File, 218 KB

Checklist 1

PRISMA-DTA checklist.

DOC File, 67 KB

  1. Ostrom QT, Cioffi G, Waite K, Kruchko C, Barnholtz-Sloan JS. CBTRUS statistical report: primary brain and other central nervous system tumors diagnosed in the United States in 2014–2018. Neuro Oncol. Oct 5, 2021;23(12 Suppl 2):iii1-iii105. [CrossRef] [Medline]
  2. Reynoso-Noverón N, Mohar-Betancourt A, Ortiz-Rafael J. Epidemiology of brain tumors. In: Principles of Neuro-Oncology. Springer; 2020:15-25. [CrossRef]
  3. Lapointe S, Perry A, Butowski NA. Primary brain tumours in adults. The Lancet. Aug 2018;392(10145):432-446. [CrossRef]
  4. Perry A, Wesseling P. Histologic classification of gliomas. Handb Clin Neurol. 2016;134:71-95. [CrossRef] [Medline]
  5. Chen R, Smith-Cohn M, Cohen AL, Colman H. Glioma subclassifications and their clinical significance. Neurotherapeutics. Apr 2017;14(2):284-297. [CrossRef] [Medline]
  6. Kalidindi N, Or R, Babak S, Mason W. Molecular classification of diffuse gliomas. Can J Neurol Sci. Jul 2020;47(4):464-473. [CrossRef] [Medline]
  7. Antonelli M, Poliani PL. Adult type diffuse gliomas in the new 2021 WHO Classification. Pathologica. Dec 2022;114(6):397-409. [CrossRef] [Medline]
  8. Hollon T, Jiang C, Chowdury A, et al. Artificial-intelligence-based molecular classification of diffuse gliomas using rapid, label-free optical imaging. Nat Med. Apr 2023;29(4):828-832. [CrossRef] [Medline]
  9. Koski E, Murphy J. AI in healthcare. Stud Health Technol Inform. Dec 15, 2021;284:295-299. [CrossRef] [Medline]
  10. van Kempen EJ, Post M, Mannil M, et al. Accuracy of machine learning algorithms for the classification of molecular features of gliomas on MRI: a systematic literature review and meta-analysis. Cancers (Basel). May 26, 2021;13(11):2606. [CrossRef] [Medline]
  11. Farahani S, Hejazi M, Moradizeyveh S, Di Ieva A, Fatemizadeh E, Liu S. Diagnostic accuracy of deep learning models in predicting glioma molecular markers: a systematic review and meta-analysis. Diagnostics (Basel). Mar 21, 2025;15(7):797. [CrossRef] [Medline]
  12. Puustinen S, Vrzáková H, Hyttinen J, et al. Hyperspectral imaging in brain tumor surgery-evidence of machine learning-based performance. World Neurosurg. Jul 2023;175:e614-e635. [CrossRef] [Medline]
  13. Singh J, Sahu S, Mohan T, et al. Current status of DNA methylation profiling in neuro-oncology as a diagnostic support tool: a review. Neuro oncol Pract. Dec 2023;10(6):518-526. [CrossRef] [Medline]
  14. Liu Y, Wu M. Deep learning in precision medicine and focus on glioma. Bioeng Transl Med. Sep 2023;8(5):e10553. [CrossRef]
  15. Zhao J, Huang Y, Song Y, et al. Diagnostic accuracy and potential covariates for machine learning to identify IDH mutations in glioma patients: evidence from a meta-analysis. Eur Radiol. Aug 2020;30(8):4664-4674. [CrossRef] [Medline]
  16. Lv Q, Liu Y, Sun Y, Wu M. Insight into deep learning for glioma IDH medical image analysis: a systematic review. Medicine (Baltimore). Feb 16, 2024;103(7):e37150. [CrossRef] [Medline]
  17. Guha A, Halder S, Shinde SH, et al. How does deep learning/machine learning perform in comparison to radiologists in distinguishing glioblastomas (or grade IV astrocytomas) from primary CNS lymphomas? A meta-analysis and systematic review. Clin Radiol. Jun 2024;79(6):460-472. [CrossRef] [Medline]
  18. McInnes MDF, Moher D, Thombs BD, et al. Preferred reporting items for a systematic review and meta-analysis of diagnostic test accuracy studies: the PRISMA-DTA statement. JAMA. Jan 23, 2018;319(4):388-396. [CrossRef] [Medline]
  19. Whiting PF, Rutjes AWS, Westwood ME, et al. QUADAS-2: a revised tool for the quality assessment of diagnostic accuracy studies. Ann Intern Med. Oct 18, 2011;155(8):529-536. [CrossRef] [Medline]
  20. Wolff RF, Moons KGM, Riley RD, et al. Probast: a tool to assess the risk of bias and applicability of prediction model studies. Ann Intern Med. Jan 1, 2019;170(1):51-58. [CrossRef] [Medline]
  21. Albuquerque T, Fang ML, Wiestler B, et al. Multimodal context-aware detection of glioma biomarkers using MRI and WSI. 2023. Presented at: Medical Image Computing and Computer Assisted Intervention – MICCAI 2023 Workshops; Oct 8-12, 2023:157-167; Vancouver, BC, Canada. [CrossRef]
  22. Chitnis SR, Liu S, Dash T, et al. Domain-specific pre-training improves confidence in whole slide image classification. 2023. Presented at: 2023 45th Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC); Jul 24-27, 2023. [CrossRef]
  23. Cui D, Liu Y, Liu G, Liu L. A multiple-instance learning-based convolutional neural network model to detect the IDH1 mutation in the histopathology images of glioma tissues. J Comput Biol. Aug 2020;27(8):1264-1272. [CrossRef] [Medline]
  24. Despotovic V, Kim SY, Hau AC, et al. Glioma subtype classification from histopathological images using in-domain and out-of-domain transfer learning: an experimental study. Heliyon. Mar 15, 2024;10(5):e27515. [CrossRef] [Medline]
  25. Fang Z, Liu Y, Wang Y, et al. Deep learning predicts biomarker status and discovers related histomorphology characteristics for low-grade glioma. arXiv. Preprint posted online on Oct 11, 2023. [CrossRef]
  26. Faust K, Lee MK, Dent A, et al. Integrating morphologic and molecular histopathological features through whole slide image registration and deep learning. Neurooncol Adv. 2022;4(1):vdac001. [CrossRef] [Medline]
  27. Hewitt KJ, Löffler CML, Muti HS, et al. Direct image to subtype prediction for brain tumors using deep learning. Neurooncol Adv. 2023;5(1):vdad139. [CrossRef] [Medline]
  28. Innani S, Baheti B, Nasrallah MP, Bakas S. Weakly supervised IDH-status glioma classification from H&E-stained whole slide images. 2024. Presented at: 2024 IEEE International Symposium on Biomedical Imaging (ISBI); May 27-30, 2024. [CrossRef]
  29. Jiang S, Zanazzi GJ, Hassanpour S. Predicting prognosis and IDH mutation status for patients with lower-grade gliomas using whole slide images. Sci Rep. Aug 19, 2021;11(1):16849. [CrossRef] [Medline]
  30. Jungo P, Hewer E. Code-free machine learning for classification of central nervous system histopathology images. J Neuropathol Exp Neurol. Feb 21, 2023;82(3):221-230. [CrossRef] [Medline]
  31. Kim GJ, Lee T, Ahn S, Uh Y, Kim SH. Efficient diagnosis of IDH-mutant gliomas: 1p/19qNET assesses 1p/19q codeletion status using weakly-supervised learning. NPJ Precis Onc. 2023;7(1):94. [CrossRef]
  32. Krebs O, Agarwal S, Tiwari P. Self-supervised deep learning to predict molecular markers from routine histopathology slides for high-grade glioma tumors. Presented at: Proceedings Volume 12471, Medical Imaging 2023: Digital and Computational Pathology; Feb 19-23, 2023. [CrossRef]
  33. Li Z, Cong Y, Chen X, et al. Vision transformer-based weakly supervised histopathological image analysis of primary brain tumors. iScience. Jan 2023;26(1):105872. [CrossRef]
  34. Liechty B, Xu Z, Zhang Z, et al. Machine learning can aid in prediction of IDH mutation from H&E-stained histology slides in infiltrating gliomas. Sci Rep. Dec 31, 2022;12(1):22623. [CrossRef] [Medline]
  35. Liu X, Hu W, Diao S, Abera DE, Racoceanu D, Qin W. Multi-scale feature fusion for prediction of IDH1 mutations in glioma histopathological images. Comput Methods Programs Biomed. May 2024;248:108116. [CrossRef] [Medline]
  36. Liu S, Shah Z, Sav A, et al. Isocitrate dehydrogenase (IDH) status prediction in histopathology images of gliomas using deep learning. Sci Rep. 2020;10(1):7733. [CrossRef]
  37. Nakagaki R, Debsarkar SS, Kawanaka H, Aronow BJ, Prasath VBS. Deep learning-based IDH1 gene mutation prediction using histopathological imaging and clinical data. Comput Biol Med. Sep 2024;179:108902. [CrossRef] [Medline]
  38. Pei L, Jones KA, Shboul ZA, Chen JY, Iftekharuddin KM. Deep neural network analysis of pathology images with integrated molecular data for enhanced glioma classification and grading. Front Oncol. 2021;11:668694. [CrossRef] [Medline]
  39. Rathore S, Iftikhar M, Nasrallah M, Gurcan M, Rajpoot N, Mourelatos Z. Tmod-35. Prediction of overall survival, and molecular markers in gliomas via analysis of digital pathology images using deep learning. Neuro Oncol. Nov 11, 2019;21(Supplement_6):vi270. [CrossRef]
  40. Wang W, Zhao Y, Teng L, et al. Neuropathologist-level integrated classification of adult-type diffuse gliomas using deep learning from whole-slide pathological images. Nat Commun. 2023;14(1):6359. [CrossRef]
  41. Wang D, Liu C, Wang X, et al. Automated machine-learning framework integrating histopathological and radiological information for predicting IDH1 mutation status in glioma. Front Bioinform. 2021;1:718697. [CrossRef] [Medline]
  42. Zhao Y, Wang W, Ji Y, et al. Computational pathology for prediction of isocitrate dehydrogenase gene mutation from whole slide images in adult patients with diffuse glioma. Am J Pathol. May 2024;194(5):747-758. [CrossRef] [Medline]
  43. Farahani S, Hejazi M, Tabassum M, Di Ieva A, Mahdavifar N, Liu S. Diagnostic performance of deep learning for predicting glioma isocitrate dehydrogenase and 1p/19q co-deletion in MRI: a systematic review and meta-analysis. Eur Radiol. Aug 16, 2025. [CrossRef] [Medline]
  44. Chen X, Lei J, Wang S, Zhang J, Gou L. Diagnostic accuracy of a machine learning-based radiomics approach of MR in predicting IDH mutations in glioma patients: a systematic review and meta-analysis. Front Oncol. 2024;14:1409760. [CrossRef]
  45. Redlich JP, Feuerhake F, Weis J, et al. Applications of artificial intelligence in the analysis of histopathology images of gliomas: a review. Npj Imaging. Jul 1, 2024;2(1):16. [CrossRef] [Medline]
  46. Nobel SMN, Swapno S, Islam MB, et al. A novel mixed convolution transformer model for the fast and accurate diagnosis of glioma subtypes. Adv Intell Syst. May 2025;7(5):2400566. [CrossRef]
  47. Henry EU, Emebo O, Omonhinmin CA. Vision transformers in medical imaging: a review. arXiv. Preprint posted online on Nov 18, 2022. [CrossRef]
  48. Alleman K, Knecht E, Huang J, Zhang L, Lam S, DeCuypere M. Multimodal deep learning-based prognostication in glioma patients: a systematic review. Cancers (Basel). Jan 16, 2023;15(2):545. [CrossRef] [Medline]
  49. d’Este SH, Nielsen MB, Hansen AE. Visualizing glioma infiltration by the combination of multimodality imaging and artificial intelligence, a systematic review of the literature. Diagnostics (Basel). Mar 25, 2021;11(4):592. [CrossRef] [Medline]
  50. Byeon Y, Park YW, Lee S, et al. Interpretable multimodal transformer for prediction of molecular subtypes and grades in adult-type diffuse gliomas. NPJ Digit Med. Mar 5, 2025;8(1):140. [CrossRef] [Medline]


AI: artificial intelligence
AUC: area under the curve
CNN: convolutional neural network
CNS: central nervous system
DBTA: Digital Brain Tumor Atlas
DenseNet121: densely connected convolutional network 121
IDH: isocitrate dehydrogenase
MIL: multiple instance learning
MRI: magnetic resonance imaging
PRISMA-DTA: Preferred Reporting Items for Systematic Reviews and Meta-Analyses–Extension for Diagnostic Test Accuracy
PROBAST: Prediction Model Risk of Bias Assessment Tool
PROSPERO: International Prospective Register of Systematic Reviews
QUADAS-2: Quality Assessment of Diagnostic Accuracy Studies 2
WHO: World Health Organization


Edited by Amy Schwartz; submitted 01.Jun.2025; peer-reviewed by Babul Salam Ibrahim, Khaldoon Dhou, Yi Lu; accepted 06.Jan.2026; published 13.Mar.2026.

Copyright

© Obada Almaabreh, Rukaya Al-Dafi, Aliya Tabassum, Ahmad Othman, Alaa Abd-alrazaq. Originally published in the Journal of Medical Internet Research (https://www.jmir.org), 13.Mar.2026.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research (ISSN 1438-8871), is properly cited. The complete bibliographic information, a link to the original publication on https://www.jmir.org/, as well as this copyright and license information must be included.