<?xml version="1.0" encoding="UTF-8"?><!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "journalpublishing.dtd"><article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" dtd-version="2.0" xml:lang="en" article-type="research-article"><front><journal-meta><journal-id journal-id-type="nlm-ta">J Med Internet Res</journal-id><journal-id journal-id-type="publisher-id">jmir</journal-id><journal-id journal-id-type="index">1</journal-id><journal-title>Journal of Medical Internet Research</journal-title><abbrev-journal-title>J Med Internet Res</abbrev-journal-title><issn pub-type="epub">1438-8871</issn><publisher><publisher-name>JMIR Publications</publisher-name><publisher-loc>Toronto, Canada</publisher-loc></publisher></journal-meta><article-meta><article-id pub-id-type="publisher-id">v27i1e66055</article-id><article-id pub-id-type="doi">10.2196/66055</article-id><article-categories><subj-group subj-group-type="heading"><subject>Original Paper</subject></subj-group></article-categories><title-group><article-title>Knowledge Graph&#x2013;Enhanced Deep Learning Model (H-SYSTEM) for Hypertensive Intracerebral Hemorrhage: Model Development and Validation</article-title></title-group><contrib-group><contrib contrib-type="author"><name name-style="western"><surname>Xia</surname><given-names>Yulong</given-names></name><degrees>MBBS</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Li</surname><given-names>Jie</given-names></name><degrees>PhD</degrees><xref ref-type="aff" rid="aff2">2</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Deng</surname><given-names>Bo</given-names></name><degrees>MBBS</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Huang</surname><given-names>Qilin</given-names></name><degrees>MAS</degrees><xref ref-type="aff" rid="aff2">2</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Cai</surname><given-names>Fenglin</given-names></name><degrees>MAS</degrees><xref ref-type="aff" rid="aff2">2</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Xie</surname><given-names>Yanfeng</given-names></name><degrees>MM</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Sun</surname><given-names>Xiaochuan</given-names></name><degrees>MD</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Shi</surname><given-names>Quanhong</given-names></name><degrees>MD</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Dan</surname><given-names>Wei</given-names></name><degrees>MD</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Zhan</surname><given-names>Yan</given-names></name><degrees>MD</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author" corresp="yes"><name name-style="western"><surname>Jiang</surname><given-names>Li</given-names></name><degrees>MD</degrees><xref ref-type="aff" rid="aff1">1</xref></contrib></contrib-group><aff id="aff1"><institution>Department of Neurosurgery, The First Affiliated Hospital of Chongqing Medical University</institution><addr-line>No. 1 Youyi Road, Yuanjiagang, Yuzhong District</addr-line><addr-line>Chongqing</addr-line><country>China</country></aff><aff id="aff2"><institution>School of Computer Science and Engineering, Chongqing University of Science and Technology</institution><addr-line>Chongqing</addr-line><country>China</country></aff><contrib-group><contrib contrib-type="editor"><name name-style="western"><surname>Sarvestan</surname><given-names>Javad</given-names></name></contrib></contrib-group><contrib-group><contrib contrib-type="reviewer"><name name-style="western"><surname>Shanthini</surname><given-names>J</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Sun</surname><given-names>Xuxue</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Hou</surname><given-names>Zhen</given-names></name></contrib></contrib-group><author-notes><corresp>Correspondence to Li Jiang, MD, Department of Neurosurgery, The First Affiliated Hospital of Chongqing Medical University, No. 1 Youyi Road, Yuanjiagang, Yuzhong District, Chongqing, China, 86 13638354200; <email>jiangli01010721@163.com</email></corresp></author-notes><pub-date pub-type="collection"><year>2025</year></pub-date><pub-date pub-type="epub"><day>12</day><month>6</month><year>2025</year></pub-date><volume>27</volume><elocation-id>e66055</elocation-id><history><date date-type="received"><day>12</day><month>09</month><year>2024</year></date><date date-type="rev-recd"><day>28</day><month>04</month><year>2025</year></date><date date-type="accepted"><day>28</day><month>04</month><year>2025</year></date></history><copyright-statement>&#x00A9; Yulong Xia, Jie Li, Bo Deng, Qilin Huang, Fenglin Cai, Yanfeng Xie, Xiaochuan Sun, Quanhong Shi, Wei Dan, Yan Zhan, Li Jiang. Originally published in the Journal of Medical Internet Research (<ext-link ext-link-type="uri" xlink:href="https://www.jmir.org">https://www.jmir.org</ext-link>), 12.6.2025. </copyright-statement><copyright-year>2025</copyright-year><license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (<ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</ext-link>), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research (ISSN 1438-8871), is properly cited. The complete bibliographic information, a link to the original publication on <ext-link ext-link-type="uri" xlink:href="https://www.jmir.org/">https://www.jmir.org/</ext-link>, as well as this copyright and license information must be included.</p></license><self-uri xlink:type="simple" xlink:href="https://www.jmir.org/2025/1/e66055"/><abstract><sec><title>Background</title><p>Although much progress has been made in artificial intelligence (AI), several challenges remain substantial obstacles to the development and translation of AI systems into clinical practice. Even large language models, which show excellent performance on various tasks, have progressed slowly in clinical practice tasks. Providing precise and explainable treatment plans with personalized details remains a big challenge for AI systems due to both the highly specialized medical knowledge required and patients&#x2019; complicated conditions.</p></sec><sec><title>Objective</title><p>This study aimed to develop an explainable and efficient decision support system named H-SYSTEM to assist neurosurgeons in diagnosing and treating patients with hypertensive intracerebral hemorrhage. The system was designed to address the limitations of existing AI systems by integrating a medical domain knowledge graph to enhance decision-making accuracy and explainability.</p></sec><sec sec-type="methods"><title>Methods</title><p>The H-SYSTEM consists of 3 main modules: the key named entity recognition (NER) module, the semantic analysis and representation module, and the reasoning module. Furthermore, we constructed a medical domain knowledge graph for hypertensive intracerebral hemorrhage, named HKG, which served as an external knowledge brain of the H-SYSTEM to enhance its text recognition and automated decision-making capability. The HKG was exploited to guide the training of the semantic analysis and representation module and reasoning module, which makes the output of the H-SYSTEM more explainable.</p><p>To assess the performance of the H-SYSTEM, we compared it with doctors and different large language models.</p></sec><sec sec-type="results"><title>Results</title><p>The outputs based on HKG showed reliable performance as compared with neurosurgical doctors, with an overall accuracy of 94.87%. The bidirectional encoder representations from transformers, inflated dilated convolutional neural network, bidirectional long short-term memory, and conditional random fields (BERT-IDCNN-BiLSTM-CRF) model was used as the key NER module of the H-SYSTEM due to its fast convergence and efficient extraction of key named entities, achieved the highest performance among 7 key NER models (precision=92.03, recall=90.22, and <italic>F</italic><sub>1</sub>-score=91.11), significantly outperforming the others. The H-SYSTEM achieved an overall accuracy of 91.74% in treatment plans, showing significant consistency with the gold standard (<italic>P</italic>&#x003C;.05), with diagnostic measures achieving 88.18% accuracy, 97.03% area under the curve (AUC), and a &#x03BA; of 0.874; surgical therapy achieving 98.53% accuracy, 98.53% AUC, and a &#x03BA; of 0.971; and rescue therapies achieving 89.50% accuracy, 94.67% AUC, and a &#x03BA; of 0.923 (all <italic>P</italic>&#x003C;.05). Furthermore, the H-SYSTEM showed high reliability and efficiency when compared to doctors and ChatGPT, achieving statistically higher accuracy (95.26% vs 91.48%, <italic>P</italic>&#x003C;.05). Additionally, the H-SYSTEM achieved a total accuracy of 92.22% (ranging from 91.14% to 95.35%) in treatment plans for 605 additional patients from 6 different medical centers.</p></sec><sec sec-type="conclusions"><title>Conclusions</title><p>The H-SYSTEM showed significantly high efficiency and generalization capacity in processing electronic medical records, and it provided explainable and elaborate treatment plans. Therefore, it has the potential to provide neurosurgeons with rapid and reliable decision support, especially in emergency conditions. The knowledge graph&#x2013;enhanced deep-learning model exhibited excellent performance in the clinical practice tasks.</p></sec></abstract><kwd-group><kwd>hypertensive intracerebral hemorrhage</kwd><kwd>artificial intelligence</kwd><kwd>knowledge graph</kwd><kwd>medical domain knowledge graph</kwd><kwd>deep learning&#x2013;based model</kwd><kwd>electronic medical record</kwd><kwd>explainability</kwd><kwd>black box</kwd></kwd-group></article-meta></front><body><sec id="s1" sec-type="intro"><title>Introduction</title><p>Acute spontaneous (nontraumatic) intracerebral hemorrhage (ICH) is one of the leading causes of death and disability worldwide, posing a massive burden on health care systems [<xref ref-type="bibr" rid="ref1">1</xref>-<xref ref-type="bibr" rid="ref3">3</xref>]. Hypertensive intracerebral hemorrhage (HICH) is one of the most serious types of ICH and affects approximately 2 million people worldwide each year [<xref ref-type="bibr" rid="ref4">4</xref>-<xref ref-type="bibr" rid="ref7">7</xref>]. Timely diagnosis and treatment are of critical clinical importance for patients with HICH, as nearly half of the resulting mortality occurs within the first 24 hours [<xref ref-type="bibr" rid="ref8">8</xref>-<xref ref-type="bibr" rid="ref10">10</xref>]. Thus, the development of methods that provide rapid and accurate diagnosis and treatment could significantly improve the prognosis of patients with HICH.</p><p>Artificial intelligence (AI) has lately made significant advances in the interpretation of sensory information, allowing machines to automatically analyze and comprehend complicated facts [<xref ref-type="bibr" rid="ref11">11</xref>-<xref ref-type="bibr" rid="ref16">16</xref>]. Although AI models like large language models (LLMs) have shown excellent performance in some general medical question&#x2013;answering tasks, they do not perform so well in the clinical practice tasks [<xref ref-type="bibr" rid="ref17">17</xref>-<xref ref-type="bibr" rid="ref22">22</xref>]. Furthermore, the need for large, well-annotated data and the black box problem remain the two main challenges to the development and translation of medical AI systems into clinical practice [<xref ref-type="bibr" rid="ref23">23</xref>]. Developing precise treatment plans with individualized details for patients with HICH still remains a challenge for AI systems due to not only the highly abstracted and professional medical knowledge and terminology, but also the complexity of the patient&#x2019;s condition.</p><p>Aiming to address the above challenges and provide substantial supports for neurosurgeons in the clinical diagnosis and treatment of HICH cases, we developed an automatic decision support system called H-SYSTEM based on multiple-center data. We also enhanced the ability of deep learning models by constructing a specialized knowledge graph that also increases the explainability of the decision-making process.</p></sec><sec id="s2" sec-type="methods"><title>Methods</title><sec id="s2-1"><title>Ethical Considerations</title><p>All study procedures were approved by the ethical committees of all the medical centers mentioned in this study. The patients involved in the study all signed informed consent forms before admission. All data involved in this study have been anonymized or deidentified. The case data involved in this study are part of a retrospective study. All participants received professional medical treatment at the time, and no additional compensation was provided.</p></sec><sec id="s2-2"><title>Study Design and Participants</title><p>Following ethical committees&#x2019; approval, electronic medical record (EMR) data from 15 medical centers were enrolled into the neurosurgery EMR database, HICH-IT, which was posted on GitHub [<xref ref-type="bibr" rid="ref24">24</xref>]. More than 8000 HICH cases were extracted from this database and divided into training, validation, and testing sets, respectively. The inclusion criteria for cases were as follows: (1) patients diagnosed with spontaneous cerebral hemorrhage in compliance with the latest stroke guidelines and (2) patients aged 10-80 years old. Patients with an incomplete medical history or physical examination or without head computed tomography (CT) results were excluded.</p></sec><sec id="s2-3"><title>Development of the H-SYSTEM</title><sec id="s2-3-1"><title>Input and Output of the H-SYSTEM: EMR Acquisition and Preprocessing</title><p>Basic demographic information, chief complaint, history of current and past illnesses, physical examination, head CT results, and other clinical variables that would be available to doctors and AI experts were abstracted from the HICH-IT. The narrative clinical notes retrieved from the EMR database needed to undergo a series of text preprocessing steps before natural language processing techniques could be applied. These steps included using the spell checker to correct misspelled words, extending abbreviations and acronyms, and tokenization. Then, the key named entities were annotated (<xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>). The first category of key named entities included the clinical and CT manifestation, such as the main complaints, vital signs, Glasgow Coma Score, abnormal pupillary sizes and reflexes, location, and nature of the hematoma. The second category included quantitative parameters of the hematoma, such as the volume of the hematoma, displacement of midline brain structures, and compression of the lateral ventricles. After performing these preprocessing steps, we finally attained a clean corpus and labels for data analysis.</p></sec><sec id="s2-3-2"><title>Treatment Plans Output by the H-SYSTEM</title><p>The treatment plans output by the H-SYSTEM were divided into diagnostic and therapeutic measures. To achieve explainability, the H-SYSTEM not only outputs the treatment plans but also provides the basis of outputting the decisions, making the internal working and decision-making process explainable (<xref ref-type="fig" rid="figure1">Figure 1</xref>). The detailed treatment plans output by the H-SYSTEM are displayed in <xref ref-type="supplementary-material" rid="app2">Multimedia Appendix 2</xref>.</p><fig position="float" id="figure1"><label>Figure 1.</label><caption><p>An example of the output results of the H-SYSTEM. CT: computed tomography; CTA: computed tomography angiography; DSA: digital subtraction angiography; GCS: Glasgow Coma Score; HICH: hypertensive intracerebral hemorrhage; MRA: magnetic resonance angiography.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="jmir_v27i1e66055_fig01.png"/></fig></sec><sec id="s2-3-3"><title>Diagnostic Measures</title><p>The purpose of the diagnostic measures was to provide more information and further evidence when the clinical manifestation and plain CT report in the EMR were not enough to make an accurate evaluation and determine proper treatment plans. The diagnostic measures included cerebrovascular examinations (such as computed tomography angiography, magnetic resonance angiography, and digital subtraction angiography), multimodal magnetic resonance imaging (such as plain and contrast-enhanced T1, T2, T2-FLAIR), and coagulation function.</p></sec><sec id="s2-3-4"><title>Therapeutic Measures</title><p>The therapeutic measures were divided into surgical therapy and nonsurgical therapy. The nonsurgical therapy was more complicated and was further divided into rescue therapies and drug therapies. The rescue therapies consisted of monitoring and maintaining vital signs, establishing a venous channel, cardiopulmonary resuscitation, respiratory tract management (including clearing the airway, trachea cannula placement, and the application of a respirator). The drug therapies included medications for antihypertension, decreasing intracranial pressure, and antigastrointestinal bleeding.</p></sec><sec id="s2-3-5"><title>Construction of the HICH Knowledge Graph</title><p>We established the HICH knowledge graph (HKG) based on the Apache Jane triplet database and the RDF structured query language SPARQL. The HKG is able to extract, integrate, and align general medical information and HICH diagnostic and treatment-related data to enhance clarity of structure and improve retrieval performance. It serves as an external knowledge brain to augment text recognition and automated decision-making capabilities.</p><p>The HKG covers 3 dimensions of medical knowledge: general medical knowledge from medical textbooks; medical subdomain knowledge from clinical guidelines and expert consensus on common neurosurgical emergencies; and medical subdomain-specific knowledge from the hypertensive intracerebral hemorrhage weight system (HWS; detailed in Supplemental Methods 4 in <xref ref-type="supplementary-material" rid="app3">Multimedia Appendix 3</xref>).</p><p>The HWS was designed based on the latest guidelines and clinical experience on HICH (<xref ref-type="fig" rid="figure2">Figure 2</xref>) [<xref ref-type="bibr" rid="ref6">6</xref>,<xref ref-type="bibr" rid="ref25">25</xref>-<xref ref-type="bibr" rid="ref30">30</xref>]. The relationships among different variables were analyzed to establish the mapping relation between key information and treatment plans. Then, a total of 2000 HICH cases were randomly selected from the HICH-IT to train, validate, and test the HWS (detailed in Figure S2 in <xref ref-type="supplementary-material" rid="app3">Multimedia Appendix 3</xref>).</p><fig position="float" id="figure2"><label>Figure 2.</label><caption><p>Some examples of HWS details. GCS: Glasgow Coma Score; HWS: hypertensive intracerebral hemorrhage weight system; Pa: partial pressure.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="jmir_v27i1e66055_fig02.png"/></fig></sec><sec id="s2-3-6"><title>Construction of the H-SYSTEM</title><p>The H-SYSTEM comprises 3 main modules (<xref ref-type="fig" rid="figure3">Figure 3</xref>): the key named entity recognition (NER) module, the HKG-enhanced semantic representation and analysis module, and HKG-enhanced reasoning module. These modules were meticulously designed and developed through collaborative work by a team of AI experts and neurosurgeons. A total of 3500 HICH cases were randomly sampled from the HICH-IT to establish the H-SYSTEM (training set=1500; validation set=1000; testing set=1000).</p><fig position="float" id="figure3"><label>Figure 3.</label><caption><p>Construction and workflow of the H-SYSTEM. BERT: bidirectional encoder representations from transformers; BiLSTM: bidirectional long short-term memory; CRF: conditional random fields; CTA: computed tomography angiography; DSA: digital subtraction angiography; EMR: electronic medical system; HKG: hypertensive intracerebral hemorrhage knowledge graph; HWS: hypertensive intracerebral hemorrhage weight system; IDCNN: inflated dilated convolutional neural network; MRA: magnetic resonance angiography; MRI: magnetic resonance imaging.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="jmir_v27i1e66055_fig03.png"/></fig></sec><sec id="s2-3-7"><title>The Key NER Module</title><sec id="s2-3-7-1"><title>Overview</title><p>In this study, 7 key NER models were constructed based on 4 major modules: bidirectional encoder representations from transformers (BERT), inflated dilated convolutional neural network (IDCNN), bidirectional long short-term memory (BiLSTM), and conditional random fields (CRF; the descriptions of these models are detailed in <xref ref-type="supplementary-material" rid="app3">Multimedia Appendix 3</xref>). The models were evaluated in the same experimental setting. Three metrics&#x2014;precision (P), recall (R), and <italic>F</italic><sub>1</sub>-score&#x2014;were used to evaluate these models. After that, the NER model with best performance was chosen as the text recognition core of the H-SYSTEM.</p></sec><sec id="s2-3-7-2"><title>BERT Model</title><p>The network architecture diagram of the BERT model is shown in <xref ref-type="fig" rid="figure3">Figure 3</xref>. The BERT achieves its pretraining objective through a masked language model and self-attention mechanism, extracting word embeddings for named entities as feature representations across the entire sequence. The calculation of this process is presented in <xref ref-type="disp-formula" rid="E1">Equation 1</xref>:</p><disp-formula id="E1"><label>(1)</label><mml:math id="eqn1"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mrow><mml:mi mathvariant="normal">A</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">t</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">t</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">e</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">n</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">t</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">i</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">o</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">n</mml:mi></mml:mrow><mml:mrow><mml:mtext>&#x00A0;</mml:mtext></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:mrow><mml:mi mathvariant="normal">Q</mml:mi></mml:mrow><mml:mo>,</mml:mo><mml:mrow><mml:mi mathvariant="normal">K</mml:mi></mml:mrow><mml:mo>,</mml:mo><mml:mrow><mml:mi mathvariant="normal">V</mml:mi></mml:mrow></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mo>=</mml:mo><mml:mrow><mml:mi mathvariant="normal">s</mml:mi><mml:mi mathvariant="normal">o</mml:mi><mml:mi mathvariant="normal">f</mml:mi><mml:mi mathvariant="normal">t</mml:mi><mml:mi mathvariant="normal">m</mml:mi><mml:mi mathvariant="normal">a</mml:mi><mml:mi mathvariant="normal">x</mml:mi></mml:mrow><mml:mrow><mml:mo>(</mml:mo><mml:mrow><mml:msup><mml:mrow><mml:mi mathvariant="normal">K</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mi mathvariant="normal">T</mml:mi></mml:mrow></mml:mrow></mml:msup><mml:mrow><mml:mi mathvariant="normal">Q</mml:mi></mml:mrow><mml:msup><mml:msqrt><mml:msub><mml:mrow><mml:mi mathvariant="normal">d</mml:mi></mml:mrow><mml:mrow><mml:mrow><mml:mi mathvariant="normal">k</mml:mi></mml:mrow></mml:mrow></mml:msub></mml:msqrt><mml:mrow><mml:mo>&#x2212;</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msup></mml:mrow><mml:mo>)</mml:mo></mml:mrow><mml:mrow><mml:mi mathvariant="normal">V</mml:mi></mml:mrow></mml:mrow></mml:mstyle></mml:math></disp-formula></sec><sec id="s2-3-7-3"><title>IDCNN Model</title><p>IDCNN uses dilated convolutions, which reduce data loss and expand the perceptual range by adjusting the dilation width of the convolution kernel, enabling the capture of longer text. The receptive field calculation formula for dilated convolutions is as follows:</p><disp-formula id="E2"><label>(2)</label><mml:math id="eqn2"><mml:msub><mml:mrow><mml:mi mathvariant="normal">F</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">i</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msup><mml:mrow><mml:mfenced separators="|"><mml:mrow><mml:msup><mml:mrow><mml:mn>2</mml:mn></mml:mrow><mml:mrow><mml:mi mathvariant="normal">i</mml:mi><mml:mo>+</mml:mo><mml:mn>2</mml:mn></mml:mrow></mml:msup><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:mfenced></mml:mrow><mml:mrow><mml:mn>2</mml:mn></mml:mrow></mml:msup></mml:math></disp-formula></sec><sec id="s2-3-7-4"><title>BiLSTM Model</title><p>BiLSTM uses gating mechanisms such as input gates, output gates, and forgetting gates to achieve bidirectional capturing of global contextual features. The update equations for the states of the 3 gates are provided in <xref ref-type="disp-formula" rid="E3 E4 E5 E6 E7 E8">Equations 3&#x2013;8</xref> below.</p><disp-formula id="E3"><label>(3)</label><mml:math id="eqn3"><mml:msub><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>&#x03C3;</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mo>(</mml:mo><mml:msub><mml:mrow><mml:mi>W</mml:mi></mml:mrow><mml:mrow><mml:mi>f</mml:mi></mml:mrow></mml:msub><mml:mi>*</mml:mi><mml:mo>[</mml:mo><mml:msub><mml:mrow><mml:mi>h</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>u</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo>]</mml:mo><mml:mo>+</mml:mo><mml:msub><mml:mrow><mml:mi>b</mml:mi></mml:mrow><mml:mrow><mml:mi>f</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:math></disp-formula><disp-formula id="E4"><label>(4)</label><mml:math id="eqn4"><mml:msub><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>&#x03C3;</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mo>(</mml:mo><mml:msub><mml:mrow><mml:mi>W</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mi>*</mml:mi><mml:mo>[</mml:mo><mml:msub><mml:mrow><mml:mi>h</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>u</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo>]</mml:mo><mml:mo>+</mml:mo><mml:msub><mml:mrow><mml:mi>b</mml:mi></mml:mrow><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:math></disp-formula><disp-formula id="E5"><label>(5)</label><mml:math id="eqn5"><mml:msub><mml:mrow><mml:mi>O</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>&#x03C3;</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mo>(</mml:mo><mml:msub><mml:mrow><mml:mi>W</mml:mi></mml:mrow><mml:mrow><mml:mi>o</mml:mi></mml:mrow></mml:msub><mml:mi>*</mml:mi><mml:mfenced open="[" close="]" separators="|"><mml:mrow><mml:msub><mml:mrow><mml:mi>h</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>u</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mfenced><mml:mo>+</mml:mo><mml:msub><mml:mrow><mml:mi>b</mml:mi></mml:mrow><mml:mrow><mml:mi>o</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:math></disp-formula><disp-formula id="E6"><label>(6)</label><mml:math id="eqn6"><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mo>˜</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>&#x03C3;</mml:mi></mml:mrow><mml:mrow><mml:mi>h</mml:mi></mml:mrow></mml:msub><mml:mo>(</mml:mo><mml:msub><mml:mrow><mml:mi>W</mml:mi></mml:mrow><mml:mrow><mml:mi>c</mml:mi></mml:mrow></mml:msub><mml:mi>*</mml:mi><mml:mo>[</mml:mo><mml:msub><mml:mrow><mml:mi>h</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>u</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo>]</mml:mo><mml:mo>+</mml:mo><mml:msub><mml:mrow><mml:mi>b</mml:mi></mml:mrow><mml:mrow><mml:mi>c</mml:mi></mml:mrow></mml:msub><mml:mo>)</mml:mo></mml:math></disp-formula><disp-formula id="E7"><label>(7)</label><mml:math id="eqn7"><mml:msub><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>*</mml:mi><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>+</mml:mo><mml:msub><mml:mrow><mml:mi>i</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mi>*</mml:mi><mml:msub><mml:mrow><mml:mover accent="true"><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mo>˜</mml:mo></mml:mover></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub></mml:math></disp-formula><disp-formula id="E8"><label>(8)</label><mml:math id="eqn8"><mml:msub><mml:mrow><mml:mi>h</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mi>O</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mi>*</mml:mi><mml:msub><mml:mrow><mml:mi>&#x03C3;</mml:mi></mml:mrow><mml:mrow><mml:mi>g</mml:mi></mml:mrow></mml:msub><mml:mfenced separators="|"><mml:mrow><mml:msub><mml:mrow><mml:mi>C</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub></mml:mrow></mml:mfenced></mml:math></disp-formula></sec><sec id="s2-3-7-5"><title>CRF Model</title><p>Rather than treating each tag independently, the modeling of tag sequences is achieved through the utilization of CRF, enabling the joint modeling of tag sequences for a given input word sequence. This distribution is formulated as follows:</p><disp-formula id="E9"><label>(9)</label><mml:math id="eqn9"><mml:mi>P</mml:mi><mml:mo>(</mml:mo><mml:mi>Y</mml:mi><mml:mo>&#x2223;</mml:mo><mml:mi>X</mml:mi><mml:mo>)</mml:mo><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>R</mml:mi><mml:mo>(</mml:mo><mml:mi>X</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:mfrac><mml:msubsup><mml:mrow><mml:mo>&#x220F;</mml:mo></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>n</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup><mml:mo>&#x200A;</mml:mo><mml:mi mathvariant="normal">e</mml:mi><mml:mi mathvariant="normal">x</mml:mi><mml:mi mathvariant="normal">p</mml:mi><mml:mo>&#x2061;</mml:mo><mml:mfenced open="{" close="}" separators="|"><mml:mrow><mml:msub><mml:mrow><mml:mi>W</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mfenced separators="|"><mml:mrow><mml:msub><mml:mrow><mml:mi>Y</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>Y</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo>&#x2223;</mml:mo><mml:mi>X</mml:mi></mml:mrow></mml:mfenced></mml:mrow></mml:mfenced></mml:math></disp-formula><disp-formula id="E10"><label>(10)</label><mml:math id="eqn10"><mml:mi>R</mml:mi><mml:mo>(</mml:mo><mml:mi>X</mml:mi><mml:mo>)</mml:mo><mml:mo>=</mml:mo><mml:msub><mml:mrow><mml:mo>&#x2211;</mml:mo></mml:mrow><mml:mrow><mml:mi>Y</mml:mi><mml:mo>&#x2208;</mml:mo><mml:mi>Y</mml:mi><mml:mo>(</mml:mo><mml:mi>x</mml:mi><mml:mo>)</mml:mo></mml:mrow></mml:msub><mml:mo>&#x200A;</mml:mo><mml:msubsup><mml:mrow><mml:mo>&#x220F;</mml:mo></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>n</mml:mi><mml:mo>+</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msubsup><mml:mo>&#x200A;</mml:mo><mml:mi mathvariant="normal">e</mml:mi><mml:mi mathvariant="normal">x</mml:mi><mml:mi mathvariant="normal">p</mml:mi><mml:mo>&#x2061;</mml:mo><mml:mfenced open="{" close="}" separators="|"><mml:mrow><mml:msub><mml:mrow><mml:mi>W</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mfenced separators="|"><mml:mrow><mml:msub><mml:mrow><mml:mi>Y</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>Y</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo>&#x2223;</mml:mo><mml:mi>X</mml:mi></mml:mrow></mml:mfenced></mml:mrow></mml:mfenced></mml:math></disp-formula><disp-formula id="E11"><label>(11)</label><mml:math id="eqn11"><mml:msub><mml:mrow><mml:mi>W</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mfenced separators="|"><mml:mrow><mml:msub><mml:mrow><mml:mi>Y</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>Y</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo>&#x2223;</mml:mo><mml:mi>X</mml:mi></mml:mrow></mml:mfenced><mml:mo>=</mml:mo><mml:msubsup><mml:mrow><mml:mo>&#x2211;</mml:mo></mml:mrow><mml:mrow><mml:mi>k</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>K</mml:mi></mml:mrow></mml:msubsup><mml:mo>&#x200A;</mml:mo><mml:msub><mml:mrow><mml:mi>w</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:msub><mml:mrow><mml:mi>f</mml:mi></mml:mrow><mml:mrow><mml:mi>k</mml:mi></mml:mrow></mml:msub><mml:mfenced separators="|"><mml:mrow><mml:msub><mml:mrow><mml:mi>Y</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi><mml:mo>-</mml:mo><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>Y</mml:mi></mml:mrow><mml:mrow><mml:mi>t</mml:mi></mml:mrow></mml:msub><mml:mo>&#x2223;</mml:mo><mml:mi>X</mml:mi></mml:mrow></mml:mfenced></mml:math></disp-formula><p>In addition, the equations for calculating R, P, and <italic>F</italic><sub>1</sub>-score are provided here:</p><disp-formula id="E12"><label>(12)</label><mml:math id="eqn12"><mml:mi mathvariant="normal">R</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi mathvariant="normal">T</mml:mi><mml:mi mathvariant="normal">P</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">T</mml:mi><mml:mi mathvariant="normal">P</mml:mi><mml:mo>+</mml:mo><mml:mi mathvariant="normal">F</mml:mi><mml:mi mathvariant="normal">N</mml:mi></mml:mrow></mml:mfrac></mml:math></disp-formula><disp-formula id="E13"><label>(13)</label><mml:math id="eqn13"><mml:mi mathvariant="normal">P</mml:mi><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mi mathvariant="normal">T</mml:mi><mml:mi mathvariant="normal">P</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">T</mml:mi><mml:mi mathvariant="normal">P</mml:mi><mml:mo>+</mml:mo><mml:mi mathvariant="normal">F</mml:mi><mml:mi mathvariant="normal">P</mml:mi></mml:mrow></mml:mfrac></mml:math></disp-formula><disp-formula id="E14"><label>(14)</label><mml:math id="eqn14"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:msub><mml:mrow><mml:mi mathvariant="italic">F</mml:mi></mml:mrow><mml:mrow><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mn>2</mml:mn><mml:mrow><mml:mi mathvariant="normal">P</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">R</mml:mi></mml:mrow></mml:mrow><mml:mrow><mml:mrow><mml:mi mathvariant="normal">P</mml:mi></mml:mrow><mml:mo>+</mml:mo><mml:mrow><mml:mi mathvariant="normal">R</mml:mi></mml:mrow></mml:mrow></mml:mfrac></mml:mrow></mml:mstyle></mml:math></disp-formula><p>Where TP is true positive, FN is false negative, and FP is false positive.</p></sec></sec></sec><sec id="s2-4"><title>The Semantic Representation and Analysis Module</title><p>We used HKG to enhance the semantic representation module in analyzing medical key named entities retrieved from the EMRs of patients with HICH, the relationships between entities, and the potential semantic associations between entities and relationships.</p><p>To further analyze the features of medical key named entities, we established a Word2vec model to vectorize relevant medical texts. We established mappings between medical key named entities and measured the similarity of medical texts (<xref ref-type="supplementary-material" rid="app3">Multimedia Appendix 3</xref>). The sentence-level representation and the cosine similarity between 2 vectors were obtained using the following formulas:</p><disp-formula id="E15"><label>(15)</label><mml:math id="eqn15"><mml:mstyle displaystyle="true" scriptlevel="0"><mml:mrow><mml:mstyle displaystyle="true" scriptlevel="0"><mml:msub><mml:mi>V</mml:mi><mml:mi>s</mml:mi></mml:msub><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:mo>&#x2211;</mml:mo><mml:mrow><mml:msubsup><mml:mi/><mml:mrow><mml:mi>i</mml:mi><mml:mo>=</mml:mo><mml:mn>1</mml:mn></mml:mrow><mml:mrow><mml:mi>m</mml:mi></mml:mrow></mml:msubsup></mml:mrow><mml:msub><mml:mi>V</mml:mi><mml:mrow><mml:mi>i</mml:mi></mml:mrow></mml:msub><mml:mo>&#x2217;</mml:mo><mml:msup><mml:mi>e</mml:mi><mml:mrow><mml:mi>w</mml:mi><mml:mrow><mml:mo>/</mml:mo></mml:mrow><mml:mo stretchy="false">(</mml:mo><mml:mi>i</mml:mi><mml:mo stretchy="false">)</mml:mo></mml:mrow></mml:msup></mml:mrow><mml:mi>m</mml:mi></mml:mfrac></mml:mstyle></mml:mrow></mml:mstyle></mml:math></disp-formula><disp-formula id="E16"><label>(16)</label><mml:math id="eqn16"><mml:mi>c</mml:mi><mml:mi>o</mml:mi><mml:mi>s</mml:mi><mml:mi>i</mml:mi><mml:mi>m</mml:mi><mml:mo>&#x2061;</mml:mo><mml:mfenced separators="|"><mml:mrow><mml:msub><mml:mrow><mml:mi>V</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">s</mml:mi><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mo>,</mml:mo><mml:msub><mml:mrow><mml:mi>V</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">s</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mfenced><mml:mo>=</mml:mo><mml:mfrac><mml:mrow><mml:msub><mml:mrow><mml:mi>V</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">s</mml:mi><mml:mn>1</mml:mn></mml:mrow></mml:msub><mml:mi mathvariant="normal">*</mml:mi><mml:msub><mml:mrow><mml:mi>V</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">s</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow><mml:mrow><mml:mfenced open="&#x2225;" close="&#x2225;" separators="|"><mml:mrow><mml:msub><mml:mrow><mml:mi>V</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">s</mml:mi><mml:mn>1</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mfenced><mml:mi>*</mml:mi><mml:mfenced open="&#x2225;" close="&#x2225;" separators="|"><mml:mrow><mml:msub><mml:mrow><mml:mi>V</mml:mi></mml:mrow><mml:mrow><mml:mi mathvariant="normal">s</mml:mi><mml:mn>2</mml:mn></mml:mrow></mml:msub></mml:mrow></mml:mfenced></mml:mrow></mml:mfrac></mml:math></disp-formula></sec><sec id="s2-5"><title>The Reasoning Module</title><p>This module combines logical reasoning with deep learning semantic similarity matching techniques, using an end-to-end pipeline model that integrates deep learning methods with prior knowledge. It uses natural language processing regularization techniques and deep learning semantic similarity matching techniques for logical diagnosis of entities. Furthermore, with the enhancement of HKG, the model gained stronger reasoning ability, and it output more explainable decisions.</p></sec><sec id="s2-6"><title>Assessment of the H-SYSTEM</title><p>The original treatment plan for the HICH cases was set as the gold standard, and 2 neurosurgeons (with more than 20 years of working experience) blinded to the processing conditions assessed the quality of the outputs of the H-SYSTEM using the following scoring standard. The whole treatment plan (full score=100 points) was divided into diagnostic measures (full score=15 points) and therapeutic measures (full score=85 points; detailed in <xref ref-type="supplementary-material" rid="app3">Multimedia Appendix 3</xref>). In the event of a disagreement between the 2 neurosurgeons, the final verdict was made by a chief neurosurgeon with more than 25 years of working experience. Interrater reliability was assessed with &#x03BA; statistics.</p><p>Then, the sensitivity, specificity, accuracy, positive predictive value, negative predictive value, and area under the curve (AUC) of the output were calculated and analyzed. Furthermore, the performance of the H-SYSTEM was also evaluated by analyzing the points distribution of treatment plans. The average time required for processing HICH cases by the H-SYSTEM was calculated and analyzed.</p><p>Meanwhile, we conducted a comparative analysis between the H-SYSTEM and ChatGPT (version 4.0) by inputting 300 randomly selected HICH cases into the 3 systems. Moreover, an additional 605 cases recently admitted to 6 medical centers from different regions were processed by the H-SYSTEM to test the practical application efficacy and generalization of the H-SYSTEM.</p></sec><sec id="s2-7"><title>Statistical Analyses</title><p>Data were analyzed with SPSS software (version 25.0 for Windows; IBM Corp) and GraphPad Prism (version 8.0 for Windows; GraphPad Software). Categorical variables are expressed as absolute numbers and percentages, and continuous variables are expressed as mean (SD). The sensitivity, specificity, positive predictive value, negative predictive value, and receiver operating characteristic curve were used to evaluate the performance of the H-SYSTEM. The AUC was used to measure the efficiency of different groups. Interrater agreement was measured using Cohen &#x03BA; value. Accuracy was calculated to evaluate the performance of the H-SYSTEM. Two-tailed <italic>P</italic>&#x003C;.05 was considered statistically significant. When calculating accuracy, sensitivity, specificity, and AUC, the following events were classified as positive events: reexamining head CT, surgical intervention, airway clearance, and intracranial pressure&#x2013;lowering treatment.</p></sec></sec><sec id="s3" sec-type="results"><title>Results</title><sec id="s3-1"><title>Reliability Analysis of HWS</title><sec id="s3-1-1"><title>Comparison With the Gold Standard</title><p>By setting the original treatment plans for the HICH cases as the gold standard, the overall accuracy of treatment plans made by nonneurosurgical doctors who rely on the HKG (HD) was 94.87%. Moreover, for diagnostic measures, rescue therapies, and surgical therapy, the accuracy, sensitivity, specificity, and AUC of HD were above 90%, while for drug therapies, these indexes of HD were above 85% (<xref ref-type="table" rid="table1">Table 1</xref> and <xref ref-type="fig" rid="figure4">Figure 4</xref>). These results indicate the HWS is accurate and reliable.</p><table-wrap id="t1" position="float"><label>Table 1.</label><caption><p>Comparison of the treatment plans of HD and ND when handling the same 800 cases. There were no statistically significant differences observed between the two in terms of diagnostic measures, surgical therapy, rescue therapies, or drug therapies.<sup><xref ref-type="table-fn" rid="table1fn1">a</xref></sup><sup>,</sup><sup><xref ref-type="table-fn" rid="table1fn2">b</xref></sup></p></caption><table id="table1" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom" rowspan="2"/><td align="left" valign="bottom" colspan="5">HD</td><td align="left" valign="bottom" colspan="5">ND</td></tr><tr><td align="left" valign="bottom">Overall treatment plans</td><td align="left" valign="bottom">Diagnostic measures</td><td align="left" valign="bottom">Surgical therapy</td><td align="left" valign="bottom">Rescue therapies</td><td align="left" valign="bottom">Drug therapies</td><td align="left" valign="bottom">Overall treatment plans</td><td align="left" valign="bottom">Diagnostic measures</td><td align="left" valign="bottom">Surgical therapy</td><td align="left" valign="bottom">Rescue therapies</td><td align="left" valign="bottom">Drug therapies</td></tr></thead><tbody><tr><td align="left" valign="top">Mean accuracy (95% CI), %</td><td align="left" valign="top">94.87 (94.58&#x2010;95.15)</td><td align="left" valign="top">93.89 (94.59&#x2010;95.14)</td><td align="left" valign="top">98.94 (98.08&#x2010;99.81)</td><td align="left" valign="top">93.14<break/>(91.99&#x2010;94.28)</td><td align="left" valign="top">86.10<break/>(76.05&#x2010;96.14)</td><td align="left" valign="top">94.86 (94.59&#x2010;95.14)</td><td align="left" valign="top">94.13 (90.63&#x2010;97.63)</td><td align="left" valign="top">98.89 (98.41&#x2010;99.37)</td><td align="left" valign="top">93.09<break/>(92.01&#x2010;94.18)</td><td align="left" valign="top">86.06 (75.50&#x2010;96.62)</td></tr><tr><td align="left" valign="top">Mean sensitivity (95% CI), %</td><td align="left" valign="top">N/A<sup><xref ref-type="table-fn" rid="table1fn3">c</xref></sup></td><td align="left" valign="top">97.77 (95.32&#x2010;100)</td><td align="left" valign="top">98.70 (96.12&#x2010;100)</td><td align="left" valign="top">90.27 (80.87&#x2010;99.66)</td><td align="left" valign="top">88.80 (84.41&#x2010;93.19)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">97.80 (94.79&#x2010;100)</td><td align="left" valign="top">98.50 (96.35&#x2010;100)</td><td align="left" valign="top">88.43 (72.57&#x2010;100)</td><td align="left" valign="top">83.33 (78.83&#x2010;87.84)</td></tr><tr><td align="left" valign="top">Mean specificity (95% CI), %</td><td align="left" valign="top">N/A</td><td align="left" valign="top">98.17 (96.05&#x2010;100)</td><td align="left" valign="top">99.10 (97.31&#x2010;100)</td><td align="left" valign="top">92.27 (83.50&#x2010;100)</td><td align="left" valign="top">86.50<break/>(77.55&#x2010;9545)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">98.03 (97.09&#x2010;98.97)</td><td align="left" valign="top">99.23 (98.23&#x2010;100)</td><td align="left" valign="top">89.60 (69.88&#x2010;100)</td><td align="left" valign="top">81.60 (66.06&#x2010;97.14)</td></tr><tr><td align="left" valign="top">Mean area under the curve (95% CI), %</td><td align="left" valign="top">N/A</td><td align="left" valign="top">97.93 (96.44&#x2010;99.43)</td><td align="left" valign="top">98.90<break/>(98.00&#x2010;99.80)</td><td align="left" valign="top">91.27 (82.75&#x2010;99.78)</td><td align="left" valign="top">87.70<break/>(81.67&#x2010;93.73)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">97.90 (95.93&#x2010;99.87)</td><td align="left" valign="top">98.87 (98.29&#x2010;99.44)</td><td align="left" valign="top">89.10 (70.98&#x2010;100)</td><td align="left" valign="top">82.47 (72.36&#x2010;92.57)</td></tr></tbody></table><table-wrap-foot><fn id="table1fn1"><p><sup>a</sup>HD: nonneurosurgical doctors who rely on the HWS (the hypertensive intracerebral hemorrhage weight scoring system) for their output results.</p></fn><fn id="table1fn2"><p><sup>b</sup>ND: senior neurosurgical doctors.</p></fn><fn id="table1fn3"><p><sup>c</sup>N/A: not applicable.</p></fn></table-wrap-foot></table-wrap><fig position="float" id="figure4"><label>Figure 4.</label><caption><p>Comparison of HD, ND, ChatGPT, and the H-SYSTEM. (A-B) For both ND and HD, the diagnostic measures, drug therapies, surgical therapy, and rescue therapies had a high accuracy, sensitivity, specificity, and AUC. (C) In the testing set, the diagnostic measures, surgical therapy, and rescue therapies output by the H-SYSTEM had a high accuracy, sensitivity, specificity, and AUC. (D) In terms of accuracy, whether in overall assessment or in individual evaluations of diagnostic measures, rescue therapies, surgical therapy, and drug therapies, the H-SYSTEM consistently outperformed ChatGPT. (E-H) The ROC curves representing rescue therapies, surgical therapies, drug therapies, and diagnostic measures in the results output by the HD. (I-L) The ROC curves representing rescue therapies, surgical therapies, drug therapies, and diagnostic measures in the results output by the ND. (M-O) The ROC curves for the rescue therapies, surgical therapies, and diagnostic measures indicated the H-SYSTEM exhibited satisfactory and reliable performance in the testing set. AUC: area under the curve; HD: nonneurosurgical doctors who rely on the medical domain knowledge graph for hypertensive intracerebral hemorrhage; ND: neurosurgical doctor; ROC: receiver operating characteristic.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="jmir_v27i1e66055_fig04.png"/></fig></sec><sec id="s3-1-2"><title>Comparison With ND</title><p>The results showed that the accuracy of overall treatment plans made by ND was lower than that of HD. For diagnostic measures, the accuracy of ND was slightly higher than that of HD, while for the surgical therapies and rescue therapies, the accuracy of HD was slightly higher than that of ND. However, the above differences were not statistically significant (94.86% vs 94.87%, 94.13% vs 93.89%, 98.94% vs 98.89%, 93.14% vs 93.09%; <italic>P</italic>&#x003E;.05; <xref ref-type="table" rid="table1">Table 1</xref>).</p></sec><sec id="s3-1-3"><title>Comparison of the Performance of HD and ND</title><p>As shown in <xref ref-type="table" rid="table2">Table 2</xref>, the concordance between the points distributions of overall treatment plans (100%) output by HD and ND was satisfactory (&#x03BA;=0.730). Moreover, for diagnostic measures, surgical therapies, rescue therapies, and drug therapies, the concordance between the points distributions of HD and ND was also satisfactory (&#x03BA;=0.805, 0.715, 0.890, 0.799; <xref ref-type="table" rid="table2">Table 2</xref> and <xref ref-type="fig" rid="figure4">Figure 4</xref>). Meanwhile, the H-SYSTEM also output the basis of treatment plans (detailed in <xref ref-type="supplementary-material" rid="app3">Multimedia Appendix 3</xref>).</p><table-wrap id="t2" position="float"><label>Table 2.</label><caption><p>The reliability analysis of the hypertensive intracerebral hemorrhage weight scoring system. The results generated by ND and HD for the same 800 cases were scored, and an analysis and comparison were conducted based on the distribution of scores.<sup><xref ref-type="table-fn" rid="table2fn1">a</xref></sup><sup>,</sup><sup><xref ref-type="table-fn" rid="table2fn2">b</xref></sup></p></caption><table id="table2" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom"/><td align="left" valign="bottom">HD, mean (SD), %</td><td align="left" valign="bottom">ND, mean (SD), %</td><td align="left" valign="bottom"><italic>P</italic> value</td><td align="left" valign="bottom">&#x03BA;</td></tr></thead><tbody><tr><td align="left" valign="top" colspan="4"><bold>Overall treatment plans (100 points)</bold></td><td align="left" valign="top">0.730</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>&#x003E;80 points</td><td align="left" valign="top">97.61 (0.29)</td><td align="left" valign="top">97.72 (0.31)</td><td align="left" valign="top">.81</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>60&#x2010;80 points</td><td align="left" valign="top">1.50 (0.17)</td><td align="left" valign="top">1.39 (0.29)</td><td align="left" valign="top">.76</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>&#x003C;60 points</td><td align="left" valign="top">0.89 (0.20)</td><td align="left" valign="top">0.89 (0.15)</td><td align="left" valign="top">&#x003E;.99</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top" colspan="4"><bold>Diagnostic measures (15 points)</bold></td><td align="left" valign="top">0.805</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>&#x003E;80% (12 points)</td><td align="left" valign="top">85.94 (2.20)</td><td align="left" valign="top">78.56 (5.42)</td><td align="left" valign="top">.28</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>60%&#x2010;80% (9&#x2010;12 points)</td><td align="left" valign="top">11.89 (2.02)</td><td align="left" valign="top">18.72 (5.64)</td><td align="left" valign="top">.32</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>&#x003C;60% (9 points)</td><td align="left" valign="top">2.17 (0.19)</td><td align="left" valign="top">2.72 (0.40)</td><td align="left" valign="top">.28</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top" colspan="4"><bold>Therapeutic measures (85 points)</bold></td><td align="left" valign="top"/></tr><tr><td align="left" valign="top" colspan="4"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content><bold>Surgical therapy (40 points)</bold></td><td align="left" valign="top">0.715</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>100% (40 points)</td><td align="left" valign="top">98.89 (0.11)</td><td align="left" valign="top">98.94 (0.20)</td><td align="left" valign="top">.82</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>0% (0 point)</td><td align="left" valign="top">1.11 (0.11)</td><td align="left" valign="top">1.06 (0.20)</td><td align="left" valign="top">.82</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top" colspan="4"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content><bold>Rescue therapies (35 points)</bold></td><td align="left" valign="top">0.890</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>&#x003E;80% (28 points)</td><td align="left" valign="top">92.67 (0.17)</td><td align="left" valign="top">92.50 (0.17)</td><td align="left" valign="top">.5</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>60%&#x2010;80% (21&#x2010;28 points)</td><td align="left" valign="top">4.28 (1.07)</td><td align="left" valign="top">4.28 (1.07)</td><td align="left" valign="top">&#x003E;.99</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>&#x003C;60% (21 points)</td><td align="left" valign="top">2.11 (0.06)</td><td align="left" valign="top">2.28 (0.20)</td><td align="left" valign="top">.47</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top" colspan="4"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content><bold>Drug therapies (10 points)</bold></td><td align="left" valign="top">0.799</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>&#x003E;80% (8 points)</td><td align="left" valign="top">72.44 (7.96)</td><td align="left" valign="top">72.00 (8.01)</td><td align="left" valign="top">.97</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>60%&#x2010;80% (6&#x2010;8 points)</td><td align="left" valign="top">23.61 (5.55)</td><td align="left" valign="top">23.78 (5.84)</td><td align="left" valign="top">.98</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>&#x003C;60% (6 points)</td><td align="left" valign="top">3.94 (2.44)</td><td align="left" valign="top">4.22 (2.34)</td><td align="left" valign="top">.94</td><td align="left" valign="top"/></tr></tbody></table><table-wrap-foot><fn id="table2fn1"><p><sup>a</sup>ND: senior neurosurgical doctors.</p></fn><fn id="table2fn2"><p><sup>b</sup>HD: nonneurosurgical doctors who rely on the HWS (the hypertensive intracerebral hemorrhage weight scoring system) for their output results.</p></fn></table-wrap-foot></table-wrap></sec></sec><sec id="s3-2"><title>Reliability Analysis of the Key NER Model</title><p>A total of 7 key NER models were constructed and evaluated, including IDCNN-CRF, BiLSTM-CRF, BiLSTM-IDCNN-CRF, BERT-CRF, BERT-IDCNN-CRF, BERT-BiLSTM-CRF, and BERT-IDCNN-BiLSTM-CRF. The P, R, and <italic>F</italic><sub>1</sub> values for the BERT-IDCNN-BiLSTM-CRF model were reported as 92.03, 90.22, and 91.11, which were significantly higher than those of the other models. These findings indicated that the BERT-IDCNN-BiLSTM-CRF model not only achieved better precision and recall in key NER but also maintained a balanced performance between the 2 metrics (<xref ref-type="fig" rid="figure5">Figure 5</xref> and <xref ref-type="table" rid="table3">Table 3</xref>).</p><fig position="float" id="figure5"><label>Figure 5.</label><caption><p>The training and evaluation process of the key named entity recognition module. (A) The validation process of the key named entity recognition module. (B-C) The change curves of P and R with the increase in iteration times. As the iterations progress, both P and R gradually reach a stable state. (D) Comparison between the BERT-IDCNN-BiLSTM-CRF model and other models in terms of P, R, and <italic>F</italic><sub>1</sub>-score. The BERT-IDCNN-BiLSTM-CRF model outperforms the other models in terms of precision, recall, and <italic>F</italic><sub>1</sub>-score, indicating it not only achieves better precision and recall in key named entity recognition, but also maintains a balanced performance between the 2 metrics. The P measures the proportion of actual positive samples among all samples predicted as positive by the model; R measures the proportion of successfully predicted positive samples among all actual positive samples. The <italic>F</italic><sub>1</sub>-score, being the harmonic mean of P and R, indicates stronger model performance as its value approaches 1. BERT: bidirectional encoder representations from transformers; BiLSTM: bidirectional long short-term memory; CRF: conditional random fields; IDCNN: inflated dilated convolutional neural network; P: precision; R: recall.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="jmir_v27i1e66055_fig05.png"/></fig><table-wrap id="t3" position="float"><label>Table 3.</label><caption><p>The comparative analysis of data across various models. All metrics of our approach are higher than those of other models.</p></caption><table id="table3" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Models</td><td align="left" valign="bottom">Precision</td><td align="left" valign="bottom">Recall</td><td align="left" valign="bottom"><italic>F</italic><sub>1</sub>-score</td></tr></thead><tbody><tr><td align="left" valign="top">IDCNN<sup><xref ref-type="table-fn" rid="table3fn1">a</xref></sup>-CRF<sup><xref ref-type="table-fn" rid="table3fn2">b</xref></sup></td><td align="left" valign="top">75.54</td><td align="left" valign="top">70.32</td><td align="left" valign="top">72.83</td></tr><tr><td align="left" valign="top">BiLSTM<sup><xref ref-type="table-fn" rid="table3fn3">c</xref></sup>-CRF</td><td align="left" valign="top">79.80</td><td align="left" valign="top">77.16</td><td align="left" valign="top">78.45</td></tr><tr><td align="left" valign="top">BiLSTM-IDCNN-CRF</td><td align="left" valign="top">78.16</td><td align="left" valign="top">72.37</td><td align="left" valign="top">75.15</td></tr><tr><td align="left" valign="top">BERT<sup><xref ref-type="table-fn" rid="table3fn4">d</xref></sup>-CRF</td><td align="left" valign="top">84.81</td><td align="left" valign="top">82.45</td><td align="left" valign="top">83.61</td></tr><tr><td align="left" valign="top">BERT-IDCNN-CRF</td><td align="left" valign="top">85.68</td><td align="left" valign="top">84.84</td><td align="left" valign="top">85.25</td></tr><tr><td align="left" valign="top">BERT-BiLSTM-CRF</td><td align="left" valign="top">87.04</td><td align="left" valign="top">85.31</td><td align="left" valign="top">86.16</td></tr><tr><td align="left" valign="top">BERT-IDCNN-BiLSTM-CRF (our approach)</td><td align="left" valign="top">92.03</td><td align="left" valign="top">90.22</td><td align="left" valign="top">91.11</td></tr></tbody></table><table-wrap-foot><fn id="table3fn1"><p><sup>a</sup>IDCNN: inflated dilated convolutional neural network.</p></fn><fn id="table3fn2"><p><sup>b</sup>CRF: conditional random fields.</p></fn><fn id="table3fn3"><p><sup>c</sup>BiLSTM: bidirectional long short-term memory.</p></fn><fn id="table3fn4"><p><sup>d</sup>BERT: bidirectional encoder representations from transformers.</p></fn></table-wrap-foot></table-wrap><p>The iteration curve of the BERT-IDCNN-BiLSTM-CRF model is depicted in <xref ref-type="fig" rid="figure5">Figures 5B and 5C</xref>. After approximately 200 iterations, there was no noteworthy enhancement observed in both P and R values for both the training and validation sets. This suggests that the model has attained its optimum and stable state. The BERT pretrained model exhibited robust semantic representation capabilities, which contributed to its excellent performance in downstream recognition tasks and significantly enhanced the effectiveness of the model in NER.</p></sec><sec id="s3-3"><title>Reliability Analysis of the H-SYSTEM</title><sec id="s3-3-1"><title>Overall Test of the H-SYSTEM</title><p>The accuracy of treatment plans output by the H-SYSTEM was 91.74%. which also showed significant consistency with the gold standard (<italic>P</italic>&#x003C;.05). For diagnostic measures, the accuracy, AUC, and &#x03BA; were 88.18%, 97.03%, and 0.874, respectively (<italic>P</italic>&#x003C;.05). For surgical therapy, these indexes were 98.53%, 98.53%, and 0.971 (<italic>P</italic>&#x003C;.05). For rescue therapies, these indexes were 89.50%, 94.67%, and 0.923, respectively (<italic>P</italic>&#x003C;.05; <xref ref-type="table" rid="table4">Tables 4</xref> and <xref ref-type="table" rid="table5">5</xref>, <xref ref-type="fig" rid="figure4">Figure 4</xref>). The detailed treatment plans of the gold standard and the H-SYSTEM are displayed in <xref ref-type="supplementary-material" rid="app3">Multimedia Appendix 3</xref><italic>.</italic></p><table-wrap id="t4" position="float"><label>Table 4.</label><caption><p>Comparison of the performance of doctors using the H-SYSTEM and ND.<sup><xref ref-type="table-fn" rid="table4fn1">a</xref></sup><sup>,</sup><sup><xref ref-type="table-fn" rid="table4fn2">b</xref></sup></p></caption><table id="table4" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom" rowspan="2"/><td align="left" valign="bottom" colspan="4"><sup><xref ref-type="table-fn" rid="table4fn3">c</xref></sup>H-SYSTEM</td><td align="left" valign="bottom" colspan="4">ND</td></tr><tr><td align="left" valign="bottom">Overall treatment plans</td><td align="left" valign="bottom">Diagnostic measures</td><td align="left" valign="bottom">Surgical therapy</td><td align="left" valign="bottom">Rescue therapies</td><td align="left" valign="bottom">Overall treatment plans</td><td align="left" valign="bottom">Diagnostic measures</td><td align="left" valign="bottom">Surgical therapy</td><td align="left" valign="bottom">Rescue therapies</td></tr></thead><tbody><tr><td align="left" valign="top">Mean accuracy (95% CI), %</td><td align="left" valign="top">94.87 (94.58&#x2010;95.15)</td><td align="left" valign="top">93.89 (94.59&#x2010;95.14)</td><td align="left" valign="top">98.94 (98.08&#x2010;99.81)</td><td align="left" valign="top">93.14 (91.99&#x2010;94.28)</td><td align="left" valign="top">94.86 (94.59&#x2010;95.14)</td><td align="left" valign="top">94.13 (90.63&#x2010;97.63)</td><td align="left" valign="top">98.89 (98.41&#x2010;99.37)</td><td align="left" valign="top">93.09 (92.01&#x2010;94.18)</td></tr><tr><td align="left" valign="top">Mean sensitivity (95% CI), %</td><td align="left" valign="top">N/A<sup><xref ref-type="table-fn" rid="table4fn4">d</xref></sup></td><td align="left" valign="top">97.77 (95.32&#x2010;100)</td><td align="left" valign="top">98.70 (96.12&#x2010;100)</td><td align="left" valign="top">90.27 (80.87&#x2010;99.66)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">97.80 (94.79&#x2010;100)</td><td align="left" valign="top">98.50 (96.35&#x2010;100)</td><td align="left" valign="top">88.43 (72.57&#x2010;100)</td></tr><tr><td align="left" valign="top">Mean specificity (95% CI), %</td><td align="left" valign="top">N/A</td><td align="left" valign="top">98.17 (96.05&#x2010;100)</td><td align="left" valign="top">99.10 (97.31&#x2010;100)</td><td align="left" valign="top">92.27 (83.50&#x2010;100)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">98.03 (97.09&#x2010;98.97)</td><td align="left" valign="top">99.23 (98.23&#x2010;100)</td><td align="left" valign="top">89.60 (69.88&#x2010;100)</td></tr><tr><td align="left" valign="top">Mean area under the curve (95% CI), %</td><td align="left" valign="top">N/A</td><td align="left" valign="top">97.93<break/>(96.44&#x2010;99.43)</td><td align="left" valign="top">98.90<break/>(98.00&#x2010;99.80)</td><td align="left" valign="top">91.27<break/>(82.75&#x2010;99.78)</td><td align="left" valign="top">N/A</td><td align="left" valign="top">97.90<break/>(95.93&#x2010;99.87)</td><td align="left" valign="top">98.87 (98.29&#x2010;99.44)</td><td align="left" valign="top">89.10 (70.98&#x2010;100)</td></tr></tbody></table><table-wrap-foot><fn id="table4fn1"><p><sup>a</sup>To enhance the validation of the performance of the H-SYSTEM, we input 1000 cases into the H-SYSTEM, compared its output with that of ND, and assessed their accuracy, sensitivity, specificity, and area under the curve through calculations and comparisons.</p></fn><fn id="table4fn2"><p><sup>b</sup>ND: senior neurosurgical doctors.</p></fn><fn id="table4fn3"><p><sup>c</sup>HD: nonneurosurgical doctors who rely on the HWS (the hypertensive intracerebral hemorrhage weight scoring system) for their output results.</p></fn><fn id="table4fn4"><p><sup>d</sup>N/A: not applicable.</p></fn></table-wrap-foot></table-wrap><table-wrap id="t5" position="float"><label>Table 5.</label><caption><p>The reliability analysis of the H-SYSTEM and the distribution of scores on different treatment plans output by the H-SYSTEM and ND.<sup><xref ref-type="table-fn" rid="table5fn1">a</xref></sup></p></caption><table id="table5" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom"/><td align="left" valign="bottom">H-SYSTEM, mean (SD), %</td><td align="left" valign="bottom">ND, mean (SD), %</td><td align="left" valign="bottom"><italic>P</italic> value</td><td align="left" valign="bottom">&#x03BA;</td></tr></thead><tbody><tr><td align="left" valign="top" colspan="4"><bold>Overall treatment plans (100 points)</bold></td><td align="left" valign="top">0.841</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>&#x003E;80 points</td><td align="left" valign="top">94.90 (0.404)</td><td align="left" valign="top">92.77 (0.73)</td><td align="left" valign="top">.06</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>60&#x2010;80 points</td><td align="left" valign="top">3.73 (0.52)</td><td align="left" valign="top">4.60 (0.30)</td><td align="left" valign="top">.06</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>&#x003C;60 points</td><td align="left" valign="top">1.37 (0.15)</td><td align="left" valign="top">2.63 (0.53)</td><td align="left" valign="top">.20</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top" colspan="4"><bold>Diagnostic measures (15 points)</bold></td><td align="left" valign="top">0.874</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>&#x003E;80% (12 points)</td><td align="left" valign="top">77.57 (0.09)</td><td align="left" valign="top">75.17 (0.90)</td><td align="left" valign="top">.11</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>60%&#x2010;80% (9&#x2010;12 points)</td><td align="left" valign="top">17.53 (0.27)</td><td align="left" valign="top">18.13 (0.93)</td><td align="left" valign="top">.64</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>&#x003C;60% (9 points)</td><td align="left" valign="top">4.90 (0.26)</td><td align="left" valign="top">6.70 (0.10)</td><td align="left" valign="top">.02</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top" colspan="5"><bold>Therapeutic measures (85 points)</bold></td></tr><tr><td align="left" valign="top" colspan="4"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content><bold><italic>Surgical therapy (40 points)</italic></bold></td><td align="left" valign="top">0.221</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>100% (40 points)</td><td align="left" valign="top">98.60 (0.12)</td><td align="left" valign="top">98.53 (0.15)</td><td align="left" valign="top">.80</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>0% (0 points)</td><td align="left" valign="top">1.40 (0.12)</td><td align="left" valign="top">1.47 (0.15)</td><td align="left" valign="top">.80</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top" colspan="4"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content><bold><italic>Rescue therapies (35 points)</italic></bold></td><td align="left" valign="top">0.923</td></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>&#x003E;80% (28 points)</td><td align="left" valign="top">88.00 (0.36)</td><td align="left" valign="top">87.13 (0.46)</td><td align="left" valign="top">.32</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>60%&#x2010;80% (21&#x2010;28 points)</td><td align="left" valign="top">8.83 (0.38)</td><td align="left" valign="top">9.17 (0.43)</td><td align="left" valign="top">.63</td><td align="left" valign="top"/></tr><tr><td align="left" valign="top"><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content><named-content content-type="indent">&#x00A0;&#x00A0;&#x00A0;&#x00A0;</named-content>&#x003C;60% (21 points)</td><td align="left" valign="top">3.17 (0.07)</td><td align="left" valign="top">3.70 (0.06)</td><td align="left" valign="top">.03</td><td align="left" valign="top"/></tr></tbody></table><table-wrap-foot><fn id="table5fn1"><p><sup>a</sup>ND: senior neurosurgical doctors.</p></fn></table-wrap-foot></table-wrap><p>The score distributions of treatment plans output by the H-SYSTEM and ND showed no significant difference. The concordance between the 2 raters on the independent testing dataset of the H-SYSTEM and ND output was satisfactory (&#x03BA;=0.841). In the evaluation of the qualification rate for the distribution of the overall points, there was no statistically significant difference between the H-SYSTEM and ND output. Furthermore, the majority of both HD and ND output results had an accuracy rate exceeding 80%. The difference between ND and HD remained statistically insignificant (<xref ref-type="table" rid="table2">Tables 2</xref> and <xref ref-type="table" rid="table5">5</xref>, <xref ref-type="fig" rid="figure6">Figure 6</xref>).</p><fig position="float" id="figure6"><label>Figure 6.</label><caption><p>Comparison of output results of HD, ND, H-SYSTEM, ChatGPT, and ChatDoctor. (A) The score distributions of treatment plans output by ND and HD showed no significant difference. For overall treatment plans, diagnostic measures, surgical therapies, rescue therapies, and drug therapies, the concordance between the points distributions of HD was satisfactory. (B) The score distributions of treatment plans output by the H-SYSTEM and ND showed no significant difference. The H-SYSTEM has an accuracy rate of over 80% in the majority of cases, and there is a high level of consistency between the results produced by the H-SYSTEM and ND. (C) Comparison of comprehensive efficiency between the H-SYSTEM and ND. (D) Quantitative comparison of ChatDoctor, ChatGPT, and the H-SYSTEM with BERTScore. HD: nonneurosurgical doctors who rely on the medical domain knowledge graph for hypertensive intracerebral hemorrhage; ND: neurosurgical doctor.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="jmir_v27i1e66055_fig06.png"/></fig></sec><sec id="s3-3-2"><title>Comprehensive Effectiveness</title><p>For ND, the total time spent on 60 cases and a single case was significantly longer as compared with the H-SYSTEM. Obviously, the speed of the H-SYSTEM in processing HICH cases was significantly higher than that of ND (<italic>P</italic>&#x003C;.001). However, no statistical difference in the sensitivity, specificity, accuracy, and AUC was found between the treatment plans of doctors and the H-SYSTEM (<italic>P</italic>&#x003E;.05; <xref ref-type="fig" rid="figure6">Figure 6</xref>, Tables S2 and S3 in <xref ref-type="supplementary-material" rid="app3">Multimedia Appendix 3</xref>).</p></sec><sec id="s3-3-3"><title>Practical Application Efficacy and Generalization of the H-SYSTEM</title><p>The results showed that ChatGPT achieved an overall accuracy of 91.48% (95% CI 90.14%&#x2010;92.77%), which is significantly lower than that of the H-SYSTEM (95.26%, 95% CI 94.43%&#x2010;96.83%). Meanwhile, for diagnostic measures, surgical therapies, rescue therapies, and drug therapies, the H-SYSTEM also had significantly higher accuracy as compared with ChatGPT (<xref ref-type="table" rid="table6">Table 6</xref>). Therefore, for patient medical text processing and decision-making, the H-SYSTEM surpassed ChatGPT with statistical significance (<italic>P</italic>&#x003C;.05).</p><table-wrap id="t6" position="float"><label>Table 6.</label><caption><p>Comparison between the H-SYSTEM and ChatGPT (N=300 cases). The accuracy of the H-SYSTEM was significantly higher than that of ChatGPT.</p></caption><table id="table6" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom"/><td align="left" valign="bottom">ChatGPT, mean (95% CI), %</td><td align="left" valign="bottom">H-SYSTEM, mean (95% CI), %</td><td align="left" valign="bottom"><italic>P</italic> value</td></tr></thead><tbody><tr><td align="left" valign="top">Overall treatment plans</td><td align="left" valign="top">91.48 (90.14&#x2010;92.77&#xFF09;</td><td align="left" valign="top">95.26 (94.43&#x2010;96.83&#xFF09;</td><td align="left" valign="top">&#x003C;.01</td></tr><tr><td align="left" valign="top">Diagnostic measures</td><td align="left" valign="top">75.00 (70.07&#x2010;79.93)</td><td align="left" valign="top">86.11 (82.19&#x2010;90.03)</td><td align="left" valign="top">&#x003C;.001</td></tr><tr><td align="left" valign="top">Surgical therapies</td><td align="left" valign="top">96.00 (93.77&#x2010;98.23&#xFF09;</td><td align="left" valign="top">99.00 (97.87&#x2010;100&#xFF09;</td><td align="left" valign="top">.02</td></tr><tr><td align="left" valign="top">Rescue therapies</td><td align="left" valign="top">96.36 (95.40&#x2010;97.31&#xFF09;</td><td align="left" valign="top">98.16 (97.64&#x2010;98.67&#xFF09;</td><td align="left" valign="top">.002</td></tr><tr><td align="left" valign="top">Drug therapies</td><td align="left" valign="top">81.08 (77.84&#x2010;84.33&#xFF09;</td><td align="left" valign="top">86.44 (84.72&#x2010;88.15&#xFF09;</td><td align="left" valign="top">.006</td></tr></tbody></table></table-wrap><p>In another study, a medical chat model fine-tuned on a large language model meta-AI (LLaMA) named ChatDoctor showed higher P, R, and <italic>F</italic><sub>1</sub>-scores (84.44, 84.51, and 84.46, respectively) as compared with ChatGPT (83.7, 84.45, and 84.06, respectively) when processing medical text tasks [<xref ref-type="bibr" rid="ref31">31</xref>]. Similarly, in our study, the P, R, and <italic>F</italic><sub>1</sub>-score of the H-SYSTEM were 98.53, 98.33, and 98.43, which were all higher than that of ChatGPT (96.00, 77.00, and 85.46, respectively).</p><p>An additional 605 patients recently admitted to 6 different medical centers from different regions were included to further test the practical application efficacy of the H-SYSTEM. The results showed that the total accuracy of treatment plans output by the H-SYSTEM was 92.22% (range: 91.14%-95.35%; <xref ref-type="table" rid="table7">Table 7</xref>).</p><table-wrap id="t7" position="float"><label>Table 7.</label><caption><p>Performance of the H-SYSTEM on HICH electronic medical records from different medical centers. The accuracies of the H-SYSTEM on HICH cases from different medical centers ranged from 91.14% to 95.35%.<sup><xref ref-type="table-fn" rid="table7fn1">a</xref></sup></p></caption><table id="table7" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom"/><td align="left" valign="bottom">Quantity of HICH cases</td><td align="left" valign="bottom">Overall treatment plans, mean (95% CI), %</td></tr></thead><tbody><tr><td align="left" valign="top">Overall medical centers</td><td align="left" valign="top">605</td><td align="left" valign="top">92.22 (91.57&#x2010;92.88)</td></tr><tr><td align="left" valign="top">Medical center 1 (Chongqing)</td><td align="left" valign="top">216</td><td align="left" valign="top">92.04 (90.95&#x2010;93.13)</td></tr><tr><td align="left" valign="top">Medical center 2 (Chongqing)</td><td align="left" valign="top">128</td><td align="left" valign="top">91.93 (90.41&#x2010;93.35)</td></tr><tr><td align="left" valign="top">Medical center 3 (Sichuan)</td><td align="left" valign="top">95</td><td align="left" valign="top">92.18 (90.22&#x2010;94.15)</td></tr><tr><td align="left" valign="top">Medical center 4 (Yunnan)</td><td align="left" valign="top">73</td><td align="left" valign="top">91.14 (89.10&#x2010;93.18)</td></tr><tr><td align="left" valign="top">Medical center 5 (Yunnan)</td><td align="left" valign="top">62</td><td align="left" valign="top">93.23 (91.90&#x2010;94.57)</td></tr><tr><td align="left" valign="top">Medical center 6 (Shanxi)</td><td align="left" valign="top">31</td><td align="left" valign="top">95.35 (93.88&#x2010;96.83)</td></tr></tbody></table><table-wrap-foot><fn id="table7fn1"><p><sup>a</sup>HICH: hypertensive intracerebral hemorrhage.</p></fn></table-wrap-foot></table-wrap></sec></sec></sec><sec id="s4" sec-type="discussion"><title>Discussion</title><sec id="s4-1"><title>Principal Findings</title><p>Despite increasing attention being directed to the research and application of AI in medicine [<xref ref-type="bibr" rid="ref23">23</xref>,<xref ref-type="bibr" rid="ref32">32</xref>], several challenges remain substantial obstacles to the development and translation of medical AI systems into clinical practice. The first challenge is access to large and well-annotated datasets. The second challenge is that the inner workings and decision-making processes of machine-learning algorithms remain opaque, which is also called the black box effect. In addition, although AI models like LLMs have shown excellent performance in various tasks, they do not perform as well as expected in clinical medical practice due to the highly abstract and specialized medical domain knowledge required and the complexity of clinical tasks.</p><p>To address these challenges, we developed a medical domain knowledge graph&#x2013;enhanced automatic decision support system, named H-SYSTEM, based on adequate data from 15 medical centers. Meanwhile, the experienced neurosurgeon-AI team ensured that the H-SYSTEM obtained large and well-annotated datasets and had a high generalization capacity. The H-SYSTEM identifies and analyzes data from patients&#x2019; medical history, physical examination, and auxiliary examination, then outputs the treatment plans; it also provides the basis for these decisions (<xref ref-type="fig" rid="figure3">Figure 3</xref>). Different from LLMs and routine deep learning systems [<xref ref-type="bibr" rid="ref23">23</xref>,<xref ref-type="bibr" rid="ref32">32</xref>-<xref ref-type="bibr" rid="ref34">34</xref>], the H-SYSTEM is enhanced by a medical domain knowledge graph (called HKG), which means that it can efficiently output both explainable and elaborate clinical decisions (<xref ref-type="supplementary-material" rid="app2">Multimedia Appendix 2</xref>). One highlight of the H-SYSTEM is its ability to explain the rationale of its decisions, helping the neurosurgeons understand its inner working process. Thus, this understandable system can gain more trust from doctors and facilitate both the development and adoption of AI systems in clinical practice.</p><p>The HKG in this study covers medical knowledge from 3 dimensions: the general medical domain knowledge, the medical subdomain knowledge, and the medical subdomain-specific knowledge. The treatment plans based on the HKG showed high reliability as compared with doctors. Therefore, the HKG was a reliable guide and external knowledge brain for the H-SYSTEM. It enhanced not only the capability of the semantic analysis and representation model to handle clinical tasks but also the explainability of the decision output by the H-SYSTEM.</p><p>The key NER model, BERT-IDCNN-BiLSTM-CRF, exhibited excellent performance with fast convergence and efficient extraction of key named entities, and it outperformed other NER models in this study (<xref ref-type="table" rid="table3">Table 3</xref> and <xref ref-type="fig" rid="figure4">Figure 4</xref>). Deep learning&#x2013;based models offer ample representability for text, enabling efficient problem-solving with high-quality outcomes. Meanwhile, the word vector space enhanced by the HKG proved to be more suitable for medical scenarios (<xref ref-type="supplementary-material" rid="app3">Multimedia Appendix 3</xref>). These models efficiently encode information from the data, facilitating the effective utilization of available data in the model.</p><p>The H-SYSTEM in this study is designed to imitate the process and thinking style of a neurosurgeon during HICH diagnosis and treatment in the emergency room. Like a neurosurgeon, the system analyzes all the medical records that can be obtained in the emergency room, including chief complaints, current medical history, important past illness history, physical examination, and important auxiliary examinations. Compared with the AI system in our previous study [<xref ref-type="bibr" rid="ref34">34</xref>], the H-SYSTEM in this study has a more powerful ability in key NER, semantic analysis, and logical decision-making. This enables it to process more complicated HICH cases and output more elaborate treatment plans that are practical in the emergency room. Besides outputting elaborate treatment plans, the H-SYSTEM also displays the basis of making these treatment plans. Our approach mimics the workflow of neurosurgeons in clinical practice and enables the users to independently review the basis of the H-SYSTEM&#x2019;s recommendations. This provides the rationale or support for its decisions and helps the human users understand the inner working process of the H-SYSTEM. Therefore, the reliability of the H-SYSTEM is significantly increased, which may help it gain the trust of neurosurgeons.</p><p>The treatment plans output by the H-SYSTEM consist of diagnostic and therapeutic measures. For patients with a complicated past medical history or unclear diagnosis, diagnostic measures are crucial for the whole treatment process. For instance, if a patient with intracerebral hemorrhage is suspected of having cerebrovascular diseases or a tumor-associated hemorrhage, then cerebrovascular-related examinations like computed tomography angiography, magnetic resonance angiography, or digital subtraction angiography, or multiple-modal magnetic resonance imaging, are included in their diagnostic measures to verify or exclude cerebrovascular diseases and brain tumors. Furthermore, for patients with a complicated past medical history, such as hemophilia, anticoagulant therapy, or hemodialysis, blood coagulation function examinations are included in their diagnostic measures to evaluate the bleeding and rebleeding risk in subsequent treatments.</p><p>Compared with the diagnostic measures, the therapeutic measures designed in the study were more complicated and elaborate. For instance, the respiratory tract management in the rescue therapy consisted of several measures, including cleaning up the airway, tracheal cannula insertion, and application of a respirator. Each measure had elaborate scoring rules and thresholds that determined whether the measure would be implemented or not. Two situations are the most concerning for a neurosurgeon in the emergency room&#x2014;the vital signs and surgical indications of patients with HICH&#x2014;and the performance of the H-SYSTEM was reliable for both rescue therapies and surgical therapies. Meanwhile, the efficiency of the H-SYSTEM in processing HICH cases was significantly higher than that of a neurosurgeon, indicating that the H-SYSTEM can provide reliable and efficient support in the diagnosis and treatment of HICH (<xref ref-type="fig" rid="figure6">Figure 6</xref>, Table S1 in <xref ref-type="supplementary-material" rid="app3">Multimedia Appendix 3</xref>).</p><p>During the comparison with 2 LLMs, the H-SYSTEM also showed excellent performance. Although ChatGPT performed well in some relatively simple HICH cases, its output treatment plans are more like general management principles. Moreover, when facing complicated HICH cases that had a changeful condition, complicated medical history, or atypical clinical and CT manifestation, the outputs of ChatGPT were not so accurate and did not offer substantial assistance and support to doctors. Meanwhile, the ChatDoctor, which has been fine-tuned on LLaMA, surpassed ChatGPT in multiple parameters by using medical domain knowledge [<xref ref-type="bibr" rid="ref31">31</xref>]. Because ChatGPT is not open source, we cannot compare it with the H-SYSTEM directly. Therefore, we took ChatGPT as the medium and calculated the P, R, and <italic>F</italic><sub>1</sub>-score. The results showed that the above values of ChatDoctor were higher than that of ChatGPT, but still lower than that of the H-SYSTEM. These results indicated that, without sufficient medical subdomain and subdomain-specific knowledge, even LLMs could not fully demonstrate their advantages in clinical text processing tasks.</p><p>In contrast, the H-SYSTEM is an HICH-specialized decision support system with dedicated diagnostic and treatment logics for HICH and a large training and validating database from multiple medical centers. Hence, it accurately processed complicated HICH cases and provided valuable assistance and support for doctors.</p><p>Furthermore, the HICH data in this study originated from 15 medical centers, covering an area of nearly 6 square kilometers in major provinces of western China, with a population of approximately 300 million. The EMRs of these HICH cases not only encompass different EMR systems of various medical centers but also reflect the diverse writing habits and academic and educational backgrounds of different doctors. This indicates that, in addition to its dependable clinical support for neurosurgeons, the H-SYSTEM also exhibits outstanding generalization capability.</p><p>Although the results of this study are promising, we have to acknowledge it does have some limitations. For the H-SYSTEM, the biggest challenge was to process and analyze the ambiguous EMRs. Due to various reasons, such as limitations of the EMR system, writing habits, and academic and educational backgrounds, some EMRs seem ambiguous, making it difficult to uncover the patient&#x2019;s real information. Therefore, we plan to continuously expand the current dataset to cover more medical centers and regions. Moreover, more complex cases, which are also challenging for neurosurgeons, are still needed for training the system. Additionally, the H-SYSTEM still has room for improvement in providing more personalized diagnosis and treatment plans. Although the H-SYSTEM identified CT reports in the EMR with high accuracy, it could not directly recognize the CT image. Thus, some possible biases caused by different radiologists might be inevitable. In future studies, these possible biases will be addressed by integrating automatic image segmentation into the H-SYSTEM.</p></sec><sec id="s4-2"><title>Conclusions</title><p>This study developed H-SYSTEM, an AI-based decision support system for HICH, which outperformed existing models like ChatGPT and ChatDoctor, and showed strong generalization capabilities for potential clinical application.</p><p>The H-SYSTEM is based on multiple-center data and shows significantly high efficiency and generalization capacity in processing EMRs; it also provides explainable and elaborate treatment plans. Therefore, it has the potential to gain more trust from doctors and facilitate both the development and adoption of AI systems in clinical practice.</p></sec></sec></body><back><ack><p>Thanks to the following medical centers for providing data support: the First Affiliated Hospital of Chongqing Medical University, the Second Affiliated Hospital of Chongqing Medical University, the First People&#x2019;s Hospital of Yunnan Province, the People&#x2019;s Hospital of Daozhen Autonomous County, the First Hospital of Shangxi Medical University, Sichuan Provincial People&#x2019;s Hospital, Yongchuan Hospital of Chongqing Medical University, University-Town Hospital of Chongqing Medical University, Chongqing Sanbo Jiangling Hospital, Chongqing Emergency Medical Center, Chongqing University Cancer Hospital, Chongqing Traditional Chinese Medicine Hospital, the People&#x2019;s Hospital of YouYang Tujia and Miao Autonomous County, Chongqing University Three Gorges Hospital, and the People&#x2019;s Hospital of Tongnan District. The study has been approved by the ethical committees of the aforementioned medical centers. This work is supported by the Chongqing Medical Scientific Research Project (joint project of Chongqing Health Commission and Science and Technology Bureau, number 2022MSXM041), the Program for Youth Innovation in Future Medicine, Chongqing Medical University (Scientific Research and Innovation Team, number W0169), Chongqing Natural Science Foundation General Project (number CSTB2022NSCQ-MSX0152), and the National Natural Science Foundation for Youth of China (number 81701226).</p></ack><notes><sec><title>Data Availability</title><p>Due to the protection of personal privacy, the dataset analyzed during the study is not publicly available, but can be obtained upon reasonable request from the corresponding author.</p></sec></notes><fn-group><fn fn-type="con"><p>All the authors contributed to the successful completion of this research. This study was designed and managed by JL and LJ. Material preparation, data collection, and analysis were performed by BD, Y Xia and LJ. The first draft of the manuscript was written by Y Xia and LJ. All authors made comments and suggestions on previous editions of the manuscript. The final manuscript was approved after all the authors had read it to their satisfaction.</p></fn><fn fn-type="conflict"><p>None declared.</p></fn></fn-group><glossary><title>Abbreviations</title><def-list><def-item><term id="abb1">AI</term><def><p>artificial intelligence</p></def></def-item><def-item><term id="abb2">BERT</term><def><p>bidirectional encoder representations from transformers</p></def></def-item><def-item><term id="abb3">BiLSTM</term><def><p>bidirectional long short-term memory</p></def></def-item><def-item><term id="abb4">CRF</term><def><p>conditional random fields</p></def></def-item><def-item><term id="abb5">CT</term><def><p>computed tomography</p></def></def-item><def-item><term id="abb6">EMR</term><def><p>electronic medical record</p></def></def-item><def-item><term id="abb7">HD</term><def><p>nonneurosurgical doctors who rely on the HKG</p></def></def-item><def-item><term id="abb8">HICH</term><def><p>hypertensive intracerebral hemorrhage</p></def></def-item><def-item><term id="abb9">HICH-IT</term><def><p>hypertensive intracerebral hemorrhage image and text database</p></def></def-item><def-item><term id="abb10">HKG</term><def><p>HICH knowledge graph</p></def></def-item><def-item><term id="abb11">HWS</term><def><p>hemorrhage weight system</p></def></def-item><def-item><term id="abb12">ICH</term><def><p>intracerebral hemorrhage</p></def></def-item><def-item><term id="abb13">IDCNN</term><def><p>inflated dilated convolutional neural network</p></def></def-item><def-item><term id="abb14">LLaMA</term><def><p>large language model meta-AI</p></def></def-item><def-item><term id="abb15">ND</term><def><p>neurosurgical doctor</p></def></def-item><def-item><term id="abb16">NER</term><def><p>named entity recognition</p></def></def-item></def-list></glossary><ref-list><title>References</title><ref id="ref1"><label>1</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lavados</surname><given-names>PM</given-names> </name><name name-style="western"><surname>Hoffmeister</surname><given-names>L</given-names> </name><name name-style="western"><surname>Moraga</surname><given-names>AM</given-names> </name><etal/></person-group><article-title>Incidence, risk factors, prognosis, and health-related quality of life after stroke in a low-resource community in Chile (&#x00D1;ANDU): a prospective population-based study</article-title><source>Lancet Glob Health</source><year>2021</year><month>03</month><volume>9</volume><issue>3</issue><fpage>e340</fpage><lpage>e351</lpage><pub-id pub-id-type="doi">10.1016/S2214-109X(20)30470-8</pub-id><pub-id pub-id-type="medline">33422189</pub-id></nlm-citation></ref><ref id="ref2"><label>2</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sarfo</surname><given-names>FS</given-names> </name><name name-style="western"><surname>Akpa</surname><given-names>OM</given-names> </name><name name-style="western"><surname>Ovbiagele</surname><given-names>B</given-names> </name><etal/></person-group><article-title>Patient-level and system-level determinants of stroke fatality across 16 large hospitals in Ghana and Nigeria: a prospective cohort study</article-title><source>Lancet Glob Health</source><year>2023</year><month>04</month><volume>11</volume><issue>4</issue><fpage>e575</fpage><lpage>e585</lpage><pub-id pub-id-type="doi">10.1016/S2214-109X(23)00038-4</pub-id><pub-id pub-id-type="medline">36805867</pub-id></nlm-citation></ref><ref id="ref3"><label>3</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Flaherty</surname><given-names>ML</given-names> </name><name name-style="western"><surname>Haverbusch</surname><given-names>M</given-names> </name><name name-style="western"><surname>Sekar</surname><given-names>P</given-names> </name><etal/></person-group><article-title>Long-term mortality after intracerebral hemorrhage</article-title><source>Neurology (ECronicon)</source><year>2006</year><month>04</month><day>25</day><volume>66</volume><issue>8</issue><fpage>1182</fpage><lpage>1186</lpage><pub-id pub-id-type="doi">10.1212/01.wnl.0000208400.08722.7c</pub-id><pub-id pub-id-type="medline">16636234</pub-id></nlm-citation></ref><ref id="ref4"><label>4</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Banzrai</surname><given-names>C</given-names> </name><name name-style="western"><surname>Bosookhuu</surname><given-names>O</given-names> </name><name name-style="western"><surname>Yadamsuren</surname><given-names>E</given-names> </name><etal/></person-group><article-title>Incidence and outcomes for stroke in Ulaanbaatar, Mongolia, during 2019-21: a prospective population-based study</article-title><source>Lancet Glob Health</source><year>2023</year><month>06</month><volume>11</volume><issue>6</issue><fpage>e942</fpage><lpage>e952</lpage><pub-id pub-id-type="doi">10.1016/S2214-109X(23)00130-4</pub-id><pub-id pub-id-type="medline">37119831</pub-id></nlm-citation></ref><ref id="ref5"><label>5</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hankey</surname><given-names>GJ</given-names> </name></person-group><article-title>The global and regional burden of stroke</article-title><source>Lancet Glob Health</source><year>2013</year><month>11</month><volume>1</volume><issue>5</issue><fpage>e239</fpage><lpage>40</lpage><pub-id pub-id-type="doi">10.1016/S2214-109X(13)70095-0</pub-id><pub-id pub-id-type="medline">25104481</pub-id></nlm-citation></ref><ref id="ref6"><label>6</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Adeoye</surname><given-names>O</given-names> </name><name name-style="western"><surname>Broderick</surname><given-names>JP</given-names> </name></person-group><article-title>Advances in the management of intracerebral hemorrhage</article-title><source>Nat Rev Neurol</source><year>2010</year><month>11</month><volume>6</volume><issue>11</issue><fpage>593</fpage><lpage>601</lpage><pub-id pub-id-type="doi">10.1038/nrneurol.2010.146</pub-id><pub-id pub-id-type="medline">20877400</pub-id></nlm-citation></ref><ref id="ref7"><label>7</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Tatlisumak</surname><given-names>T</given-names> </name><name name-style="western"><surname>Cucchiara</surname><given-names>B</given-names> </name><name name-style="western"><surname>Kuroda</surname><given-names>S</given-names> </name><name name-style="western"><surname>Kasner</surname><given-names>SE</given-names> </name><name name-style="western"><surname>Putaala</surname><given-names>J</given-names> </name></person-group><article-title>Nontraumatic intracerebral haemorrhage in young adults</article-title><source>Nat Rev Neurol</source><year>2018</year><month>04</month><volume>14</volume><issue>4</issue><fpage>237</fpage><lpage>250</lpage><pub-id pub-id-type="doi">10.1038/nrneurol.2018.17</pub-id><pub-id pub-id-type="medline">29521335</pub-id></nlm-citation></ref><ref id="ref8"><label>8</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Chen</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Wright</surname><given-names>N</given-names> </name><name name-style="western"><surname>Guo</surname><given-names>Y</given-names> </name><etal/></person-group><article-title>Mortality and recurrent vascular events after first incident stroke: a 9-year community-based study of 0&#x00B7;5 million Chinese adults</article-title><source>Lancet Glob Health</source><year>2020</year><month>04</month><volume>8</volume><issue>4</issue><fpage>e580</fpage><lpage>e590</lpage><pub-id pub-id-type="doi">10.1016/S2214-109X(20)30069-3</pub-id><pub-id pub-id-type="medline">32199124</pub-id></nlm-citation></ref><ref id="ref9"><label>9</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Koutkias</surname><given-names>V</given-names> </name><name name-style="western"><surname>Bouaud</surname><given-names>J</given-names> </name><collab>Section Editors for the IMIA Yearbook Section on Decision Support</collab></person-group><article-title>Contributions on Clinical Decision Support from the 2018 Literature</article-title><source>Yearb Med Inform</source><year>2019</year><month>08</month><volume>28</volume><issue>1</issue><fpage>135</fpage><lpage>137</lpage><pub-id pub-id-type="doi">10.1055/s-0039-1677929</pub-id><pub-id pub-id-type="medline">31419825</pub-id></nlm-citation></ref><ref id="ref10"><label>10</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Longde</surname><given-names>W</given-names> </name></person-group><article-title>Brief report on stroke prevention and treatment in China, 2019</article-title><source>Chinese Journal of Cerebrovascular Diseases</source><year>2020</year><fpage>272</fpage><lpage>281</lpage></nlm-citation></ref><ref id="ref11"><label>11</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Brownstein</surname><given-names>JS</given-names> </name><name name-style="western"><surname>Rader</surname><given-names>B</given-names> </name><name name-style="western"><surname>Astley</surname><given-names>CM</given-names> </name><name name-style="western"><surname>Tian</surname><given-names>H</given-names> </name></person-group><article-title>Advances in artificial intelligence for infectious-disease surveillance</article-title><source>N Engl J Med</source><year>2023</year><month>04</month><day>27</day><volume>388</volume><issue>17</issue><fpage>1597</fpage><lpage>1607</lpage><pub-id pub-id-type="doi">10.1056/NEJMra2119215</pub-id><pub-id pub-id-type="medline">37099342</pub-id></nlm-citation></ref><ref id="ref12"><label>12</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Kim</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Genevriere</surname><given-names>E</given-names> </name><name name-style="western"><surname>Harker</surname><given-names>P</given-names> </name><etal/></person-group><article-title>Telerobotic neurovascular interventions with magnetic manipulation</article-title><source>Sci Robot</source><year>2022</year><month>04</month><day>13</day><volume>7</volume><issue>65</issue><fpage>eabg9907</fpage><pub-id pub-id-type="doi">10.1126/scirobotics.abg9907</pub-id><pub-id pub-id-type="medline">35417201</pub-id></nlm-citation></ref><ref id="ref13"><label>13</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lapidus</surname><given-names>D</given-names> </name></person-group><article-title>Strengths and limitations of new artificial intelligence tool for rare disease epidemiology</article-title><source>J Transl Med</source><year>2023</year><month>04</month><day>30</day><volume>21</volume><issue>1</issue><fpage>292</fpage><pub-id pub-id-type="doi">10.1186/s12967-023-04152-0</pub-id><pub-id pub-id-type="medline">37122037</pub-id></nlm-citation></ref><ref id="ref14"><label>14</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Saeidi</surname><given-names>H</given-names> </name><name name-style="western"><surname>Opfermann</surname><given-names>JD</given-names> </name><name name-style="western"><surname>Kam</surname><given-names>M</given-names> </name><etal/></person-group><article-title>Autonomous robotic laparoscopic surgery for intestinal anastomosis</article-title><source>Sci Robot</source><year>2022</year><month>01</month><day>26</day><volume>7</volume><issue>62</issue><fpage>eabj2908</fpage><pub-id pub-id-type="doi">10.1126/scirobotics.abj2908</pub-id><pub-id pub-id-type="medline">35080901</pub-id></nlm-citation></ref><ref id="ref15"><label>15</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Monteiro</surname><given-names>M</given-names> </name><name name-style="western"><surname>Newcombe</surname><given-names>VFJ</given-names> </name><name name-style="western"><surname>Mathieu</surname><given-names>F</given-names> </name><etal/></person-group><article-title>Multiclass semantic segmentation and quantification of traumatic brain injury lesions on head CT using deep learning: an algorithm development and multicentre validation study</article-title><source>Lancet Digit Health</source><year>2020</year><month>06</month><volume>2</volume><issue>6</issue><fpage>e314</fpage><lpage>e322</lpage><pub-id pub-id-type="doi">10.1016/S2589-7500(20)30085-6</pub-id><pub-id pub-id-type="medline">33328125</pub-id></nlm-citation></ref><ref id="ref16"><label>16</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Guo</surname><given-names>Y</given-names> </name><name name-style="western"><surname>He</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Lyu</surname><given-names>J</given-names> </name><etal/></person-group><article-title>Deep learning with weak annotation from diagnosis reports for detection of multiple head disorders: a prospective, multicentre study</article-title><source>Lancet Digit Health</source><year>2022</year><month>08</month><volume>4</volume><issue>8</issue><fpage>e584</fpage><lpage>e593</lpage><pub-id pub-id-type="doi">10.1016/S2589-7500(22)00090-5</pub-id><pub-id pub-id-type="medline">35725824</pub-id></nlm-citation></ref><ref id="ref17"><label>17</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Gotbaum</surname><given-names>R</given-names> </name><name name-style="western"><surname>Poterucha</surname><given-names>T</given-names> </name></person-group><article-title>Is medicine ready for AI?</article-title><source>N Engl J Med</source><year>2023</year><month>04</month><day>6</day><volume>388</volume><issue>14</issue><fpage>e49</fpage><pub-id pub-id-type="doi">10.1056/NEJMp2301939</pub-id></nlm-citation></ref><ref id="ref18"><label>18</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Martel</surname><given-names>S</given-names> </name></person-group><article-title>Beyond imaging: macro- and microscale medical robots actuated by clinical MRI scanners</article-title><source>Sci Robot</source><year>2017</year><month>02</month><day>15</day><volume>2</volume><issue>3</issue><fpage>eaam8119</fpage><pub-id pub-id-type="doi">10.1126/scirobotics.aam8119</pub-id><pub-id pub-id-type="medline">33157863</pub-id></nlm-citation></ref><ref id="ref19"><label>19</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Loftus</surname><given-names>TJ</given-names> </name><name name-style="western"><surname>Altieri</surname><given-names>MS</given-names> </name><name name-style="western"><surname>Balch</surname><given-names>JA</given-names> </name><etal/></person-group><article-title>Artificial intelligence-enabled decision support in surgery: state-of-the-art and future directions</article-title><source>Ann Surg</source><year>2023</year><month>07</month><day>1</day><volume>278</volume><issue>1</issue><fpage>51</fpage><lpage>58</lpage><pub-id pub-id-type="doi">10.1097/SLA.0000000000005853</pub-id><pub-id pub-id-type="medline">36942574</pub-id></nlm-citation></ref><ref id="ref20"><label>20</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Graham</surname><given-names>F</given-names> </name></person-group><article-title>Daily briefing: a new Turing test for artificial intelligence</article-title><source>Nature New Biol</source><year>2023</year><pub-id pub-id-type="doi">10.1038/d41586-023-02439-2</pub-id></nlm-citation></ref><ref id="ref21"><label>21</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Rajpurkar</surname><given-names>P</given-names> </name><name name-style="western"><surname>Chen</surname><given-names>E</given-names> </name><name name-style="western"><surname>Banerjee</surname><given-names>O</given-names> </name><name name-style="western"><surname>Topol</surname><given-names>EJ</given-names> </name></person-group><article-title>AI in health and medicine</article-title><source>Nat Med</source><year>2022</year><month>01</month><volume>28</volume><issue>1</issue><fpage>31</fpage><lpage>38</lpage><pub-id pub-id-type="doi">10.1038/s41591-021-01614-0</pub-id><pub-id pub-id-type="medline">35058619</pub-id></nlm-citation></ref><ref id="ref22"><label>22</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>He</surname><given-names>J</given-names> </name><name name-style="western"><surname>Baxter</surname><given-names>SL</given-names> </name><name name-style="western"><surname>Xu</surname><given-names>J</given-names> </name><name name-style="western"><surname>Xu</surname><given-names>J</given-names> </name><name name-style="western"><surname>Zhou</surname><given-names>X</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>K</given-names> </name></person-group><article-title>The practical implementation of artificial intelligence technologies in medicine</article-title><source>Nat Med</source><year>2019</year><month>01</month><volume>25</volume><issue>1</issue><fpage>30</fpage><lpage>36</lpage><pub-id pub-id-type="doi">10.1038/s41591-018-0307-0</pub-id><pub-id pub-id-type="medline">30617336</pub-id></nlm-citation></ref><ref id="ref23"><label>23</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lee</surname><given-names>H</given-names> </name><name name-style="western"><surname>Yune</surname><given-names>S</given-names> </name><name name-style="western"><surname>Mansouri</surname><given-names>M</given-names> </name><etal/></person-group><article-title>An explainable deep-learning algorithm for the detection of acute intracranial haemorrhage from small datasets</article-title><source>Nat Biomed Eng</source><year>2019</year><month>03</month><volume>3</volume><issue>3</issue><fpage>173</fpage><lpage>182</lpage><pub-id pub-id-type="doi">10.1038/s41551-018-0324-9</pub-id><pub-id pub-id-type="medline">30948806</pub-id></nlm-citation></ref><ref id="ref24"><label>24</label><nlm-citation citation-type="web"><article-title>CYBUS123456/HICH-IT-datasets</article-title><source>GitHub</source><access-date>2025-06-06</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://github.com/CYBUS123456/HICH-IT-Datasets">https://github.com/CYBUS123456/HICH-IT-Datasets</ext-link></comment></nlm-citation></ref><ref id="ref25"><label>25</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Hanley</surname><given-names>DF</given-names>  <suffix>Jr</suffix></name></person-group><article-title>Intracerebral haemorrhage: prognostic scales versus clinical judgment in ICH</article-title><source>Nat Rev Neurol</source><year>2016</year><month>04</month><volume>12</volume><issue>4</issue><fpage>192</fpage><lpage>193</lpage><pub-id pub-id-type="doi">10.1038/nrneurol.2016.11</pub-id><pub-id pub-id-type="medline">26868679</pub-id></nlm-citation></ref><ref id="ref26"><label>26</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Brown</surname><given-names>DL</given-names> </name><name name-style="western"><surname>Morgenstern</surname><given-names>LB</given-names> </name></person-group><article-title>Stopping the bleeding in intracerebral hemorrhage</article-title><source>N Engl J Med</source><year>2005</year><month>02</month><day>24</day><volume>352</volume><issue>8</issue><fpage>828</fpage><lpage>830</lpage><pub-id pub-id-type="doi">10.1056/NEJMe048362</pub-id><pub-id pub-id-type="medline">15728817</pub-id></nlm-citation></ref><ref id="ref27"><label>27</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Siebe</surname><given-names>C</given-names> </name></person-group><article-title>Rapid blood-pressure lowering in patients with acute intracerebral hemorrhage</article-title><source>J Emerg Med</source><year>2013</year><month>12</month><volume>45</volume><issue>6</issue><fpage>979</fpage><pub-id pub-id-type="doi">10.1016/j.jemermed.2013.10.021</pub-id></nlm-citation></ref><ref id="ref28"><label>28</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Steiner</surname><given-names>T</given-names> </name><name name-style="western"><surname>Al-Shahi Salman</surname><given-names>R</given-names> </name><name name-style="western"><surname>Beer</surname><given-names>R</given-names> </name><etal/></person-group><article-title>European Stroke Organisation (ESO) guidelines for the management of spontaneous intracerebral hemorrhage</article-title><source>Int J Stroke</source><year>2014</year><month>10</month><volume>9</volume><issue>7</issue><fpage>840</fpage><lpage>855</lpage><pub-id pub-id-type="doi">10.1111/ijs.12309</pub-id><pub-id pub-id-type="medline">25156220</pub-id></nlm-citation></ref><ref id="ref29"><label>29</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Greenberg</surname><given-names>SM</given-names> </name><name name-style="western"><surname>Ziai</surname><given-names>WC</given-names> </name><name name-style="western"><surname>Cordonnier</surname><given-names>C</given-names> </name><etal/></person-group><article-title>2022 guideline for the management of patients with spontaneous intracerebral hemorrhage: a guideline from the American Heart Association/American Stroke Association</article-title><source>Stroke</source><year>2022</year><month>07</month><volume>53</volume><issue>7</issue><fpage>e282</fpage><lpage>e361</lpage><pub-id pub-id-type="doi">10.1161/STR.0000000000000407</pub-id><pub-id pub-id-type="medline">35579034</pub-id></nlm-citation></ref><ref id="ref30"><label>30</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Yu</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Tao</surname><given-names>C</given-names> </name><name name-style="western"><surname>Xiao</surname><given-names>A</given-names> </name><etal/></person-group><article-title>Chinese multidisciplinary guideline for management of hypertensive intracerebral hemorrhage</article-title><source>Chin Med J (Engl)</source><year>2022</year><month>10</month><day>5</day><volume>135</volume><issue>19</issue><fpage>2269</fpage><lpage>2271</lpage><pub-id pub-id-type="doi">10.1097/CM9.0000000000001976</pub-id><pub-id pub-id-type="medline">36315009</pub-id></nlm-citation></ref><ref id="ref31"><label>31</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Li</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Li</surname><given-names>Z</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>K</given-names> </name><name name-style="western"><surname>Dan</surname><given-names>R</given-names> </name><name name-style="western"><surname>Jiang</surname><given-names>S</given-names> </name><name name-style="western"><surname>Zhang</surname><given-names>Y</given-names> </name></person-group><article-title>ChatDoctor: a medical chat model fine-tuned on a large language model meta-AI (LLaMA) using medical domain knowledge</article-title><source>Cureus</source><year>2023</year><month>06</month><volume>15</volume><issue>6</issue><fpage>e40895</fpage><pub-id pub-id-type="doi">10.7759/cureus.40895</pub-id><pub-id pub-id-type="medline">37492832</pub-id></nlm-citation></ref><ref id="ref32"><label>32</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Sutton</surname><given-names>RT</given-names> </name><name name-style="western"><surname>Pincock</surname><given-names>D</given-names> </name><name name-style="western"><surname>Baumgart</surname><given-names>DC</given-names> </name><name name-style="western"><surname>Sadowski</surname><given-names>DC</given-names> </name><name name-style="western"><surname>Fedorak</surname><given-names>RN</given-names> </name><name name-style="western"><surname>Kroeker</surname><given-names>KI</given-names> </name></person-group><article-title>An overview of clinical decision support systems: benefits, risks, and strategies for success</article-title><source>NPJ Digit Med</source><year>2020</year><volume>3</volume><fpage>17</fpage><pub-id pub-id-type="doi">10.1038/s41746-020-0221-y</pub-id><pub-id pub-id-type="medline">32047862</pub-id></nlm-citation></ref><ref id="ref33"><label>33</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Lakshminarayan</surname><given-names>K</given-names> </name><name name-style="western"><surname>Rostambeigi</surname><given-names>N</given-names> </name><name name-style="western"><surname>Fuller</surname><given-names>CC</given-names> </name><name name-style="western"><surname>Peacock</surname><given-names>JM</given-names> </name><name name-style="western"><surname>Tsai</surname><given-names>AW</given-names> </name></person-group><article-title>Impact of an electronic medical record-based clinical decision support tool for dysphagia screening on care quality</article-title><source>Stroke</source><year>2012</year><month>12</month><volume>43</volume><issue>12</issue><fpage>3399</fpage><lpage>3401</lpage><pub-id pub-id-type="doi">10.1161/STROKEAHA.112.662536</pub-id><pub-id pub-id-type="medline">23033346</pub-id></nlm-citation></ref><ref id="ref34"><label>34</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Deng</surname><given-names>B</given-names> </name><name name-style="western"><surname>Zhu</surname><given-names>W</given-names> </name><name name-style="western"><surname>Sun</surname><given-names>X</given-names> </name><etal/></person-group><article-title>Development and validation of an automatic system for intracerebral hemorrhage medical text recognition and treatment plan output</article-title><source>Front Aging Neurosci</source><year>2022</year><volume>14</volume><fpage>798132</fpage><pub-id pub-id-type="doi">10.3389/fnagi.2022.798132</pub-id><pub-id pub-id-type="medline">35462698</pub-id></nlm-citation></ref></ref-list><app-group><supplementary-material id="app1"><label>Multimedia Appendix 1</label><p>The annotated keywords in the electronic medical records of different hypertensive intracerebral hemorrhage cases.</p><media xlink:href="jmir_v27i1e66055_app1.docx" xlink:title="DOCX File, 18 KB"/></supplementary-material><supplementary-material id="app2"><label>Multimedia Appendix 2</label><p>Treatment plans: gold standard, ChatGPT, and the H-SYSTEM.</p><media xlink:href="jmir_v27i1e66055_app2.docx" xlink:title="DOCX File, 25 KB"/></supplementary-material><supplementary-material id="app3"><label>Multimedia Appendix 3</label><p>Supplemental methods and results.</p><media xlink:href="jmir_v27i1e66055_app3.docx" xlink:title="DOCX File, 404 KB"/></supplementary-material></app-group></back></article>