Abstract
Accurately predicting the prognosis of ischemic stroke patients after discharge is crucial for physicians to plan for long-term health care. Although previous studies have demonstrated that machine learning (ML) shows reasonably accurate stroke outcome predictions with limited datasets, to identify specific clinical features associated with prognosis changes after stroke that could aid physicians and patients in devising improved recovery care plans have been challenging. This study aimed to overcome these gaps by utilizing a large national stroke registry database to assess various prediction models that estimate how patients’ prognosis changes over time with associated clinical factors. To properly evaluate the best predictive approaches currently available and avoid prejudice, this study employed three different prognosis prediction models including a statistical logistic regression model, commonly used clinical-based scores, and a latest high-performance ML-based XGBoost model. The study revealed that the XGBoost model outperformed other two traditional models, achieving an AUROC of 0.929 in predicting the prognosis changes of stroke patients followed for 3 months. In addition, the XGBoost model maintained remarkably high precision even when using only selected 20 most relevant clinical features compared to full clinical datasets used in the study. These selected features closely correlated with significant changes in clinical outcomes for stroke patients and showed to be effective for predicting prognosis changes after discharge, allowing physicians to make optimal decisions regarding their patients’ recovery.
Graphical Abstract
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
Stroke is the second leading cause of mortality in the global population and is also a significant burden within the global health community [1]. Many patients are unable to promptly receive necessary treatments, leading to post-discharge disability. This critically impacts their quality of life with excessive burdens on both their families and society.
Recent studies have demonstrated that machine learning (ML) is a powerful tool for predicting stroke functional outcomes [2,3,4]. For instance, one study explored various ML models to predict outcomes for patients in a national stroke registry with their best model achieved an area under the receiver operating characteristic curve (AUROC) of 0.94 for both ischemic and hemorrhagic stroke patients using admission assessment, treatment, and inpatient data. The study also found a strong association between the prediction results and the future recovery and health conditions of the stroke patients [3]. Fernandez-Lozano et al. utilized a random forest-based outcome prediction model to forecast the prognosis of ischemic stroke patients with pre-hospital and hospitalization data, along with 2-day follow-up data, to project the patients’ 30-day prognosis outcomes. The study achieved AUROC values of 0.90 and 0.75 for predicting patients’ mortality and morbidity, respectively [5].
Recently, Monteiro et al. utilized logistic regression (LR), decision tree (DT), support vector machine (SVM), and extreme gradient boosting (XGboost) models to predict the modified Rankin Scale (mRS) at the 3-month follow-up. The ML approach outperformed models based on clinical scores, such as the Acute Stroke Registry and Analysis of Lausanne (ASTRAL) [6]. The superb performance of the XGBoost model has attracted many ML researchers, such as latest studies reported by Chen, et al. [7], Price, et al. [8], and again Moore, et al. [9] that further demonstrated the XGBoost model for its exceptional performance in various clinical cohorts and studies in terms of both speed and accuracy for clinical applications.
Although previous ML-based studies have demonstrated a reasonable level of accuracy in predicting stroke outcomes, they have not identified any associated clinical features that could assist physicians and patients in developing improved recovery care plans. In addition, a recent review conducted by Wang et al. [10] found that many ML-based studies on stroke outcome prediction utilized small sample populations only ranging from 70 to 3184 patients. These studies also considered a limited number of clinical features, ranging from 4 to 152, respectively, which may produce biased predictions and conclusions. As far as our knowledge extends, no studies have been reported that employed prediction models specifically focusing on analyzing clinical factors that may influence the prognosis of ischemic stroke patients within a large cohort, with particular emphasis on changes in functional outcomes after discharge which is crucial to the stroke recovery and toward the development of clinical applications.
In this study, we aimed to establish a predictive model that accurately predicts patient’s prognosis changes and identifies associated key clinical factors that may potentially impact changes in subsequent functional outcomes following patient discharge. To achieve this, we utilized the Taiwan Stroke Registry (TSR), a large nationwide multicenter stroke registry database, with a substantial number of data variables [11]. to investigate various prognosis prediction models with associated clinical factors affecting prognoses of ischemic stroke patients at 1-month and 3-month follow-ups after discharge. To avoid prejudice in conducting a comprehensive evaluation of different predictive models based on changes in functional outcomes with associated clinical features, we examined and utilized both traditional and the latest ML-based approaches for comparison. These approaches included the latest proven high-performance machine learning-based XGboost model, a traditional statistical model based on logistic regression [12], as well as commonly used clinical score-based models that incorporate patient age and the NIH Stroke Scale Index (a.k.a. SPAN-100 index) [13], and the totaled health risks in vascular events (THRIVE) [14]. Both SPAN-100 index and THRIVE are commonly used clinical feature-based scores by physicians and health researcher to associate and predict clinical response, outcome, mortality, and risk of in ischemic stroke patients.
2 Key characteristics of functional outcome changes of ischemic stroke patients after discharge in the TSR database
To explore and illustrate the trend of ischemic stroke patient’s functional outcome measures, a flow diagram of TSR patients was created from over 60,000 ischemic stroke patients’ modified Rankin Scale (mRS) at discharge, 1-month and 3-month follow-ups from 2006 to 2020. The diagram is shown in Fig. 1. In this study, the good outcome was defined as mRS score < 3 and the poor outcome as mRS score ≥ 3 as commonly accepted by the stroke research community [15]. Figure 1 demonstrated that a majority of patients remained at the same outcome status during the entire observation period (up to 3 months), while only 8.33% of total cases exhibited a change in their functional outcome at 1-month, and 7.33% of cases changed at 3-month follow-up. In addition, four distinct types of outcome changes were noticed from discharge to 1-month follow-up, with different alterations of outcome changes were observed between 1- and 3-month follow-ups. For example, in Fig. 1, the dark blue color described fewer cases transitioning from good to poor outcome from discharge to 1-month follow-up (denoted as Gd → P1m) with most cases remaining at a poor outcome status between 1- and 3-month follow-up (Gd → P1m → P3m), however, with about 30% of cases improved back to a better health condition at 3-month follow-up (Gd → P1m → G3m). Moreover, there were significant poor outcome cases (6.37% of total studied population) at discharge that turned into good outcome (Pd → G1m) at 1-month follow-up with the majority of cases remaining at the same health condition. This is the first time a detailed flow diagram was presented that clearly demonstrated the case distributions of outcome changes from discharge to 3-month follow-up from a large population-based stroke registry. The objective of this study was to identify the best prediction model for the outcome change prognosis at 1-month and 3-month follow-ups with associated clinical features that may promote good outcome changes for stroke patients after discharge. The study also aimed to identify clinical risk factors that may trigger poor or promote good outcome changes at long-term follow-ups. The findings of this study will help physicians better understand and plan for their patient recovery and health care management after stroke.
3 Materials and methods
3.1 Ethical approval
The TSR database encrypts personal information of the patients to protect privacy and provides researchers with anonymous identification numbers associated with relevant claims information, including sex, date of birth, medical services received, and prescriptions. This study was approved to fulfill the condition for exemption by the Institutional Review Board (IRB) of China Medical University, Taichung, Taiwan (CMUH102-REC1-086(CR-7) that waived the consent requirement. All investigations employed in this study were performed in accordance with the relevant clinical research guidelines and regulations.
3.2 Patient population and cohort selections for the study
The datasets supporting the investigation and findings of this study are available upon requests from the corresponding authors. For the development and evaluation of predictive models, 155,030 stroke patient records from TSR with 531 clinical variables recorded between 2006 and 2020 were retrospectively selected. TSR mainly collected stroke patients’ admission and assessment data, clinical assessments during hospitalization, functional outcome measurements, in-hospital complications, medical history, laboratory test results, electrocardiography, computed tomography findings, magnetic resonance imaging (MRI) findings, medications during admission, functional outcome measures, discharge status, and follow-up information for retrospective research use. As presented in Fig. 2, to select represented high quality datasets for this study, 64,379 cases without any available follow-up information and 97 duplicated records were processed and removed. The remaining 90,554 records were employed for the further screening of 1- and 3-month cohort sets. For the 1-month cohort dataset, 82,413 cases were kept by removing 4418 patients who died before discharge and 3723 patients without a 1-month follow-up mRS record. In addition, 135 patients less than 18 years old, and 13,083 with other identified stroke diagnosis (e.g., hemorrhage stroke) were removed, leaving 69,195 adult patients being selected for 1-month follow-up cohort. For the 3-month cohort dataset, similar selection and extraction steps were performed to delete those lack of 1- and 3-month mRS records that left with 3723 and 9965 cases. After dropping 116 non-adult cases and 10,662 cases diagnosed as other strokes, 61,128 ischemic stroke patients were retained as the 3-month follow-up dataset. Those collected adult cases with ischemic stroke would then be applied with the data preprocessing for missing values and amputation steps (see Supplementary Fig. S1). After data preprocessing steps were performed, two high quality 1- and 3-month follow-up ischemic stroke cohorts with 46,198 and 41,604 subjects, respectively, were used in the study.
3.3 Definitions of the prediction targets
The prediction target of this study was to determine whether each case’s 1- and 3-month follow-up mRS conditions (good or poor outcome) were changed from that of at discharge. The study divided patients into two condition subgroups, good and poor outcomes at discharge. For the good outcome (mRS < 3) at discharge group, the patient’s health conditions could remain unchanged or become deteriorated (i.e., poor outcome) at 1-month follow-up. Similarly, the patients in poor outcome (mRS ≥ 3) at discharge group could remain as unchanged or improved. Similarly, the 3-month outcome groups followed the same definitions as those of at 1-month follow-up. The prediction models were trained to predict changes in two condition groups’ functional outcomes and validated for such models’ predictive power at 1- and 3-month, respectively. The good outcome change from discharge to 1-month follow-up dataset was denoted as Gd → 1 m and the poor outcome change from discharge to 1-month follow-up dataset was denoted as Pd → 1 m. Similarly, the good outcome change from discharge to 3-month follow-up dataset was denoted as Gd → 3 m and the poor outcome change from discharge to 3-month follow-up dataset was denoted as Pd → 3 m.
3.4 Predictive models and cross validation
The predictive model building steps with cross validation processes are demonstrated in Fig. 3. Based on the admission years, the TSR cohort data sets were designed and split into three sets: training (2006–2011), validation (2012–2013), and test (2014–2020) datasets. The cross-validation based on time series splits mimics and simulates the “real world” clinical forecasting environment, that is, using prior years’ populations to predict future stroke patient’s outcome changes. Training and validation datasets were used in each cross-validation round for model evaluation and feature selections. As shown in Fig. 1, the distributions of outcome changes were largely unbalanced; thus, Tomek Links[16, 17], a known under-sampling method, was applied on majority distributions to tackle this challenge before building the prediction models. For model fine tuning steps, hyperparameter optimization with randomized search [18] and probability calibration with isotonic regression [19] were applied. In each cross-validation round, individual model was evaluated by the validation dataset for the best model selection.
To compare and validate predictability obtained from each built model, this study employed two well-validated and simple-to-use clinical outcome prediction tools, the SPAN-100 index and the THRIVE score, for ischemic stroke with test datasets as the baseline performance [20,21,22,23]. The SPAN-100 index is the sum of age (in years) and admission National Institutes of Health Stroke Scale (NIHSS) (0–42) with a decision split point of 100. The THRIVE score employs a 0 to 9 point scale that consists of age, admission NIHSS, and medical history such as hypertension (HT), diabetes mellitus (DM), and atrial fibrillation (AF). The THRIVE total score ranges from 0 to 9 with some previously reported studies split the score into three groups (0–2, 3–5, and 6–9) [14, 20, 24,25,26,27], and others split the score into two groups (0–5 or 6–9) [28]. This study split the THRIVE score into two subgroups by two different cutoff points to examine good or poor outcome subgroups based on mRS [29]. One was THRIVE-3 (whether greater or equal to three) and the other was THRIVE-6 (whether greater or equal to six) for subsequent analyses. Taking the THRIVE-6 as an example, the predicted outcome is poor when the THRIVE is no less than 6, while patients with THRIVE score less than 6 are anticipated to have good outcome at 1- and 3-month follow-ups. Subsequently, according to the definition of the prediction targets, 1- and 3-month outcome’s labels were converted into unchanged and changed (either improved or deteriorated) classes for model evaluations and comparisons.
For the statistical analysis model, differences within two cohort (1-month and 3-month) sets’ training, validation, and test datasets were investigated by standardized mean difference (SMD). In this study, an SMD of 0 indicates no inconsistencies among the subsets, whereas that of over 0.2, 0.5, and 0.8 represent small, medium, and large disparity, respectively [30]. Model performance was evaluated using well-accepted AUROC, specificity, sensitivity, and positive predictive value (PPV). If the predictions were not derived from probabilities, the AUROC was acquired from a confusion matrix which is \(\frac{1}{2}\left[\left(\frac{{\text{tp}}}{{\text{tp}}+{\text{fn}}}\right)+\left(\frac{{\text{tn}}}{{\text{tn}}+{\text{fp}}}\right)\right]\); Specificity is \(\frac{{\text{tn}}}{{\text{tn}}+{\text{fp}}}\). Sensitivity is \(\frac{{\text{tp}}}{{\text{tp}}+{\text{fn}}}\); PPV is \(\frac{{\text{Sensitivity}}\times \mathrm{ Prevalence}}{{\text{Sensitivity}}*{\text{Specificity}}+\left(1-{\text{Specificity}}\right)\times (1-{\text{Prevalence}})}\) where \({\text{prevalence}}= \frac{{\text{tp}}+{\text{fn}}}{{\text{tp}}+{\text{fp}}+{\text{fn}}+{\text{tn}}}\)[31]. Note that tp was a case that truly changed the status, fp was a case that not truly changed the status, fn was a case that not truly remained the status, and tn was a case that truly remained the status. Statistical analysis was performed with R tableone package (version 0.12.0) and Python sklearn package (version 0.24.2).
Feature selection is a crucial step in identifying a small subset of input features that can still achieve reasonable predictive performance. Reducing the number of required feature inputs makes prediction models more practical since it is often difficult to collect hundreds of input features during stroke triage without missing or making errors in real-world clinical practices. In the XGboost model used in this study, feature importance was computed from Gini impurity: \(1- \sum_{i=1}^{j}{p}_{i}^{2}\), where \(j\) is the number of classes, \({p}_{i}\) is the proportion of samples labeled with class i. All features were ranked in descending order by their importance. Two types of thresholds were adopted to select the most important features. One was to extract a specific number of the top influential features, while the other was to establish \({{\text{threshold}}}_{{\text{min}}}={\text{minimum}}\left(\upsigma \right)+\mathrm{standard deviation}\left(\sigma \right)\)[32] where \(\sigma\) is the list of feature importance in which the zeros were removed. The selected features were further used to re-train a pruned XGboost model (see Fig. 3).
4 Results
4.1 Patient population characteristics
The study included 46,198 and 41,604 ischemic stroke patients from TSR with respective 1-month and 3-month follow-up information as described in the “Materials and methods”. These two cohorts were further divided into training, validation, and test datasets for further investigations. The summary of patient characteristics of three datasets from two cohorts based on their case numbers, age, sex, NIHSS, and mRS as well as their corresponding outcomes is shown in Table 1 with no significant variations across datasets. However, the authors did notice subtle but important observations in each cohort and individual dataset that may provide additional insights on the prediction results. For example, male population is about 60% of ischemic stroke cases in TSR with mean age of about 69. The mean mRS at 1-month follow-up was found about 0.13 – 0.18 scale lower than that at discharge (i.e., improvement) versus 0.27 – 0.34 lower at 3-month follow-up. These findings indicate overall positive trends with improvement of outcomes 90 days after discharge. The same positive trends were reflected in the numbers of cases with good outcome changes at 3-month follow-up (e.g., more than 50% of patient population in each dataset).
4.2 Comparison of the models for the prediction of prognosis changes
As shown in Table 2, the AUROCs of clinical scores model, statistical analysis model, and ML-based XGboost model were obtained from two different cohort datasets with two discharge outcome conditions (i.e., good vs poor outcomes). The ML-based XGboost model significantly outperformed both statistical and clinical score-based models. The ML model’s performance had higher predictability in poor outcome subgroups of 1- and 3-month follow-ups (e.g., Pd → 1 m and Pd → 3 m) than those with good outcome changes at discharge subgroups (e.g., Gd → 1 m and Gd → 3 m). In addition, the AUROCs of both logistic regression and XGboost models obtained from the 3-month cohort datasets were found 8–10% better than those of the 1-month cohort datasets. For the clinical score-based model, all AUROCs were merely around 0.5 that is, with nearly no predictability [33]. These results indicated that there were more important clinical features in the clinical score model other than age, NIHSS, HT, DM, or AF combined that contributed to the outcome changes in ischemic stroke patients. Compared to the clinical score model, statistical (logistic regression) and XGboost models outperformed with AUROCs by approximately 20 to 35%, some with AUROCs over 0.9, which implies its high predictably in outcome prognosis. Between the two models, XGboost provided better performance and predictability, which the authors chose to further investigate clinical features relevant to prognosis changes that could be utilized in future clinical applications. The details of each model’s performance other than AUROC, including their sensitivities, specificities, and PPVs, were provided in Supplementary Table S1, S2, and S3 for references.
4.3 Predictive performance with selected features based on the XGboost model
Given that XGboost was the best prediction model, we further explored its predictive performance for selecting a small subset of clinical features at different thresholds that could still maintain its reasonable predictability (i.e., similar AUROCs) compared to that with all clinical features employed. This allowed them to elucidate and assess crucial clinical features that physicians can focus on during the stroke triage while maintaining their confidence in the model predictability. Supplementary Fig. S2 demonstrates the AUROCs of four outcome change subgroups at discharge from two cohorts (1-month vs 3-month) with numbers of selected features ranked by their importance obtained by the XGboost model. The total number of all features used for the 1-month cohort data set was 228, while those of the 3-month cohort data set was 229. The number of clinical features selected at the thresholdmin for Gd → 1 m, Pd → 1 m, Gd → 3 m, and Pd → 3 m subgroups were 6, 8, 8, and 3 respectively. The AUROCs did not change significantly when top 20 features were selected for the prediction model compared to those with all features used. All the AUROCs using the XGboost model with the thresholdmin predictors (i.e., crucial clinical features) showed satisfying predictability from all four subgroup datasets except for the Gd → 1 m. The selected thresholdmin predictors were mostly from the Barthel index assessments which indicated patients’ mobility and functional independence such as transfering and bathing were highly associated with prognosis changes after stroke. In addition, those patients with large artery atherosclerosis, previous cerebrovascular accident, and hypertriglyceridemia history were likely to show improvement over time as evidenced in their good 3-month outcome changes. More selected features with detailed information such as ranking of feature importance and the ROC curves are available in Supplementary Table S4 and Figures S3, S4, S5, and S6).
5 Discussion
The study focused on predicting changes in functional outcomes at two key follow-up time points with associated key clinical features instead of directly predicting their mRS outcome scores. This approach allowed for better clinical applications to assist physicians caring for ischemic stroke patients after discharge. As presented in the study results, the XGboost and LR models were clearly better discriminators than the clinical scores-based model. Between the two, XGboost showed the best performance to investigate key clinical features associated with the prognosis changes. Compared to the clinical scores model with only a few variables collected at admission, the XGboost model employed clinical variables obtained at different time points throughout the triage ranging from admission, assessments, treatment to discharge, and follow-up outcomes. Similar observations and performance were reported by M. Monteiro et al. [6] who also verified that great AUROC could be achieved to some extent as long as features at various time points were added to the analysis.
The study further employed SHapley Additive exPlanations (SHAP) algorithm [34, 35] to elucidate and assess XGboost selected clinical features or risk factors that caused positive and/or negative prognosis changes. SHAP is a unified framework that helps explain the origins of predictions by assigning each predictor (in this case clinical feature) an impact value to a specific prediction significance (i.e., SHAP value). In this study, only the top 20 predictors of each cohort group were selected and shown in summary plots of SHAP value obtained (Fig. 4). The summary plots of four prognosis outcome change subgroups (i.e., Gd → 1 m, Pd → 1 m, Gd → 3 m, and Pd → 3 m) were presented with SHAP value (SV) and feature value (FV). When the SV is greater than 0 (positive), it implies that the follow-up mRS would change; on the other hand, SV less than 0 signifies that the follow-up mRS would remain consistent or unchanged. In this study, the FV is presented as a blue to red gradient with blue, purple, and red representing low, middle, and high FV, respectively, relative to their degrees of importance. Taking age in each follow-up subgroup as an example, the model indicated that older age patients were more likely to alter their outcome status from good to poor in both Gd-1 m and Gd-3 m subgroups but remained in poor outcomes status in both Pd → 1 m and Pd → 3 m subgroups. That is, older age was found to be associated with a high probability of having poor prognosis at the follow-ups regardless of their discharged mRS in TSR population.
This study clearly demonstrated that the presented ML-based XGboost model can be utilized to predict the long-term outcome changes after ischemic stroke. Rather than directly predicting the outcome score, stroke care physicians can now use patients’ outcomes at discharge with our prediction model to evaluate their follow-up mRS changes for a personalized recovery plan. In addition, from the summary plots of SHAP values obtained from 1- and 3-month follow-up subgroups, these important risk factors associated with each prognosis change can be used to better triage patients following each outcome change subgroup. It is important to note that our predictive model should be used in conjunction with physician’s clinical knowledge and judgment, but not as a substitute.
Many studies have previously shown the high predictability of stroke functional outcome at discharge using various ML models. However, as shown in our study with a large and diverse population, most patients were found to have the same outcome status during the entire observation period (i.e., up to 3-month follow-up); thus, functional outcome alone could not be predicted reliably. In this study, we demonstrated the changes in functional outcomes after discharge as a better prognosis indicator and compared their predictabilities and performance with different computational approaches (e.g., clinical score-based model, statistical LR model, vs XGboost ML model). We utilized the best prediction model to ascertain clinical features crucial to the recovery and care management after stroke based on a large nationwide multi-center stroke registry database. Compared with clinical scores or simple statistical LR models, our predictive model based on XGboost ML algorithm had significantly higher performance in predicting ischemic stroke patient’s outcome changes. In addition, selected few key clinical features such as age, functional independence, and mobility were shown to have profound impacts on the recovery of ischemic stroke patients.
In summary, this study allows physicians to use the predicted results in their clinical practices for planning an optimal personalized care plan aiming for improved recovery. Future studies will focus on building an effective prediction tool to help clinicians select an optimal treatment plan with positive prognosis changes for stroke patients and to prevent any potential deterioration in their future outcomes after discharge. This study has several limitations that should be noted. The majority of the TSRs used in this study were from Asian patients; thus, the model’s performance for other populations is uncertain. Patients with missing data may have recovered to a point where follow-up is deemed unnecessary or could have been relocated to a nursing home due to poor functional status, rendering them unable to attend follow-up appointments.
References
WHO (2020) The top 10 causes of death. https://www.who.int/news-room/fact-sheets/detail/the-top-10-causes-of-death. Accessed 3 Nov 2023
Xu Y et al (2019) Extreme gradient boosting model has a better performance in predicting the risk of 90-day readmissions in patients with ischaemic stroke. J Stroke Cerebrovasc Dis 28(12):104441
Lin C-H et al (2020) Evaluation of machine learning methods to stroke outcome prediction using a nationwide disease registry. Comput Methods Programs Biomed 190:105381
Van Os HJ et al (2018) Predicting outcome of endovascular treatment for acute ischemic stroke: potential value of machine learning algorithms. Front Neurol 9:784
Fernandez-Lozano C et al (2021) Random forest-based prediction of stroke outcome. Sci Rep 11(1):10071. https://doi.org/10.1038/s41598-021-89434-7
Monteiro M et al (2018) Using machine learning to improve the prediction of functional outcome in ischemic stroke patients. IEEE/ACM Trans Comput Biol Bioinf 15(6):1953–1959
Chen R et al (2023) A study on predicting the length of hospital stay for Chinese patients with ischemic stroke based on the XGBoost algorithm. BMC Med Inform Decis Mak 23(1):1–10
Price J, Yamazaki T, Fujihara K, Sone H (2022) XGBoost: interpretable machine learning approach in medicine. In: 2022 5th World Symposium on Communication Engineering (WSCE). IEEE, pp 109–113
Moore A, Bell M (2022) XGBoost, a novel explainable ai technique, in the prediction of myocardial infarction: a UK biobank cohort study. Clin Med Insights: Cardiol 16:11795468221133612
Wang W et al (2020) A systematic review of machine learning models for predicting outcomes of stroke with structured data. PLoS ONE 15(6):e0234722
Hsieh C-Y, Wu DP, Sung S-F (2018) Registry-based stroke research in Taiwan: past and future. Epidemiol Health 40
Hosmer JrDW, Lemeshow S, Sturdivant RX (2013) Applied logistic regression. John Wiley & Sons
Saposnik G, Guzik AK, Reeves M, Ovbiagele B, Johnston SC (2013) Stroke prognostication using age and NIH stroke scale: SPAN-100. Neurology 80(1):21–28
Flint AC, Cullen SP, Faigeles BS, Rao VA (2010) Predicting long-term outcome after endovascular stroke treatment: the totaled health risks in vascular events score. AJNR Am J Neuroradiol 31(7):1192–1196. https://doi.org/10.3174/ajnr.A2050. (in eng)
Saver JL et al (2016) Time to treatment with endovascular thrombectomy and outcomes from ischemic stroke: a meta-analysis. JAMA 316(12):1279–1289
Tomek I (1976) Two modifications of CNN. IEEE Trans Syst Man Cybern
Viadinugroho RAA Imbalanced classification in python: SMOTE-Tomek Links method combining SMOTE with Tomek Links for imbalanced classification in python. https://towardsdatascience.com/imbalanced-classification-inpython-smote-tomek-links-method-6e48dfe69bbc. Accessed 3 Nov 2023
Bergstra J, Bengio Y (2012) Random search for hyper-parameter optimization. J Mach Learn Res 13(2)
Zadrozny B, Elkan C (2001) Obtaining calibrated probability estimates from decision trees and naive bayesian classifiers. In: Icml, pp 609–616
Flint AC et al (2013) THRIVE score predicts ischemic stroke outcomes and thrombolytic hemorrhage risk in VISTA. Stroke 44(12):3365–3369. https://doi.org/10.1161/strokeaha.113.002794. (in eng)
Matsumoto K, Nohara Y, Soejima H, Yonehara T, Nakashima N, Kamouchi M (2020) Stroke prognostic scores and data-driven prediction of clinical outcomes after acute ischemic stroke. Stroke 51(5):1477–1483. https://doi.org/10.1161/strokeaha.119.027300. (in eng)
Nishi H et al (2019) Predicting clinical outcomes of large vessel occlusion before mechanical thrombectomy using machine learning. Stroke 50(9):2379–2388
Ospel JM, Brown S, Kappelhof M et al (2021) Comparing the prognostic impact of age and baseline National Institutes Of Health Stroke Scale in acute stroke due to large vessel occlusion. Stroke 52(9):2839–2845. https://doi.org/10.1161/strokeaha.120.032364. (in eng)
Flint AC, Kamel H, Rao VA, Cullen SP, Faigeles BS, Smith WS (2014) Validation of the Totaled Health Risks In Vascular Events (THRIVE) score for outcome prediction in endovascular stroke treatment. Int J Stroke 9(1):32–39. https://doi.org/10.1111/j.1747-4949.2012.00872.x. (in eng)
Flint AC et al (2013) “THRIVE score predicts outcomes with a third-generation endovascular stroke treatment device in the TREVO-2 trial. Stroke 44(12):3370–3375. https://doi.org/10.1161/strokeaha.113.002796. (in eng)
Kamel H et al (2013) The totaled health risks in vascular events (THRIVE) score predicts ischemic stroke outcomes independent of thrombolytic therapy in the NINDS tPA trial. J Stroke Cerebrovasc Dis 22(7):1111–1116. https://doi.org/10.1016/j.jstrokecerebrovasdis.2012.08.017. (in eng)
Lei C et al (2014) Totaled health risks in vascular events score predicts clinical outcomes in patients with cardioembolic and other subtypes of ischemic stroke. Stroke 45(6):1689–1694. https://doi.org/10.1161/strokeaha.113.004352. (in eng)
Chen B et al (2019) Predictive value of the THRIVE score for outcome in patients with acute basilar artery occlusion treated with thrombectomy. Brain and behavior 9(10):e01418. https://doi.org/10.1002/brb3.1418. (in eng)
Drozdowska BA, Singh S, Quinn TJ (2019) Thinking about the future: a review of prognostic scales used in acute stroke. Front Neurol 10:274. https://doi.org/10.3389/fneur.2019.00274. (in eng)
Faraone SV (2008) Interpreting estimates of treatment effects: implications for managed care. Pharm Ther 33(12):700
Sokolova M, Lapalme G (2009) A systematic analysis of performance measures for classification tasks. J Inf Process Manag 45(4):427–437
Kouwaye B (2016) Regression trees and random forest based feature selection for malaria risk exposure prediction. arXiv preprint arXiv:1606.07578
Mandrekar JN (2010) Receiver operating characteristic curve in diagnostic test assessment. J Thorac Oncol 5(9):1315–1316. https://doi.org/10.1097/JTO.0b013e3181ec173d. (in eng)
Antwarg L, Miller RM, Shapira B, Rokach L (2021) Explaining anomalies detected by autoencoders using shapley additive explanations. Expert Syst Appl 186:115736.
Lundberg SM, Lee S-I (2017) A unified approach to interpreting model predictions. In: Proceedings of the 31st International Conference on Neural Information Processing Systems. pp 4768–4777
Acknowledgements
Authors wish to dedicate this work to Professor Chung Y. Hsu who devoted his life to stroke research and led to the creation of TSR. We would like to acknowledge and thank TSR participated PIs who contributed their patient data to the registry (see Supplementary Appendix I for a completed list of TSR Investigators)
Funding
This work is supported by funding from Division of Intramural Research of National Institute of Neurological Disorders and Stroke, National Institutes of Health, USA (YCF); funding from the Chang Gung Memorial Hospital Research Project [grant number CGRPG3L0011] (CHL); funding from the Ministry of Science and Technology, Taiwan [grant numbers MOST 110–2222-E-182A-001-MY2, MOST 110–2321-B-039–003] (CHL); and funding from China Medical University Hospital [grant number DMR-111–105] (KCH).
Author information
Authors and Affiliations
Consortia
Contributions
CHL conceived the idea of the study, implemented the machine learning approaches, and drafted the manuscript. YAC conducted experiments, implemented the machine learning approaches, and drafted the manuscript. JSJ, YS, CYW, PYY, WLC, JTL, and Taiwan Stroke Registry investigators collected or provided datasets. KCH interpreted the result and provided practical suggestion to this study. YCF provided advices and inputs to the study design and manuscript. All authors contributed to the review of the manuscript and approved the final version.
Corresponding authors
Ethics declarations
Conflict of interest
The authors declare no competing interests.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary Information
Below is the link to the electronic supplementary material.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Lin, CH., Chen, YA., Jeng, JS. et al. Predicting ischemic stroke patients’ prognosis changes using machine learning in a nationwide stroke registry. Med Biol Eng Comput 62, 2343–2354 (2024). https://doi.org/10.1007/s11517-024-03073-4
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11517-024-03073-4