AI-based Assessment of Risk Factors for Coronary Heart Disease in Patients With Diabetes Mellitus and Construction of a Prediction Model for a Treatment Regimen

Zhen Gao , Qiyuan Bai , Mingyu Wei , Hao Chen , Yan Yan , Jiahao Mao , Xiangzhi Kong , Yang Yu

Reviews in Cardiovascular Medicine ›› 2025, Vol. 26 ›› Issue (6) : 36293

PDF (22222KB)
Reviews in Cardiovascular Medicine ›› 2025, Vol. 26 ›› Issue (6) :36293 DOI: 10.31083/RCM36293
Original Research
research-article
AI-based Assessment of Risk Factors for Coronary Heart Disease in Patients With Diabetes Mellitus and Construction of a Prediction Model for a Treatment Regimen
Author information +
History +
PDF (22222KB)

Abstract

Background:

This study aimed to construct a prediction model for a treatment plan for patients with coronary artery disease combined with diabetes mellitus using machine learning to efficiently formulate the treatment plan for special patients and improve the prognosis of patients, provide an explanation of the model based on SHapley Additive exPlanation (SHAP), explore the related risk factors, provide a reference for the clinic, and concurrently, to lay the foundation for the establishment of a multicenter prediction model for future treatment plans.

Methods:

To investigate the relationship between concomitant coronary heart disease (CHD) and diabetes mellitus (DM), this study retrospectively included patients who attended the Beijing Anzhen Hospital of Capital Medical University between 2022 and 2023. The processed data were then input into five different algorithms for model construction. The performance of each model was rigorously evaluated using five specific evaluation indicators. The SHAP algorithm also provided clear explanations and visualizations of the model's predictions.

Results:

The optimal set of characteristics determined by the least absolute shrinkage and selection operator (LASSO) regression were 15 features of general information, laboratory test results, and echocardiographic findings. The best model identified was the eXtreme Gradient Boost (XGBoost) model. The interpretation of the model based on the SHAP algorithm suggests that the feature in the XGBoost model that has the greatest impact on the prediction of the results is the glycated hemoglobin level.

Conclusions:

Using machine-learning algorithms, we built a prediction model of a treatment plan for patients with concomitant DM and CHD by integrating patients' information and screened the best feature set containing 15 features, which provides help and strategies to develop the best treatment plan for patients with concomitant DM and CHD.

Graphical abstract

Keywords

coronary heart disease / diabetes mellitus / machine learning / predictive modeling / SHapley Additive exPlanation

Cite this article

Download citation ▾
Zhen Gao, Qiyuan Bai, Mingyu Wei, Hao Chen, Yan Yan, Jiahao Mao, Xiangzhi Kong, Yang Yu. AI-based Assessment of Risk Factors for Coronary Heart Disease in Patients With Diabetes Mellitus and Construction of a Prediction Model for a Treatment Regimen. Reviews in Cardiovascular Medicine, 2025, 26(6): 36293 DOI:10.31083/RCM36293

登录浏览全文

4963

注册一个新账户 忘记密码

1. Introduction

In recent years, the morbidity and mortality rates of coronary heart disease (CHD) have been on the rise, with the age of onset decreasing annually [1, 2, 3, 4, 5, 6]. Meanwhile, diabetes mellitus (DM) has reached epidemic proportions worldwide, and its prevalence is also on the rise [7, 8]. CHD and DM, as two separate pathological entities, can enhance each other’s disease progression, and the mortality rate of patients suffering from both is higher than that of patients with just one [9, 10, 11, 12, 13, 14, 15, 16, 17]. Therefore, the development of treatment regimens for patients with both DM and CHD needs to take into account the common factors and influences of the two diseases [18, 19, 20, 21, 22]. At this time, treatment options for patients with both DM and CHD can be broadly divided into two categories: conservative treatment with medication after glycemic control and surgical treatment (including percutaneous coronary interventions and coronary artery bypass grafting) [23, 24]. Due to human errors and imperfections in examination and testing indicators, many patients are still unable to receive appropriate treatment plans, such that the prognosis and recovery of patients cannot be optimized [25, 26]. In recent years, machine learning has often been used to deal with this kind of data involving magnanimous samples and data mining [27, 28].

Machine learning is dedicated to the study of how computers can simulate or implement human learning behaviors to acquire new knowledge or skills and reorganize existing knowledge structures to continuously enhance their performance [29, 30, 31, 32, 33]. Machine learning can systematically process and classify much clinical data on its own and ultimately obtain information of clinical interest from the system’s output [34, 35], which can help to reveal the essential features of the disease and elucidate the potential correlation between the information of different variables [36, 37, 38, 39, 40, 41, 42, 43, 44]. In recent years, it has been shown to provide useful insights into cardiovascular diseases and has begun to have clinical applications [45, 46, 47, 48, 49, 50, 51].

In this study, we aimed to develop and validate a prediction model for treatment regimens of patients with CHD combined with DM by conducting a retrospective study using a single-center database. We used five machine-learning algorithms for model construction, from which we identified the eXtreme Gradient Boost (XGBoost) algorithm as the best algorithm and used it as a basis for mining the relevant risk factors.

2. Materials and Methods

2.1 Data Source and Study Population

The Coronary Heart Disease Database of the Anzhen Hospital of the Capital Medical University is a platform-type operation and management system for disease resource sharing customized and developed for the Coronary Heart Disease Database platform of the Anzhen Hospital on the basis of the Jiahemeikang Disease Resource Sharing Management System. In this study, we exported 3171 patients diagnosed with coronary heart disease combined with DM from the Coronary Heart Disease Database of the Beijing Anzhen Hospital of the Capital Medical University in 2022 and 2023 and retrospectively included 3153 patients with coronary heart disease combined with DM in the internal cohort after filtering out useless data that did not meet the criteria for nullclassification or with missing features greater than 30 or more items.

Inclusion criteria:

(1) Patients with a clear diagnosis of coronary artery disease (CAD) combined with DM;

(2) After rigorous history taking, important data were complete;

(3) Age 18 years.

Exclusion criteria:

(1) Pregnant patients;

(2) Combination of malignant tumors and long-term use of chemotherapy drugs;

(3) Combination of diseases that can significantly affect routine blood and biochemical indices.

2.2 Data Collection and Preprocessing

Restricted pre-processing of the collected data was performed by coding categorical variables such as heart failure, atrial fibrillation, or cardiogenic shock using 0 and 1, representing that the sample in which 0 did not have this characteristic, and 1 did have this characteristic. In addition, factorization was performed. All data were divided into positive (Group P) and negative (Group N) groups based on the presence or absence of the treatment (i.e., percutaneous stenting and coronary artery bypass grafting). All continuous variables comparing the clinical data of the two groups of patients were described using either x¯ ± S (satisfying normal distribution) or M (Q1, Q3) (not satisfying normal distribution). Categorical variables (count data) were described using percentages and frequencies. Statistical analysis used the R language (R 4.3.2, R Core Team, Vienna, Austria) for subsequent predictive model construction and data visualization.

To make each feature in the results comparable, all the data are first standardized, and all the data was randomly divided in the ratio of 8:2, i.e., 80% of the data were used as the training set and 20% of the data were used as the test set. The screening of features, model construction, and parameter tuning were all done in the training set, and it is guaranteed that data leakage in the test set. Based on the training set data, dummy variables (DVs) were introduced for variables that did not need to be classified and then regressed by the least absolute shrinkage and selection operator (LASSO). To optimize the regularization strength of the LASSO regression model, a grid search was carried out to determine the optimal alpha value.

For missing data, features with greater than 15% missing data were deleted, and features with no more than 15% missing data were inputed into the random forest algorithm. The random forest imputation process consists of the following steps: First, for each feature with a missing value, a random forest regression model was constructed using the other features as inputs [52]. Second, simple statistics (e.g., mean or median) were used to estimate the missing values. The model is then trained in the complete case (samples without missing values) and used to predict missing values, replacing the initial imputations. This process is repeated until the model converges or a predetermined number of iterations is reached to ensure stable input results [53]. If the data are unbalanced samples, the Synthetic Minority Over-Sampling Technique (SMOTE) algorithm is introduced to eliminate the effect of imbalance before splitting the data, where SMOTE generates new synthetic samples in the vicinity of the minority class instances with the aim of enhancing their representativeness [54].

2.3 Model Construction

Five different machine learning algorithms, namely Random Forest (RF), Logistic Regression (LR), XGBoost, Support Vector Machine (SVM), and K-nearest neighbor (KNN), were used in this experiment.

2.3.1 RF

Random forest is an integrated machine learning algorithm that improves the accuracy and robustness of a model by combining the predictions of multiple decision trees. For the classification task, the prediction result of random forest is usually obtained through the majority voting mechanism, that is, each tree gives a prediction result, and the majority category is finally selected as the prediction result of random forest. This process can be expressed as: for the input feature vector x, each tree ti in the random forest gives a prediction result yi (for classification tasks, yi is a category label). The final prediction class y* is the class that maximizes the following expression:

Where, I()is the indicator function, which takes the value 1 when the parenthesis condition is true, otherwise it is 0; k is the number of trees in the random forest; y is the possible category label.

2.3.2 LR

LR belongs to probabilistic nonlinear regression, which is mainly used to study the relationship between the outcome index of binary classification (dependent variable) and some influencing factors (independent variable) (can be extended to multiple categories). It is commonly used in epidemiology to analyze quantitative relationships between diseases and associated risk factors. The LR model can be expressed as:

In the formula, P is the probability when a positive result occurs, β0 is the constant term, β1, β2…, βm is the independent variable regression coefficient of X1, X2, …, Xm. Logarithmic conversion of the formula can be expressed in linear form:

The logitPfor positive results with negative results occur when probability of the natural logarithm and the value oflogitP ranges have no numerical bounds.

2.3.3 XGBoost

XGBoost is an ensemble learning algorithm based on gradient-raising decision trees that optimizes the loss function by adding prediction trees, each attempting to correct the error of the previous tree‌. The core idea of XGBoost is to combine multiple weak classifiers (decision trees) into one strong classifier. Its mathematical formula mainly involves the definition and optimization of the objective function. The objective function of XGBoost can be expressed as:

L (yi, y^) represents a loss function that measures the difference between the model’s predicted value y^ and the actual value yi. Ω(fk) represents the complexity of the k-th tree and is used to control the complexity of the model to prevent overfitting. θ represents the parameters of the model. XGBoost performs a second-order expansion of the loss function using Taylor’s formula to better approximate and optimize the loss function. By adding prediction trees, XGBoost gradually reduces residuals and improves the predictive performance of the model. Because of its efficiency, flexibility and powerful performance, XGBoost has been widely used in a variety of machine learning tasks such as classification, regression, and sequencing.

2.3.4 SVM

SVM is a two-class classification model, its basic model is defined as the linear classifier with the largest interval on the feature space, its learning strategy is to maximize the interval, and finally can be transformed into a convex quadratic programming problem. The goal of SVM is to find a hyperplane:

w is the weight vector and b is the bias term. The sample points of different classes are separated, and the distance from the nearest point (i.e., support vector) to the line is maximized as far as possible. This distance is called the margin, and the SVM attempts to maximize this gap.

2.3.5 KNN

KNN algorithm is a simple and intuitive classification and regression method, namely the K nearest neighbor algorithm. The core idea is that a sample belongs to a class if most of the K nearest neighbors of the sample in the feature space belong to that class. The general flow of the KNN algorithm is as follows:

First determine the size of the K, that is, how many neighbors to choose to participate in the decision. The choice of K value has great influence on the performance of the algorithm. Then calculate the distance between the test object and all objects in the training set: Euclidean distance is generally adopted, and the formula is:

Where x and y are two points in n-dimensional space, and xi and yi are their coordinates on the i-th dimension, respectively. According to the calculated distance, the K training samples closest to the test sample are found, which are the K nearest neighbors of the test sample. Then take a vote or weighted average and finally use the prediction category or predicted value as the output of the algorithm.

First, it is necessary to sample N times from the original data set (Bootstrap sampling method) to form a training set with the same size N (the comparison with the original data set is not completely consistent). If each sample in the data set has T attributes, t (t T) attributes will be randomly selected when the RF internal decision tree is split, and then the split attributes of the node will be selected according to some strategy, and finally a decision tree will be grown on the training set with size N. This is repeated m times, and a random forest of m decision trees is trained.

After completing the model parameter tuning, the predictive ability of each model was verified using a test machine, and the receiver operating characteristic curve (ROC) of each model was plotted, and precision, accuracy, and recall were selected, and the F1 score and area under curve (AUC) were selected as the evaluation indexes of model effectiveness. The model with the largest AUC was selected as the best model, and the Hosmer-Lemeshow (HL) was further performed to assess the degree of correspondence between the predicted probabilities and observations using the Hosmer-Lemeshow goodness-of-fit test [55]. p-values less than 0.05 indicate that there may be a model fitting problem, such as overfitting or underfitting [56]. We used the SHapley Additive exPlanation (SHAP) algorithm to explain the prediction model, which provides a globally consistent explanation of the model from the theory of game theory, can explain each feature output of the machine learning model at the group level as well as at the individual level, and visualize the output results to study the relative importance of each feature, in which the SHAP value of the edible oil bar graphs and scatter plots composed of summary graphs in a graphical representation, are used to illustrate the importance of individual features and their overall impact on model predictions [57].

2.4 Model Evaluation

2.4.1 Accuracy

Accuracy is a measure of the percentage of all predicted samples that the model correctly predicts. In this study, accuracy provides an intuitive evaluation criterion to help us understand the predictive power of the model as a whole. By evaluating the accuracy, it is possible to determine whether the model’s performance on the test data meets expectations, thus providing a basis for further optimization and the calculation formula is:

TP (True Positive) is the number of samples correctly predicted as a positive class, TN (True Negative) is the number of samples correctly predicted as a negative class, FP (False Positive) is the number of samples incorrectly predicted as a positive class, FN (False Negative) is the number of samples incorrectly predicted as a negative class.

2.4.2 Precision

The accuracy rate measures the proportion of all samples predicted to be positive that are actually positive. In this study, the accuracy rate reflects the accuracy of the model when predicting positive classes, such as patients with cardiovascular disease. The higher accuracy indicates that the model can identify the real positive samples well, which is of great significance for avoiding false positive prediction and reducing misdiagnosis, and the formula is:

Where TP is the number of samples correctly predicted as positive, and FP is the number of samples incorrectly predicted as positive.

2.4.3 Recall

The recall represents the percentage of all samples that are actually positive that are correctly predicted to be positive. Recall rates in this study were used to assess the model’s ability to identify positive samples, especially in high-risk patients. The higher recall rate means that the model can capture more actual positive samples, which is crucial for early detection of diseases and reducing missed diagnoses and is calculated as:

Where TP is the number of samples correctly predicted as a positive class, FN is the number of samples incorrectly predicted as a negative class.

2.4.4 F1 Score

The F1 score is the harmonic average of the accuracy rate and the recall rate, which takes into account the accuracy and completeness of the model in the positive prediction. In this study, F1 scores provide a way to balance accuracy and recall, especially when dealing with data imbalances. With F1 scores, we are able to evaluate the classification performance of the model more comprehensively and its calculation formula is:

2.4.5 AUC

The AUC value is calculated by plotting the area under the ROC curve, and the calculation formula is usually obtained by numerical integration. AUC is an important index to evaluate the performance of binary classification models, which measures the ability of models to distinguish between positive and negative classes. In this study, the AUC values reflect the comprehensive performance of the model under different thresholds. A higher value of AUC means that the model can distinguish positive and negative samples more effectively, and has a strong classification ability. Especially in the case of unbalanced categories, AUC is a very useful performance evaluation standard.

2.4.6 Matthews correlation coefficient (MCC)

MCC is a comprehensive index considering all classification results, which can fully reflect the classification performance of the model. In this study, MCC is used to evaluate the performance of the model in the face of unbalanced data. The closer the value of MCC is to 1, the better the prediction results of the model are. Especially when dealing with small samples of positive or negative classes, MCC provides a more stable performance evaluation and its calculation formula is:

2.4.7 Hosmer-Lemeshow Test Statistic

The Hosmer-Lemeshaw test is used to evaluate the goodness of fit of a model, and it tests the agreement between the predicted values of the model and the actual observed values. In this study, the test is used to determine whether the model can accurately fit the data and whether there are systematic errors. Through this test, we can confirm the model’s consistency across different data sets, thereby enhancing its reliability for clinical application and the statistics of the Hosmer-Lemeshaw test are usually calculated by the following formula:

Here, G represents the number of groups (usually 10 groups), O1kis the actual number of events observed in group kk (i.e., the number of samples with the dependent variable taking the value of 1), E1k is the number of events predicted by the model in group kk (that is, the sum of the predicted probabilities of all samples in this group), Nkis the total sample size of Group k, π¯kis the average of the predicted probabilities of the k group.

2.4.8 Hosmer-Lemeshow Test p-value

The p-value is calculated from the statistics of the Hosmer-Lemeshaw test, which is usually tested based on the Chi-square distribution. The p-value of Hosmer-Lemeshaw test is used to evaluate the fitting effect of the model, and a higher p-value indicates a better match between the model and the actual data. In this study, the p-value helped us judge the applicability of the model to clinical data, ensuring that its prediction results have a high degree of confidence in practical applications.

2.4.9 Confusion Matrix

The confusion matrix intuitively reveals the classification results of the model by showing the comparison between the actual categories and the predicted categories of the model. In this study, the confusion matrix helps us to understand the predictive performance of the model for various samples, especially whether it can correctly identify positive and negative classes. With this tool, we are able to evaluate the specific performance of the model in each category and provide specific directions for subsequent improvement. The confusion matrix usually looks like this:

Where TP, TN, FP and FN represent true example, true counter example, false positive and false counter example respectively.

2.4.10 Calibration Curve

The calibration curve evaluates the calibrability of the model by calculating the difference between the actual incidence and the predicted probability for each predicted probability interval. Specific formulas usually involve calculating by ratios or differences. The calibration curve shows the agreement between the probability predicted by the model and the actual results. In this study, the accuracy of the model’s prediction probabilities was evaluated by calibration curves to ensure that the model was not only able to classify, but also to provide reliable probability predictions. The good calibration curve shows that the probabilistic prediction of the model is consistent with the actual incidence rate, which enhances its operability and reliability in the actual clinical environment.

2.5 Research Quality Control

Strict quality control is used in order to ensure the reliability and accuracy of the study: (1) the collection process of the samples is carried out by two investigators in strict accordance with the development of the inclusion and exclusion criteria, and controversial cases are discussed and resolved with the intervention of a third person; (2) the sample data collection is completed for comparison, to ensure that there are no data extraction mismatches and omissions; (3) the inclusion of the data is carried out once again before inputing the data, sample set features containing missing values are deleted; (4) data are processed using the R language, and after the code is written, the code is repeated several times to check the code to ensure the accuracy of the results.

2.6 Literature Search Strategy

To ensure the comprehensiveness and rigor of the literature, we adopted a multi-dimensional literature search strategy in this study, which includes the following aspects: First, we searched for literature related to the risk factors and treatment strategies specifically for diabetic and CAD populations. The focus was on the complications of cardiovascular diseases in diabetic patients, risk assessments for CAD, the impact of diabetes on cardiovascular health, the effectiveness of drug treatments, and the role of lifestyle interventions (such as diet and exercise) in the prevention and treatment of CAD. These sources provided theoretical support for feature selection and helped identify key risk factors related to both diabetes and CAD.

Next, we searched for commonly used machine learning models in clinical prediction, particularly those applied in cardiovascular disease prediction. The search included traditional machine learning algorithms, such as logistic regression, SVM, and random forests, which are widely used in clinical data prediction modeling. Additionally, we explored the application of deep learning algorithms, such as neural networks, convolutional neural networks (CNN), and Transformer architectures, and assessed their applicability, advantages, and limitations in cardiovascular disease prediction. These sources provided valuable theoretical and practical guidance for model selection and optimization.

We then conducted further searches on the risk factors and treatment strategies for diabetes patients with CAD. Specifically, we focused on the unique characteristics of diabetic patients with CAD, such as the impact of diabetes on vascular health, the comorbid mechanisms between diabetes and CAD, and the effectiveness of combined treatments for diabetes and cardiovascular diseases. These studies contributed to a deeper understanding of the risks and treatment strategies specific to diabetic patients with CAD.

3. Results

3.1 Population Characteristics Overview

In this study, we retrospectively included patients with concomitant DM and CHD who attended the Beijing Anzhen Hospital of the Capital Medical University from 2023 to 2024. We included 3153 patients after strictly screening potential participants per the nadir criteria. Among them, 2056 patients received treatment (percutaneous coronary stenting or coronary artery bypass grafting), and 1097 patients received conservative treatment.

After excluding entries containing missing values, feature screening was performed with the LASSO regression within the training set (the differences between the training and test sets were not statistically significant), and the optimal parameter alpha of 0.02 was obtained after validation through grid search, corresponding to the smallest error. The features with non-zero coefficients could be included at this time, and the final set of the best features was obtained, which included: sex, age, and the randomized glucose level, positivity or weak positivity of fecal occult blood, free thyroxine level, erythrocyte distribution width, coefficient of variation, glycosylated hemoglobin level, high-density lipoprotein cholesterol level, hemoglobin level, glomerular filtration rate, alanine aminotransferase titer, pulmonary artery trunk internal diameter, maximum left ventricular diastolic E-wave flow velocity, maximum aortic flow velocity, and hospitalization period, and activated partial thromboplastin time. This optimal feature set was used in the construction of the subsequent five predictive models. The flowchart is shown in Fig. 1.

The general data of treatment group (n = 2056) and conservative treatment group (n = 1097) were compared: the ages (years) were 62.59 ± 8.67 and 64.73 ± 9.28, respectively. There were 1527 (74.27%) and 803 (73.20%) males, respectively. Fecal occult blood was positive or weak positive in 258 (12.55%) cases and 166 (15.13%) cases. Heart failure occurred in 381 (18.53%) cases and 303 (27.62%) cases, respectively. Comparison of relevant data of laboratory examination results: the results of random blood glucose examination (mmol/L) were 6.49 ± 2.58 and 7.05 ± 2.84, respectively. The results of low density lipoprotein cholesterol (mmol/L) were 2.19 ± 0.85 and 2.26 ± 0.87, respectively. The relative data of echocardiography were compared: left ventricular ejection fraction (%) was 58.99 ± 8.71 and 58.22 ± 10.95, respectively. The maximum diastolic A-wave velocity (cm/s) of left ventricular were 89.95 ± 20.69 and 91.45 ± 24.73, respectively. The main internal diameters of pulmonary artery (mm) were 23.16 ± 2.54 and 23.6 ± 3.1, respectively. The maximum velocity of pulmonary artery (cm/s) was 90.25 ± 15.36 and 91.17 ± 18.73, respectively. The maximum E-wave velocity (cm/s) in left ventricular diastolic period were 72.21 ± 24.92 and 80.79 ± 35.42, respectively. Comparisons of general data between the positive group (n = 2056) and the negative group (n = 1097) are presented in Table 1.

Subsequently, all included data were randomly split into training and testing sets at a ratio of 8:2. The data were preprocessed and fed into various algorithms for model construction. Upon completion of the construction, each algorithmic model was evaluated for its performance.

3.2 Prognostic Implication and Predictive Performance of the Five Models

In this study, five models were constructed based on machine learning, and the calibration curves of its five algorithms are shown in Fig. 2A–E; the ROC curves of the five algorithms are shown in Fig. 2L. Among them, the AUC of the RF model was 0.87; the AUC of the LR model was 0.70; the AUC of the XGBoost model was 0.89; the AUC of the SVM model was 0.75; and the AUC of the KNN model was 0.74. To alleviate the bias caused by the data imbalance, this study calculates additional metrics, including precision, accuracy, recall, MCC and the F1 score to comprehensively evaluate the model’s predictive performance based on the four basic metrics TP, TN, FP and FN in the model confusion matrix. Precision represents the percentage of samples that are actually positive out of all the samples predicted by the model to be positive. It measures the accuracy of the model when the prediction is a positive example, and the formula is Precision=TPTP+Fp; Accuracy represents the proportion of samples correctly predicted by the model to the total number of samples. It measures the accuracy of the overall classification of the model. The calculation formula is Accuracy=TP+TNTP+TN+FP+FN; Recall represents the percentage of samples that are correctly predicted to be positive by the model out of all samples that are actually positive. It measures the model’s ability to identify positive samples and is calculated as Recall=TPTP+FN; MCC is a measure of the quality of a binary classification model that takes into account all four classification outcomes (TP, TN, FP, FN) and returns a value between –1 and 1. An MCC of 1 is a perfect prediction, 0 is an average random prediction and –1 is a completely inconsistent prediction and its calculation formula is MCC=TP×TN-FP×FN(TP+FP)(TP+FN)(TN+FP)(TN+FN); The F1 score is the harmonic average of Precision and Recall and is used to weigh between the two. It provides a single metric to evaluate the overall performance of the model and its calculation formula is F1Score=2×TP2×TP+FP+FN.

Based on the evaluation results of the five algorithms, it can be concluded that the XGBoost model shows better performance in all the evaluation metrics, demonstrating the best prediction. A confusion matrix (CM) is a specific table layout used in machine learning and statistics to describe the performance of supervised learning algorithms. It provides a visual representation of the comparison between the predicted and actual results of a classification model on a test dataset. The CM for the five models is shown in Fig. 2G–K. A comparison of the performances of the different models is presented in Fig. 2F and Table 2. To further assess model calibration, we used the HL test. The results, including the associated p-values, are shown in Table 3. RF and KNN were the poorly fitted models, whereas LR, XGBoost, and SVM were the well-fitted ones.

3.3 Clinical Interpretability Underlying XGBoost

To visualize feature selection and treatment options and further elaborate the correlation between features, we used the XGBoost algorithm as a template to draw a scatter plot of the relationship between continuous variables and ending variables in the features (Fig. 3). At the same time, we constructed pairwise plots and Spearman correlation heatmaps using the original dataset. Pairwise plots use a color-coding system to differentiate the choice of treatment regimen, thus facilitating the observation of correlations and distributions among features (Fig. 4A). Heatmaps indicate correlations between features, with the color intensity indicating the Spearman correlation coefficient (Fig. 4B). These visualizations provide insights into the relationships between features and reveal differences in the distribution of treatment regimen choices. At the same time, we plotted the Pearson correlation coefficient, which measures the linear relationship between continuous variables in the features (Fig. 4C).

To gain a deeper understanding of the predictive power of the XGBoost model, this study used SHAP values, which elucidate the contribution of each feature to the prediction, thus enabling the identification of the key features that influence the model’s decision. To visualize the influence of the importance of each feature on the individual predictions, we used TreeExplainer from the SHAP library to calculate the SHAP values and generated SHAP summary plots and feature importance plots (Fig. 5A). From the plots, it can be seen that the feature with the greatest impact on the predictions in the XGBoost model is the glycated hemoglobin level, followed by the free thyroxine level result and the pulmonary artery trunk internal diameter. The SHAP correlation summary plots, conversely, demonstrate the impact of the XGBoost predictions on the population. The SHAP correlation summary plot shows the contribution of each feature in the XGBoost prediction model at the population level. Each point in the plot represents a sample, and the color of the point reflects the magnitude of the corresponding feature value of the sample, with red indicating that the value of the feature is relatively high and blue indicating that the value of the feature is low. Points to the right of the baseline (i.e., the dotted line) in the figure are meant to have a positive impact on the model predictions, while the opposite is true for those on the left side, with the impact increasing with the distance from the baseline (Fig. 5B).

The SHAP analysis provides a comprehensive understanding of the decision-making process of the XGBoost model and identifies relevant predictors. These findings are essential for further optimizing the model and interpreting its predictions. Meanwhile, the relevant features mined highlight their potential clinical applications, which can provide a more comprehensive assessment and a scientific basis for personalized medicine.

4. Discussion

In this study, we retrospectively included patients with both CHD and DM who attended the Beijing Anzhen Hospital of the Capital Medical University in 2022–2023 based on five machine-learning algorithms for prediction modeling. We found that compared with other algorithms, the XGBoost algorithm model had better predictive ability and better comprehensive performance, suggesting that machine-learning algorithms and data mining and analysis have unique advantages, which are more easily reflected in large-sample data. This algorithm is characterized by its robustness and proficiency in dealing with high-latitude features and its excellent ability to capture complex nonlinear relationships. Meanwhile, we identified 65 predictors among more than 100 features, including biological data, laboratory results, and imaging data, in which 15 features should be given sufficient attention in clinical work and provide new ideas for the treatment plan of the disease.

Age and gender became two of the 15 key factors in the predictive model for treatment options of coronary heart disease in diabetic patients, mainly due to their significant roles in physiological mechanisms, pathological changes, and treatment responses. Age influences the model through physiological changes such as increased vascular stiffness, impaired endothelial function, and exacerbated atherosclerosis with aging. These changes make elderly diabetic patients more susceptible to cardiovascular complications, requiring more cautious treatment decisions. For example, elderly patients are more likely to undergo interventional treatments (such as coronary stent implantation) rather than relying solely on medication because their blood vessels are less elastic, and the effects of medication may not be as effective as in younger patients. Additionally, elderly patients often have other chronic conditions, such as hypertension and kidney disease, which further influence treatment choices. Gender plays a crucial role due to the differences in the pathophysiology and treatment response between males and females. Men tend to exhibit more atherosclerosis and coronary artery disease at a younger age, while women experience a sharp increase in cardiovascular risk after menopause due to the loss of estrogen’s protective effects, particularly in diabetic women. Gender also affects drug metabolism and treatment adherence. Women may have different responses to certain medications (such as antihypertensive and lipid-lowering drugs) compared to men, and their treatment adherence may be lower. In summary, age and gender directly influence treatment decisions by affecting the patient’s physiological condition, disease progression, and drug response, which explains why these factors are key predictors in our machine learning model.

It is important to accurately predict the optimal treatment regimen for different individuals, as CHD and DM are two comorbid conditions with interdependent disease progressions [58, 59]. In recent years, many risk assessment and disease prognosis prediction models have been developed but not specifically for the prediction of treatment regimens in patients with both DM and CHD. A study using machine learning to predict atrial fibrillation in elderly patients with coronary heart disease and type 2 diabetes showed that the best model XGBoost had a sensitivity of 0.833, a specificity of 0.562, an accuracy of 0.587, and an AUC of 0.743, compared to existing superior models [60]. Another study of CHD diagnosis model for elderly diabetic patients based on machine learning algorithm showed that the optimal random forest model had an AUC of 0.845, an accuracy of 0.789, an accuracy of 0.778, an F1 score of 0.735, a sensitivity of 0.688, and a specificity of 0.851 [61]. Compared to these findings, our model outperforms on several key metrics: accuracy of 0.801, accuracy of 0.822, recall rate of 0.779, F1 score of 0.8, AUC of 0.893, and MCC of 0.603. These results show that our model not only makes a breakthrough in prediction accuracy, but also outperforms existing prediction models in overall performance, especially in terms of AUC and F1 scores, which further validates the effectiveness and potential clinical application value of our method. It is worth noting that in such machine-learning prediction models for predicting the risk of diabetic patients with comorbid CHD, there are the same features as in the present study, and such features show a consistent tendency and match. Sometimes, identical features have different weights in different models [62, 63, 64] suggesting that there is heterogeneity among features in disease assessment models constructed for different populations and even the same population. We also need to note that there are still barriers to the development of artificial intelligence (AI) and its integration into clinical practice [65, 66, 67], such as the need to maximize accuracy while avoiding overfitting and determine what clinical and general data should be included, taking into account convenience and the patient’s financial burden. Also, in the use of machine-learning algorithms, researchers often choose algorithms based on their own preferences and knowledge limitations, resulting in algorithms that may not be the best ones, having suboptimal prediction accuracy [68, 69]. In this regard, we utilize a standardized approach that employs retrospective studies to ensure the credibility of the basic factual results. Using multiple machine-learning models can reduce the underlying uncertainty and ultimately identify the best prediction model based on the evaluation indexes, such as the area under the AUC curve of each algorithm, accuracy, and precision, to reduce the bias of the results caused by human error.

Although this study provides valuable insights, several limitations should be acknowledged. Since this is a retrospective study, some relevant information may have been omitted during the data inclusion process. Individual cases with missing data were excluded, which resulted in a reduction of the sample size. Additionally, features with higher rates of missing values were removed, which may have led to the exclusion of potentially stronger predictive factors.

Furthermore, the model was validated solely on data from coronary heart disease patients at the Beijing Anzhen Hospital, affiliated with the Capital Medical University. It is important to note that when the model is trained on datasets with different data patterns (such as those from different hospitals, regions, or ethnic backgrounds), it may face challenges in generalizing to external populations. This can lead to incorrect predictions or an overfitting to the specificity of the training data, thereby limiting its ability to generalize to heterogeneous data not represented in the training set. To address these issues, we plan to develop related software programs or websites to support multicenter networking and improve the model’s accuracy and broad applicability.

In the future, in addition to conducting multi-center external validation, prospective validation of the model’s accuracy and universality with data from diverse and broader patient populations will be necessary. Additionally, an automatic data extraction system could be established within the database to significantly improve the efficiency of sample collection, further enhancing the robustness and scalability of the model.

In the future deployment of clinical AI models, first, the model should be integrated with existing electronic health record (EHR) systems, allowing doctors to quickly obtain real-time risk assessments and treatment recommendations based on patient data during daily care. By seamlessly connecting with clinical workflows, doctors can directly refer to the output of AI models at every stage of diagnosis and treatment, thus making more accurate decisions. Second, AI models can set thresholds and alert mechanisms to help doctors identify high-risk patients, especially in emergency or intensive care Settings, and automatically alert doctors to intervene in a timely manner. Finally, to ensure the continued effectiveness of AI models, they need to be regularly updated and optimized to respond to changing patient population characteristics and disease progression in clinical practice, ensuring the accuracy and clinical adaptability of the models. Through these measures, the proposed AI model will not only enhance personalized treatment, but also improve clinical work efficiency and ultimately optimize patient outcomes.

As the dataset continues to expand, we will further explore the application of this model in clinical settings and validate its effectiveness in real-world clinical decision-making, aiming to provide more precise and personalized treatment options for coronary heart disease patients.

5. Conclusion

We retrospectively included patients with concomitant DM and CHD who attended the Beijing Anzhen Hospital of the Capital Medical University in 2022–2023 based on a machine-learning algorithm and established a prediction model for the establishment of a treatment plan for these patients. We identified the XGBoost algorithm as the best one by incorporating the patients’ general information, laboratory test results, and echocardiographic findings, and screened for the optimal feature set. The optimal feature set, which contained 15 features, was selected to assist us in choosing the treatment plan. It provides help and ideas for the development of the optimal treatment plan for patients with concomitant DM and CHD.

References

[1]

Correction to: 2024 Heart Disease and Stroke Statistics: A Report of US and Global Data From the American Heart Association. Circulation. 2024; 149: e1164. https://doi.org/10.1161/CIR.0000000000001247.

[2]

Virani SS, Alonso A, Benjamin EJ, Bittencourt MS, Callaway CW, Carson AP, et al. Heart Disease and Stroke Statistics-2020 Update: A Report From the American Heart Association. Circulation. 2020; 141: e139–e596. https://doi.org/10.1161/CIR.0000000000000757.

[3]

Ding H, Hou X, Gao Z, Guo Y, Liao B, Wan J. Challenges and Strategies for Endothelializing Decellularized Small-Diameter Tissue-Engineered Vessel Grafts. Advanced Healthcare Materials. 2024; 13: e2304432. https://doi.org/10.1002/adhm.202304432.

[4]

Safiri S, Karamzad N, Singh K, Carson-Chahhoud K, Adams C, Nejadghaderi SA, et al. Burden of ischemic heart disease and its attributable risk factors in 204 countries and territories, 1990-2019. European Journal of Preventive Cardiology. 2022; 29: 420–431. https://doi.org/10.1093/eurjpc/zwab213.

[5]

Dai H, Much AA, Maor E, Asher E, Younis A, Xu Y, et al. Global, regional, and national burden of ischaemic heart disease and its attributable risk factors, 1990-2017: results from the Global Burden of Disease Study 2017. European Heart Journal. Quality of Care & Clinical Outcomes. 2022; 8: 50–60. https://doi.org/10.1093/ehjqcco/qcaa076.

[6]

Florek K, Kübler M, Górka M, Kübler P. New Modifiable Risk Factors Influencing Coronary Artery Disease Severity. International Journal of Molecular Sciences. 2024; 25: 7766. https://doi.org/10.3390/ijms25147766.

[7]

Senyigit A, Uzun H, Gultepe I, Konukoglu D. The relationship between carotid intima-media thickness and serum secreted frizzled-related protein-4 and dipeptidyl peptidase-4 in diabetic patients with cardiovascular diseases. Bratislavske Lekarske Listy. 2019; 120: 188–194. https://doi.org/10.4149/BLL_2019_032.

[8]

Bielka W, Przezak A, Molęda P, Pius-Sadowska E, Machaliński B. Double diabetes-when type 1 diabetes meets type 2 diabetes: definition, pathogenesis and recognition. Cardiovascular Diabetology. 2024; 23: 62. https://doi.org/10.1186/s12933-024-02145-x.

[9]

Zheng Y, Ley SH, Hu FB. Global aetiology and epidemiology of type 2 diabetes mellitus and its complications. Nature Reviews. Endocrinology. 2018; 14: 88–98. https://doi.org/10.1038/nrendo.2017.151.

[10]

Liu Y, Ning X, Zhang L, Long J, Liang R, Peng S, et al. Prevalence of long-term complications in inpatients with diabetes mellitus in China: a nationwide tertiary hospital-based study. BMJ Open Diabetes Research & Care. 2022; 10: e002720. https://doi.org/10.1136/bmjdrc-2021-002720.

[11]

Wang Y, Yu Q, Fan D, Cao F. Coronary heart disease in type 2 diabetes: mechanisms and comprehensive prevention strategies. Expert Review of Cardiovascular Therapy. 2012; 10: 1051–1060. https://doi.org/10.1586/erc.12.52.

[12]

Song SH, Hardisty CA. Type 2 diabetes mellitus: a high-risk condition for cardiovascular disease irrespective of the different degrees of obesity. QJM: Monthly Journal of the Association of Physicians. 2008; 101: 875–879. https://doi.org/10.1093/qjmed/hcn109.

[13]

Fan R, Zhang N, Yang L, Ke J, Zhao D, Cui Q. AI-based prediction for the risk of coronary heart disease among patients with type 2 diabetes mellitus. Scientific Reports. 2020; 10: 14457. https://doi.org/10.1038/s41598-020-71321-2.

[14]

Zhang H, Dellsperger KC, Zhang C. The link between metabolic abnormalities and endothelial dysfunction in type 2 diabetes: an update. Basic Research in Cardiology. 2012; 107: 237. https://doi.org/10.1007/s00395-011-0237-1.

[15]

Hong SJ, Lee YJ, Kang WC, Hong BK, Lee JY, Lee JB, et al. Effect of rosuvastatin versus atorvastatin on new-onset diabetes mellitus in patients treated with high-intensity statin therapy for coronary artery disease: a post-hoc analysis from the LODESTAR randomized clinical trial. Cardiovascular Diabetology. 2024; 23: 287. https://doi.org/10.1186/s12933-024-02386-w.

[16]

Betteridge DJ, Carmena R. The diabetogenic action of statins - mechanisms and clinical implications. Nature Reviews. Endocrinology. 2016; 12: 99–110. https://doi.org/10.1038/nrendo.2015.194.

[17]

Thompson PD, Panza G, Zaleski A, Taylor B. Statin-Associated Side Effects. Journal of the American College of Cardiology. 2016; 67: 2395–2410. https://doi.org/10.1016/j.jacc.2016.02.071.

[18]

Yoshida N, Okumura KI, Aso Y. High serum pentosidine concentrations are associated with increased arterial stiffness and thickness in patients with type 2 diabetes. Metabolism: Clinical and Experimental. 2005; 54: 345–350. https://doi.org/10.1016/j.metabol.2004.09.014.

[19]

Sedlis SP, Hartigan PM, Teo KK, Maron DJ, Spertus JA, Mancini GBJ, et al. Effect of PCI on Long-Term Survival in Patients with Stable Ischemic Heart Disease. The New England Journal of Medicine. 2015; 373: 1937–1946. https://doi.org/10.1056/NEJMoa1505532.

[20]

Melamed RJ, Tillmann A, Kufleitner HE, Thürmer U, Dürsch M. Evaluating the efficacy of an education and treatment program for patients with coronary heart disease. Deutsches Arzteblatt International. 2014; 111: 802–808. https://doi.org/10.3238/arztebl.2014.0802.

[21]

Fox KAA, Metra M, Morais J, Atar D. The myth of ’stable’ coronary artery disease. Nature Reviews. Cardiology. 2020; 17: 9–21. https://doi.org/10.1038/s41569-019-0233-y.

[22]

Bertsimas D, Orfanoudaki A, Weiner RB. Personalized treatment for coronary artery disease patients: a machine learning approach. Health Care Management Science. 2020; 23: 482–506. https://doi.org/10.1007/s10729-020-09522-4.

[23]

Mensah GA, Fuster V, Murray CJL, Roth GA, Global Burden of Cardiovascular Diseases and Risks Collaborators. Global Burden of Cardiovascular Diseases and Risks, 1990-2022. Journal of the American College of Cardiology. 2023; 82: 2350–2473. https://doi.org/10.1016/j.jacc.2023.11.007.

[24]

Fuster V, Badimon L, Badimon JJ, Chesebro JH. The pathogenesis of coronary artery disease and the acute coronary syndromes (2). The New England Journal of Medicine. 1992; 326: 310–318. https://doi.org/10.1056/NEJM199201303260506.

[25]

Fihn SD, Blankenship JC, Alexander KP, Bittl JA, Byrne JG, Fletcher BJ, et al. 2014 ACC/AHA/AATS/PCNA/SCAI/STS focused update of the guideline for the diagnosis and management of patients with stable ischemic heart disease: a report of the American College of Cardiology/American Heart Association Task Force on Practice Guidelines, and the American Association for Thoracic Surgery, Preventive Cardiovascular Nurses Association, Society for Cardiovascular Angiography and Interventions, and Society of Thoracic Surgeons. Journal of the American College of Cardiology. 2014; 64: 1929–1949. https://doi.org/10.1016/j.jacc.2014.07.017.

[26]

Warnes CA. Adult congenital heart disease: the challenges of a lifetime. European Heart Journal. 2017; 38: 2041–2047. https://doi.org/10.1093/eurheartj/ehw529.

[27]

Khanna NN, Maindarkar MA, Viswanathan V, Puvvula A, Paul S, Bhagawati M, et al. Cardiovascular/Stroke Risk Stratification in Diabetic Foot Infection Patients Using Deep Learning-Based Artificial Intelligence: An Investigative Study. Journal of Clinical Medicine. 2022; 11: 6844. https://doi.org/10.3390/jcm11226844.

[28]

Munjral S, Maindarkar M, Ahluwalia P, Puvvula A, Jamthikar A, Jujaray T, et al. Cardiovascular Risk Stratification in Diabetic Retinopathy via Atherosclerotic Pathway in COVID-19/Non-COVID-19 Frameworks Using Artificial Intelligence Paradigm: A Narrative Review. Diagnostics (Basel, Switzerland). 2022; 12: 1234. https://doi.org/10.3390/diagnostics12051234.

[29]

Leem J, Kim HY. Action-specialized expert ensemble trading system with extended discrete action space using deep reinforcement learning. PloS One. 2020; 15: e0236178. https://doi.org/10.1371/journal.pone.0236178.

[30]

Zhang Z, written on behalf of AME Big-Data Clinical Trial Collaborative Group. Reinforcement learning in clinical medicine: a method to optimize dynamic treatment regime over time. Annals of Translational Medicine. 2019; 7: 345. https://doi.org/10.21037/atm.2019.06.75.

[31]

Marcos-Pasero H, Colmenarejo G, Aguilar-Aguilar E, Ramírez de Molina A, Reglero G, Loria-Kohen V. Ranking of a wide multidomain set of predictor variables of children obesity by machine learning variable importance techniques. Scientific Reports. 2021; 11: 1910. https://doi.org/10.1038/s41598-021-81205-8.

[32]

Heng T, Yang D, Wang R, Zhang L, Lu Y, Du G. Progress in Research on Artificial Intelligence Applied to Polymorphism and Cocrystal Prediction. ACS Omega. 2021; 6: 15543–15550. https://doi.org/10.1021/acsomega.1c01330.

[33]

Kaul V, Enslin S, Gross SA. History of artificial intelligence in medicine. Gastrointestinal Endoscopy. 2020; 92: 807–812. https://doi.org/10.1016/j.gie.2020.06.040.

[34]

Cook NR, Ridker PM. Calibration of the Pooled Cohort Equations for Atherosclerotic Cardiovascular Disease: An Update. Annals of Internal Medicine. 2016; 165: 786–794. https://doi.org/10.7326/M16-1739.

[35]

Mintz Y, Brodie R. Introduction to artificial intelligence in medicine. Minimally Invasive Therapy & Allied Technologies: MITAT: Official Journal of the Society for Minimally Invasive Therapy. 2019; 28: 73–81. https://doi.org/10.1080/13645706.2019.1575882.

[36]

Sánchez-Cabo F, Rossello X, Fuster V, Benito F, Manzano JP, Silla JC, et al. Machine Learning Improves Cardiovascular Risk Definition for Young, Asymptomatic Individuals. Journal of the American College of Cardiology. 2020; 76: 1674–1685. https://doi.org/10.1016/j.jacc.2020.08.017.

[37]

Kimber TB, Chen Y, Volkamer A. Deep Learning in Virtual Screening: Recent Applications and Developments. International Journal of Molecular Sciences. 2021; 22: 4435. https://doi.org/10.3390/ijms22094435.

[38]

Garcia-Sanchez AJ, Garcia Angosto E, Llor JL, Serna Berna A, Ramos D. Machine Learning Techniques Applied to Dose Prediction in Computed Tomography Tests. Sensors (Basel, Switzerland). 2019; 19: 5116. https://doi.org/10.3390/s19235116.

[39]

Fernández-Ortiz A, Jiménez-Borreguero LJ, Peñalvo JL, Ordovás JM, Mocoroa A, Fernández-Friera L, et al. The Progression and Early detection of Subclinical Atherosclerosis (PESA) study: rationale and design. American Heart Journal. 2013; 166: 990–998. https://doi.org/10.1016/j.ahj.2013.08.024.

[40]

Ambale-Venkatesh B, Yang X, Wu CO, Liu K, Hundley WG, McClelland R, et al. Cardiovascular Event Prediction by Machine Learning: The Multi-Ethnic Study of Atherosclerosis. Circulation Research. 2017; 121: 1092–1101. https://doi.org/10.1161/CIRCRESAHA.117.311312.

[41]

Kakadiaris IA, Vrigkas M, Yen AA, Kuznetsova T, Budoff M, Naghavi M. Machine Learning Outperforms ACC / AHA CVD Risk Calculator in MESA. Journal of the American Heart Association. 2018; 7: e009476. https://doi.org/10.1161/JAHA.118.009476.

[42]

Motwani M, Dey D, Berman DS, Germano G, Achenbach S, Al-Mallah MH, et al. Machine learning for prediction of all-cause mortality in patients with suspected coronary artery disease: a 5-year multicentre prospective registry analysis. European Heart Journal. 2017; 38: 500–507. https://doi.org/10.1093/eurheartj/ehw188.

[43]

Weng SF, Reps J, Kai J, Garibaldi JM, Qureshi N. Can machine-learning improve cardiovascular risk prediction using routine clinical data? PloS One. 2017; 12: e0174944. https://doi.org/10.1371/journal.pone.0174944.

[44]

Alaa AM, Bolton T, Di Angelantonio E, Rudd JHF, van der Schaar M. Cardiovascular disease risk prediction using automated machine learning: A prospective study of 423,604 UK Biobank participants. PloS One. 2019; 14: e0213653. https://doi.org/10.1371/journal.pone.0213653.

[45]

van der Ploeg T, Austin PC, Steyerberg EW. Modern modelling techniques are data hungry: a simulation study for predicting dichotomous endpoints. BMC Medical Research Methodology. 2014; 14: 137. https://doi.org/10.1186/1471-2288-14-137.

[46]

Collins GS, Moons KGM. Reporting of artificial intelligence prediction models. Lancet (London, England). 2019; 393: 1577–1579. https://doi.org/10.1016/S0140-6736(19)30037-6.

[47]

Omar AMS, Narula S, Abdel Rahman MA, Pedrizzetti G, Raslan H, Rifaie O, et al. Precision Phenotyping in Heart Failure and Pattern Clustering of Ultrasound Data for the Assessment of Diastolic Dysfunction. JACC. Cardiovascular Imaging. 2017; 10: 1291–1303. https://doi.org/10.1016/j.jcmg.2016.10.012.

[48]

Deo RC. Machine Learning in Medicine. Circulation. 2015; 132: 1920–1930. https://doi.org/10.1161/CIRCULATIONAHA.115.001593.

[49]

Beck AH, Sangoi AR, Leung S, Marinelli RJ, Nielsen TO, van de Vijver MJ, et al. Systematic analysis of breast cancer morphology uncovers stromal features associated with survival. Science Translational Medicine. 2011; 3: 108ra113. https://doi.org/10.1126/scitranslmed.3002564.

[50]

Huo D, Kou B, Zhou Z, Lv M. A machine learning model to classify aortic dissection patients in the early diagnosis phase. Scientific Reports. 2019; 9: 2701. https://doi.org/10.1038/s41598-019-39066-9.

[51]

Kwon JM, Kim KH, Jeon KH, Lee SE, Lee HY, Cho HJ, et al. Artificial intelligence algorithm for predicting mortality of patients with acute heart failure. PloS One. 2019; 14: e0219302. https://doi.org/10.1371/journal.pone.0219302.

[52]

Zhang G, Wang Z, Tong Z, Qin Z, Su C, Li D, et al. AI hybrid survival assessment for advanced heart failure patients with renal dysfunction. Nature Communications. 2024; 15: 6756. https://doi.org/10.1038/s41467-024-50415-9.

[53]

Feng ZH, Wang LY, Yang ZQ, Zhang YY, Li X, Song L, et al. Hyperspectral Monitoring of Powdery Mildew Disease Severity in Wheat Based on Machine Learning. Frontiers in Plant Science. 2022; 13: 828454. https://doi.org/10.3389/fpls.2022.828454.

[54]

Nguyen TTD, Nguyen DK, Ou YY. Addressing data imbalance problems in ligand-binding site prediction using a variational autoencoder and a convolutional neural network. Briefings in Bioinformatics. 2021; 22: bbab277. https://doi.org/10.1093/bib/bbab277.

[55]

Vanchiere C, Thirumal R, Hendrani A, Dherange P, Bennett A, Shi R, et al. Association Between Atrial Fibrillation and Occupational Exposure in Firefighters Based on Self-Reported Survey Data. Journal of the American Heart Association. 2022; 11: e022543. https://doi.org/10.1161/JAHA.121.022543.

[56]

Würtz P, Havulinna AS, Soininen P, Tynkkynen T, Prieto-Merino D, Tillin T, et al. Metabolite profiling and cardiovascular event risk: a prospective study of 3 population-based cohorts. Circulation. 2015; 131: 774–785. https://doi.org/10.1161/CIRCULATIONAHA.114.013116.

[57]

Sun R, Wang X, Jiang H, Yan Y, Dong Y, Yan W, et al. Prediction of 30-day mortality in heart failure patients with hypoxic hepatitis: Development and external validation of an interpretable machine learning model. Frontiers in Cardiovascular Medicine. 2022; 9: 1035675. https://doi.org/10.3389/fcvm.2022.1035675.

[58]

Kannel WB, McGee DL. Diabetes and cardiovascular disease. The Framingham study. JAMA. 1979; 241: 2035–2038. https://doi.org/10.1001/jama.241.19.2035.

[59]

de Grauw WJ, van de Lisdonk EH, van den Hoogen HJ, van Weel C. Cardiovascular morbidity and mortality in type 2 diabetic patients: a 22-year historical cohort study in Dutch general practice. Diabetic Medicine: a Journal of the British Diabetic Association. 1995; 12: 117–122. https://doi.org/10.1111/j.1464-5491.1995.tb00441.x.

[60]

Xu Q, Peng Y, Tan J, Zhao W, Yang M, Tian J. Prediction of Atrial Fibrillation in Hospitalized Elderly Patients With Coronary Heart Disease and Type 2 Diabetes Mellitus Using Machine Learning: A Multicenter Retrospective Study. Frontiers in Public Health. 2022; 10: 842104. https://doi.org/10.3389/fpubh.2022.842104.

[61]

Xu H, Cao WZ, Bai YY, Dong J, Che HB, Bai P, et al. Establishment of a diagnostic model of coronary heart disease in elderly patients with diabetes mellitus based on machine learning algorithms. Journal of Geriatric Cardiology: JGC. 2022; 19: 445–455. https://doi.org/10.11909/j.issn.1671-5411.2022.06.006.

[62]

Ma CY, Luo YM, Zhang TY, Hao YD, Xie XQ, Liu XW, et al. Predicting coronary heart disease in Chinese diabetics using machine learning. Computers in Biology and Medicine. 2024; 169: 107952. https://doi.org/10.1016/j.compbiomed.2024.107952.

[63]

Xiao S, Dong Y, Huang B, Jiang X. Predictive nomogram for coronary heart disease in patients with type 2 diabetes mellitus. Frontiers in Cardiovascular Medicine. 2022; 9: 1052547. https://doi.org/10.3389/fcvm.2022.1052547.

[64]

Eliasson B, Gudbjörnsdottir S, Zethelius B, Eeg-Olofsson K, Cederholm J, National Diabetes Register (NDR). LDL-cholesterol versus non-HDL-to-HDL-cholesterol ratio and risk for coronary heart disease in type 2 diabetes. European Journal of Preventive Cardiology. 2014; 21: 1420–1428. https://doi.org/10.1177/2047487313494292.

[65]

Samad MD, Ulloa A, Wehner GJ, Jing L, Hartzel D, Good CW, et al. Predicting Survival From Large Echocardiography and Electronic Health Record Datasets: Optimization With Machine Learning. JACC. Cardiovascular Imaging. 2019; 12: 681–689. https://doi.org/10.1016/j.jcmg.2018.04.026.

[66]

Kwon JM, Kim KH, Jeon KH, Park J. Deep learning for predicting in-hospital mortality among heart disease patients based on echocardiography. Echocardiography (Mount Kisco, N.Y.). 2019; 36: 213–218. https://doi.org/10.1111/echo.14220.

[67]

Hearn J, Ross HJ, Mueller B, Fan CP, Crowdy E, Duhamel J, et al. Neural Networks for Prognostication of Patients With Heart Failure. Circulation. Heart Failure. 2018; 11: e005193. https://doi.org/10.1161/CIRCHEARTFAILURE.118.005193.

[68]

Volovici V, Syn NL, Ercole A, Zhao JJ, Liu N. Steps to avoid overuse and misuse of machine learning in clinical research. Nature Medicine. 2022; 28: 1996–1999. https://doi.org/10.1038/s41591-022-01961-6.

[69]

Khan MS, Arshad MS, Greene SJ, Van Spall HGC, Pandey A, Vemulapalli S, et al. Artificial intelligence and heart failure: A state-of-the-art review. European Journal of Heart Failure. 2023; 25: 1507–1525. https://doi.org/10.1002/ejhf.2994.

Funding

Beijing Hospitals Authority's Ascent Plan(DFL20220605)

PDF (22222KB)

0

Accesses

0

Citation

Detail

Sections
Recommended

/