Тема: Research on Diabetes Prediction Model Based on Machine Learning from the Perspective of Cooperative Game Theory
Закажите новую по вашим требованиям
Представленный материал является образцом учебного исследования, примером структуры и содержания учебного исследования по заявленной теме. Размещён исключительно в информационных и ознакомительных целях.
Workspay.ru оказывает информационные услуги по сбору, обработке и структурированию материалов в соответствии с требованиями заказчика.
Размещение материала не означает публикацию произведения впервые и не предполагает передачу исключительных авторских прав третьим лицам.
Материал не предназначен для дословной сдачи в образовательные организации и требует самостоятельной переработки с соблюдением законодательства Российской Федерации об авторском праве и принципов академической добросовестности.
Авторские права на исходные материалы принадлежат их законным правообладателям. В случае возникновения вопросов, связанных с размещённым материалом, просим направить обращение через форму обратной связи.
📋 Содержание
2 Introduction 4
3 Diabetes data set 7
3.1 Description of the dataset 7
3.2 Data preprocessing 8
4 Medical detection model 10
4.1 One-Class SVM algorithm 10
4.2 Prediction Model Results and Analysis 14
4.3 Experimental Results and Analysis of the Prediction Model 14
4.3.1 Evaluation Metrics 14
4.3.2 Experimental Results 16
5 Medical detection model interpretability methods 18
5.1 SHAP algorithm 18
5.1.1 Shapley values in cooperative games 18
5.1.2 SHAP algorithm in machine learning 19
5.1.3 Feature function of SHAP algorithm 20
5.1.4 Compute SHAP values 23
5.2 тalgorithm 25
5.2.1 т-values in cooperative games 25
5.2.2 т-values algorithm 26
6 Diabetes detection model interpretation methods 28
6.1 SHAP model and result analysis 28
6.1.1 Visualization of Predictions 28
6.1.2 SHAP Feature Importance 29
6.1.3 SHAP Summary Plot 31
6.2 т-values model and result analysis 33
6.2.1 т-values Feature Importance 33
6.3 Model Comparison 34
7 Predictive model based on XGBoost 36
7.1 XGBoost 36
7.2 XGBoost prediction model 39
7.2.1 A prediction model based on Shapley 39
7.2.2 A prediction model based on т-values 41
7.2.3 Interpretable Model Comparison 42
8 Conclusion 44
References 46
9 appendix 50
📖 Введение
✅ Заключение
The process begins with the creation of a diabetes prediction model based on ONE-CLASS SVM. It is crucial to analyze and preprocess the dataset to avoid any adverse effects on the prediction results. Following this, the model is trained, prediction results are generated, and the model is evaluated. After the predictions are made, the results are interpreted to explain the feasibility of the model, thereby increasing the confidence of doctors and patients in the model’s predictions.
The explanation process is primarily divided into two steps. Firstly, Shapley values and т-values are computed based on the predictive model from the perspective of cooperative game theory. Both of these values can reflect the importance of features. After obtaining specific values, statistical graphs regarding instances and features are plotted to intuitively observe the influence of different features on the overall prediction results. The feature importance can then explain where the results of the ”black- box” predictive model come from.
Shapley values are a commonly used method for studying inter¬pretability, but т-values are being used for the first time. To investigate whether т-values can serve as a method for interpretability research, it is necessary to study their feasibility as an explanatory tool. Comparing the feature importance obtained by т-values with that obtained by the SHAP method reveals similar conclusions, indicating that т-values can be used as a method for researching interpretable AI. However, their performance in feature importance values is not significant.
I reconstructed medical detection models using т-values and SHAP values based on XGBoost. Both of these detection models exhibit high accuracy and predictive value, with similar results. Therefore, т -values can be fully applied to research on interpretable AI, and they can also be used to construct medical detection models along with SHAP values.





