Sirocchi, C., Suffian, M., Sabbatini, F., Bogliolo, A., & Montagna, S. (2024). Evaluating Machine Learning Models against Clinical Protocols for Enhanced Interpretability and Continuity of Care. arXiv preprint arXiv:2411.03105.
This research paper investigates the integration of clinical protocols into machine learning models to enhance their interpretability and alignment with established medical guidelines, focusing on diabetes prediction as a case study. The authors aim to develop and evaluate metrics for assessing the accuracy and explainability of such integrated models compared to traditional data-driven approaches.
The study utilizes the Pima Indians Diabetes dataset and a set of rules derived from public health guidelines on type-2 diabetes risk factors. Two neural network models are trained: a data-driven model (DD-ML) and an integrated model (KB-ML) incorporating domain knowledge through a custom loss function. The authors propose novel metrics: Relative Accuracy (RA) to measure the model's adherence to the clinical protocol's predictions and Explanation Similarity to quantify the overlap between explanations provided by the protocol and the models. Rule extraction using CART is employed to generate interpretable rule sets from the black-box models.
The integrated model (KB-ML) demonstrates superior performance compared to the data-driven model in terms of balanced accuracy, ROC AUC, and recall, while also exhibiting significantly higher relative accuracy. Explanation similarity metrics reveal that the integrated model provides explanations more aligned with the clinical protocol than the data-driven model, particularly when using XNOR similarity. Additionally, the integrated model shows greater robustness in its explanations across different cross-validation folds.
Integrating domain knowledge from clinical protocols into machine learning models improves their adherence to established guidelines, resulting in more reliable and interpretable predictions, especially in diabetes risk assessment. The proposed RA and Explanation Similarity metrics effectively evaluate the alignment of integrated models with clinical knowledge.
This research contributes to the field of Informed Machine Learning in healthcare by proposing novel evaluation metrics and demonstrating the benefits of integrating domain knowledge for enhanced interpretability and continuity of care.
The study is limited to the Pima Indians Diabetes dataset and a specific set of clinical rules. Future research should explore the generalizability of these findings to other datasets, medical domains, and knowledge integration techniques. Further investigation into different rule extraction methods and explanation similarity metrics is also warranted.
To Another Language
from source content
arxiv.org
Key Insights Distilled From
by Christel Sir... at arxiv.org 11-06-2024
https://arxiv.org/pdf/2411.03105.pdfDeeper Inquiries