Mixed-variable graphical modeling framework towards risk prediction of hospital-acquired pressure injury in spinal cord injury individuals.

Autor: Li Y; Department of Health Science and Technology, ETH Zürich, 8006, Zürich, Switzerland. yanke.li@hest.ethz.ch.; Swiss Paraplegic Research (SPF), 6207, Nottwil, Switzerland. yanke.li@hest.ethz.ch., Scheel-Sailer A; Swiss Paraplegic Research (SPF), 6207, Nottwil, Switzerland.; Universitätsspital Bern, 3010, Bern, Switzerland., Riener R; Department of Health Science and Technology, ETH Zürich, 8006, Zürich, Switzerland.; Medical Faculty, University of Zürich, 8006, Zürich, Switzerland., Paez-Granados D; Department of Health Science and Technology, ETH Zürich, 8006, Zürich, Switzerland. diego.paez@hest.ethz.ch.; Swiss Paraplegic Research (SPF), 6207, Nottwil, Switzerland. diego.paez@hest.ethz.ch.
Jazyk: angličtina
Zdroj: Scientific reports [Sci Rep] 2024 Oct 23; Vol. 14 (1), pp. 25067. Date of Electronic Publication: 2024 Oct 23.
DOI: 10.1038/s41598-024-75691-9
Abstrakt: Developing machine learning (ML) methods for healthcare predictive modeling requires absolute explainability and transparency to build trust and accountability. Graphical models (GM) are key tools for this but face challenges like small sample sizes, mixed variables, and latent confounders. This paper presents a novel learning framework addressing these challenges by integrating latent variables using fast causal inference (FCI), accommodating mixed variables with predictive permutation conditional independence tests (PPCIT), and employing a systematic graphical embedding approach leveraging expert knowledge. This method ensures a transparent model structure and an explainable feature selection and modeling approach, achieving competitive prediction performance. For real-world validation, data of hospital-acquired pressure injuries (HAPI) among individuals with spinal cord injury (SCI) were used, where the approach achieved a balanced accuracy of 0.941 and an AUC of 0.983, outperforming most benchmarks. The PPCIT method also demonstrated superior accuracy and scalability over other benchmarks in causal discovery validation on synthetic datasets that closely resemble our real dataset. This holistic framework effectively addresses the challenges of mixed variables and explainable predictive modeling for disease onset, which is crucial for enabling transparency and interpretability in ML-based healthcare.
(© 2024. The Author(s).)
Databáze: MEDLINE
Nepřihlášeným uživatelům se plný text nezobrazuje