Autor: |
Xiong X, Sweet SM, Liu M, Hong C, Bonzel CL, Panickan VA, Zhou D, Wang L, Costa L, Ho YL, Geva A, Mandl KD, Cheng S, Xia Z, Cho K, Gaziano JM, Liao KP, Cai T, Cai T |
Jazyk: |
angličtina |
Zdroj: |
MedRxiv : the preprint server for health sciences [medRxiv] 2023 Oct 02. Date of Electronic Publication: 2023 Oct 02. |
DOI: |
10.1101/2023.09.29.23296239 |
Abstrakt: |
Though electronic health record (EHR) systems are a rich repository of clinical information with large potential, the use of EHR-based phenotyping algorithms is often hindered by inaccurate diagnostic records, the presence of many irrelevant features, and the requirement for a human-labeled training set. In this paper, we describe a knowledge-driven online multimodal automated phenotyping (KOMAP) system that i) generates a list of informative features by an online narrative and codified feature search engine (ONCE) and ii) enables the training of a multimodal phenotyping algorithm based on summary data. Powered by composite knowledge from multiple EHR sources, online article corpora, and a large language model, features selected by ONCE show high concordance with the state-of-the-art AI models (GPT4 and ChatGPT) and encourage large-scale phenotyping by providing a smaller but highly relevant feature set. Validation of the KOMAP system across four healthcare centers suggests that it can generate efficient phenotyping algorithms with robust performance. Compared to other methods requiring patient-level inputs and gold-standard labels, the fully online KOMAP provides a significant opportunity to enable multi-center collaboration. |
Databáze: |
MEDLINE |
Externí odkaz: |
|