WeChat Mini Program
Old Version Features

A Multi-Center Study on the Adaptability of a Shared Foundation Model for Electronic Health Records

NPJ DIGITAL MEDICINE(2024)

Hosp Sick Children

Cited 2|Views38
Abstract
Foundation models are transforming artificial intelligence (AI) in healthcare by providing modular components adaptable for various downstream tasks, making AI development more scalable and cost-effective. Foundation models for structured electronic health records (EHR), trained on coded medical records from millions of patients, demonstrated benefits including increased performance with fewer training labels, and improved robustness to distribution shifts. However, questions remain on the feasibility of sharing these models across hospitals and their performance in local tasks. This multi-center study examined the adaptability of a publicly accessible structured EHR foundation model (FMSM), trained on 2.57 M patient records from Stanford Medicine. Experiments used EHR data from The Hospital for Sick Children (SickKids) and Medical Information Mart for Intensive Care (MIMIC-IV). We assessed both adaptability via continued pretraining on local data, and task adaptability compared to baselines of locally training models from scratch, including a local foundation model. Evaluations on 8 clinical prediction tasks showed that adapting the off-the-shelf FMSM matched the performance of gradient boosting machines (GBM) locally trained on all data while providing a 13% improvement in settings with few task-specific training labels. Continued pretraining on local data showed FMSM required fewer than 1% of training examples to match the fully trained GBM’s performance, and was 60 to 90% more sample-efficient than training local foundation models from scratch. Our findings demonstrate that adapting EHR foundation models across hospitals provides improved prediction performance at less cost, underscoring the utility of base foundation models as modular components to streamline the development of healthcare AI.
More
Translated text
Key words
Electronic Health Records,Medical Concept Embedding,Patient Similarity
PDF
Bibtex
AI Read Science
Must-Reading Tree
Example
Generate MRT to find the research sequence of this paper
Data Disclaimer
The page data are from open Internet sources, cooperative publishers and automatic analysis results through AI technology. We do not make any commitments and guarantees for the validity, accuracy, correctness, reliability, completeness and timeliness of the page data. If you have any questions, please contact us by email: report@aminer.cn
Chat Paper

要点】:本研究探讨了公开可用的结构化电子健康记录(EHR)基础模型(FMSM)在不同医院的适应性,该模型基于斯坦福医学院257万患者的编码医疗记录训练而成。研究发现,通过在本地数据上持续预训练,该模型在8项临床预测任务上的表现可与在全部数据上本地训练的梯度提升机(GBM)相媲美,在标签较少的环境中还能提供13%的改进。

方法】:研究通过比较在本地数据上持续预训练的FMSM与从零开始训练的本地基础模型等基线方法,评估了FMSM的适应性。

实验】:实验使用了来自多伦多儿童医院(SickKids)和医疗信息重症监护数据库(MIMIC-IV)的EHR数据。结果显示,FMSM在本地数据上进行少量训练样本即可达到与完全训练的GBM相匹配的表现,且比从零开始训练本地基础模型更加节省样本。