Unknown

Dataset Information

0

Med-BERT: pretrained contextualized embeddings on large-scale structured electronic health records for disease prediction.


ABSTRACT: Deep learning (DL)-based predictive models from electronic health records (EHRs) deliver impressive performance in many clinical tasks. Large training cohorts, however, are often required by these models to achieve high accuracy, hindering the adoption of DL-based models in scenarios with limited training data. Recently, bidirectional encoder representations from transformers (BERT) and related models have achieved tremendous successes in the natural language processing domain. The pretraining of BERT on a very large training corpus generates contextualized embeddings that can boost the performance of models trained on smaller datasets. Inspired by BERT, we propose Med-BERT, which adapts the BERT framework originally developed for the text domain to the structured EHR domain. Med-BERT is a contextualized embedding model pretrained on a structured EHR dataset of 28,490,650 patients. Fine-tuning experiments showed that Med-BERT substantially improves the prediction accuracy, boosting the area under the receiver operating characteristics curve (AUC) by 1.21-6.14% in two disease prediction tasks from two clinical databases. In particular, pretrained Med-BERT obtains promising performances on tasks with small fine-tuning training sets and can boost the AUC by more than 20% or obtain an AUC as high as a model trained on a training set ten times larger, compared with deep learning models without Med-BERT. We believe that Med-BERT will benefit disease prediction studies with small local training datasets, reduce data collection expenses, and accelerate the pace of artificial intelligence aided healthcare.

SUBMITTER: Rasmy L 

PROVIDER: S-EPMC8137882 | biostudies-literature |

REPOSITORIES: biostudies-literature

Similar Datasets

| S-EPMC8441576 | biostudies-literature
| S-EPMC7069636 | biostudies-literature
| S-EPMC6703965 | biostudies-literature
| S-EPMC10865188 | biostudies-literature
| S-EPMC6746103 | biostudies-literature
| S-EPMC9280463 | biostudies-literature
| PRJNA158491 | ENA
| S-EPMC6913757 | biostudies-literature
| S-EPMC8647222 | biostudies-literature
| S-EPMC7055671 | biostudies-literature