Unknown

Dataset Information

0

Study on the semi-supervised learning-based patient similarity from heterogeneous electronic medical records.


ABSTRACT:

Background

A new learning-based patient similarity measurement was proposed to measure patients' similarity for heterogeneous electronic medical records (EMRs) data.

Methods

We first calculated feature-level similarities according to the features' attributes. A domain expert provided patient similarity scores of 30 randomly selected patients. These similarity scores and feature-level similarities for 30 patients comprised the labeled sample set, which was used for the semi-supervised learning algorithm to learn the patient-level similarities for all patients. Then we used the k-nearest neighbor (kNN) classifier to predict four liver conditions. The predictive performances were compared in four different situations. We also compared the performances between personalized kNN models and other machine learning models. We assessed the predictive performances by the area under the receiver operating characteristic curve (AUC), F1-score, and cross-entropy (CE) loss.

Results

As the size of the random training samples increased, the kNN models using the learned patient similarity to select near neighbors consistently outperformed those using the Euclidean distance to select near neighbors (all P values < 0.001). The kNN models using the learned patient similarity to identify the top k nearest neighbors from the random training samples also had a higher best-performance (AUC: 0.95 vs. 0.89, F1-score: 0.84 vs. 0.67, and CE loss: 1.22 vs. 1.82) than those using the Euclidean distance. As the size of the similar training samples increased, which composed the most similar samples determined by the learned patient similarity, the performance of kNN models using the simple Euclidean distance to select the near neighbors degraded gradually. When exchanging the role of the Euclidean distance, and the learned patient similarity in selecting the near neighbors and similar training samples, the performance of the kNN models gradually increased. These two kinds of kNN models had the same best-performance of AUC 0.95, F1-score 0.84, and CE loss 1.22. Among the four reference models, the highest AUC and F1-score were 0.94 and 0.80, separately, which were both lower than those for the simple and similarity-based kNN models.

Conclusions

This learning-based method opened an opportunity for similarity measurement based on heterogeneous EMR data and supported the secondary use of EMR data.

SUBMITTER: Wang N 

PROVIDER: S-EPMC8323210 | biostudies-literature |

REPOSITORIES: biostudies-literature

Similar Datasets

| S-EPMC9729175 | biostudies-literature
| S-EPMC6788002 | biostudies-literature
| PRJNA158491 | ENA
| S-EPMC10507898 | biostudies-literature
| S-EPMC3261909 | biostudies-literature
2019-11-13 | GSE140262 | GEO
| S-EPMC5771885 | biostudies-literature
| S-EPMC9653502 | biostudies-literature
| S-EPMC5519723 | biostudies-literature
| S-EPMC10858490 | biostudies-literature