Project description:Integrating machine learning (ML) models into clinical practice presents a challenge of maintaining their efficacy over time. While existing literature offers valuable strategies for detecting declining model performance, there is a need to document the broader challenges and solutions associated with the real-world development and integration of model monitoring solutions. This work details the development and use of a platform for monitoring the performance of a production-level ML model operating in Mayo Clinic. In this paper, we aimed to provide a series of considerations and guidelines necessary for integrating such a platform into a team's technical infrastructure and workflow. We have documented our experiences with this integration process, discussed the broader challenges encountered with real-world implementation and maintenance, and included the source code for the platform. Our monitoring platform was built as an R shiny application, developed and implemented over the course of 6 months. The platform has been used and maintained for 2 years and is still in use as of July 2023. The considerations necessary for the implementation of the monitoring platform center around 4 pillars: feasibility (what resources can be used for platform development?); design (through what statistics or models will the model be monitored, and how will these results be efficiently displayed to the end user?); implementation (how will this platform be built, and where will it exist within the IT ecosystem?); and policy (based on monitoring feedback, when and what actions will be taken to fix problems, and how will these problems be translated to clinical staff?). While much of the literature surrounding ML performance monitoring emphasizes methodological approaches for capturing changes in performance, there remains a battery of other challenges and considerations that must be addressed for successful real-world implementation.
Project description:BackgroundThe coronavirus disease (COVID-19) pandemic has spread rapidly across the world, creating an urgent need for predictive models that can help healthcare providers prepare and respond to outbreaks more quickly and effectively, and ultimately improve patient care. Early detection and warning systems are crucial for preventing and controlling epidemic spread.ObjectiveIn this study, we aimed to propose a machine learning-based method to predict the transmission trend of COVID-19 and a new approach to detect the start time of new outbreaks by analyzing epidemiological data.MethodsWe developed a risk index to measure the change in the transmission trend. We applied machine learning (ML) techniques to predict COVID-19 transmission trends, categorized into three labels: decrease (L0), maintain (L1), and increase (L2). We used Support Vector Machine (SVM), Random Forest (RF), and XGBoost (XGB) as ML models. We employed grid search methods to determine the optimal hyperparameters for these three models. We proposed a new method to detect the start time of new outbreaks based on label 2, which was sustained for at least 14 days (i.e., the duration of maintenance). We compared the performance of different ML models to identify the most accurate approach for outbreak detection. We conducted sensitivity analysis for the duration of maintenance between 7 days and 28 days.ResultsML methods demonstrated high accuracy (over 94%) in estimating the classification of the transmission trends. Our proposed method successfully predicted the start time of new outbreaks, enabling us to detect a total of seven estimated outbreaks, while there were five reported outbreaks between March 2020 and October 2022 in Korea. It means that our method could detect minor outbreaks. Among the ML models, the RF and XGB classifiers exhibited the highest accuracy in outbreak detection.ConclusionThe study highlights the strength of our method in accurately predicting the timing of an outbreak using an interpretable and explainable approach. It could provide a standard for predicting the start time of new outbreaks and detecting future transmission trends. This method can contribute to the development of targeted prevention and control measures and enhance resource management during the pandemic.
Project description:Machine learning (ML), as an advanced domain of artificial intelligence (AI), is progressively changing our view of the world. By implementing its algorithms, our ability to detect previously undiscoverable patterns in data has the potential to revolutionize predictive analytics. Scoliosis, as a relatively specialized branch in the spine field, mainly covers the pediatric, adult and the elderly populations, and its diagnosis and treatment remain difficult. With recent efforts and interdisciplinary cooperation, ML has been widely applied to investigate issues related to scoliosis, and surprisingly augment a surgeon's ability in clinical practice related to scoliosis. Meanwhile, ML models penetrate in every stage of the clinical practice procedure of scoliosis. In this review, we first present a brief description of the application of ML in the clinical practice procedures regarding scoliosis, including screening, diagnosis and classification, surgical decision making, intraoperative manipulation, complication prediction, prognosis prediction and rehabilitation. Meanwhile, the ML models and specific applications adopted are presented. Additionally, current limitations and future directions are briefly discussed regarding its use in the field of scoliosis. We believe that the implementation of ML is a promising revolution to assist surgeons in all aspects of clinical practice related to scoliosis in the near future.
Project description:Gene expression profiles were generated from 199 primary breast cancer patients. Samples 1-176 were used in another study, GEO Series GSE22820, and form the training data set in this study. Sample numbers 200-222 form a validation set. This data is used to model a machine learning classifier for Estrogen Receptor Status. RNA was isolated from 199 primary breast cancer patients. A machine learning classifier was built to predict ER status using only three gene features.
Project description:The contact and interaction of human is considered to be one of the important factors affecting the epidemic transmission, and it is critical to model the heterogeneity of individual activities in epidemiological risk assessment. In digital society, massive data makes it possible to implement this idea on large scale. Here, we use the mobile phone signaling to track the users' trajectories and construct contact network to describe the topology of daily contact between individuals dynamically. We show the spatiotemporal contact features of about 7.5 million mobile phone users during the outbreak of COVID-19 in Shanghai, China. Furthermore, the individual feature matrix extracted from contact network enables us to carry out the extreme event learning and predict the regional transmission risk, which can be further decomposed into the risk due to the inflow of people from epidemic hot zones and the risk due to people close contacts within the observing area. This method is much more flexible and adaptive, and can be taken as one of the epidemic precautions before the large-scale outbreak with high efficiency and low cost.
Project description:Background:The number of cases from the coronavirus disease 2019 (COVID-19) global pandemic has overwhelmed existing medical facilities and forced clinicians, patients, and families to make pivotal decisions with limited time and information. Main body:While machine learning (ML) methods have been previously used to augment clinical decisions, there is now a demand for "Emergency ML." Throughout the patient care pathway, there are opportunities for ML-supported decisions based on collected vitals, laboratory results, medication orders, and comorbidities. With rapidly growing datasets, there also remain important considerations when developing and validating ML models. Conclusion:This perspective highlights the utility of evidence-based prediction tools in a number of clinical settings, and how similar models can be deployed during the COVID-19 pandemic to guide hospital frontlines and healthcare administrators to make informed decisions about patient care and managing hospital volume.
Project description:BackgroundData drift can negatively impact the performance of machine learning algorithms (MLAs) that were trained on historical data. As such, MLAs should be continuously monitored and tuned to overcome the systematic changes that occur in the distribution of data. In this paper, we study the extent of data drift and provide insights about its characteristics for sepsis onset prediction. This study will help elucidate the nature of data drift for prediction of sepsis and similar diseases. This may aid with the development of more effective patient monitoring systems that can stratify risk for dynamic disease states in hospitals.MethodsWe devise a series of simulations that measure the effects of data drift in patients with sepsis. We simulate multiple scenarios in which data drift may occur, namely the change in the distribution of the predictor variables (covariate shift), the change in the statistical relationship between the predictors and the target (concept shift), and the occurrence of a major healthcare event (major event) such as the COVID-19 pandemic. We measure the impact of data drift on model performances, identify the circumstances that necessitate model retraining, and compare the effects of different retraining methodologies and model architecture on the outcomes. We present the results for two different MLAs, eXtreme Gradient Boosting (XGB) and Recurrent Neural Network (RNN).ResultsOur results show that the properly retrained XGB models outperform the baseline models in all simulation scenarios, hence signifying the existence of data drift. In the major event scenario, the area under the receiver operating characteristic curve (AUROC) at the end of the simulation period is 0.811 for the baseline XGB model and 0.868 for the retrained XGB model. In the covariate shift scenario, the AUROC at the end of the simulation period for the baseline and retrained XGB models is 0.853 and 0.874 respectively. In the concept shift scenario and under the mixed labeling method, the retrained XGB models perform worse than the baseline model for most simulation steps. However, under the full relabeling method, the AUROC at the end of the simulation period for the baseline and retrained XGB models is 0.852 and 0.877 respectively. The results for the RNN models were mixed, suggesting that retraining based on a fixed network architecture may be inadequate for an RNN. We also present the results in the form of other performance metrics such as the ratio of observed to expected probabilities (calibration) and the normalized rate of positive predictive values (PPV) by prevalence, referred to as lift, at a sensitivity of 0.8.ConclusionOur simulations reveal that retraining periods of a couple of months or using several thousand patients are likely to be adequate to monitor machine learning models that predict sepsis. This indicates that a machine learning system for sepsis prediction will probably need less infrastructure for performance monitoring and retraining compared to other applications in which data drift is more frequent and continuous. Our results also show that in the event of a concept shift, a full overhaul of the sepsis prediction model may be necessary because it indicates a discrete change in the definition of sepsis labels, and mixing the labels for the sake of incremental training may not produce the desired results.
Project description:Cells modify their internal organization during continuous state-transitions, supporting functions from cell division to differentiation. However, tools to measure dynamic physiological states of individual transitioning cells are lacking. We combined live-cell imaging and machine learning to monitor ERK1/2-inhibited primary murine skeletal muscle precursor cells, that transition rapidly and robustly from proliferating myoblasts to post-mitotic myocytes and then fuse, forming multinucleated myotubes. Our model, trained using motility and actin intensity features from single cell tracking data, effectively tracked real-time continuous differentiation, revealing that differentiation occurs 7.5-14.5 hours post-induction, followed by fusion ~3 hours later. Co-inhibition of ERK1/2 and p38 led to differentiation without fusion. Our model inferred co-inhibition leads to terminal differentiation, indicating that p38 is specifically required for transitioning from terminal differentiation to fusion. Our model also predicted that co-inhibition leads to changes in actin dynamics. Mass spectrometry supported these in silico predictions and suggested novel fusion and maturation regulators downstream of differentiation. Collectively, this approach can be adapted to various biological processes to uncover novel links between dynamic single-cell states and their functional outcomes.
Project description:ObjectiveThis review introduces interpretable predictive machine learning approaches, natural language processing, image recognition, and reinforcement learning methodologies to familiarize end users.BackgroundAs machine learning, artificial intelligence, and generative artificial intelligence become increasingly utilized in clinical medicine, it is imperative that end users understand the underlying methodologies.MethodsThis review describes publicly available datasets that can be used with interpretable predictive approaches, natural language processing, image recognition, and reinforcement learning models, outlines result interpretation, and provides references for in-depth information about each analytical framework.ResultsThis review introduces interpretable predictive machine learning models, natural language processing, image recognition, and reinforcement learning methodologies.ConclusionsInterpretable predictive machine learning models, natural language processing, image recognition, and reinforcement learning are core machine learning methodologies that underlie many of the artificial intelligence methodologies that will drive the future of clinical medicine and surgery. End users must be well versed in the strengths and weaknesses of these tools as they are applied to patient care now and in the future.