Abstract
Although in-hospital cardiac arrest is uncommon, it has a high mortality rate. Risk identification of at-risk patients is critical for post-cardiac arrest survival rates. Early warning scoring systems are generally used to identify hospitalized patients at risk of deterioration. However, these systems often require clinical data that are not always regularly measured. We developed a more accurate, machine learning-based model to predict clinical deterioration. The time series early warning score (TEWS) used only heart rate, systolic blood pressure, and respiratory data, which are regularly measured in general wards. We tested the performance of the TEWS in two tasks performed with data from the electronic medical records of 16,865 adult admissions and compared the results with those of other classifications. The TEWS detected more deteriorations with the same level of specificity as the different algorithms did when inputting vital signs data from 48 h before an event. Our framework improved in-hospital cardiac arrest prediction and demonstrated that previously obtained vital signs data can be used to identify at-risk patients in real-time. This model may be an alternative method for detecting patient deterioration.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Introduction
Although uncommon, in-hospital cardiac arrest (IHCA) has a high mortality rate. In the United States, approximately 200,000 cardiac arrests are treated annually in hospitalized patients, and the event rate of cardiac arrest is about 0.92 per 1000 bed days1. The post-discharge survival rate for patients with cardiac arrest is nearly 25% in the United States and less than 20% worldwide2, 3. Intra-arrest factors, including whether the arrest was witnessed or monitored and whether it occurred during daytime hours, were reported to be associated with increased odds of survival4. Identifying at-risk patients has become critical because post-cardiac arrest survival rates are associated with care team awareness.
Various early warning scoring systems have been used to detect hospitalized patients at high risk of deterioration5,6,7,8,9,10,11. Such scoring systems are generally developed by selecting appropriate variables associated with prediction outcomes. Most early warning scoring systems, such as the Modified Early Warning Score (MEWS)12, use vital signs, including temperature, heart rate, respiratory rate, and blood pressure, for clinical judgments. Because the areas under the receiver operating characteristic curve (AUROC) for MEWS have been below 0.7 in many studies, research has been conducted to include more types of clinical data, including laboratory results, demographics, and heart rate variability, to increase prediction performance5,6,7, 13,14,15. These methods have been demonstrated to be more accurate, resulting in fewer false alarms and more favorable detection. However, the methods may not be feasible in care units where these clinical data are not regularly measured. Information about the research to predict IHCA is presented in Table 1.
The increasing application of artificial intelligence and machine learning (ML) systems has fundamentally altered the biomedical field from molecule to disease level16. ML facilitates the automatic analysis of highly complex data and produces meaningful results. ML models can improve prediction accuracy with the same data or reduce features with the same performance17. Cho and Kwon used vital signs over the past 8 h to develop a deep learning-based early warning score to predict deterioration in patients in general wards accurately. Some research used ML with continuous vital signs to predict deterioration in intensive care units (ICUs)9. However, continuous vital signs measurements may not be available in general wards.
Therefore, in this study, we sought to develop a more accurate machine-learning model (the time series early warning score [TEWS]) for predicting clinical deterioration using only heart rate, systolic blood pressure, and respiratory data. These vital signs which are regularly measured in general wards. This model may be an alternative to the MEWS system.
Methods
Ethics declarations
This retrospective cohort study was approved by the Institutional Review Board (IRB) of the En-Chu-Kong Hospital (IRB number: ECKIRB1071001). We confirm that all experiments were performed in accordance with relevant guidelines and regulations. The data retrieved from electronic health records (EHRs) were de-identified by an IT specialist and could not be linked to the patients’ identity by the research team. The need for written informed consent was waived and confirmed by the En-Chu-Kong Hospital IRB (ECKIRB1071001) because this was a retrospective cohort study with de-identified data.
Study setting and population
The study population comprised inpatients from a community general hospital. The study data set was EHRs of the adult inpatients (aged ≥ 20 years) who visited the hospital between August 2016 and September 2019. Each patient’s information was anonymized and de-identified before analysis.
Data sources
We used five vital signs as predictor features: systolic blood pressure (SBP), diastolic blood pressure (DBP), heart rate (HR), respiratory rate (RR), and body temperature (BT). Medical staff measured these vital signs regularly, at least two to three times per day during the day, night, and early morning. We defined the time window (TW) measurement as 8 h. Therefore, 1 day comprised three TWs. We considered features measured during each TW; each TW had five vital signs measurements.
The hospital data were divided according to date into a derivation (August 2016–November 2018) and a validation set (December 2018–September 2019). The derivation and validation sets were used to develop the TEWS and to determine the TEWS parameters, respectively. We used AUROC and area under the precision-recall curve (AUPRC) values for binary classification. The characteristics of the study population are listed in Table 2.
Outcomes
The primary outcome of interest was cardiac arrest, defined as a loss of pulse with attempted resuscitation. We examined the collected EHRs to identify the exact time of each outcome. We categorized the selected inpatients into positive and negative groups. The positive group contained inpatients with a cardiac arrest event in the general wards. For patients with several cardiac arrest events during their stay at the hospital, we used only the first event. The negative group contained inpatients who did not stay in the ICU and had no cardiac arrest event during the study period.
The TEWS was compared with the MEWS and other classifiers. We then performed a time analysis of the vital signs and predicted whether a patient would be IHCA-positive by using the features recorded in one, three, or six TWs (i.e., 8, 24, or 48 h, respectively).
Model development
Data preprocessing
Because the collected EHRs may have contained human or system errors, our data had the potential to have missing values. For example, medical staff may have failed to measure vital signs during some TWs, leading to missing data in the TWs. To compensate for the missing values, we applied the multiple imputation by chained equations approach18. The advantage of this approach is that it not only restores the natural variability of missing values but also incorporates the uncertainty resulting from the missing data, thus enabling a valid statistical inference. In the event of duplicate data for the same TW, we used the maximum value.
Values of the features in our data were distributed over a wide range, which increased the difficulty of training the classifier. Therefore, we used standard scores (commonly referred to as z-scores) to adjust the values of all features.
Handing imbalanced data
In many real-life problems, especially in the medical field, data sets are imbalanced; that is, the class distribution in such data sets is severely skewed. Similarly, our data set was imbalanced. However, most machine-learning algorithms are most effective when the number of samples in each class is nearly balanced. Failure to manage imbalanced data sets can adversely affect the performance of classifiers19; in machine-learning classifiers, biases in training data sets can lead to minority classes being ignored entirely. Accordingly, imbalanced data sets can be managed through under-sampling (in which samples are deleted from the majority class) and oversampling (in which samples from the minority class are duplicated). In our previous study, we under-sampled IHCA-negative samples for detection. The results indicated that our approach effectively solved the imbalance in the data set used for detecting cardiac arrest20.
Accordingly, in the present study, we used a modified weight balancing method in place of an oversampling or under-sampling approach to balance our data set. We used this method when the number of samples in one of our classes was substantially higher than in the other. This method modified the class weights according to the ratio of IHCA-positive to IHCA-negative samples to ensure that all classes contributed equally to the loss. Furthermore, we applied focal loss to balance the weight of our training samples21. When an imbalanced data set is used for classification, the majority class is adequately represented in the classification because more data are available for this class; however, the minority class is not sufficiently represented. We applied “focal loss” to prevent this situation. Focal loss assigns relatively high weight to the minority class during training to ensure that the class is adequately represented in the classification. Therefore, we applied focal loss and the weight-balancing method to our imbalanced data in developing the TEWS.
We used features obtained in 1, 3, and 6 TWs (i.e., 8, 24, and 48 h, respectively). Each TW contained one set of features. The study workflow is illustrated in Fig. 1.
Time series early warning score [TEWS] model
The proposed TEWS comprises three recurrent neural network (RNN) layers with LSTM22, 23. An RNN is a neural network with feedback loops, which enable it to process sequential data, such as EHRs24. The architecture of TEWS and LSTM is illustrated in Fig. 2. The LSTM unit comprises a cell, an input gate, an output gate, and a forget gate. The cell remembers values over arbitrary time intervals, and three gates regulate the flow of information into and out of the cell. LSTM deals with the time series data well. Therefore, TEWS can adequately process time-series data.
We train the TEWS model using the training data set and assess TEWS model performance using the validation data set. In our TEWS model, the training and validation data set are split into an 8:2 ratio. Figure 3 presents the algorithm to create six time windows for each vital sign of all inpatients.
Performance evaluation
Benchmarking with contemporary algorithms
We implemented our LSTM-based system by using the scikit-learn package in Python25; the neural networks were implemented in Keras, with TensorFlow serving as the backend engine. The scikit-learn package was also used to implement some classification for comparison26, namely naïve Bayes27, support vector machine (SVM)28, 29, AdaBoost30, 31, k-nearest neighbor32, 33, classification and regression tree34, and C4.5 decision tree35. We also used gradient boosting36, logistic regression37, and random forest38 algorithms. Gradient boosting produces a prediction model in the form of an ensemble of weak prediction models, and it can be considered an optimization algorithm on a suitable cost function39. Logistic regression is a statistical model used to model the probability of a specific class. It uses a logistic function to model a binary dependent variable in its basic form. Random forest is an ensemble learning method for classification, regression, and other tasks; it involves constructing a multitude of decision trees during training and outputting a class (that is, the mode of the classes or the mean prediction of the individual trees). Because of our imbalanced data, our proposed TEWS sets class weights according to the ratio of IHCA-positive to IHCA-negative samples. We compared the prediction performance of these classifications and TEWS.
Predicted probabilities were calculated for each observation of validation data set from each derived model to understand the accuracy of results within the context of the literature. The result of MEWS was also calculated. The AUROCs and the AUPRCs were determined according to whether an event occurred within eight hours of each observation because these are standard early warning score comparisons metrics.
Feature selection
Feature selection involves selecting the best features from a set of valuable features for discriminating between classes. Feature selection can be completed through an elimination process. Feature elimination methods can be broadly classified into filter and wrapper methods. In wrapper methods, the feature selection criterion is the predictor’s performance (i.e., the predictor is wrapped on a search algorithm that will identify the subset with the highest predictor performance). Sequential backward selection (SBS) algorithms are straightforward and greedy search algorithms. An SBS algorithm can be used for feature selection. The algorithm removes one feature from a complete set of features at a time, leading to a minimal decrease in predictor performance. SBS performs most favorably when the optimal subset has fewer features40, 41.
Results
A total of 16,865 adult admissions were included in this study. 118 (0.7%) of these patients experienced cardiac arrest in a general ward (Table 2).We further describe the characteristics of IHCA-positive and IHCA-negative data in Fig. 4.
We used two tasks to test the performance of our proposed TEWS. We then compared the results of TEWS and these classifications. The tasks are detailed as follows.
First task
Prediction of IHCA Using Features (Vital Signs) Recorded in One Time Window (8 h), Three TWs (24 h), and Six TWs (48 h).
In the 1TW (8 h) group, we applied one set of five vital signs (i.e., features obtained in one TW) to predict IHCA events using the proposed TEWS. The performance of the TEWS model was then compared with that of the MEWS and other classifiers, as displayed in Fig. 5. ROC and PR curve are illustrated in supplementary files. The support vector machine (SVM) and logistic regression algorithms had the highest AUROC values (0.729 and 0.721, respectively), followed by gradient boosting (0.712) and the TEWS (0.688). However, no classifier adequately outperformed the MEWS.
In the 3TW (24 h) group, we applied features recorded in three TWs (24 h) to predict IHCA events using the TEWS. Each TW included a single set of vital signs; therefore, three TWs with five vital signs' measurements contained 15 features. The AUROC value of the TEWS (0.762) was superior to those of the logistic regression (0.730), random forest (0.676), MEWS (0.649), and other algorithms.
In the 6TW group (48 h), we applied features recorded in six TWs (48 h) to predict IHCA events using the TEWS. The AUROC value of the TEWS (0.808) was superior to those of gradient boosting (0.768), SVM (0.747), random forest (0.733), and other algorithms. TEWS performed well regardless of the 1TW, 3TW, and 6TW groups.
Most classification algorithms exhibited similar performance levels when we used features from a single TW. The AUROCs of these models were within 0.62–0.73 (AUROC of MEWS: 0.65). When we used more TWs, some classifications exhibited improved performance levels. Our TEWS demonstrated more favorable prediction in the 6TW (AUROC = 0.808, area under the precision-recall curve [AUPRC] = 0.052) than the MEWS did (AUROC = 0.649, AUPRC = 0.015).
Second task
Prediction of IHCA Using Features Selected Through Sequential Backward Selection (SBS) algorithm.
The TEWS had the most favorable performance in the first task when six TWs (48 h) were included. Because six TWs comprise 30 features, we sought a means of reducing the required features without compromising performance. We selected the most relevant features in the six TWs by using an SBS algorithm. These selected features are presented in Fig. 6. The first TW was the time window closest to the cardiopulmonary resuscitation time for patients who were IHCA-positive. Heart rate, respiratory rate, and systolic blood pressure were the most relevant features for predicting IHCA events. The top five features were heart rate in the first, fourth, and fifth TWs, respiratory rate, and systolic blood pressure in the first TW.
Furthermore, we applied the five selected features to the TEWS model and the other algorithms. The performance of the algorithms using the five features was then compared with that of the MEWS and other classifiers, as displayed in Fig. 6. The TEWS (AUROC = 0.875, AUPRC = 0.087), Adaboost (AUROC = 0.958, AUPRC = 0.110), and logistic regression (AUROC = 0.845, AUPRC = 0.050) achieved their highest performance.
Discussion
In this study, we used only vital signs in two days to predict cardiac arrest. Our results revealed that the TEWS model using features from six TWs outperformed the other classification algorithms. When the TEWS was implemented using features from six TWs, its prediction performance (AUROC = 0.808, AUPRC = 0.052) was higher than that when it was implemented using features from a single TW (AUROC = 0.688, AUPRC = 0.041) and that of the MEWS (AUROC = 0.649, AUPRC = 0.015). The improved model performance suggests that more information on vital signs data could be obtained from different TWs.
Similar studies have also reported that the essential predictor variables for clinical deterioration are respiratory rate, heart rate, age, and systolic blood pressure5. Our study proposed TEWS with only five features from six TWs (respiratory rate, systolic blood pressure in the most recent TW, and three heart rate values in different TWs; AUROC = 0.875, AUPRC = 0.087) outperformed the other classifications. The result indicates that trends in heart rate variation rather than absolute heart rate value alone are essential data.
Our study has several strengths compared with others. First, although some deep learning-based early warning systems can accurately predict patients’ deterioration in intensive care settings, our TEWS can be implemented in general wards or long-term care units. Second, we used a longer observation time (48 h) for vital signs and deep learning-based method to increase the accuracy of predicting cardiac arrest without additional variables. Third, we developed our model using only vital signs. This model can be widely implemented in any system that is equipped for MEWS. The minimum requirement for TEWS is a single personal computer with the capacity for manual entry of vital signs or automatic extraction from EHRs.
Our study has several limitations. First, this was a single-center study at a community general hospital. Therefore, the results may not be generalizable to other settings. Second, our TEWS had the best performance when vital signs from 48 h were included; it did not have a higher prediction performance on the first day of admission compared with other early warning systems. However, prehospital heart rate collected through wearable devices may be an alternative data source for the model. Third, we could not accurately predict several cases of cardiac arrest in our data set. Some cases involved sudden collapses, such as a pulmonary embolism after cesarean section or postoperative airway obstruction with hematoma, which were not predicted. In addition, TEWS could not detect deterioration between two time windows. This is a limitation of noncontinuous vital signs-based prediction models.
Conclusion
We developed a LSTM-based model using vital signs data from 48 h to predict IHCA. TEWS detected more deteriorations with the same level of specificity as the other algorithms. Our results demonstrate that the 6TW-TEWS and five feature–TEWS more favorably predicted deterioration than the different algorithms did with 1TW (Supplementary Information).
Our framework improved IHCA prediction and demonstrated the feasibility of using only previously obtained vital signs data to detect critical illness in ward patients in real-time. Our TEWS model may be an alternative method for detecting patient deterioration.
Data availability
The dataset analyzed in this study is available from the corresponding authors on reasonable request and upon approval by the Institutional Review Board (IRB) of the first authors’ institution to share the data. We will share our dataset including vital signs data with 6 time windows. The hyperlink of our training dataset is https://www.cs.nccu.edu.tw/~sichiu/allz_train_6tw.csv. The hyperlink of our test dataset is https://www.cs.nccu.edu.tw/~sichiu/allz_test_6tw.csv.
References
Merchant, R. M. et al. Incidence of treated cardiac arrest in hospitalized patients in the United States. Crit. Care Med. 39, 2401–2406. https://doi.org/10.1097/CCM.0b013e3182257459 (2011).
Andersen, L. W., Holmberg, M. J., Berg, K. M., Donnino, M. W. & Granfeldt, A. In-hospital cardiac arrest: A review. JAMA 321, 1200–1210. https://doi.org/10.1001/jama.2019.1696 (2019).
Chen, L. M., Nallamothu, B. K., Spertus, J. A., Li, Y. & Chan, P. S. Association between a hospital’s rate of cardiac arrest incidence and cardiac arrest survival. JAMA Intern. Med. 173, 1186–1195. https://doi.org/10.1001/jamainternmed.2013.1026 (2013).
Fernando, S. M. et al. Pre-arrest and intra-arrest prognostic factors associated with survival after in-hospital cardiac arrest: Systematic review and meta-analysis. BMJ (Clin. Res. Ed.) 367, 6373. https://doi.org/10.1136/bmj.l6373 (2019).
Churpek, M. M. et al. Multicenter comparison of machine learning methods and conventional regression for predicting clinical deterioration on the wards. Crit. Care Med. 44, 368–374. https://doi.org/10.1097/CCM.0000000000001571 (2016).
Green, M. et al. Comparison of the between the flags calling criteria to the MEWS, NEWS and the electronic cardiac arrest risk triage (eCART) score for the identification of deteriorating ward patients. Resuscitation 123, 86–91. https://doi.org/10.1016/j.resuscitation.2017.10.028 (2018).
Bartkowiak, B. et al. Validating the electronic cardiac arrest risk triage (eCART) score for risk stratification of surgical inpatients in the postoperative setting: Retrospective cohort study. Ann. Surg. https://doi.org/10.1097/sla.0000000000002665 (2018).
Kwon, J. M., Lee, Y., Lee, Y., Lee, S. & Park, J. An algorithm based on deep learning for predicting in-hospital cardiac arrest. J. Am. Heart Assoc. 7, 8678. https://doi.org/10.1161/JAHA.118.008678 (2018).
Kim, J., Chae, M., Chang, H. J., Kim, Y. A. & Park, E. Predicting cardiac arrest and respiratory failure using feasible artificial intelligence with simple trajectories of patient data. J. Clin. Med. https://doi.org/10.3390/jcm8091336 (2019).
Cho, K.-J. et al. Detecting patient deterioration using artificial intelligence in a rapid response system. Crit. Care Med. 48, e285–e289. https://doi.org/10.1097/ccm.0000000000004236 (2020).
Kim, S. H. et al. Predicting severe outcomes using national early warning score (NEWS) in patients identified by a rapid response system: A retrospective cohort study. Sci. Rep. 11, 18021. https://doi.org/10.1038/s41598-021-97121-w (2021).
Subbe, C. P., Kruger, M., Rutherford, P. & Gemmel, L. Validation of a modified early warning score in medical admissions. QJM Month. J. Assoc. Phys. 94, 521–526. https://doi.org/10.1093/qjmed/94.10.521 (2001).
Churpek, M. M., Yuen, T. C., Park, S. Y., Gibbons, R. & Edelson, D. P. Using electronic health record data to develop and validate a prediction model for adverse outcomes in the wards*. Crit. Care Med. 42, 841–848. https://doi.org/10.1097/ccm.0000000000000038 (2014).
Sessa, F. et al. Heart rate variability as predictive factor for sudden cardiac death. Aging (Albany NY) 10, 166–177. https://doi.org/10.18632/aging.101386 (2018).
Liu, N. et al. An intelligent scoring system and its application to cardiac arrest prediction. IEEE Trans. Inf. Technol. Biomed. 16, 1324–1331. https://doi.org/10.1109/TITB.2012.2212448 (2012).
Zhang, Y., Lei, X., Fang, Z. & Pan, Y. CircRNA-disease associations prediction based on metapath2vec++ and matrix factorization. Big Data Mining Anal. 3, 280–291. https://doi.org/10.26599/BDMA.2020.9020025 (2020).
Akel, M. A., Carey, K. A., Winslow, C. J., Churpek, M. M. & Edelson, D. P. Less is more: Detecting clinical deterioration in the hospital with machine learning using only age, heart rate, and respiratory rate. Resuscitation 168, 6–10. https://doi.org/10.1016/j.resuscitation.2021.08.024 (2021).
van Buuren, S. & Groothuis-Oudshoorn, K. Mice: Multivariate imputation by chained equations in R. J. Stat. Softw. 45, 1–67. https://doi.org/10.18637/jss.v045.i03 (2011).
Wei, X. et al. An ensemble model for diabetes diagnosis in large-scale and imbalanced dataset. In Proc. Computing Frontiers Conference, 71–78. https://doi.org/10.1145/3075564.3075576 (2017).
Chang, H.-K. et al. Early detecting in-hospital cardiac arrest based on machine learning on imbalanced data. In 2019 IEEE International Conference on Healthcare Informatics (ICHI), 1–10. https://doi.org/10.1109/ICHI.2019.8904504 (2019).
Lin, T.-Y., Goyal, P., Girshick, R., He, K. & Dollar, P. Focal loss for dense object detection. IEEE Trans. Pattern Anal. Mach. Intell. 42, 318–327. https://doi.org/10.1109/tpami.2018.2858826 (2020).
Hochreiter, S. & Schmidhuber, J. R. Long short-term memory. Neural Comput. 9, 1735–1780. https://doi.org/10.1162/neco.1997.9.8.1735 (1997).
Hou, C., Wu, J., Cao, B. & Fan, J. A deep-learning prediction model for imbalanced time series data forecasting. Big Data Mining Anal. 4, 266–278. https://doi.org/10.26599/BDMA.2021.9020011 (2021).
Esteban, C., Staeck, O., Baier, S., Yang, Y. & Tresp, V. Predicting clinical events by combining static and dynamic information using recurrent neural networks. In 2016 IEEE International Conference on Healthcare Informatics (ICHI), 93–101. https://doi.org/10.1109/ICHI.2016.16 (2016).
Pedregosa, F. et al. Scikit-learn: Machine learning in python. J. Mach. Learn. Res. 12, 2825–2830. https://doi.org/10.48550/arXiv.1201.0490 (2011).
Wu, X. et al. Top 10 algorithms in data mining. Knowl. Inf. Syst. 14, 1–37. https://doi.org/10.1007/s10115-007-0114-2 (2008).
Hand, D. J. & Yu, K. Idiot’s Bayes: Not so stupid after all? Int. Stat. Rev. 69, 385–398. https://doi.org/10.2307/1403452 (2001).
Cortes, C. & Vapnik, V. Support-vector networks. Mach. Learn. 20, 273–297. https://doi.org/10.1007/BF00994018 (1995).
Abe, S. Support Vector Machines for Pattern Classification (Springer, 2005).
Freund, Y. & Schapire, R. E. Experiments with a new boosting algorithm. In Proc. Thirteenth International Conference on International Conference on Machine Learning (ICML '96), 148–156. https://doi.org/10.5555/3091696.3091715 (1996).
Schapire, R. E. A brief introduction to boosting. In Proc. 16th International Joint Conference on Artificial Intelligence—Volume 2, 1401–1406. https://doi.org/10.5555/1624312.1624417 (1999).
Fix, E. & Hodges, J. L. Discriminatory analysis. Nonparametric discrimination: Consistency properties. Int. Stat. Rev. 57, 238–247. https://doi.org/10.2307/1403797 (1989).
Tan, P.-N., Steinbach, M., Karpatne, A. & Kumar, V. Introduction to Data Mining 2nd edn. (Pearson, 2018).
Breiman, L., Friedman, J., Olshen, R. & Stone, C. Classification and Regression Trees (2017).
Quinlan, J. R. C4.5: Programs for Machine Learning (Elsevier, 2014).
Friedman, J. H. Stochastic gradient boosting. Comput. Stat. Data Anal. 38, 367–378. https://doi.org/10.1016/S0167-9473(01)00065-2 (2002).
Böhning, D. Multinomial logistic regression algorithm. Ann. Inst. Stat. Math. 44, 197–200. https://doi.org/10.1007/BF00048682 (1992).
Tin Kam, H. Random decision forests. In Proc. 3rd International Conference on Document Analysis and Recognition, Vol. 271, 278–282. https://doi.org/10.1109/ICDAR.1995.598994 (1995).
Breiman, L. Arcing the Edge Technical Report 486 (Statistics Department, University of California, 1997).
Chandrashekar, G. & Sahin, F. A survey on feature selection methods. Comput. Electr. Eng. 40, 16–28. https://doi.org/10.1016/j.compeleceng.2013.11.024 (2014).
Khalid, S., Khalil, T. & Nasreen, S. A survey of feature selection and feature extraction techniques in machine learning. In 2014 Science and Information Conference, 372–378. https://doi.org/10.1109/SAI.2014.6918213 (2014).
Acknowledgements
The authors express their sincere gratitude to the anonymous reviewers for their constructive comments on this manuscript. The work presented in this paper was partly supported by the Ministry of Science and Technology, Taiwan [Grant Number 10X-62634-F-002-015]. The authors acknowledge the support.
Author information
Authors and Affiliations
Contributions
C.F.S. initiated the project. C.F.S., S.I.C., J.S.J., and F.P.L. contributed to the overall experimental design. C.F.S. and S.I.C. created the dataset. S.I.C. and J.S.J. contributed to software engineering. C.F.S. and S.I.C. analyzed the results and made the first article draft. F.P.L. supervised, reviewed, and revised the draft and final manuscripts. All authors contributed significantly to the revision of the first article draft and approval of the final version of the manuscript.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary Information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Su, CF., Chiu, SI., Jang, JS.R. et al. Improved inpatient deterioration detection in general wards by using time-series vital signs. Sci Rep 12, 11901 (2022). https://doi.org/10.1038/s41598-022-16195-2
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41598-022-16195-2
- Springer Nature Limited