About the Authors:
Wenshuo Liu
Roles Conceptualization, Data curation, Formal analysis, Methodology, Writing – original draft, Writing – review & editing
Affiliation: Michigan Integrated Center for Health Analytics and Medical Prediction, University of Michigan, Ann Arbor, MI, United States of America
Cooper Stansbury
Roles Data curation, Formal analysis, Writing – review & editing
Affiliations Department of Computational Biology and Bioinformatics, University of Michigan Medical School, Ann Arbor, MI, United States of America, Department of Systems, Populations, and Leadership, University of Michigan School of Nursing, Ann Arbor, MI, United States of America
Karandeep Singh
Roles Writing – review & editing
Affiliations Michigan Integrated Center for Health Analytics and Medical Prediction, University of Michigan, Ann Arbor, MI, United States of America, Division of Learning and Knowledge Systems, Department of Learning Health Sciences, University of Michigan Medical School, Ann Arbor, MI, United States of America, Division of Nephrology, Department of Internal Medicine, University of Michigan Medical School, Ann Arbor, MI, United States of America
Andrew M. Ryan
Roles Writing – review & editing
Affiliation: Department of Health Management and Policy, University of Michigan School of Public Health, Ann Arbor, MI, United States of America
Devraj Sukul
Roles Data curation, Writing – review & editing
Affiliation: Division of Cardiology, Department of Internal Medicine, University of Michigan Medical School, Ann Arbor, MI, United States of America
ORCID logo http://orcid.org/0000-0003-4709-3390
Elham Mahmoudi
Roles Writing – review & editing
Affiliations Michigan Integrated Center for Health Analytics and Medical Prediction, University of Michigan, Ann Arbor, MI, United States of America, Department of Family Medicine, University of Michigan Medical School, Ann Arbor, MI, United States of America
ORCID logo http://orcid.org/0000-0002-9746-8165
Akbar Waljee
Roles Writing – review & editing
Affiliations Michigan Integrated Center for Health Analytics and Medical Prediction, University of Michigan, Ann Arbor, MI, United States of America, Division of Gastroenterology and Hepatology, Department of Internal Medicine, University of Michigan Medical School, Ann Arbor, MI, United States of America, VA Center for Clinical Management Research, VA Ann Arbor Health Care System, Ann Arbor, MI, United States of America
ORCID logo http://orcid.org/0000-0003-1964-8790
Ji Zhu
Roles Conceptualization, Formal analysis, Methodology, Writing – original draft, Writing – review & editing
Affiliations Michigan Integrated Center for Health Analytics and Medical Prediction, University of Michigan, Ann Arbor, MI, United States of America, Department of Statistics, University of Michigan, Ann Arbor, MI, United States of America
Brahmajee K. Nallamothu
Roles Conceptualization, Data curation, Formal analysis, Project administration, Supervision, Visualization, Writing – original draft, Writing – review & editing
* E-mail: [email protected]
Affiliations Michigan Integrated Center for Health Analytics and Medical Prediction, University of Michigan, Ann Arbor, MI, United States of America, Division of Cardiology, Department of Internal Medicine, University of Michigan Medical School, Ann Arbor, MI, United States of America, VA Center for Clinical Management Research, VA Ann Arbor Health Care System, Ann Arbor, MI, United States of America
ORCID logo http://orcid.org/0000-0003-4331-6649
Introduction
Approximately 15% of patients discharged after an acute hospitalization are readmitted within 30 days, leading to potentially worse clinical outcomes and billions of dollars in healthcare costs [1]. Given these concerns, multiple quality efforts have been instituted in recent years to reduce readmissions in the United States. For example, the Medicare Hospital Readmission Reduction Program (HRRP) was created as part of the Patient Protection and Affordable Care Act and financially penalizes U.S. hospitals with excess 30-day readmission rates among Medicare beneficiaries [2,3]. Similar programs are being launched for patients with commercial insurance with the goal of further incentivizing hospitals to reduce readmissions [4,5].
Not surprisingly, the development of these programs has led to an increased demand for statistical models that accurately predict readmissions using available healthcare claims data. As the likelihood of readmission is related to key input features of patients (e.g., age and co-morbidities), differences in the distribution of patients across hospitals based on such features may lead to unfair penalization of hospitals that care for more at-risk individuals. Therefore, using statistical prediction models to adjust for patient risk across hospitals is a major priority for accountability programs [6]. However, the performance of prediction models for readmissions have been generally poor. For example, existing methods that rely on regression-based models report area under the curve (AUC) for the receiver operating characteristic in the range of 0.63 to 0.65, suggesting limited discrimination for prediction [7,8]. Recent use of more flexible prediction models that leverage machine learning algorithms, such as random forest and traditional artificial neural network (ANN) models, have attempted to address this limitation with minimal improvements [9–11].
The purpose of this study is to explore whether advances in ANN models and numerical embedding techniques could improve prediction of 30-day readmission using administrative claims data and how this potential improvement may impact calculation of risk-standardized hospital readmission rates. Administrative claims data such as diagnosis code are key to describe a patient’s condition and other characteristics, but are often not in the easiest or most straightforward format for statistical analysis. We exploit a word embedding technique classically used in Natural Language Processing (NLP) to convert each diagnosis code into a numerical vector such that the “distance” between diagnosis codes is related to “semantic” similarity. Further, using these numerical vectors as input, we employ a newly developed deep set architecture ANN model to accommodate varying numbers of diagnosis codes across different patients and the fact that the prediction should be invariant with respect to the ordering of the diagnosis codes. ANN models abstract input features from large-scale datasets to assign output probability by approximating a combination of non-linear functions over the input feature-space [12, 13]. Modern deployment of ANN models, including deep learning models, have been used successfully in a range of applications that include image classification and natural language processing [14–17], as well as prediction from electronic heath records [18,19]. We apply embedding algorithms from NLP and a new deep set architecture ANN model to a large United States administrative claims data source focusing on 3 common conditions that were targeted under the HRRP: acute myocardial infarction (AMI), heart failure (HF) and pneumonia (PNA).
Methods
We conducted this study following the Transparent Reporting of a Multivariable Prediction Model for Individual Prognosis or Diagnosis (TRIPOD) reporting guidelines (see S1 Checklist). All statistical code for replicating these analyses are available on the following GitHub repository: https://github.com/wenshuoliu/DLproj/tree/master/NRD. Data used for these analyses are publicly available at: https://www.hcup-us.ahrq.gov/tech_assist/centdist.jsp.
Study cohort
We used the 2014 Nationwide Readmissions Database (NRD) developed by the Agency for Healthcare Research and Quality (AHRQ) Healthcare Cost and Utilization Project (HCUP), which includes data on nearly 15 million admissions from 2,048 hospitals [20–22]. The NRD has the advantage of including all payers, including government and commercial insurers. We identified patients hospitalized for AMI, HF, and PNA. We created a separate cohort for each condition using strategies for identifying patients that were adopted from prior published work [8, 23]. The cohort of index admissions for each condition was based on principal International Classification of Diseases-9 (ICD-9) diagnosis codes at discharge (e.g. in the case of AMI we used 410.xx, except for 410.x2) while excluding the following cases: (1) records with zero length of stay for AMI patients (n = 4,926) per standards for constructing that cohort (as patients with AMI are unlikely to be discharged the same day); (2) patients who died in the hospital (n = 13,896 for AMI, n = 14,014 for HF, n = 18,648 for PNA); (3) patients who left the hospital against medical advice (n = 2,667 for AMI, 5,753 for HF, n = 5,057 for PNA); (4) patients with hospitalizations and no 30-day follow up (i.e. discharged in December, 2014 (n = 23,998 for AMI, n = 44,264 for HF, n = 47,523 for PNA)); (5) patients transferred to another acute care hospital (n = 8,400 for AMI, n = 5,393 for HF, n = 4,839 for PNA); (6) patients of age < 18 years old at the time of admission (n = 12 for AMI, n = 409 for HF, n = 28,159 for PNA); and (8) patients discharged from hospitals with less than 10 admissions (n = 1,956 for AMI, n = 1,221 for HF, n = 418 for PNA). Given that such facilities (<10 admissions) are not generally considered a part of typical quality assurance or performance measurement programs for readmissions, we were not interested in these facilities. In circumstances where the same patient was admitted several times during the study period, we selected only the first admission. Flow diagrams for the cohort selection are shown in S1 Fig.
Study variables
Our outcome was 30-day unplanned readmission created using the NRD Planned Readmission Algorithm [23]. The NRD also includes patient-level information on demographics and up to 30 ICD-9 diagnosis codes and 15 procedure codes from each hospitalization. Among the diagnosis codes, the principal diagnosis code at discharge represents the primary reason for the hospitalization while the rest represent comorbidities for the patient. To improve computational efficiency, we only included codes that appeared at least 10 times in the whole NRD database, reducing the number of ICD-9 diagnosis and ICD-9 procedure codes for inclusion in our analyses from 12,233 to 9,778 diagnosis codes and from 3,722 to 3,183 procedure codes, respectively.
Statistical models and analysis
We evaluated four statistical models: 1) a hierarchical logistic regression model; 2) gradient boosting (using the eXtreme Gradient Boosting [XGBoost] [24] approach, a widely-used, decision tree-based machine learning algorithm) using ICD-9 diagnosis and procedure codes represented as dummy variables (1 if present, 0 if absent); 3) an ANN model using a feed-forward neural network with ICD-9 codes represented as dummy variables; and 4) an ANN model in which ICD-9 codes were represented as latent variables learned through a word embedding algorithm. We used hierarchical logistic regression as a baseline comparator given its ubiquitous use in health services and outcomes research. XGBoost is based on gradient boosted decision trees and it is designed for speed and performance. We used it given its rising popularity in recent years as a flexible machine learning algorithm for structured data. The intuition behind our model comparisons was to explore the differences between sophisticated non-linear statistical models and traditional, “off-the-shelf” machine learning techniques. A more detailed explanation for the statistical models and ANN approaches as well as accompanying statistical code are available in the S1 Information.
To provide a reasonable baseline prediction against which to compare more sophisticated models we constructed a hierarchical logistic regression model trained on account age, gender and co-morbidity data. For co-morbidities, we used the well-established Elixauser Comorbitidy Index [25] to identify 29 variables to include as independent features in the model, with a hospital-specific intercept to account for patient clustering [7]. We implemented this model using the R function glmer from the package lme4.
For the second model we trained an XGBoost model on ICD-9 codes and age and gender information in order to provide a comparison to logistic regression. XGBoost has been well-recognized as an “off-the-shelf” ensemble algorithm that extends classical decision trees by iteratively fitting decision trees on the gradient of previous decision trees. XGBoost has been shown to be highly effiecient on large datasets and require little hyper-parameter tuning to achieve state-of-the-art performance in a variety of tasks [26]. We implemented this model using the Python package XGBoost with a learning rate of of 0.0002 to prevent potential overfitting.
For the third model we trained a shallow feed-forward ANN on the same set of features as the gradient boosted tree. Our motivation for the ANN architecture was to use a simple design with widely adopted parameters. We employed two fully-connected hidden layers with relu activation functions and a single fully-connected output layer (softmax). We chose the ADAM optimizer with a categorical cross-entropy loss function with a conservative learning rate of 0.0002. We reduced the dimensionality of the input feature space between the fully connected layers from 1,024 to 256 to learn complex patterns from the input features instead of using human-engineered selection of variables (i.e., the Elixhauser Comorbidity Index. ANN models require human parameter specification and may be prone to overfitting. For this reason we kept the architecture of the ANN relatively simple. As such, the ANN model represents a reasonable “off-the-shelf” analogy to the XGBoost model. To further mitigate chances of overfitting we included a dropout layer (0.3). Hyper-parameters were selected through cross-validation to give the best prediction accuracy on a hold-out validation set and evaluated on testing data.
In the fourth model, we encoded 9,778 ICD-9 diagnosis and 3,183 procedure codes into 200- and 50-dimensional latent variable space, using the Global Vector for Word Representations (GloVe) algorithm [27], i.e. each diagnosis code is represented by a 200-dimensional numerical vector and each procedure code is represented by a 50-dimensional numerical vector. We used GloVe, an unsupervised embedding algorithm to project ICD-9 co-occurrences to a numerical feature-space where semantic relations between codes are preserved. The purpose of exploring GloVe embeddings and their potential impact on predictive readmission models is to discover if radical changes from current practices in feature-space and model design impact risk-standarization scores. The prescence of two ICD-9 diagnosis or procedure codes in a patient record during hospitalization was considered as a co-occurrence. We then counted the number of co-occurrences for each pair of ICD-9 diagnosis and/or procedure codes in the NRD training data, (excluding the testing set) and constructed embedding vectors according to the GloVe algorithm, which uses the global co-occurrence of ICD-9 codes along with a local context. A two-dimensional t-SNE visualization of the embedding vectors of the ICD-9 diagnosis codes is shown in the S2 Fig. The visualization demonstrates that word embedding resulted in related diseases clustering closer to each other and is consistent with the application of word embedding algorithms in other administrative claims data [28, 29].
We used the deep set structure proposed by Zaheer et al [30] to incorporate ICD-9 diagnosis and procedure codes into the ANN model. This allowed us to account for varying counts of secondary ICD-9 diagnosis and procedure codes across patients and allow our model to be invariant to the ordering of these codes (e.g., the 2nd and the 10th code are interchangeable). The hospital ID was embedded into a 1-dimensional variable–conceptually this is similar to the hospital-level random intercept used in the hierarchical logistic regression models. The architectures of the two ANN models are shown in S3 Fig. The implementation of the ANN models was done using the Python packages Keras and Tensorflow.
To avoid the risk of overfitting, each of the study cohorts were divided into training, validation (for parameter tuning), and final testing sets at a proportion of 80%, 10%, and 10%, stratified by hospitals (i.e., within each hospital). We calculated AUC for the standard hierarchical logistic regression model, the XGBoost model and both ANN models on the final testing set, with the 95% confidence interval given from a 10-fold cross-validation. Once the models were developed, we then calculated risk-standardized hospital readmission rates for both the hierarchical logistic regression and the ANN model trained on diagnosis code embeddings. We calculated these using predictive margin, which is a generalization of risk adjustment that can be applied for both linear and non-linear models (like ANN models) [31, 32]. Specifically, the predictive margin for a hospital is defined as the average predicted readmission rate if everyone in the cohort had been admitted to that hospital. Benefits of predictive margins over conditional approaches have been discussed in Chang et al [33]. We compared this approach to the traditional approach for calculating risk-standardized hospital readmission rates in hierarchical logistic regression models that uses the predicted over expected readmission ratio for each hospital and then multiplying by the overall unadjusted readmission rate [7]; importantly, we found similar results (see S4 Fig).
Results
Study cohort
Our study cohort included 202,038 admissions for AMI, 303,233 admissions for HF, and 327,833 admissions for PNA, with unadjusted 30-day readmission rates of 12.0%, 17.7% and 14.3% respectively. The mean (standard deviation) age was 66.8 (13.7) for AMI, 72.5 (14.2) for HF and 69.2 (16.8) for PNA, with the proportion of females 37.6%, 48.9% and 51.8%, respectively. Summary baseline characteristics are shown in Table 1 with additional details of the ICD-9 diagnosis and procedure codes in S1 Table. In these cohorts, we noticed an extremely skewed prevalence of ICD-9 diagnosis and procedure codes that were used to identify features for training related to comorbidities. For example, in the AMI cohort, three quarters of the 5,614 distinct secondary ICD-9 diagnosis codes appear less than 49 times (prevalence 0.02%), while the most frequent ICD-9 diagnosis code (i.e., 41.401 for coronary atherosclerosis of native coronary artery) appears 152,602 times (prevalence 75.5%). See S1 Table for details.
[Figure omitted. See PDF.]
Table 1. Summary statistics of the predictors for each cohort assessed in this study population.
https://doi.org/10.1371/journal.pone.0221606.t001
Performance of prediction models
Results of prediction of 30-day readmission as assessed by AUC are reported in Table 2 for each model and each cohort. The gradient boosting model utilizing XGBoost performed slightly better than the hierarchical logistic regression model and similar to the basic feed-forward ANN model. In general, the medical code embedding deep set architecture model generated the best results on all cohorts relative to the other three models. Compared with hierarchical logistic regression (as a reasonable baseline), the medical code embedding deep set architecture model improved the AUC from 0.68 (95% CI 0.678, 0.683) to 0.72 (95% CI 0.718, 0.722) for the AMI cohort, from 0.60 (95% CI 0.592, 0.597) to 0.64 (95% CI 0.635, 0.639) for the HF cohort, from 0.63 (95% CI 0.624, 0.632) to 0.68 (95% CI 0.678, 0.683) for the PNA cohort. One possible explanation for this performance increase is that the embeddings capture the co-occurrence relationship between diagnosis codes, which is not enjoyed by the other three models, and the ANN is able to learn non-linear mapping patterns between the embeddings and the outcome. In a sensitivity analysis, we repeated the same analysis on elderly patients (65 years old and above) and these are provided in Table 2. Not unexpectedly, the overall AUCs decreased in the sensitivity analysis due to restriction of the cohort by age (which is a powerful predictor of readmission for patients); however, the margins for differences in AUCs across the four different statistical models increased slightly with this restriction by age.
[Figure omitted. See PDF.]
Table 2. Summary statistics of ICD-9CM diagnosis and procedure codes for each cohort.
https://doi.org/10.1371/journal.pone.0221606.t002
Risk-standardized hospital readmission rates
Given its overall higher performance, we compared risk-standardized hospital readmission rates calculated from the medical code embedding deep set architecture model with those calculated using the hierarchical logistic regression model. The histograms and summaries of these results are shown in (Fig 1). Distributions of the risk-standardized hospital readmission rates from the two models were similar with just a modest shift downward in the mean for the medical code embedding deep set architecture model. We observed substantial differences in terms of rankings of individual hospitals between the two models. For both models, we divided the hospitals into three groups based on quintiles of predicted risk-standardized hospital readmission rates: top 20%, middle 60% and bottom 20%. For AMI, the medical code embedding deep set architecture model classified 72 (6.4%) hospitals in the middle 60% that the hierarchical model classified in the top 20% and classified 37 (3.3%) hospitals in the middle 60% that the hierarchical model classified in the bottom 20%. Results were similar for the HF and PNA cohorts (Table 3). Given the differences in risk-standardization it is worth investigating whether traditional approaches (like logistic regression) are the best models on which to base risk-standardization scores.
[Figure omitted. See PDF.]
Fig 1. Distribution of risk-standardized hospital readmission rates.
This figure shows differences in the distribution of risk-standardized hospital readmission rates for acute myocardial infarction (AMI), congestive health failure (HF), and pneumonia (PNA) generated by the hierarchical logistic regression (HLR) model and the medical code embedding Deep Set architecture ANN (ME-DS) model. Standardized readmission rates are generated by comparing model predictions to expected readmission rates for each hospital. This figure illustrates that despite having comparable predictive accuracy LHR and MS-DS lead to differences in hospital risk stratification.
https://doi.org/10.1371/journal.pone.0221606.g001
[Figure omitted. See PDF.]
Table 3. Cross tabulation of divided groups between the Hierarchical Logistic Regression (HLR) and the medical code embedding deep set architecture (ME-DS) model for each cohort.
https://doi.org/10.1371/journal.pone.0221606.t003
Discussion
In recent years, ANN models have shown advantages over traditional statistical models in a variety of medical tasks [18, 19]. Whether the application of such models to administrative claims data brings similar improvement in specific tasks related to prediction is worth exploring. This is especially important given the ubiquitous nature of claims data for assessing quality and hospital performance. In this paper, we applied ANN models towards the task of predicting 30-day readmission after AMI, HF, and PNA hospitalizations with and without diagnosis code embeddings. We compared “more sophisticated” statistical models to existing approaches that use input features from classification systems that rely on expert knowledge like hierarchical logistic regression models as well as gradient boosting. Our findings suggest ANN models trained on medical code embeddings provide more accurate predictions of readmission and generate risk-standardized hospital readmission rates that vary from commonly used hierarchical logistic regression models.
There has been substantial work performed on constructing risk prediction models to predict readmissions after a hospitalization. The most frequent way these models are employed is through regression-based models that include age, gender and co-morbidities as input features [7]. For co-morbidities, ICD-9 diagnosis and procedure codes obtained from administrative claims data are used as input features to adjust for differences in individual patient risk in these models; however, not all of the thousands of potential ICD-9 diagnosis and procedure codes are included in the models and selecting which to incorporate is an important step. The selection has been based largely on expert input and empirical studies that have been used to generate fixed classification systems like the Hierarchical Condition Categories [34] or Elixhauser Comorbidity Index [25]. Our findings suggest that more attention should be paid to risk-stratification methods based on non-linear classification systems, as they lead to substantial differences in risk-scores.
An advantage of ANN models is their ability as a statistical model to capture potential non-linear effects and interactions of an abstract feature space. By first representing the cooccurance patterns in diagnosis codes using GloVe, followed by a deep set ANN model, one may not need to rely on human-generated classification systems, instead learning to automate extraction of relevant features from the data. Yet few studies to date have explored this type of model towards administrative claims data. We believe a primary reason for this is that ANN models can be difficult to train due to the issues related to parameter optimization and memory consumption in the setting of a large number of parameters–sometimes in the order of millions. In the few studies that have used ANN models with administrative claims data [9, 35, 36], their use also may not have fully captured their full potential for risk prediction. For example, the use of binary “1/0” input features for ICD-9 diagnosis and procedure codes may ignore hidden relationships across comorbidities, limiting the ability of ANN models to improve on traditional hierarchical logistic regression or other methods like gradient boosting.
Of course, there has been some work on predicting readmissions using ANN models in the published literature. Futoma et al. implemented the basic architecture of feed-forward ANN models and showed modest advantages over conventional methods [9]. A number of researchers proposed to embed medical concepts (including but not limited to ICD-9 diagnosis and procedural codes) into a latent variable space to capture their co-relationships [28, 29, 37]; however, these investigators used this approach largely for cohort creation rather than predicting clinical outcomes or risk-adjustment. Krompass et al [36] used Hellinger-distance based principal components analysis [38] to embed ICD-10 codes and then built a logistic regression model using the embedded codes as input features. They found marginal improvements in prediction of readmissions over a feed-forward neural network but were restricted by their limited sample size. Choi et al. [35] designed a graph-based attention model to supplement embedding with medical ontologies for various prediction tasks, including readmission. However, their model did not explicitly consider the fact that the medical codes are permutation invariant. In this paper, we took advantage of a novel word embedding approach, Global Vector for Word Representations (GloVe) [27], as well as a new and recently proposed deep set architecture [30] to fully capture the inter-relationship and the permutation-invariant nature of the diagnosis and procedure codes at a local and global level. These choices–which were purposeful and driven by our intuition on the benefits of ANN models for this specific task–resulted in improved accuracy of prediction for readmission for a word embedding deep set architecture model across all 3 conditions under consideration.
Our study should be interpreted in context of the following limitations. First, although we found ANN models outperformed hierarchical logistic regression models, it is uncertain whether these improvements will justify their use more broadly as this requires consideration of other issues. For example, ANN models require large-scale data sources to train. Even though such data were available given the NRD for our current work, these are not always available. But the widespread availability and application of administrative claims data in assessing quality and hospital performance justifies the need to explore ANN models (and other approaches) and alternative feature representation further. Second, ANN models are computationally intensive and retain a “blackbox” feel with its findings difficult to understand and explain to users (similar to other models like gradient boosting). These issues may make it less attractive to policymakers and administrators when there may be a need to justify why performance is lacking in a public program (e.g., HRRP). Third, ANN models may not work for applications beyond 30-day readmission in these 3 common conditions. Work is needed to compare the performance of ANN models with traditional approaches for other outcomes (e.g., mortality), rare diseases, or populations (i.e., non-hospitalized patients).
In summary, ANN models with medical code embeddings have higher predictive accuracy for 30-day readmission when compared with hierarchical logistic regression models and gradient boosting. Furthermore, ANN models generate risk-standardized hospital readmission rates that lead to differing assessments of hospital performance when compared to these other approaches. The role of ANN models in clinical and health services research warrants further investigation.
Supporting information
[Figure omitted. See PDF.]
S1 Checklist. TRIPOD checklist: Prediction model development.
https://doi.org/10.1371/journal.pone.0221606.s001
(DOCX)
S1 Fig. Flow diagrams of the cohort creation for Acute Myocardial Infarction (AMI), congestive Health Failure (HF), and pneumonia (PNA).
Cohort selection for AMI, HF, and PNA are shown in (a), (b) and (c) respectively.
https://doi.org/10.1371/journal.pone.0221606.s002
(TIFF)
S2 Fig. Visualization of embedding vectors of the principal diagnosis codes in two dimensions.
This visualization was done using t-SNE. Each point represents a diagnosis code (disease). The size of the points represents the prevalence of that code. (a) The points are coloured by the Clinical Classifications Software (CCS)10 level 1 categories of the multi-level classification system. The frequent codes with the same CCS level 1 categories form clusters, while the infrequent codes form a “cloud” without a clear pattern. (b) As examples, two CCS level 1 categories, “7 Diseases of the circulatory system” and “8 Diseases of the respiratory system” are highlighted in the visualization, with all other diseased represented in grey. (c) The principal diagnosis codes as the inclusion criterion of the three cohorts, acute myocardial infarction, congestive health failure and pneumonia are highlighted.
https://doi.org/10.1371/journal.pone.0221606.s003
(TIFF)
S3 Fig. Architectures of Artificial Neural Network (ANN) models.
(a) Feed-forward neural network ANN model. (b) Medical code embedding deep set architecture model. This model looks up the medical code embedding of each ICD-9 codes that are pretrained by the GloVe model, and aggregates variable number of secondary diagnosis and procedure codes into a final representation vector using the deep set architecture.
https://doi.org/10.1371/journal.pone.0221606.s004
(TIFF)
S4 Fig. Comparison of risk-adjustment methods.
The plots shows the risk-standardized hospital readmission rates and the hospital rankings calculated by the two risk-adjustment methods, predicted over expected readmission rate ratio (e.g., the method employed by the Centers for Medicare & Medicaid Services [CMS]) and predictive margins, for the hierarchical logistic regression (HLR) model on the acute myocardial infarction cohort.
https://doi.org/10.1371/journal.pone.0221606.s005
(TIFF)
S1 Table. Summary statistics of ICD-9CM diagnosis and procedure codes for each cohort.
https://doi.org/10.1371/journal.pone.0221606.s006
(DOCX)
S1 Information.
https://doi.org/10.1371/journal.pone.0221606.s007
(DOCX)
Citation: Liu W, Stansbury C, Singh K, Ryan AM, Sukul D, Mahmoudi E, et al. (2020) Predicting 30-day hospital readmissions using artificial neural networks with medical code embedding. PLoS ONE 15(4): e0221606. https://doi.org/10.1371/journal.pone.0221606
1. Jencks SF, Williams MV, Coleman EA. Rehospitalizations among Patients in the Medicare Fee-for-Service Program. N Engl J Med. 2009;360(14):1418–1428. pmid:19339721
2. Rosenbaum S. The Patient Protection and Affordable Care Act: Implications for Public Health Policy and Practice. Public Health Rep. 2011;126(1):130–135. pmid:21337939
3. Patient Protection and Affordable Care Act—HealthCare.gov Glossary. HealthCare.gov. https://www.healthcare.gov/glossary/patient-protection-and-affordable-care-act/. Accessed January 7, 2019.
4. Chen C, Scheffler G, Chandra A. Readmission penalties and health insurance expansions: a dispatch from Massachusetts. J Hosp Med. 2014;9(11):681–687. pmid:24945696
5. Chakraborty H, Axon RN, Brittingham J, Lyons GR, Cole L, Turley CB. Differences in Hospital Readmission Risk across All Payer Groups in South Carolina. Health Serv Res. 2017;52(3):1040–1060. pmid:27678196
6. Kansagara D, Englander H, Salanitro A, et al. Risk prediction models for hospital readmission: A systematic review. JAMA. 2011;306(15):1688–1698. pmid:22009101
7. Krumholz HM, Lin Z, Drye EE, et al. An Administrative Claims Measure Suitable for Profiling Hospital Performance Based on 30-Day All-Cause Readmission Rates Among Patients With Acute Myocardial Infarction. Circulation Cardiovascular Quality and Outcomes. 2011;4(2):243–252. pmid:21406673
8. Sukul D, Sinha SS, Ryan AM, Sjoding MW, Hummel SL, Nallamothu BK. Patterns of Readmissions for Three Common Conditions Among Younger US Adults. Am J Med. 2017;130(10):1220.e1–1220.e16. pmid:28606799
9. Futoma J, Morris J, Lucas J. A comparison of models for predicting early hospital readmissions. Journal of Biomedical Informatics. 2015;56:229–238. pmid:26044081
10. Mortazavi Bobak J., Downing Nicholas S., Bucholz Emily M., et al. Analysis of Machine Learning Techniques for Heart Failure Readmissions. Circulation: Cardiovascular Quality and Outcomes. 2016;9(6):629–640. pmid:28263938
11. Frizzell JD, Liang L, Schulte PJ, et al. Prediction of 30-Day All-Cause Readmissions in Patients Hospitalized for Heart Failure: Comparison of Machine Learning and Other Statistical Approaches. JAMA Cardiol. 2017;2(2):204–209. pmid:27784047
12. LeCun Y, Bengio Y, Hinton G. Deep learning. Nature. 2015;521:436. pmid:26017442
13. Goodfellow I, Bengio Y, Courville A, Bengio Y. Deep Learning. Vol 1. MIT press Cambridge; 2016.
14. Bengio Y. Learning Deep Architectures for AI. MAL. 2009;2(1):1–127.
15. Deng L, Yu D. Deep Learning: Methods and Applications. SIG. 2014;7(3–4):197–387.
16. Voulodimos A, Doulamis N, Doulamis A, Protopapadakis E. Deep Learning for Computer Vision: A Brief Review. Comput Intell Neurosci. 2018;2018:7068349. pmid:29487619
17. Young T, Hazarika D, Poria S, Cambria E. Recent Trends in Deep Learning Based Natural Language Processing [Review Article]. IEEE Computational Intelligence Magazine. 2018;13(3):55–75.
18. Shickel B, Tighe P, Bihorac A, Rashidi P. Deep EHR: A Survey of Recent Advances on Deep Learning Techniques for Electronic Health Record (EHR) Analysis. Vol PP.; 2017. pmid:29989977
19. Rajkomar A, Oren E, Chen K, et al. Scalable and accurate deep learning with electronic health records. npj Digital Medicine. 2018;1(1):18.
20. Barrett M, Wier L, Jiang H, Steiner C. All-cause readmissions by payer and age, 2009–2013: statistical brief# 199. 2006.
21. Healthcare Cost and Utilization Project (HCUP. NRD overview. 2015.
22. NRD Database Documentation. https://www.hcup-us.ahrq.gov/db/nation/nrd/nrddbdocumentation.jsp. Accessed January 7, 2019.
23. 2016 Condition-Specific Measures Updates and Specifications Report Hospital-Level 30-Day Risk-Standardized Readmission Measures: Acute Myocardial Infarction–Version 9.0, Chronic Obstructive Pulmonary Disease–Version 5.0, Heart Failure–Version 9.0, Pneumonia–Version 9.0, Stroke–Version 5.0. 2016:112.
24. Chen T, Guestrin C. XGBoost: A Scalable Tree Boosting System. In: Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining—KDD ‘16. San Francisco, California, USA: ACM Press; 2016:785–794. https://doi.org/10.1145/2939672.2939785
25. Elixhauser A, Steiner C, Harris DR, Coffey RM. Comorbidity measures for use with administrative data. Medical care. 1998:8–27. pmid:9431328
26. Scalable, Portable and Distributed Gradient Boosting (GBDT, GBRT or GBM) Library, for Python, R, Java, Scala, C++ and More. Runs on Single Machine, Hadoop, Spark, Flink and DataFlow: Dmlc/Xgboost. Distributed (Deep) Machine Learning Community; 2019. https://github.com/dmlc/xgboost. Accessed May 22, 2019.
27. Pennington J, Socher R, Manning CD. GloVe: Global Vectors for Word Representation. In: Empirical Methods in Natural Language Processing (EMNLP).; 2014:1532–1543. http://www.aclweb.org/anthology/D14-1162.
28. Choi Y, Chiu CY-I, Sontag D. Learning Low-Dimensional Representations of Medical Concepts. AMIA Jt Summits Transl Sci Proc. 2016;2016:41–50. pmid:27570647
29. Finlayson SG, LePendu P, Shah NH. Building the graph of medicine from millions of clinical narratives. Scientific Data. 2014;1:140032. pmid:25977789
30. Zaheer M, Kottur S, Ravanbakhsh S, Poczos B, Salakhutdinov RR, Smola AJ. Deep Sets. In: Guyon I, Luxburg UV, Bengio S, et al., eds. Advances in Neural Information Processing Systems 30. Curran Associates, Inc.; 2017:3391–3401. http://papers.nips.cc/paper/6931-deep-sets.pdf.
31. Lee J. Covariance adjustment of rates based on the multiple logistic regression model. Journal of Clinical Epidemiology. 1981;34(8):415–426.
32. Lane PW, Nelder JA. Analysis of covariance and standardization as instances of prediction. Biometrics. 1982:613–621. pmid:7171691
33. Chang I-M, Gelman R, Pagano M. Corrected group prognostic curves and summary statistics. Journal of chronic diseases. 1982;35(8):669–674. pmid:7096530
34. Pope GC, Kautter J, Ellis RP, et al. Risk adjustment of Medicare capitation payments using the CMS-HCC model. Health Care Financ Rev. 2004;25(4):119–141. pmid:15493448
35. Choi E, Bahadori MT, Song L, Stewart WF, Sun J. GRAM: graph-based attention model for healthcare representation learning. In: ACM; 2017:787–795.
36. Krompass D, Esteban C, Tresp V, Sedlmayr M, Ganslandt T. Exploiting Latent Embeddings of Nominal Clinical Data for Predicting Hospital Readmission. Vol 29.; 2014.
37. Beam AL, Kompa B, Fried I, et al. Clinical Concept Embeddings Learned from Massive Sources of Multimodal Medical Data. April 2018. https://arxiv.org/abs/1804.01486v2. Accessed February 22, 2019.
38. Lebret R, Collobert R. Word emdeddings through hellinger PCA. arXiv preprint arXiv:13125542. 2013.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
This is an open access article, free of all copyright, and may be freely reproduced, distributed, transmitted, modified, built upon, or otherwise used by anyone for any lawful purpose. The work is made available under the Creative Commons CC0 public domain dedication: https://creativecommons.org/publicdomain/zero/1.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
Reducing unplanned readmissions is a major focus of current hospital quality efforts. In order to avoid unfair penalization, administrators and policymakers use prediction models to adjust for the performance of hospitals from healthcare claims data. Regression-based models are a commonly utilized method for such risk-standardization across hospitals; however, these models often suffer in accuracy. In this study we, compare four prediction models for unplanned patient readmission for patients hospitalized with acute myocardial infarction (AMI), congestive health failure (HF), and pneumonia (PNA) within the Nationwide Readmissions Database in 2014. We evaluated hierarchical logistic regression and compared its performance with gradient boosting and two models that utilize artificial neural networks. We show that unsupervised Global Vector for Word Representations embedding representations of administrative claims data combined with artificial neural network classification models improves prediction of 30-day readmission. Our best models increased the AUC for prediction of 30-day readmissions from 0.68 to 0.72 for AMI, 0.60 to 0.64 for HF, and 0.63 to 0.68 for PNA compared to hierarchical logistic regression. Furthermore, risk-standardized hospital readmission rates calculated from our artificial neural network model that employed embeddings led to reclassification of approximately 10% of hospitals across categories of hospital performance. This finding suggests that prediction models that incorporate new methods classify hospitals differently than traditional regression-based approaches and that their role in assessing hospital performance warrants further investigation.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer