ArticlePDF Available

Enhancement of Claims Data to Improve Risk Adjustment of Hospital Mortality

Authors:

Abstract

Comparisons of risk-adjusted hospital performance often are important components of public reports, pay-for-performance programs, and quality improvement initiatives. Risk-adjustment equations used in these analyses must contain sufficient clinical detail to ensure accurate measurements of hospital quality. To assess the effect on risk-adjusted hospital mortality rates of adding present on admission codes and numerical laboratory data to administrative claims data. Comparison of risk-adjustment equations for inpatient mortality from July 2000 through June 2003 derived by sequentially adding increasingly difficult-to-obtain clinical data to an administrative database of 188 Pennsylvania hospitals. Patients were hospitalized for acute myocardial infarction, congestive heart failure, cerebrovascular accident, gastrointestinal tract hemorrhage, or pneumonia or underwent an abdominal aortic aneurysm repair, coronary artery bypass graft surgery, or craniotomy. C statistics as a measure of the discriminatory power of alternative risk-adjustment models (administrative, present on admission, laboratory, and clinical for each of the 5 conditions and 3 procedures). The mean (SD) c statistic for the administrative model was 0.79 (0.02). Adding present on admission codes and numerical laboratory data collected at the time of admission resulted in substantially improved risk-adjustment equations (mean [SD] c statistic of 0.84 [0.01] and 0.86 [0.01], respectively). Modest additional improvements were obtained by adding more complex and expensive to collect clinical data such as vital signs, blood culture results, key clinical findings, and composite scores abstracted from patients' medical records (mean [SD] c statistic of 0.88 [0.01]). This study supports the value of adding present on admission codes and numerical laboratory values to administrative databases. Secondary abstraction of difficult-to-obtain key clinical findings adds little to the predictive power of risk-adjustment equations.
ORIGINAL CONTRIBUTION
Enhancement of Claims Data to Improve
Risk Adjustment of Hospital Mortality
Michael Pine, MD, MBA
Harmon S. Jordan, ScD
Anne Elixhauser, PhD
Donald E. Fry, MD
David C. Hoaglin, PhD
Barbara Jones, MA
Roger Meimban, PhD
David Warner, MS
Junius Gonzales, MD, MBA
RISK-ADJUSTED HOSPITAL MOR-
tality rates for specified con-
ditions and procedures fre-
quently are used in public
reports and pay-for-performance pro-
grams as indicators of the quality of hos-
pital care.1-3 Risk adjustment often is
based solely on administrative claims
data from uniform bills that hospitals
submit to payers. These data lack clini-
cally important pathophysiological in-
formation and do not distinguish be-
tween conditions that were present on
admission (POA; ie, potential risk fac-
tors) and complications that occurred
during hospitalization. The validity of
risk-adjustment systems that use only
administrative data has been chal-
lenged repeatedly,4-9 and there is gen-
eral agreement10-23 that additional data
are required to predict accurately an in-
dividual patient’s risk of dying.
Physicians are particularly con-
cerned that inadequate risk adjust-
ment penalizes the practitioners and fa-
cilities that care for the sickest patients
and may result in the denial of needed
care to high-risk patients.24-26 Con-
sumer advocates and payers, through
initiatives such as the Consumer-
Purchaser Disclosure Project,27 are at-
tempting to expand administrative data
sets to include clinical information to
ensure that incentives reward high-
quality clinical care. On the other hand,
many hospital administrators have com-
plained that the cost of retrieving
supplementary clinical data from medi-
cal records is prohibitive, and some re-
searchers have argued that risk adjust-
ment using only administrative data can
be made sufficiently accurate to sup-
port valid comparisons among hospi-
tals.28,29
The addition of a POA modifier for
secondary diagnosis codes first was pro-
posed in 199130 and was successfully
adopted in New York State in 1994 and
in California in 1996.31 The planned
implementation in March 2007 of new
standards for hospital claims data in-
cludes nationwide adoption of this
Author Affiliations: Michael Pine and Associates Inc,
Chicago, Ill (Drs Pine, Fry, and Meimban, and Ms
Jones); Department of Medicine, Pritzker School of
Medicine, University of Chicago, Chicago, Ill (Dr Pine);
Abt Associates Inc, Cambridge, Mass (Drs Jordan, Hoag-
lin, Gonzales, and Mr Warner); School of Medicine,
Tufts University, Boston, Mass (Dr Jordan); and Agency
for Healthcare Research and Quality, Rockville, Md
(Dr Elixhauser).
Corresponding Author: Michael Pine, MD, MBA, 1210
Chicago Ave, Suite 503, Evanston, IL 60202 (mpine
@aol.com).
Context Comparisons of risk-adjusted hospital performance often are important com-
ponents of public reports, pay-for-performance programs, and quality improvement
initiatives. Risk-adjustment equations used in these analyses must contain sufficient
clinical detail to ensure accurate measurements of hospital quality.
Objective To assess the effect on risk-adjusted hospital mortality rates of adding present
on admission codes and numerical laboratory data to administrative claims data.
Design, Setting, and Patients Comparison of risk-adjustment equations for in-
patient mortality from July 2000 through June 2003 derived by sequentially adding
increasingly difficult-to-obtain clinical data to an administrative database of 188 Penn-
sylvania hospitals. Patients were hospitalized for acute myocardial infarction, conges-
tive heart failure, cerebrovascular accident, gastrointestinal tract hemorrhage, or pneu-
monia or underwent an abdominal aortic aneurysm repair, coronary artery bypass graft
surgery, or craniotomy.
Main Outcome Measures Cstatistics as a measure of the discriminatory power
of alternative risk-adjustment models (administrative, present on admission, labora-
tory, and clinical for each of the 5 conditions and 3 procedures).
Results The mean (SD) cstatistic for the administrative model was 0.79 (0.02). Add-
ing present on admission codes and numerical laboratory data collected at the time of
admission resulted in substantially improved risk-adjustment equations (mean [SD] c
statistic of 0.84 [0.01] and 0.86 [0.01], respectively). Modest additional improve-
ments were obtained by adding more complex and expensive to collect clinical data
such as vital signs, blood culture results, key clinical findings, and composite scores
abstracted from patients’ medical records (mean [SD] cstatistic of 0.88 [0.01]).
Conclusions This study supports the value of adding present on admission codes
and numerical laboratory values to administrative databases. Secondary abstraction
of difficult-to-obtain key clinical findings adds little to the predictive power of risk-
adjustment equations.
JAMA. 2007;297:71-76 www.jama.com
©2007 American Medical Association. All rights reserved. (Reprinted) JAMA, January 3, 2007—Vol 297, No. 1 71
at Attn Clinical Affairs, on January 3, 2007 www.jama.comDownloaded from
modifier, which distinguishes condi-
tions that develop during hospital stays
(potential complications of care) from
conditions that were present at admis-
sion (potential treatment-indepen-
dent risk factors). Inclusion of POA
codes in administrative data sets should
permit analysts to incorporate impor-
tant predictors of inpatient mortality
into administrative risk-adjustment
equations without improperly desig-
nating patients as having high intrin-
sic risks at admission when their in-
creased vulnerability resulted from
hospital-acquired complications.31
The adoption of Logical Observa-
tion Identifiers Names and Codes32 for
laboratory data and advances in elec-
tronic health data technology have low-
ered the cost of retrieving numerical
laboratory data at many hospitals.33 Be-
cause of substantial differences in the
cost of obtaining various types of clini-
cal data, limited enhancement of ad-
ministrative data sets appears to be both
practical and desirable. Ideally, clini-
cal data elements selected for this pur-
pose will be relatively inexpensive to
obtain and will be useful predictors of
mortality for multiple conditions and
procedures.
This study was designed to test the
hypothesis that the combination of POA
modifiers for secondary diagnoses and
a limited set of numerical laboratory
data would improve risk adjustment of
inpatient mortality for a diverse set of
clinical conditions and procedures. We
also hypothesized that further addi-
tions of highly specific, difficult-to-
obtain clinical data sometimes consid-
ered important predictors of inpatient
mortality by clinicians would add little
to the accuracy of predictive models.
METHODS
Risk-adjustment models were created
and analyzed using data from July 2000
through June 2003 from 188 Pennsyl-
vania hospitals supplied by the Penn-
sylvania Health Care Cost Contain-
ment Council.34 Case-level claims data
were supplemented with clinical data
abstracted from medical records by spe-
cially trained personnel using Medi-
Qual’s proprietary Atlas clinical infor-
mation system.35 This system defines a
broad array of clinical data elements,
including historical information, labo-
ratory results, vital signs, clinical symp-
toms and signs, pathophysiological
abnormalities, and composite patho-
physiological scores, which are col-
lected and stored along with the hos-
pital day on which each clinical finding
was observed.
Risk-adjusted mortality rates were
analyzed for 5 health conditions (acute
myocardial infarction, congestive heart
failure, acute cerebrovascular acci-
dent, gastrointestinal tract hemor-
rhage, or pneumonia) and 3 surgical
procedures (abdominal aortic aneu-
rysm repair, coronary artery bypass graft
surgery, or craniotomy). The Agency for
Healthcare Research and Quality’s In-
patient Quality Indicator software ver-
sion 2.1 was used to identify cases that
met criteria for inclusion in each
group.36
Four models were constructed for
each condition and procedure (1 set of
data for each of the 5 conditions and 1
set of data for each of the 3 proce-
dures). The first model termed admin-
istrative used standard claims data. The
second model termed POA used data ab-
stracted from medical records to deter-
mine whether coded secondary diag-
noses had been present at admission.
The third model termed laboratory used
POA codes and numerical laboratory
data (often available in electronic form;
eg, creatinine, hematocrit level) docu-
mented on the first day of hospitaliza-
tion prior to a procedure requiring gen-
eral or regional anesthesia. The fourth
model termed clinical used the criteria
in the third model plus vital signs, other
laboratory data not included in the third
model (eg, bacterial culture results), At-
las key clinical findings abstracted from
medical records (eg, immunocompro-
mised, lethargy), and composite clini-
cal scores (ie, American Society of An-
esthesiologists classification, Glasgow
Coma Score) documented on the first
day of hospitalization prior to a proce-
dure requiring general or regional an-
esthesia.
The administrative model was based
solely on data from hospital bills (ie,
age, sex, and principal diagnoses, sec-
ondary diagnoses, and procedures
coded according to the International
Classification of Diseases, Ninth Revi-
sion, Clinical Modification [ICD-9-
CM]). To avoid using hospital-
acquired complications as risk factors,
hospital bills from New York and Cali-
fornia (secondary diagnoses were modi-
fied by POA codes in these states) were
used to help identify which secondary
diagnoses were generally present at ad-
mission. Secondary diagnoses were eli-
gible for inclusion as risk factors in the
administrative model only when they
were coded as hospital-acquired com-
plications in fewer than 20% of cases
in which they occurred.
The POA model included addi-
tional secondary diagnoses excluded
from the administrative model be-
cause of their association with unac-
ceptably high rates of complications. Be-
cause Pennsylvania claims data do not
include POA codes, clinical data in the
Atlas database were used to determine
whether coded secondary diagnoses
were present at admission. In the cre-
ation of surrogate POA codes, the At-
las database served as a substitute for
the complete medical record available
to coders in New York and California.
For example, posthemorrhagic ane-
mia was excluded from the adminis-
trative model for congestive heart fail-
ure because hospitals in New York and
California coded it as acquired during
hospitalization in more than 30% of the
cases in which it occurred. However,
posthemorrhagic anemia was eligible
for inclusion as a risk factor in the POA
model for congestive heart failure when
the Atlas database documented that
anemia was present on the day of ad-
mission.
For each condition or procedure,
candidate risk factors were con-
structed from principal diagnosis codes,
up to 8 secondary diagnosis codes, up
to 6 procedure codes, and clinical data
elements associated with higher than
average mortality rates. Infrequently oc-
curring codes were combined with
CLAIMS DATA FOR IMPROVING RISK-ADJUSTED HOSPITAL MORTALITY
72 JAMA, January 3, 2007—Vol 297, No. 1 (Reprinted) ©2007 American Medical Association. All rights reserved.
at Attn Clinical Affairs, on January 3, 2007 www.jama.comDownloaded from
codes for clinically similar conditions
or procedures that had similar mortal-
ity rates. Continuous measures (eg, age,
creatinine level) were transformed into
1 or more categorical variables based
on clinical judgment and empirical
evaluation of associated mortality rates.
For each condition or procedure,
stratified random samples of live dis-
charges and fatalities were combined to
create 3 mutually exclusive data sets:
a training set (50%), a validation set
(25%), and a test set (25%). Partition-
ing the data in this way facilitated the
construction of more robust models.
A preliminary predictive equation
was developed on the training set for
each condition or procedure using only
age categories and individual hospital
identifiers. (Including hospitals as risk
factors during model development is a
standard technique for reducing pos-
sible bias caused by the associations be-
tween the prevalence of potential risk
factors at individual hospitals and the
quality of care provided by those hos-
pitals.) For the administrative, POA,
and laboratory models, additional po-
tential risk factors were added in a se-
quence determined by forward step-
wise logistic regression.37 To avoid
overfitting, variables added after the
minimum value of the Schwarz crite-
rion38 was attained were removed from
models. (This criterion weighs the
trade-off between the fit of a model and
its complexity.) The remaining predic-
tive variables and their coefficients were
evaluated for clinical plausibility. On
rare occasions, clinically problematic
variables were eliminated or modi-
fied. To avoid substituting more ex-
pensive clinical variables for less costly
ones with almost equivalent predic-
tive power, predictive variables se-
lected for the laboratory model were re-
tained and additional clinical risk
factors were added in sequence as de-
scribed above.
For each of the 4 models (adminis-
trative, POA, laboratory, and clinical)
for each condition or procedure, a
nested sequence of models was cre-
ated first with 1 variable selected us-
ing the training data set, then with 2
variables, and lastly with all the vari-
ables. Variables were added to succes-
sive models in the order in which they
were entered in the minimum Schwarz
criterion model. From each nested se-
quence of models, the validation set was
used to select the model with the small-
est average prediction error39 as the fi-
nal validated model. Finally, the coef-
ficients of the variables in the validated
models were retained, the hospital vari-
ables were removed, and the inter-
cepts were recalculated to equate ob-
served and predicted mortality rates.
Case-level discriminatory power (ie,
the ability of a model to distinguish
cases that died from those that sur-
vived) was computed on the test set us-
ing cstatistics.40 All data management
and statistical analyses were per-
formed using SAS software versions 8
and 9.1 (SAS Institute Inc, Cary, NC).
The study design was approved by the
Abt Associates’ institutional review
board.
RESULTS
The numbers of cases ranged from 5309
(abdominal aortic aneurysm repair) to
200 506 (congestive heart failure)
(TABLE). Mortality rates ranged from
3.2% for coronary artery bypass graft
surgery to 10.8% for acute cerebrovas-
cular accident.
Designating secondary diagnoses as
present at admission increased the av-
erage number of secondary diagnosis
variables included from 8.6 in the ad-
ministrative model to 15.4 in the POA
model. This increase occurred be-
cause secondary diagnoses such as acute
renal failure in patients admitted to the
hospital with pneumonia, who also had
elevated creatinine levels on the day of
admission, were eligible for inclusion
as risk factors in the POA model. Com-
parison of the POA model and the labo-
ratory model revealed that the addi-
tion of an average of 11.1 numerical
laboratory values present on the first
hospital day was accompanied by an av-
erage reduction of 4.5 secondary diag-
nosis variables. This reduction re-
flected the substitution of more specific
laboratory values for less specific sec-
ondary diagnosis variables (eg, pH
7.25 or pH 7.25 but 7.35 re-
placed the ICD-9-CM secondary diag-
nosis code for acidosis in acute myo-
cardial infarction). Compared with the
laboratory model, an average of 9 ad-
ditional clinical findings present on the
first hospital day were incorporated into
the clinical model.
The final models included a total of
20 numerical laboratory determina-
tions, 3 other laboratory determina-
tions (eg, blood cultures), 5 vital signs,
22 key clinical findings, and 2 compos-
ite scores. Many individual numerical
laboratory results and vital signs ap-
peared in the clinical models for 4 or
more conditions or procedures (eg, pH
and prothrombin time were risk fac-
tors in the clinical models for all 5 con-
ditions and 3 procedures). On the other
hand, few key clinical findings ap-
peared in the models for more than 2 of
the conditions and procedures. The
Glasgow Coma Score was a risk factor
for 3 of the 5 conditions and 1 of the 3
procedures and the American Society of
Table. Number of Hospitals, Cases, and Fatalities and Mortality Rate for Each Condition and
Procedure
Condition or Procedure
No. of
Hospitals
No. of
Cases
No. of
Deaths
Mortality
Rate, %
Pneumonia 188 176 696 14 552 8.2
Congestive heart failure 187 200 506 8739 4.4
Acute cerebrovascular accident 187 82 682 8960 10.8
Gastrointestinal tract hemorrhage 187 75 392 2507 3.3
Acute myocardial infarction 184 104 110 9821 9.4
Abdominal aortic aneurysm repair 139 5309 557 10.5
Craniotomy 100 16 928 1169 6.9
Coronary artery bypass graft surgery 63 58 879 1890 3.2
CLAIMS DATA FOR IMPROVING RISK-ADJUSTED HOSPITAL MORTALITY
©2007 American Medical Association. All rights reserved. (Reprinted) JAMA, January 3, 2007—Vol 297, No. 1 73
at Attn Clinical Affairs, on January 3, 2007 www.jama.comDownloaded from
Anesthesiologists classification was a risk
factor for the 2 other procedures.
The receiver operating characteris-
tic curves reflecting the average csta-
tistics of alternative models are shown
in the FIGURE. The average cstatistic
increased from 0.50 for no risk adjust-
ment to a mean (SD) of 0.79 (0.02) for
the administrative model, to 0.84 (0.01)
for the POA model, to 0.86 (0.01) for
the laboratory model, to 0.88 (0.01) for
the clinical model.
COMMENT
This study was designed to guide the
selection of a cost-effective set of clini-
cal data elements to improve the valid-
ity of comparisons of risk-adjusted hos-
pital mortality rates. Because these
comparisons often are important com-
ponents of public reports, pay-for-
performance programs, and quality im-
provement initiatives, it is essential that
they accurately reflect the quality of care
provided by each facility.41 Unlike most
previous studies that attempted to de-
rive the most parsimonious or most so-
phisticated risk-adjustment model for
a single condition or procedure or to
compare models based on administra-
tive data to corresponding models based
on clinical data, the principal goal of this
study was to evaluate the relative per-
formances of alternative equations
based on progressively more detailed
data sets and identify one that could
meet the sometimes conflicting needs
of physicians, hospital administrators,
and payers. Therefore, a diverse sample
of conditions and procedures was evalu-
ated, and methodological uniformity
was emphasized to minimize the con-
founding effects of differences in ana-
lytic technique and to obtain precise es-
timates of improvements in the risk
adjustment directly attributable to
changing the type of data available for
use in the predictive equations.
In deriving the administrative and
POA models, care was taken to avoid
using risk factors based on conditions
or procedures that reflected poten-
tially avoidable hospital-acquired com-
plications rather than intrinsic patient
risks at the time of admission. For both
the administrative and POA models, the
use of procedure codes as risk factors
was limited to situations in which they
were found to be irreplaceable surro-
gates for intrinsic patient risk. In the ad-
ministrative model, secondary diag-
noses were eligible for use as risk factors
only when a separate analysis docu-
mented that they only rarely reflected
hospital-acquired complications. In the
POA model, secondary diagnoses in-
eligible for the administrative model
were considered as potential risk fac-
tors only if the Atlas clinical data sub-
stantiated their presence on the first day
of hospitalization.
A national standard for adding a POA
code to administrative claims data in the
UB-04 (the uniform bill used to sub-
mit all hospital claims to payers) is
planned as part of the revised ICD-
9-CM coding modifications for 2007.31,42
In this study, the use of a surrogate for
this code resulted in noteworthy im-
provements in the performance of the
risk-adjustment models, confirming the
value of this new coding convention.
Substantial additional improvements in
the performance of risk-adjustment
models occurred when numerical labo-
ratory values were added to the POA
codes. The sum of all further improve-
ments from adding other clinical data
elements was substantially less than
improvements achieved by adding sur-
rogate POA coding and numerical labo-
ratory values to the standard admin-
istrative data.
Data collected by the Pennsylvania
Health Care Cost Containment Coun-
cil43 demonstrated that when hospi-
tals routinely collect a specified set of
clinical data elements on a large num-
ber of discharges, they can reduce the
cost of retrieving these data by invest-
ing in standardized record formats and
electronic aids to data collection and by
training less expensive personnel to ob-
tain required data quickly and accu-
rately. In addition, a recent study44 by
HIMSS Analytics found that 80.8% of
hospitals had the computerized labo-
ratory systems required to support the
laboratory models. Therefore, many
Figure. Receiver Operating Characteristic Curves for the Models
No Risk Adjustment
Laboratory Model
Administrative Model
POA Model
Clinical Model
100
60
40
80
20
0
0
100 80 60 40 20 0
0%
2%
4%
6%
8%
100%
20 40 60 80 100
100 – Specificity, %
Specificity, %
Sensitivity, %
The 4 receiver operating characteristic curves represent data from each of the 4 risk-adjustment models: stan-
dard administrative model; present at admission (POA) model, standard administrative model with POA modi-
fiers; laboratory model, POA plus numerical laboratory data; and clinical model, laboratory model plus all avail-
able clinical data. The diagonal dotted line indicates no risk adjustment. The data markers represent cut points
of mortality risk predicted by each model in increments of 2%. Mortality rate cut points shown in the plot
include 0% (upper right), 2% (gray), 4% (blue), 6% (pink), 8% (black), and 100% (bottom left). Each model
is based on 8 data sets (1 set of data for each of the 5 conditions and 1 set of data for each of the 3 proce-
dures). Predicted mortality rates from the 8 data sets in each model were averaged and compared with the
observed values to calculate the true positive rate (sensitivity, where positive equals dead) and false-positive
rate (100 minus specificity) at each mortality risk cut point.
CLAIMS DATA FOR IMPROVING RISK-ADJUSTED HOSPITAL MORTALITY
74 JAMA, January 3, 2007—Vol 297, No. 1 (Reprinted) ©2007 American Medical Association. All rights reserved.
at Attn Clinical Affairs, on January 3, 2007 www.jama.comDownloaded from
hospitals currently should be capable
of electronically merging administra-
tive and numerical laboratory data,
thereby reducing their costs of acquir-
ing laboratory data by eliminating the
need for manual abstraction. In con-
trast, only 10.6% of hospitals cur-
rently have computerized nursing docu-
mentation required to support models
that include vital signs and other clini-
cal data, although rapid improve-
ments in health information technol-
ogy are anticipated over the next few
years.33
The present study was limited by its
use of only 1 indicator of hospital qual-
ity (ie, mortality) and by its failure to
evaluate directly the effects of varia-
tions in coding practices and in the
number of secondary diagnosis codes
included in the centralized databases.
Recommendations about the inclu-
sion of specific data elements within
each level of clinical data may not ap-
ply in all circumstances because some
data elements not identified in this
study might prove to be important for
outcomes and conditions outside the
scope of this investigation. In addi-
tion, measures of function were not
available but have been shown to have
value in predicting outcomes.45
In summary, this analysis strongly
supports the value of enhancing ad-
ministrative claims data with POA codes
and a limited set of numerical labora-
tory values obtained at admission.
These data provide information re-
quired to avoid errors in the designa-
tion of hospitals and their medical staffs
as delivering better than average or
worse than average care. On the other
hand, secondary abstraction of difficult-
to-obtain key clinical findings appears
to add little to the risk adjustment of
inpatient mortality rates.
Author Contributions: Dr Pine had full access to all
of the data in the study and takes responsibility for
the integrity of the data and the accuracy of the data
analysis.
Study concept and design: Pine, Jordan, Elixhauser,
Hoaglin.
Acquisition of data: Jordan, Elixhauser, Jones.
Analysis and interpretation of data: Pine, Jordan, Fry,
Hoaglin, Jones, Meimban, Warner, Gonzales.
Drafting of the manuscript: Pine, Jordan, Fry, Jones,
Meimban, Warner.
Critical revision of the manuscript for important in-
tellectual content: Jordan, Elixhauser, Hoaglin,
Gonzales.
Statistical analysis: Pine, Jordan, Hoaglin, Jones,
Meimban.
Obtained funding: Pine, Jordan, Elixhauser.
Administrative, technical, or material support: Pine,
Jordan, Elixhauser, Warner.
Study supervision: Pine, Jordan, Elixhauser.
Financial Disclosures: Dr Hoaglin reported owning
shares of stock in Cardinal Health Inc. None of the other
authors reported any disclosures.
Funding/Support: This study was supported by the
Agency for Healthcare Research and Quality con-
tract 233-02-0088 (task order 13).
Role of the Sponsor: The Agency for Healthcare Re-
search and Quality (AHRQ) specified the overarch-
ing study design in a request for proposal. Data were
collected by Pennsylvania hospitals, which were re-
quired by law to submit these data to the Pennsylva-
nia Health Care Cost Containment Council. Data were
transmitted to the AHRQ by the Council and the AHRQ
transmitted them to the authors. Further data man-
agement, analyses, interpretation, and preparation of
the manuscript were the independent work of the au-
thors. The manuscript was reviewed by the Council
and by the director of the Center for Delivery, Orga-
nizations, and Markets at AHRQ prior to the initial sub-
mission. Complete specifications of data elements, po-
tential risk factors, and risk-adjustment equations are
available in the final report submitted to the AHRQ.
Administrative and Atlas clinical data were provided
by the Council, an independent state agency respon-
sible for addressing the problem of escalating health
costs, ensuring the quality of health care, and increas-
ing access to health care for all citizens regardless of
ability to pay. The Council provided data in an effort
to further its mission of educating the public and con-
taining health care costs in Pennsylvania. The Coun-
cil, its agents and staff, have made no representa-
tion, guarantee, or warranty, express or implied, that
the data are error-free, or that the use of the data will
avoid differences of opinion or interpretation. Analy-
ses reported in this article were not prepared by the
Council.
Disclaimer: The Pennsylvania Health Care Cost Con-
tainment Council, its agents and staff, bear no re-
sponsibility or liability for the results of these analy-
ses, which are solely the opinion of the authors.
REFERENCES
1. Agency for Healthcare Research and Quality. In-
patient quality indicators, technical specifications [Feb-
ruary 20, 2006]. http://www.qualityindicators.ahrq
.gov/downloads/iqi/iqi_technical_specs_v30.pdf.
Accessed November 6, 2006.
2. HealthGrades Inc Web site. http://www
.healthgrades.com. Accessed November 6, 2006.
3. McKay NL, Deily ME. Comparing high- and low-
performing hospitals using risk-adjusted excess mor-
tality and cost inefficiency. Health Care Manage Rev.
2005;30:347-360.
4. Hannan EL, Kilburn H Jr, Lindsey ML, et al. Clini-
cal versus administrative data bases for CABG sur-
gery: does it matter? Med Care. 1992;30:892-907.
5. Jollis JG, Ancukiewicz M, DeLong ER, et al. Dis-
cordance of databases designed for claims payment
versus clinical information systems: implications for out-
comes research. Ann Intern Med. 1993;119:844-850.
6. Iezzoni LI, Ash AS, Coffman GA, et al. Predicting
in-hospital mortality: a comparison of severity mea-
surement approaches. Med Care. 1992;30:347-359.
7. Iezzoni LI, Ash AS, Shwartz M, et al. Predicting who
dies depends on how severity is measured: implica-
tions for evaluating patient outcomes. Ann Intern Med.
1995;123:763-770.
8. Hannan EL, Racz MJ, Jollis JG, et al. Using Medi-
care claims data to assess provider quality for CABG
surgery: does it work well enough? Health Serv Res.
1997;31:659-678.
9. Pine M, Norusis M, Jones B, et al. Predictions of
hospital mortality rates: a comparison of data sources.
Ann Intern Med. 1997;126:347-354.
10. Knaus WA, Zimmerman JE, Wagner DP, et al.
APACHE-acute physiology and chronic health evalu-
ation: a physiologically based classification system. Crit
Care Med. 1981;9:591-597.
11. Brewster AC, Karlin BG, Hyde LA, et al. MEDIS-
GRPS: a clinically based approach to classifying hos-
pital patients at admission. Inquiry. 1985;22:377-387.
12. Daley J, Jencks S, Draper D, et al. Predicting hos-
pital-associated mortality for Medicare patients: a
method for patients with stroke, pneumonia, acute
myocardial infarction, and congestive heart failure.
JAMA. 1988;260:3617-3624.
13. Hannan EL, Kilburn H Jr, O’Donnell JF, et al. Adult
open heart surgery in New York State: an analysis of
risk factors and hospital mortality rates. JAMA. 1990;
264:2768-2774.
14. Smith DW, Pine M, Bailey RC, et al. Using clini-
cal variables to estimate the risk of patient mortality.
Med Care. 1991;29:1108-1129.
15. Higgins TL, Estafanous FG, Loop FD, et al. Strati-
fication of morbidity and mortality outcome by
preoperative risk factors in coronary artery bypass pa-
tients: a clinical severity score. JAMA. 1992;267:2344-
2348.
16. Lemeshow S, Teres D, Klar J, et al. Mortality prob-
ability models (MPM II) based on an international co-
hort of intensive care unit patients. JAMA. 1993;270:
2478-2486.
17. Van Ruiswyk J, Hartz A, Kuhn E, et al. A mea-
sure of mortality risk for elderly patients with acute
myocardial infarction. Med Decis Making. 1993;13:
152-160.
18. Khuri SF, Daley J, Henderson W, et al. The De-
partment of Veterans Affairs’ NSQIP [National VA Sur-
gical Quality Improvement Program]: the first na-
tional, validated, outcome-based, risk-adjusted, and
peer-controlled program for the measurement and en-
hancement of the quality of surgical care. Ann Surg.
1998;228:491-507.
19. Pine M, Jones B, Lou YB. Laboratory values im-
prove predictions of hospital mortality. Int J Qual
Health Care. 1998;10:491-501.
20. Gibbs J, Cull W, Henderson W, et al. Preopera-
tive serum albumin level as a predictor of operative
mortality and morbidity. Arch Surg. 1999;134:36-42.
21. Ghali WA, Faris PD, Galbraith PD, et al. Sex dif-
ferences in access to coronary revascularization after
cardiac catheterization: importance of detailed clini-
cal data. Ann Intern Med. 2002;136:723-732.
22. Grover FL, Shroyer AL, Hammermeister K, et al.
A decade’s experience with quality improvement in
cardiac surgery using the Veterans Affairs and Soci-
ety of Thoracic Surgeons national databases. Ann Surg.
2001;234:464-474.
23. Lee DS, Austin PC, Rouleau JL, et al. Predicting
mortality among patients hospitalized for heart fail-
ure: derivation and validation of a clinical model. JAMA.
2003;290:2581-2587.
24. Schneider EC, Epstein AM. Influence of cardiac
surgery performance reports on referral practices and
access to care: a survey of cardiovascular specialists.
N Engl J Med. 1996;335:251-256.
25. Moscucci M, Eagle KA, Share D, et al. Public re-
porting and case selection for percutaneous coro-
nary interventions: an analysis from two large multi-
center percutaneous coronary intervention databases.
J Am Coll Cardiol. 2005;45:1759-1765.
26. Narins CR, Dozier AM, Ling FS, et al. The influ-
ence of public reporting of outcome data on medical
decision making by physicians. Arch Intern Med. 2005;
165:83-87.
27. National Health Care Purchasing Institute.
Status report: consumer and purchaser disclosure
CLAIMS DATA FOR IMPROVING RISK-ADJUSTED HOSPITAL MORTALITY
©2007 American Medical Association. All rights reserved. (Reprinted) JAMA, January 3, 2007—Vol 297, No. 1 75
at Attn Clinical Affairs, on January 3, 2007 www.jama.comDownloaded from
project [October 2002]. http://www.pbgh
.org/programs/standardization/documents
/DisclosureProjectNewsletter-October2002.pdf. Ac-
cessed November 6, 2006.
28. Krumholz HM, Wang Y, Mattera JA, et al. An ad-
ministrative claims model suitable for profiling hospital
performance based on 30-day mortality rates among
patients with an acute myocardial infarction. Circulation.
2006;113:1683-1692.
29. Krumholz HM, Wang Y, Mattera JA, et al. An ad-
ministrative claims model suitable for profiling hospital
performance based on 30-day mortality rates among
patients with heart failure. Circulation. 2006;113:1693-
1701.
30. Naessens JM, Brennan MD, Boberg CJ, et al. Ac-
quired conditions: an improvement to hospital dis-
charge abstracts. Qual Assur Health Care. 1991;3:257-
262.
31. Coffey R, Milenkovic M, Andrews RM; US Agency
for Healthcare Research and Quality. The case for the
present on admission (POA) indicator: HCUP methods
series report 2006-01 [June 26, 2006]. http://www
.hcup-us.ahrq.gov/reports/2006_1.pdf. Accessed No-
vember 6, 2006.
32. Regenstreif Institute. Logical observation identifi-
ers names and codes (LOINC). http://www.regenstrief
.org/loinc. Accessed November 6, 2006.
33. Berner ES, Detmer DE, Simborg D. Will the wave
finally break? a brief view of the adoption of electronic
medical records in the United States. J Am Med Inform
Assoc. 2005;12:3-7.
34. Pennsylvania Health Care Cost Containment Coun-
cil Web site. http://www.phc4.org. Accessed Novem-
ber 6, 2006.
35. Cardinal Health. Atlas outcomes system. http:
//www.mediqual.com/products/atlas.asp. Accessed No-
vember 6, 2006.
36. Agency for Healthcare Research and Quality. Qual-
ity indicators software for SAS, zipped. http://www
.qualityindicators.ahrq.gov/iqi_download.htm. Ac-
cessed November 6, 2006.
37. Hosmer DW, Lemeshow S. Logistic regression. In:
Applied Logistic Regression. 2nd ed. New York, NY: John
Wiley and Sons; 2000:91-142.
38. Schwarz G. Estimating the dimension of a model.
Ann Stat. 1978;6:461-464.
39. Hastie T, Tibshirani R, Friedman J. The Elements of
Statistical Learning. New York, NY: Springer-Verlag; 2001.
40. Hanley JA, McNeil BJ. The meaning and use of the
area under a receiver operating characteristic (ROC)
curve. Radiology. 1982;143:29-36.
41. Lee TH, Meyer GS, Brennan TA. A middle ground
on public accountability. N Engl J Med. 2004;350:2409-
2412.
42. National Uniform Billing Committee. UB-04 Data
specifications manual (Beta3): present on admission
indicator. http://www.nubc.org/public/whatsnew/POA
.pdf. Accessed November 6, 2006.
43. Pennsylvania Healthcare Cost Containment Council.
Independent accountant’s [Parente Randolph] report
on applying agreed-upon procedures [August 5, 2004].
http://www.phc4.org. Accessed December 1,
2006.
44. Garets D, Davis M. Electronic medical records vs elec-
tronic health records: yes, there is a difference: a HIMSS
analytics white paper [2006]. http://www.himssanalytics
.org/docs/WP_EMR_EHR.pdf. Accessed November 6,
2006.
45. Lee SJ, Lindquist K, Segal MR, Covinsky KE. De-
velopment and validation of a prognostic index for 4-year
mortality in older adults. JAMA. 2006;295:801-
808.
Of all the inanimate objects, of all of man’s creations,
books are the nearest to us, for they contain our very
thought, our ambitions, our indignations, our illu-
sions, our fidelity to truth, and our persistent lean-
ing toward error.
—Joseph Conrad (1857-1924)
CLAIMS DATA FOR IMPROVING RISK-ADJUSTED HOSPITAL MORTALITY
76 JAMA, January 3, 2007—Vol 297, No. 1 (Reprinted) ©2007 American Medical Association. All rights reserved.
at Attn Clinical Affairs, on January 3, 2007 www.jama.comDownloaded from
... Our study had several important limitations. Although administrative data contain information across all healthcare sectors over several years, such data lack clinical predictors that might add validity to the index [50]. For instance, we did not consider the time-to-balloon and the revascularization path for STEMI and NSTEMI [51]. ...
Article
Full-text available
Objective Development of an aggregate quality index to evaluate hospital performance in cardiovascular events treatment. Methods We applied a two-stage regression approach using an accelerated failure time model based on variance weights to estimate hospital quality over four cardiovascular interventions: elective coronary bypass graft, elective cardiac resynchronization therapy, and emergency treatment for acute myocardial infarction. Mortality and readmissions were used as outcomes. For the estimation we used data from a statutory health insurer in Germany from 2005 to 2016. Results The precision-based weights calculated in the first stage were higher for mortality than for readmissions. In general, teaching hospitals performed better in our ranking of hospital quality compared to non-teaching hospitals, as did private not-for-profit hospitals compared to hospitals with public or private for-profit ownership. Discussion The proposed approach is a new method to aggregate single hospital quality outcomes using objective, precision-based weights. Likelihood-based accelerated failure time models make use of existing data more efficiently compared to widely used models relying on dichotomized data. The main advantage of the variance-based weights approach is that the extent to which an indicator contributes to the aggregate index depends on the amount of its variance.
... 20,21 History of tobacco use versus non-use in the last 12 months, 18,22,23 and significant alcohol use, defined as two or more drinks daily (averaged over a week), were also recorded. 24,25 Characteristics of the ICU patients were also collected, including admission APACHE II score, Elixhauser comorbidity score, 26,27 age, sex, ICU and hospital length of stay, in-hospital mortality, 3-month mortality, and number of prior ICU admissions for the patient in the last 5 years, based on family member report. ...
Article
Full-text available
Objective To determine whether heart rate variability (HRV; a physiological measure of acute stress) is associated with persistent psychological distress among family members of adult intensive care unit (ICU) patients. Methods This prospective study investigated family members of patients admitted to a study ICU. Participants’ variability in heart rate tracings were measured by low frequency (LF)/high frequency (HF) ratio and detrended fluctuation analysis (DFA). Questionnaires were completed 3 months after enrollment to ascertain outcome rates of anxiety, depression, and post-traumatic stress disorder (PTSD). Results Ninety-nine participants were enrolled (median LF/HF ratio, 0.92 [interquartile range, 0.64–1.38]). Of 92 participants who completed the 3-month follow-up, 29 (32%) had persistent anxiety. Logistic regression showed that LF/HF ratio (odds ratio [OR] 0.85, 95% confidence interval [CI] 0.43, 1.53) was not associated with 3-month outcomes. In an exploratory analysis, DFA α (OR 0.93, 95% CI 0.87, 0.99), α 1 (OR 0.97, 95% CI 0.94, 0.99), and α 2 (OR 0.94, 95% CI 0.88, 0.99) scaling components were associated with PTSD development. Conclusion Almost one-third of family members experienced anxiety at three months after enrollment. HRV, measured by LF/HF ratio, was not a predictor of psychologic distress, however, exploratory analyses indicated that DFA may be associated with PTSD outcomes.
Article
Objectives: To assess their construct validity, we compared results from 2 models used for estimating hospital standardized mortality ratios (HSMRs) in Belgium. The method of the Flemish Hospital Network (FHN) is based on a logistic regression for each of the 64 All Patient Refined Diagnosis-Related Groups that explain 80% of mortality and uses the Elixhauser score to correct for comorbidities. (H)SMRs published on the 3M-Benchmark-Portal are calculated by a simpler indirect standardization for All Patient Refined Diagnosis-Related Groups and risk of mortality (ROM) at discharge. Methods: We used administrative data from all eligible hospital admissions in 22 Flemish hospitals between 2016 and 2019 (FHN, n = 682,935; 3M, n = 2,122,305). We evaluated model discrimination and accuracy and assessed agreement in estimated HSMRs between methods. Results: The Spearman correlation between HSMRs generated by the FHN model and the standard 3M model was 0.79. Although 2 of 22 hospitals showed opposite classification results, that is, an HSMR significantly <1 according to the FHN method but significantly >1 according to the 3M model, classification agreement between methods was significant (agreement for 59.1% of hospitals, κ = 0.45). The 3M model (c statistic = 0.96, adjusted Brier score = 26%) outperformed the FHN model (0.87, 17%). However, using ROM at admission instead of at discharge in the 3M model significantly reduced model performance (c statistic = 0.94, adjusted Brier score = 21%), but yielded similar HSMR estimates and eliminated part of the discrepancy with FHN results. Conclusions: Results of both models agreed relatively well, supporting convergent validity. Whereas the FHN method only adjusts for disease severity at admission, the ROM indicator of the 3M model includes diagnoses not present on admission. Although diagnosis codes generated by complications during hospitalization have the tendency to increase the predictive performance of a model, these should not be included in risk adjustment procedures.
Article
Objective: To measure hospital quality based on routine data available in many health care systems including the United States, Germany, the United Kingdom, Scandinavia, and Switzerland. Data Sources and Study Setting: We use the Swiss Medical Statistics of Hospitals, an administrative hospital dataset of all inpatient stays in acute care hospitals in Switzerland for the years 2017–2019. Study Design: We study hospital quality based on quality indicators used by leading agencies in five countries (the United States, the United Kingdom, Germany, Austria, and Switzerland) for two high-volume elective procedures: inguinal hernia repair and hip replacement surgery. We assess how the least absolute shrinkage and selection operator (LASSO), a supervised machine learning technique for variable selection, and Mundlak corrections that account for unobserved heterogeneity between hospitals can be used to improve risk adjustment and correct for imbalances in patient risks across hospitals. Data Collection/Extraction Methods: The Swiss Federal Statistical Office collects annual data on all acute care inpatient stays including basic socio-demographic patient attributes and case-level diagnosis and procedure codes. Principal Findings: We find that LASSO-selected and Mundlak-corrected hospital random effects logit models outperform common practice logistic regression models used for risk adjustment. Besides the more favorable statistical properties, they have superior in- and out-of-sample explanatory power. Moreover, we find that Mundlak-corrected logits and the more complex LASSO-selected models identify the same hospitals as high or low-quality offering public health authorities a valuable alternative to standard logistic regression models. Our analysis shows that hospitals vary considerably in the quality they provide to patients. Conclusion: We find that routine hospital data can be used to measure clinically relevant quality indicators that help patients make informed hospital choices.
Article
Objective: To explore the relationship between novel, time-varying predictors for healthcare delivery strain (eg, counts of patient orders per hour) and imminent discharge and in-hospital mortality. Materials and methods: We conducted a retrospective cohort study using data from adults hospitalized at 21 Kaiser Permanente Northern California hospitals between November 1, 2015 and October 31, 2020 and the nurses caring for them. Patient data extracted included demographics, diagnoses, severity measures, occupancy metrics, and process of care metrics (eg, counts of intravenous drip orders per hour). We linked these data to individual registered nurse records and created multiple dynamic, time-varying predictors (eg, mean acute severity of illness for all patients cared for by a nurse during a given hour). All analyses were stratified by patients' initial hospital unit (ward, stepdown unit, or intensive care unit). We used discrete-time hazard regression to assess the association between each novel time-varying predictor and the outcomes of discharge and mortality, separately. Results: Our dataset consisted of 84 162 161 hourly records from 954 477 hospitalizations. Many novel time-varying predictors had strong associations with the 2 study outcomes. However, most of the predictors did not merely track patients' severity of illness; instead, many of them only had weak correlations with severity, often with complex relationships over time. Discussion: Increasing availability of process of care data from automated electronic health records will permit better quantification of healthcare delivery strain. This could result in enhanced prediction of adverse outcomes and service delays. Conclusion: New conceptual models will be needed to use these new data elements.
Article
Coding variation distorts performance/outcome statistics not eliminated by risk adjustment. Among 1596 community-acquired pneumonia patients hospitalized from 1998 to 2012 identified using an evidence-based algorithm, the authors measured the association of principal diagnosis (PD) with 30-day readmission, stratified by Pneumonia Severity Index risk class. The 152 readmitted patients were more ill (Pneumonia Severity Index class V 38.8% versus 25.8%) and less likely to have a pneumonia PD (52.6% versus 69.9%). Among patients with PDs of pneumonia, respiratory failure, sepsis, and aspiration, mortality/readmission rates were 3.9/8.5%, 28.8/14.0%, 24.7/19.6%, and 9.0/15.0%, respectively. The nonpneumonia PDs were associated with a greater risk of adjusted 30-day readmission: respiratory failure odds ratio (OR) 1.89 (95% confidence interval [CI], 1.13-3.15), sepsis OR 2.54 (95% CI, 1.52-4.26), and possibly aspiration OR 1.73 (95% CI, 0.88-3.41). With increasing use of alternative PDs among pneumonia patients, quality reporting must account for variations in condition coding practices. Rigorous risk adjustment does not eliminate the need for accurate, consistent case definition in producing valid quality measures.
Article
Full-text available
Background: Administrative claims databases have been used widely in studies because they have large sample sizes and are easily available. However, studies using administrative databases lack the severity of the disease, so a risk adjustment method needs to be developed. Objective: To develop and validate deep learning-based prediction models for in-hospital mortality of acute-care patients. Methods: The main model was developed using only administrative claims data (age, sex, diagnoses, and procedures on the day of admission). We also constructed disease-specific models for acute myocardial infarction, heart failure, stroke, or pneumonia using common severity indices for these diseases. Using the Japanese Diagnosis Procedure Combination data from July 2010 to March 2017, we identified 46,665,933 inpatients and divided them into derivation and validation cohorts in a ratio of 95:5. The main model was developed using a 9-layer deep neural network with four hidden dense layers that had 1000 nodes and were fully connected to adjacent layers. We evaluated model discrimination ability by an area under the receiver operating characteristics curve (AUC) and calibration ability by calibration plot. Results: Among the eligible patients, 2,005,035 (4.3%) died. Discrimination and calibration of the models were satisfactory. The AUC of the main model in the validation cohort was 0.954 (95% confidential interval 0.9537-0.9547). The main model had higher discrimination ability than the disease-specific models. Conclusions: Our deep learning-based model using diagnoses and procedures produced valid predictions of in-hospital mortality. Clinicaltrial:
Article
Objectives: To develop a model to benchmark mortality in hospitalized patients using accessible electronic medical record data. Design: Univariate analysis and multivariable logistic regression were used to identify variables collected during the first 24 hours following admission to test for risk factors associated with the end point of hospital mortality. Models were built using specific diagnosis (International Classification of Diseases, 9th Edition or International Classification of Diseases, 10th Edition) captured at discharge, rather than admission diagnosis, which may be discordant. Variables were selected based, in part, on prior the Acute Physiology and Chronic Health Evaluation methodology and included primary diagnosis information plus three aggregated indices (physiology, comorbidity, and support). A Physiology Index was created using parsimonious nonlinear modeling of heart rate, mean arterial pressure, temperature, respiratory rate, hematocrit, platelet counts, and serum sodium. A Comorbidity Index incorporates new or ongoing diagnoses captured by the electronic medical record during the preceding year. A Support Index considered 10 interventions such as mechanical ventilation, selected IV drugs, and hemodialysis. Accuracy was determined using area under the receiver operating curve for discrimination, calibration curves, and modified Brier score for calibration. Setting and patients: We used deidentified electronic medical record data from 74,434 adult inpatients (ICU and ward) at 15 hospitals from 2010 to 2013 to develop the mortality model and validated using data for additional 49,752 patients from the same 15 hospitals. A second revalidation was accomplished using data on 83,684 patients receiving care at six hospitals between 2014 and 2016. The model was also validated on a subset of patients with an ICU stay on day 1. Interventions: None. Measurements and main results: This model uses physiology, comorbidity, and support indices, primary diagnosis, age, lowest Glasgow Coma Score, and elapsed time since hospital admission to predict hospital mortality. In the initial validation cohort, observed mortality was 4.04% versus predicted mortality 4.12% (Student t test, p = 0.37). In the revalidation using a different set of hospitals, predicted and observed mortality were 2.66% and 2.99%, respectively. Area under the receiver operating curve were 0.902 (0.895-0.909) and 0.884 (0.877-0.891), respectively, and calibration curves show a close relationship of observed and predicted mortalities. In the evaluation of the subset of ICU patients on day1, the area under the receiver operating curve was 0.87, with an observed mortality of 8.78% versus predicted mortality of 8.93% (Student t test, p = 0.52) and a standardized mortality ratio of 0.98 (0.932-1.034). Conclusions: Variables considered by traditional ICU prognostic models accurately benchmark patient mortality for patients receiving care in multiple hospital locations, not only the ICU. Unlike Acute Physiology and Chronic Health Evaluation, this model relies on electronic medical record data alone and does not require personnel to collect the independent predictor variables. Assessing the model's utility for benchmarking hospital performance will require prospective testing in a larger representative sample of hospitals.
Article
Full-text available
Introduction to the Logistic Regression Model Multiple Logistic Regression Interpretation of the Fitted Logistic Regression Model Model-Building Strategies and Methods for Logistic Regression Assessing the Fit of the Model Application of Logistic Regression with Different Sampling Models Logistic Regression for Matched Case-Control Studies Special Topics References Index.
Article
Hospital mortality statistics derived from administrative data may not adjust adequately for patient risk on admission. Using clinical data collected from the medical record, this study compared the ability of six models to predict in-hospital death, including one model based on administrative data (age, sex, and principal and secondary diagnoses), one on admission MedisGroups score, and one on an approximation of the Acute Physiology Score (APS) from the revised Acute Physiology and Chronic Health Evaluation (APACHE II), as well as three empirically derived models. The database from 24 hospitals included 16,855 cases involving five medical conditions, with an overall in-hospital mortality rate of 15.6%. The administrative data model fit least well (R-squared values ranged from 1.9-5.5% across the five conditions). Admission MedisGroups score and the proxy APS score did better, with R-squared values ranging from 4.9% to 25.9%. Two empirical models based on small subsets of explanatory variables performed best (R-squared values ranged from 18.5-29.9%). The preceding models had the same relative performances after cross-validation using split samples. However, the high R-squared values produced by the full empirical models (using 40 or more explanatory variables) were not preserved when they were cross-validated. Most of the predictive clinical findings were general physiologic measures that were similar across conditions; only a fifth of predictors were condition-specific. Therefore, an efficient approach to risk-adjusting in-hospital mortality figures may involve adding a small subset of condition-specific clinical variables to a core group of acute physiologic variables. The best predictive models employ condition-specific weighting of even the generic clinical findings.
Article
Objective To improve the precision and reliability of estimates of the association between preoperative serum albumin concentration and surgical outcomes. Design Prospective observational study. Patients were followed up for 30 days postoperatively. Multiple logistic regression models were developed to evaluate serum albumin level as a predictor of operative mortality and morbidity in relation to 61 other preoperative patient risk variables. Setting Forty-four tertiary care Veterans Affairs (VA) medical centers. Patients A total of 54,215 major noncardiac surgery cases from the National VA Surgical Risk Study. Main Outcome Measures Thirty-day operative mortality and morbidity. Results A decrease in serum albumin from concentrations greater than 46 g/L to less than 21 g/L was associated with an exponential increase in mortality rates from less than 1% to 29% and in morbidity rates from 10% to 65%. In the regression models, albumin level was the strongest predictor of mortality and morbidity for surgery as a whole and within several subspecialties selected for further analysis. Albumin level was a better predictor of some types of morbidity, particularly sepsis and major infections, than other types. Conclusions Serum albumin concentration is a better predictor of surgical outcomes than many other preoperative patient characteristics. It is a relatively low-cost test that should be used more frequently as a prognostic tool to detect malnutrition and risk of adverse surgical outcomes, particularly in populations in whom comorbid conditions are relatively frequent.
Article
. Objective : To determine whether assessments of illness severity, defined as risk for in-hospital death, varied across four severity measures. . Design : Retrospective cohort study. . Setting : 100 hospitals using the MedisGroups severity measure. . Patients : 11 880 adults managed medically for acute myocardial infarction ; 1574 in-hospital deaths (13.2%). . Measurements : For each patient, probability of death was predicted four times, each time by using patient age and sex and one of four common severity measures : 1) admission MedisGroups scores for probability of death scores ; 2) scores based on values for 17 physiologic variables at time of admission ; 3) Disease Staging's probability-of-mortality model ; and 4) All Patient Refined Diagnosis Related Groups (APR-DRGs). Patients were ranked according to probability of death as predicted by each severity measure, and rankings were compared across measures. The presence or absence of each of six clinical findings considered to indicate poor prognosis in patients with myocardial infarction (congestive heart failure, pulmonary edema, coma, low systolic blood pressure, low left ventricular ejection fraction, and high blood urea nitrogen level) was determined for patients ranked differently by different severity measures. . Results : MedisGroups and the physiology score gave 94.7% of patients similar rankings. Disease Staging, MedisGroups, and the physiology score gave only 78% of patients similar rankings. MedisGroups and APR-DRGs gave 80% of patients similar rankings. Patients whose illnesses were more severe according to MedisGroups and the physiology score were more likely to have the six clinical findings than were patients whose illnesses were more severe according to Disease Staging and APR-DRGs. . Conclusions : Some pairs of severity measures assigned very different severity levels to more than 20% of patients. Evaluations of patient outcomes need to be sensitive to the severity measures used for risk adju
Article
Background: Comparing hospital mortality rates requires accurate adjustment for patients' intrinsic differences. Commercial severity systems require either administrative data that omit vital clinical facts about patients' conditions at hospital admission or costly, time-consuming ion of medical records. The validity of supplementing administrative data with laboratory data has not been assessed. Objective: To compare risk-adjusted mortality predictions using administrative data alone; administrative data plus laboratory values; and the combination of administrative, laboratory, and clinical data. Design: Retrospective cohort study. Setting: 30 acute care hospitals. Patients: 46 769 patients hospitalized with acute myocardial infarction, cerebrovascular accident, congestive heart failure, or pneumonia. Measurements: Each patient's probability of dying was estimated by using administrative data only (unrestricted administrative models), administrative data restricted to secondary diagnoses that are unlikely to be hospital-acquired complications (restricted administrative models), restricted administrative data plus laboratory data (laboratory models), and restricted administrative data plus laboratory and abstracted clinical data (clinical models). Results: The unrestricted administrative models predicted death better than the restricted administrative models (average areas under the receiver-operating characteristic [ROC] curves, 0.87 and 0.75, respectively) and as well as the laboratory models and the clinical models (average areas under the ROC curves, 0.86 and 0.87, respectively). The good mortality predictions obtained by using the unrestricted administrative models result from inclusion of hospital-acquired complications that commonly precede death. The laboratory models ranked 93% of patients and 95% of hospitals in a manner similar to the clinical models; in comparison, rankings provided by the laboratory models were similar to those provided for 75% of patients and 69% of hospitals by the unrestricted administrative models and for 72% of patients and 77% of hospitals by the restricted administrative models. Conclusions: Adding laboratory data (often available electronically) to restricted administrative data sets can provide accurate predictions of inpatient death from acute myocardial infarction, cerebrovascular accident, congestive heart failure, or pneumonia. This alternative avoids the cost of data abstraction and the serious errors associated with using administrative data alone.