ICU-acquired infection rates are not an indication of patients’ mortality risk, according to researchers the University of Pennsylvania, undermining a central tenet of many pay-for-performance initiatives.
Public reporting of quality data is increasingly common in healthcare. These “report cards” are designed to improve the quality of care by helping patients choose the best hospitals. Yet, they only work if they successfully identify high performers, and may be misleading if they steer patients toward poor performers.
The findings will be reported at the American Thoracic Society 2010 International Conference in New Orleans.
To examine whether or not publicly-reported infection rates actually identify the best hospitals, Kate Courtright, MD, resident physician at the University of Pennsylvania and colleagues looked at patients in Pennsylvania hospitals especially at risk for two types of infections: pneumonia and bloodstream infections. They calculated hospital death rates accounting for differences in illness severity across 158 hospitals, which included nearly 19,000 admissions involving mechanical ventilation and over 16,000 ICU admissions involving central venous catheterization, and compared them to ICU-acquired infection rates obtained from a public state website. They then used rank correlation and linear regression to determine the relationship between infections and death.
“We found that ICU-acquired infection rates as reported on a state Web site did not correlate with death rates for at-risk patients.” said Courtright, lead author of the abstract. “In fact, hospitals with lower rates of ICU-acquired infection did not also have lower death rates for at-risk patients.” For example, the 43 hospitals that reported no cases of ICU-acquired pneumonia had an average death rate of 35.7 percent for patients receiving mechanical ventilation; hospitals with high infection rates (ranging from 1 to 8 cases per 1,000 ventilator days) had an average death rate of 34.6 percent. These numbers were not statistically different.
Despite their limitations, Courtright noted, ICU-acquired infections rates are likely to continue to be a part of hospital report cards.
However, “both policy-makers and the public should recognize that these rates, at least as reported by hospitals, provide limited information about the quality of the hospital, and may misidentify high and low performers,” she said. “More comprehensive report cards that report both complications like ICU-acquired infections and overall survival rates are needed to help patients make correct decisions. In the meantime, more care is needed to make sure that hospital report cards don’t do more harm than good. This is especially important because under upcoming health care reform, infection rates are also to be used for hospital reimbursement—hospitals with high infection rates will not be reimbursed as well for their care. Such a strategy, known as ‘pay-for-performance’, may actually penalize good hospitals with low mortality rates.”
Research on the efficacy of ‘report cards’ in predicting mortality rates must be expanded to other states or in a national study, said Courtright.
Additionally, she said, infection rates as reported by the hospitals may be incorrect as they have an incentive to report low infection rates.
“Report cards only work if they successfully identify the best hospitals,” concluded Courtright. “We were surprised to find that many hospitals with good report cards from an infection standpoint are not that good from a more important standpoint—patient survival. Additionally, many hospitals with high infection rates actually had very good survival rates. Using these report cards to choose a hospital may be misleading and potentially harmful.”