News Release

Hospital 'report cards' found not effective for quality improvement

Peer-Reviewed Publication

JAMA Network

Hospitals that were given feedback on their performance on certain quality indicators for treating heart attack patients did not show more improvement in those areas than hospitals that were provided with the feedback at a later date, according to a study in the July 20 issue of JAMA.

Acute myocardial infarction (AMI; heart attack) patients often do not receive recommended evidence-based treatments, according to background information in the article. There is increasing interest in implementing quality improvement strategies for AMI care. One quality improvement strategy that has been suggested is feedback on "quality indicators" to hospitals and clinicians treating AMI patients. Quality indicators are defined as a summary of clinical performance over a specified time. It is suggested that "report cards" presenting a summary of quality indicators relevant to care provided by individual hospitals can catalyze quality improvement at these hospitals. Ideally, hospital report cards provide clinicians with an accurate picture of the care they deliver and provide benchmarks for comparison, such as the care delivered at other hospitals or recommended target rates. Hospital report cards are increasingly being implemented in the United States and some parts of Canada as a strategy for quality improvement in many areas of health care, despite lack of strong evidence to support their use.

Christine A. Beck, M.Sc., of McGill University Health Centre, Montreal, and colleagues conducted a study to determine whether hospital report cards produced using linked administrative databases are effective for improving AMI care. The study included patients with AMI who were admitted to 76 acute care hospitals in Quebec that treated at least 30 AMI patients per year between April 1, 1999, and March 31, 2003. The hospitals were randomly assigned to receive rapid (immediate; n = 38 hospitals and 2,533 patients) or delayed (14 months; n = 38 hospitals and 3,142 patients) confidential feedback on quality indicators developed using administrative data. The quality indicators pertained to processes of care and outcomes of patients admitted between 4 and 10 months after randomization. The primary indicator was the proportion of elderly survivors of AMI at each study hospital who filled a prescription for a beta-blocker within 30 days after discharge.

The researchers found that at follow-up, adjusted prescription rates within 30 days after discharge were similar in the early vs. late groups for beta-blockers, angiotensin-converting enzyme (ACE) inhibitors, lipid-lowering drugs and aspirin. In addition, the adjusted death rate was similar in both groups, as were length of in-hospital stay, physician visits after discharge, waiting times for invasive cardiac procedures, and readmissions for cardiac complications.

"In this cluster randomized controlled trial, confidential feedback provided to hospitals in the form of report cards constructed using linked administrative data was not effective in improving quality of AMI care. Our results suggest that even if the United States eventually acquires these types of administrative data through the Medicare program, confidential feedback based on these data are unlikely to be a sufficient strategy for health care quality improvement," the authors write. "More intensive interventions, which could include chart review and continuous and/or public data feedback accompanied by other multimodal interventions, such as team workshops and standard orders, may be effective, but a need remains to study these interventions and their cost-benefit ratios in well-controlled randomized trials."

The researchers speculate that there could be several reasons for the lack of effectiveness of the study intervention, including "that the administrative data were perceived as invalid or irrelevant to practice. It is possible that report cards constructed using chart review data may be more effective than those constructed using administrative data because physicians are less skeptical of their data quality."

(JAMA. 2005;294:309–317. Available pre-embargo to the media at http://www.jamamedia.org).
Editor's Note: For funding/support information, please see the JAMA article.

Editorial: Optimizing the Science of Quality Improvement

In an accompanying editorial, Eric David Peterson, M.D., M.P.H., of Duke University Medical Center, Durham, N.C., comments on the study in this week's JAMA on quality improvement (QI).

"The report by Beck et al and other similar studies exemplify valuable prototypes for the emerging science of QI. Better means are needed to ensure that medical discoveries are adopted rapidly and completely into community practice. Like the medical discovery process, QI strategies need to be subject to rigorous evaluation to determine what works, at what intensity, and when coupled with what. While the current study intervention was insufficient to motivate change, it does provide important lessons for optimizing the process of translating evidence to practice in the future."
(JAMA. 2005;294:369–371. Available pre-embargo to the media at http://www.jamamedia.org).

###


Disclaimer: AAAS and EurekAlert! are not responsible for the accuracy of news releases posted to EurekAlert! by contributing institutions or for the use of any information through the EurekAlert system.