November 2018
National Implementation of the Hospital Consumer Assessment of Healthcare Providers and Systems
(HCAHPS) Survey
CMS 10102
OMB Supporting Statement - Part B
Prepared by
Division of Consumer Assessment & Plan Performance Centers for Medicare & Medicaid Services
7500 Security Boulevard
Baltimore, MD 21244
1
Section Page
Introduction…………………………………………………………………………………………. 4
2. Data Collection Procedures…………………………………………………………………. 5
Statistical Methodology for Stratification and Sample Selection……………………… 5
b. Estimation Procedures…………………………………………………………………. 7
Degree of Accuracy Needed for the Purpose Described in the Justification…………… 7
Unusual Problems Requiring Specialized Sampling Procedures……………………….. 7
Any Use of Periodic Data Collection Cycles to Reduce Burden……………………….. 7
Maximizing Response Rates/Non-response and Issues of Accuracy,
Reliability and Validity........................................................................................................... 8
Past and Ongoing Tests of Procedures, Training, and Quality Improvement Activities; Collaborator and Contractor Participation…………………………………………………… 12
Names and Telephone Numbers of Individuals Consulted on Statistical Aspects
of the Survey and Implementation Design and Names of Agency Units, Contractor(s), Grantee(s), or Other Persons(s) who Collect and/or Analyze the Information for the
Agency…………………………………………………………….………………………… 18
TABLES
Table 1. Hospital-Level Spearman-Brown Reliabilities of HCAHPS Top-Box Scores of HCAHPS Measures at 300 Completed Surveys, 3.1 million discharges, January – December 2016. ……………………… 9
Table 5. HCAHPS Survey Development and Implementation Timeline, 2002-2008…………….. 13
Attachment A -- HCAHPS Survey Instrument (Mail) and Supporting Material (included with Supporting Statement - Part A)
Attachment B -- HCAHPS Survey Instrument (Telephone) and Supporting Material (included with Supporting Statement - Part A)
Attachment C -- HCAHPS Survey Instrument (Interactive Voice Response) and Supporting Material (included with Supporting Statement - Part A)
National Implementation of the Hospital Consumer Assessment of Healthcare Providers and Systems (HCAHPS) Survey
CMS-10102
The Hospital Consumer Assessment of Healthcare Providers and Systems (HCAHPS) survey is administered to a random sample of adult inpatients between 48 hours and six weeks after discharge. Patients admitted in the medical, surgical and maternity care service lines are eligible for the survey. HCAHPS is not restricted to Medicare patients. Hospitals may use an approved survey vendor or collect their own HCAHPS data, if approved by CMS to do so. HCAHPS can be implemented in four survey modes: mail, telephone, mail with telephone follow-up, or active interactive voice recognition (IVR), each of which requires multiple attempts to contact patients. Hospitals must survey patients throughout each month of the year. IPPS hospitals must achieve at least 300 completed surveys over four calendar quarters. HCAHPS is available in official English, Spanish, Chinese, Russian, Vietnamese, and Portuguese versions. The HCAHPS Survey sampling protocol promotes the following: 1) standardized administration of the HCAHPS Survey by hospital/survey vendors and 2) comparability of resulting data across all participating hospitals. The survey and its protocols for sampling, data collection, coding and submission can be found in the HCAHPS Quality Assurance Guidelines manual (Version 13.0, March 2018) on the official HCAHPS On-Line Web site, www.hcahpsonline.org.
The HCAHPS Survey is broadly intended for patients of all payer types who meet the following criteria:
Eighteen (18) years or older at the time of admission
Admission includes at least one overnight stay in the hospital
Non-psychiatric MS-DRG/principal diagnosis at discharge
Alive at the time of discharge
There are a few categories of otherwise eligible patients who are excluded from the sample frame. These are:
“No-Publicity” patients – Patients who request that they not be contacted
Court/Law enforcement patients (i.e., prisoners); patients residing in halfway houses are included
Patients with a foreign home address (U.S. territories – Virgin Islands, Puerto Rico, Guam, American Samoa, and Northern Mariana Islands are not considered foreign addresses and are not excluded)
Patients discharged to hospice care (Hospice-home or Hospice-medical facility)
Patients who are excluded because of state regulations
Patients discharged to nursing homes and skilled nursing facilities
Hospitals/Survey vendors must retain documentation that verifies all exclusions and ineligible patients for a minimum of three years. This documentation is subject to review.
Hospitals/Survey vendors participating in HCAHPS are responsible for generating complete, accurate, and valid sample frame data files each month that contain all administrative information on all patients who meet the eligible population criteria. The following steps must be followed when creating the sample frame:
The sample frame for a particular month must include all eligible hospital discharges between the first and last days of the month (e.g., for January, any qualifying discharges between the 1st and 31st)
If a hospital is conducting sampling at the end of each month, they must create the sample frame in a timely manner in order to initiate contact for all sampled patients within 42 days of discharge
Patients with missing or incomplete addresses and/or telephone numbers must not be removed from the sample frame. Instead, every attempt must be made to find the correct address and/or telephone number.
Patients whose eligibility status is uncertain must be included in the sample frame
The hospital/survey vendor must retain the sample frame (i.e., the entire list of eligible HCAHPS patients from which each hospital’s sample is pulled) for 3 years. Confidentiality note: Patient-identifying information within the sample frame will not be part of the final data submitted to CMS, nor will any other PHI.
Hospitals must submit at least 300 completed HCAHPS Surveys in a rolling four-quarter period (unless the hospital is too small to obtain 300 completed surveys). The absence of a sufficient number of HCAHPS eligible discharges is the only acceptable reason for submitting fewer than 300 completed HCAHPS Surveys in a rolling four-quarter period. In that not all sampled patients who are contacted to complete the survey will actually do so, guidance is provided hospitals/survey vendors as to how many discharges are needed to reach the required 300 completed surveys per four rolling quarters of data (a 12- month reporting period).
Statistical Methodology for Stratification and Sample Selection.
The basic sampling procedure for HCAHPS entails drawing a random sample of all eligible discharges from a hospital on a monthly basis. Sampling may be conducted either continuously throughout the month or at the end of the month, as long as a random sample is generated from the entire month. The HCAHPS sample must be drawn according to this uninterrupted random sampling protocol and not according to any “quota” system. Hospitals/Survey vendors must sample from every month throughout the entire 12-month reporting period and not stop sampling or curtail ongoing survey administration activities even if 300 completed surveys have been attained.
Sampling for HCAHPS is based on the eligible discharges (HCAHPS sample frame) for a calendar month. If every eligible discharge for a given month has the same probability of being sampled, then an equiprobable approach is being used. Stratified sampling is where eligible discharges are divided into non-overlapping subgroups, referred to as strata, before sampling.
There are three options for sampling patients for the HCAHPS Survey: Simple Random Sampling (SRS), Proportionate Stratified Random Sampling (PSRS), and Disproportionate Stratified Random Sampling
(DSRS). Once a sample type is used within a quarter, it must be maintained throughout that quarter; “Sample Type” can only be changed at the beginning of a quarter. For more information about HCAHPS sampling, please see HCAHPS Quality Assurance Guidelines, V13.0, pp. 55-78, at www.hcahpsonline.org/en/quality-assurance/.
SRS: Simple Random Sampling is the most basic sampling type; patients are randomly selected from all eligible discharges for a month. Strata are not used when employing SRS and each patient has equal opportunity of being selected into the sample, making SRS equiprobable. Census sampling is considered a form of simple random sampling.
PSRS: Proportionate Stratified Random Sampling uses strata definitions and random sample selection from all strata at equal rates. Since the sampling rates of the strata are “proportionate,” PSRS is also considered equiprobable.
DSRS: Disproportionate Stratified Random Sampling involves sampling within strata at different rates, and thus, DSRS requires information about the strata. By definition, DSRS is not an equiprobable sampling approach as DSRS allows for dissimilar sampling rates across strata. DSRS means that all eligible discharges do not have an equal chance of being selected for inclusion in the monthly sample. To account for this, CMS requires additional information from hospitals and survey vendors who choose to use DSRS as a sampling type. Hospitals/survey vendors must submit an Exceptions Request Form and then be approved to use DSRS.
Estimation Procedures.
Not applicable to the HCAHPS Survey.
Degree of Accuracy Needed for the Purpose Described in the Justification.
IPPS hospitals are expected to achieve at least 300 completed surveys over a 12 month period, if possible, to attain the desired degree of accuracy; please see below. HCAHPS scores based on fewer than 100 or 50 completed surveys are publicly reported but the lower reliability of these scores is noted with an appropriate footnote in public reporting. HCAHPS scores based on fewer than 25 completed surveys are not reported on Hospital Compare. However, these hospitals do receive their HCAHPS scores in their confidential Preview Reports. IPPS hospitals must achieve at least 100 completed HCAHPS surveys during the 12 month Performance Period in order for a Patient and Caregiver Centered Experience of Care/Care Coordination domain score to be calculated in the Hospital Value-Based Purchasing program. The HCAHPS Quality Assurance Guidelines, V13.0, pp. 61-63, describe how to calculate the sample size necessary to attain at least 300 completed surveys, www.hcahpsonline.org/en/quality-assurance/. Elliot, et al. provide additional information about the validity and reliability of the HCAHPS Survey; see Elliott, Lehrman, et al. (2010). “Do Hospitals Rank Differently on HCAHPS for Different Patient Subgroups?” Medical Care Research and Review, 67(1):56- 73.
Unusual Problems Requiring Specialized Sampling Procedures.
CMS recognizes that some small hospitals may not have 300 HCAHPS-eligible discharges in the course of a year and so are unable to obtain 300 completed surveys in a 12-month report period. In such cases, IPPS hospitals must sample all eligible discharges (that is, conduct a census) and attempt to obtain as many completes as possible
Any Use of Periodic Data Collection Cycles to Reduce Burden.
There is no use of periodic (less frequent than annual) data collection cycles for the HCAHPS Survey. Great effort was expended considering how often HCAHPS data should be collected. We solicited and received much comment on this issue when the HCAHPS Survey was being developed. Two options for the frequency of data collection were suggested: once during the year or continuous sampling. The majority of hospitals/vendors suggested continuous sampling would be easier to integrate into their current data collection processes. Thus we decided to require sampling of discharges on a continuous basis (i.e., a monthly basis) and cumulate these samples to create rolling estimates based on 12-months of data. We chose to pursue the continuous sampling approach for the following reasons:
It is more easily integrated with many existing survey processes used for internal improvement,
Improvements in hospital care can be more quickly reflected in hospital scores (e.g., 12-month estimates could be updated on a quarterly or semi-annual basis),
Hospital scores are less susceptible to unique events that could affect hospital performance at a specific point in time,
It is less susceptible to gaming (e.g., hospitals being on their best behavior at the time of an annual survey), and
There is less variation in time between discharge and data collection.
Maximizing Response Rates/Non-response and Issues of Accuracy, Reliability, and Validity.
Implementation of the HCAHPS Survey according to the protocols contained in the current HCAHPS Quality Assurance Guidelines, V13.0, helps hospital to attain the greatest response rate. CMS examines hospital response rates every quarter and works with hospitals and survey vendors whose response rate is significantly below the national average. Among other tactics, CMS strongly encourages hospitals to offer the HCAHPS Survey in the language spoken at home by their patients.
Analysis of HCAHPS data indicates that the patient-mix adjustment applied to survey results adequately addresses the non-response bias that would exist without patient-mix adjustment; see “The Effects of Survey Mode, Patient Mix, and Nonresponse on CAHPS Hospital Survey Scores.” Elliott, Zaslavsky et al. (2009) Health Services Research, 44 (2): 501-518.
Information on statistical tests of the original 27-item HCAHPS Survey can be found in the original 2006 OMB package. Here we focus on the most recent tests of the validity and reliability of the current, 29-item HCAHPS Survey at the recommended 300 completed surveys in a 12-month public reporting period.
In terms of hospital-level reliability, the signal-to-noise ratio indicates how much of what you measure is “signal” (true variation in performance), rather than “noise” (measurement error). Unit or hospital-level reliability (Spearman-Brown reliability) is the proportion of variance in hospital-level scores that reflect true variation among hospitals, not noise due to limited numbers of patient surveys; see Table 1.
Table 1: Hospital-Level Spearman-Brown Reliabilities of HCAHPS Top-Box Scores of HCAHPS Measures at 300 Completed Surveys, 3.1 million discharges, January – December 2016.
Communication with Nurses |
0.87 |
Communication with Doctors |
0.86 |
Responsiveness of Hospital Staff |
0.92 |
Communication about Medicines |
0.83 |
Cleanliness |
0.87 |
Quietness |
0.93 |
Discharge Information |
0.83 |
Overall Rating |
0.91 |
Recommend Hospital |
0.93 |
Care Transition |
0.86 |
Generally accepted standards for reliability are: 0.7: Adequate (all HCAHPS measures exceed this); 0.8: Good (all exceed this); 0.85 Very good (8 of 10 reach this); and 0.9: Excellent (4 of 10 reach this level). It should be noted that at more than the recommended 300 completed surveys in 12 month reporting period, HCAHPS reliability is even higher.
Evidence of the internal consistency of the HCAHPS Survey can be found in Table 2, which displays the Cronbach’s Alpha statistics for the six publicly reported HCAHPS composite measures, which are made up of two or three survey items.
Communication with Nurses |
0.87 |
Communication with Doctors |
0.88 |
Responsiveness of Hospital Staff |
0.73 |
Communication about Medicines |
0.69 |
Discharge Information |
0.51 |
Care Transition |
0.811 |
Criterion validity indicates the extent to which HCAHPS measures what it purports to measure: patient experience of care. The patient-level correlation of specific HCAHPS measures with “Hospital Rating” and “Recommend the Hospital” can be used to assess validity (criterion validity); see Table 3. We expect moderate, positive correlations of the other eight HCAHPS measures with patient experience of care constructs. If the correlations were too high, this may indicate redundancy or halo effect. If too low, would indicate lack of validity. We find moderate, positive correlations between the eight specific HCAHPS measures and the two global measures (Rating and Recommendation). All 16 of these correlations are between 0.28 and 0.64; 13 of 16 correlations equal or exceed 0.35.
|
Communication with Nurses |
Communication with Doctors |
Responsiveness of Hosp. Staff |
Comm. About Medicines |
Cleanliness of Hospital Env. |
Quietness of Hospital Env. |
Discharge Information |
Care Transition |
Hospital Rating |
Recommend the Hospital |
Communication with Nurses |
1 |
0.53 |
0.56 |
0.50 |
0.39 |
0.32 |
0.27 |
0.43 |
0.64 |
0.58 |
Communication with Doctors |
|
1 |
0.38 |
0.44 |
0.27 |
0.26 |
0.28 |
0.40 |
0.52 |
0.47 |
Responsiveness of Hosp. Staff |
|
|
1 |
0.41 |
0.35 |
0.32 |
0.20 |
0.35 |
0.51 |
0.45 |
Comm. About Medicines |
|
|
|
1 |
0.33 |
0.29 |
0.35 |
0.45 |
0.48 |
0.43 |
Cleanliness of Hospital Env. |
|
|
|
|
1 |
0.28 |
0.18 |
0.27 |
0.41 |
0.36 |
Quietness of Hospital Env. |
|
|
|
|
|
1 |
0.13 |
0.25 |
0.35 |
0.29 |
Discharge Information |
|
|
|
|
|
|
1 |
0.30 |
0.30 |
0.28 |
Care Transition |
|
|
|
|
|
|
|
1 |
0.47 |
0.45 |
Hospital Rating |
|
|
|
|
|
|
|
|
1 |
0.76 |
Recommend the Hospital |
|
|
|
|
|
|
|
|
|
1 |
The correlations of the eight specific measures with the two global measures are larger at the hospital level than at the patient level, in part because of the better measurement at the hospital level (as intended) than at the patient level; see Table 4. These hospital-level correlations range from 0.40 to 0.73, with all but four equal to or exceeding 0.50.
|
Nurse |
Doctor |
Staff |
RX Comm |
Clean |
Quiet |
Discharge |
CTM |
Recommend |
Rating |
Nurse |
1 |
|
|
|
|
|
|
|
|
|
Doctor |
0.75 |
1 |
|
|
|
|
|
|
|
|
Staff |
0.78 |
0.66 |
1 |
|
|
|
|
|
|
|
RX Comm |
0.71 |
0.62 |
0.65 |
1 |
|
|
|
|
|
|
Clean |
0.68 |
0.56 |
0.68 |
0.55 |
1 |
|
|
|
|
|
Quiet |
0.57 |
0.58 |
0.60 |
0.52 |
0.52 |
1 |
|
|
|
|
Discharge |
0.48 |
0.36 |
0.38 |
0.41 |
0.33 |
0.18 |
1 |
|
|
|
CTM |
0.70 |
0.60 |
0.60 |
0.57 |
0.56 |
0.43 |
0.47 |
1 |
|
|
Recommend |
0.65 |
0.55 |
0.50 |
0.49 |
0.49 |
0.40 |
0.49 |
0.72 |
1 |
|
Rating |
0.72 |
0.61 |
0.63 |
0.58 |
0.58 |
0.51 |
0.51 |
0.73 |
0.86 |
1 |
The goal of HCAHPS is to collect information from patients using a standardized, national survey and to present the information based on those surveys to consumers, providers and hospitals. One of the methodological issues associated with making comparisons between hospitals is the need to adjust appropriately for patient-mix differences. Patient-mix refers to patient characteristics that are not under the control of the hospital that may affect measures of patient experiences, such as demographic characteristics and health status. The basic goal of adjusting for patient-mix is to estimate how different hospitals would be rated if they all provided care to comparable groups of patients.
CMS applies patient-mix adjustment to control for patient characteristics that affect ratings and that are differentially distributed across hospitals. Most of the patient-mix items are included in the “About You” section of the instrument, while others are taken from administrative records. Based on the mode experiment, and consistent with previous studies of patient-mix adjustment in CAHPS and in previous hospital patient surveys, we employ the following variables in the patient-mix adjustment model:
Self-reported general health status (specified as a linear variable)
Education (specified as a linear variable)
Type of service (medical, surgical, or maternity care) by gender
Age (specified as a categorical variable)
Lag time between discharge and survey
Age by service line interaction
Language (English, Spanish, Chinese, Russian/Vietnamese/Portuguese/Other) spoken at home
Once the data are adjusted for patient-mix, there is a fixed adjustment for each of the reported measures for mode of administration (discussed in detail below). The patient-mix adjustment employs a regression methodology, also referred to as covariance adjustment.
On the survey there are two items that capture the race and ethnicity of the respondent. These items are not included in the patient-mix adjustment model but are used as analytic variables to support the congressionally-mandated report: “The Quality Disparities Report.” This report provides annual, national-level breakdowns of HCAHPS scores by race and ethnicity. Many hospitals collect information on race and ethnicity through their administrative systems, but coding is not standard. Thus it was determined that administrative data are not adequate to support the analyses needed for the reports and the items should be included in the questionnaire.
Past and Ongoing Tests of Procedures, Trainings, and Quality Improvement Activities; Collaborator and Contractor Participation (collaborators and contractors are bolded).
Hospitals began using the HCAHPS Survey in 2006 under the auspices of the Hospital Quality Alliance, a private/public partnership that includes CMS and the American Hospital Association, the Federation of American Hospitals, and the Association of American Medical Colleges; the Joint Commission on Accreditation of Healthcare Organizations; the National Quality Forum; AARP; and CMS’ Agency for Healthcare Research and Quality (AHRQ).
Beginning in 2002, HCAHPS has undertaken extensive research, planning, development and testing activities. CMS originally partnered with AHRQ to develop a standard instrument and data collection and reporting procedures that capture patients’ perspectives of their hospital care. AHRQ is a leader in developing instruments for measuring patient perspectives on health care. Table 5 summarizes the process of creating and implementing the HCAHPS survey throughout the period starting July 2002 through the first public reporting of HCAHPS results in March of 2008.
Activity |
Timeframe |
Published a “call for measures” in the Federal Register and received 7 submissions from Avatar, Edge Health Care Research Healthcare Financial Management Association, Jackson Organization, Press Ganey Associates, National Research Corporation, Peace Health, Professional Research Consultants, and SSM Health Care. |
July 2002 |
Completed literature review. |
Sept-Nov 2002 |
Held a Web chat to answer questions about HCAHPS. |
Oct 2002 |
Provided draft domains to CMS. |
Oct 2002 |
Reviewed measures submitted in response to Federal Register Notice (FRN). |
Nov 2002 |
Held Stakeholders Meeting to solicit suggestions and comments. |
Nov 2002 |
Held vendors meeting to solicit suggestions and comments. |
Nov 2002 |
AHRQ delivered 66 item draft survey to CMS for use in pilot test. |
Jan 2003 |
Developed data collection and sampling methods, and developed analysis plan. |
Jan-Feb 2003 |
Published FRN soliciting comments on draft HCAHPS. |
Feb 2003 |
Completed hospital recruitment for pilot. |
Mar 2003 |
Began data collection for CMS 3-state pilot test. |
June 2003 |
Published a FRN soliciting comments on draft HCAHPS and asked for input about implementation issues. |
June 2003 |
Analyzed data from CMS pilot test. |
Sept–Nov 2003 |
Review of instrument by CAHPS Cultural Comparability team. |
Fall 2003 |
Began CT pilot test of the HCAHPS instrument. |
Fall 2003 |
Held HCAHPS Stakeholders’ Meeting at AHRQ. |
Nov 2003 |
Revised HCAHPS instrument to 32 items. |
Nov 2003 |
AHRQ submitted revised 32-items HCAHPS Instrument to CMS. |
Dec 2003 |
Published a FRN soliciting input for 32-item HCAHPS instrument and implementation strategy. |
Dec 2003-Feb 2004 |
Started coordination of national implementation with HSAG, the AZ QIO. |
January 2004 |
Completed CT pilot test of HCAHPS. |
Jan 2004 |
AHRQ submitted Analysis Report of the CMS 3-state pilot to CMS. |
Jan 2004 |
Continued discussions with hospitals, vendors, consumers to follow-up on FRN comments from February. |
March – Sept 2004 |
Revised 25-item HCAHPS Instrument submitted by AHRQ to CMS. |
Oct 2004 |
Submitted HCAHPS to NQF for its consensus development process. |
November 2004 |
Started developing training documents for national implementation. |
December 2004 |
Started discussions regarding data transmission via QNET & other issues with the IFMC, the IA QIO. |
April 2005 |
Formed the Data Integrity Group. |
June 2005 |
Received endorsement for 27-item HCAHPS from the National Quality |
May 2005 |
Modified survey instrument and protocol as 27-items. |
May 2005 |
Abt Associates, Inc. receives OMB approval for cost-benefit analysis. |
June 2005 |
Established the Data Submission and Reporting Group. |
July 2005 |
Abt Associates, Inc. submits final report of the cost-benefit analysis. |
October 2005 |
Published FRN soliciting comments on draft CAHPS Hospital Survey. |
November 2005 |
Received final approval. |
December 2005 |
Mode Experiment. |
Feb – May 2006 |
National Implementation begins. |
October 2006 |
HCAHPS participation linked to RHQDAPU program (“pay for reporting”). |
July 2007 |
First public reporting of HCAHPS results. |
March 2008 |
Throughout the HCAHPS development process, CMS solicited and received a great deal of public input. As a result, the HCAHPS questionnaire and methodology went through several iterations prior to national implementation. The accumulated lessons learned from the pilot testing, public comments, input from stakeholders, numerous team discussions, and the National Quality Forum’s review and endorsement (NQF #0166) through their consensus development process led to the national implementation in 2006 of the 27-item HCAHPS Survey and the HCAHPS data collection protocol that allows hospitals to integrate their own supplemental questions. The resulting core questionnaire is comprised of questions in several dimensions of primary importance to the target audience: doctor communication, responsiveness of hospital staff, cleanliness of the hospital environment, quietness of the hospital environment, nurse communication, pain management, communication about medicines, and discharge information. The HCAHPS Survey was re-endorsed by the NQF in 2012 and 2015.
The HCAHPS implementation plan changed significantly as a result of the public input we received prior to national implementation. CMS made the following major changes in the implementation approach, which have served to lessen the burden on hospitals/survey vendors:
reduced the number of mailings for the “mail only” survey protocol from three to two;
reduced the number of follow-up phone calls for the “telephone only” survey protocol from ten to five;
added active interactive voice response (IVR) as a mode of survey administration;
eliminated the 50% response rate requirement;
reduced the number of patient discharges to be surveyed.
Since national implementation began in 2006, CMS has continually refined and clarified HCAHPS survey protocols, created new translations of the mail version of the survey in Chinese, Russian, Vietnamese, and Portuguese in addition to the original English and Spanish versions (in the telephone and IVR modes, the HCAHPS Survey is available in both English and Spanish), added five new survey items, including three items that form the new composite measure, Care Transition, annually updated the HCAHPS Quality Assurance Guidelines (currently version 13.0), improved the appearance and accessibility of HCAHPS results on the Hospital Compare website: https://www.medicare.gov/hospitalcompare/, and made information about HCAHPS quickly and easily available through its official HCAHPS On-Line website, www.hcahpsonline.org. There has been steady and significant improvement in HCAHPS scores since national implementation; see Elliott, Lehrman, Goldstein et al. (2010), “Hospital Survey Shows Improvements in Patient Experience.” Health Affairs, 29 (11): 2061-2067. The HCAHPS Project Team has published a number of analyses of HCAHPS results in peer-reviewed scientific journals and made numerous presentations at professional conferences; a bibliography of the publications can be found at: http://www.hcahpsonline.org/globalassets/hcahps/home/october_2016_home_bibliography.pdf.
Since public reporting of hospitals HCAHPS results was inaugurated in March 2008, a growing number of healthcare, consumer and professional organizations, state governments, media outlets and others have adopted or incorporated HCAHPS scores, in part or in whole, for their own purposes. These activities, external to CMS, have had the effect of extending knowledge about HCAHPS and increasing the impact of survey results. The content of the HCAHPS Survey, its methodology and administration protocols, and its ambition to measure and publicly report consumers’ experiences in a uniform and standardized manner have influenced other surveys developed within CMS as well as those undertaken by hospitals and healthcare systems in the United States and abroad.
There are distinct roles for hospitals or their survey vendors and the federal government in the national implementation of HCAHPS. The federal government is responsible for support and public reporting, including:
conducting training on data collection and submission procedures,
providing on-going technical assistance,
ensuring the integrity of data collection,
accumulating HCAHPS data from individual hospitals,
producing patient-mix adjusted hospital-level estimates,
conducting research on the presentation of data for public reporting, and
publicly reporting the comparative hospital data.
Hospitals or their survey vendors are responsible for data collection, including: developing a sampling frame of relevant discharges, drawing the sample of discharges to be surveyed, collecting survey data from sampled discharges, and submitting HCAHPS data to CMS in a standard format. We have formatted the data files so hospitals/vendors will submit to CMS de-identified data files following 45 CFR Section §164.514. Hospitals maintain business associate agreements with their contracted survey vendors to collect and submit HCAHPS survey data through the secure QualityNet Exchange portal and data warehouse.
CMS began its collaboration with the Health Services Advisory Group (HSAG) in 2003 to coordinate the national implementation of the Hospital CAHPS Survey. HSAG’s role is to provide technical assistance and training for vendors and hospitals, data validation, data processing, analysis, and adjustment, and oversight of self-administering hospitals and survey vendors. HSAG also produces electronic data files and a hospital level extract file for public reporting of the HCAHPS scores.
In the spring of 2006, CMS conducted a large-scale experiment to assess the impact of mode of survey administration, patient characteristics and patient non-response on HCAHPS results. This first mode experiment was based on a nationwide random sample of short-term acute care hospitals. Hospitals from each of CMS’ ten geographic regions participated in in the Mode Experiment. A hospital's probability of being selected for the sample was proportional to its volume of discharges, which guaranteed that each patient would have an equal probability of being sampled for the experiment. The participating hospitals contributed patient discharges from a four-month period: February, March, April, and May 2006. Within each hospital, an equal number of patients was randomly assigned to each of the four modes of survey administration. Sample selection and surveying were conducted by the National Opinion Research Center of the University of Chicago, and the data was analyzed by the RAND Corporation.
A randomized mode experiment of 27,229 discharges from 45 hospitals was used to develop adjustments for the effects of survey mode (Mail Only, Telephone Only, Mixed mode, or Active Interactive Voice Response) on responses to the HCAHPS Survey. In general, patients randomized to the Telephone Only and Active Interactive Voice Response provided more positive evaluations than patients randomized to Mail Only and Mixed (Mail with Telephone follow-up) modes. These mode effects varied little by hospital, and were strongest for global items (rating and recommendation), and the Cleanliness & Quiet, Responsiveness, Pain Management, and Discharge Information composites. Adjustments for these mode effects are necessary to make the reported scores independent of the survey mode that was used. These adjustments are applied to HCAHPS results before they are publicly reported on the Hospital Compare website. The mode adjustments can be found in the “Mode Adjustment” section of the HCAHPS website, http://www.hcahpsonline.org.
The Mode Experiment also provided valuable information on the impact of salient patient characteristics and non-response bias on HCAHPS results. This analysis was needed because hospitals do not provide care for comparable groups of patients but, as demonstrated in the HCAHPS Three-State Pilot Study, some patient characteristics may affect measures of patient experiences of care. The goal of patient-mix adjustment, which is also known as case-mix adjustment, is to estimate how different hospitals would be rated if they provided care to comparable groups of patients. As suggested by the Three-State Pilot Study, a set of patient characteristics not under control of the hospital was selected for analysis. In summary, the most important patient-mix adjustment items were patients’ self-reported health status, education, service line (maternity, medical or surgical care) and age. In addition, after mode and patient-mix adjustments have been made, non-response effects were found to be negligible. A report on patient-mix and non- response adjustments, as well as the mode adjustments and current and past patient-mix adjustments, is available on the HCAHPS On-Line website, http://www.hcahpsonline.org/en/mode--patient-mix-adj/.
We also looked at the extent to which each domain contributes to measurement in priority areas established by an independent, expert body on quality measurement, the National Quality Forum (NQF). The HCAHPS domains “communication with doctors,” “communication with nurses,” “communication about medications” contribute to the NQF’s priority on improving care coordination and communication. The HCAHPS “discharge information” domain contributes to the priority on improving self-management and health literacy.
CMS, with assistance from HSAG and its sub-contractor, the National Committee on Quality Assurance (NCQA), has developed and conducted annual training sessions for self-administering hospitals and survey vendors participating in the HCAHPS Survey, as well as others interested in this program. HCAHPS Introductory Training was first offered at the CMS headquarters in January 2006, and then by webinar in January and April 2006. Since then, HCAHPS Introductory Training has been held annually, by webinar. In addition, CMS developed the HCAHPS Update Training program. HCAHPS Update Training was first offered in May 2007 and has been offered annually by webinar since then.
HCAHPS Introductory Training is required for self-administering hospitals and survey vendors that wish to join HCAHPS. HCAHPS Update Training provides information on important changes to the HCAHPS program and is required for all self-administering hospitals and survey vendors participating in HCAHPS.
As part of an initiative to add five-star quality ratings to its Compare Web sites, CMS added HCAHPS Star Ratings to the Hospital Compare Web site in April 2015. Star ratings make it easier for consumers to use the information on the Compare Web sites and spotlight excellence in healthcare quality. Eleven HCAHPS Star Ratings appear on Hospital Compare: one for each of the 10 publicly reported HCAHPS measures, plus the HCAHPS Summary Star Rating. HCAHPS Star Ratings were the first star ratings to appear on Hospital Compare. CMS updates the HCAHPS Star Ratings each quarter. In support of the addition of HCAHPS Star Ratings on Hospital Compare, we created a special section of our HCAHPS On-Line Web site on which we post information and documents, as well as archives, including HCAHPS Star Rating Technical Notes that: explain the calculation of the star ratings; distribution of HCAHPS Star Ratings, HCAHPS Summary Star Rating distributions by state, slides from a national provider call on HCAHPS Star Ratings, and frequently asked questions; see http://www.hcahpsonline.org/en/hcahps-star-ratings/ .
Marc N. Elliott, PhD
Distinguished Chair in Statistics; Senior Principal Researcher RAND Corporation
(310) 393-0411, x7931
Christopher Cohea, MS
Director, Research and Analysis
Health Services Advisory Group, Inc. (602) 471-3673
File Type | application/vnd.openxmlformats-officedocument.wordprocessingml.document |
File Modified | 0000-00-00 |
File Created | 0000-00-00 |