Supporting Statement – Part B
Collections of Information Employing Statistical Methods
1. Describe (including a numerical estimate) the potential respondent universe and any sampling or other respondent selection method to be used. Data on the number of entities (e.g., establishments, State and local government units, households, or persons) in the universe covered by the collection and in the corresponding sample are to be provided in tabular form for the universe as a whole and for each of the strata in the proposed sample. Indicate expected response rates for the collection as a whole. If the collection had been conducted previously, include the actual response rate achieved during the last collection.
The Data Validation Contractor will randomly sample 300 facilities (5% of all dialysis facilities), per contract and Quality Incentive Program (QIP) rule guidelines, for participation in the validation project. As a random sample, this should be a representative sample of all included facilities nationally. The sample pool will consist of Medicare-certified dialysis facilities that are required to submit administrative and clinical data into CROWNWeb in order to meet Section 494.108(h) of the 2008 updated Conditions for Coverage for ESRD Dialysis Facilities. The 300 facilities will be asked to submit records that will be validated for CMS-designated Critical Performance Measures (CPMs). The patient sample size is limited to 10 patients per facility, as per contract and QIP rule guidelines. The Data Validation Contractor will sample 10 patients (or the maximum patients possible) from each selected facility for CPM reviews. Historically, facility response rates have been solid. The response rate for the 2016 validation study was 100%; of the 300 facilities selected for participation, all eligible participating facilities responded and complied with our records request.
Using data from previous CROWNWeb validation work, we estimate the numbers of facilities per three strata by the numbers of patients reported per year. Using the ESRD QIP rule guidelines of 10 records per facility and no more than 300 facilities sampled, the Validation Contractor calculated the estimated sample size from each of three (3) strata of facilities. Table 1 below provides the sample size estimates for the validation effort where approximately five (5) percent (n=300) of all dialysis facilities will be sampled. We expect facility-specific error rates for the CPM validation to have relatively small confidence intervals as displayed in Table 2.
Table 1: Sample Size Estimates
Number of Patients per Quarter |
% Total Facilities |
% Total Patients |
Number of Facilities |
Number of Records to Sample per Facility |
Number of Records |
0 to 35 |
17.28% |
4.66% |
52 |
10 |
520 |
26 to 95 |
48.03% |
36.53% |
144 |
10 |
1,440 |
>95 |
34.69% |
58.81% |
104 |
10 |
1,040 |
Total |
100.00% |
100.00% |
300 |
|
3,000 |
Table 2: Confidence Interval for CROWNWeb Data
Agreement Rate |
Sample Size |
95% Confidence Interval |
90% |
5% (300) |
±3.4 |
80% |
5% (300) |
±4.5 |
70% |
5% (300) |
±5.2 |
Some smaller facilities may have less than 10 patients treated for the period; in these cases, we will select all of the patients treated at the facility during the study period for validation. We will randomly sample 10 medical records without replacement per facility. Table 3 depicts the methodology to be used when sampling for patients for CPM reviews.
Table 3: Sampling Methodology for CPM Reviews
Sampling Source |
Sample to be Taken |
CROWNWeb Extract |
Maximum number of patients available, up to 10 |
The 300 facilities to be sampled for validation will be chosen within 10 days of receiving the corresponding Facility/Patient data file from CROWNWeb. The Validation Contractor will receive a CROWNWeb extract that contains all data reported into CROWNWeb during the selected second quarter time frame (April – June 2018).
The Validation Contractor intends on validating records for the second quarter of 2018. This timeframe was selected after considering several factors. To ensure that the validation can be completed during the period of performance, the Validation Contractor considered the data reporting periods allowed to facilities to submit clinical data into CROWNWeb. Facilities are given 60 days from the end of any particular month to enter CROWNWeb clinical data. The mandated reporting period limits the time frame we can validate expeditiously, as we will not be able to obtain an extract until after the close of the data-reporting period. Another important consideration is the ESRD QIP rule that makes it mandatory for us to give facilities up to 60 days to submit records. Taking into consideration these factors as well as the need to ensure that there is adequate time to perform analysis and prepare reports, we decided on the second quarter of 2018 validation time frame. A breakdown of the mandated reported deadlines that were taken into consideration is displayed in Table 4.
Table 4: Mandated Reporting Deadlines
Submission Type |
Mandated Reporting Deadlines |
CROWNWeb Data Submission |
60 days after month close (Q2 – August 31, 2018) |
Facility Record Submission Deadline |
60 days after request receipt per QIP rule |
Assuming the CROWNWeb data team will need at least one (1) week to export and send the data, the Validation Contractor has estimated preliminary dates for data availability. Table 5 provides the Validation Contractor’s estimates for when the data will be received for each corresponding data set.
Table 5: Mandated Reporting Deadlines
Type of Data |
Data Reporting Period |
Estimated Receive Date |
CPM |
April, May, June 2017 |
Starting mid-November through beginning of December |
Due to the tight timeframe for data abstraction, effective coordination and management as well as adherence to established schedules will be crucial to the success of the project.
2. Describe the procedures for the collection of information including:
- Statistical methodology for stratification and sample selection,
- Estimation procedure,
- Degree of accuracy needed for the purpose described in the justification,
- Unusual problems requiring specialized sampling procedures, and
- Any use of periodic (less frequent than annual) data collection cycles to reduce burden.
Please see response to question 1 for statistical methodology for stratification and sample selection, including estimation procedure and degree of accuracy needed for the purpose of this work. As noted below in response to question 4, there are no unusual problems requiring specialized sampling procedures as our previous experience on past CMS CROWNWeb CPM validation efforts have shown near universal compliance by the hospitals with medical record requests. The period for data collection cycles is expected to be no more frequently than annually.
3. Describe methods to maximize response rates and to deal with issues of non-response. The accuracy and reliability of information collected must be shown to be adequate for intended uses. For collections based on sampling, a special justification must be provided for any collection that will not yield 'reliable' data that can be generalized to the universe studied.
As part of our previous work to collect medical records for the 2016 CROWNWeb CPM data validation, facilities were contacted via certified letter in January 2017 and were asked to participate in the validation effort. The letter provided instructions on the types of records to be submitted, methods to submit records to the Validation Contractor, and identified patients selected for validation. Facilities that did not respond to the initial request for records were contacted via phone by the Validation Contractor and received a final request letter in March 2017. To aid in maximizing facility response rates, our team presented our project details during multiple ESRD community forums (Town Halls, NICE Calls, etc.) to increase facility exposure to our validation study. We also communicated/coordinated extensively with LDOs to facilitate on-time medical records submission by their participating clinics. Facilities that did not respond to the request for records were subject to a 10-point reduction to their Total Performance Score (TPS). The response rate for the 2016 validation study was 100%; of the 300 facilities selected for participation, all eligible participating facilities responded and complied with our records request. For future validations, we plan to follow the same records request methodology, follow-up, and ESRD community outreach approach we’ve used in the past since it has been effective in producing desired response rates.
The main objective of this analysis is to perform a single comparison of the CROWNWeb system data against CPM element data obtained from the facilities’ records, leading to an evaluation of the reliability (i.e., the data are reasonably complete and accurate) and validity (i.e., the data actually represent what is being measured) of CROWNWeb data.
Completeness refers to the extent that relevant records are present and the fields in each record are populated appropriately, and,
Accuracy refers to the extent recorded data reflect the actual underlying information. In this instance, that the data in the CROWNWeb system accurately reflects the data contained within the source documents, i.e. the facilities’ medical records.
Validity: (as used here) refers to whether the data actually represent what one believes is being measured. A number of measures are commonly used to assess validity of any measure.
In our interpretation of these measures, we identify the key sources of overall disagreement between the CROWNWeb data and the patients’ medical records, which would serve as the “gold standard.” Typical sources of disagreement include missing information about events, inaccurate dates and inaccurate dialysis catheter information.
We propose to use Cohen’s Kappa () because it is an overall measure of agreement between the test and reference databases. The kappa coefficient, κ, is calculated as κ = (po – pe) * (1 – pe), where po is the observed agreement between two classifications and pe is the expected agreement between two classifications based on the marginal distributions.
Cohen’s Kappa is usually presented as a proportion; we will convert key measures of agreement from proportions to percentages for easier interpretation. All measures will range from -100 % (perfect disagreement) to +100% (perfect agreement). Landis and Koch (1977) have suggested overall agreement between two classifications using Cohen’s Kappa be interpreted as shown in Table 6.
Kappa () |
Interpretation |
< 0% |
Poor |
0% to 20% |
Slight |
20% to 40% |
Fair |
40% to 60% |
Moderate |
60% to 80% |
Substantial |
80% to 100% |
Almost Perfect |
Although the Landis and Koch interpretation is an old one, it is still widely referenced and is still the dominant one used to indicate the strength of agreement (Cunningham, 2009).
Other coefficients of agreement have been suggested but as Fleiss pointed out (1981), has a number of qualities that has made it an attractive option for the measurement of agreement:
If there is complete agreement, = +1,
If observed agreement is greater than or equal to chance agreement, ≥ 0, and
If the observed agreement is less than or equal to chance agreement, < 0.
However, over the years a number of difficulties in the interpretation of Cohen’s Kappa have been pointed out and several statistical fixes have been proposed. Kappa not only measures agreement, but it is affected in complex ways by the distribution of data across categories that are used (“prevalence”) and by bias that may be inherent in the measures used. These are the problems associated with Kappa (Feinstein and Cicchetti, 1990):
If the expected agreement (pe) is large, the correction process can convert a relatively high value of the observed agreement (po) into a low value of Kappa ().
Unbalanced marginal totals produce higher values of than balanced totals.
Kappa is also affected both by any bias between the two measures of gender and by the overall prevalence (the relative probability of the responses – the “Yes” and “No” responses). Byrt, Bishot, and Carlin (1993) introduced measures of prevalence (prevalence index – pindex) and bias (bias index – bindex) that can be used to compensate for the biases and suggest that these measures are reported together with . The pindex and bindex can then be used to produce a prevalence-adjusted, bias-adjusted kappa (PABAK) that takes on the values of -100% when there is no agreement and +100% when there is perfect agreement and 0 when the agreement is equal to 50%. Additionally, PABAK is linearly related to po.
Given the following standard table with two measures and a dichotomous response, we can therefore add the following measures of agreement shown in Table 7 below (Byrt, Bishot and Carlin, 1993; Cunningham, 2009).
Table 7: Standard Fourfold Table
Measurement A |
Measurement B, Yes |
Measurement B, No |
Total |
Yes |
a |
b |
a + b |
No |
c |
d |
c + d |
Total |
a +c |
b +d |
N |
The observed proportion of agreement, po = (a+d) N
Expected proportion of agreement, pe = ((a + c)(a + b)+(b + d)(c + d)) N2
Proportion of positive agreement, ppos = (2a) (N + a – d)
Proportion of negative agreement, pneg = (2d) (N – a + d)
Prevalence Index, pindex = (a – d) N
Bias Index, bindex = (b – c) N
Prevalence-adjusted, bias-adjusted Kappa, PABAK = 2po – 1
4. Describe any tests of procedures or methods to be undertaken. Testing is encouraged as an effective means of refining collections of information to minimize burden and improve utility. Tests must be approved if they call for answers to identical questions from 10 or more respondents. A proposed test or set of tests may be submitted for approval separately or in combination with the main collection of information.
As noted above, the sample pool will consist of Medicare-certified dialysis facilities that are required to submit administrative and clinical data into CROWNWeb in order to meet Section 494.108(h) of the 2008 updated Conditions for Coverage for ESRD Dialysis Facilities. Our previous experience on past CMS CROWNWeb validation efforts have shown near universal compliance with medical record requests. No additional tests of procedures or methods to be undertaken are expected.
5. Provide the name and telephone number of individuals consulted on statistical aspects of the design and the name of the agency unit, contractor(s), grantee(s), or other person(s) who will actually collect and/or analyze the information for the agency.
Jon Mark Hirshon, University of Maryland-Baltimore, 1 (410) 328-7474
Joseph Kufera, University of Maryland-Baltimore, 1 (410) 328-4161
Darren Childers, Allegheny Science & Technology, 1 (304) 612-1571
Heather Duvall, Healthcare Management Solutions, 1 (304) 816-5262
File Type | application/msword |
File Title | Supporting Statement – Part B |
Author | CMS |
Last Modified By | SYSTEM |
File Modified | 2018-01-12 |
File Created | 2018-01-12 |