NAGB OMB Section B (Revised 1-10-11)

NAGB OMB Section B (Revised 1-10-11).docx

Evaluating Student Need for Developmental or Remedial Courses at Postsecondary Education Institutions (Formerly titled Survey of Placement Tests and Cut-Scores in Higher Education Institutions).

OMB: 3098-0006

Document [docx]
Download: docx | pdf

Request for OMB Clearance

Evaluating Student Need for Developmental or Remedial Courses

at Postsecondary Education Institutions

(Formerly titled Survey of Placement Tests and Cut-Scores in Higher Education Institutions).



Supporting Statement for Paperwork Reduction Act Submissions


REVISED 1-10-11



Supporting Statement B: Statistical Methodology for Pilot Test and Full-Scale Survey



B.1. Respondent Universe for Pilot Study and Full-Scale Survey

As with the pilot test, the sampling frame or respondent universe from which the sample of institutions for the full-scale survey will be drawn, will be constructed from the 2008-09 Institutional Characteristics (IC) component of the Integrated Postsecondary Education data System (IPEDS) maintained by the National Center for Education Statistics (NCES). The 2008-09 IC file contains over 7,100 Title IV postsecondary institutions. However, only 2-year and 4-year degree granting institutions located in the United States with a bachelors’ program or lower are eligible for the study. Thus, post graduate institutions offering only a first professional degree will be excluded from the sampling frame. Table B-1 summarizes how the frame of institutions will be constructed. As indicated in the last row of the table, 4,148 institutions meeting the eligibility criteria for the study will be included in the sampling frame, of which 2,458 will be 4-year institutions and 1,690 will be 2-year institutions.


Table B-1. Summary of frame construction for pilot test and full-scale survey







Description of institutions in 2008-09 IPEDS IC File

No. of institutions

4-year

2-year

Other






All institutions

7,126

2,906

2,299

1,921

Institutions in United States (excluding outlying territories)

6,961

2,837

2,274

1,850

Institutions in United States offering undergraduate classes

6,649

2,541

2,274

1,834

Non-degree granting

2,377

6

537

1,834

Degree granting

4,272

2,535

1,737

n/a

Non Title IV

45

32

13

n/a

Title IV*

4,227

2,503

1,724

n/a

Administrative unit†

79

45

34

n/a

Non-administrative unit **

4,148

2,458

1,690

n/a

 

 

 

 

 






* Includes 4 US Service academies.


Pending review, all or most of these will be deleted from the sampling frame.


** Count includes 52 cases with missing enrollment data. Unless deemed ineligible, all will be included in the final sampling frame.



Table B-2 summarizes the distribution of eligible institutions by level and control.


Table B-2. Distribution of eligible postsecondary institutions to be included in the sampling frame, by level and control


Level

Control

Number of institutions in sampling frame*

4-year

Public

636


Private, not-for-profit

1,311


Private, for-profit

472

2-year

Public

1,022


Private, not-for-profit

92


Private, for-profit

563

Total institutions (with non missing enrollment)

4,096

* The counts given in this table are based on initial tabulations from the 2008-09 IPEDS Institutional Characteristics (IC) file. Fifty-two eligible institutions in the 2008-09 IC file have missing values for enrollment and are not included in this table. However, such institutions will be included in the final sampling frame and given appropriate chances of selection.





B.2. Statistical Methodology



Sample Design and Selection for the Pilot Test


The pilot test was conducted in fall 2010. Since a primary purpose of the pilot test was to identify potential data collection problems during the full-scale survey, the goal of the sample design was to ensure that (a) a diverse cross-section of postsecondary instititions in the United States were included in the sample, and (b) the sample sizes for major subgroups defined by type of control and level were sufficiently large to provide a range of responses that can usefully inform questionnaire design for the full-scale survey. Thus, in addition to institutional control and level, the sample was stratified by characteristics such as size of institution and highest level of degree offering. Within a sampling stratum, institutions was selected at rates that are approximately proportional to the square root of their enrollment. The sample size of 120 institutions for the pilot test was designed to yield about 100 respondents assuming a response rate of 85 percent. Although measures of selectivity was not be used as a stratifier in sample selection, the sample of 4-year institutions to be included in the pilot test was expected to cover a broad range of institutions with respect to selectivity. Table B-3 summarizes the proposed sample sizes for the pilot test by major subgroups defined by institutional level and control.


Table B-3. Target sample sizes for the pilot test, by control and level


Subgroup

Number sampled

Respondents*

Total sample

120

102

Public

60

51

4-year

30

26

2-year

30

26

Private, not-for-profit

40

34

4-year

30

26

2-year

10

9

Private, for-profit

20

17

4-year

10

9

2-year

10

9

* Assumes 85 percent response rate.

Note: Details may not sum to total due to rounding.



Sample Design and Selection for the Full-Scale Survey



Sample Allocation for the Full-Scale Survey


One of the goals of the sample design is to ensure that the sample sizes are sufficiently large to enable the detection of significant differences of 0.2 standard deviations in assessment score scales between major subgroups to the extent feasible. Table B-4 summarizes the minimum detectable effect size (i.e., the difference expressed as a multiple of the standard deviation) between subgroups of various sizes and a range of design effects. The design effects will arise primarily because of the varying probabilities of selection with which the institutions will be drawn into the sample. Depending on the magnitude of the design effect, it can be seen that each subgroup would have to include between 400-500 respondents to be able to detect an effect size of 0.20 (i.e., 0.20 standard deviations). Larger effect sizes of 0.40 can be detected with subgroups sample sizes of 100-200. Only very large effect sizes of 0.60 or greater can be detected with subgroup sample sizes as small as 50 to 60.


Table B-4. Minimum detectable effect sizes* between subgroups for selected sample sizes and design effects



Design effect

Subgroup sample size

1.10

1.20

1.30

800

0.15

0.15

0.16

700

0.16

0.16

0.17

600

0.17

0.18

0.18

500

0.19

0.19

0.20

400

0.21

0.22

0.23

300

0.24

0.25

0.26

200

0.29

0.31

0.32

150

0.34

0.35

0.37

100

0.42

0.43

0.45

60

0.54

0.56

0.58

* Entries are detectable effect sizes for a two-sided test with alpha = 0.05 and power = 0.80.

Sample size per subgroup.

In addition to the specified precision goals, disaggregation of the sample into subgroups defined by control and level will also be important for analysis. However, this will not be possible for very small subgroups without unduly reducing the overall efficiency of the sample. For example, it can be seen in Table B-2 that most of the private 2-year institutions are for-profit institutions. Since the private 2-year institutions tend to be small (fewer than 500 students), it does not seem useful to separate the 2-year private for-profit and not-for-profit institutions from either a statistical or policy perspective. Among the private 4‑year institutions, on the other hand, there are several extremely large for-profit institutions (e.g., DeVry, University of Phoenix) that are inherently different from the traditional private not-for-profit institutions. In this case, separating the for-profit institutions from the others would be desirable to provide meaningful comparisons.

In view of the above considerations, we propose to select the numbers of institutions indicated in Table B-5. The proposed sample sizes assume an overall 85 percent response rate and are designed to permit separate analysis of most major subgroups, while giving appropriate representation in the sample of the rarer subgroups. For example, under the proposed design, it will be possible to detect an effect size of 0.20 for comparisons between public vs. private not-for-profit institutions, between public 4-year and public 2-year institutions, and between public 4-year and private not-for-profit 4-year institutions. Larger effect sizes of 0.40 can be detected between smaller subgroups; e.g., comparisons between private not-for-profit and private for-profit institutions or between private 4-year not-for-profit institutions and private 4-year for-profit institutions.



Table B-5. Proposed target sample sizes for the full-scale survey, by control and level


Subgroup

Number sampled

Respondents*

Total sample

1,668

1,418

Public

936

796

4-year

468

398

2-year

468

398

Private, not-for-profit

490

417

4-year

468

398

2-year

22

19

Private, for-profit

242

206

4-year

160

136

2-year

82

70

* Assumes 85 percent response rate.

Note: Details may not sum to total due to rounding.



Sample Selection Procedures for the Full-Scale Survey


To permit the efficient allocation of the sample to meet the goals stated above, the sampling frame will be stratified by size of institution, level, and type-of-control, along with other institutional characteristics such as highest level of degree offering. Implicit stratification within major strata by geography and minority status of institution will also be employed to ensure that all regions and types of institutions are appropriately represented in the sample. Within a stratum we will sample institutions at rates that are approximately proportional to the square root of their enrollment. Table B-6 shows the distribution of the proposed sample by each of the 39 sampling strata using this method. A similar design has been used to select institutions for the Postsecondary Quick Information System (PEQIS) and is expected to be efficient for the proposed survey on the use of placement tests and cut scores. Note that institutions with enrollments of 10,000 or more are selected with certainty under this design.



Table B-6. Distribution of postsecondary institutions to be included in the sampling frame


Stratum

Level

Type of

Control

Highest

level of

offering

Enrollment

size class

Number of

institutions

in frame*

Number of institutions in sample

1

4-year

Public

Doctorate

<3,000

20

7

2




3,000 to 9,999

63

42

3




10,000 to 19,999

94

94

4




20,000+

106

106

5



Masters

<3,000

57

19

6




3,000 to 4,999

51

34

7




5,000 to 9,999

87

65

8




10,000+

42

42

9



Bachelors

<1,000

19

5

10




1,000 to 2,999

45

15

11




3,000 to 9,999

38

25

12




10,000+

14

14

13

4-year

Private

Doctorate

<500

47

5

14


(not-for-profit)


500 to 2,999

100

33

15




3,000 to 9,999

152

91

16




10,000+

53

53

17



Masters

<500

81

9

18




500 to 999

92

31

19




1,000 to 2,999

269

90

20




3,000 to 9,999

92

55

21




10,000+

8

8

22



Bachelors

<500

151

17

23




500 to 999

107

21

24




1,000 to 2,999

146

49

25




3,000+

13

7

26

4-year

Private

––

<100

28

6

27


(for-profit)


100 to 499

163

41

28




500+

281

112

29

2-year

Public

––

<1,000

113

16

30




1,000 to 2,999

288

72

31




3,000 to 4,999

201

80

32




5,000 to 9,999

241

121

33




10,000+

179

179

See notes at end of table.


Table B-6. Distribution of postsecondary institutions to be included in the sampling frame (continued)


Stratum

Level

Type of

Control

Highest

level of

offering

Enrollment

size class

Number of

institutions

in frame*

Number of institutions in sample

34

2 year

Private

––

<500

67

13

35


(nonprofit)


500 to 999

20

7

36




1,000+

5

3

37

2 year

Private

––

<500

369

41

38


(for profit)


500 to 999

154

31

39




1,000+

40

10

Total





4,096

1,668


* The population counts given in this table are based on initial tabulations from the 2008-09 IPEDS Institutional Characteristics (IC) file. Fifty-two eligible institutions in the 2008-09 IC file have missing values for enrollment and are not included in this table. However, such institutions will be included in the final sampling frame and given appropriate chances of selection.

Note: Details may not sum to total due to rounding.



For the 4-year institutions, measures of selectivity will be used, as appropriate for the interpretation of the results from the survey. Thus, it is important that the sample include institutions covering a broad range of selectiveness. Applying the algorithm developed by Cunningham (2005), 1 we have estimated the numbers of institutions to be included in the sample by various categories of selectivity. The numbers of institutions in the frame and the corresponding expected sample sizes are shown in Table B-7. As can be seen in this table, a wide range of institutions with respect to selectivity is expected to be included in the sample. Note that other methods of defining selectivity can also be employed. The results in table B-7 are intended to provide rough orders of magnitude of the sample sizes to be expected under the proposed sample design.
























Table B-7. Number of institutions in the frame and expected sample sizes, by selectivity among 4-year institutions


Selectivity (based on Cunningham method)

Level

Number of institutions in frame

Number of institutions in sample

Most/Very Selective

Public

106

93


Private

249

129





Moderately Selective

Public

316

248


Private

560

213





Minimally Selective

Public

96

58


Private

523

155





Open Admissions

Public

118

69


Private

451

131

Total


2,419

1,095

Note: Details may not sum to total due to rounding.




Expected Levels of Precision for the Full-Scale Survey


Table B-8 summarizes the approximate sample sizes and standard errors to be expected under the proposed design for selected subgroups. Since the sample sizes in Table B-8 are based on preliminary tabulations of the 2008-09 IC file, the actual sample sizes to be achieved may differ from those shown. Also, it is important to note that the sample sizes represent the expected numbers of completed questionnaires with eligible institutions, and not the initial numbers of institutions to be sampled. The standard errors in Table B-8 reflect an overall design effect of 1.3. The design effect arises primarily from the use of variable sampling fractions across the major sampling strata and differential nonresponse weighting adjustments (see description under Estimation and Calculation of Sampling Errors). In particular, the design effect reflects the fact that under the proposed stratified design, large institutions will be sampled at relatively higher rates (i.e., have smaller sampling weights) than small institutions. In fact, some very large institutions will be selected with certainty or near certainty. Subgroups consisting solely of such institutions will have negligible standard errors because the finite population correction (fpc) for these subgroups will be 0 or close to 0. Hence, the entries in the table should be viewed as rough upper bounds on the standard errors to be expected from the survey.


The standard errors in Table B-8 can be converted to 95 percent confidence bounds by multiplying the entries by 2. For example, an estimated proportion of the order of 20 percent (P = 0.20) for public 4-year institutions will be subject to a margin of error of ±4.6 percent at the 95 percent confidence level. Similarly, an estimated proportion of the order of 50 percent (P = 0.50) for 4-year institutions requiring test scores for admission will be subject to a margin of error of ±4.4 percent at the 95 percent confidence level.



Table B-8. Expected standard error of an estimated proportion under proposed design for selected analytic domains, by control and level




Standard error† of an estimated



proportion equal to ...


Domain (subset)

Expected sample size*


P = 0.20


P = .33


P = .50


Total

1,418

0.012

0.014

0.015

Public

796

0.016

0.019

0.020

4-year

398

0.023

0.027

0.029

2-year

398

0.023

0.027

0.029

Private, 4-year

534

0.020

0.023

0.025

Not-for-profit

399

0.023

0.027

0.029

For-profit

135

0.039

0.046

0.049

Private, 2-year

89

0.048

0.057

0.060

4-year schools

931

0.015

0.018

0.019

Requires test scores

620

0.018

0.022

0.023

Has open admissions


144


0.038


0.045


0.048


Most/Very Selective**

188

0.023

0.027

0.028

Moderately Selective

392

0.023

0.027

0.029

Minimally Selective

181

0.034

0.040

0.042

Open Admissions

170

0.035

0.041

0.044







* Expected number of responding eligible institutions, assuming response rate of 85 percent. The standard errors given in this table are given for illustration. Actual standard errors may differ from those shown.


Assumes unequal weighting design effect of 1.3. For subgroups consisting of institutions selected with certainty, the standard errors will be smaller than those shown.


** Standard errors include an approximate finite population correction to reflect the fact these institutions will be selected at relatively high rates.




Estimation and Calculation of Sampling Errors for the Full-Scale Survey


For estimation purposes, sampling weights reflecting the overall probabilities of selection and adjustments for nonresponse will be attached to each data record. To properly reflect the complex features of the sample design, standard errors of the survey-based estimates will be calculated using jackknife replication. Under the jackknife replication approach, 50 subsamples or "replicates" will be formed in a way that preserves the basic features of the full sample design. A set of estimation weights (referred to as "replicate weights") will then be constructed for each jackknife replicate. Using the full sample weights and the replicate weights, estimates of any survey statistic can be calculated for the full sample and each of the 50 jackknife replicates. The variability of the replicate estimates is used to obtain a measure of the variance (standard error) of the survey statistic. Previous surveys, using similar sample designs, have yielded relative standard errors (i.e., coefficients of variation) in the range of 2 to 10 percent for most national estimates. Similar results are expected for this survey.


B.3. Methods for Maximizing the Response Rate for the Full-Scale Survey


As in the pilot test, the President’s office of each sampled postsecondary institution will be contacted to solicit the institution’s participation in the survey and to identify the appropriate survey respondent. Survey respondents will have the option of completing the survey on paper or on the web. The paper version of the questionnaire is limited to four pages of questions. Telephone followup for nonresponse, which will be conducted by Westat staff, will begin about 3 weeks after mailout for each type of collection. Experienced telephone interviewers will be trained to conduct followup and will be monitored by Westat supervisory personnel during all interviewing hours. Collection procedures will follow standard methods developed by Westat for the pilot test and previous surveys on postsecondary institutions.



B.4. Tests of Procedures and Methods


A pretest of the survey was conducted with seven respondents to determine what problems respondents might have in providing the requested information and to make appropriate changes to the questionnaire. Responses and comments on the questionnaire were collected by fax and telephone during the pretest, and the results were summarized as part of the documentation for the study.


A pilot test was conducted with a sample of 120 postsecondary institutions. Respondents had the option of completing a web version that will be accessed through the Internet or a traditional paper and pencil questionnaire. The pilot test data was analyzed to identify potential problems that may emerge from data collection during the full-scale survey. In addition, findings from the pilot test provided insights into the extent to which institutions use various tests, either independently or in combination with other evaluation criteria to identify student need for remediation in mathematics and reading. Results of the pilot test, including survey management procedures for contacting respondents and obtaining cooperation, was summarized as part of the documentation for the study.



B.5. Reviewing Statisticians for the Pilot Test and Full-Scale Survey


Westat’s statisticians, Adam Chu (telephone: 301-251-4326) and Greg Norman were consulted about the statistical aspects of the design.


1Cunningham, A.F. (2005). Changes in Patterns of Prices and Financial Aid (NCES 2006-153). U.S. Department of Education. Washington, DC: National Center for Education Statistics.



File Typeapplication/vnd.openxmlformats-officedocument.wordprocessingml.document
File Title2001CU_REPORT.DOC
AuthorStat Group
File Modified0000-00-00
File Created2021-02-01

© 2024 OMB.report | Privacy Policy