Part B NPSAS24 FS Student Collection

Part B NPSAS24 FS Student Collection.docx

2023-24 National Postsecondary Student Aid Study (NPSAS:24) Full-Scale Study - Student Data Collection and Student Records

OMB: 1850-0666

Document [docx]
Download: docx | pdf

2023–24 NATIONAL POSTSECONDARY STUDENT AID STUDY (NPSAS:24) FULL-SCALE STUDY

Student Data Collection and Student Records

Supporting Statement Part B

OMB # 1850-0666 v. 36

Submitted by

National Center for Education Statistics

U.S. Department of Education

September 2023




Tables and Figures





Figure 1. Institution contacting 12



B. Collection of Information Employing Statistical Methods

This submission requests clearance for the 2023–24 National Postsecondary Student Aid Study (NPSAS:24) full-scale student data collection materials and procedures, which include the institution student record data abstraction and student survey. It carries over respondent burden, procedures, and materials related to the NPSAS:24 institution contacting, enrollment list collection, list sampling, and administrative matching activities previously approved in September 2023 (OMB#1859-0666 v. 35). Specific plans are provided below.

  1. Respondent Universe

  1. Institution Universe

NPSAS:24 will be nationally representative for both undergraduate and graduate students and will use a two-stage sampling design. The first stage involves the selection of institutions. In the second stage, students are selected from within sampled institutions. Also, the NPSAS:24 sample is designed to serve as the base year for a 2024 cohort of the Baccalaureate and Beyond (B&B) Longitudinal Study and, therefore, will include a nationally representative sample of baccalaureate recipients. Although no funding is available to field the follow-up surveys, this allows NCES to add the collections later should Congress appropriate additional funding. To construct the full-scale institution sampling frame for NPSAS:24, we used institution data collected from various surveys of the Integrated Postsecondary Education Data System (IPEDS). The student sampling frame includes all students who meet eligibility requirements from the participating institutions.

The NPSAS:24 institution (first stage) sampling frame includes all levels (less-than-2-year, 2-year, and 4-year) and control classifications (public, private nonprofit, and private for-profit) of Title IV eligible postsecondary institutions in the 50 states, the District of Columbia, and Puerto Rico. To be eligible for NPSAS:24, an institution must do the following during the 2023–24 academic year:

  • offer an educational program designed for persons who have completed secondary education;

  • offer at least one academic, occupational, or vocational program of study lasting at least 3 months or 300 clock hours;

  • offer courses that are open to more than the employees or members of the company or group (e.g., union) that administer the institution;

  • be located in at least one of the 50 states, the District of Columbia, or Puerto Rico;

  • be other than a U.S. service academy;1 and

  • have a signed Title IV participation agreement with the U.S. Department of Education.2

As indicated above, institutions providing only avocational, recreational, or remedial courses or only in-house courses for their own employees will be excluded.

  1. Student Universe

The student (second stage) sampling frame is described below. NPSAS-eligible undergraduate and graduate students are those who were enrolled in the NPSAS institution in any term or course of instruction between July 1, 2023 and April 30, 20243 for the full-scale and who are:

  • enrolled in either (1) an academic program; (2) at least one course for credit that could be applied toward fulfilling the requirements for an academic degree; (3) exclusively noncredit remedial coursework that has been determined by their institution to be eligible for Title IV aid; or (4) an occupational or vocational program that requires at least 3 months or 300 clock hours of instruction to receive a degree, certificate, or other formal award; and

  • not concurrently enrolled in high school; and

  • not enrolled solely in a General Educational Development (GED®)4 or other high school completion program.


  1. Statistical Methodology

  1. Institution Sample

The NPSAS:24 full-scale institution frame was constructed prior to the start of the field test data collection from the IPEDS 2021-22 Institutional Characteristics (IC) header, 2021-22 IC, 2020-21 Completions, and 2020-21 Full-year Enrollment files. Prior to the start of full-scale data collection, the institution sample was freshened using the IPEDS:2022-23 header, 2022-23 IC, 2021-22 Completions, and 2021-22 Full-year Enrollment files. This ensures that all potentially eligible institutions have a probability of selection. As part of this process, we also used the current IPEDS files to identify full-scale institutions no longer eligible for NPSAS:24 due to closure. Because it is possible that some for-profit institutions and large chains of for-profit institutions will have closed or been sold after the latest IPEDS data collection, we will conduct web searches for information about closed institutions to identify them. For the small number of institutions on the frame that have missing enrollment information because they are not imputed as part of IPEDS, we will impute the enrollment.

The institution strata for the full-scale study will be the eleven sectors, used since the NPSAS:16 full-scale study, which are based on institution level, control, and highest level of offering:

  • Public less-than-2-year;

  • Public 2-year;

  • Public 4-year, non-doctorate-granting, primarily subbaccalaureate;

  • Public 4-year, non-doctorate-granting, primarily baccalaureate;

  • Public 4-year, doctorate-granting;

  • Private nonprofit less-than-4-year;

  • Private nonprofit 4-year, non-doctorate-granting;

  • Private nonprofit 4-year, doctorate-granting;

  • Private for-profit less-than-2-year;

  • Private for-profit 2-year; and

  • Private for-profit 4-year.

The full-scale sample of 2,020 institutions was selected first using a variation of probability proportional to size (PPS) sampling called sequential probability minimum replacement (PMR) sampling.5 This method selects institutions sequentially with probability proportional to size and with minimum replacement. Selection with minimum replacement means that the actual number of hits for an institution can equal the integer part of the expected number of hits for that institution, or the next largest integer, that is, institutions have a chance of being selected more than once.6 Instead of the PMR sampling algorithm selecting some institutions multiple times, prior to the PMR sample selection, we set aside for inclusion in the sample with certainty all institutions with a probability of being selected more than once, that is, adjusting their probability of selection to be one. Then, the probabilities of selection for other institutions were adjusted accordingly prior to PMR selection, so that the total institution sample size target was met. A composite size measure7 was used to help achieve self-weighting samples8 for student-by-institution strata and to allow flexibility to change sampling rates in selected strata without losing the self-weighting attribute of the sampling method. Institution composite measures of size were determined using undergraduate and graduate student enrollment counts and baccalaureate recipient counts from the 2020-21 IPEDS Full-year Enrollment and Completion files.

Additionally, four institutions were added to the sample during the freshening process. The freshened sample size was determined based on the enrollment of the 66 new institutions such that the probabilities of selection are similar to the originally sampled institutions in the same strata. The total institution sample size is 2,025.

The institution sample sizes by the eleven institution strata, prior to sample freshening, are presented in table 1. We expect to obtain an overall 97 percent student eligibility rate and at least an overall 80 percent institution participation (response) rate, which will yield approximately 1,570 institutions providing lists.


Within each institution stratum, additional implicit stratification will be accomplished by sorting the sampling frame by the following classifications, as appropriate:


  1. Level of institution;

  2. Historically Black Colleges and Universities (HBCU) indicator;

  3. Hispanic-serving institutions (HSI) indicator;9

  4. Carnegie classification of postsecondary institutions;10 and

  5. The institution measure of size.

The objective of this implicit stratification is to approximate proportional representation of institutions on these measures.


Table 1. Number of full-scale institutions in the population and sampled, by control and level of institution1

Control and level of institution

Population estimate

Sample size

Total

5,920

2,025




Public less-than-2-year

224

22

Public 2-year

896

378

Public 4-year, non-doctorate-granting, primarily sub-baccalaureate

186

83

Public 4-year, non-doctorate-granting, primarily baccalaureate

214

116

Public 4-year, doctorate-granting

391

340

Private nonprofit 2-year or less

178

20

Private nonprofit 4-year, non-doctorate-granting

854

325

Private nonprofit 4-year, doctorate-granting

751

268

Private for-profit less-than-2-year

1,369

71

Private for-profit 2-year

530

122

Private for-profit 4-year

327

280

1 The number of institutions in the population and sample is prior to sample freshening.

NOTE: Details may not sum to totals due to rounding.

SOURCE: Population estimates based on IPEDS 2022-23 data.


When necessary, substitutions for sampled, eligible institutions not providing student enrollment lists may be used so that we have sufficient institution participation and sampled students. To do so, we will recreate the institution sampling frame in the same order as used for sample selection described above. Then, within the institution strata, we will identify institutions on the frame immediately before and after the sampled institution as potential substitutes. Substitutes will not include institutions already selected for either the full-scale or field test sample. Of the two substitute institutions identified, we will use the one that has the closest measure of size to the sampled institution. Any institutions included in the sample with certainty will not have substitutes because they do not have neighboring institutions with a similar measure of size that are not already in the sample. We plan to identify substitute institutions in the following four strata:

  • public less-than-2-year;

  • private nonprofit 2-year or less;

  • private for-profit less-than-2-year; and

  • private for-profit 2-year.

These strata have had historically low institution participation rates. They also have low institution sampling rates, which allows for substitutes to be available for many sampled institutions, and there are not many institutions in these strata that were selected with certainty.

  1. Student Sample

To begin NPSAS data collection, sampled institutions are asked to provide a list of their NPSAS-eligible undergraduate and graduate students enrolled in the targeted academic year, covering July 1 through June 30 (methods for contacting the sampled institutions are described below in section B.3, and student list data elements are described in appendix D1). Since NPSAS:04, institutions have been asked to limit listed students to those enrolled through April 30 and, starting in NPSAS:20, continuous enrollment institutions were asked to limit listed students to those enrolled through March 31.11 These truncated enrollment periods exclude students who first enrolled in May or June (and additionally in April for continuous enrollment institutions), but they allow lists to be collected earlier and, as a result, data collection to be completed in less than 12 months. Any lack of coverage resulting from the truncated enrollment period will be accounted for by the poststratification weight adjustment.

We will request that high school students be included on the enrollment list even though these students are not eligible for NPSAS. While these students will be excluded from sampling, high school student counts are needed later for the weighting poststratification adjustment. We will poststratify the NPSAS students to IPEDS enrollment counts (used as control totals), which include high school students. As dual enrollment becomes more prominent, it is important that we adjust the IPEDS counts downward to account for dual enrollment. Since dual enrollment counts are not currently readily available, using high school student counts from the enrollment lists may be the best source for adjusting IPEDS counts and ensuring accurate control totals.

In addition to collecting typical enrollment lists from institutions, we will also attempt to obtain enrollment lists directly from the National Student Clearinghouse (NSC) for some institutions unable to provide enrollment lists but willing to provide permission for us to obtain their lists from the NSC. This may be done, within the strata that do not have substitute institutions (see above), for those institutions that provide data to NSC. Obtaining enrollment lists directly from NSC has the potential to reduce institution burden and help with refusal conversion. See section B.4 for results of evaluating enrollment lists obtained from the NSC during the field test.

Student Stratification

The student sampling strata will be:

  1. potential baccalaureate recipients who are veterans;

  2. potential baccalaureate recipients from science, technology, engineering, and mathematics (STEM) programs, White;

  3. potential baccalaureate recipients from STEM programs, all other race/ethnicity categories (Black or African American, Hispanic or Latino, Asian, American Indian or Alaska Native, Native Hawaiian or Pacific Islander);

  4. potential baccalaureate recipients from teacher education programs, White;

  5. potential baccalaureate recipients from teacher education programs, all other race/ethnicity categories (Black or African American, Hispanic or Latino, Asian, American Indian or Alaska Native, Native Hawaiian or Pacific Islander);

  6. potential baccalaureate recipients from business programs, White;

  7. potential baccalaureate recipients from business programs, all other race/ethnicity categories (Black or African American, Hispanic or Latino, Asian, American Indian or Alaska Native, Native Hawaiian or Pacific Islander);

  8. potential baccalaureate recipients from other programs, White;

  9. potential baccalaureate recipients from other programs, all other race/ethnicity categories (Black or African American, Hispanic or Latino, Asian, American Indian or Alaska Native, Native Hawaiian or Pacific Islander);

  10. other undergraduate students who are veterans;

  11. other undergraduate students, Hispanic or Latino;

  12. other undergraduate students, White;

  13. other undergraduate students, Black or African American;

  14. other undergraduate students, other;

  15. graduate students who are veterans;

  16. master’s degree students in STEM programs;

  17. master’s degree students in education and business programs;

  18. master’s degree students in other programs;

  19. doctoral-research/scholarship/other students in STEM programs;

  20. doctoral-research/scholarship/other students in education and business programs;

  21. doctoral-research/scholarship/other students in other programs;

  22. doctoral-professional practice students; and

  23. other graduate students.



We are keeping the graduate student strata similar to the sampling strata used since NPSAS:16. If students fall into multiple strata, such as graduate students who are veterans, the ordering of the strata above will be used to prioritize the stratification.


Several student subgroups will be intentionally sampled at rates different than their natural occurrence within the population due to specific analytic objectives. The following groups will be oversampled:


  • potential baccalaureate recipients overall and who are:

    • veterans,12

    • in STEM or teacher education programs,

    • all race/ethnicity categories other than White (Black or African American, Hispanic or Latino, Asian, American Indian or Alaska Native, Native Hawaiian or Pacific Islander),

    • in public 4-year, non-doctorate-granting, primarily sub-baccalaureate institutions, or

    • in private for-profit 4-year institutions;

  • other undergraduate students who are:

    • veterans,

    • Hispanic or Latino,

    • Black or African American, or

    • all race/ethnicity categories other than White (Black or African American, Hispanic or Latino, Asian, American Indian or Alaska Native, Native Hawaiian or Pacific Islander);

  • graduate students in STEM programs.


Similarly, we anticipate the following groups will be undersampled:

  • undergraduate White students;

  • potential baccalaureate recipients who are White, in business programs, in other non-STEM and non-teacher education programs, or in public 4-year, doctorate-granting institutions;

  • master’s degree students in education and business programs; and

  • doctoral-research/scholarship/other students in education and business programs.


Because these groups are so large, sampling in proportion to the population would make it difficult to draw inferences about the experiences of some other types of students.


We will match the student enrollment lists to two supplemental databases prior to sampling (pre-sampling matching), as has been done since NPSAS:16. Because the veterans identified by institutions on the lists are incomplete, in order to identify veterans, we will match the student enrollment lists with a list of veterans from the Veterans Benefits Administration (VBA). This veterans’ information will be used with the veteran status from the enrollment lists for full-scale stratification. We will also match the student lists to the National Student Loan Data System (NSLDS) data and use the available financial aid data for student-implicit stratification. Within the student-explicit strata, we will sort the students by federally aided/unaided, and this will allow the sample proportions of aided and unaided students to approximately match the population within institution and student strata.

Identification of baccalaureate recipients

NPSAS:24 may serve as the base year for a 2024 cohort of B&B and will include a nationally representative sample of baccalaureate recipients, hence the stratification described above. This allows NCES to decide later if they will conduct a B&B study should funding become available. We will ask all institutions that award baccalaureate degrees to identify baccalaureate recipients. Instead of waiting until July for institutions to positively identify these students and send in lists, we will request that a baccalaureate indicator be included on the enrollment lists to flag whether students have completed requirements for or received a bachelor’s degree between July 1 and the date the enrollment list is provided. In NPSAS:16, we additionally requested that institutions provide a second indicator on the lists to flag students who had not yet received their bachelor’s degree but were expected to receive it by June 30.


Because of the difficulty institutions experienced in providing this second indicator in NPSAS:16, we developed a baccalaureate proxy algorithm for the NPSAS:24 field test based on analysis of NPSAS:16 data from student lists and the Central Processing System (CPS). We established the proxy separately for independent and dependent students within baccalaureate-granting institutions. Dependency status was determined based on pre-sampling matching to CPS. We classified as independent any students who, according to the enrollment list, were 24 years or older or a veteran, and did not fill out the Free Application for Federal Student Aid. We flagged all other students for whom there was no CPS data as having an unknown dependency status.


We performed logistic modeling to predict the probability of a student in baccalaureate-granting institutions being a baccalaureate recipient. Prior to running the logistic models, we divided the data into two groups – those to include in and those to exclude from the models. Dependent students were excluded from the model if they were 20 years old or younger; independent students were excluded from the model if they were doctoral students or other graduate students not enrolled for the master’s degree (graduate students at the master’s level were retained).


With the students flagged for inclusion, we then ran two logistic models – one for independent students and one for dependent students. The dependent variable in each model is baccalaureate receipt, based on enrollment list data, rather than survey data, because the baccalaureate proxy is meant to replace what an institution would provide on the list. Table 2 below lists the predictor variables used in the models and for which model it was included, independent and/or dependent. For continuous variables, we tried both the continuous and categorical versions in the models and determined that, except for age for independent students, the categorical versions performed better at predicting baccalaureate status.


We evaluated the criteria used to include or exclude students from the baccalaureate proxy model using NPSAS:24 field test list and survey data, and included in the evaluation a second indicator, provided by sampled institutions, for students who were expected to complete requirements for or receive a bachelor’s degree between the date the enrollment list was provided and June 30, 2023 (similar to what was requested in NPSAS:16). We calculated both false-positive and false-negative rates for all baccalaureate recipients based on the modeling. Evaluation results are presented below in section B.4. Based on results of the evaluation, we will adjust and test a new model using field test results. During full-scale sampling, we will continue to ask institutions to provide the second baccalaureate indicator described above. If an institution indicates not knowing a student’s baccalaureate status (selects “don’t know”), we will run the proxy to estimate the student’s status. Having the proxy as a supplement will help to prevent underestimation of the number of baccalaureates for sampling.


Table 2. NPSAS:24 field test predictor variables for the logistic models predicting baccalaureate status

Variable description

Model for independent students

Model for dependent students

Control and level of institution

Class level of student (year of enrollment – 1st, 2nd, 3rd,…)

Months since high school graduation (categorical – 0 or missing, 1-47, 48-61, 62 or more)


Indicator of having graduated high school at least 47 months ago


Indicator of enrollment date at least 33 months ago


Indicator of enrollment date at least 45 months ago


Indicator of having at least 105 credit hours

Age (continuous)


Age (categorical – 21, 22, 23, missing)


Dependency status (dependent and unknown)


NOTE: Months since high school graduation and since enrollment at NPSAS institution are based on June 30 of the NPSAS year.

Sample Sizes and Student Sampling

NPSAS:24 will be designed to sample about 162,000 students. We expect to obtain, minimally, 95 percent eligibility rates and at least a 70 percent response rate overall. This will yield approximately 107,730 student surveys. The graduate student sample is 25,000, matching the sample size targets in both NPSAS:18-AC and NPSAS:20. Table 3 shows the preliminary population and sample sizes, respectively, by control and level of institution and student type.


In setting the NPSAS:24 sample sizes, we need to determine the sample size of potential baccalaureate recipients, who will be part of both NPSAS and the B&B 2024 cohort, if a B&B study is conducted. The B&B cohort would include 30,000 baccalaureate recipients who respond to the NPSAS:24 survey and confirm that they have received their bachelor’s degree in the appropriate time frame. The NPSAS:24 potential baccalaureate recipient preliminary sample size will be approximately 53,040, assuming a 95 percent eligibility rate, a 70 percent survey response rate, a 19.7 percent false-positive rate, and a 3.0 percent false-negative rate among other undergraduate students, as in NPSAS:16.13 and this will be updated after the sample optimization.


Institution-level student sampling rates will be set based on frame data and adjusted to account for the overestimation of enrollment counts in IPEDS data that has been found in prior NPSAS list collections. Based on these adjusted rates, students will be sampled on a flow basis as enrollment lists are received using stratified systematic sampling procedures. As mentioned above, student strata will be sorted by federally aided/unaided students to maintain proportionality between the sample and frame. Within the graduate-student stratum for veterans, the students will be sorted by master’s and doctoral degree levels to ensure that the sample will be roughly proportional to the frame. Sample yield will be monitored by institution and student sampling strata, and the sampling rates will be adjusted early, if necessary, to achieve the desired sample yields.

Quality Control Checks for Lists and Sampling

The number of enrollees on each institution’s student list will be checked against the latest IPEDS full-year enrollment and completions data for each student level: baccalaureate, undergraduate, and graduate. As has been




Table 3. NPSAS:24 preliminary student population and sample sizes, by control and level of institution and student type

Potential baccalaureate recipients

Other undergraduate students

Graduate students

Control and level of institution

Population estimate

Sample size

Population estimate

Sample size

Population estimate

Sample size

Total

2,134,236

53,043

18,928,868

83,957

4,081,808

25,000








Public less-than-2-year

0

0

84,920

848

0

0

Public 2-year

0

0

6,705,311

26,556

0

0

Public 4-year, non-doctorate-granting, primarily sub-baccalaureate

26,686

3,694

2,167,527

3,702

2,708

97

Public 4-year, non-doctorate-granting, primarily baccalaureate

157,581

3,694

766,915

3,470

132,748

2,015

Public 4-year, doctorate-granting

1,217,550

13,745

4,787,445

13,905

1,817,471

6,074

Private nonprofit 2-year or less

0

0

73,432

1,084

0

0

Private nonprofit 4-year, non-doctorate-granting

205,571

7,050

970,369

5,333

212,716

3,358

Private nonprofit 4-year, doctorate-granting

423,620

7,852

1,747,074

4,643

1,549,338

5,878

Private for-profit less-than-2-year

0

0

374,444

4,193

0

0

Private for-profit 2-year

0

0

355,346

8,661

0

0

Private for-profit 4-year

103,228

17,008

896,085

11,562

366,827

7,578

SOURCE: Population estimates based on IPEDS 2021-22 data.


done in past rounds of NPSAS, only counts within 50 percent of non-imputed IPEDS counts will pass QC and will be allowed to move to student sampling.


Institutions that fail QC will be re-contacted to resolve the discrepancy and to verify that the institution coordinator who prepared the student list clearly understood our request and provided a list of the appropriate students and data. If we determine that the initial list provided by the institution was not satisfactory, we will request a replacement list. We will proceed with selecting sample students when we have either confirmed that the list received is correct or have received a corrected list.


QC is very important for sampling and all statistical activities. All statistical procedures will undergo thorough QC checks, following the Quality Management Plan. We will employ a checklist for all statisticians to use to make sure that all appropriate QC checks are done for student sampling. Some specific student sampling QC checks include, but are not limited to, checking that the:


  • students on the sampling frames all have a known, non-zero probability of selection;

  • high school students are excluded;

  • student strata are populated, as expected, based on institutions stratum;

  • email addresses match student names; and

  • number of students selected match the target sample sizes.



  1. Methods for Maximizing Response Rates

Achieving high response rates in NPSAS:24 will depend on successfully recruiting institutions, obtaining institution data and other administrative data, and identifying and locating sample members and being able to gain their cooperation. The following sections outline methods for maximizing institution and student response to in the NPSAS full-scale data collection.

  1. Collection of Data from Institutions

Establishing and maintaining contact with sampled institutions throughout the data collection process is vital to the success of NPSAS:24. Institution participation is required in order to collect enrollment lists and select the student sample. The process by which institutions will be contacted is depicted in figure 1 and described below.


The data collection contractor will be responsible for contacting institutions on behalf of NCES. Each staff member will be assigned a set of institutions that is their responsibility for the duration of data collection. This allows contractor staff to establish rapport with institution staff and provides a reliable point of contact for the institution. Staff members are thoroughly trained in basic financial aid concepts and in the purposes and requirements of the study, which helps them establish credibility with the institution staff.


The first step in the process is verification of the chief administrator’s contact information. Web searches and verification calls will be conducted to confirm eligibility and confirm contact information obtained from the IPEDS header files before study information is mailed. The Higher Ed Directory (https://hepinc.com/) may also be used to verify information. Once the contact information is verified, we will prepare and send an information packet to the chief administrator of each sampled institution. A copy of the letter and brochure can be found in appendix D1. The materials provide information about the purpose of the study and the nature of subsequent requests. Two versions of the chief administrator letter will be used, tailored to the institution’s situation: (1) one letter for institutions for which we identify and recommend a potential campus coordinator from previous NPSAS participation; (2) another letter for institutions for which we cannot identify and recommend a potential campus coordinator. For institutions without a recommended coordinator, institution contactors will conduct follow-up calls to the chief administrator to secure study participation and identify a campus coordinator. If the coordinator is not already a Postsecondary Data Portal user, they will be added as a user.


NCES and its contractor will identify relevant multi-campus systems within the sample because these systems can supply enrollment list data at the system level, minimizing burden on individual campuses. Even when it is not possible for a system to supply data from a centralized office, the system can lend support in other ways, such as by prompting institutions under its jurisdiction to participate. NCES and its contractor will undertake additional outreach activities, such as engaging with higher education organizations and networking within the postsecondary community at conferences and professional meetings. These activities are intended to promote the value of NPSAS both to data providers and data users thereby increasing interest and participation in NPSAS:24.

Once a campus coordinator has been identified for an institution, the contractor will send the coordinator study materials with a request to complete the online Registration Page as the first step. The materials include a letter, the study brochure, and a quick guide to participation in the study (see appendix D1). The primary functions of the Registration Page are to confirm the date the institution will be able to provide the student enrollment list and to determine how they will report student records data, by term or by month. Based on the information provided, a customized timeline for collecting the enrollment list will be created for each institution.

After the Registration Page is completed, the campus coordinator will be sent a letter or email requesting an electronic enrollment list of all students enrolled during the academic year. Enrollment lists will be collected from January 2024 to July 2024. As described above, the lists will serve as the frame from which the student samples will be drawn. Follow-up contacts with institutions include telephone prompts, reminder emails and mailers, typically sent prior to a deadline, and touch-base emails typically sent after a period of no outbound contact from study staff (see appendix D1). After enrollment lists are received and validated by the contractor for completeness and quality, the campus coordinator will be sent a “thank you” email acknowledging appreciation for their time and effort.

Figure 1. Institution contacting



Student Enrollment List Template

In the NPSAS:24 field test, we offered institutions a new option to submit student enrollment lists using a pre-formatted Excel template, similar to the Excel template offered during the student records collection. The enrollment list template option was offered for a few purposes: first, to reduce burden on institutions by making it easier to format their enrollment lists; second, to increase the uniformity of lists by encouraging institutions to submit data in a single format; and third to increase data quality by performing error checks on the list data at the time the list is uploaded. Institutions were given a choice of using the new template option or preparing the student enrollment list file in their own format (consistent with prior rounds of NPSAS). Of the 184 institutions submitting a student enrollment list in the field test, 135 (73 percent) submitted lists using the template option. Of the 42 institutions that received data errors when uploading, 41 percent uploaded a revised file that resolved the errors. Based on the number of institutions that chose the template option and the number that were able to immediately resolve data errors, we will continue to offer both the template option and the “create your own” option in the full-scale data collection.


Alternate Enrollment List Submission Method

As previously described, in addition to collecting typical enrollment lists from institutions, we will also attempt to obtain enrollment lists directly from the National Student Clearinghouse (NSC) for some institutions unable to provide enrollment lists. This may be done in the strata that do not have substitute institutions (see above) for institutions that provide data to NSC. Obtaining enrollment lists directly from NSC has the potential to reduce institution burden and help with refusal conversion. Permission from the institution will be required before obtaining enrollment list data directly from NSC.

Postsecondary Data Portal (PDP)

The NPSAS:24 institution data collection will utilize NCES’ Postsecondary Data Portal (PDP) website. The flexible design of the website allows it to be used for multiple NCES postsecondary institution sample studies in data collection at the same time, even when those studies collect different types of data. Currently, there are no plans for other postsecondary data collections to be underway using the PDP during the NPSAS:24 full-scale.

The PDP provides to users both general-purpose and study-specific content. General-purpose pages provide overview information about NCES postsecondary studies and use of the website. These pages are identified in appendix D1 as the “pre-login” pages. Once a user logs in, they see pages with study-specific content. These pages are identified in appendix D as the “after login” content. The NPSAS:24 study-specific content includes FAQs about NPSAS:24, instructions and resources, and pages for providing data (appendix D1). Institutions see study-specific PDP content only for the study or studies for which they have been sampled.

The PDP was updated for the NPSAS:24 field test to add a new option for institutions to submit the student enrollment list using a pre-formatted Excel template file. This new feature was designed for institutions that had requested an Excel template in prior rounds of NPSAS and also facilitates real-time error checking during the enrollment list upload process. Enrollment list error checking will provide institution staff with immediate feedback about potential enrollment list data problems, rather than waiting for NPSAS staff to review the enrollment list and follow-up with feedback at a later date. Institutions that prefer to create their own enrollment list files will still be able to do so.

Student Records

After students are sampled from an institution’s enrollment list, the institution coordinator will receive a mailing containing a letter requesting student records data for those sampled students. Institutional contactors will follow up after the mailing to ensure receipt of the package and to answer any questions. Follow-up contacts include telephone prompts, reminder emails that are typically sent 2 weeks prior to a deadline, and touch-base emails typically sent when 3–4 weeks have passed with no outbound contact from study staff. Contact materials are included in Appendix D2. Staff will also be available by telephone and email to help when institution staff have questions or encounter problems.

As with the enrollment list collection, the student record collection will utilize the PDP. The content of the PDP specific to student records collection is included in Appendix F (the student records instrument content) and Appendix D2 (student records communication materials). The following options will be offered to institutions for collecting student records:

  • Web-based data entry interface. The web-based data entry interface allows the coordinator to enter data by student, by year.

  • Excel workbook. An Excel workbook will be created for each institution and will be preloaded with each sampled student’s ID, name, date of birth, and last four digits of SSN (if available). To facilitate simultaneous data entry by different offices within the institution, the workbook contains a separate worksheet for each of the following topic areas: Student Information, Financial Aid, Enrollment, and Budget. The user will download the Excel worksheet from the PDP, enter the data, and then upload the data. Validation checks will occur both within Excel as data are entered and when the data are uploaded. Data will be imported into the web application so that institution staff can check their submission for quality control purposes.

  • CSV (comma separated values) file. Institutions with the means to export data from their internal database systems to a flat file may use this method of supplying student records. Institutions that select this method will be provided with detailed import specifications, and all data uploading will occur through the PDP. Like the Excel workbook option, data will be imported into the web application so that institution staff can check their submission before finalizing.

  • Alternate method. Institutions will be offered an alternate submission format allowing staff to upload data in any format or file type that is convenient, rather than making their data conform to our template or CSV specifications, as a refusal aversion strategy (described below).


Refusal Aversion Strategies with Institutions

If institution staff report a lack of time or resources needed to provide student records data, the following additional accommodations will be offered:

  • reimbursement to help offset labor or staffing costs;

  • a reduced set of the most critical data elements (see data elements marked with an asterisk in Appendix D2); and/or

  • an alternate submission format allowing staff to upload data in any format or file type that is convenient, rather than making their data conform to our template or CSV specifications.

Data Security on the PDP

Because of the risks associated with transmitting confidential data on the internet, the latest technology systems will be incorporated into the web application to ensure strict adherence to NCES confidentiality guidelines. The web server will include a Secure Sockets Layer (SSL) certificate and will be configured to force encrypted data transmission over the Internet. All data-entry modules on this site require the user to log in before accessing confidential data. Logging in requires entering an assigned ID number and two-factor authentication with a code sent via email and a password. Through the PDP, the campus coordinator at the institution will be able to use a “Manage Users” link to add and delete users, as well as reset passwords and assign roles. Each user will have a unique username and will be assigned to one e-mail address. Upon account creation, the new user will be sent a temporary password by the PDP. When logging in for the first time, the new user will be required to create a new password. The system automatically will log out after 20 minutes of inactivity. Files uploaded to the secure website will be stored in a secure project folder that is accessible and visible to authorized project staff only.



  1. Matching to Administrative Databases

Information about NPSAS:24 sampled students will be matched with their data from several administrative databases, including NSLDS, CPS including FAFSA, NSC, VBA, and student records obtained directly from postsecondary institutions. Further details about these matches are provided in the Supporting Statement Part A (sections A.1, A.2, A.10, and A.11) and in appendix C. We continue to explore matches to other potential data sources to be added to the full-scale collection, such as the Supplemental Nutrition Assistance Program (SNAP) recipients’ data from the U.S. Department of Agriculture.



  1. Collection of Student Survey Data

Tracing of Sample Members

To yield the maximum number of located cases with the least expense, we designed an integrated tracing approach, with the following elements.

  • Advance tracing activities, which will occur prior to the start of data collection, include initial batch database searches, such as to the National Change of Address (NCOA) databases, for cases with sufficient contact information to be matched. To handle cases for which contact information is invalid or unavailable, additional advance tracing through proprietary interactive databases will expand on leads found.

  • Hard copy mailings, emails, and text messages will be used to maintain ongoing contact with sample members, prior to and throughout data collection. The student contacting materials, which will be developed with a design appealing to students in 2024, are provided in Appendix J. The initial mailing to sample members will include a letter announcing the start of data collection, requesting that the sample member complete the web survey, and including a toll-free number, the study website address, a Study ID and password, and a study brochure. We will send a similar email message mirroring information provided in the mailing.

  • Sample members will have a variety of means to provide updated contact information and contact preferences. Students can use an Update Contact Information page on the secure NPSAS:24 website to provide their contact information, including cell phone number, as well as provide contacting preferences with respect to phone calls, mailings, emails, and text messages. Help Desk calls and emails providing information about a sample member’s text message preferences will be monitored and the sample member’s data record updated as soon as the information becomes known.

  • The telephone locating and surveying stage includes calling all available telephone numbers and following up on leads provided by parents and other contacts.

  • The pre-intensive batch tracing stage consists of the LexisNexis SSN and Premium Phone batch searches that will be conducted between the telephone locating and surveying stage and the intensive tracing stage.

  • Once all known telephone numbers are exhausted, a case will move into the intensive tracing stage during which tracers will conduct interactive database searches using all known contact information for a sample member. With interactive tracing, a tracer assesses each case on an individual basis to determine which resources are most appropriate and the order in which each should be used. Sources that may be used, as appropriate, include credit database searches, such as Experian, various public websites, and other integrated database services.

  • Other locating activities will take place as needed, including a LexisNexis email search conducted for nonrespondents toward the end of data collection.

Training for Data Collection Staff

Telephone data collection will include supervisors and interviewers. Training programs for these staff members are critical to maximizing response rates and collecting accurate and reliable data.

Team supervisors, who are responsible for all supervisory tasks, will attend their own project-specific training, in addition to the interviewer training. They will receive an overview of the study, background and objectives, and the data collection instrument through a question-by-question review. Supervisors will also receive training in the following areas: providing direct supervision during data collection; handling refusals; monitoring interviews and maintaining records of monitoring results; problem resolution; case review; specific project procedures and protocols; reviewing reports generated from the ongoing Computer Assisted Telephone Interviewing (CATI); and monitoring data collection progress.

Training for interviewers is designed to help staff become familiar with and practice using the CATI case management system and the survey instrument, as well as to learn project procedures and requirements. Particular attention will be paid to quality control initiatives, including refusal avoidance and methods to ensure that quality data are collected. Interviewers will receive project-specific training on telephone interviewing and answering questions from web participants regarding the study or related to specific items within the survey. At the conclusion of training, all NPSAS data collection staff must meet certification requirements by successfully completing a certification interview. This evaluation consists of a full-length interview with project staff observing and evaluating interviewers, as well as an oral evaluation of interviewers’ knowledge of the study’s Frequently Asked Questions.

Case Management System

Surveys will be conducted using a single web-based survey instrument for both web (including mobile devices) and CATI data collection. Control of data collection activities will be accomplished through a CATI case management system, which is equipped with the numerous capabilities, including: on-line access to locating information and histories of locating efforts for each case; a questionnaire administration module with full “front-end cleaning” capabilities (i.e., editing as information is obtained from respondents); sample management module for tracking case progress and status; and automated scheduling module which delivers cases to interviewers. The automated scheduling module incorporates the following features:

  • Automatic delivery of appointment and call-back cases at specified times. This reduces the need for tracking appointments and helps ensure the interviewer is punctual. The scheduler automatically calculates the delivery time of the case in reference to the appropriate time zone.

  • Sorting of non-appointment cases according to parameters and priorities set by project staff. For instance, priorities may be set to give first preference to cases within certain sub-samples or geographic areas; cases may be sorted to establish priorities between cases of differing status. Furthermore, the historic pattern of calling outcomes may be used to set priorities (e.g., cases with more than a certain number of unsuccessful attempts during a given time of day may be passed over until the next time period). These parameters ensure that cases are delivered to interviewers in a consistent manner according to specified project priorities.

  • Restriction on allowable interviewers. Groups of cases (or individual cases) may be designated for delivery to specific interviewers or groups of interviewers. This feature is most commonly used in filtering refusal cases, locating problems, or foreign language cases to specific interviewers with specialized skills.

  • Complete records of calls and tracking of all previous outcomes. The scheduler tracks all outcomes for each case, labeling each with type, date, and time. These are easily accessed by the interviewer upon entering the individual case, along with interviewer notes.

  • Flagging of problem cases for supervisor action or supervisor review. For example, refusal cases may be routed to supervisors for decisions about whether and when a refusal letter should be mailed, or whether another interviewer should be assigned.

  • Complete reporting capabilities. These include default reports on the aggregate status of cases and custom report generation capabilities.

The integration of these capabilities reduces the number of discrete stages required in data collection and data preparation activities and increases capabilities for immediate error reconciliation, which results in better data quality and reduced cost. Overall, the scheduler provides an efficient case assignment and delivery function by reducing supervisory and clerical time, improving execution on the part of interviewers and supervisors by automatically monitoring appointments and call-backs, and reducing variation in implementing survey priorities and objectives.

Survey Instrument Design

The survey will employ a web-based instrument and deployment system, which has been in use since NPSAS:08. The system provides multimode functionality that can be used for self-administration, including on mobile devices, CATI, or data entry.

In addition to the functional capabilities of the case management system and web instruments described above, our efforts to achieve the desired response rate will include using established procedures proven effective in other large-scale studies we have completed. These include:

  • Providing multiple response modes, including mobile-friendly self-administered and interviewer-administered options.

  • Offering incentives to encourage response.

  • Assigning experienced CATI interviewers who have proven their ability to contact and obtain cooperation from a high proportion of sample members.

  • Training the interviewers thoroughly on study objectives, study population characteristics, and approaches that will help gain cooperation from sample members.

  • Maintaining a high level of monitoring and direct supervision so that interviewers who are experiencing low cooperation rates are identified quickly and corrective action is taken.

  • Making every reasonable effort to obtain an interview during the initial contact, but allowing respondent flexibility in scheduling appointments to be interviewed.

  • Thoroughly reviewing all refusal cases and making special conversion efforts whenever feasible (see next section).

  • Implementing and assuring participants of confidentiality procedures, including restricting the ability for the respondent to view survey responses from prior log in sessions (i.e. no ability to use navigation buttons to go to “Previous” survey questions from another log in session) and the survey automatically logging out of a session after 10 minutes of inactivity.

  • For the NSAS:24 full-scale collection, item-by-item toggling between English and Spanish languages at the discretion of the web respondent, or telephone interviewer when warranted.

Refusal Aversion and Conversion of Student Sample Members

Recognizing and avoiding refusals is important to maximize the response rate. We will emphasize this, and other topics related to obtaining cooperation during interviewer training. Supervisors will monitor interviewers intensely during the early days of outbound calling and provide retraining as necessary. In addition, the supervisors will review daily interviewer production reports produced by the CATI system to identify and retrain any data collectors who are producing unacceptable numbers of refusals or other problems.

Refusal conversion efforts will be delayed for at least 1 week to give the respondent time after the initial refusal. Attempts at refusal conversion will not be made with individuals who become verbally aggressive or who threaten to take legal or other action. Refusal conversion efforts will not be conducted to a degree that would constitute harassment. We will respect a sample member’s right to decide not to participate and will not impinge this right by carrying conversion efforts beyond the bounds of propriety.

Offer of Incentives

Sample members will be offered $30 for a completed survey, paid by the respondent’s choice of check or PayPal. In addition, we anticipate that a boost incentive will be needed for key student subgroups to further encourage participation and reduce the potential for nonresponse bias.


Researchers have commonly used incentive boosts as a nonresponse conversion strategy for sample members who have implicitly or explicitly refused to complete the survey (e.g., Groves and Heeringa 2006; Singer and Ye 2013).14 15 These boosts are especially common in large federal surveys during their nonresponse follow-up phase and have been implemented successfully in other postsecondary education surveys (e.g., HSLS:09 second follow-up; BPS:12/17; NPSAS:20). In NPSAS:20, a $10 incentive boost increased the overall response rate by about 3.2 percentage points above the projected response rate.


To determine whether a boost incentive is needed, we will monitor response rates overall, within data collection

wave, and within key student subgroups, particularly those that will be oversampled because of historic underrepresentation. At least at two points in time (by June and September), we will run logistic regression models to predict likelihood of response for nonrespondents, prioritizing members of the oversampled groups (see section B.2), with their sampling characteristics and any potential interactions as predictors, identifying those most likely to be nonrespondents. Results will be shared with OMB with a recommendation on whether these cases should be offered a $10 boost. Below we provide an example of the data collection flow for the oversampled cases for the earliest data collection waves:


  1. Feb-March: All oversampled members are invited to complete the NPSAS:24 survey and offered a $30 promised incentive.

  2. June (Sept for later waves): Predictive models identify cases likely to be missing from the respondent pool. Those cases receive a $10 incentive boost once approved.16

  3. August: Data collection reminders are paused for 2-3 weeks.

  4. September: Nonresponding oversample members are reinvited to complete the NPSAS:24 survey ($40 promised incentive).

  5. October: All nonresponding cases are offered an abbreviated survey.

  6. November: Final push and end of data collection.

For the rest of the sample, the data collection plan will be as follows:


  1. All sample members are invited to complete the NPSAS:24 survey and offered a $30 promised incentive.

  2. Data collection reminders are paused for 2-3 weeks for cases who have been in data collection for about 8-10 weeks.

  3. Cases who have been in data collection for about 5 months (3-4 for later sample releases) are offered the abbreviated survey.

  4. Begin end of data collection notifications for cases who have been in data collection for 6 months; such cases might be recontacted for a final push at the beginning of November, if needed.

  5. End of data collection for all cases in November.


  1. Tests of Procedures or Methods

The NPSAS:24 field test was used to evaluate several new procedures designed to improve data quality and decrease burden on institutions: the collection of enrollment lists from National Student Clearinghouse (described above in section B.2), the bachelor’s degree recipient proxy used for identifying potential B&B cohort members (described above in section B.2), and the new pre-formatted enrollment list template file (described above in section B.3). In addition, the student data collection tested three incentive approaches to determine the effects on student participation. Results of these tests are described below.

a. Collection of enrollment lists from National Student Clearinghouse

For a subset of field test institutions, we obtained enrollment list data directly from NSC, in addition to collecting standard NPSAS enrollment lists directly from the institutions. By comparing institutions’ NSC data with their regular NPSAS lists, the viability of collecting lists from NSC on a wider scale in the full-scale study could be assessed. After discussing the plan with staff from the institutions identified for this initiative and obtaining their permission, we were able to obtain NSC enrollment list data for over 110,000 students from 21 institutions.

To evaluate the quality of the NSC lists, student counts were compared between the lists from NSC and the institutions, as shown in Table 4.17 The NSC counts were higher than institution counts for some institutions and lower for others. Overall, the differences in the counts were minimal. When comparing the NSC and institution list counts to IPEDS counts, as described in section B.2, the results of lists passing or failing the QC checks were the same for all but one institution.

Table 4. Comparison of National Student Clearinghouse (NSC) and institution enrollment list counts1

Student type

Median absolute relative percent difference2

Percent of institutions with same QC results (compared to IPEDS counts)

Baccalaureate recipients

30.0

94.0

Undergraduate students

11.0

94.0

Graduate students

8.0

100.0

1 The comparison was done for 18 institutions that had sufficient data.

2 The median absolute relative percent difference is the median of the absolute value of (institution list count - NSC list count) / institution list count. Baccalaureate counts were included if the institution provided the baccalaureate indicator on the list. Baccalaureate and graduate counts were included only for the 4-year institutions.

NOTE: QC = Quality Control; IPEDS = Integrated Postsecondary Education Data System.

SOURCE: U.S. Department of Education, National Center for Education Statistics, 2023–24 National Postsecondary Student Aid Study (NPSAS:24) Field Test.


Additionally, the NSC lists were examined to determine if there were sufficient non-missing data for the most important data elements for student sampling and contacting, as shown in Table 5. The percentage of data received from NSC was generally high, except for race/ethnicity. Also, the NSC data received for the field test subset of institutions were more complete than the institution data for both Social Security number and baccalaureate indicator.

Table 5. Percentage of non-missing important data elements

Data element

Percent of non-missing data

From NSC

From institutions

Social Security number

96.1

88.5

Date of birth

99.9

99.9

Race/ethnicity

9.6

86.2

Degree program

92.4

100.0

Received bachelor’s since July 1

100.0

65.1

Contact information

100.0

100.0

NOTE: NSC = National Student Clearinghouse.

SOURCE: U.S. Department of Education, National Center for Education Statistics, 2023–24 National Postsecondary Student Aid Study (NPSAS:24) Field Test.

While the NSC lists contained data that were fairly comparable to the lists provided by the institutions, there are some noteworthy disadvantages to the NSC lists, including the low amount of race/ethnicity data; the inclusion of dually enrolled high school students, especially for less-than-4-year institutions; and the lack of data elements to identify students expected to receive their bachelor’s degree before June 30.

As discussed in sections B.2, we plan to obtain enrollment lists from NSC for some institutions refusing to provide enrollment lists but willing to provide permission for us to obtain their lists from NSC. Use of NSC lists will help us obtain student sample size targets, as well as reduce institution burden and help with refusal conversion.

b. Bachelor’s degree recipient proxy

As discussed in section B.2, we developed a baccalaureate proxy in the field test to identify students who had not yet received their bachelor’s degree but were expected to receive it by June 30. The institutions provided a baccalaureate indicator on the enrollment lists to flag whether students have completed requirements for or received a bachelor’s degree between July 1 and the date the enrollment list is provided. We used this flag and the proxy to identify students to sample as potential baccalaureate recipients. For use in evaluating the proxy, we additionally requested that institutions provide a second indicator on the lists to flag students who had not yet received their bachelor’s degree but were expected to receive it by June 30.

We have analyzed preliminary field test data to begin evaluating the baccalaureate proxy. Table 6 shows that about 51 percent of students who were both sampled as baccalaureates using the proxy and surveyed were confirmed to have received or expected to receive their bachelor’s degree between July 1, 2022, and June 30, 2023. This is a false positive rate of 49 percent (100-51).


Table 6. Baccalaureate status determination, by student type1

Student type

Students surveyed in field test

Confirmed B&B eligibility

Number

Unweighted percent

Total

3,470

1,260

36.4

Total undergraduate

3,270

1,260

38.4

Potential Baccalaureate

2,320

1,240

53.5

Identified on student lists2

200

170

84.7

Identified by proxy definition

2,110

1,070

50.5

Other undergraduate

950

20

1.7

Graduate

200

#

2.0

# Rounds to zero.

2 Identified on student lists using baccalaureate indicator on the enrollment lists for students who have completed requirements for or received a bachelor’s degree between July 1 and the date of the enrollment list.

NOTE: B&B = Baccalaureate and Beyond Longitudinal Study. Detail may not sum to totals due to rounding.

SOURCE: U.S. Department of Education, National Center for Education Statistics, 2023–24 National Postsecondary Student Aid Study (NPSAS:24) Field Test.

Table 7 shows that the baccalaureate proxy matches the expected baccalaureate flag from the enrollment lists for about 61 percent of the surveyed students. The baccalaureate proxy correctly identifies about 67 percent of baccalaureate recipients, as compared to the survey, and the expected baccalaureate flag from the enrollment lists correctly identifies about 80 percent of baccalaureate recipients, as compared to the survey.

At the end of field test student data collection, we began a final evaluation of the baccalaureate proxy, including determining if different models would perform better, based on field test data. Given the equivocal results, we plan to use the proxy to supplement, not replace, institutions reporting an unknown degree status.



Table 7. Counts and percentages of surveyed students by baccalaureate status1

Confirmed baccalaureate recipient in survey2

Baccalaureate recipient by proxy

Baccalaureate recipient by expected flag on lists3

Number of surveyed students

Percent of surveyed students

Yes

Yes

Yes

760

23.3

Yes

Yes

No

310

9.5

Yes

No

Yes

10

0.3

Yes

No

No

10

0.3

No

Yes

Yes

210

6.4

No

Yes

No

840

25.7

No

No

Yes

130

3.9

No

No

No

1,000

30.7

2 Baccalaureate recipients not confirmed in the survey include responses of no and missing responses.

3 Baccalaureate recipients not flagged as expected on the list include responses of no, does not apply, and unknown and missing responses.

SOURCE: U.S. Department of Education, National Center for Education Statistics, 2023–24 National Postsecondary Student Aid Study (NPSAS:24) Field Test.



  1. Incentive Experiment

During the field test, we conducted an experiment to determine the optimal amount and delivery timing of incentives, given fixed overall data collection costs (e.g., the costs of incentives, nonresponse follow-up). The field test sample of 6,000 students was divided at random into three experimental groups. Group 1 received an initial incentive offer of $25, followed by an offer of an additional $10 to those who remained nonrespondents after four weeks of data collection (i.e., a boosted incentive). Group 2 received an initial offer of $25 as well, but with an additional $20 boost offered to nonrespondents at the eight-week point. Group 3, a control group, received an incentive offer of $30 throughout the entire period of data collection. We found no significant advantage of the incentive boost or timing of the boost over the control condition ($30 promised incentive). We also found no differences across key demographics across the experimental groups; nor cost per complete advantages (see Appendix I for detailed results).


As indicated in section B.3., sample members will be offered $30 for a completed survey, in line with the control condition from the field test incentive experiment. While the two experimental groups in the NPSAS:24 field test, which initially offered sample members $25, achieved numerically higher response rates (58.5 percent for the +$10 boost and 58.9 percent for the +$20 boost relative to the 57.0 percent for the $30 control condition), a $25 incentive without a boost was not tested. Figure 2 provides some limited insight into the performance of the $25 incentive before a boost is introduced. While response rates were fairly comparable to that of the $30 control group, the $25 groups did not show the same sustained level of participation except after the offer of the boost.


The field test sample size, originally specified as 3,500, was increased to 6,000 sample members to accommodate the incentive experiment with 3 equally sized groups of 2,000 but, due to cost constraints, could not be increased further to detect smaller differences in groups. The design yielded sufficient power to detect differences in response rates across conditions of no less than 4 percentage points at alpha=0.05 and 80% power; thus, we failed to reject the null hypothesis of no difference across conditions with observed differences of 1.5 and 1.9 percentage points. Consequently, we recommend using the control condition incentive amount for the NPSAS:24 data collection but monitoring data collection throughout for groups that are underrepresented in the respondent pool and may benefit from an incentive boost. As noted in part A.9, our experience on NPSAS:20


Figure 2. NPSAS:24 field test completes, by experimental group





showed that a small boost of $10 can yield an overall increase in response rates, over projected response rates, of 17.5 percent. By data collection wave, the boost resulted in increases over expected rates ranging from 8.05 percent to 25.19 percentage points.18



    1. Reviewing Statisticians and Individuals Responsible for Designing and Conducting the Study

NPSAS:24 is being conducted by NCES. The following statisticians at NCES are responsible for the statistical aspects of the study: Dr. Tracy Hunt-White, Dr. David Richards, Dr. Sean Simone, and Dr. Chris Chapman. NCES’s prime contractor for NPSAS:24 is RTI International (Contract# 91990022C0017), and subcontractors include Activate Research; ARSIEM Corporation; EurekaFacts; Forum One Communications; HR Directions; KEN Consulting, Inc.; Leonard Resource Group; Research Support Services; Strategic Communications, Inc.; The Equity Paradigm; and Whitworth Kee Consulting, LLC. Dr. Anthony Jones, Dr. Vincent Castano, Dr. Eric Atchison, Richard Reeves, Dr. Sandy Baum, Dr. Matt Springer, and Dr. Shelly Steward are consultants on the study. The following staff members at RTI are working on the statistical aspects of the study design: Dr. Jennifer Wine, Dr. Josh Pretlow, Peter Siegel, Stephen Black, Ruby Johnson, Jennifer Cooney, Dr. T. Austin Lacy, and Dr. Emilia Peytcheva. Principal professional RTI staff, not listed above, who are assigned to the study include: Kristin Dudley, Jamie Wescott, Ashley Wilson, Austin Caperton, Jeff Franklin, Dr. Jerry Timbrook, and Dr. Erin Velez.



1 The U.S. service academies (the U.S. Air Force Academy, the U.S. Coast Guard Academy, the U.S. Military Academy, the U.S. Merchant Marine Academy, and the U.S. Naval Academy) are not eligible for this financial aid study because of their unique funding/tuition base.

2 A Title IV eligible institution is an institution that has a written agreement (program participation agreement) with the U.S. Secretary of Education that allows the institution to participate in any of the Title IV federal student financial assistance programs other than the State Student Incentive Grant and the National Early Intervention Scholarship and Partnership programs.

3 The end date is March 31, 2024 for continuous enrollment institutions.

4 The GED® credential is a high school equivalency credential earned by passing the GED® test, which is administered by GED Testing Service. For more information on the GED test and credential, see https://ged.com/about_test/test_subjects/.

5 Chromy, J.R. (1979). Sequential Sample Selection Methods. In Proceedings of the Survey Research Methods Section of the American Statistical Association (pp. 401–406). Alexandria, VA: American Statistical Association.

7 Folsom, R.E., Potter, F.J., and Williams, S.R. (1987). Notes on a Composite Size Measure for Self-Weighting Samples in Multiple Domains. In Proceedings of the Section on Survey Research Methods of the American Statistical Association. Alexandria, VA: American Statistical Association, 792–796.

8 Self-weighting samples have equal weights within sampling domains.

9 A Hispanic-serving institutions indicator is no longer available from IPEDS, so we will create an HSI proxy following the definition of HSI as provided by the U.S. Department of Education (https://www2.ed.gov/programs/idueshsi/definition.html) and using IPEDS Hispanic or Latino enrollment data.

10 We used the 2018 version of Carnegie classification available on the IPEDS files.

11 Many institutions know their enrolled students prior to April 30 and provide lists in February, March, or April. However, continuous enrollment institutions, including many of the for-profit institutions, typically cannot provide enrollment lists until mid-May, at the earliest, given that the lists include students enrolled through April 30.

12 Oversampled veterans in this document refer to veterans who receive veteran’s benefits.

13 30,000 = (53,043 potential baccalaureate recipients * .95 * .70 * .803) + (83,957 other undergraduate students * .95 * .70 * .03), where .803 = 1-.197. 

The false-positive and false-negative rates will be updated for the full-scale based on field test results. The false negative rate will also be updated to account for a small percentage (less than two percent) of students who are sampled as graduate students but are baccalaureate recipients.

14 Groves R.M., and Heeringa, S.G. (2006). Responsive Design for Household Surveys: Tools for Actively Controlling Survey Errors and Costs. Journal of the Royal Statistical Society Series A-Statistics in Society, 169(3): 439-457.

15 Singer, E. and Ye, C. (2013). The Use and Effects of Incentives in Surveys. Annals. Annals of the American Academy of Political and Social Science, 645(1): 112-141.

16 It is possible all nonresponding oversampled cases are eligible for the $10 boost.

17 The comparison was done for 18 institutions that had sufficient data on degree program to compare counts by student type.

18 Memorandum summarizing results submitted by NCES to OMB on April 20, 2021.


File Typeapplication/vnd.openxmlformats-officedocument.wordprocessingml.document
File TitleChapter 2
Authorspowell
File Modified0000-00-00
File Created2023-12-20

© 2024 OMB.report | Privacy Policy