2020/22 Beginning Postsecondary Students Longitudinal Study (BPS:20/22) Full-scale
Supporting Statement Part B
	
	
2020/22 Beginning Postsecondary Students (BPS:20/22) Full-Scale Study
Supporting Statement Part B
OMB # 1850-0631 v.19
Submitted by
National Center for Education Statistics
U.S. Department of Education
November 2021
Section 1 – Respondent Universe 1
Section 2 – Statistical Methodology 1
Section 3 – Methods for Maximizing Response Rate 7
Section 4 – Tests of Procedures and Methods 11
	
	
This submission requests clearance for the 2020/22 Beginning Postsecondary Students Longitudinal Study (BPS:20/22) full-scale study data collection materials and procedures. BPS:20/22 is the first follow-up of sample members from the 2019-20 National Postsecondary Student Aid Study (NPSAS:20) who began their postsecondary education during the 2019-20 (full-scale sample) or 2018-19 (field test sample) academic year. For details on the NPSAS:20 sampling design see NPSAS:20 Supporting Statement Part B (OMB# 1850-0666 v.25). Specific plans are provided below for the BPS:20/22 cohort.
Included in this section is information describing the respondent universe and any sampling or other respondent selection method that will be used.
The
	respondent universe for BPS:20/22 consists of all students who began
	their postsecondary education for the first time during the 2019-20
	academic year at any Title IV-eligible postsecondary institution in
	the United States. 
The BPS:20/22 full-scale cohort will
	be comprised of students who first enroll in postsecondary education
	after high school during the 2019-20 academic year. The BPS:20/22
	full-scale sample will include students from the NPSAS:20 full-scale
	sample who were identified as confirmed or potential 2019-20
	academic year first-time beginner students based on survey,
	institution, or other administrative data.
	
	
Statistical
	Methodology
The target population for the 2020/22
	Beginning Postsecondary Students Longitudinal Study (BPS:20/22)
	full-scale consists of all students who began their postsecondary
	education for the first time during the 2019–20 academic year
	at any Title IV-eligible postsecondary institution in the United
	States. Identification of the BPS:20/22 full-scale sample requires a
	multi-stage process that began, first, with selection of the 2019-20
	National Postsecondary Student Aid Study (NPSAS:20) full-scale
	sample of institutions and was followed next by selection of
	students within those institutions. The BPS:20/22 full-scale sample
	is comprised of students from the NPSAS:20 full-scale sample who
	were determined to be first-time beginners (FTBs), or were potential
	FTBs, as indicated by the NPSAS institution or administrative data.
	
NPSAS:20 Full-scale Sample
The NPSAS:20
	institution (first stage) sampling frame included all levels
	(less-than-2-year, 2-year, and 4-year) and control classifications
	(public, private nonprofit, and private for-profit) of nearly all
	Title IV eligible postsecondary institutions in the 50 states, the
	District of Columbia, and Puerto Rico. The institution sampling
	frame used institution data collected from various surveys of the
	Integrated Postsecondary Education Data System (IPEDS). An
	institution was NPSAS-eligible if, during the 2019-20 academic year,
	the institution:
•offered an educational program designed
	for persons who have completed secondary education;
•offered
	at least one academic, occupational, or vocational program of study
	lasting at least 3 months or 300 clock hours;
•offered
	courses that were open to more than the employees or members of the
	company or group (e.g., union) that administered the
	institution;
•was located in the 50 states, the District
	of Columbia, or Puerto Rico;
•was not a U.S. service
	academy (the U.S. Air Force Academy, the U.S. Coast Guard Academy,
	the U.S. Military Academy, the U.S. Merchant Marine Academy, and the
	U.S. Naval Academy) due to their unique funding/tuition base);
	and
•had a signed Title IV participation agreement with
	the U.S. Department of Education (an institution that has a written
	program participation agreement with the U.S. Secretary of Education
	that allows the institution to participate in any of the Title IV
	federal student financial assistance programs other than the State
	Student Incentive Grant and the National Early Intervention
	Scholarship and Partnership programs). 
The NPSAS:20
	institution sampling frame was constructed from the Integrated
	Postsecondary Education Data System (IPEDS) 2018-19 Institutional
	Characteristics Header, 2018-19 Institutional Characteristics,
	2017-18 12-Month Enrollment, and 2017 Fall Enrollment files.
The
	institution strata used for the sampling design were based on the
	following three sectors within each state and territory, for a total
	of 156 (52 x 3) sampling strata:
•public 2-year;
•public
	4-year (includes all eligible institutions that IPEDS classifies as
	public 4-year institutions, including those that are
	non–doctorate-granting, primarily sub-baccalaureate
	institutions);and
•all other institutions,
	including:
-public less-than-2 year;
-private nonprofit
	(all levels); and
-private for-profit (all levels).
The
	sample design allowed NPSAS:20 to have state-representative
	undergraduate student samples for public 2-year and public 4-year
	institutions as well as overall. From this point forward, the word
	“state” will refer to the 50 states, the District of
	Columbia, and Puerto Rico. In addition, the sample was nationally
	representative for both undergraduate and graduate students. The
	NPSAS:20 institution sample consisted of a census of all public
	2-year and all public 4-year institutions and a random sample of
	institutions from the “all other institutions” stratum.
	Within the “all other institutions” stratum, the goal
	was to sample at least 30 institutions per state, when there are 30
	institutions in this stratum in a state, so that institutions in
	this stratum were sufficiently represented within the state and
	national samples. 
The following criteria were used to
	determine institution sample sizes within the “all other
	institutions” stratum:
•In states with 30 or
	fewer institutions in the “all other institutions”
	strata, a census of these institutions was selected. 
•In
	states with more than 30 institutions in the “all other
	institutions” strata and where selecting only 30 institutions
	would result in a very high sampling fraction, we selected a census
	of institutions. We arbitrarily chose 36 institutions as the cutoff
	to avoid high sampling fractions. This cutoff resulted in taking a
	census of institutions in states that had between 31 and 36
	institutions in the “all other institutions” strata.
	Based on the latest IPEDS data, there were only three states
	(Mississippi, Nebraska, and Nevada) that had between 31 and 36
	institutions in the “other” stratum. 
•In
	states with more than 36 institutions in the “all other
	institutions” strata, a sample of 30 institutions was
	selected.
Within the “all other institutions”
	stratum, institutions were selected using sequential probability
	minimum replacement (PMR) sampling (Chromy 1979), which resembles
	stratified systematic sampling with probabilities proportional to a
	composite measure of size. This is the same methodology that has
	been used since NPSAS:96. Institution measure of size was determined
	using undergraduate and graduate student enrollment counts and FTB
	counts from the IPEDS 2017-18 12-Month Enrollment and 2017 Fall
	Enrollment files, respectively (OMB# 1850-0666 v.25.). Composite
	measure of size (Folsom, Potter, and Williams 1987) sampling was
	used to ensure that target sample sizes were achieved within
	institution and student sampling strata, while also achieving
	approximately equal student weights across institutions. All
	eligible students from sampled institutions comprised the student
	sampling frame.
Within the “all other institutions”
	stratum, additional implicit stratification was accomplished by
	sorting the sampling frame by the following classifications, as
	appropriate (OMB# 1850-0666 v.25.): 
•Control and level of
	institution;
•Historically Black Colleges and Universities
	(HBCUs) indicator; 
•Hispanic-serving institutions (HSIs)
	indicator (no longer available from IPEDS, so we created an HSI
	proxy following the definition of HSI as provided by the U.S.
	Department of Education
	(https://www2.ed.gov/programs/idueshsi/definition.html) and using
	IPEDS Hispanic enrollment data); 
•Carnegie
	classifications of postsecondary institutions; and
•the
	institution measure of size. 
The objective of this implicit
	stratification was to approximate proportional representation of
	institutions on these measures.
From the approximately
	3,110 institutions selected for the NPSAS:20 full-scale data
	collection, 98 percent met eligibility requirements; of those,
	approximately 72 percent provided enrollment lists.
The
	second stage of the NPSAS:20 sample specification was the selection
	of a stratified sample of individuals within sampled institutions.
	NPSAS-eligible undergraduate and graduate students were those who
	were enrolled in the NPSAS institution in any term or course of
	instruction between July 1, 2019 and April 30, 2020 and who
	were:
•enrolled in either (1) an academic program; (2) at
	least one course for credit that could be applied toward fulfilling
	the requirements for an academic degree; (3) exclusively noncredit
	remedial coursework but had been determined by their institution to
	be eligible for Title IV aid; or (4) an occupational or vocational
	program that requires at least 3 months or 300 clock hours of
	instruction to receive a degree, certificate, or other formal award;
	and
•not concurrently enrolled in high school; and
•not
	enrolled solely in a General Educational Development (GED®) or
	other high school completion program. The GED® credential is a
	high school equivalency credential earned by passing the GED®
	test, which is administered by GED Testing Service (
	http://www.gedtestingservice.com/ged-testing-service).
There
	were 11 student sampling strata as follows:
•undergraduate
	students who were potential FTBs; 
•other undergraduate
	students not classified as potential FTBs;
•graduate
	students who were veterans;
•master's degree students in
	science, technology, engineering, and mathematics (STEM)
	programs;
•master's degree students in education and
	business programs;
•master's degree students in all other
	programs;
•doctoral-research/scholarship/other graduate
	students in STEM programs;
•doctoral-research/scholarship/other
	graduate students in education and business
	programs;
•doctoral-research/scholarship/other graduate
	students in other programs;
•doctoral-professional
	practice students; and
•other graduate students not
	captured in the above categories.
When students met the
	criteria to be classified into multiple stratum, they were assigned
	following the list in hierarchical order (e.g., a STEM master's
	student who was also a veteran would fall into the “graduate
	students who are veterans” category). Several student
	subgroups were sampled at rates different than their natural
	occurrence within the population due to specific analytic
	objectives. The following groups were oversampled:
•undergraduate
	students who are potential FTBs;
•graduate students who
	are veterans;
•master's degree students in STEM
	programs;
•doctoral-research/scholarship/other graduate
	students in STEM programs; and
•master's degree students
	enrolled in for-profit institutions.
The NPSAS:20
	full-scale sample was randomly selected from the frame with students
	sampled at fixed rates according to student sampling strata and
	institution sampling strata. Sample yield was monitored, and
	sampling rates were adjusted when necessary. Sampling rates were
	adjusted to maintain sample yield targets, such as when institution
	enrollment lists were trending larger or smaller than expected. The
	full-scale sample achieved a size of approximately 380,100 students;
	about 173,360 of which were asked to complete a survey, and around
	206,740 of which were not asked to complete a survey. The
	administrative sample of 380,100 students was randomly selected
	first and the 173,360 students requested to complete a survey were
	subsampled from the larger administrative sample. The latter group
	of 206,740 are referred to as an admin-only sample since only
	administrative data were collected for them (they were not fielded
	for the survey data collection). Student records and administrative
	data were attempted to be collected for all sampled students.
	
Identification of FTBs
Correctly classifying
	FTBs is important because unacceptably high rates of
	misclassification (i.e., false positives and false negatives) can
	and have resulted in (1) excessive cohort loss with too few eligible
	sample members to sustain the longitudinal study, (2) excessive cost
	to “replenish” the sample with little value added, and
	(3) inefficient sample design (excessive oversampling of “potential”
	FTBs) to compensate for anticipated misclassification error. To
	address this concern, participating institutions were asked to
	provide additional information for all eligible students and
	matching to administrative databases was utilized to further reduce
	false positives and false negatives prior to sample selection.
In
	addition to an FTB indicator, we requested that enrollment lists
	provided by institutions (or institution systems) include degree
	program, class level, date of birth, enrollment in high school (or
	completion program) indicator, and high school completion date.
	Students identified by the institution as FTBs, but also identified
	as in their third year or higher and/or not an undergraduate
	student, were not classified as potential FTBs for sampling.
	Additionally, students who were dually enrolled at the postsecondary
	institution and in high school based on the enrollment in high
	school (or completion program) indicator and the high school
	graduation date were not eligible for sampling. If the FTB indicator
	was not provided for a student on the list but the student was 18
	years old or younger and did not appear to be dually enrolled, the
	student was classified as a potential FTB for sampling. Otherwise,
	if the FTB indicator was not provided for a student on the list and
	the student was over the age of 18, then the student was sampled as
	an “other undergraduate,” (but such students would be
	included in the BPS cohort if identified during the student survey
	as an FTB).
Prior to sampling, all students listed as
	potential FTBs were matched to National Student Loan Data System
	(NSLDS) records to determine if any had a federal financial aid
	history pre-dating the NPSAS year (earlier than July 1, 2019). Since
	NSLDS maintains current records of all Title IV grant and loan
	funding, any students with data showing disbursements from prior
	years could be reliably excluded from the sampling frame of FTBs.
	Given that about 68 percent of FTBs receive some form of Title IV
	aid in their first year, this matching process could not exclude all
	listed FTBs with prior enrollment but significantly improved the
	accuracy of the listing prior to sampling, yielding fewer false
	positives. 
Simultaneously with NSLDS matching, all
	potential FTBs were also matched to the Central Processing System
	(CPS) to identify students who, on their Free Application for
	Federal Student Aid (FAFSA), indicated that they had attended
	college previously. After NSLDS and CPS matching, a subset of the
	remaining potential FTBs were matched to the National Student
	Clearinghouse (NSC) for further narrowing of FTBs based on the
	presence of evidence of earlier enrollment. Due to the cost of
	matching individuals to the NSC we only targeted individuals in
	institution sectors that historically had high false positive rates.
	Potential FTBs over the age of 18 in the public 2-year and
	for-profit sectors were targeted for this match because these
	sectors either had high false-positive rates in NPSAS:12 or had
	large NPSAS:20 sample sizes.
In setting the NPSAS:20 FTB
	selection rates, we considered the false-positive rates, based on
	the NPSAS:12 survey, which had an overall unweighted false positive
	weight of 22 percent and an unweighted false negative rate of 4.6
	percent. NPSAS:12 was examined as the reference as it was the most
	recent NPSAS administration with a BPS cohort. Based on confirmed
	FTB status from the NPSAS:20 survey, we found the observed false
	positive rate of the final potential FTB indicator to be 21.7
	percent with a false negative rate of 2.2 percent. These rates are
	similar to those observed from NPSAS:12 which followed the same
	administrative matching approach to refine the institution-provided
	FTB flags. 
After NPSAS:20 data collection, the potential
	FTB flag was further refined by matching all potential FTBs to the
	NSLDS again to catch any individuals who did not match through the
	early cycle (generally due to missing student identifiers at the
	time of the first match). In addition, this second NSLDS match was
	used to remove any students who responded to the NPSAS:20 student
	survey and self-identified as an FTB, but the NSLDS match indicated
	otherwise.
BPS:20/22 Full-Scale Sample
NPSAS:20
	consisted of an administrative sample of 380,100 students. The
	NPSAS:20 survey sample was a subset of approximately 173,360
	students from the total student sample. The remaining individuals
	who were part of the administrative sample but were not selected for
	the survey sample are referred to as administrative-only (hereafter
	referred to as ‘admin-only') sample or students.
	For the BPS:20/22 cohort, the full-scale sampling frame is based on
	the larger NPSAS:20 administrative sample. This frame contains
	survey confirmed FTBs as well as survey nonrespondent and admin-only
	potential FTBs. 
The BPS:20/22 frame only includes
	confirmed and potential FTB students who are defined as study
	respondents in NPSAS:20 and who were not found to be deceased during
	NPSAS:20 data collection. A study respondent is any individual who
	is a survey or administrative student respondent. All confirmed FTB
	students are also NPSAS:20 study respondents since NPSAS survey
	respondents are also NPSAS:20 study respondents. The BPS:20/22
	full-scale sample consists of three different groups based on their
	NPSAS:20 sample along with their administrative and survey response
	statuses. The groups are (1) NPSAS:20 survey respondents who are
	confirmed FTBs, (2) NPSAS:20 survey nonrespondents who are potential
	FTBs and NPSAS administrative student respondents, and (3) NPSAS:20
	admin-only students who are potential FTBs and NPSAS administrative
	student respondents. As stated earlier, all three groups are
	considered study respondents. Table 1 shows the distribution of the
	BPS:20/22 frame. 
	
Table 1. BPS:20/22 full-scale frame by NPSAS:20 sample and response status
Some
	NPSAS:20 admin-only potential FTB students (approximately 370) who
	are considered administrative student respondents were sampled from
	institutions that were unable to provide accurate enrollment list
	information necessary for student survey contacting. These students
	were therefore not able to be included in the student survey portion
	of the sample and will not be included in the BPS:20/22 sampling
	frame. These institutions were not found to significantly differ
	from the remaining institutions or admin-only potential FTBs and
	thus should not impact sample representativeness. The decision to
	include admin-only FTB students is based on two primary factors: 1)
	field test data showing significantly higher response rates for the
	admin-only group compared to survey nonrespondents and 2) analysis
	results suggesting minimal loss of precision which will still allow
	the sample to meet precision goals. 
The BPS:20/22
	full-scale sample consists of approximately 37,330 total students
	including approximately 26,470 NPSAS:20 survey respondents and
	approximately 10,860 potential FTB students who did not complete the
	survey but who were administrative student respondents. The
	potential FTB students are split nearly evenly between NPSAS:20
	survey nonrespondents and admin-only students. 
In
	addition, the BPS:20/22 sample is designed with a goal to be state
	representative for a subset of states. The minimum desired sample
	sizes for each state subgroup are based on the ability to measure a
	relative change of 20 percent in proportions across rounds and an
	assumed design effect of 3. To meet this goal, 741 survey
	respondents per desired state are necessary by the end of BPS:20/25
	data collection.
The base NPSAS:20 administrative sample
	that BPS:20/22 follows was designed to be state representative. The
	distribution of confirmed and potential FTBs from NPSAS:20 was
	reviewed by state to arrive at the final decision to target
	representativeness for California, Florida, Georgia, New York, North
	Carolina, Pennsylvania, and Texas. An oversample of approximately
	330 students is required to achieve the desired sample size for
	Georgia. The decision to oversample students for Georgia was made as
	Georgia was close to the target sample size, had enough additional
	administrative student respondents to be oversample and was
	considered an important state to target.
All NPSAS:20
	survey respondents who are confirmed FTBs are sampled with
	certainty. The approximately 25,440 potential FTBs who did not
	complete a survey, but who are administrative respondents, are
	explicitly stratified by “survey nonrespondent” and
	“admin-only student” as well as control and level of
	institution. A sample of approximately 10,530 students is split
	evenly between survey nonrespondents and admin-only students and was
	proportionally allocated across control and level of institution.
	Additionally, the oversample of approximately 330 students for
	Georgia is allocated by sampling all potential FTBs within Georgia.
	Within the explicit strata, a simple random sample of students is
	selected. Table 2 displays the final BPS:20/22 sample by control and
	level of institution including the oversample of approximately 330
	students within Georgia while table 3 details the sample by state.
	
	
Table 2. BPS:20/22 sample by control and level of institution
| 
 | 
 | Confirmed and potential FTB administrative student respondents from NPSAS:20 Administrative Sample | ||
| Institution characteristics | Total | NPSAS:20 Survey Respondents | NPSAS:20 Survey Nonrespondents | NPSAS:20 Admin-Only Respondents | 
| Total | 37,330 | 26,470 | 5,510 | 5,350 | 
| 
					 | 
					 | 
					 | 
					 | 
					 | 
| Institution type | 
					 | 
					 | 
					 | 
					 | 
| Public | 
					 | 
					 | 
					 | 
					 | 
| Less-than-2-year | 400 | 270 | 120 | <5 | 
| 2-year | 12,370 | 9,000 | 2,210 | 1,160 | 
| 4-year non-doctorate-granting primarily sub-baccalaureate | 2,610 | 2,010 | 550 | 50 | 
| 4-year non-doctorate-granting primarily baccalaureate | 2,290 | 1,850 | 260 | 180 | 
| 4-year doctorate-granting | 7,840 | 4,750 | 790 | 2,300 | 
| Private nonprofit | 
					 | 
					 | 
					 | 
					 | 
| Less-than-4-year | 290 | 190 | 100 | <5 | 
| 4-year non-doctorate- granting | 2,940 | 2,000 | 220 | 720 | 
| 4-year doctorate-granting | 3,600 | 2,400 | 330 | 880 | 
| Private for-profit | 
					 | 
					 | 
					 | 
					 | 
| Less-than-2-year | 980 | 780 | 150 | 50 | 
| 2-year | 1,600 | 1,280 | 300 | 10 | 
| 4-year | 2,430 | 1,940 | 480 | <5 | 
| NOTE: Detail may not sum to totals because of rounding. Potential FTB’s are individuals who did not complete a NPSAS survey who appeared to be FTB’s in enrollment or NPSAS:20 admin data. SOURCE: U.S. Department of Education, National Center for Education Statistics, 2020/22 Beginning Postsecondary Students Longitudinal Study (BPS:20/22) Full-scale. | ||||
	
	
	
Table 3. BPS:20/22 sample by state
| 
 | 
 | Confirmed and potential FTB administrative student respondents from NPSAS:20 Administrative Sample | ||
| State | Total | NPSAS Survey Respondents | NPSAS Survey Nonrespondents | NPSAS Admin-Only Respondents | 
| Total | 37,330 | 26,470 | 5,510 | 5,350 | 
| 
					 | 
					 | 
					 | 
					 | 
					 | 
| California | 2,750 | 2,230 | 500 | 20 | 
| Florida | 1,970 | 1,580 | 350 | 40 | 
| Georgia1 | 1,410 | 860 | 390 | 170 | 
| North Carolina | 1,370 | 970 | 230 | 180 | 
| New York | 2,030 | 1,660 | 280 | 90 | 
| Pennsylvania | 1,340 | 940 | 170 | 230 | 
| Texas | 2,340 | 1,810 | 410 | 120 | 
| 
					 | 
					 | 
					 | 
					 | 
					 | 
| All other states | 24,120 | 16,410 | 3,190 | 4,510 | 
| 1 Georgia includes an oversample of an additional 250 NPSAS survey nonrespondents and 80 NPSAS admin-only respondents. NOTE: Detail may not sum to totals because of rounding. Potential FTB’s are individuals who did not complete a NPSAS survey who appeared to be FTB’s in enrollment or NPSAS:20 admin data. SOURCE: U.S. Department of Education, National Center for Education Statistics, 2020/22 Beginning Postsecondary Students Longitudinal Study (BPS:20/22) Full-scale. | ||||
Based
	on observations from the BPS:20/22 field test and previous
	full-scale BPS data collections, we expect a response rate of
	roughly 82 percent for NPSAS:20 survey respondents, 22 percent for
	survey nonrespondents and approximately 57 percent for the
	admin-only students. Table 4 displays the BPS:20/22 full-scale
	sample size by NPSAS:20 data collection outcome group and the
	expected yield of completed interviews.
	
Table 4. BPS:20/22 expected completes by NPSAS:20 data collection outcome
| NPSAS:20 Outcome | Sample Size | Eligibility Rate | Expected Response Rate | Expected Completes | 
| Overall | 37,330 | 0.94 | 0.72 | 25,030 | 
| 
					 | 
					 | 
					 | 
					 | 
					 | 
| NPSAS Survey Respondents | 26,470 | 1.00 | 0.82 | 21,710 | 
| NPSAS Survey Nonrespondents | 5,510 | 0.78 | 0.22 | 950 | 
| NPSAS Admin-Only Respondents | 5,350 | 0.78 | 0.57 | 2,380 | 
| 
					NOTE:
					Detail may not sum to totals because of rounding.  | ||||
	
	
Included in this section is information describing the methods to be used to maximize response and to deal with issues of non-response.
Achieving
	high response rates in the BPS:20/22 full-scale study data
	collection will depend on successfully identifying and locating
	sample members and being able to contact them and gain their
	cooperation. As was used successfully in prior NCES longitudinal
	studies, shortly before data collection begins, we will send an
	address update/initial contact mailing/e-mail to remind sample
	members of their inclusion in the study. The following sections
	outline additional methods for maximizing response to the BPS:20/22
	full-scale data collection.
a.
	Tracing of Sample Members
To
	yield the maximum number of located cases with the least expense, we
	designed an integrated tracing approach with the following
	elements.
•Tracing activities conducted prior to the
	start of data collection will include batch database searches, such
	as to the National Change of Address (NCOA), for cases with enough
	contact information to be matched. To handle cases for which contact
	information is invalid or unavailable, project staff will conduct
	additional advance tracing through proprietary interactive databases
	to expand on leads found.
•Hard copy mailings,
	e-mails, and text messages will be used to maintain ongoing contact
	with sample members, prior to and throughout data collection. We
	will send a panel maintenance mailing in October 2021 to request
	that sample members update their contacting information (previously
	approved under the BPS:20/22 field test clearance package, OMB #
	1850-0631 v.18). 
•Immediately prior to the start of
	data collection, we will send an initial contact mailing to sample
	members to request that they update their contact information. A
	follow-up reminder e-mail will be sent approximately 2 weeks after
	that mailing to remind them to respond. In addition, we will send a
	letter to announce the start of data collection. The announcement
	will include a request that sample members complete the web survey
	and will provide each sample member a Study ID and password, the
	study website address, and a toll-free number to the help desk.
	Sample members who did not participate in the NPSAS:20 survey will
	receive $2 cash (or PayPal if a good address is not available) with
	the data collection announcement. After the data collection
	announcement mailing, an e-mail message with the same information
	will also be sent.
•The telephone locating and
	interviewing stage will include calling all available telephone
	numbers and following up on leads provided by parents and other
	contacts.
•The pre-intensive batch tracing stage
	consists of the LexisNexis SSN and Premium Phone batch searches that
	will be conducted between the telephone locating and interviewing
	stage and the intensive tracing stage.
•Once all
	known telephone numbers are exhausted, a case will move into the
	intensive tracing stage during which tracers will conduct
	interactive database searches using all known contact information
	for a sample member. With interactive tracing, a tracer assesses
	each case on an individual basis to determine which resources are
	most appropriate and the order in which each should be used. Sources
	that may be used, as appropriate, include credit database searches,
	such as Experian, various public websites, and other integrated
	database services.
•Other locating activities will
	take place as needed, including conducting a LexisNexis e-mail
	search for nonrespondents toward the end of data collection.
b.
	Training for Data Collection Staff
Telephone data
	collection will be conducted using the contractor's virtual call
	center, which allows the contractor to retain experienced staff who
	can successfully work from home. Telephone data collection staff
	will include Performance Team Leaders (PTLs) and Data Collection
	Interviewers (DCIs). Training programs, administered through Zoom,
	are critical to maximizing response rates and collecting accurate
	and reliable data.
PTLs, who are responsible for all
	supervisory tasks, will attend project-specific training for PTLs,
	in addition to the interviewer training. They will receive an
	overview of the study, background and objectives, and the data
	collection instrument through a question-by-question review. PTLs
	will also receive training in the following areas: providing direct
	supervision of virtual staff during data collection; handling
	refusals; monitoring interviews and maintaining records of
	monitoring results; problem resolution; case review; specific
	project procedures and protocols; reviewing reports generated from
	the ongoing Computer Assisted Telephone Interviewing (CATI); and
	monitoring data collection progress.
Training for DCIs is
	designed to help staff become familiar with and practice using the
	CATI case management system and the survey instrument, as well as to
	learn project procedures and requirements. Particular attention will
	be paid to quality control initiatives, including refusal avoidance
	and methods to ensure that quality data are collected. DCIs will
	receive project-specific training on telephone interviewing and
	answering questions from web participants regarding the study or
	related to specific items within the interview. Bilingual
	interviewers will receive a supplemental training that will focus on
	Spanish contacting and interviewing procedures. At the conclusion of
	training, all BPS data collection staff must meet certification
	requirements by successfully completing a certification interview.
	This evaluation consists of a full-length interview with project
	staff observing and evaluating interviewers, as well as an oral
	evaluation of interviewers' knowledge of the study's Frequently
	Asked Questions.
c. Case Management System
The
	BPS:20/22 full-scale survey will be conducted using a single
	web-based survey instrument for both web (including mobile devices)
	and CATI data collection. Data collection activities will be
	monitored through a CATI case management system, which is equipped
	with the numerous capabilities, including: online access to locating
	information and histories of locating efforts for each case; a
	questionnaire administration module with full “front-end
	cleaning” capabilities (i.e., editing based upon information
	obtained from respondents); sample management module for tracking
	case progress and status; and an automated scheduling module which
	delivers cases to interviewers. The automated scheduling module
	incorporates the following features:
•Automatic
	delivery of appointment and call-back cases at specified times
	reduces the need for tracking appointments and helps ensure punctual
	interviewing. The scheduler automatically calculates the delivery
	time of the case in reference to the appropriate time zone.
•Sorting
	non-appointment cases according to parameters and priorities set by
	project staff is another feature of the scheduling module. For
	instance, priorities may be set to give first preference to cases
	within certain sub-samples or geographic areas; or cases may be
	sorted to establish priorities based on prior round response status.
	Furthermore, the historic pattern of calling outcomes may be used to
	set priorities (e.g., cases with more than a certain number of
	unsuccessful attempts during a given time of day may be passed over
	until the next time period). These parameters ensure that cases are
	delivered to interviewers in a consistent manner according to
	specified project priorities.
•Groups of cases, or
	individual cases, may be designated for delivery to specific
	interviewers or groups of interviewers. This feature is most
	commonly used in filtering refusal cases, locating problems, or
	cases with language barriers that require interviewers with
	specialized skills.
•The scheduler tracks all outcomes for
	each case, labeling each with type, date, and time. These are easily
	accessed by the interviewer upon entering the individual case, along
	with interviewer notes.
•The scheduler can flag problem
	cases for supervisor attention. For example, refusal cases may be
	routed to supervisors for decisions about whether and when a refusal
	letter should be mailed, or whether another interviewer should be
	assigned.
•Complete reporting capabilities include default
	reports on the aggregate status of cases and custom report
	generation capabilities.
The integration of these
	capabilities reduces the number of discrete stages required in data
	collection and data preparation activities and increases
	capabilities for immediate error reconciliation, which results in
	better data quality and reduced cost. Overall, the scheduler
	provides an efficient case assignment and delivery function by
	reducing supervisory and clerical time, improving execution on the
	part of interviewers and supervisors by automatically monitoring
	appointments and call-backs, and reducing variation in implementing
	survey priorities and objectives.
d. Survey Instrument
	Design
The survey will employ a web-based instrument and
	deployment system, which has been in use since NPSAS:08. The system
	provides multimode functionality that can be used for
	self-administration, including on mobile devices, CATI, or data
	entry. The survey instrument can be found in Appendix E.
In
	addition to the functional capabilities of the case management
	system and web survey instruments described above, our efforts to
	achieve the desired response rate will include using established
	procedures proven effective in other large-scale studies. These
	include:
•Providing multiple response modes,
	including mobile-friendly self-administered and
	interviewer-administered options.
•Offering incentives to
	encourage response.
•Employing experienced DCIs who have
	proven their ability to contact and obtain cooperation from a high
	proportion of sample members.
•Training the DCIs
	thoroughly on study objectives, study population characteristics,
	and approaches that will help gain cooperation from sample
	members.
•Maintaining a high level of monitoring and
	direct supervision so that interviewers who are experiencing low
	cooperation rates are identified quickly and corrective action is
	taken.
•Making every reasonable effort to obtain a
	completed interview at the initial contact while allowing respondent
	flexibility in scheduling appointments to be interviewed.
•Providing
	assurance of confidentiality procedures, including requiring
	respondents to answer security questions before obtaining and
	resuming access to the survey and the survey automatically logging
	out of a session after 10 minutes of inactivity.
•Thoroughly
	reviewing all refusal cases and making special conversion efforts
	whenever feasible (see next section e).
e. Refusal
	Aversion and Conversion
Recognizing and avoiding refusals
	is important to maximize the response rate, and interviewer training
	will cover this and other topics related to obtaining cooperation.
	PTLs will closely monitor DCIs at the beginning of outbound calling,
	and provide re-training, as necessary. In addition, supervisors will
	review daily interviewer production reports produced by the CATI
	system to identify and retrain any DCIs who are producing
	unacceptable numbers of refusals or other problems.
Refusal
	conversion efforts will not be made with individuals who become
	verbally aggressive or who threaten to take legal or other action.
	Refusal conversion efforts will not be conducted to a degree that
	would constitute harassment. We will respect a sample member's right
	to decide not to participate and will not impinge this right by
	carrying conversion efforts beyond the bounds of propriety.
Included in this section is information describing any tests of procedures or methods that will be undertaken.
During
	the course of this data collection, the following experiment(s) will
	be undertaken. The BPS:20/22 field test included two sets of
	experiments: data collection experiments focused on survey
	participation to reduce nonresponse error and the potential for
	nonresponse bias, and questionnaire design experiments focused on
	minimizing measurement error to improve data quality. The full-scale
	data collection design described below will implement the tested
	approaches, revised based on the BPS:20/22 field test results
	described in Section 4.a. 
a. Summary of BPS:20/22 Field
	Test Data Collection Design and Results
The BPS:20/22
	field test contained two data collection experiments and two
	questionnaire design experiments. The results of these field test
	experiments are summarized below. For detailed results of the
	BPS:20/22 field test experiments, see Appendix D.
The
	data collection experiments explored the effectiveness of 1)
	offering an extra incentive for early survey completion, and 2)
	sending survey reminders via text messages. Results from these data
	collection experiments provide insight in preparation for the
	full-scale study regarding the effectiveness of these interventions
	across three data quality indicators: survey response
	(operationalized using response rates), sample representativeness
	(assessed across age, sex, ethnicity, race, and institutional
	control), and data collection efficiency (operationalized as the
	number of the days between the start of the experiment and survey
	completion). 
The “early bird” incentive
	experiment investigated the effectiveness of giving respondents an
	additional $5 incentive if they completed the survey within the
	first three weeks of data collection (experimental group) versus no
	additional incentive (control group). Response rates at the end of
	data collection did not differ across the early bird group (63.9
	percent) and the control group (63.4 percent; X2 = 0.08, p = .78).
	Both the early bird and control groups had similar
	representativeness across age, sex, ethnicity, race, and
	institutional control. At the end of data collection, respondents in
	the early bird group took significantly fewer days (28.1 days) than
	respondents in the control group (30.9 days) to complete the survey
	(t(2,231.7) = 2.09, p < 0.05). However, this difference is small
	(2.8 days), and not long enough to allow for any significant cost
	savings in the data collection process (e.g., via fewer reminder
	calls, texts, or mailings). Therefore, the use of an early bird
	incentive in the BPS:20/22 full-scale data collection is not
	recommended.
The reminder mode experiment compared the
	effectiveness of using text message reminders (experimental group)
	versus telephone call reminders (control group). Response rates at
	the end of data collection for the text message group (29.6 percent)
	and the telephone group (31.5 percent) did not significantly differ
	(X2 = 0.83, p = 0.36). In the telephone reminder group, the
	percentage of white respondents (74.1 percent) significantly
	differed from the percentage of white nonrespondents (64.5 percent;
	X2 = 7.40, p < 0.01), indicating a potential source of
	nonresponse bias. For the text message reminder group, there was not
	a significant difference between the percentage of White respondents
	(65.3 percent) and nonrespondents (63.3 percent) (X2 = 0.28, p =
	0.60), indicating better sample representativeness. The text message
	and telephone groups had similar representativeness across the
	remaining respondent characteristics: age, sex, ethnicity, and
	institutional control. 
Finally, the number of days it
	took for respondents in the text message reminder group to complete
	the survey (75.5 days) was not significantly different from the
	telephone reminder group (77.0 days; t(542.8) = 0.62, p = 0.27). As
	text message reminders achieved response rates, representativeness,
	and efficiency that was comparable to more expensive telephone
	reminders, the use of text reminders (coupled with telephone
	reminders as described in Section 4.b) is recommended as a part of
	the BPS:20/22 full-scale data collection.
The
	questionnaire design experiments explored the effectiveness of 1)
	different methods for collecting enrollment data, and 2) using a
	predictive search database on a question collecting address
	information. In addition, information about the impacts of the
	coronavirus pandemic was collected by randomly assigning respondents
	one of two separate topical modules to maximize the number of
	questions fielded without increasing burden. Results from these
	questionnaire collection experiments provide insight in preparation
	for the full-scale study regarding the effectiveness of these
	methods across three data quality indicators: missingness
	(operationalized as item- and question-level nonresponse rate),
	administrative data concordance (operationalized as agreement rates
	between self-reported enrollment and administrative records;
	month-level enrollment intensity experiment only), and timing burden
	(operationalized as the mean complete time at the question level).
	
The month-level enrollment intensity experiment compared
	two methods for collecting enrollment information in the 2020-21
	academic year: a single forced-choice grid question that displayed
	all enrollment intensities (i.e., full-time, part-time, mixed, and
	no enrollment) on one form (control group) and separate yes/no radio
	gates for full-time and part-time enrollment (experimental group).
	There were no statistically significant differences across the
	control and treatment conditions on rates of missingness (0 percent
	and 0.03 percent missing, respectively (t(636) = 1.42, p = 0.1575))
	or agreement rates with administrative enrollment data (70.0 percent
	agreement and 70.6 percent agreement, respectively (t(1311.1) =
	0.24, p = 0.8119)). On average, the treatment group took
	significantly longer to complete the enrollment question (17.2
	seconds) than the control group (10.5 seconds; t(1312.8) = 15.47, p
	< .0001), though this difference is expected given the additional
	screen respondents must navigate in the experimental group. As the
	experimental question did not represent a clear improvement over the
	original forced-choice grid, the use of the original question is
	recommended BPS:20/22 full-scale data collection. 
The
	predictive search address database experiment explored the utility
	of suggesting USPS-standardized addresses to respondents as they
	entered their address into the survey. This analysis compares
	address entry for the same set of respondents across the BPS:20/22
	field test (using the database-assisted predictive search method)
	and the NPSAS:20 full-scale survey (using traditional, manual
	address entry). Overall, 98 percent of respondents provided a
	complete permanent address using the manual method in NPSAS:20,
	compared to 85 percent using the predictive search method in
	BPS:20/22 field test (t(2023.5) = 13.88, p < .0001). However, it
	should be noted that addresses obtained using the predictive search
	method were error-free (0 percent of FTB check addresses were
	undeliverable), while 1.2 percent of addresses obtained using the
	manual method were undeliverable. Also, additional improvements to
	the survey instrument (e.g., soft check validations for incomplete
	addresses) may further reduce rates of missingness for the
	predictive search method. Finally, on average, respondents took
	longer to provide their address using manual entry (29.5 seconds)
	compared to the predictive search system (27.1 seconds; t(3055.7) =
	4.09, p < .0001). Given the higher quality data resulting from
	the predictive search method, the potential to improve the
	predictive search method via instrument adjustments, and the
	significant reduction in completion time compared to manual entry,
	the continuation of predictive search method is proposed for
	BPS:20/22 full-scale.
Given the impact of the coronavirus
	pandemic on higher education, researchers have expressed interest in
	using BPS:20/22 data to examine these impacts on postsecondary
	students. BPS:20/22 field test respondents were randomly assigned
	into two groups that received one of two modules. Each module
	measured similar constructs, however, module one consisted of survey
	questions from NPSAS:20 that measured student academic, social, and
	personal experiences related to the coronavirus pandemic, and module
	two collected a new set of constructs, including changes in
	enrollment and borrowing, changes in academic engagement, and access
	to support resources, that may be of analytic value to researchers
	and policymakers. Across both modules, the average item nonresponse
	rate was 2 percent. Module one had an average nonresponse rate of 3
	percent, significantly higher than the 0.6 percent nonresponse rate
	of module two (t(836.72) = 5.16, p < .0001). Regardless of module
	assignment, the coronavirus pandemic questions took respondents an
	average of 2.7 minutes to complete. The BPS:20/22 full-scale survey
	instrument will administer a subset of the questions from both field
	test coronavirus pandemic modules, based upon field test performance
	and TRP feedback. The coronavirus pandemic module for the full-scale
	maintains the burden goal of three minutes.
b. BPS:20/22
	Full-scale Data Collection Design
The data collection
	design proposed for the BPS:20/22 full-scale study builds on the
	designs implemented in past BPS studies, as well as the National
	Postsecondary Student Aid Study (NPSAS) and the Baccalaureate and
	Beyond (B&B) studies. Additionally, results from the BPS:20/22
	field test Data Collection Experiments (Appendix D) inform
	recommendations for the BPS:20/22 full-scale data collection design.
	
A primary goal of the full-scale design is to minimize
	the potential for nonresponse bias that could be introduced into
	BPS:20/22, especially bias that could be due to lower response rates
	among NPSAS:20 nonrespondents. Another important goal is to reduce
	the amount of time and cost of data collection efforts. 
To
	accomplish these goals, the plan is to achieve at least a 70 percent
	response rate. Doing so will minimize potential nonresponse bias,
	optimize statistical power, and enable sub-group analyses. The
	sample will be divided into two groups and differential data
	collection treatments will be implemented based on prior round
	response status. A similar approach was successfully implemented in
	the BPS:20/22 field test, and the latest B&B studies where more
	reluctant sample members received a more aggressive protocol (for an
	experimental comparison see B&B:16/17 field test
	https://nces.ed.gov/pubsearch/pubsinfo.asp?pubid=2020441). 
For
	the BPS:20/22 full-scale design, the following sample groupings will
	be used:
•NPSAS:20 survey respondents: Sample members who
	responded to NPSAS:20 and self-identified that they began their
	postsecondary education between July 1, 2019 and April 30, 2020 will
	receive a default data collection protocol (n = 26,470). 
•NPSAS:20
	survey nonrespondents and administrative-only cases: NPSAS:20
	administrative student respondents who are potential 2019-20
	academic year FTBs will receive an aggressive data collection
	protocol. This group includes NPSAS:20 survey nonrespondents (n =
	5,510) and NPSAS:20 administrative-only sample (cases who were never
	invited to complete the NPSAS:20 survey; n = 5,350) who are
	potential 2019-20 academic year FTBs based on administrative data.
	The goal of this treatment is to convert reluctant sample members
	(i.e., NPSAS:20 survey nonrespondents) and sample members who have
	never been contacted (i.e., administrative-only cases) to
	participate in the study as early in data collection as possible.
	Table 5 below presents the type and timing of interventions to be
	applied in data collection by groups and protocol. The details of
	these interventions are described below.
Table 5. 2020/22 Beginning Postsecondary Students Full-Scale data collection protocols, by data collection phase and group assignment
| 
					 | Data Collection Group Assignments | |
| 
					 | Default Protocol | Aggressive Protocol | 
| Sample | 
 | 
 | 
| Data Collection Protocols | ||
| Prior to data collection | 
 | 
 | 
| Early completion phase | 
 | 
 | 
| Production phase 1 | 
 | 
 | 
| Production phase 2 | 
 | 
 | 
| Nonresponse Conversion Phase | 
 | 
					 | 
| Total incentives | 
 
 
 | 
 | 
	
	
The
	duration of each phase of data collection will be determined based
	on phase capacity—the time at which a subgroup's estimates
	remain stable regardless of additional data collection efforts. For
	example, during the early completion phase, key metrics are
	continually monitored, and when they stabilize over a period of
	time, cases are then transferred to the next phase. Phase capacity
	will be determined based on a series of individual indicators within
	each data collection protocol. For example, response rates and other
	level of effort indicators over time accounting for covariates, such
	as institution control, will be assessed.
Incentives.
	The baseline incentive for the default protocol will be $30 with a
	$10 incentive boost in Production Phase 2, leading to a maximum
	possible total incentive of $40. The baseline incentive for the
	aggressive protocol will be $45. An experiment conducted in
	BPS:12/14 showed that a $45 baseline incentive yielded the highest
	response rates (Hill et al. 2016). However, this experiment was
	underpowered to detect differences from $30 in the lower propensity
	response groups (Wilson et al. 2015). Nonetheless, implementing a
	higher baseline incentive is recommended given the $30 baseline
	incentive and the $10 incentive boost from NPSAS:20 was not enough
	to encourage prior year nonrespondents to participate. Further, the
	$40 BPS:20/22 field test incentive yielded a response rate of only
	25.3 percent among these “aggressive protocol” sample
	members. The baseline incentive will be paid in addition to a
	possible $2 prepaid incentive (see prepaid incentive section below),
	and a $20 incentive boost (see nonresponse conversion incentive
	section below). The maximum possible total incentive is $67 in the
	aggressive data collection protocol. Results from the BPS:20/22
	field test showed that offering sample members an early bird
	incentive did not significantly improve response rates or
	representativeness by the end of data collection, nor did it
	practically improve data collection efficiency (Appendix D).
	Therefore, early bird incentives will not be used in the full-scale
	study for either the default or aggressive protocols.
Beyond
	the baseline incentives, both data collection protocols employ
	similar interventions, although the timing and intensity of these
	interventions differ across groups. Interventions occur sooner in
	the aggressive protocol and are more intense.
Prenotification.
	The first mailing that individuals in the default and aggressive
	data collection protocols will receive is a greeting card. This
	mailing is aimed to increase the perceived legitimacy of the
	upcoming survey request (e.g., Groves et al. 1992) in both data
	collection groups and announce the incentive amounts. Greeting
	cards, in particular, have been shown to significantly increase
	response rates in longitudinal studies (Griggs et al. 2019) and this
	method will be used as a precursor to the invitation letter. The
	greeting card will be mailed a few weeks in advance of data
	collection.
$2
	prepaid incentive.
	Cash prepaid incentives have been shown to significantly increase
	response rates in both interviewer-administered and
	self-administered surveys. These prepaid incentives increase the
	perceived legitimacy of the survey request and therefore reduce the
	potential for nonresponse bias (e.g., Church 1993; Cantor et al.
	2008; Goeritz 2006; Medway and Tourangeau 2015; Messer and Dillman
	2011; Parsons and Manierre 2014; Singer 2002). During the early
	completion phase in the B&B:16/17 field test, prepaid incentives
	($10 via check or PayPal) in combination with telephone prompting
	also significantly increased response rates by 4.4 percentage points
	in the aggressive protocol group. Given these positive findings
	combined with general recommendations in the literature (e.g.,
	Singer and Ye 2013; DeBell et al. 2019), a small $2 cash prepaid
	‘visible' incentive, or, where necessary due to
	low address quality, a $2 prepaid PayPal incentive announced on a
	separate index card will be sent to all cases in the aggressive
	protocol for BPS:20/22 full-scale (see results from the B&B:16/20
	calibration experiment – Kirchner et al. 2021). Sample members
	will be notified of this prepaid incentive in the data collection
	prenotification, and it will be included in the data collection
	announcement letter. 
Mode
	tailoring.
	The leverage-saliency theory suggests that respondents have
	different hooks that drive their likelihood of survey participation
	(Groves et al. 2000); thus, offering a person the survey mode (e.g.,
	web, mail, telephone) that they prefer may increase their likelihood
	of responding. This is further supported by empirical evidence that
	shows offering people their preferred mode speeds up their response
	and is associated with higher participation rates (e.g., Olson et
	al. 2012). Using the NPSAS:20 survey completion mode as a proxy for
	mode preference, the BPS:20/22 full-scale early completion phase
	will approach sample members in the default protocol with their mode
	of completion for NPSAS:20. Specifically, while all sample members
	in the default protocol will receive identical data collection
	announcement letters and e-mails, those who completed the NPSAS:20
	survey by telephone (4.3 percent) will be approached by telephone
	from the start of data collection. Likewise, those who completed the
	NPSAS:20 main study survey online will not be contacted by telephone
	before a preassigned outbound telephone data collection
	date.
(Light)
	outbound CATI calling and text messaging.
	The results from the BPS:20/22 field test showed that there were no
	statistically significant differences in the response rates for the
	text message reminder and the telephone only group at the end of the
	experimental period (Appendix D). As a result, both data collection
	groups will receive early text message reminders combined with
	prioritized telephone calls. Telephone calls will be prioritized to
	individuals for whom no cell phone number exists, those who opt out
	of the text message reminders, and those sample members who will be
	prioritized based on other criteria (e.g., from lower performing
	sectors). Text messages from sample members will be answered with an
	automated text response, with the possibility of two-way text
	messaging (i.e., interviewers respond to text message questions sent
	by sample members) in some cases. 
Sample members in the
	default group who qualify for telephone calls will receive a light
	CATI protocol. Light CATI involves a minimal number of phone calls,
	used mainly to prompt web response (as opposed to regular CATI
	efforts that involve more frequent phone efforts, with the goal to
	locate sample members and encourage their participation). In the
	B&B:16/17 field test, introduction of light CATI interviewing
	appeared to increase production phase response rates in the default
	protocol. Although one should use caution when interpreting these
	results – group assignment in B&B:16/17 field test was not
	random but instead compared NPSAS:16 “early” and “late”
	respondents– the findings are consistent with the literature
	which has shown that web surveys tend to have lower response rates
	compared to interviewer-administered surveys (e.g., Lozar Manfreda
	et al. 2008). Attempting to survey sample members by telephone also
	increases the likelihood of initiating locating efforts sooner.
	B&B:16/17 field test results showed higher locate rates in the
	default protocol (93.7 percent), which had light CATI, compared to a
	more relaxed protocol without light CATI (77.8 percent; p <
	0.001). For the BPS:20/22 full-scale data collection, light CATI
	will be used in the default protocol once CATI begins in Production
	Phase 1. Additionally, all cases in the aggressive protocol will
	receive earlier and more intense telephone prompting than eligible
	cases in the default group.
Incentive
	boosts.
	Researchers have commonly used incentive boosts as a nonresponse
	conversion strategy for sample members who have implicitly or
	explicitly refused to complete the survey (e.g., Groves and Heeringa
	2006; Singer and Ye 2013). These boosts are especially common in
	large federal surveys during their nonresponse follow-up phase
	(e.g., The Center for Disease Control and Prevention's National
	Survey of Family Growth) and have been implemented successfully in
	other postsecondary education surveys (e.g., HSLS:09 second
	follow-up; BPS:12/17; NPSAS:20). In NPSAS:20, a $10 incentive boost
	increased the overall response rate by about 3.2 percentage points
	above the projected response rate. Therefore, a $10 incentive boost
	increase to the BPS:20/22 baseline incentive is planned during
	Production Phase 2 for all remaining nonrespondents in the default
	data collection protocol, before the abbreviated survey is offered
	in the nonresponse conversion phase. Remaining nonrespondents in the
	aggressive data collection protocol will be offered a $20 incentive
	boost increase to the baseline incentive before the abbreviated
	survey (both offered in Production Phase 2). This is because the $10
	incentive boost in NPSAS:20 did not show any effect on this group.
	If necessary, incentive boosts may be targeted only at certain
	groups of nonrespondents to achieve response goals (e.g., targeting
	nonrespondents from certain states to ensure representativeness,
	targeting aggressive group nonrespondents to reduce the potential
	for nonresponse bias).
Abbreviated
	survey.
	Obtaining responses from all sample members is an important
	assumption of the inferential paradigm. The leverage-saliency theory
	(Groves et al. 2000) and the social exchange theory (Dillman et al.
	2014) suggest that the participation decision of an individual is
	driven by different survey design factors or perceived cost of
	participating. As such, reducing the perceived burden of
	participating by reducing the survey length may motivate sample
	members to participate.
During the B&B:16/17 field
	test, prior round nonrespondents were randomly assigned to one of
	two groups: 1) prior round nonrespondents who were offered the
	abbreviated survey during the production phase (i.e., before the
	nonresponse conversion phase), and 2) prior round nonrespondents who
	were offered the abbreviated survey during the nonresponse
	conversion phase (i.e., after the production phase). At the end of
	the production phase, prior round nonrespondents who received the
	abbreviated survey had a higher overall response rate (22.7 percent)
	than those who were not offered the abbreviated during that phase
	(12.1 percent; t(2,097) = 3.67, p < 0.001). Further, at the end
	of data collection, prior round nonrespondents who were offered the
	abbreviated survey during the earlier production phase had a
	significantly higher response rate (37 percent) than prior round
	nonrespondents who were not offered the abbreviated survey until the
	nonresponse conversion phase (25 percent) (t(2,097) = 3.52, p
	=.001). These results indicate that offering an abbreviated survey
	to prior round nonrespondents during the production phase (i.e.,
	earlier in data collection) significantly increases response rates.
	The B&B:08/12 and B&B:08/18 full-scale studies also
	demonstrated the benefit of an abbreviated survey. Offering the
	abbreviated survey to prior round nonrespondents increased overall
	response rates of that group by 18.2 (B&B:08/12) and 8.8
	(B&B:08/18) percentage points (Cominole et al. 2015). In
	NPSAS:20, 14.4 percent of those offered the abbreviated survey
	completed it. Therefore, an abbreviated survey option will be
	offered to all sample members in the BPS:20/22 full-scale study. For
	the aggressive protocol, the abbreviated survey will be offered
	during Production Phase 2, which is the latter half of the
	production phase of data collection. For the default protocol, the
	abbreviated survey will be offered as the last step in nonresponse
	conversion.
Other
	interventions.
	While all BPS studies are conducted by NCES, the data collection
	contractor, RTI International, has typically used the study-specific
	e-mail “@rti.org” to contact and support sample members.
	Changing the e-mail sender to the NCES project officer or the RTI
	project director may increase the perceived importance of the survey
	and help personalize the contact materials, thereby potentially
	increasing relevance. Switching the sender during data collection
	also increases the chance that the survey invitation is delivered to
	the sample member rather than to a spam filter.
As a result of the above experiment(s), detailed information on field results can be found in Appendix D.
Included in this section is the name and telephone number of individuals consulted on statistical aspects of the design and the name of the agency unit, contractor(s), grantee(s), or other persons who will actually collect and/or analyze the information for the agency.
BPS:20/22
	is being conducted by NCES/ED. The following statisticians at NCES
	are responsible for the statistical aspects of the study: Dr. David
	Richards, Dr. Tracy Hunt-White, Dr. Elise Christopher, and Dr. Gail
	Mulligan. NCES's prime contractor for BPS:20/22 is RTI International
	(RTI). The following staff members at RTI are working on the
	statistical aspects of the study design: Dr. Joshua Pretlow, Dr.
	Jennifer Wine, Dr. Nestor Ramirez, Mr. Darryl Cooney, Mr. Michael
	Bryan, Dr. T. Austin Lacy, Dr. Emilia Peytcheva, and Mr. Peter
	Siegel.
Principal professional RTI staff not listed above, who
	are assigned to the study include: Ms. Ashley Wilson, Ms. Kristin
	Dudley, Mr. Jeff Franklin, Ms. Chris Rasmussen, and Ms. Donna
	Anderson.
References
	
Cantor,
	D., O'Hare, B.C., and O'Connor, K.S. (2008). The Use of Monetary
	Incentives to Reduce Nonresponse in Random Digit Dial Telephone
	Surveys. In Lepkowski, J.M., Tucker, N.C., Brick, J.M., de Leeuw,
	E., Japec, L., Lavrakas, P.J., Link, M.W., and Sangster, R.L.
	(eds.). Advances in Telephone Survey Methodology. New York:
	Wiley.
Chromy, J.R. (1979). Sequential Sample Selection
	Methods. In Proceedings of the Survey Research Methods Section of
	the American Statistical Association (pp. 401–406).
	Alexandria, VA: American Statistical Association.
Church,
	A.H. (1993). Estimating the Effect of Incentives on Mail Survey
	Response Rates: A Meta-Analysis. Public Opinion Quarterly, 57(1):
	62-79.
Cominole, M., Shepherd, B. and Siegel, P. (2015).
	2008/12 Baccalaureate and Beyond Longitudinal Study (B&B:08/12).
	Data file documentation. NCES 2015-141. Washington, DC: National
	Center for Education Statistics.
DeBell, M., Maisel, N.,
	Edwards, B., Amsbary, M., and Meldener, V. (2019). Improving Survey
	Response Rates with Visible Money. Journal of Survey Statistics and
	Methodology. Online First. Retrieved from:
	https://academic.oup.com/jssam/advance-
	article/doi/10.1093/jssam/smz038/5610622
Dillman, D.,
	Smyth, J., and Christian, L. (2014). Internet, Phone, Mail, and
	Mixed Mode Surveys: The Tailored Design Method. John Wiley &
	Sons Inc. Hoboken, New Jersey.
Folsom, R.E., Potter,
	F.J., and Williams, S.R. (1987). Notes on a Composite Size Measure
	for Self-Weighting Samples in Multiple Domains. In Proceedings of
	the Section on Survey Research Methods of the American Statistical
	Association. Alexandria, VA: American Statistical Association,
	792–796.
Goeritz, A.S. (2006). Incentives in web
	studies: Methodological issues and review. International Journal of
	Internet Science, 1: 58-70.
Griggs, A., Powell, R.,
	Keeney, J., Waggy, M., Harris, K., Halpern, C. and Dean, S. (2019)
	Research Note: A Prenotice Greeting Card's Impact on Response Rates
	and Response Time, Longitudinal and Life Course Studies, 10(4):
	421-432.
Groves, R.M., Cialdini, R., and Couper, M. 1992.
	Understanding the Decision to Participate in a Survey. Public
	Opinion Quarterly, 56(4):, 475-495.
Groves R.M., and
	Heeringa, S.G. (2006). Responsive Design for Household Surveys:
	Tools for Actively Controlling Survey Errors and Costs. Journal of
	the Royal Statistical Society Series A-Statistics in Society,
	169(3): 439-457.
Groves, R.M., Singer, E. and Corning, A.
	(2000). Leverage-Saliency Theory of Survey Participation.
	Description and Illustration. Public Opinion Quarterly, 64:
	299-308.
Hill, J., Smith, N., Wilson, D., and Wine, J.
	(2016). 2008/12 2012/14 Beginning Postsecondary Students
	Longitudinal Study (BPS:12/14). NCES 2016-062. Washington, DC:
	National Center for Education Statistics.
Kirchner, A.,
	Emilia, P., and N. Tate. (2021, May). Can Prepaid PayPal Incentives
	Be as Effective as Prepaid Cash Incentives? Paper presented at the
	Annual Conference of the American Association for Public Opinion
	Research.
Lozar Manfreda, K., Bosnjak, M., Berzelak, J.,
	Haas, I., and Vehovar, V. (2008). Web Surveys Versus Other Survey
	Modes. A Meta-Analysis Comparing Response Rates. International
	Journal of Market Research, 50(1): 79-104.
Medway, R.L.
	and Tourangeau, R. (2015). Response Quality in Telephone Surveys. Do
	Prepaid Incentives Make a Difference? Public Opinion Quarterly,
	79(2): 524-543.
Messer, B.L., and Dillman, D.A. (2011).
	Surveying the General Public Over the Internet Using Address-Based
	Sampling and Mail Contact Procedures. Public Opinion Quarterly,
	75(3): 429–457.
Olson, K., Smyth, J. D., and Wood,
	H. (2012). Does Giving People Their Preferred Survey Mode Actually
	Increase Survey Participation? An Experimental Examination. Public
	Opinion Quarterly, 76: 611–635.
Parsons, L., and
	Manierre, M.J. (2014). Investigating the Relationship among Prepaid
	Token Incentives, Response Rates, and Nonresponse Bias in a Web
	Survey. Field Methods, 26(2): 191-204.
Singer, E. (2002).
	The Use of Incentives to Reduce Nonresponse in Household Surveys. In
	Groves, R.M., Dillman, D. A., Eltinge, J.L., Little, R.J.A. (eds.),
	Survey Nonresponse. New York: Wiley.
Singer, E. and Ye,
	C. (2013). The Use and Effects of Incentives in Surveys. Annals.
	Annals of the American Academy of Political and Social Science,
	645(1): 112-141.
Wilson, D., Shepherd, B., and J. Wine.
	(2015, May). The Use of a Calibration Sample in a Responsive Survey
	Design. Paper presented at the Annual Conference of the American
	Association for Public Opinion Research, Hollywood, Florida.
	
	
	
	
| File Type | application/vnd.openxmlformats-officedocument.wordprocessingml.document | 
| Author | William West | 
| File Modified | 0000-00-00 | 
| File Created | 2021-12-16 |