Supporting Statement for OMB Clearance Request Part A
Family and Medical Leave Act, Wave 4 Surveys
OMB Control Number 1290-0NEW
July 2017
Table of Contents
A.1. Necessity for the Data Collection 1
A.2. Purpose of Survey and Data Collection Procedures 2
A.2.1 Target Population, Mode of Administration, Survey Content and Objectives 2
A.2.2 How the Survey Findings Will Be Used 9
A.3. Improved Information Technology to Reduce Burden 9
A.4. Efforts to Identify Duplication 10
A.5. Involvement of Small Organizations 10
A.6. Consequences of less Frequent Data Collection 10
A.8. Federal Register Notice and Consultation 11
A.8.1 Federal Register Notice and Comments 11
A.8.2 Consultation with Experts 11
A.9. Incentives of Respondents 12
A.10. Privacy of Respondents 13
A.12. Estimation of Information Collection Burden 16
A.13. Cost Burden to Respondents of Record Keepers 18
A.14. Estimate of Cost to the Federal Government 18
A.16. Plan and Time Schedule for Information Collection, Tabulation and Publication 19
A.17. Reasons Not to Display OMB Expiration Date 20
A.18. Exceptions of Certification for Paperwork Reduction Act Submissions 20
List of Attachments
Attachment A: Employee Survey
Attachment B: Employer Survey
Attachment C: Advance Letter for the Employer Survey
Attachment D: 60-Day Federal Register Notice
Attachment E: Summary of Public Comments and DOL Response
Attachment F: Privacy Statement
Attachment G: Weighting and Estimation Procedures
Attachment H: IRB Approval Letter
As required under the Paperwork Reduction Act (PRA), the Chief Evaluation Office (CEO) at the U.S. Department of Labor (DOL) is seeking approval from the Office of Management and Budget (OMB) for a study of the federal Family and Medical Leave Act (FMLA). Consistent with its commitment to tracking and analyzing user feedback, DOL periodically conducts surveys of employees and employers who are covered by the provisions of the FMLA. Surveys were conducted in 1996 (DOL 1996), 2000 (Cantor et al. 2001), and again in 2012 (Klerman, Daley, and Pozniak 2013). DOL has contracted with Abt Associates Inc. (in partnership with Abt SRBI1) to conduct a fourth “wave” of Employee and Employer Surveys.
Since 2012, both the labor market and the policy context in which the FMLA operates have continued to evolve. In particular, as of December 2016, building on the FMLA, 43 states, counties, and cities have implemented one or more paid leave policies.
This information collection request covers a fourth “wave” of nationally representative surveys of employees and employers that will update and expand upon the knowledge gained from prior studies in the current policy landscape. Specifically, the surveys will help DOL understand current workplace policies and practices related to family and medical leave, allowing the Department to determine how those policies affect the work-life balance of workers and the productivity and workflow of employers. The study will also enable the Wage and Hour Division (WHD) at DOL to shape future regulatory options, craft interpretive guidance (such as plain-language fact sheets), develop compliance programs (employer outreach and investigation policies), and establish regulatory priorities based on sound, current data, rather than outdated or anecdotal information. Finally, the study will provide a data set by which WHD can evaluate the effect of a range of FMLA activities—regulatory, educational, investigative, and legal—on employer compliance.
CEO undertakes a learning agenda process each year to identify Departmental priorities for program evaluations. This evaluation was prioritized as part of that process in FY 2016. Division H, Title I, section 107 of Public Law 114-113, the “Consolidated Appropriations Act, 2016” authorizes the Secretary of Labor to reserve not more than 0.75 percent from special budget accounts for transfer to and use by the Department’s Chief Evaluation Office (CEO) for departmental program evaluation. Further, 29 USC 3224a (1), authorizes the Secretary of Labor to conduct ongoing evaluation of programs and activities to improve the management and effectiveness of these programs.
DOL’s Chief Evaluation Office requests clearance to administer two surveys:
A survey of employees that primarily examines leave usage patterns and unmet needs; and
A survey of employers that examines their leave policies, as well as their experiences with and perceptions of the FMLA.
These surveys are based on the 2012, Wave 3 FMLA surveys. However, the instruments in this information collection request have been revised (with the input of substantive and methodological experts) to reflect changes in the external environment, build on past data collection experience, and improve our understanding of the FMLA and leave-taking behavior and needs.
As part of the questionnaire updating process, the contractor solicited input from various sources. First, the contractor’s team conducted several listening sessions and in-depth interviews with numerous stakeholders, including DOL staff, scholars who study the FMLA, policy and advocacy professionals working in the area of family and medical leave, and employers and third-party administrators who administer FMLA. Second, the contractor’s team also consulted with the study’s Technical Working Group (TWG) (see Section A8 below for a list of TWG members). Feedback from these informative sessions has been incorporated into the Wave 4 surveys. The content, sample, and administration of the Employee and Employer Surveys are described in detail below.
The following section describes the target population for each survey, as well as the method by which the surveys will be administered. It then describes the content of each survey and how that aligns with the survey objectives. This information is provided first for the Employee Survey and then for the Employer Survey.
The Employee Survey (Attachment A) will examine the need for and use of leave among employees. The survey will be conducted by computer-assisted telephone interview (CATI) in both English and Spanish. The target number of completed interviews is 4,000. The target population for the Employee Survey will be individuals aged 18 or older who live in the United States, have a telephone (landline or cellular), and have been employed for pay (private or public sector) in the 12 months prior to the interview. Employed workers will be screened into the survey using a question adapted from the Current Population Survey: “in the past 12 months were you employed by government, by a private company, a nonprofit organization, or were you self-employed?” Respondents who say that they were employed by government, a private company, or a nonprofit organization are classified as wage and salary workers and are included in the survey. Those who report that they were self-employed are not included in the survey.
The proposed Employee Survey is also designed to yield representative samples of major demographic groups defined by race, ethnicity, age, gender, and income. Consistent with the 1995, 2000, and 2012 waves of this survey, leave-takers (employees who took leave) and leave-needers (employees who needed leave but did not take one) will be oversampled to achieve reliable subsample estimates. New to the current wave, low-wage workers and employees working in a state with a paid family and medical leave statute will also be oversampled.
The survey will gather information across seven different domains: (1) respondent characteristics, (2) FMLA coverage and eligibility, (3) leave benefits, (4) leave usage, (5) leave experience, (6) the decision not to use FMLA benefits, and (7) leave requirements and practices. Exhibit 1, below, describes the rationale for including each domain (and their subdomains) in the Employee Survey for each section of the survey (“Survey Section”).
The Employee Survey has four different types of target respondents: those who participate in the interview screening, those who are selected to complete the survey after screening (referred to as “all respondents” in the table), leave-takers (employees who took leave for a covered reason during the screening period), and leave-needers (employees who needed to take leave for a covered family or medical reason, but did not). Exhibit 1 also indicates the target respondents for each section.
In addition to the domain specific rationales listed below, several survey questions have been included in the survey to maintain comparability with the 2012 Employee Survey. For additional detail about comparability, see Section B.2.1.
Exhibit 1. Employee Survey Content and Objectives
Domain |
Rationale |
Subdomain |
Survey Section2 |
Survey Question |
Respondent Characteristics |
To facilitate understanding of how FMLA coverage, use (and non-use), need, and experiences vary by employee characteristics. |
|
|
S1-10, T1-6 S13, E0, E9-10, E15-17, D2-11, ZIP |
FMLA Coverage & Eligibility |
To document current levels of FMLA worksite coverage and employee eligibility. Also, employees’ awareness and familiarity with the law, as well as how they learn about the FMLA. |
|
|
E2-6, E11-12 |
Leave Benefits |
To provide information in the Employee Survey on the types of paid and unpaid benefits available to employees during leave, the breadth of those benefits, and whether those benefits vary by type of employee. |
|
|
E7-8 |
Leave Usage |
To better understand leave usage from the employee perspectives. That is, how often, how long, and for what reasons employees take family and medical leave. This domain also covers how employers process FMLA requests, when they deny requests, and their perceptions about the frequency of FMLA misuse. |
|
|
S11a-b, A1-20a, C1 |
Leave Experience |
To provide information on employees’ financing of leave-taking and their transition back into the workplace, and on how employers accommodate their employees’ need for leave. |
|
|
A23a-g, A43-59, A62-64, E5-6 |
Decision to Not Use FMLA Benefits |
To understand the circumstances of those who needed but did not take leave, including reasons the leave was needed and reasons why it was not taken |
|
|
S12a-b, B1-20 |
Leave Requirements and Practices |
To determine the length of notice employees give to employers before taking leave (both according to company policy and in practice), and examine the process of obtaining medical certification and clearance to return to work. |
|
|
A26-42, A60 |
A detailed description of the respondent universe, sampling methods, and data collection procedures, as well as discussion of efforts to maximize response rates for the Employee Survey, can be found in Part B.
The Employer Survey (Attachment B) will explore human resource/benefits administrator experience with and perception of implementing FMLA leave by employees at their worksite. The survey will be conducted using a sequential multi-mode (web and CATI) methodology. The target number of completed surveys is 2,000. This sample will support nationally representative estimates of results in the aggregate, as well as estimates by major industry and employer size.
The target population for the Employer Survey is U.S. private-sector business worksites,3 excluding self-employed persons without employees and also excluding government and quasi-government units (federal, state, and local governments; public educational institutions; and post offices). This universe differs from that of the Employee Survey, which includes both private- and public-sector employees. As in 2012, the Dun & Bradstreet Dun’s Market Identifiers (DMI) file will serve as the sampling frame for the Employer Survey.
As with the Employee Survey, the Employer Survey collects information across seven domains: (1) worksite and firm characteristics, (2) FMLA coverage and eligibility, (3) leave benefits, (4) leave usage, (5) leave experience, (6) leave requirements and practices, and (7) perceptions of the FMLA. Exhibit 2, below, describes the rationale for collecting information across these domains (and subdomains) and the target respondents for each section of the survey (“Survey Section”).
Exhibit 2. Employer Survey Content and Objectives
Domain |
Rationale |
Subdomain |
Survey Section4 |
Survey Question |
Worksite and Firm Characteristics |
To facilitate understanding of how FMLA coverage, use (and non-use), need, and experiences vary by employer characteristics. |
|
|
Q1-1.2, Q4-7, Q14-14a |
FMLA Coverage & Eligibility |
To document current levels of FMLA worksite coverage and employee eligibility. Also, employers' awareness and familiarity with the law, as well as how they learn about the FMLA. |
|
|
Q2-3.1, Q8-10, Q17, Q47 |
Leave Benefits |
To provide information in the Employer Survey on the types of paid and unpaid benefits available to employees during leave, the breadth of those benefits, and whether those benefits vary by type of employee. |
|
|
Q11-11.3, Q16_A-H, Q16X_1, Q16X_4, Q44D |
Leave Usage |
To better understand leave usage from the employer perspectives. That is, how often, how long, and for what reasons employees take family and medical leave. This domain also covers how employers process FMLA requests, when they deny requests, and their perceptions about the frequency of FMLA misuse. |
|
|
Q18-22, Q27-28, Q40-43, Q58-60, Q68 |
Leave Experience |
To provide information on employees’ financing of leave-taking and their transition back into the workplace, as well as how employers accommodate their employees’ need for leave. |
|
|
Q16X_5, Q57_1-3, Q23-24, Q44C, Q61X-bX |
Leave Requirements and Practices |
To determine the length of notice employees give to employers before taking leave (both according to company policy and in practice), and examine the process of obtaining medical certification and clearance to return to work. |
|
|
Q15-15a, Q16X_2-3, Q25-26a, Q34-39, Q44A-B, Q48-48a |
Perceptions of the FMLA |
To provide information on the perceived impact on employers of their employees taking leave under the FMLA. |
|
|
Q49-56, Q67A-E |
Additional information about the respondent universe, sampling methods, and data collection procedures, as well as efforts to maximize response rates for the Employer Survey, can be found in Part B.
DOL’s contractor will compile an analytical research report on the results of these surveys. We anticipate that information will be used by policy makers and researchers both inside and outside of government to analyze the current state of family and medical leave as well as make comparisons with 2012. The results will be primarily descriptive in nature since the study cannot provide evidence of causal effects of FMLA. To the extent feasible however, regression-adjusted comparisons will be reported for some outcomes to provide results that are “closer to causal.”
The use of electronic technology potentially provides the opportunity to facilitate the administration and completion of the survey instruments, ultimately leading to a reduction in respondent burden. This section discusses this issue first for the Employee Survey and then for the Employer Survey.
The study team will field the Employee Survey using computer-assisted telephone interview (CATI) methods. The study team has taken several steps to minimize burden on respondents. Respondents will be allowed to provide the information in a short telephone survey. In addition, respondents will be offered the opportunity to conduct the survey at a time most convenient to them, and given a toll-free number to call to schedule or conduct it. Following standard survey protocol, respondents will be informed that they may choose not to answer any question and that they may end the survey at any time. Furthermore, respondents will be offered the opportunity to begin the survey in one session and finish it in another session.
The Employer Survey will use a combination of internet administration and CATI methods. The study team has attempted to minimize the burden on respondents by allowing them to provide the information either through the internet or over the phone.
To determine the eligibility of a worksite and the name and contact information of the person who is the most appropriate to complete the survey questionnaire given the nature of the information requested (the “key informant”), the study team will first call each worksite and conduct a short telephone screener survey (see Attachment B). Once the appropriate employer respondent has been identified (and prior to actual survey administration), the study team will give that key respondent the option of receiving an informational packet by email or mail. The informational packet explains the survey and provides a list of information about leave-takers that the employer should gather prior to the telephone interview (Attachment C). If the study team is not able to reach the identified respondent on the telephone, it will mail (not email) the materials. For those who request email, that email will also contain a secure link to take the respondent directly to his/her own landing page to log into the survey. During the telephone screener, interviewers will first walk respondents through how to access the web-based survey on their own. If they are unwilling or unable to do the survey on the internet, the study team will offer the option to complete the survey on the telephone.
The study team has designed the survey to provide comparison data for the FMLA Employee Survey conducted by the DOL Commission on Family and Medical Leave in 1995 and by the Department in 2000 and 2012. There are no other surveys of a large, random sample of individuals or households that asked questions specifically related to the required topics regarding family and medical leave.
The study team has designed the survey to provide comparison data for the FMLA Employer Survey conducted by the DOL Commission on Family and Medical Leave in 1995 and by the Department in 2000 and 2012. There are no other surveys of a large, random sample of businesses of diverse sizes that ask questions specifically related to the required topics regarding family and medical leave.
The CATI technique is being used for the Employee Survey to minimize burden on leave taking and leave needing employees. Small businesses and other small entities are not involved.
About one-third of the Employer Survey respondents are expected to represent small businesses. According to the Small Business Association, the Office of Advocacy defines a small business as an independent enterprise having fewer than 500 employees. The data collection procedures have been designed to minimize the burden on those small employers as well as representatives from larger organizations through the following:
The advance letter and accompanying materials (Attachment C) inform the respondents that survey questions may require them to consult administrative records in order to accurately complete the instrument. Being notified in advance will minimize the need to pause in completing the survey in order to consult the proper records.
The Employer Survey seeks information about a 12-month period, and allows employer respondents to select the 12-month period most convenient for them.
The internet survey will further enhance this opportunity in so far as respondents may end any given session on the survey and return to their previous answers at their discretion. This will allow respondents to complete the survey at a time and place most convenient to them.
This is a one-time collection.
There are no special circumstances for the proposed data collection efforts.
In accordance with the Paperwork Reduction Act of 1995 (P.L. 104-13) and Office of Management and Budget (OMB) regulations at 5 CFR Part 1320 (60 FR 44978, August 29, 1995), DOL published a notice in the Federal Register announcing the agency’s intention to request an OMB review of this information collection activity. This notice was published on Friday, October 28, 2016, in Volume 81, Number 209, pages 75161–75163, and provided a 60-day period for public comment. A copy of this notice is included as Attachment D. During the notice and comment period, the government received 13 comments from 12 entities. The comments included statements of support for the data collection, requests for the survey instruments, and substantive comments about the surveys. DOL provided the survey instruments to all entities that requested them. See Attachment E for a full discussion of the substantive comments and our responses.
The following people were consulted in developing the design, the data collection plan, and the forms for which clearance is requested.
Abt Associates
Ms. Radha Roy (301) 347-5722
Mr. Glen Schneider (617) 349-2471
Dr. Jane Leber Herr (617) 520-3042
Mr. Jacob Klerman (617) 520-2613
Dr. Kelly Daley (312) 529-9703
Abt Associates assembled a technical working group consisting of five experts in the following areas: (1) studies of the federal Family and Medical Leave Act (2) workforce and work-family policies (3) survey methodology, and (4) human resources administration and employment law. These experts have reviewed and commented on the Employee and Employer Surveys and will review the study findings. Due to her particular expertise, Dr. Jolene Smyth will also review the study design.
TWG Members
Dr. Ann Bookman University of Massachusetts Boston
Dr. Christopher Ruhm University of Virginia
Dr. Jolene Smyth University of Nebraska-Lincoln
Dr. Jane Waldfogel Columbia University
Ms. Christine Walters FiveL Company
The surveys will update results from the three previous surveys conducted on this topic in 1995, 2000, and 2012. Many of the items in the Employee and Employer Surveys are from the previously approved 2012 FMLA Employer and Employee Surveys (OMB No. 1235-0026).
This section discusses incentives for respondents, first for the Employee Survey and then for the Employer Survey.
In order to cover extenuating costs (e.g., per minute phone charges), participants completing the Employee Survey on a cellphone will be issued $15 as a token of appreciation for their participation. The ultimate objective of offering incentive payments is to improve our estimates by reducing nonresponse bias. In accordance with OMB guidance, incentives are most appropriately used in federal statistical surveys to reduce nonresponse bias.5 The incidence of taking leave or needing to take leave is relatively rare and the failure to include these hard to find participants could jeopardize the quality of the survey data. Furthermore, the incentive will improve our estimates for low-wage workers, a group that is traditionally difficult to reach and survey, and also a group that is of special interest to DOL. The study team plans to oversample low-wage workers by oversampling prepaid cellphones (see Part B, Section B.1.1 for more information). The $15 incentive will cover the costs of per minute phone charges incurred by these individuals and their propensity to use pre-paid cell phones.
The Employee Survey features both a screener and an extended interview. In theory, incentives could be offered at one or both stages. The research literature and the survey sample design indicate that extended interview incentives would be cost effective for this survey while screener incentives would not (O’Hare and Frederick 2003; Cantor et al. 1998, 2007; Singer et al. 2000). A related issue is whether incentives should be prepaid or post-paid (“promised”). Based on the prominence of cellphones and unlisted landlines in the sample, we estimate that fewer than half of the numbers sampled for the survey (about 40 percent) could be matched to an address. This makes prepaid incentives a moot issue. The interview incentives, thus, will need to be post-paid. Respondents will be notified at the beginning of the survey that they will receive the incentive after completion of the survey. Respondents to the survey who provide their address will receive a check for the incentive amount within three weeks of completing the telephone survey.
A nonresponse follow-up survey (NRFU), which the study team will use to evaluate nonresponse in the Employee Survey, will include incentive payments. The NRFU will collect information on employees who fail to respond to the survey and provide insight into whether the nonrespondents differ from the respondents on the characteristics of interest (e.g., family and medical leave). Specifically, interviewers will call back a subsample (n=1,200) of households that declined the original survey and attempt to recruit an eligible employee to complete a shortened interview with a $40 incentive. Respondents to the NRFU who provide their address will be mailed a check for $40 within three weeks of completing the NRFU survey.
Incentives are a common feature in NRFU surveys because, by definition, the NRFU sample did not cooperate with the original survey, and so a major change in the recruitment protocol is required to elicit cooperation. Zimowski and colleagues (1997) noted in their report to the Federal Highway Administration (FHWA-PL-98-029) that large monetary incentives (e.g., $20 to $50) are a common element of NRFU designs for household surveys. For example, Peytchev et al. (2009) documented how a $20 incentive was used in a successful NRFU to the National Intimate Partner and Sexual Violence Survey for the Centers for Disease Control and Prevention. As part of the 2012, Wave 3 FMLA surveys, the contractor offered a $20 incentive to respondents of the NRFU survey and achieved a response rate of 23 percent. More recently, in the Worker Classification Knowledge Survey project for DOL, respondents were offered $20 as remuneration for completing a NRFU, and the response rate was just seven percent (Daley et al. 2016). To improve on this recent experience, the Wave 4 FMLA NRFU study team has increased the amount of the incentive from $20 to $40.
No payments or incentives will be provided for participation in the Employer Survey.
Verbal assurance of privacy will be provided to all respondents of the Employee and Employer Surveys. Respondents who elect to take the online version of the Employer Survey will receive written assurance of privacy. All participants will be informed that their participation is voluntary and they may skip any questions they do not wish to answer, their responses will be kept private, and the risks are minimal. The exact statement indicating the privacy of respondents’ answers is attached (Attachment F).
The contractor has obtained approval from its IRB (Attachment H). All staff working on the study will have completed basic human protections as well as project-specific IRB training.
Prior to the analysis of the collected survey data, the contractor will assign generic study identifiers—not based on PII—for each study participant to link participant data. The contractor will then remove personal identifiers (e.g., name of company, name of employee). As a result, this analysis file (and the Public Use File which will be derived from it) will not include information that will allow direct linking to a specific individual or employer. Thus external analysts will never see or be able to access these individual-level identifiers. Reverse identification (not from identifiers) remains possible. Later in this section, we discuss the additional steps we will take to prevent reverse identification from variables other than identifiers.
The survey data will be stored on an contractor’s computer that is protected by a firewall which monitors and evaluates all attempted connections from the Internet. Private information on each survey respondent (e.g.., name, telephone number, and for those who elect to receive an incentive, mailing address) will be maintained in a file separate from the survey data. The contractor uses a LAN and WAN network infrastructure to facilitate computer-assisted telephone interviewing. Access to the LAN and WAN are restricted by IP address. Access to the WAN is restricted by IP address, firewalls, and security rules on the contractor’s routers. All servers’ access is restricted by username and password authentication. The perimeter is secured by an active IPS (Intrusion Prevention System) with state packet inspection. Once the survey is completed, all private data on each respondent will be destroyed. Any data used for analysis by the contractor or the Department will be completely de-identified. The entire database will be encrypted so that any data stored will be further protected. Finally, access to data with identifying information will be limited to contractor staff directly working on the survey. Those doing analysis will not have access to the identifying information.
All analyses, summaries or briefings will be presented at the aggregate level. As a result, it will not be possible to link specific responses to individual respondents in any way. The database delivered to DOL and subsequently available for public use will not include any identifying information such as names, addresses, telephone numbers, or social security numbers, nor any other information that might support reverse identification of respondents.
Public use files (PUFs) for both the Employee Survey and Employer Survey will be made available after completion of the data collection. Following standard privacy protocols, in the Wave 4 files the study team will coarsen the data to protect respondent privacy, while still providing sufficient information to allow other researchers using the PUFs to replicate the analyses presented in the Final Report. In particular, the study team will review the final data for rare responses in both survey data sets. As necessary, the study team will re-code so that no single response category or combination of closely related response categories has an unweighted frequency below five. For other rare responses, or for combinations of responses that may susceptible to re-identification, the contractor will coarsen the data by some combination of suppression, rounding, top-coding, bottom-coding, or collapsing response groups as needed so that no identifying values are released in the PUFs. However, for some variables of special interest, the contractor may instead maintain privacy by using an alternative approach, such as data swapping.
The study team also will prepare and deliver tables that describe how to use and access the files, each variable, any editing strategies employed, the data source, the sample and population represented, and information such as frequencies, means, and any additional information agreed on with DOL. The collection, maintenance, and dissemination of data (including personal identifiers) will be compliant with all requirements articulated in federal statutes. The contractor complies with the E-Government Act of 2002, including Title III: Federal Information Security Management Act (FISMA), which covers site security, security control documentation, access control, change management, incident response, and risk management. The contractor follows a risk management process to manage security risks to the company and projects, in order to assess risks and implement secure mitigation strategies accordingly to secure Personally Identifiable Information (PII) and Protected Health Information (PHI).
With multimodal data collection via phone and web surveys, the privacy of the individuals rests heavily on the interviewers and the systems doing the data collection. All staff are trained in how to protect sensitive information and are monitored by multiple supervisors. For any IT used, the contractor’s Information Security team reviews the design and security requirements of the survey system to identify gaps in implementation and privacy weaknesses. In the event of an incident, the contractor has developed an incident response capability to support the proper investigation and reporting of any PII related incident.
The contractor will collaborate with DOL in determining options for presenting the Wave 4 public use data files in order to support replication of key results and data usage while preserving respondent privacy.
The content of both surveys is limited to information needed for policy purposes as was true in the 2012 and earlier waves of the surveys. This includes asking the earlier questions again (to estimate trends) and asking new questions based on changing regulations, changes in regulation under consideration, changes in the external environment, and lessons learned in the analysis and use of the earlier surveys.
Before starting the survey, respondents will be told that the survey pertains to family and medical leave and will contain questions about their experience taking leave, or needing to take leave, from work. Interviewers will inform respondents that they may skip or refuse to answer any question and may also end the interview at any time. Respondents will also be notified that their participation and responses will be kept private, to the extent allowable under law.
The Employee Survey asks respondents to identify whether or not they have taken leave, or needed to take leave, from work for a variety of reasons which may be sensitive. The survey also asks respondents about the relationship (e.g., parent, spouse, child, etc.) between themselves and the person for whom they took leave. These questions are central to the goal and focus of this study. This information is vital for determining what working families need to help them balance the demands of work and family. All data will be aggregated so that information about individual respondents will not be identifiable.
The Employee Survey also asks respondents about their sexual orientation. This question wording is taken verbatim from the sexual orientation and gender identity module of the Behavioral Risk Factor Surveillance System (BRFSS) coordinated by the Centers for Disease Control and Prevention (CDC) at the U.S. Department of Health and Human Services. The CDC developed this module in 2013 with input from the Sexual Minority Assessment and Research Team and the Gender Identity in U.S. Surveillance, and it allowed states to begin using the module the following year. Twenty states and territories added these questions in 2014 and 25 did so in 2015 (Baker and Hughes 2016).
The issue of coverage of same sex partners for family and medical leave has been an area of specific policy activity with respect to the FMLA. In particular, in June 2010, DOL clarified the FMLA definition of “son and daughter” to include an employee standing “in loco parentis” to a child, regardless of the employee’s legal or biological relationship to the child. Combined with other questions for the survey, this information will help DOL understand the magnitude, needs, and experiences of same sex partners regarding family and medical leave from work.
The study team does not anticipate substantial nonresponse for any of the survey items described above. Similar questions were asked in the 2012 survey and there was less than 5 percent nonresponse across these items. Also, the study team conducted cognitive tests of the Employee Survey with nine volunteer respondents from a range of different backgrounds and leave experiences. The cognitive interview research protocol used by the study team followed guidelines recommended in the Statistical Policy Directive No. 2: Standards and Guidelines for Statistical Surveys, section on Cognitive Interviewing6. Further information is included in Part B, Section B.4.1. The testing did not reveal any issues with the questions. The survey included in this package reflects the findings from those interviews.
There are no sensitive questions in the Employer Survey.
The proposed data collection does not impose a financial burden on respondents nor will respondents incur any expense other than the time spent completing the surveys.
Exhibit 3. Estimated Annualized Respondent Hour and Cost Burden
Type of Instrument |
Number of Respondents |
Number of Reponses Per Respondent |
Number of Responses |
Average Burden Per Response (in Hours) |
Estimated Burden Hours |
Average Hourly Wage1,2 |
Total Burden Costs |
EMPLOYEE SURVEY |
|||||||
Screener |
8,864 |
1 |
8,864 |
2/60 |
295 |
$25.62 |
$7,557 |
Extended Interview Overall |
1,333 |
1 |
1,333 |
14/60 |
311 |
$25.62 |
$7,968 |
Nonresponse Follow-Up |
167 |
1 |
167 |
10/60 |
28 |
$25.62 |
$717 |
Duplicated Total |
10,364 |
-- |
10,364 |
-- |
634 |
-- |
$16,242 |
EMPLOYER SURVEY |
|||||||
Screener |
3,044 |
1 |
3,044 |
4/60 |
203 |
$56.29 |
$11,427 |
Data Gathering + Extended Interview Overall |
667 |
1 |
667 |
1 |
667 |
$56.29 |
$37,545 |
Duplicated Total |
3,711 |
|
3,711 |
|
870 |
-- |
$48,972 |
Duplicated Totals |
14,075 |
-- |
14,075 |
-- |
1,504 |
-- |
$65,214 |
1Hourly wage reflects the June 2016 (seasonally adjusted) average hourly earnings for private nonfarm payrolls, as reported by the U.S. Department of Labor, Bureau of Labor Statistics, Table B-3. "Average hourly and weekly earnings of all employees on private nonfarm payrolls by industry sector, seasonally adjusted" (accessed from the following web site as of September 26, 2016: http://www.bls.gov/news.release/empsit.t19.htm).
2Hourly wage reflects the May 2015 average hourly earnings for human resources managers, as reported by the U.S. Department of Labor, Bureau of Labor Statistics, Table 1. "National employment and wage data from the Occupational Employment Statistics survey by occupation, May 2015" (accessed from the following web site as of January 11, 2017: https://www.bls.gov/news.release/ocwage.t01.htm).
There are no costs to respondents other than their time, and no recordkeeping is required. There are no capital and start-up costs borne by respondents for this information collection. There are no costs to respondents associated with operation, maintenance, or the purchase of services.
The estimated annualized cost to the Federal government is $920,132. The estimated total cost to the Federal government over the 3 year clearance period is $2,760,396.7 This cost estimate is comprised of two components:
The estimated total cost to the Federal government for the contractor to carry out the tasks for which this justification seeks approval is $2,702,000. This is comprised of $378,881 for instrument design and $2,323,119 for data collection. Annualized over three years, the estimated annual cost is $900,667.8
DOL expects the annual level of effort for Federal government technical staff to oversee the contract will require 200 hours for one Washington D.C.-based l GS-14, Step 5 employee earning $60.83 per hour.9 To account for fringe benefits and other overhead costs the agency applies a multiplication factor of 1.6. Thus, the estimated annual cost borne by DOL for these duties is $19,466.10 The data collection period covered by this justification is three years, so the estimated total cost for performance of these duties is $58,397.
This is a new data collection.
Consistent with the overall intent and design of the study, data will be analyzed and presented primarily in a descriptive statistical manner, employing cross-tabulations. No complex analytical techniques will be used in assessing these data. Survey results will be weighted to be representative of the U.S. non-institutional population based on figures from the American Community Survey and the Current Population Survey. The analysis team will use SAS to estimate level differences in key outcomes across these subgroups (e.g., FMLA eligible and ineligible employees), reporting confidence intervals and testing for statistical significance. The study team will also supplement these simple weighted comparisons with regression-adjusted comparisons for selected outcomes. See Part B for more detail.
As with the Employee Survey, analysis of the results from the Employer Survey will also rely primarily on descriptive statistics. Again, no complex analytical techniques will be used in assessing these data. Survey results will be weighted to adjust for differential probabilities of selection and nonresponse. The analysis team will use SAS to estimate level differences in key outcomes across these subgroups (e.g., FMLA eligible and ineligible employees), reporting confidence intervals and testing for statistical significance. The study team will also supplement these simple weighted comparisons with regression-adjusted comparisons for selected outcomes. See Part B for more detail.
CEO is seeking OMB approval beginning October 2017 and ending October 2020. Exhibit 4 presents an overview of the project schedule for data collection, analysis, and publications.
Exhibit 4. Project Timeline
Timeline |
Activity |
Details of Activity |
August 2016 – January 2017 |
Develop and refine surveys |
|
February – March 2017 |
Cognitive testing of surveys |
|
April – October 2017 |
Office of Management and Budget (OMB) package under review |
|
November 2017 – April 2018 |
Field surveys (assuming OMB clearance obtained by October 31, 2017) |
|
May 2018 – April 2019 |
Prepare findings |
|
April – August 2019 |
Disseminate findings |
|
All instruments will display the expiration date for OMB approval.
No exceptions are necessary for this information collection
Baker, Kellan E., and Margaret Hughes. 2016. “Sexual Orientation and Gender Identity Data collection in the Behavioral Risk Factor Surveillance System.” Washington, DC: Center for American Progress.
Cantor, David, Patricia Cunningham, and Pamela Giambo. 1998. “Testing the Effects of a Pre-Paid Incentive and Express Delivery to Increase Response Rates on a Random Digit Dial Telephone Survey.” Paper presented at the Annual Meeting of the American Association for Public Opinion Research, St. Louis, MO.
Cantor, David, Jane Waldfogel, Jeffrey Kerwin, Mareena McKinley Wright, Kerry Levin, John Rauch, Tracey Hagerty, and Martha Stapleton Kudela. 2001. Balancing the Needs of Families and Employers: Family and Medical Leave Surveys, 2000 Update. Rockville, MD: Westat.
Cantor, D., Cunningham, P., Triplett, T., and Steinbach, R. 2003. “Comparing Incentives at Initial and Refusal Conversion Stages on a Screening Interview for a Random Digit Dial Survey.” Paper presented at the Annual Meeting of the American Association for Public Opinion Research, Nashville, TN.
Cantor, David, Barbara O'Hare, and Kathleen O'Connor. 2007. “The Use of Monetary Incentives to Reduce Nonresponse in Random Digit Dial Telephone Surveys.” In Advances in Telephone Survey Methodology, edited by James M. Lepkowski, Clyde Tucker, J. Michael Brick, Edith D. De Leeuw, Lilli Japec, Paul J. Lavrakas, Michael W. Link, and Roberta L. Sangster, 471–498. New York: Wiley.
Daley, Kelly, Stas Kolenikov, Marci Schalk, Mehera Baugher, Julie Pacer, Jacob Klerman, and Lauren Dunton. 2016. Worker Classification Knowledge Survey: Volume II – Methodology Report. Cambridge, MA: Abt Associates.
DOL (U.S. Department of Labor, Commission on Leave). 1996. A Workable Balance; Report to Congress on Family and Medical Leave Policies. Washington, DC: U.S. Department of Labor. https://www.dol.gov/whd/fmla/1995Report/family.htm
Klerman, Jacob, Kelly Daley, and Alyssa Pozniak. 2013. Family and Medical Leave in 2012: Technical Report. Cambridge, MA: Abt Associates.
O’Hare, Barbara and Christina Frederick. 2003. “The Effect of a Pre-Paid and Promised Incentive on Response Rate.” Paper presented at the Annual Meeting of the American Association for Public Opinion Research, Nashville, TN.
Peytchev, Andy, Rodney K. Baxter, and Lisa R. Carley-Baxter. 2009. “Not All Survey Effort Is Equal: Reduction of Nonresponse Bias and Nonresponse Error.” Public Opinion Quarterly 17: 785–806. doi: 10.1093/poq/nfp03.
Singer, Eleanor, John Van Hoewyk, and Mary P. Maher. 2000. “Experiments with Incentives in Telephone Surveys.” Public Opinion Quarterly 64: 171–188.
Strouse, R. and Hall, J. 1997. “Incentives in Population-based Health Surveys.” Proceedings of the American Statistical Association, Survey Research Section, 952-957.
Zimowski, Michele, Roger Tourangeau, Rashna Ghadialy, and Steven Pedlow. 1997. “Nonresponse in Household Travel Surveys.” Prepared for the Federal Highway Administration. Chicago, IL: NORC.
1 Effective April 1, 2017, Abt SRBI will be consolidated into the existing Abt Associates Inc. government segment. Abt SRBI will cease to function as an independent entity and will become a part of the parent company. Abt Associates is in the process of executing this transition, including integrating current Abt SRBI employees into the parent company.
2 See Attachment A for the Employee Survey
3 Not firms; i.e., in a multi-site firm, the respondent should answer only for the sample worksite, not for the firm as a whole.
4 See Attachment B for the Employer Survey
5 The research literature indicates that we can expect about an 8 percentage point increase in the response rate if $15 is promised as opposed to $0 (Cantor et al. 2003; Strouse and Hall 1997).
6 See guidelines here: https://obamawhitehouse.archives.gov/sites/default/files/omb/inforeg/directive2/final_addendum_to_stat_policy_dir_2.pdf
7 Total costs include the full cost of instrument design and data collection activities. Annual costs include the total costs annualized over a three year period.
8 ($378,881+$2,323,119)/3=$900,667
9 See Office of Personnel Management 2017 Hourly Salary Table: https://www.opm.gov/policy-data-oversight/pay-leave/salaries-wages/salary-tables/pdf/2017/DCB_h.pdf.
10 200 hours x $60.83 x 1.6=$19,466
File Type | application/vnd.openxmlformats-officedocument.wordprocessingml.document |
File Modified | 0000-00-00 |
File Created | 2021-01-15 |