Part
A: Justification for the Collection of Data
for the National
Evaluation of the Performance Partnership Pilots for Disconnected
Youth (P3)
March, 2018
Submitted to:
Office of Management and Budget
Submitted
by:
Chief Evaluation Office
Office of the Assistant Secretary for Policy
United States Department of Labor
200 Constitution Avenue, NW
Washington, DC 20210
This page has been left blank for double-sided copying.
CONTENTS
Part A: Justification for the Study 1
A.1. Circumstances making the collection of information necessary 1
1. Overview of P3 2
2. Overview of the P3 evaluation 6
3. Data collection activities requiring clearance 7
A.2. Purposes and use of the information 8
1. Site visit interviews 8
2. Youth focus groups 9
3. Survey of partner managers 13
4. Survey of partner service providers 14
A.3. Use of technology to reduce burden 14
A.4. Efforts to avoid duplication 15
A.5. Methods to minimize burden on small entities 15
A.6. Consequences of not collecting data 15
A.7. Special circumstances 16
A.8. Federal Register announcement and consultation 16
1. Federal Register announcement 16
2. Consultation outside of the agency 16
A.9. Payments or gifts 16
A.10. Assurances of privacy 17
1. Privacy 17
2. Data security 19
A.11. Justification for sensitive questions 19
A.12. Estimates of hours burden 19
1. Hours by activity 19
2. Total estimated burden hours 21
A.13. Estimates of cost burden to respondents 22
A.14. Annualized costs to the federal government 22
A.15. Reasons for program changes or adjustments 24
A.16. Plans for tabulation and publication of results 24
1. Data analysis 24
2. Publication plan and schedule 25
A.17. Approval not to display the expiration date for OMB approval 26
A.18. Explanation of exceptions 26
REFERENCES 27
TABLES
Table A.1: Description of the 15 pilots 2
Table A.2. Data collection activity and instruments included in the request 8
Table A.3. Examples of topics for site visit interviews and focus groups, by respondent 11
Table A.4. Annual burden estimates for data collection activities (over 36 months) 20
Table A.5. Monetized burden hours, over 36 months 21
Table A.6. Estimated annual federal costs for the National Evaluation of P3 24
Table A.7. Schedule for the National Evaluation of P3 25
FIGURES
Figure A.1. P3 program model 6
The Chief Evaluation Office (CEO) of the U.S. Department of Labor (DOL) has contracted with Mathematica Policy Research and Social Policy Research Associates (hereafter “the study team”) to evaluate the Performance Partnership Pilots for Disconnected Youth (P3). In partnership, the P3 federal agencies— DOL; the Departments of Education (ED), Health and Human Services (HHS), Housing and Urban Development (HUD), and Justice (DOJ); the Corporation for National and Community Service (CNCS); and the Institute of Museum and Library Services (IMLS)—awarded grants to 15 P3 pilots to test innovative, cost-effective, and outcome-focused strategies for improving results for disconnected youth. This study provides information to policymakers and administrators that they can use to determine whether allowing states, localities, and Indian tribes greater flexibility to pool funds and waive programmatic requirements will help them overcome significant hurdles in providing effective services to and improving outcomes for disconnected youth.
This package request updates to the prior submitted and approved clearance package (Control number 1290-0013). In addition to two rounds of site visits to the P3 Cohort 1 pilots, this revised package includes follow-up telephone interviews with the Cohort 1 pilots, one round of site visits to the Cohort 2/3 pilots, and one site visit to each of six exemplary pilots as part of the evaluation’s implementation and systems analyses. The data collection activities included in this request are: (1) site visit and telephone interviews; (2) focus group discussions with P3 youth participants; (3) a survey of partner managers; and (4) a survey of partner service providers.
It is vital that young people today develop the skills, knowledge, and behaviors to help them successfully transition to adulthood, fulfilling their potential and advancing our nation’s social and economic prospects. Many, but not all, youth will gain these skills, knowledge, and behaviors and become self-sufficient, productive members of society and families (Dion et al. 2013). Youth who need additional supports for this transition have been called “at risk” or “disconnected,” but they also have been called “opportunity youth” by the White House Council for Community Solutions (2012) when it highlighted their promise and drew attention to them as a top national priority.
Although improving the outcomes of these youth has been a priority, stakeholders have identified critical barriers to serving these youth, including multiple federally funded programs with different eligibility and reporting requirements and multiple data systems across the local network of youth providers. P3 is testing whether granting the flexibility to blend and braid discretionary program funding and seek appropriate waivers will ameliorate the barriers to effective services identified by the field and ultimately improve the outcomes of these youth (U.S. Government 2014). Through a competitive grant process, the federal agencies awarded grantees this flexibility, starting with a cohort in fall 2015 as authorized in the 2014 Consolidated Appropriations Act and awarding additional pilots as authorized by Congress in 2015 and 2016. The evaluation of P3 pilots represents an important opportunity to study the implementation outcomes and system changes that the pilots can achieve and the pilot programs’ outcomes of and impacts on youth participants. This request focuses on data collection for the implementation and systems analyses.
In October 2015, nine competitively awarded grantees were announced as the first P3 cohort. They received up to $700,000 in start-up funds and the flexibility to blend or braid discretionary funds from fiscal year 2014 and 2015 to improve the outcomes of disconnected youth. The first cohort grantees are located in eight states (California, Florida, Illinois, Indiana, Kentucky, Louisiana, Oklahoma, and Washington) and a federally recognized Indian tribe located in Texas. Subsequently, six Cohort 2/3 grantees representing four states (Arizona, California, Connecticut, and New York) were awarded P3 grants.
As required in the legislation authorizing P3, the 15 grantees are serving disconnected youth, defined as low-income youth ages 14 to 24 and are either homeless, in foster care, involved in the juvenile justice system, unemployed, or not enrolled in or at risk of dropping out of school. Several grantees are serving in-school and out-of-school youth, and some are focusing on specific populations such as youth in foster care or public housing. Almost all of the Cohort 1 grantees are relying on their Workforce Innovation and Opportunity Act Title I Youth funds along with other DOL, ED, HHS, CNCS, or IMLS funds. Table A.1 provides additional information about the 15 pilots.
Table A.1: Description of the 15 pilots
Pilot name |
Location of pilot services |
Anticipated number of participants |
Estimated number of partners |
Target population |
Brief description of intervention |
Cohort 1 |
|||||
Baton Rouge P3 |
Baton Rouge, Louisiana |
84 |
6 |
14 to 24 year-olds who are 2 or more years behind in school |
Youth will develop an individual success plan. Program staff will develop training activities, and will encourage youth to participate in other training and education programs provided by partners. |
Best Opportunities to Shine and Succeed (BOSS) |
Broward County, Florida |
420 |
4 |
At-risk youth in six high schools |
Students will be provided a case manager. The case managers will have a 1:35 ratio of case manager to youth. The case manager will connect each participant with “evidence-based and evidence-informed” educational, employment, and personal development services that specifically address the needs of the student in regard to graduation and post-secondary success. The BOSS program will provide intensive, comprehensive, and sustained service pathways via a coordinated approach that helps youth progress seamlessly from high school to post-secondary opportunities. |
Chicago Young Parents Program (CYPP) |
Chicago, Illinois |
140 |
3 |
Low-income women ages 16 to 24 with at least one child younger than six |
CYPP is a parent engagement, education and employment program that combines two successful, research-based program models: employment and mentoring for youth and high quality comprehensive Head Start programming for children and families. All participants receive basic Head Start services plus additional mentoring, home visits, socializations, education planning, enrichment sessions, and employment. |
Indy P3 |
Indianapolis, Indiana |
80 |
8 |
At-risk, low-income youth ages 14-24; target youth in public housing |
Indy P3 will provide comprehensive, concentrated, and coordinated services for cohorts of very high-risk disconnected youth. Staff members called connectors (each serving 40 youth and families at a time) will develop individual service and success plans, link participants to core service providers, and share data across programs. Partners will emulate best practices and lessons learned from evidence-based models. |
Los Angeles P3 (LAP3) |
Los Angeles, California |
8,000 |
24 |
Youth ages 16 to 24 |
LA P3 is comprehensive service delivery system that coordinates and integrates the delivery of education, workforce and social services to disconnected youth. Partner agencies and WIOA youth contractors in the city of Los Angeles provide the program services. These are existing services: the aim of LAP3 is to enhance the availability of these services through the enhanced coordination of partner agencies. |
P3-OKC |
Oklahoma County, Oklahoma |
60-70 |
12
|
Foster youth ages 14 to 21 |
Youth will receive: (1) modified wraparound services more consistent with child welfare services; (2) an integrated plan of services to promote service integration and foster partnerships across nonprofit and public organizations; (3) the Check and Connect intervention designed to monitor school attendance, participation, and performance; and (4) enhanced vocational development, work, and/or career opportunities achieved through wraparound, educational options, and career aspects of students enrolled in career academies. |
Seattle-King County Partnership to Reconnect |
Seattle-King County, Washington |
200 |
3
|
Youth ages 16 to 24 |
The program will have three components: (1) strategic coordination of workforce development services with the state’s unique Open Doors policy, which provides K-12 funding for reengagement programs; (2) using AmeriCorps members to develop a regional outreach strategy aimed at placing the hardest-to-serve youth in programs that best reflect their interests and needs; and (3) advancing efforts toward a shared data system and common intake process that will enhance the coordination and targeting of services across Seattle-King County. |
Southeast Kentucky Promise Zone P3 |
7 rural southeast Kentucky counties |
1,000 |
3 |
At-risk youth ages 14 to 24 |
The program will include a teen pregnancy prevention program, career assessments and exploration trips, academic and career mentoring and tutoring, professional development for teachers and community members, two generations of family engagement focused on youths’ parents, and paid work experience. |
Tigua Institute of Academic and Career Development Excellence |
Ysleta Del Sur Pueblo tribe (Texas) |
45-50 |
2 |
Tribal youth members ages 14 to 17 enrolled in two local high schools |
Youth will receive group sessions of an integrated Leadership curriculum based on nation building theory and the Pueblo Revolt Timeline, which includes the Tigua lecture series to teach youth about their history, language and tribal government and the various services offered by the departments. Youth will also receive individually based wraparound services. |
Cohort 2/3 |
|||||
Partnerships Advancing Youth Together in Hartford |
Hartford, Connecticut |
100 |
5 |
Disconnected youth ages 16 to 24 residing in North Hartford Promise Zone |
Promise Zone PATH tests an innovative strategy of reconnecting youth to success by creating a virtual network between programs and connecting databases and applications of existing evidence-based programs. The dashboard identifies disconnected/ opportunity youth experiencing gaps in service, matches them to needed services and programs, and tracks their progress on a pathway to success that is sustainable. |
LA County P3 |
Los Angeles County, California |
963 |
4 |
18 to 24 year olds |
Public Service Career Mentoring for System Involved Youth has two levels. Level 1 focuses on preparing young adults for employment. It consists of curriculum and paid work experience that will be completed in about 6 months. Young adults who complete Level 1 may apply for Level 2, which is a competitive one-year internship with a county office. The young adult also is enrolled in WIOA and has a career mentor. |
New York City P3 |
Brooklyn, NY |
50 |
3 |
Young parents ages 17 to 24 and their children |
The program seeks to improve participants’ access to multiple career pathways through activities including guaranteed childcare, high school equivalency preparation, work readiness and occupational skills training (including paid work experiences), supportive services, and parenting support. The program also provides youth with personalized enrichment and support plans through a Child Care Navigator. |
New York State |
Albany, New York |
120 |
2 |
14 to 24 year olds who are unemployed, not in school, or reentering the community from juvenile justice services. |
The Connecting Youth in Transition intervention will use integrated case management, working with transition coordinators to jointly develop individual service plans with youth. Youth will be linked with existing educational and employment services, and the Transition Coordinator will continue to follow up and track progress for up to four months. |
Phoenix Manufacturing Apprenticeship Program |
Phoenix, Arizona |
96 |
6 |
Young adults ages 17 to 24 disconnected from school and work settings |
The program is a workforce development program where participants receive two curricula (the Gateway College Manufacturing Curriculum and the TCI Solutions soft skills curriculum). |
Sacramento Promise Zone Performance Partnership Pilot |
Sacramento, California |
100 |
9 |
Youth ages 16 to 24, who are homeless, aging out of foster care, and/or on probation |
The pilot will provide young adults with a housing voucher, continued case management from the referring agency, and be offered additional services to help them accomplish their education and employment goals. They will receive mental health services to support their transition to a new living situation. |
Sources: Cohort 1: Grantees’ presentations at annual P3 conference in June 2016, grantees’ draft evaluation plans, and grant applications. Cohort 2/3: grant application and telephone discussions.
The National Evaluation of P3 includes the provision of evaluation technical assistance to the P3 grantees and their local evaluators, an outcomes analysis based on administrative data already being collected by the grantees, and systems and process analyses. These latter two analyses, as depicted in the program model (Figure A.1), will seek to determine how the pilots operate at both the systems and program levels. They are the focus of this data collection package.
At the systems level, grantees seek to establish partnerships and work on goals such as integrating data systems and procedures or seeking approaches with more established evidence of effectiveness. Ideally these activities will promote effective collaboration and produce cost efficiencies, among other outputs—or achieve the “collective impact” model of broad participation and intensive focus of resources—leading to better system coordination and alignment, more integrated data systems, fewer barriers to effective supports for disconnected youth, and greater knowledge of what works to improve youth outcomes.
At the program/participant level, pilots may implement or expand programs or services for youth based on evidence-based models, recruit participants, and engage and retain youth. The goal of these activities and outputs are improved outcomes for youth such as employment, engagement or retention in education, and well-being. Activities, outputs, and outcomes at both levels will be influenced by contextual factors, such as the local economy and community conditions, and a set of challenges and opportunities that limit or enhance their progress.
The data collection activities described in this package will provide data for a systems analysis and an implementation evaluation of the first cohort of P3 grantees. This information will address five main research questions:
How do the pilots use the financial and programmatic flexibilities offered by P3 to design and implement interventions to improve the outcomes of disconnected youth?
How has each pilot aimed to leverage the P3 flexibilities to enhance its partnerships and work across partners to provide effective services to disconnected youth? To what extent have these aims materialized?
Who are the youth who participate in P3 and what services do they receive?
What systems and programmatic changes and efficiencies resulted from P3?
What lessons can be drawn for developing integrated governance and service strategies to improve the outcomes of disconnected youth?
This package requests clearance for four data collection activities conducted as part of the evaluation’s implementation and systems analyses: (1) site visit and phone interviews; (2) focus group discussions with program participants; (3) a survey of partner managers; and (4) a survey of partner service providers. The data collection instruments associated with these activities that require Office of Management and Budget (OMB) approval are:
Site visit and phone interview master protocol. In-person visits to the 15 grantees will provide information on the implementation of P3. The study team will conduct interviews with grantee administrators and staff, partner leaders and managers, and frontline staff. In addition to the site visits that occurred to the nine Cohort 1 pilots in 2017, we will conduct second round visits to Cohort 1 grantees in Spring 2018 and final telephone interviews with program directors in Fall 2018. In addition, we will conduct site visits to the six Cohort 2/3 pilots in Fall 2018, as well as visits to six exemplary grantees in 2018. Depending on the scope of the pilot, the visits will be from two to three days. The master interview protocol is provided as Instrument 1.
Youth focus group discussion guide. Focus group discussions with P3 youth participants will provide important information about youths’ program experiences, views on whether the program is meeting their needs, challenges that interfere with their participation, and suggestions for program improvements. To collect this information, the study team will conduct an average of three focus groups at each of the nine Cohort 1 pilots across the two site visits conducted to those pilots. The study team will conduct a single focus group during each of the Cohort 2/3 visits. The study team anticipates including an average of eight youth in each focus group, and that each interview will last about one hour. The youth focus group protocol is provided as Instrument 2, with Instruments 3 and 4 providing consent forms.
Partner manager survey. A survey of partner managers will provide systematic information about how partner managers perceive the P3 collaboration and the relationships across P3 partners. During all site visits except exemplar visits, the site visitor will administer a short survey (about 5 minutes) to partner managers after concluding the site visit interview with the partner manager. We anticipate administering this survey to an average of 10 partner managers per grantee. We provide the survey as Instrument 5.
Partner network survey. A survey of partners will provide systematic information about the relationship across providers. During the 2018 Cohorts 1 and 2/3 visits, the study team will administer a short survey (about 10 minutes) to direct service staff of each pilot partner. We anticipate conducting this survey once to up to 10 provider staff at each grantee. The survey will use social network analysis questions to focus on partners’ interactions with one another. The study team will field the survey in each pilot shortly after completing its first site visit. We provide the partner network survey as Instrument 6.
Table A.2 lists each instrument included in this request.
Table A.2. Data collection activity and instruments included in the request
Data collection activity/instrument |
Site visit and phone interviews 1. P3 master interview protocol |
Youth focus groups 2-4. Focus group discussion protocol and consents |
Survey of partner managers 5. Partner manager survey |
Survey of partner service providers 6. Partner network survey
|
The study team will use the data collected through activities described in this request to thoroughly document and analyze (1) the grantees’ local networks or systems for serving disconnected youth and how these systems changed as a result of P3; and (2) the grantees’ implementation of program services under P3. In Section A.16, Plans for tabulation and publication of results, we provide an outline of how the study team will analyze and report on all data collected.
The most important source of data for understanding local systems and program services will be in-person interviews with the staff of P3 grantees, partners, and service providers. During the first visit to each pilot (the visits to Cohort 1 pilots were conducted in 2017 and the ones to Cohort 2/3 pilots to be conducted in Fall 2018), the study team is collecting information on key topics. The topics include pilots’ start-up efforts, including initial discussions about the potential areas for blending and braiding of funds and programmatic waivers; the work of the lead agency in managing the P3 collaboration; the planned system changes—for example, those related to partners’ sharing of customers’ information, performance agreements, and management information systems; the process for mobilizing and communicating across key partners; and early implementation of the P3 program, including successes and challenges.
The second visit to the Cohort 1 pilots, planned for Spring 2018, will collect information on how systems, service models, and partnerships are evolving, including the process for identifying modifications to the grants, any new waivers, or funding changes (whether they required P3 authority or not); the pilots’ cost implications; and pilots’ sustainability plans, ongoing implementation challenges, and their solutions.
In addition, in Fall 2018, the study researchers will conduct telephone interviews with each Cohort 1 program director as the P3 grant ends. This interview will focus on sustainability of activities and lessons learned. As indicated, they also will conduct site visits to up to six pilots for which P3 has engendered the most change in community systems. The focus of these visits will be on best practices and scalability of program efforts.
The site visits, lasting from two to three days, will include semi-structured interviews with administrators and staff from the grantees and partners. The researchers conducting the visits will use a modular interview guide, organized by major topics that they can adapt based on the respondent’s knowledge base, to prompt discussions on topics of interest to the study.
Researchers with substantial experience in qualitative data collection of programs serving youth in different settings, including tribal communities, will conduct all interviews. Although experienced, all researchers will be trained before each data collection activity to ensure they have a common understanding of the P3 program, data collection goals, the data collection instruments, and the procedures for consistent collection and documentation of the data. A senior member of the P3 study team, who also has extensive qualitative research experience related to youth programs, will lead each pilot’s initial visit. An experienced qualitative researcher will support the lead visitor on the first round and conduct subsequent visits and telephone interviews.
The P3 Site Visits Master Protocol will guide these interviews (Instrument 1), and Table A.2 displays the topics that the study team will address with each activity.
The study team will conduct focus groups during each round of pilot site visits to learn about participants’ initial interest and enrollment in P3, program experiences, views on whether the program is meeting their needs, challenges that interfere with their participation, suggestions for program improvement, and expectations for the future. Across the two visits to each Cohort 1 pilot, the study team will conduct three focus groups with participating youth. Cohort 2/3 visits will include one focus group. Focus groups will not be conducted during exemplar visits.
We will coordinate with the lead agency to ensure that we invite and recruit a diverse set of program participants. Possible dimensions of diversity include: gender, race/ethnicity, age, and length of time in the program. Even though the focus group participants will not be representative of all pilot participants, they will offer perspectives on program operations and experiences that differ from those of staff members. The study team anticipates that each focus group will include an average of eight youth and will last about one hour. We will offer a $20 incentive to encourage participation. The Youth Focus Group Protocol will guide these on-site discussions (Instrument 2), and Table A.3 displays the topics that the study team will address with focus group participants.
Table A.3. Examples of topics for interviews and focus groups, by respondent
|
Interviews |
|
||||
Topics of interest |
Grantee lead |
Pilot manager |
Data systems manager |
Partner managers |
Front-line staff |
Focus groups of youth |
Community context |
||||||
Local labor market conditions |
|
|
|
|
|
|
Local network of organizations |
|
|
|
|
|
|
Network factors affecting delivery of services to youth |
|
|
|
|
|
|
Defining the pilots |
||||||
Organizational structure/system |
||||||
Lead agencies and roles |
|
|
|
|
|
|
Partners and roles |
|
|
|
|
|
|
Other community organizations/ stakeholders |
|
|
|
|
|
|
Program model and stage of development |
||||||
Theory of change |
|
|
|
|
|
|
Program model |
|
|
|
|
|
|
Needs identified |
|
|
|
|
|
|
Stage of development |
|
|
|
|
|
|
Flexibility |
||||||
Identified areas for flexibility |
|
|
|
|
|
|
Role of state and federal governments |
|
|
|
|
|
|
Funding |
||||||
Discretionary and other funding used for P3 |
|
|
|
|
|
|
Braiding and blending of funds |
|
|
|
|
|
|
Leveraged or other funding |
|
|
|
|
|
|
Pilot’s use of start-up funds |
|
|
|
|
|
|
Funding challenges |
|
|
|
|
|
|
Waivers |
||||||
Waivers requested and those approved |
|
|
|
|
|
|
Requests or considerations for additional waivers |
|
|
|
|
|
|
Partnerships, management, and communications |
||||||
Partner network |
||||||
Collaboration on designing P3 |
|
|
|
|
|
|
Prior (pre-P3) relationship |
|
|
|
|
|
|
Type(s) of partnership arrangements |
|
|
|
|
|
|
P3 partners’ shared vision |
|
|
|
|
|
|
P3 effect on partnerships |
|
|
|
|
|
|
P3 partnership effect on youth outcomes |
|
|
|
|
|
|
Partnership strengths and weaknesses |
|
|
|
|
|
|
Management and continuous program improvement |
||||||
Decision-making processes |
|
|
|
|
|
|
Management tools |
|
|
|
|
|
|
Assessment of performance |
|
|
|
|
|
|
Communications |
||||||
Mode and frequency |
|
|
|
|
|
|
P3-engendered changes |
|
|
|
|
|
|
Communications strengths and weaknesses |
|
|
|
|
|
|
The P3 program |
||||||
Development of P3 program/intervention |
||||||
Origins of intervention |
|
|
|
|
|
|
Use of evidence-based practices |
|
|
|
|
|
|
Differences in partners’ vision of program design |
|
|
|
|
|
|
Staff |
||||||
Structure and number |
|
|
|
|
|
|
Training, cross-training |
|
|
|
|
|
|
Communication across frontline staff |
|
|
|
|
|
|
Youth participants |
||||||
Target population |
|
|
|
|
|
|
Participant characteristics |
|
|
|
|
|
|
Eligibility criteria |
|
|
|
|
|
|
Recruitment process |
|
|
|
|
|
|
Any changes since design |
|
|
|
|
|
|
Intake and enrollment |
||||||
Intake/enrollment process and integration across partners |
|
|
|
|
|
|
Information sharing, access, and use |
|
|
|
|
|
|
P3-engendered changes to process, and challenges |
|
|
|
|
|
|
Youth services |
||||||
Menu of services and for which youth |
|
|
|
|
|
|
Roles of partners in service delivery |
|
|
|
|
|
|
Length and dosage of participation |
|
|
|
|
|
|
Definition of “completion” of the program |
|
|
|
|
|
|
Follow-up services |
|
|
|
|
|
|
Data systems and sharing |
||||||
Context |
||||||
Existing climate for sharing data |
|
|
|
|
|
|
Prior (pre-P3) data sharing and agreements |
|
|
|
|
|
|
Existing challenges |
|
|
|
|
|
|
P3 data systems and data |
||||||
Systems used to track participation and outcomes |
|
|
|
|
|
|
Systems shared and how |
|
|
|
|
|
|
Data agreements |
|
|
|
|
|
|
Data collected on participants |
|
|
|
|
|
|
Length, type, and methods of data collection |
|
|
|
|
|
|
How data are shared |
|
|
|
|
|
|
Follow-up data collected |
|
|
|
|
|
|
Data reports generated |
|
|
|
|
|
|
P3-engendered changes in systems, data, and sharing |
|
|
|
|
|
|
P3-related systems and data challenges |
|
|
|
|
|
|
Federal role and technical assistance |
||||||
Awareness of P3 |
|
|
|
|
|
|
Federal role and supports |
|
|
|
|
|
|
Interactions with federal agencies |
|
|
|
|
|
|
Interactions with other pilots |
|
|
|
|
|
|
Satisfaction with assistance |
|
|
|
|
|
|
P3-related interactions compared with other grant programs |
|
|
|
|
|
|
Assessing P3 |
||||||
Assessing P3’s potential for local change and innovation |
||||||
Ways P3 has affected community, network |
|
|
|
|
|
|
Factors hindering/facilitating innovation |
|
|
|
|
|
|
Service delivery and systems efficiencies |
||||||
Indications of improved efficiency |
|
|
|
|
|
|
Factors hindering/facilitating efficiencies |
|
|
|
|
|
|
Impact of P3 funding flexibility on number of youth served |
|
|
|
|
|
|
Sustainability |
||||||
Plans |
|
|
|
|
|
|
Potential for sustainability |
|
|
|
|
|
|
Perceptions of P3 and pilot’s success |
||||||
Assessment of overall initiative |
|
|
|
|
|
|
Perception of P3 as a governance model |
|
|
|
|
|
|
Perception of youth response |
|
|
|
|
|
|
Perception that pilot goals achieved |
|
|
|
|
|
|
Lessons learned |
||||||
Overall challenges and strategies to address them |
|
|
|
|
|
|
Plans for applying lessons learned |
|
|
|
|
|
|
The survey of partner managers will provide important systematic information about how the partner managers view collaboration within their communities. Although site visit interviews will provide important information on the P3 system and how the P3 partners have worked together since the initial application design through the implementation of the P3 pilot, the survey, which draws upon the Wilder Collaboration Factors Inventory (Wilder 2013), will provide a unique opportunity to assess key aspects of the partnership using quantitative data.
Following the conclusion of the on-site interview with a partner manager, we will request that the manager complete the short paper survey. The site visitor will strive to collect the survey while on site; but, if that is not feasible, he or she will provide a pre-addressed, pre-stamped envelope for the respondent to return the completed survey. With this in-person contact, we will seek a response rate of 100 percent.
During the 2018 site visits to the nine Cohort 1 and six Cohort 2/3 pilots, the study team will administer by email a short (about 10 minutes) survey to staff of partners who provide direct services to the P3 youth participants (we estimate up to 10 respondents per pilot). Drawing on prior network surveys the study team has conducted, the survey will ask about partner interactions with one another, enabling us to obtain more systematic and discrete information about partner relationships than we could obtain through site visit interviews alone. During the site visit, the study team will introduce the survey, talk with grantee and partner staff about each partner’s appropriate respondent for the survey, and encourage the partners’ participation when they receive the emailed survey. They will seek a response rate of 90 percent or higher by relying on support from the grantee, developing relationships with pilot staff during the first visit and grantee conferences, and encouraging survey completion during on-site interviews.
The National Evaluation of P3 will use multiple methods to collect study information and, when feasible and appropriate, will use technology to reduce the burden of the data collection activities on providers of the data.
Site visits and telephone interview. These interviews have relatively low burden, and the qualitative data to be collected do not benefit from technology, other than digitally recording interviews upon approval. We will avoid unnecessary data collection burden by covering topics not available from other sources.
Youth focus group. The study team will conduct youth focus group discussions in person without the use of information technology, other than digital recordings.
Partner manager survey. The site visit team will administer the survey on-site without the use of information technology.
Partner network survey. We will distribute the partner network survey by email. It does not contain or request sensitive or personally identifiable information (PII). Given the instrument’s brevity and the fact that it does not request or contain PII, using a PDF document attached to email is the least burdensome and most accessible means of collecting the data. Partner respondents can open the PDF attachment to the introductory email, enter their responses, and forward the email back to the sender with the completed document attached. They can do so at a convenient time and not be held to a scheduled appointment, as would be the case if data collection took place by phone or in person. The survey will ask each partner staff the same three questions about the other partners at that grantee.
The use of email enables self-administration of the P3 partner network survey, as well as tracking survey completes. We will use partner contact information, gathered during the P3 site visit, to distribute the survey to the partner staff identified as respondents for the survey. We will preload the full list of partners into the PDF document to obtain a response that relates to each partner. The PDF will allow respondents to enter responses (only check marks or Xs are necessary) but prevent them from revising any other text or information in the questionnaire. The survey does not contain complex skip patterns, and respondents will be able to view the question matrix with each possible category of response (across the top) and the full range of partners (down the side) on one sheet. This approach is commonly used for network analysis data collection to help respondents consider their levels of connectivity with all partners of the network and assess their relationships using a common set of considerations regarding the question of interest. The approach can only be used when the network is known ahead of time and the number of partners is relatively small, and it has the added advantage of facilitating data entry and analysis in that respondents provide information about all partners in the network. If respondents are not able to complete the survey in one sitting, they may save the document and return to it at another time, further reducing the burden on the respondent.
The data the study team is collecting from the site visit and telephone interviews, youth focus group, and partner network survey for the National Evaluation of P3 are not otherwise available from existing sources. We will conduct interviews with some of the same staff during the repeated data collections with Cohort 1 pilots but the questions will be new or will gather updated information of pilot implementation and sustainability plans. The first visit will focus on pilots’ planning, partnership building, and early program implementation. The second visit will focus on understanding any changes that have occurred in the system since the first visit and program service levels, challenges and successes, and plans for sustainability. The exemplar visits will focus on particular aspects of pilots’ activities that have promising practices and lessons for other performance partnership pilots. We will not ask youth participants to participate in more than one focus group, and we will conduct the partner network survey after the second Cohort 1 visit and the Cohort 2/3 visits. We will conduct the partner manager survey in both rounds of site visits in order to track any changes in P3 partnerships.
We have developed the instruments to minimize burden and collect only critical evaluation information.
The federal government is intent on learning how P3 has helped local communities overcome barriers to effectively serve disconnected youth and improve youths’ outcomes. Without the information collected as part of the study, federal policymakers will not have an analysis of how P3 has affected local systems for serving disconnected youth and P3-engendered programmatic changes to improve the outcomes of disconnected youth. Information collected will be important for informing future rounds of P3 as well as other federal initiatives granting administrative flexibilities to grantees of discretionary funding.
No special circumstances apply to this data collection. In all respects, we will collect the data in a manner consistent with federal guidelines. No plans require respondents to report information more often than quarterly, submit more than one original and two copies of any document, retain records, or submit proprietary trade secrets.
The 60-day notice (81 FR 31664) to solicit public comments was published in the Federal Register on May 19, 2016. No comments were received.
Consultations on the research design, sample design, and data needs were part of the study design phase of the National Evaluation of P3. These consultations ensured the technical soundness of study sample selection and the relevance of study findings and verified the importance, relevance, and accessibility of the information sought in the study.
Mathematica Policy Research
Jeanne Bellotti
Cay Bradley
Karen Needels
Linda Rosenberg
Social Policy Research Associates
Andrew Wiegand
The study team plans to offer gift cards to youth participating in the focus groups respondents as part of the data collection activities described in this clearance request. Each youth participant will receive a $20 gift card in appreciation of his or her contributions toward the research. Previous research has shown that sample members with certain socioeconomic characteristics, particularly those with low incomes and/or low educational attainment, have proven responsive to incentives (Duffer et al. 1994; Educational Testing Service 1991). We will not offer site visit interview respondents and partner network survey respondents any payments/gifts because their participation can be considered part of their regular work responsibilities, given the grant.
We are conducting the National Evaluation of P3 in accordance with all relevant regulations and requirements, including the Privacy Act of 1974 (5 U.S.C. 552a); the Privacy Act Regulations (34 CFR Part 5b); and the Freedom of Information Act (5 CFR 552) and related regulations (41 CFR Part 1-1, 45 CFR Part 5b, and 40 CFR 44502). Before they participate in a focus group, we will ask youth 18 or older for their consent to participate. For potential participants who are younger than 18, we will ask for their parent’s or guardian’s consent for their child to participate and then collect the youths’ assent to participate before proceeding with the focus group.
We will notify all interview respondents that the information they provide is private, that all data reported in project reports will be de-identified, and that the study team will carefully safeguard study data. All study team site visitors and interviewers will receive training in privacy and data security procedures.
Site visits, interviews, and youth focus groups. No reports shall identify P3 sites and youth focus group participants, nor will the study team share interview notes with DOL or anyone outside of the team, except as otherwise required by law. Site visit interviews and focus groups will take place in private areas, such as offices or conference rooms. At the start of each interview, the study team will read the following statement to assure respondents of privacy and ask for their verbal consent to participate in the interview:
Everything that you say will be kept strictly private within the study team. The study report will include a list of the P3 grantees and their partners. All interview data, however, will be reported in the aggregate and, in our reports, we will not otherwise identify a specific person, grantee, or partner agency. We might identify a pilot by name or a type of organization or staff position if we identify a promising practice.
This discussion should take about <duration> minutes. Do you have any questions before we begin? Do you consent to participate in this discussion?
<If recording interview>: I would like to record our discussion so I can listen to it later when I write up my notes. No one outside the immediate team will listen to the recording. We will destroy the recording after the study is complete. If you want to say something that you do not want recorded, please let me know and I will be glad to pause the recorder. Do you have any objections to being part of this interview or to my recording our discussion?
This statement is available at the top of the P3 Site Visit Master Protocol (Instrument 1).
We will ask youth recruited for and participating in focus groups to sign a consent form. If the youth is younger than 18, we will ask program staff to collect parent/guardian consent before allowing the minor to participate. The consent forms are attached (Instruments 3 and 4). At the start of the focus group, the facilitator will also indicate that the comments will be kept private:
To help us better understand how [PROGRAM NAME] is working, we would like to ask you some questions about how you came to participate in it and your experiences. This discussion will be kept private. We will not share any information you provide with staff from [PROGRAM NAME]. In addition, our reports will never identify you by name. Instead, we will combine information from this discussion with information from discussion groups in other programs. Participants’ comments will be reported as, “One person felt that. . .” or “About half of the participants did not agree with…”
I hope you will feel free to talk with us about your experiences. I ask that none of you share what you hear with others outside the group. It will also help me if you speak clearly and if you will speak one at a time. The discussion should last about one hour.
I’d like to record the discussion so we don’t have to take detailed notes and can listen carefully to what you are saying. The recording is just to help me remember what you say. No one outside of the research team will have access to the tape. Are there any objections?
Let’s get started. [HIT THE RECORD BUTTON].
I have hit the record button. Any objections to recording this discussion?
This statement can is available at the top of the P3 Youth Focus Group Protocol (Instrument 2).
Partner manager survey. We will administer the survey so that we maintain respondents’ privacy. The introduction to the survey contains the following statement: “All of your responses will remain private and will not be shared with anybody from outside the evaluation team; nobody from the grantee, the community partners, or federal partners will see your responses.” This statement is available at the beginning of the survey (Instrument 5).
However, for the study team’s analysis of partners’ perceptions of collaboration and how it changes over time, it will be important to identify the respondent and his or her partner organization. Thus, prior to the site visit, the survey team will generate identification numbers for each partner manager and affix labels with the number onto the survey handed to the respondent. The site visitor will be responsible for handing the appropriate survey to each partner manager. We will further protect respondents’ privacy by providing them with a pre-addressed, pre-stamped envelope to return the survey in the event that they are unable to hand it directly to the site visitor upon completion.
Partner network survey. No reports shall identify the survey respondents. The survey instrument will request only the name of the organization and the respondent’s job title and responsibilities. All other data items that identify survey respondents—respondent name and contact information—will be stored in a restricted file that only the study team can access. As the study team is not requesting respondent names as part of the survey, analysis files will also not contain respondent names. The introduction to the partner network survey contains the following statement assuring respondents of privacy: “Your name and responses will be kept private to the extent of the law. Findings from the survey will be reported in aggregate form only so that no person can be identified.” This statement is available at the beginning of the survey (Instrument 6).
To further remove any connection between individuals and their partner network analysis survey responses, we will save each completed survey immediately upon receipt in a secure project folder on Mathematica’s restricted network drives. The saved survey will indicate only the organizational affiliation of the respondent and the P3 partner. We will then delete the survey document from the return email to prevent it from being backed-up on the email servers.
Mathematica’s security staff and the study team will work together to ensure that all data collected as part of the study—including data collected as part of site visits, including interviews, focus groups, and partner manager surveys; and through the partner network survey (including interview recordings)—are handled securely. As a frequent user of data obtained from and on behalf of federal agencies, Mathematica has adopted federal standards for the use, protection, processing, and storage of data. These safeguards are consistent with the Privacy Act, the Federal Information Security Management Act, OMB Circular A-130, and National Institute of Standards and Technology security standards. Mathematica strictly controls access to information on a need-to-know basis. Data are encrypted in transit and at rest using Federal Information Processing Standard 140-2-compliant cryptographic modules. Mathematica will retain the data collected on the National Evaluation of P3 for the duration of the study. We will completely purge data processed for the National Evaluation of P3 from all data storage components of the computer facility in accordance with instructions from DOL. Until this takes place, Mathematica will certify that any data remaining in any storage component will be safeguarded to prevent unauthorized disclosure.
The instruments associated with the National Evaluation of P3 do not contain questions of a sensitive or personal nature. We will not request any personal information from respondents interviewed during site visits, other than the number of years served in their current employment position (interviews) or their age (focus groups). The interviews focus on respondents’ knowledge, experiences, and impressions of P3. Nonetheless, we will inform respondents that they do not have to respond to any questions they do not feel comfortable answering.
Table A.4 provides the annual burden estimates for each of the four data collection activities for which this package requests clearance. All of the activities will take place over 36 months. Total annual burden is 295 hours.
Site visit and telephone interviews with administrators and staff. Interviews with P3 administrators/managers and frontline staff will last, on average, 1.25 hours. Most will be one-on-one interviews, but we anticipate that some of the frontline-staff interviews will be with small groups of two to three staff. We estimate the maximum total hours for P3 data collection during the nine Cohort 1 site visits, excluding the P3 program director, will be 315, which includes 157.77 hours per each round of site visits (9 sites × 14 respondents × 1.25 hour per interview). We will interview the Cohort 1 P3 program directors three times (twice on site and once by phone) for maximum total hours of 333.75 hours (9 sites × 1 respondent × 3 responses × 1.25 hour per interview). We estimate the maximum total hours for P3 data collection at the six Cohort 2/3 sites will be 112.5 hours (6 sites × 15 respondents × 1.25 hour per interview). Finally, we estimate the maximum total hours for P3 data collection during exemplar site visits will be 112.5 hours (6 sites × 15 respondents × 1.25 hour per interview).
Youth focus groups. We expect to conduct an average of three one-hour focus group discussions with youth participants at each of the nine P3 Cohort 1 sites across the two rounds of site visits. We expect to conduct one focus group at each of the six Cohort 2/3 sites. We expect that eight youth will attend each focus group. Thus, we estimate the total maximum reporting burden for the Cohort 1 youth focus groups will be 216 hours (9 sites × 3 interviews per pilot × 8 respondents in each group × 1 hour per group discussion). The burden for the Cohort 2/3 pilots will be 48 hours (6 sites × 1 interview per pilot × 8 respondents in each group × 1 hour per group discussion). We estimate that half of the youth will be younger than 18. The total burden for parents across cohorts to provide consent is 2.2 hours (132 parents * 1 minute).
Partner manager survey. We expect to conduct the survey with approximately 10 partner managers (15 P3 sites with an average of 10 partner managers per site). The survey will take an average of 5 minutes (.08 hours) for the partner manager to complete. We will administer the survey in the two Cohort 1 site visits and the Cohort 2/3 visits.
The total estimated reporting burden for the Cohort 1 P3 partner managers participating in the survey is 15 hours (9 P3 sites × 10 partners × 0.08 hours (5 mins.) per survey × 2 survey rounds). The burden for Cohort 2/3 is 5 hours (6 P3 sites × 10 partners × 0.08 hours (5 mins.) per survey × 1 survey rounds).
Partner network survey. We expect to conduct the survey with approximately 150 partners (15 P3 sites with up to 10 partner staff per site). We expect the survey will take about 10 minutes (0.17 hours) to complete, on average, per respondent. We will stagger survey administration, enabling us to test administration procedures in one pilot before administering to all sites.
The total estimated reporting burden for the P3 partners participating in the survey is 25 hours (15 P3 sites × 10 partners × 0.17 hours per survey × 1 survey round).
Table A.4. Annual burden estimates for data collection activities (over 36 months)
Respondents |
Total number of respondents over evaluation |
Number of responses per respondent |
Annual Number of Responses |
Average burden time per response |
Total burden hours over evaluation |
Annual burden hours |
|
Site visit/phone interviews |
|
|
|
|
|
|
|
Administrators and staff (Cohort 1) |
126 |
2 |
84 |
1.25 hours |
315 |
105 |
|
P3 Program directors (Cohort 1) |
9 |
3 |
9 |
1.25 hours |
33.75 |
11.25 |
|
Administrators and staff (Cohort 2/3) |
90 |
1 |
30 |
1.25 hours |
112.5 |
37.5 |
|
Administrators and staff (Exemplar) |
90 |
1 |
30 |
1.25 hours |
112.5 |
37.5 |
|
Focus group discussions |
|
|
|
|
|
|
|
Youth (Cohort 1) |
216 |
1 |
72 |
1 hour |
216.0 |
72.0 |
|
Youth (Cohort 2/3) |
48 |
1 |
16 |
1 hour |
48.0 |
16.0 |
|
Partner manager survey |
|
|
|
|
|
|
|
Partner managers (Cohort 1) |
90 |
2 |
60 |
5 minutes |
15.0 |
5.0 |
|
Partner managers (Cohort 2/3) |
60 |
1 |
20 |
5 minutes |
5.0 |
1.7 |
|
Partner network survey |
|
||||||
Frontline staff (Cohort 1) |
90 |
1 |
30 |
10 minutes |
15.0 |
5.0 |
|
Frontline staff (Cohort 2/3) |
60 |
1 |
20 |
10 minutes |
10.0 |
3.3 |
|
Total |
879 |
-- |
371 |
-- |
883 |
294 |
The total estimated maximum hours of burden for the data collection included in this request for clearance is 885 hours (see Table A.4), which equals the sum of the estimated burden for the semi-structured site visit/phone interviews, youth focus groups, the partner manager survey, and the partner network survey (573.75 + 266.2 + 20+ 25 = 885).
The total monetized burden estimate for this data collection is $17,694.37 (Table A.5). Using the average hourly wage of social and community service managers taken from the U.S. Bureau of Labor Statistics, National Compensation Survey, 2015 (http://www.bls.gov/ncs/ncspubs_2015.htm), the cost estimate for this staff burden is $33.38. Therefore, the maximum cost estimate for grantee and partner managers to participate in interviews is $4,506.30 (135 × $33.38). The cost for partner managers to participate in the partner manager survey is $501 (15 × $33.38).The average hourly wage of miscellaneous community and social service specialists taken from the U.S. Bureau of Labor Statistics, National Compensation Survey, 2015, is $19.36. Therefore, the cost estimate for frontline staff from across different grantees and partners to participate in site visit interviews is $3,920.40 (202.5 hours × $19.36). The cost for frontline staff to participate in the survey is $290.10 (15 hours × $19.36). We assume that cost for youth participation is the federal minimum wage ($7.25 per hour) for a cost of $1,566 (216 hours × $7.25). We assume that cost for parents’ participation is the average minimum wage in the 12 states with a P3 pilot ($8.85 per hour) for a cost of $19.47.
Table A.5. Monetized burden hours, over 36 months
Respondents |
Total maximum burden (hours) |
Type of respondent |
Estimated hourly wages |
Total indirect cost burden |
Annual monetized burden hours |
Semi-structured interviews |
|||||
Grantee and partner managers |
249.75 |
Manager |
$33.38 |
$8,336.66 |
$2,778.89 |
Frontline staff |
324 |
Frontline staff |
$19.36 |
$6,272.64 |
$2,090.88 |
Subtotal |
573.75 |
-- |
-- |
$14,609.30 |
$4,869.77 |
Youth focus groups |
|||||
Youth |
264 |
Youth |
$7.25 |
$1,914.00 |
$638.00 |
Parents |
2.2 |
Parents |
$8.85* |
$19.47 |
$6.49 |
Subtotal |
266.6 |
-- |
-- |
$1,933.47 |
$644.49 |
Partner manager survey |
|||||
Partner managers |
20 |
Partner manager |
$33.38 |
$667.60 |
$222.53 |
Partner network survey |
|||||
Frontline staff |
25 |
Partner staff |
$19.36 |
$484 |
$161.33 |
Total |
885 |
-- |
-- |
$17,694.37 |
$5,898.12 |
*Average minimum wage in 12 states with a P3 pilot.
There will be no direct costs to respondents for the National Evaluation of P3.
DOL, like most other federal agencies, uses contracts with firms that have proven experience with program evaluation to conduct all evaluation activities. Federal employees will rely on contract staff to perform the majority of the work described in this package, and have no direct role in conducting site visit discussions or focus groups, developing study protocols or designs, the direct collection of data using these instruments, or the analysis or production of reports using these data. The role of federal staff is almost entirely restricted to managing these projects. The costs incurred by contractors to perform these activities are essentially direct federal contract costs associated with conducting site visits, discussions, and focus groups.
This estimate of federal costs is a combination of (1) direct contract costs for planning and conducting this research and evaluation project, including any necessary information collection and (2) salary associated with federal oversight and project management.
Estimates of direct contract costs. There are two categories of direct costs to the federal government associated with conducting this project. These costs are routine and typical for studies such as this. The first category is design and planning, including external review of the design by a technical working group of outside subject matter experts, and development of instruments. This work is estimated to cost $677,850. The second category is data collection and reporting, which will occur through the project period, and is estimated to cost $2,534,703. The total estimated direct costs are:
$677,850 (design) + $2,534,703 (data collection and reporting) = $3,212,553
Although this project is expected to last five years, an accurate estimate of the annualized direct contract cost will vary considerably from year to year because the tasks are focused on specific periods in the project life cycle. The design and planning costs are obviously front-loaded, the data collection costs will be incurred throughout the project, and the analysis and reporting costs will occur close to the end of the project. As a very basic estimate, the total estimated direct costs can cost can be divided by the five years of the study to produce an estimate of the average annualized cost (see Table A.6):
$3,212,553 / 5 years of study = $642,510.60 per year in estimated direct contract costs.
Estimates of federal oversight and project management costs. Staff in the Office of the Chief Evaluation Officer have regular duties and responsibilities for initiating, overseeing, and administering contracts to perform research and evaluation on behalf of agency programs and offices. In the event that OMB approves this information collection request, federal staff would need to perform certain functions that, although clearly part of their normal duties, would be directly attributable to this specific research and evaluation project. For purposes of calculating federal salary costs, DOL assumes:
An Evaluation Team Manager, GS-15, step 2, based in the Office of the Chief Evaluation Officer in Washington, D.C.,1 who would earn $66.966 per hour to perform this work, and would spend approximately one-eighth of his or her annual time (2,080 hours / 8 = 260 hours) on this project. Total estimated annual federal costs for this individual are 260 hours × $66.96hour = $17,409.60.
A Senior Evaluation Specialist, GS-14, step 2, based in the Office of the Chief Evaluation Officer in Washington, D.C., who would earn $56.93 per hour to perform this work, and would spend approximately one-fourth of his or her annual time (2,080 hours / 4 = 520 hours) on this project. Total estimated annual federal costs for this individual are 520 hours × $56.93/hour = $29,603.60.
Table A.6. Estimated annual federal costs for the National Evaluation of P3
Estimates of annual federal costs |
|
Direct contracts costs |
$642,510.60 |
Federal oversight and management |
|
1 GS-15 (1/8 time) |
$17,409.60 |
1 GS-14 (1/4 time) |
$29,603.60 |
Subtotal for federal oversight and management |
$47,013 |
Total annual cost |
$689,523 |
Note: Federal staff costs are drawn from the most current available estimates of wages and salaries available at https://www.opm.gov/policy-data-oversight/pay-leave/salaries-wages/salary-tables/16Tables/html/DCB_h.aspx.
In 2017, the Federal P3 partners awarded P3 grants to six additional pilots (known as the Cohort 2/3 pilots). With the goal of producing a comprehensive study of P3’s implementation and systems work, the U.S. Department of Labor (DOL) modified the P3 evaluation contract with Mathematica to add one site visit to each of the six Cohort 2/3 pilots as well as visits to each of six exemplar pilots. In addition, phone interviews to each Cohort 1 program director were also added to the evaluation plan. The additional site visits and phone interviews will use instruments already approved that were used in the first round of visits to cohort 1 pilots (OMB Control No. 1290-0013). No other changes to the implementation and systems study are planned.
The National Evaluation of P3 will use the rich information collected from all sources to describe the P3 grantees’ systems and models for delivery services to disconnected youth. The analysis plan consists of a mixed-method approach with three steps: (1) organize the qualitative data from the site visits and focus groups; (2) identify themes and patterns in the data within and across grantees; and (3) conduct a network analysis using data from the partner survey.
Organize the qualitative data. Analyzing qualitative data is inherently challenging because it requires combining information from different sources, a great deal of which is unstructured. Our first strategy to manage the volume of data will be to develop structured templates and checklists for site visitors to distill the information they collect during site visit interviews and focus groups. Through these templates, site visitors will respond to specific questions and avoid long narratives on particular topics of interest. Our second strategy will be to lay an analytic foundation by organizing the data from the site visits and focus groups using qualitative data analysis software, such as NVivo.
Identify themes and patterns in the data. A critical part of the analytic approach will be to draw on multiple sources, including different respondents within a P3 pilot, and interview and programmatic data, to triangulate the data. Both agreements and discrepancies in respondents’ responses or across data sources can provide useful information on pilots’ implementation experiences and their successes and challenges. Within each pilot, we will analyze information from interviews on the effects P3 has had on its system or network for providing services to disconnected youth and how the provision of services to these youth has changed. After the study team has analyzed and organized all of the site visit data across all pilots, it will examine the data across the pilots to look for similarities in system changes and models of organization, service delivery, or other characteristics.
Analyze the partner manager survey data. The survey will explore the quality of the P3 partnerships from the partner manager perspective. We will tabulate the responses of the survey by pilot and also explore responses by partner types, for example, public and private partners, to analyze differences between them. We also will conduct simple tabulations and analyses to analyze changes in collaboration between the first and second site visits.
Conduct a network analysis using data from the partner network survey. The survey will explore the structure and strength of the networks that P3 grantees created to serve disconnected youth by assessing a number of specific characteristics of each grantee network. The survey will gather information about the frequency of communication and the change in communication over time, and the helpfulness of various partners in serving disconnected youth. The study team will not request respondents’ names on the survey instrument, only organization names. Further, although the study team will conduct the analysis separately for each P3 grantee, individual partners will not be identified in the presentation of findings. Instead, we will discuss partner networks by types of partners, not specific partner entities. In this way, results from the partner network survey will not reveal identities of any respondents.
The study team will use two primary measures to describe and depict service delivery networks within and across the P3 grantees: density (interconnectedness) and centrality (prominence). Density is the proportion of possible relationships that are actually present and measures the extent to which each partner is connected with all others across the network as a whole. Centrality examines the prominence of individual entities within the network by identifying the partner entities that are most sought after (indegree centrality). The study team will examine the measures of prominence for specific partners within the select networks across the two measures for comparison. We expect that we may find differences in the network interconnectedness and centrality of partners based on any communication and based specifically on changes in communication.
Using sociograms, the study team will illustrate the patterns in the size of partner networks, the strength of the relationships across partners, and the direction of partnerships. These sociograms will depict the density and centrality of P3 networks based on (1) contact frequency and (2) change in contact frequency since becoming involved in P3.
In addition to sociograms, the study will produce tables that present network-level characteristics such as overall density and centralization (measures discussed above). The study team will also present figures of helpfulness ratings in the P3 pilots, illustrating the centrality of specific partners in each network.
We will present findings from the evaluation in interim and final reports. Table A.7 shows the schedule for the study.
Table A.7. Schedule for the National Evaluation of P3
Activity |
Date |
Conduct Cohort 1 round 1 data collection (site visit interviews and focus groups, and partner manager survey) |
Completed (May– September 2017) |
Interim report |
March 2018 |
Conduct Cohort 1 round 2 data collection (site visit interviews, focus groups, partner manager survey, and partner network survey) |
April – June 2018 |
Conduct Cohort 2/3 data collection (site visit interviews, focus groups, partner manager survey, and partner network survey) |
September –November 2018 |
Conduct exemplar grantees data collection (site visit interviews) |
Late 2018 |
Conduct Cohort 1 phone interviews |
Fall 2018 |
Final implementation study report |
June 2019 |
Final evaluation report |
December 2020 |
The OMB approval number and expiration date will be displayed or cited on all forms completed as part of the data collection.
No exceptions are necessary for this information collection.
Dion, R., M.C. Bradley, A. Gothro, M. Bardos, J. Lansing, M. Stagner, and A. Dworsky. “Advancing the Self-Sufficiency and Well-Being of At-Risk Youth: A Conceptual Framework.” OPRE Report 2013-13. Washington, DC: U.S. Department of Health and Human Services, Office of Planning, Research and Evaluation, Administration for Children and Families, 2013.
U.S. Government. “Consultation Paper. Changing the Odds for Disconnected Youth: Initial Design Considerations for the Performance Partnership Pilots, Washington, DC: U.S. Government, 2014.
White House Council for Community Solutions. “Final Report: Community Solutions for Opportunity Youth.” 2012. Available at http://www.serve.gov/sites/default/files/ctools/12_0604whccs_finalreport.pdf. Accessed February 8, 2016.
Wilder Foundation. “Wilder Collaboration Factors Inventory.” 2013. Available at https://www.wilder.org/Wilder-Research/Research-Services/Documents/Wilder%20Collaboration%20Factors%20Inventory.pdf. Accessed February 9, 2016.
1 https://www.opm.gov/policy-data-oversight/pay-leave/salaries-wages/salary-tables/pdf/2018/DCB.pdf
File Type | application/vnd.openxmlformats-officedocument.wordprocessingml.document |
File Title | Part A: Justification for the Collection of Data for the Institutional Analysis of American Job Centers |
Author | MATHEMATICA |
File Modified | 0000-00-00 |
File Created | 2021-01-21 |