This template is designed to help evaluators plan and write evaluation reports using high quality methods. An analysis plan, developed before beginning any analyses, demonstrates an evaluator’s commitment to being objective and not being swayed by the data in hand—in other words, following a pre-specified, systematic, and scientific approach. The analysis plan both ensures funders, the program staff, the tribal communities, and potential skeptics that the outcomes and approaches to analyzing outcomes were pre-selected and justifies those approaches. The analysis plan serves as a road map for the planned analyses, which is helpful should you experience any changes in key project staff. We encourage evaluators to share analysis plans with a colleague who can carefully review the methods and approaches.
Descriptive evaluations of Tribal PREP programs include implementation (or process) studies and outcomes studies. An implementation study aims to answer questions about how the program was delivered in a particular context. Generally, an outcomes study seeks to answer questions about whether and how the outcomes for program participants changed during the program.
Both studies might also investigate questions related to the cultural relevance or significance of the program, including any adaptations that were made to address the needs of the communities involved in the projects. To the extent that your project aims to answer questions like these, be sure to include the relevant information throughout the analysis plan. For example, if you plan to measure how well certain program adaptations were implemented, be sure the program description includes those adaptations.
The analysis plan template is organized as follows. Section A discusses the elements needed for the analysis plan for the implementation study. Section B discusses the elements needed for the analysis plan for the outcomes study. Within each section, subsections 1–3 revisit proposed research questions, program characteristics, and the study design. Subsection 4 provides the blueprint for the analyses that will be conducted for each evaluation. Your responses should reflect your approved evaluation plan and research questions but may include additional details and information. Highlight any changes from the approved evaluation plan and research questions. The general structure of the template should work for all grantees regardless of the details of their evaluations, but some grantees might need to adapt some subsections to fit their design.
Please email your analysis plan to your Federal Project Officer (FPO) and copy your rigorous evaluation technical assistance (RETA) liaison by [date]. For consistency, use this naming convention when submitting your analysis plan: [Grantee Name] Evaluation Analysis Plan. If you have made substantial changes to your evaluation (for instance, sample size, eligible population, or data collection plan) since your last abstract submission, submit an updated abstract as well, with tracked changes. The latest version of your abstract is available in your grantee SharePoint folder, in the Abstract subfolder. Your FPO and RETA liaison will review the analysis plan, along with the latest version of your abstract for context, and will provide comments and suggested edits and return to you for revisions. FYSB would like your FPO to be able to approve your analysis plan and updated abstract (if applicable) by [date] so that you can begin work on the final report.
The research questions articulate the main hypotheses of your implementation evaluation. Research questions can refer to implementation elements such as fidelity, dosage, quality of implementation, engagement, and context. Use the research questions from your approved evaluation plan and make note of any revisions to those questions. Present the research questions in a table, organized by implementation elements. Table 1 provides an example.
Implementation elements:
Fidelity: Whether the program (and each of its activities, if applicable) was delivered as intended
Dosage: The amount of programming (or activities, if applicable) youth received
Quality: How well the program was delivered to youth
Engagement: Whether youth engaged in program services, and if so, how engaged they were
Context: Other types of services available to program participants, external events, or unplanned adaptations that might have affected implementation.
Table 1. Examples of research questions for each implementation element
Implementation element |
Research question |
Fidelity |
|
Dosage |
|
Quality |
|
Engagement |
|
Context |
|
Use a table to clearly and succinctly summarize your program’s components. See Tables 2 and 3 for examples (sample text included in italics). If you are implementing different program models or curricula with different youth (for example, Curriculum A is offered to 6th graders and Curriculum B is offered to 8th graders), repeat the content in this section for each program model or curriculum being evaluated as part of your implementation study. You can use text from your approved evaluation plan or abstract to complete the tables. These definitions offer guidance on what to include in each section of the tables.
Intended program activities. List each of the key structural elements of the program (for example, group classes, one-on-one services) (see Table 2 for an example).
Intended program content. For each activity, list the topics to be covered and the resources/materials provided. Also, indicate whether specific content is based on a particular curriculum. Include any modifications planned before the program was implemented and the reasons for modification. Include a list of the APS topics covered in each activity.
Planned dosage and implementation schedule. Describe the number of sessions and the duration of each component of the program. Include the length of each session and how frequently sessions occur. Describe variation in the frequency or length of sessions across sites, if applicable.
Intended delivery (location and setting). Describe where the program component takes place.
Characteristics, education and training of staff. Provide information on the required education, gender, cultural background, and hiring requirements of the providers/facilitators of each component, and on the training and technical assistance activities that were planned for providers to maintain fidelity.
Table 2. Description of intended program activities for [Program Name]
Program activity |
Curriculum and content |
Dosage and schedule |
Delivery |
Workshop |
Youth Pride curriculum
APS topics: Adolescent development |
8 hours, one hour per week |
Group workshop in classrooms |
Informational text messages |
Reminders from lessons on strategies to prevent HIV infection APS topics: Adolescent development |
12 text messages sent over one year |
Text messages sent by a centrally managed service with reminder facts from lessons |
Table 3. Staff training and development to support program activities for [Program Name]
Program activity |
Education and initial training of staff |
Ongoing training of staff |
Workshop |
Facilitators are male and female and hold at least a bachelor’s degree and receive four days of initial training. |
Facilitators receive a half-day of semiannual refresher training in the program’s curricula from study staff. |
Informational text messages |
N/A |
N/A |
The evaluation design should explain the sample formation, including the consent process, data sources, and outcomes, for your implementation evaluation. Aim to provide a detailed account of your methods, sufficiently transparent so that someone could replicate them. Include the following sections:
Describe the sample you will use to answer each research question. For example, you might have interviewed staff about curriculum delivery or conducted focus groups of a subset of participants in the program. For each sample, provide any required characteristics for sample inclusion (for example, age, AI/AN status, attending a particular school, geographical area).
Provide this information for each of the samples used to address research questions for the implementation study.
Describe the data sources you will use to answer the research questions (for example, youth surveys, fidelity protocols, attendance logs). Describe the timing and frequency of each data collection effort (for example, during all sessions, once a week, annually), and the party responsible for collecting the data. Use a table to link the information on data collection to the research questions. See Table 4 for an example of such a table (sample text included in italics). Finally, please provide a copy of your data collection instruments in an appendix to your analysis plan (at a minimum, provide the instruments that you used to collect the data you will use for your implementation study).
Table 4. Examples of data for addressing the research questions
Implementation element |
Research question |
Data source |
Timing/frequency of data collection |
Party responsible for data collection |
Fidelity |
Were all intended program activities offered and for the expected duration? |
Fidelity tracking log or protocol; session observations |
Every session for fidelity tracking; two times a year for session observations |
Program staff for fidelity tracking; study staff for session observations |
Fidelity |
What content did the youth receive? |
Fidelity tracking log or protocol; attendance logs; session observations |
Every session for fidelity tracking and attendance logs; two times a year for session observations |
Program staff for fidelity tracking and attendance logs; study staff for session observations |
Fidelity |
Who delivered services to youth? |
Staff applications; hiring records; training logs |
One time X months after start of implementation; annually |
Program staff |
Fidelity |
What were the unplanned adaptations to key program components? |
Adaptation request; work plan; 6-month progress report; annual progress report |
Annually; ad hoc |
Program staff; study staff |
Dosage |
How often did youth participate in the program on average? |
Workshop sessions and individual service contacts; attendance logs |
All sessions delivered |
Program staff |
Quality |
What was the quality of staff–participant interactions? |
Observations of interaction quality, using protocol developed by study staff |
X percent of sessions selected at random for observation |
Study staff |
Engagement |
How engaged were youth in the program? |
Observations of engagement, possibly using an engagement assessment tool; ratings from facilitator fidelity logs; engagement ratings from participant satisfaction surveys |
Y percent of sessions selected at random for observation |
Study staff |
Context |
What other pregnancy prevention programming was available to study participants? |
Interviews with staff from partnering agencies in the community; survey items on baseline and follow-up assessments; websites of other agencies in the community providing programming |
Once a year; ad hoc |
Study staff |
Context |
What external events affected implementation? |
Interviews with community/county representatives; list of site/school closures |
Once a year; ad hoc |
Study staff |
Describe the specific measures you will construct to address each research question. Use a table to link the description of the measures to the research questions. Table 5 presents an example of such table (sample text included in italics). Define each measure being examined in the implementation evaluation using data sources included in Table 4. For each measure, briefly explain how you will construct it (that is, survey item source or, if multiple item sources, how those items will be combined). In addition, describe any targets you pre-specified and used, if applicable, to assess how well the program was implemented relative to program or developer standards.
Table 5. Examples of measures for addressing the research questions
Implementation element |
Research question |
Measures |
Target |
Fidelity |
Were all intended program activities offered and for the expected duration? |
|
|
Fidelity |
What content did the youth receive? |
|
|
Fidelity |
Who delivered services to youth? |
|
|
Fidelity |
What were the unplanned adaptations to key program components? |
|
|
Dosage |
How often did youth participate in the program on average? |
|
|
Quality |
What was the quality of staff–participant interactions? |
|
|
Engagement |
How engaged were youth in the program? |
|
|
Context |
What other pregnancy prevention programming was available to study participants? |
|
|
Context |
What external events affected implementation? |
|
|
This section should lay out the approach you will use to answer the research questions. Describe your approach to cleaning and analyzing data to answer the research questions.
Describe how you will clean and prepare the data for analysis. For quantitative data, include information on how you will handle missing or inconsistent data. For qualitative data, include information on your approaches to examine and summarize qualitative data from interviews, focus groups, and observations. Provide the name of any software used to organize and code data. Provide details on the coding system you will used, who will code the data, and how you will ensure consistency across coders.
Describe how you will analyze measures to answer the research questions. For example, for quantitative data, describe whether you will calculate averages, percentages, and frequencies using the data collected and whether you will compare them to established targets for implementation. For qualitative data, this may include describing how you will identify common themes from interview data, or identify lessons learned from focus groups conducted with youth or staff.
The research questions state the main hypotheses for the outcomes evaluation. Use the research questions from your approved evaluation plan and make note of any revisions to those questions. You can present the research questions in this section as a list. If you have different research questions for different samples, or subgroups, make that clear in the presentation. For example, you might be interested in different outcomes for middle school youth in your study than for high school youth. Present the research questions in a table, organized by implementation elements. Table 6 provides an example.
Table 6. Examples of research questions for each outcome
Outcome |
Research question |
Refusal skills |
What percentage of youth reported an increase in refusal skills from pre-test to post-test? |
… |
|
Cultural connectedness |
What percentage of youth reported an increase in their reported connection to their tribal community from pre-test to post-test? |
Specify which program or programs described in Section A.2 are part of your outcomes evaluation, and for which samples.
The evaluation design section should explain the design overview, target population, enrollment and consent process, data collection, and outcomes for your evaluation. Aim to provide a detailed account of your methods, sufficiently transparent so that someone could replicate them.
As part of the overview for this section, briefly describe the research design used in the outcomes evaluation (for example, comparing people over time using a pre-post approach with linked data or comparing groups over time using a pre-post approach with unlinked data).
Describe the characteristics of the youth to be served by the project, including the study setting, context, and eligibility criteria. Include information about the type and number of sites that you aimed to enroll (for example, the number of schools, classrooms, or clinics, and the target location and grade) and the total number of youth intended to enroll in the evaluation. Broadly describe the intended characteristics of the target sample (for example, youth at risk for teen pregnancy or STIs due to high rates of sexual initiation). Describe any required characteristics for sample inclusion (for example, age, grade level, attending a particular school, geographical area, AI/AN status). You can expand on text from your approved evaluation plan and from the “sample and setting” section of the approved evaluation abstract.
Describe the process for enrolling the sample and obtaining consent. Include information on recruitment and enrollment procedures, such as the number and type of staff who were responsible for recruiting youth and details about each type of recruitment strategy used (for example, flyers home to parents, referrals from social service agencies, direct outreach, advertisements), how you formed partnerships with the sites, how the sites identified eligible participants, and how you enrolled the youth into the evaluation and program. Also, include the time frame for recruitment and enrollment in the evaluation. In addition, describe the timing, process, and materials used during the consent process (such as the consent forms, incentives, and so on). Provide the start date of sample enrollment and the estimated date (month and year) for completing sample enrollment.
This section should be a high-level overview of the data sources and data collection methods you will use for the research questions your study will address. Describe the timing of each data collection point (for example, at enrollment, at the first workshop X weeks after enrollment, at the final workshop Y weeks after the first workshop). Describe the modes and methods of collecting data at each data collection point (for example, in-person paper surveys, online surveys, phone interviews). Provide estimated dates (month and year) for starting and completing each planned data collection effort. Attach the survey instrument(s) as an appendix to the submitted plan.
Use text from your approved evaluation plan or evaluation abstract, and use a table to link data sources to research questions. See Table 7 for a sample table. You do not need to explain how you will construct specific measures in this section.
Table 7. Sources of data to address the research questions
Research question to be addressed |
Data source |
Timing of data collection |
Mode of data collection |
Start and end date of data collection |
Research question 1 |
Program participants |
At the first lesson |
In-person hard-copy survey |
September 2018 through March 2020 |
… |
|
|
|
|
Research question 4 |
Program participants |
Six months after the last lesson |
Telephone interview |
September 2019 through March 2021 |
This section will discuss the specific outcome measures you will use to answer the research questions. For each outcome being examined in the outcomes evaluation, briefly explain how you will construct each outcome measure (that is, survey item source or, if multiple item sources, how those items will be combined). If you will construct measures from multiple questions, describe the survey items you will use to create each construct and how you will code them to create the measure. Use the outcomes provided in the “outcomes study” section of the approved evaluation plan.
Use the table shell provided in Table 8 to describe all measures that will answer the research questions. Describe how you are constructing the outcome measure and the time periods when you will measure outcomes. In the “outcome name” column, include the name of the outcome that will be used. In the “source item” column, provide the data source for the measure. In the “constructed measure” column, describe the outcome and how it is constructed. If the outcome is from a published measure, provide the name of the measure. In the “timing of measure” column, indicate when the measure was measured, relative to the end of the program.
Table 8. Outcomes used to answer the research questions
Outcome name |
Source item |
Constructed measure |
Timing of measure |
Refusal skills |
Would you say that being in the program has made you more likely, about the same, or less likely to… resist or say no to peer pressure?” |
Dichotomous variable coded as 1 if answered “more likely” and zero otherwise. |
A post-test (immediately after program ends) |
This section should lay out the approach you will use to examine youth outcomes, including data preparation and statistical methods. Note that we provide a recommended approach to constructing your analytic sample. If you choose to adopt this approach, address the issues discussed below in your analysis plan. If you propose an alternative approach to constructing your sample, such as through dealing with youth nonresponse or attrition though multiple imputation or nonresponse weights, carefully articulate your proposed approach for each section that follows.
Recommended analytic approach
In order to look at changes over time in youth attitudes, intentions, or knowledge, we recommend limiting your analytic sample to youth who have responded to both the pre- and post-program survey. Using this approach, you can examine how a person’s outcomes have changed over time, controlling for factors that might be associated with those changes (for instance, age or sex).
You will also need to deal with item nonresponse, when youth have skipped questions that you plan to incorporate in your analyses. We recommend dropping those youth from your analyses as well (that is, a complete case analysis), unless they are a sizable percentage of the sample. If so, you could consider imputing data using a process such as multiple imputation. Feel free to discuss options with your RETA liaison.
In addition to conducting your outcome analysis, we recommend you conduct a nonresponse analysis in order to understand the representativeness of the analytic sample with your original enrolled population. You would compare the profile (for example, baseline demographics and risk characteristics) of those in your analysis sample to the full sample who responded to the baseline survey by assessing differences in baseline means using a t-test of mean differences.
Note: If you are unable to link pre- and post-program survey data for an individual, we recommend you aggregate the data to the smallest cluster possible (such as classroom-year) and look at differences within that cluster over time. When using cluster-level data, it will be important to give the reader a sense of the proportion of the youth in the program who are in your analytic sample (for example, 90 percent of youth in the served classrooms provided pre- and post-program data). Instead of a nonresponse analysis, describe how the sample analyzed might reflect the population that received the program. For example, report on the percentage of the enrolled sample who took the pre-test and the percentage who took the post-test. In addition, estimate the number of youth who completed the post-test who were not enrolled in the program at the time of the pre-test, and conversely, the number of youth who left the program after completing the pre-test but before completing the post-test.
Describe how you will clean the baseline and post-program data and prepare them for analysis. For example, if you find a person has reported sexual activity at baseline and reported never having had sex at follow-up, what rules will you use to reconcile these responses?
Indicate whether you will use a complete case analysis. If you will not (see “Recommended analytic approach” above), describe how you will construct your analytic sample.
If you will use a complete case analysis, please state that. If you will impute missing data, rather than dropping cases that are missing their entire survey or are missing selected measures you will analyze, describe how you will impute data and account for the imputation in your analysis.
Describe how you will analyze outcome measures to answer the research questions. For instance, explain whether you will describe summary statistics (for example, means or percentages) of outcome measures at the follow-up(s); compare means and other indicators at baseline and the follow-up(s) (that is, a pre-post analysis); conduct correlational analyses of the associations between implementation variables (such as dosage/participation) and outcome measures; or conduct growth-curve analyses. Provide the following information:
Model specification. Describe the type of model you will use to examine outcomes (for example, linear regression, logistic regression). Define the criteria you will use to assess the statistical significance of the findings (for example, “findings are considered statistically significant based on p < .05, two-tailed test”). Specify the statistical software package you will use.
Covariates. Using the table shell provided in Table 9, list all potential covariates (control variables in the regression) you plan to include in the models you will use to examine the outcomes of interest. If you have not determined the covariates yet, describe a plan for determining the covariates you will include in your analyses.
Table 9. Covariates included in outcomes analysis
Covariate |
Description of covariate |
Age |
Age (in years) as of the baseline data collection |
|
|
Nonresponse analysis. If following the recommended approach, list the measures you will examine to determine how similar the sample with post-program survey data is to the full sample with pre-program survey data. Discuss the method you will use to assess the magnitude or statistical significance of the differences. If you will not be using a complete case analysis and instead take a different approach to defining your analytic sample, describe how you will assess how similar the analytic sample is to the original sample.
Subgroup analysis. If you plan on conducting any subgroup analyses, list those subgroups and discuss any differences in your analytic approach, if applicable. See Table 10 for a sample table.
Table 10. Subgroup analyses in outcomes analysis
Subgroup |
Outcome |
Description of deviation from full-sample analytic methods |
Girls |
Refusal skills |
Same as full-sample analysis |
|
|
|
File Type | application/vnd.openxmlformats-officedocument.wordprocessingml.document |
File Title | Mathematica Standard Report Template |
Author | Cindy Castro |
File Modified | 0000-00-00 |
File Created | 2021-01-12 |