PREP PAS+Baseline - Attachment B - Analysis Plan for PREP Impact Study - 6-7-12

PREP Eval - P+B - Attachment B - Analysis Plan PREP Impact - Clean - 11-12.docx

Personal Responsibility Education Program (PREP) Multi-Component Evaluation

PREP PAS+Baseline - Attachment B - Analysis Plan for PREP Impact Study - 6-7-12

OMB: 0970-0398

Document [docx]
Download: docx | pdf







attachment b
analysis plan for prep Impact study



Analysis Plan for PREP Impact Study

The purpose of the PREP impact study is to rigorously assess the impacts of funded programs in four or five selected PREP sites. In each site, ACF expects to recruit and enroll a sample of 1,200 to 1,500 youth (for a total of 6,000 youth across four or five sites). Youth will be randomly assigned to a treatment group that receives the program being tested or to a control group that does not. Program impacts will be analyzed with survey data collected at baseline and at 6 and 18 months after program completion. Impacts will be analyzed separately for each site.

Our analysis plan for the impact study has four main components: (1) a random assignment evaluation of each site using either a individual random assignment or cluster-based random assignment approach, (2) an early analysis of baseline data, (3) a primary impact analysis of key behavioral outcome measures, and (4) exploratory analyses of secondary research questions. These are described below.

Random assignment. The impact study is expected to include four to five different sites. We expect that each site will focus on a different program model, population, and setting, reflecting the diversity of programs funded under PREP. Each site will be analyzed separately, so we are not constrained to using the exact same research design in each site. Rather, to achieve the most rigorous evaluation of each site, we will customize our approach to random assignment to the unique circumstances of each site. In some sites, we anticipate that random assignment of individuals will offer the best approach. In others, random assignment of schools or other types of clusters may be the best approach. Regardless of the specific approach, we will conduct random assignment only within each of the four or five selected sites. Because each site will be analyzed separately, we do not plan to conduct random assignment or make comparisons across sites.

Baseline analysis. As soon as baseline data collection has been completed in each site, we will begin preliminary analyses of the baseline data. We will use these analyses to describe the study sample in each site and compare it with the target population. We will also assess whether random assignment successfully generated treatment and control groups balanced on important baseline characteristics. To support this analysis, our baseline survey will collect key measures of demographics (such as age, gender, race, and ethnicity) and other personal characteristics (such as prior sexual experience) needed to describe the study sample and examine the equivalence of the treatment and control groups.

Primary impact analysis. Impact analysis will begin after the completion of follow-up data collection in each site. With a random assignment design, unbiased impact estimates can be obtained from the difference in unadjusted mean outcomes at follow up between the treatment and control groups. However, we can improve the precision of the estimates by using regression models to control for covariates, especially baseline measures of outcomes. Regression adjustment can also account for any strata or blocking variables used in conducting random assignment, or for any differences between the treatment and control groups in baseline characteristics that arise by chance or from survey nonresponse.

The empirical specification for the model will depend on the unit of random assignment. With random assignment of youth, our model can be expressed as

(1) yi =β′xi+λTii

where yi is the outcome of interest for youth i; xi is a vector of baseline characteristics; Ti is an indicator equal to one for youth in the treatment group and zero for youth in the control group; and εi is a random error term. The vector of baseline characteristics xi will include demographic characteristics such as age, gender, race/ethnicity, and baseline measures of the outcomes. These baseline characteristics will be gathered on baseline surveys. The parameter estimate for λ is the estimated impact of the program.

If clusters, rather than individual youth, are the unit of assignment, the estimation must account for the correlation of outcomes among youth in the same cluster, as they will all be randomly assigned as a single unit, and each sample member cannot be considered statistically independent. To account for this dependence, we can modify the previous regression model as

(2) yis =β′xis+λTissis .

The general structure of the model is the same, but now yis is the outcome measure for individual i in cluster s (and similarly for the treatment status indicator, Tis, vector of baseline characteristics, xis and the error term εis). Most important, the error term in Equation (2) accounts for the clustering of youth within clusters because of the inclusion of the cluster-level error term ηs—a cluster “random effect.” If this error term is excluded, the precision of the impact estimates could be seriously overstated. As in Equation (1), the estimated impact of the program is λ.

Equation (1) or (2) will be estimated separately for each primary outcome in each site. The specific method for estimating the parameters of the models will depend on the form of the dependent variable. Ordinary least squares will be used for continuous variables (such as number of partners), whereas logistic regression procedures will be specified for binary outcomes (such as ever had sexual intercourse). Weights will be created for each site to account for any differences in random assignment, sampling, consent, or nonresponse probabilities among study participants. For any sites that pool data from multiple sub-awardees, we will weight each sub-awardee proportionate to size, so that each study participant is weighted equally. However, we will also run sensitivity analyses with each sub-awardee given equal weight.

To control for multiple hypothesis testing (the increased chance of falsely identifying an impact as statistically significant when examining effects on many outcomes), we will limit the primary analyses for each site to a small set of key outcomes. In selecting these outcomes, we will rely on the program logic model and data needs table developed for each site. We anticipate that most of these outcomes will be measures of sexual risk behavior and its health consequences (pregnancy, STIs, or birth), though the exact outcomes selected will vary by site. Within this small set of key outcomes, we will also consider applying a formal statistical correction for multiple hypothesis testing.

To support these analyses, the follow-up surveys will include measures of all key outcomes—primarily pregnancy, STIs, and associated sexual risk behaviors. We will also include these measures on the baseline survey, so that we can include them as covariates in the regression models used to estimate program impacts.

Analysis of secondary research questions. In addition to our primary impact analysis, we will also define and answer additional secondary research questions for each site:

  • Subgroup analyses. To examine whether the programs were more effective for some youth than for others, we will estimate impacts for subgroups of youth by adding a term to Equations (1) and (2) that interacts the treatment indicator by a binary indicator of a particular subgroup. The regression coefficient on this term provides an estimate of the difference in the program effect across the subgroups. Subgroups of particular interest include gender and baseline sexual experience. To support these analyses, we will include these subgroup variables on the baseline survey.

  • Impacts on mediating variables. In addition to primary analysis of program impacts on outcomes of most central importance, as part of secondary analysis we will also examine program impacts on key mediating variables specified in the program logic model for each site (for example, refusal skills, attitudes, or engagement in after-school or community activities). We will estimate impacts on these outcomes following the same approach described in Equations (1) and (2). These mediating variables will be drawn primarily from the short-term follow-up survey, which will be conducted six months after program completion. We will also include selected mediating variables on the baseline survey, to include as covariates in the regression models.

  • Variation in impacts by participation levels. Our primary impact analysis will include the full study sample, yielding intent-to-treat (ITT) estimates that do not account for varying participation rates among youth assigned to the treatment group. As exploratory analyses, we will consider adjusting for participation levels in two ways. First, to account for youth who do not attend any program sessions or activities, we can make the standard Bloom adjustment to calculate estimates of the treatment on the treated (TOT). Second, to explore the association between program dosage—the degree of program participation—and impacts, we can conduct propensity score analyses, whereby youth with the highest program attendance are matched to a subset of control group youth with similar demographic and baseline characteristics. To support these analyses, our baseline survey will include a broad range of demographic and other personal characteristics to consider as potential matching variables.


File Typeapplication/vnd.openxmlformats-officedocument.wordprocessingml.document
Authorbgoesling
File Modified0000-00-00
File Created2021-01-30

© 2024 OMB.report | Privacy Policy