SPF-SIG Community Outcomes SS-A_3_3_16

SPF-SIG Community Outcomes SS-A_3_3_16.docx

Strategic Prevention Framework State Incentive Grant (SPF SIG) Program, Cohorts IV and V

OMB: 0930-0358

Document [docx]
Download: docx | pdf

Strategic Prevention Framework State Incentive Grant (SPF SIG) Program: Community Outcomes

Supporting Statement

Part A. Justification

A1. Circumstances Necessitating Data Collection

The Substance Abuse and Mental Health Services Administration’s (SAMHSA) Center for Substance Abuse Prevention (CSAP) requests OMB approval to collect a new community outcomes data for the cross-site evaluation of the Strategic Prevention Framework State Incentive Grant (SPF SIG) program, Cohorts IV and V. The community outcome data will inform the broader cross-site evaluation. The instruments seeking approval are the: 1) SPF SIG National Outcome Measures and 2) Community Outcomes Data Fields.

CSAP has funded two cross-site evaluations of the Strategic Prevention Framework State Incentive Grant (SPF SIG), one focused on Cohorts I and II and the other on Cohorts III, IV, and V. Collectively, these evaluations provide an important opportunity to inform the prevention field on current practices and their association with community- and state-level outcomes.

The primary cross-site evaluation objective is to determine the impact of SPF SIG on building prevention capacity and infrastructure, and preventing the onset and reducing the progression of substance abuse, as measured by the SAMHSA National Outcomes Measures (NOMs). Data are collected at the grantee, community, and participant levels. The collection of community

outcomes data is the focus of the current request.

Historical context

A1a. The SPF SIG Program

The SPF SIG is a major SAMHSA grant program that supports an array of activities to help grantees and subrecipient communities build a solid foundation for delivering and sustaining substance abuse prevention services that are effective in reducing the incidence and prevalence of substance use as well as improving risk and protective factors associated with substance use. CSAP provides funding to states/territories, jurisdictions, and tribal entities to implement the five steps of the strategic prevention framework (SPF), which are:

Step 1: Profile population needs, resources, and readiness to address the problems and gaps in service delivery;

Step 2: Mobilize and/or build capacity to address needs;

Step 3: Develop a comprehensive strategic plan;

Step 4: Implement evidence-based prevention programs, policies, practices and infrastructure development activities; and

Step 5: Monitor process, evaluate effectiveness, sustain effective programs/activities, and improve or replace those that fail.

For the purposes of this document, the word grantee will refer to all funded states, jurisdictions, and tribal entities. SPF SIG grantees select and fund their own community subrecipients who actually implement the prevention interventions. The selection of these subrecipients is not guided by CSAP.

A1b. The Cross-site Evaluations

Information on both cross-site evaluations is presented below as context for understanding SAMHSA’s request to collect community outcomes data from Cohorts IV and V.

A1b1. Cohorts I, II, and III Cross-site Evaluation

The National Institute on Drug Abuse (NIDA) provided support to CSAP to evaluate the impact of the SPF SIG project for Cohorts I and II. Specifically, data were collected from the 26 states and territories receiving grants initiated in 2004 (Cohort I sites) and 2005 (Cohort II sites) and up to 32 non-Cohort I and II grantee states and territories that served as a comparison group. Data from a Community-Level Instrument was used to determine the impact of the SPF SIG on all of the NOMs domains related to prevention (i.e., Abstinence, Education/Employment, Crime and Criminal Justice, Access/Capacity, Retention, Cost Efficiency and Use of Evidence-based Practices). The evaluation has also measured the effect of establishing and sustaining infrastructure at the state and community-levels to allow for data-based decision-making; the implementation of the SPF; and the environmental factors that affect substance abuse.

A1b2. Cohorts IV, and V Cross-site Evaluation

The contractor is implementing a multi-level evaluation design encompassing process and outcome data collection at the grantee, community, and participant levels. Data have been gathered from the 16 states, jurisdictions, and tribal entities receiving grants in FY 2006 (Cohort III), the 25 Cohort IV grantees funded in FY 2009 (one subsequently relinquished funding), and the 10 Cohort V grantees funded in FY 2010. Data collection for Cohort III has ended; therefore, this data collection request pertains only to Cohorts IV and V. Funding ended for most Cohort IV grantees in June 2015.

Collection of community data was a part of grantees’ strategic plans and most collected (or, for Cohort V, are continuing to collect) these data for their grantee-level evaluations. Grantees gathered existing archival or survey data (e.g., student risk behavior survey, required Federal reporting of arrests and motor vehicle crashes) for their data where available. Some tribal and jurisdiction grantees lacked such data and conducted primary data collection for their SPF SIG evaluations.

A2. Purpose and Use of Information

The purpose of the SPF SIG grant program is to improve substance abuse prevention systems and enhance the quality of prevention programs, primarily through the implementation of the SPF process. The goal of this initiative is to provide states, jurisdictions, tribal entities, and the communities within them with the tools necessary to develop an effective prevention system with attention to the processes, directions, goals, expectations, and accountabilities necessary for functionality. SAMHSA/CSAP needs to collect information over the course of the remaining grant period to monitor the progress of the SPF SIG initiative. CSAP will use the findings from the analysis of the community outcomes data in the cross-site evaluations to assess the impact of SPF activities on community-level outcomes. Without these data, the extended impact of the SPF process across various types of communities would remain unknown. Additionally, findings from this cross site evaluation will assist CSAP policymakers and program developers as they design and implement future initiatives.

Cohorts IV and V grantees will provide aggregate data for at least one substance use consumption, consequence, and/or mediating/intervening variables for each of their subrecipient communities. (In the case of twelve single-community grantees, data are reported for the entire jurisdiction or tribal entity.) Grantees that have data sources available annually will provide approximately six annual data points per measure: a baseline time point prior to subrecipient or intervention funding, and follow-up time points at least 11 months after the previous time point for the duration of funding. At a minimum, grantees with less available data (e.g., tribal grantees that conducted primary data collection for their evaluations) will provide a baseline time point and a follow-up time point close to the end of the grant period.

Reported measures must be a SPF SIG National Outcome Measure (see Attachment A) or a substitute measure. Community outcomes data are to be drawn from existing survey or administrative data within the state, tribe, or jurisdiction that grantees have used for their SPF SIG grantee-level evaluations. In addition to providing the measured value and variability estimates for each outcome measure, the contractor will request measure descriptions, including targeted SPF SIG priority, data source name and type, data collection time frame, item and response option wording, and population and sample information. See Attachment B for a full list of proposed data collection fields.

The contractor team will create an Excel data entry spreadsheet for each grantee. Given the non-confidential nature of aggregate community outcomes data, the data entry spreadsheet will be distributed to grantees and returned to the contractor via email. This spreadsheet will be pre-populated with any community outcomes data that the contractor has in our cross-site database. Each time point for each outcome for each subrecipient community will be entered in its own row in the spreadsheet.

A3. Use of Information Technology

Information technology has been used to minimize respondent burden for Cohorts IV and V in the SPF SIG Cross-site Evaluation. An Excel data entry spreadsheet will be created for each grantee. The Excel format will allow grantees to copy and paste data elements that are applicable across multiple time points and multiple subrecipients. Closed-ended fields will contain drop-down menus that allow for the quick selection from among response options. Given the non-confidential nature of aggregate community outcomes data, the data entry spreadsheet will be distributed to grantees and returned to the contractor via email, which will be both efficient and expedient.

Technology is also used to facilitate communication and provide updates to SPF SIG personnel. Through a SPF SIG listserv, grantee evaluators, project directors, coordinators, and other key staff will have the opportunity to exchange valuable advice; find guidance and resource materials; and receive announcements and clarifications from CSAP, other SPF SIG grantees, and the cross-site evaluation team. In addition to the listserv, grantees can access community outcomes guidance materials on the contractor’s Knowledge Base and the cross-site evaluation team will also send electronic copies of guidance materials via email to SPF SIG grantees upon request.

A4. Efforts to Identify Duplication

The information being collected for the cross-site evaluations for SPF SIG Cohorts IV and V is specific to the program and is not available elsewhere.

A5. Involvement of Small Entities

SPF SIG grantees are mainly State agencies, tribal organizations and other jurisdictions; they are not typically small entities. Some subrecipients may be small entities; however, the impact of reporting pre-existing community outcomes data should not be significant.

A6. Consequences If Information is Not Collected or is Collected Less Frequently

The requested Cohort IV and V community outcomes data collection will be a one-time data collection, occurring for approximately two months following the receipt of OMB approval.

If these data are not collected from the grantees, SAMHSA will have an incomplete dataset for the entire program and will lose critical information. This is particularly problematic for tribal grantees (who comprise a significant portion of the Cohort IV and V grantees) and jurisdictional grantees, neither of whom are represented in the national datasets (reported at the state level) that the cross-site evaluation is using to assess outcomes at the grantee level (e.g., National Survey on Drug Use and Health; Fatality Analysis Reporting System).

A7. Consistency with Guidelines in 5 CFR 1320.5(d) (2)

This information collection fully complies with 5 CFR 1320.5(d) (2).

A8. Consultation Outside the Agency

A8a. Federal Registry Announcement

The notice required in 5 CFR 1320.8(d) was published in the Federal Register on December 4, 2015 (80 FR 75869).



A8b. Consultations Outside the Agency

The community outcomes data fields were reviewed by SAMHSA staff and contractors. These experts provided feedback on the data collection instrument and the instrument was revised based on their feedback. The following individuals consulted with CSAP staff on the community outcomes data fields:


Nilufer Isvan, Ph.D.

DACCC Data Analysis Team Lead

Human Services Research Institute

Cambridge, MA


Virginia Mulkern, Ph.D.

DACCC Project Director

Human Services Research Institute

Cambridge, MA


Kelly Vander Ley, Ph.D.

Senior Research Associate

RMC

Portland, OR


Kelly Jarvis, Ph.D.

Senior Research Associate

RMC

Portland, OR


Gillian Leichtling

Senior Research Associate

RMC

Portland, OR


A9. Payment to Respondents

There is no payment to any respondent.

A10. Assurance of Confidentiality

No individual-level or personal data will be collected; grantee staff will provide aggregate community data. Although not collecting individual-level data, evaluation staff are trained on the importance of privacy and in handling sensitive data. SAMHSA and its contractors will not receive identifiable client records. A PIA and SORN application have been submitted for this data collection for review at the U.S. Department of Health and Human Services

A11. Questions of a Sensitive Nature

Grantees and subrecipients will potentially report data derived from sensitive questions, such as those about substance use behavior. However, these data are generally collected for a purpose other than SPF SIG (e.g., student risk behavior survey, required Federal reporting of arrests and motor vehicle crashes) and will only be reported to the contractor in the aggregate.

A12. Estimates of Average Annualized Hour Burden

All 34 SPF SIG Cohort IV (n=24) and Cohort V (n=10) grantees are expected to provide community data. Grantees will provide one response that includes data for all available time points. The community outcome instrument (completed in the form of a spreadsheet) is estimated to take approximately 4 hours to complete; this includes time to look up and compile information (3 hours) and time to complete the spreadsheet (1 hour). The estimated burden time is based on test instruments completed by evaluation staff members that have experience working with SPF SIG grantees (see Section B.4 for more detail). There are no direct costs to respondents other than their time to complete the instrument. Table 1 provides the details of the total burden for the instrument (submitted only once): burden hours estimated at 136, and respondent cost of $5,559.68 (total burden hours × the average hourly wage for State government managers, as reported in the 2014 Occupational Employment Statistics [OES] by the Bureau of Labor Statistics [BLS]; see http://www.bls.gov/oes/current/naics4_999200.htm#11-0000).

Table 1. Estimates of Annualized Hour and Cost Burden to Respondents for Cohorts IV and V

Respon-dent

No. of Respon-dents

No. of Responses per Respondent

Total No. of Responses

Burden per Response (Hrs.)

Total Burden (Hrs.)

Hourly Wage Cost

Total Cost

Grantee

34

1

34

4

136

$40.88

$5,559.68



A13. Estimates of Annualized Cost Burden to Respondents

There are no capital/startup costs or operational/maintenance of services costs associated with this project.

A14. Estimates of Annualized Cost to the Government

The total estimated cost to the government for this one time data collection is $256,472.68. This includes approximately $89,711 for developing the data collection spreadsheets; collecting data from grantees; processing, cleaning, and housing data; and analyzing and reporting data. Approximately $55,602 per year represents SAMHSA costs to manage/administer the data collection and analysis for 25% each of two employees (GS-14-10, $111,203 annual salary). Approximately $105,600 per year represents SAMHSA costs to monitor and approve grantee reporting (10% time of 10 Project Officers at $105,600 annual salary) and approximately $5,559.68 for grantee data collection.

A15. Changes in Burden

This is a new collection of information.

A16. Time Schedule, Analysis and Publication Plans

Time Schedule

Table 2 outlines the key time points for the SPF SIG community outcomes data collection.

Table 2. Time Schedule for SPF SIG Community Outcomes Data Collection

Activity

Time Schedule

Prepare for data collection, including preparation of data collection spreadsheets

April 2016

Obtain OMB approval for data collection

April 2016

Collect data

May 2016-June 2016

Process data

July 2016-September 2016

Analyze data and disseminate findings

October 2016–September 2018

Note. OMB, Office of Management and Budget; SPF SIG, Strategic Prevention Framework State Incentive Grant

Analysis

Community outcomes data will be used to address two evaluation research questions in the SPF SIG cross-site evaluation:

  1. Did the implementation of the SPF SIG lead to community-level improvement on NOMs or other outcomes?

  2. What factors accounted for variation in performance on NOMs and other outcomes across funded communities?1

Evaluation Research Question 1 will be examined using meta-analysis. Meta-analytic techniques will be used to calculate effect sizes that will provide descriptive information regarding the magnitude of change in outcomes. We will calculate effect sizes for community outcomes as follows: For each of the outcome measures at each time point, we will calculate odds to reflect the occurrence of a “successful event.” For instance, if the baseline percentage of arrests due to alcohol/drugs is 25%, then the percentage of arrests that are not due to alcohol/drugs is 75%. Thus, the odds of a successful event (i.e., an arrest not related to alcohol or drugs) are calculated as 3.0 (3 successes to every 1 failure). From these odds, an odds ratio will be computed to reflect the amount of change in outcomes over time, specifically by dividing the odds of a successful event at follow-up by the odds of a successful event at baseline. Odds ratios are centered at 1.0; odds ratios greater than 1.0 indicate greater odds of success at follow-up compared with baseline (i.e., improvement), whereas odds ratios less than 1.0 indicate greater odds of success at baseline (i.e., worsening). All subsequent analyses will be conducted on the natural log of the odds ratio,2 and variance estimates will be calculated.

Evaluation Research Question 2 will be examined via hierarchical linear modeling (HLM) or meta-regression. SPF SIG data represent data in which observations are nested—for instance, communities nested within grantees. To address the nested nature of these data, the contractor plans to use HLM, a multilevel approach that can accommodate the nested design as well as the addition of covariates. The HLM approach develops regression equations at each level of nesting to account for variation at different levels of the model. Each subrecipient will have its own estimated trajectory for the outcome variable, which will vary across subrecipients and grantees. This approach has a number of advantages over traditional general linear models (GLMs), including the ability to include individual- and cluster-level covariates while adjusting for random effects associated with each cluster and greater flexibility regarding covariance structures.3

When HLM is not possible because of small sample sizes or failure to meet statistical assumptions, meta-regression will be the technique of choice. Meta-regression is a technique in which regression analyses are conducted on effect sizes (e.g., odds ratios) rather than on raw data. Random-effects meta-regression models allow for heterogeneity in the “true effect” and thus allow for the likelihood that differences in grantee characteristics may result in differing effect sizes from one grantee to another.4 This method treats each effect size as the result of a different “study” conducted on a grantee or on a community, where each study (i.e., grantee mean or community mean) is weighted by the inverse of its variance. Modified regression analyses are then conducted on these weighted means, in a manner conceptually similar to that of a traditional regression. One limitation to the meta-regression approach (similar to traditional regression) is that it is not recommended when the number of studies (i.e., grantees or subrecipient communities) is small. Just as with traditional regression analyses, a typical rule of thumb is at least 10 studies per covariate, although more are better. As a result, meta-regression models will have to remain relatively simple for cohort-level analyses, particularly at the grantee level, and may be more complex when the sample size is larger, such as for community-level or cross-cohort analyses. If inferential statistics are underpowered, we will either cautiously conduct inferential statistics and focus interpretation on the effect size estimates and confidence intervals rather than p-values or restrict our analyses to descriptive statistics (in the event of very small sample sizes).

Publications

In combination with other cross-site evaluation data sources, the contractor will use SPF SIG community outcomes data to help SAMHSA reach its diverse stakeholders through targeted products and innovative dissemination venues. The objective for all reports and dissemination products is to provide user-friendly documents and presentations that help SAMHSA successfully disseminate and explain the findings. The dissemination plan includes products in a variety of formats for a variety of target audiences. Audiences for these reports will include Congress, SAMHSA Centers, the evaluation’s SAMHSA Contracting Officer’s Representatives (CORs), SPF SIG grantees, and the broader substance abuse prevention field (e.g., academia, researchers, policy-makers, providers). The contractor and SAMHSA recognize that different audiences are best reached by different types of report formats. For example, reports to Congress will require materials that are concise but offer policy-relevant recommendations. Reports created for SAMHSA Centers and the CORs will require more in-depth information, such as substantive background and discussion sections, to supplement the analytic approach. Reports created for SPF SIG grantees will be concise handouts with helpful and easy-to-read graphics on performance data rather than lengthy text. The assortment of disseminations products developed using the PFS MRT data will include short and long analytic reports, congressional briefings, annual evaluation reports, research and policy briefs, ad hoc analytic reports, journal articles, best practice summaries, and conference or other presentations. Certain dissemination products may be made available on SAMHSA’s website.

A17. Display of Expiration Date

The expiration date for OMB approval will be displayed.

A18. Exceptions to Certification Statement

This collection of information involves no exceptions to the Certification for Paperwork Reduction Act Submissions.



1 The “factors” in this question will be assessed with the Community-Level Instrument, previously approved under OMB No. 0930-0279, expiration 02/28/2017.

2 Lipsey, M. W., & Wilson, D. B. (2001). Practical meta-analysis. Thousand Oaks, CA: Sage.

3 West, B. T., Welch, K. B., & Galecki, A. T. (2007). Linear mixed models: A practical guide using statistical software. Boca Raton, FL: Chapman & Hall/CRC.

4 Borenstein, M., Hedges, L. V., Higgins, J. P. T., & Rothstein, H. R. (2009). Introduction to meta-analysis. Chichester, England: John Wiley & Sons.

6


File Typeapplication/vnd.openxmlformats-officedocument.wordprocessingml.document
File TitleRMC Basic Report & Proposal Template
AuthorRMC Research Corporation
File Modified0000-00-00
File Created2021-01-24

© 2024 OMB.report | Privacy Policy