SupSta B (0493)

SupSta B (0493).docx

2013 - 2015 Youth Risk Behavior Surveys

OMB: 0920-0493

Document [docx]
Download: docx | pdf


SUPPORTING STATEMENT FOR THE

2013 and 2015 NATIONAL YOUTH RISK BEHAVIOR SURVEY



PART B

















Submitted by:

Danice K. Eaton, MPH, PhD, Project Officer
Division of Adolescent and School Health
National Center for Chronic Disease Prevention and Health Promotion

4770 Buford Hwy, NE, MS K-33
Atlanta, GA 30341
770-488-6143 (voice); 770-488-6156 (fax)

[email protected]

Centers for Disease Control and Prevention
Department of Health and Human Services

May 2012

TABLE OF CONTENTS



B. COLLECTIONS OF INFORMATION EMPLOYING STATISTICAL METHODS


1. Respondent Universe and Sampling Methods

2. Procedures for the Collection of Information


a. Statistical Methodology for Stratification and Sample Selection

b. Estimation and Justification of Sample Size

c. Estimation and Statistical Testing Procedures

d. Use of Less Frequent than Annual Data Collection

e. Survey Instrument

f. Data Collection Procedures

g. Obtaining Access to and Support from Schools

h. Informed Consent

i. Quality Control


3. Methods to Maximize Response Rates and Deal with Nonresponse

a. Expected Response Rates

b. Methods for Maximizing Response and Handling Non-Response


4. Tests of Procedures or Methods to be Undertaken


5. Individuals Consulted on Statistical Aspects and Individuals Collecting and/or

Analyzing Data


a. Statistical Review

b. Agency Responsibility

c. Responsibility for Data Collection


REFERENCES

LIST OF APPENDICES

  1. Authorizing Legislation


  1. 60-Day Federal Register Notice

  2. 60-Day Federal Register Notice Comment

  3. Rationale for Survey Questions


E. Youth Risk Behavior Survey Questionnaire


F. Youth Risk Behavior Survey Questionnaire Supplemental Documents

F1. Parental Permission Form Distribution Script

F2. Parental Permission Form and Fact Sheet (English Version)

F3. Parental Permission Form and Fact Sheet (Spanish Version)

F4. Parental Permission Form Reminder Notice (English Version)

F5. Parental Permission Form Reminder Notice (Spanish Version)

F6. Questionnaire Administration Guide

F7. Data Collector Confidentiality Agreement

G. Recruitment Scripts for the Youth Risk Behavior Survey

G1. State-level Recruitment Script for the Youth Risk Behavior Survey

G2. District-level Recruitment Script for the Youth Risk Behavior Survey

G3. School-level Recruitment Script for the Youth Risk Behavior Survey


H. Recruitment Scripts for the Youth Risk Behavior Survey Supplemental Documents

H1. School-level Recruitment Script for the Youth Risk Behavior Survey Supplemental Documents - State Letter of Invitation

H2. School-level Recruitment Script for the Youth Risk Behavior Survey Supplemental Documents - District Letter of Invitation

H3. School-level Recruitment Script for the Youth Risk Behavior Survey Supplemental Documents

H3a. School Letter of Invitation and YRBS Fact Sheet for Schools

H3b. Letter to Agreeing Schools


I. Data Collection Checklist for the Youth Risk Behavior Survey


J. Data Collection Checklist for the Youth Risk Behavior Survey Supplemental Documents

J1. Letter to Teachers in Participating Schools

J2. Make-up List and Instructions


K. Expert Reviewers for 1989 Consultations

L. IRB Approval Letter


M. Sample Table Shells


N. Detailed Sampling and Weighting Plan

B. COLLECTIONS OF INFORMATION EMPLOYING STATISTICAL METHODS


B.1 RESPONDENT UNIVERSE AND SAMPLING METHODS


The universe for the study will consist of public and private school students in grades 9, 10, 11, and 12 in the 50 states and the District of Columbia.


The sampling frame for schools has been obtained from MDR, Inc. The MDR data encompass both private and public schools and include the latest data from the Common Core of Data from the National Center for Education Statistics. MDR school-level files also include data on enrollments by grade and minority enrollments. For more than 40 years, MDR has provided information on K-12, higher education, library, early childhood, and related education organizations. The commercial sampling frame provided by MDR is nationally recognized as the most complete, current, and accurate education databases available in the industry.


Table B-1 displays the current distribution of schools nationally by urban status and type of school.


Table B-1

Distribution of Schools Nationally by Urban Status and School Type


Metro Status

File Type

Total

Frequency

Percent

Public

Private

Catholic


Large Central City


3,170

9.96%

1,039

3.26%

291

0.91%

4,500

14.14%


Mid-size Central City


3,082

9.68%

1,083

3.40%

266

0.84%

4,431

13.92%

Urban Fringe of Large City

4,892

15.37%

1,549

4.87%

303

0.95%

6,744

21.19%

Urban Fringe of Mid-sized City

2658

8.35%

760

2.39%

80

0.25%

3498

10.99%


Large Town


269

0.85%

78

0.25%

26

0.08%

373

1.17%


Small Town


2,74

8.53%

617

1.94%

74

0.23%

3,405

10.70%

Rural, Outside MSA

4,962

15.59%

546

1.72%

28

0.09%

5,536

17.39%


Rural, Inside MSA


2,791

8.77%

524

1.65%

25

0.08%

3,340

10.49%


Total

24,538

77.10%

6,196

19.47%

1,093

3.43%

31,827

100.00%



Sampling or other respondent selection method used: Students will be selected using the procedures described in detail below. To briefly summarize, for each YRBS cycle, a nationally representative sample of students will be selected using a three-stage stratified cluster sample. Primary Sampling Units (counties) and Secondary Sampling Units (schools) within selected counties will be selected with probability proportional to size selection methods. Schools will be sampled from a frame that includes all US public, private and Catholic schools. In each selected school, one class will be selected in each grade to participate, except in high minority schools where two classes per grade will be selected. All students in selected classes are eligible to participate.


Expected response rates for the data collection: The average participation rates over the 12 prior cycles of YRBS are 78% for schools and 86% for students. We assume these historical average participation rates in preparing the sample design for the 2013 and 2015 YRBS.


Actual response rates achieved during the last collection period: During the most recent cycle of the YRBS, conducted in 2011, the participation rates were 81% for schools and 87% for students, which resulted in an overall response rate of 70%.


Statistical justification for all sample sizes: The expected student sample size of approximately 22,673students before nonresponse is necessary to meet study precision requirements. The sample size is calculated by inflating the sample size that would be required under the assumptions of simple random sampling by historical design effects (to account for the complex sampling design) and participation rates to account for nonresponse at both the student and school levels.


B.2 PROCEDURES FOR THE COLLECTION OF INFORMATION


B.2.a Statistical Methodology for Stratification and Sample Selection


For each YRBS cycle, a probability sample will be selected that will support national estimates among students in grades 9-12 overall, and by age or grade, sex, and race/ethnicity (white, black, Hispanic). The design also will support sex-specific estimates by grade and race/ethnicity and racial/ethnic-specific estimates by grade. A detailed description of the sampling design may be found in Appendix N.


Sampling Frame. The sampling frame will stratify the 50 states and the District of Columbia by region, urbanicity, and minority composition. The sampling frame is structured into geographically defined units, called primary sampling units (PSUs) defined as a county or groups of contiguous counties (except when they are unaffiliated cities). The stratification by minority composition will divide the PSUs into eight groups on the percentages of blacks and Hispanics in the PSU. This is accomplished in 2 steps. First, each PSU is stratified into either the Hispanic strata or black strata based on whether there is a higher percentage of Hispanic or black enrolled students in the PSU. Each stratum will then be subdivided into four strata depending on the percentage of black or Hispanic enrolled students, as appropriate, in the PSU. The eight racial/ethnic-oriented strata will each be further divided by urban status defined as being in one of the 54 largest Metropolitan Statistical Areas (MSA) versus not. In addition, the first-stage PSU sample will be implicitly stratified by geography using 5-digit zip code areas.

Selection of PSUs. Fifty-four PSUs will be selected with probability proportional to the student enrollment in the PSU within strata. The allocation of PSUs to the first-stage strata will be approximately in proportion to the total enrollment in the PSU. A proportional allocation tends to maximize the precision of overall survey estimates.


Selection of Secondary Sampling Unites (SSUs). SSUs are comprised of either a single school (if the school includes each of grades 9-12) or multiple schools “linked” together. An SSU is comprised of multiple “linked” schools when the physical schools do not include all of grades 9-12. This is done to form school-based SSUs that provide coverage for all four grades in each unit. SSUs will be grouped by size as either large or small, depending upon whether or not they have 28 students or more per grade. In each selected PSU, at least three large SSUs will be selected with probability proportional to an aggregate enrollment measure, resulting in 162 selected SSUs (54 PSUs * 3 SSUs). In addition, from a sub-sample of 15 PSUs, one small SSU will be randomly selected to represent what is about 5.7% of the students nationwide (those attending small schools). A total of 177 SSUs will be selected (162 large and 15 small).These 177 SSUs will include approximately 200 physical schools.

Selection of Classes. Classes in each school are randomly selected based on two very specific scientific parameters to ensure a nationally representative sample.  First, classes have to be selected in such a way that all students in the school have a chance to participate.  Second, all classes must be mutually exclusive so that no student is selected more than once.   In each school, once we have determined the type of class or time period from which classes will be selected, we randomly select the appropriate number of classes within each grade.  To maintain acceptable school participation rates, it is essential that each school have input in the decision of which classes will be sampled in their school.  Examples of class sampling frames that have been used in past surveys include all 2nd period classes or a required physical education class.  As long as the scientific sampling parameters are met, we work with each school to identify a classroom sampling frame that will work best for each school. One class will be selected in each eligible grade for all schools except for those schools with the highest percentages of black and Hispanic students. In those schools, two classes per grade will be selected.


Selection of Students. All students in a selected classroom are eligible for the study. Based on historical averages, in large SSUs each selected class will include 28 students on average and each small SSU will supply a total of 63 students. Of the 162 large schools, approximately 20% (n=32) will be “high minority” and 2 classes will be selected from each grade. In the remaining 130 selected large schools, 1 class per grade will be selected. Therefore, we expect to select approximately 21,728 students from large SSUs [( 32 SSUs * 8 classes * 28 students = 7,168 students) + (130 + 4 classes *28 students = 14,560 students) = 21,728 students] and approximately 945 students (15 SSUs * 63 students) from small SSUs.


Refusals. School districts, schools, and students who refuse to participate in the study, and students whose parents refuse to give permission, will not be replaced in the sample. We will record the characteristics of schools that refuse for analysis of potential study biases. Accounting for school and student nonresponse, we expect approximately 15,194 participating students.




B.2.b Estimation and Justification of Sample Size


Overview. The YRBS is designed to produce estimates with error margins of ±5 percent:


  • 95 percent confidence for domains defined by grade, sex, or race/ethnicity;

  • 95 percent confidence for domains defined by crossing grade by sex, and race/ethnicity by sex; and

  • 90 percent confidence for domains formed by crossing grade with race/ethnicity.

The 2013 and 2015 YRBS sample size and design will be consistent with the sample size and design used in the 2009 and 2011 YRBS cycles because these design parameters met the levels of precision required for CDC's purposes. Minor design refinements are made to account for the changing demographics of the in-school population of students. Specifically, the design is adjusted to account for the increasing percentage of Hispanic students.

For the 2013 and 2015 YRBS cycles, we performed an extensive simulation study that assessed whether it would be possible to achieve target sample sizes for minority groups without the oversampling of black and Hispanic students that has been done in previous YRBS cycles. In previous cycles, oversampling was accomplished in three ways:

  1. Disproportional allocation of PSUs to strata based on the density of minority enrollment

  2. Weighting coefficients in the measure of size used in sampling PSU’s with probabilities proportional to size (PPS) so that high minority PSUs and schools had an increased probability for selection

  3. In schools with a high percentage of black or Hispanic students, selecting two classes in each grade rather than one.

Specifically, the simulation study investigated whether required sample sizes could be achieved with a proportional allocation to strata and an unweighted measure of size (student enrollment in the eligible grades). Both of these simplifications would lead to a design that is more efficient statistically, and would lead to more precise survey estimates overall. The simulation results indicated that the required sample sizes will be achieved with the modified, more efficient sampling design.


The proposed samples for the 2013 and 2015 YRBS will each consist of 54 PSUs. At each grade level, at least three different schools will contribute classes of approximately 28 students each and at least three schools will be selected within each PSU. The actual number of physical schools will be more than 3 times 54, however, for two reasons. Schools that only span part of the grades of interest are combined during sampling to form sampling units, such that 1 sampled school could be composed of 2 physical schools. One small school will be selected in 15 subsample PSUs separately from large schools. As a result, approximately 200 schools will be selected into the sample. We will select one class per school per eligible grade, except for schools with the largest concentrations of minority students where we will select two classes per grade. We expect that the final sample will include approximately 22,673 selected students which will yield approximately 15,194 participating students.


School and Student Non-response. The average participation rates over the 12 prior cycles of YRBS are 78% for schools and 86% for students. We are assuming these historical average participation rates in preparing the sample design for the 2013 and 2015 YRBS.


B.2.c Estimation and Statistical Testing Procedures


Sample data will be weighted by the reciprocal of the probability of case selection and adjusted for non-response. The resulting weights will be trimmed to reduce mean-squared error. Next, the strata weights will be adjusted to reflect true relative enrollments. Finally, the data will be post-stratified to match national distributions of high school students by race/ethnicity and grade.

Variances will be computed using linearization methods. YRBS data are also used for trend analyses where data for successive cycles are compared with statistical testing techniques. Statistical testing methods are used also to compare subgroup prevalence rates (e.g., male versus female students) for each cross-sectional survey.


Confidence intervals vary from estimate to estimate depending upon whether the estimate is for the full population or for a subset such as a particular grade or sex. Within a grouping, confidence intervals also vary depending on the level of the estimate and the design effect associated with the measure. Based on prior YRBS’s which had similar designs and sample sizes, we can expect the following:

  • Estimates among students overall or by grade or age, sex, and race/ethnicity (white, black, Hispanic) will be accurate at ±5 percent at 95 percent confidence.

  • For racial/ethnic estimates by grade (e.g., 11th grade Hispanics), about 70% will be accurate to within ±5 percent at 90 percent confidence.

The experience in using these data is that the levels of sampling errors involved are appropriate given the uses of the data for descriptive reporting and trend analysis.


B.2.d Use of Less Frequent Than Annual Data Collection


As stated in A.6, the YRBS originally was planned and twice approved by OMB as an annual survey. Based on experience, it was determined in 1992 that it would be sufficient to address the programmatic needs of CDC and other Federal agencies to conduct the YRBS biennially. By shifting from an annual to biennial survey starting with the 1993 YRBS, burden has been reduced by half.


It is important that data be collected biennially to detect any changes in health risk behaviors by high school students that need to be addressed in school health programs, public education campaigns, demonstrations, and professional education/training, especially those sponsored by CDC. Due to the speed with which many of these problems, including the AIDS epidemic and tobacco use, will take an increasing toll in human suffering and financial burden, which will be heavily borne by the Federal government, it is imperative to conduct the survey biennially. School systems have the capacity to change their school health programs to help prevent health risk behaviors, such as tobacco use, that contribute to the leading causes of mortality and morbidity readily on an annual basis, if circumstances require.


B.2.e Survey Instrument


The YRBS questionnaire (Appendix E), contains 92 items which can be roughly divided into seven categories. The first category includes four demographic questions. The remaining questions address health risk behaviors in six categories: unintentional injuries and violence; tobacco use; alcohol and other drug use; sexual behaviors that contribute to HIV infection, other sexually transmitted diseases and unintended pregnancies; unhealthy dietary behaviors; and physical inactivity. Obesity (assessed by self-reported height and weight) and asthma also are assessed. The questions are all in a multiple-choice format and will be administered as a 12-page optically scannable questionnaire booklet.

B.2.f Data Collection Procedures


Data will be collected by a small staff of professional data collectors, specially trained to conduct the YRBS. The time during the school day in which the survey is administered varies by school.  This decision is made in coordination with each school to ensure that the type of class or period of the day selected for sampling 1) meets the scientific sampling parameters to ensure a nationally representative sample and 2) results in the least burden/highest possible acceptability for the school. The data collector will have direct responsibility for administering the survey to students. Data collectors will follow a questionnaire administration guide (Appendix F6).


Teachers will be asked to remain at the front or back of the classroom and not to walk around the room monitoring the aisles during survey administration because doing so could affect honest responses and compromise anonymity. Teachers also will be asked to identify students allowed to participate in the survey and to make sure non-participating students have appropriate alternative activities. The rationale for this is to increase the candor and comfort level of students. The only direct responsibility of teachers in data collection is to distribute and follow up on parental permission forms sent out prior to the scheduled date for data collection in the school. Teachers are provided with a parental permission form distribution script (Appendix F1) to follow when distributing permission forms to students.


The Data Collection Checklist (Appendix I) is completed by teachers to track which students have received parental permission to participate in the data collection. The teachers receive instructions on completing the Data Collection Checklist in the “Letter to Teachers in Participating Schools” (Appendix J1). The data collector will utilize the information on the Data Collection Checklist to identify students eligible for a make-up survey administration; this information will be recorded by the data collector on the “Make-up List and Instructions” document (Appendix J2).


In general, our data collection procedures have been designed to ensure that:


  • Protocol is followed in obtaining access to schools.

  • Everyday school activity schedules are disrupted minimally.

  • Administrative burden placed on teachers is minimal.

  • Parents give informed permission to participate in the survey.

  • Anonymity of student participation is maintained, with no punitive actions against nonparticipants.

  • Alternative activities are provided for nonparticipants.

  • Control over the quality of data is maintained.


B.2.g Obtaining Access to and Support From Schools


All initial letters of invitation will be on CDC letterhead from the Department of Health and Human Services and signed by Howell Wechsler, Ed.D, M.P.H., Director, Division of Adolescent and School Health, Centers for Disease Control and Prevention.  The procedures for gaining access to schools will have three major steps:


  • Notify state education agencies (SEAs) in states with sampled schools and invite states to participate. Obtain written approval for participation at the SEA level. Verify existence and grade range of selected schools. Obtain names of school districts in which schools are located, school district addresses, names of district superintendents, names of supportive school district contacts, and general guidance on working with the selected school districts and schools in the state. Request that the state notify school districts that they may anticipate being contacted about the survey.

  • Once cleared at the state level, invite school districts in which selected schools are located to participate in the study. For Catholic schools and other private schools, invite the office comparable to the school district office (e.g., diocesan office of education). Obtain written approval for participation at the district level. Verify existence of school, grade range, and other information provided by the state. Request that the school district notify schools that they may anticipate being contacted about the survey. Request general guidance on working with the selected schools.

  • Once cleared at the school district level, invite selected schools to participate. Verify information previously obtained about the school. Present the burden and benefits of participation in the survey. After a school agrees to participate, develop a tailor-made plan for collection of data in the school (e.g., select classes; determine whether survey will be administered to selected class sections simultaneously or in serial). Obtain written approval for participation at the school level. Ensure that all materials reach the school well in advance of when they are needed. Maintain contact with schools until all data collection activities have been completed.

Prior experience suggests the process of working with each state education agency, school district, and school will have unique features. Discussions with each education agency will recognize the organizational constraints and prevailing practices of the agency. Scripts for use in guiding these discussions may be found in Appendices G1 (state-level), G2 (district-level), and G3 (school-level). Appendix H contains copies of letters of invitation to states (Appendix H1), school districts (Appendix H2), and school administrators (Appendix H3a). Appendix H also contains the YRBS Fact Sheet for Schools (H3a). A copy of the letter to be sent to schools once they have agreed to participate is found in Appendix H3b.

B.2.h Informed Consent


The permission form (Appendices F2 & F3) informs both the student and the parent about an important activity in which the student has the opportunity to participate. By providing adequate information about the activity, it ensures that permission will be informed. In accord with the No Child Left Behind Act, the permission form indicates that a copy of the questionnaire will be available for review by parents at their child’s school. The parental permission forms will be made available in both English and Spanish.


A waiver of written student assent was obtained for the participation of children because this research presents no more than minimal risk to subjects, parental permission is required for participation, the waiver will not adversely affect the rights and welfare of the students because they are free to decline to take part, and it is thought that some students may perceive their responses are not anonymous if they are required to provide stated assent and sign a consent/assent document. Students are told “Participating in this survey is voluntary and your grade in this class will not be affected, whether or not you answer the questions.” Completion of the survey implies student assent.


B.2.i Quality Control


Table B-2 lists the major means of quality control. As shown, the task of collecting quality data begins with a clear and explicit study protocol and ends with procedures for the coding, entry, and verification of collected data. In between these activities, and subsequent to data collector training, measures must be taken to reinforce training, to assist field staff who run into trouble, and to check on data collection techniques. Because the ultimate aim is production of a high quality database and reports, various quality assurance activities will be applied during the data collection phase.

Table B-2

Major Means of Quality Control


Survey Step


Quality Control Procedures


Mail Out


Check inner vs. outer label for correspondence (5% sample).

Verify that any errors in packaging were not systematic (100%).


Previsit Logistics

Verification


Review data collection procedures with school personnel in each school to ensure that all preparatory activities were performed properly (100%).


Receipt Control


Verify that a sample of forms received the prior day were logged in and are stored in the proper location (5%).

Require entry of staff ID in receipt control and all other transactions 100%).


Telephone Contacts


Monitor early sample of scheduling and follow-up telephone calls to ensure

that the caller follows procedures, elicits proper information, and has proper

demeanor (10%).


Manual Editing


Verify initial editing by all editors until standards are achieved (100%).

Spot check editing by editor (5%).


Computer Scanning


Key enter questionnaires that are not scannable (100%).

Remove any scannable forms that reflect intentional misuse by respondent (100%).


B.3 METHODS TO MAXIMIZE RESPONSE RATES AND DEAL WITH NONRESPONSE


B.3.a Expected Response Rates

While we aim for an 80% school participation rate, the historical average for this study over the last 12 cycles of YRBS has ranged between 70% and 81%, with an average of 78%. We have conservatively assumed the average school participation rate for the purposes of sample design. . The addition of a $500 incentive for each school (as suggested by OMB in 1999) has helped maintain and perhaps slightly increase school participation rates. Even before the addition of the school incentive, the YRBS set the standard for response rates among federally-funded national, school-based, health related surveys of high school students. For example, the widely cited Monitoring the Future survey (formerly known as the High School Senior Survey), achieves substantially lower participation rates than the YRBS, even though Monitoring the Future contains less sensitive questions. The participation rates established by the YRBS are the product of the application of proven and tested procedures for maximizing school and student participation.

As indicated in A.16.c, it is highly desirable to complete data collection before the final 2 months of school. Schools are very busy then with testing and attendance can be very unstable, especially among twelfth grade students.


B.3.b Methods for Maximizing Responses and Handling Non-Response


We distinguish among six potential types of nonresponse problems: refusal to participate by a selected school district, school, teacher, parent, or student; and collection of incomplete information from a student.


To minimize refusals at all levels--from school district to student--we will use a variety of techniques, emphasizing the importance of the survey. Given the high visibility and subject matter of the survey, we expect that some school districts or schools will need to place the issue of survey participation before the school board. To increase the likelihood of an affirmative decision, we will: (1) work through the SEA to communicate its support of the survey to school districts and schools; (2) indicate that the survey is being sponsored by CDC and has the support of Federal and state agencies; (3) convey to school districts and schools that the survey has the endorsement of many key national educational and health associations, such as the American Academy of Pediatrics, American Association of School Administrators, Association of State and Territorial Health Officials, National Association of Secondary School Principals, National Catholic Educational Association, National PTA, National Association of State Boards of Education, Council of Chief State School Officers, the National Education Association, and the National School Boards Association; (4) maintain a toll-free hotline to answer questions from school district and school officials, teachers, parents, and students throughout the process of recruiting schools and obtaining parental permission for the student’s participation; (5) comply with all requirements from school districts in preparing written proposals for survey clearance; (6) convey a willingness to appear in person, if needed, to present the survey before a school board, research committee, or other local entity tasked with reviewing the survey; (7) offer a package of educational products to each participating school, as recommended and approved by OMB in approving the 1998 YRBS in alternative schools (OMB No. 0920-0416, expiration 12/98), and continued ever since; and (8) offer schools a monetary incentive of $500, consistent with recommendations OMB previously made, and implemented in the national YRBS since 2001.


The sampling plan does not allow for the replacement of schools that refuse to participate due to concern that replacing schools would introduce bias. All participating SEAs, school districts, and schools also will be promised and sent a copy of the published survey results.


Maximizing responses and dealing with refusals from parents, teachers, and students require different strategies. Parental permission form reminders (Appendices F4 & F5) will be sent to parents who have not returned parental permission forms within an agreed upon time period (e.g., 3 days); those who do not respond to the reminder will be sent a second and final reminder. The permission form will provide a telephone number at CDC that parents may call to have questions answered before agreeing to give permission for their child's participation. Permission forms will be available in English, Spanish, and other languages as required by dominant languages spoken by parents in selected schools. Field staff will be available on location to answer questions from parents who remain uncertain of permission. Bilingual field staff will be used in locations with high Hispanic concentrations (e.g., California, Florida, New York City, and Texas).


Teacher refusals to cooperate with the study are not expected to be a problem because schools already will have agreed to participate and burden to teachers is minimal. Refusals by students whose parents have consented also are expected to be minimal. No punitive action will be taken against a nonconsenting student. Nonconsenting students will not be replaced. Data will be analyzed to determine if student nonresponse introduces any biases.


To minimize the likelihood of missing values on the questionnaire, students will be reminded in writing in the questionnaire booklet and verbally by the survey administrator to review the optically scannable questionnaire before turning it in to verify that: (1) each question has been answered, (2) only one oval is filled in for each question with the exception of the question on race/ethnicity, and (3) each response has been entered with a No. 2 pencil, fills the oval, and is dark. A No. 2 pencil will be provided to each survey participant to reduce the likelihood that responses will not scan properly, which would produce missing values. In addition, when completed questionnaires are visually scanned later at project headquarters, any oval that is lightly filled in will be darkened (unless they appear to be erasures) and stray marks will be erased before the forms are scanned. Missing values for an individual student on the survey will not be imputed.


B.4 TESTS OF PROCEDURES OR METHODS TO BE UNDERTAKEN

YRBS questionnaire items were originally tested by the NCHS laboratories. The 1993 special issue of Public Health Reports on the development of the Youth Risk Behavior Surveillance System describes the development and testing process. A limited pretest of the questionnaire on nine respondents was conducted in November 1989 by the contractor in the Prince George's County, Maryland school system in accord with OMB guidelines. The pretest was conducted to:

  • Quantify respondent burden.

  • Test survey administrator instructions and procedures.

  • Verify the overall feasibility of the survey approach.

  • Identify needed changes in the instruments or instructions to control/reduce burden.

The pilot test sharpened the articulation of certain survey questions and produced an empirical estimate of the survey burden.

The YRBS questionnaire has been used extensively in ten prior national school-based surveys approved by OMB, as well as at the state and local levels. Further pilot testing in accord with OMB guidelines has been performed on new questions.


B.5 INDIVIDUALS CONSULTED ON STATISTICAL ASPECTS AND INDIVIDUALS COLLECTING AND/OR ANALYZING DATA

B.5.a Statistical Review

Statistical aspects of the study have been reviewed by the individuals listed below.

  • Michael T. Errecart, Ph.D. (deceased)

Macro International Inc.


  • Ronaldo Iachan, Ph.D.

ICF International Inc. (formerly Macro International Inc.)

11785 Beltsville Drive, Suite 300

Beltsville, MD 20705

Phone: (301) 572-0538

Fax: (301) 572-0986

E-mail: [email protected]


  • William Robb, MS, MBA

ICF International Inc. (formerly Macro International Inc.)

126 College Street

Burlington, VT 05401

Phone: (802) 264-3713

E-mail: [email protected]


B.5.b Agency Responsibility

Within the agency, the following individual will be responsible for receiving and approving contract deliverables and will have primary responsibility for data analysis:


  • Danice K. Eaton, MPH, Ph.D.

Commander
United States Public Health Service
SBSB/DASH/NCHHSTP/CDC
4770 Buford Highway NE, MS K-33
Atlanta, GA  30341
Voice: 770-488-6143
Fax: 770-488-6156
E-mail: [email protected]


B.5.c Responsibility for Data Collection


The representative of the contractor responsible for conducting the planned data collection is:


  • Katherine H. Flint, M.A.

Principal

ICF International Inc. (formerly Macro International Inc.)

11785 Beltsville Drive, Suite 300

Beltsville, Maryland 20705

Phone: (301) 572-0333

Fax: (301) 572-0986

E-mail: [email protected]

REFERENCES




Bouchery EE, Harwood HJ, Sacks J, Simon C, Brewer R (2011). Economic Costs of Excessive Alcohol Consumption in the U.S., 2006. American Journal of Preventive Medicine; 41(5):516-524.


CDC (1994). Guidelines for school health programs to prevent tobacco use and addiction. MMWR; 43(No. RR‑2).


CDC (1996). Guidelines for school health programs to promote lifelong healthy eating. MMWR; 45(No. RR‑9).


CDC (1997). Guidelines for school and community programs to promote lifelong physical activity among young people. MMWR; 46(No. RR‑6).


CDC (2001). School health guidelines to prevent unintentional injuries and violence. MMWR; 50(No. RR-22).


CDC (2004). Medical expenditures attributable to injuries - United States, 2000. MMWR; 53(No.1):1‑4.


CDC (2005). School Health Index: A Self-Assessment and Planning Guide. Atlanta, Georgia.


CDC (2006). Physical Education Curriculum Analysis Tool. Atlanta, Georgia.


CDC (2007a). Best Practices for Comprehensive Tobacco Control Programs—2007. Atlanta: U.S. Department of Health and Human Services, Centers for Disease Control and Prevention, National Center for Chronic Disease Prevention and Health Promotion, Office on Smoking and Health.


CDC (2007b). Health Education Curriculum Analysis Tool. Atlanta: CDC.


CDC (2008). Smoking-Attributable Mortality, Years of Potential Life Lost, and Productivity Losses --- United States, 2000--2004. MMWR; 57(45):1226-1228.


CDC (2011a) Vital signs: teen pregnancy — United States, 1991 – 2009. MMWR 60(No 13):414-420.


CDC (2011b) Sexually Transmitted Disease Morbidity for selected STDs by age, race/ethnicity and gender 1996-2009, CDC WONDER On-line Database. Accessed at http://wonder.cdc.gov/std-std-race-age.html on Jan 5, 2012 10:20:50 AM.


CDC (2011b) HIV surveillance report, 2009; vol. 21. Available at http://www.cdc.gov/hiv/topics/surveillance/resources/reports/. Published February 2011. Accessed November 28, 2011.


CDC (2011d) FY 2012 Online Performance Appendix. Atlanta, GA: US Department of Health and Human Services, CDC. Available at http://www.cdc.gov/fmo/topic/Performance/performance_docs/FY2012_CDC_Online_Performance_Appendix.pdf.


CDC (2011e). School Health Guidelines to Promote Healthy Eating and Physical Activity. MMWR; 60(RR#5):1-76.


CDC (2012). Indicators for chronic disease surveillance. Available at http://apps.nccd.cdc.gov/cdi/Default.aspx. Accessed on February 4, 2012.


CDC, NCHS (2009). Public use data file and documentation: multiple cause of death for ICD-10 2006 data [CD-ROM].


Chesson HW, Blandford JM, Gift TL, Tao G, and Irwin KL (2004). The estimated direct medical cost of sexually transmitted diseases among American youth, 2000. Perspectives on Sexual and Reproductive Health; 36(1):11-19.


Colditz G. Economic costs of obesity and inactivity. (1999). Medicine and Science in Sports and Exercise; 31(11):S663-S667.


Denniston M, Brener ND, Kann L, Eaton DK, McManus T, Kyle TM, Roberts AM, Flint KH, Gross JG (2010). Comparison of paper-and-pencil versus web administration of the Youth Risk Behavior Survey (YRBS): participation, data quality, and perceived privacy and anonymity. Computers in Human Behavior; 26:1054-1060.


Eaton DK, Brener ND, Kann L, Denniston M, McManus T, Kyle TM, Roberts AM, Flint KH, Gross JG. (2010a) Comparison of paper-and-pencil versus web administration of the Youth Risk Behavior Survey (YRBS): risk behavior prevalence estimates. Evaluation Review; 34(2): 137-153.


Eaton DK, Kann L, Kinchen S, Shanklin S, Ross J, Hawkins J, et al (2010b). Youth Risk Behavior Surveillance—United States, 2009. MMWR; 59(SS-5):1-142.


Eaton DK, ND Brener, L Kann, AM Roberts, TM Kyle, KH Flint, ALR Ross (2011). Computer availability and principals’ perceptions of online surveys. Journal of School Health; 81(7):365-373.


Finkelstein EA, Fiebelkorn IC, Wang G (2003). National medical spending attributable to overweight and obesity: How much, and who’s paying? Health Affairs; W3:219-226.


Foti K, Balaji A, Shanklin S (2011). Uses of Youth Risk Behavior Survey and School Health Profiles Data: Applications for Improving Adolescent and School Health. Journal of School Health; 81(6):345–354


Frazão E (Ed) (1999). America's Eating Habits: Changes and Consequences. Food and Rural Economics Division, Economic Research Service, U.S. Department of Agriculture. Agriculture Information Bulletin No. 750 (AIB‑750), p. 5-32.


Hoffman SD (2006). By the Numbers: The Public Costs of Teen Childbearing. Washington, DC: National Campaign to Prevent Teen Pregnancy.


National Center for Health Statistics (2011). Health, United States, 2010: With Special Feature on Death and Dying. Hyattsville, Maryland.


National Coalition for Food-Safe Schools (2004). Food-Safe Schools Action Guide. Available at: http://www.foodsafeschools.org/.


National Drug Intelligence Center (2011). The Economic Impact of Illicit Drug Use on American Society. Washington D.C.: United States Department of Justice.


Office of the Assistant Secretary for Planning and Evaluation, U.S. Department of Health and Human Services (2004). Trends in the Well-being of America’s Children and Youth, 2003. Available on-line: http://aspe.hhs.gov/hsp/03trends/index.htm.


U.S. Department of Agriculture, U.S. Department of Health and Human Services, U.S. Department of Health and Human Services; and U.S. Department of Education (2005). Making It Happen! School Nutrition Success Stories. Alexandria, VA.


U.S. Department of Health and Human Services (1994). Preventing tobacco use among young people: a report of the Surgeon General. Atlanta GA.


U.S. Department of Health and Human Services (1996). Physical activity and health: a report of the Surgeon General. Atlanta GA.


U.S. Department of Health and Human Services (1998). Tobacco use among U.S. racial/ethnic minority groups: A report of the Surgeon General. Atlanta GA.


U.S. Department of Health and Human Services (1999). Mental health: a report of the Surgeon General. Atlanta GA.


U.S. Department of Health and Human Services (1999). The Surgeon General’s call to action to prevent suicide. Atlanta GA.


U.S. Department of Health and Human Services. Office of Disease Prevention and Health Promotion. (2011) Healthy People 2020. Washington, DC. Available at http://www.healthypeople.gov. Accessed February 6, 2012.


U.S. Department of Health and Human Services (2001). Women and smoking: a report of the Surgeon General. Atlanta GA.


U.S. Department of Health and Human Services (2001). Youth violence: a report of the Surgeon General. Atlanta GA.


U.S. Department of Health and Human Services (2004). The health consequences of smoking: a report of the Surgeon General. Atlanta GA.




File Typeapplication/vnd.openxmlformats-officedocument.wordprocessingml.document
AuthorWindows User
File Modified0000-00-00
File Created2021-01-30

© 2024 OMB.report | Privacy Policy