SUPPORTING STATEMENT
Part B
National Survey of Teachers of Algebra
for the
NATIONAL MATHEMATICS ADVISORY PANEL
National Mathematics Advisory Panel
U.S. Department of Education
Section B: Description of Statistical Methodology
B.1. Universe and Sampling Procedures 4
B.2. Statistical Methodology 6
B.3. Methods for Maximizing Response Rate 10
B.4. Tests of Procedures and Methods 11
B.5. Reviewing Statisticians……………………………………...........................................11
The contractor has designed a fully representative random sample of public school algebra teachers. The contractor has acquired a full universe of public schools (from the the USED Common Core of Data). In order to represent a cross-section of public school algebra teachers across the country, the survey sampling should stratify all of the schools nationwide based on type of locale, percentage of students who receive free or reduced-price lunch, and minority or foreign language density, and graded configuration of the school.
Schools sampled for the study and which agree to participate will be asked to provide comprehensive lists of the algebra I teachers, and all listed teachers will be included in the sample.
Each stratum of schools will operate as its own random sample and will be fielded independently. Respondents will be selected via a random selection process where each stratum is divided into representative mini-samples or replicates. Standard best practices of replicate control measures will be employed to ensure highest possible response rates.
Clarification of Statistical Methods (Per OMB’s Questions of March 27, 2007)
Question: The NSAT sample frame should include schools with missing data on any of the stratification variables. This could be done by defining a new stratum consisting of schools missing FRPL eligibility data, or redefining the current definitions of the strata to incorporate the missing schools (e.g., 40% or greater FRPL or missing FRPL). Because mailout preparations using the original sample are sufficiently far along at this point, NORC opted to select a supplemental sample of schools from the set of schools missing data on the stratification variables.
Because
the USED NCES Common Core of Data were
missing data across over 1200 stratification variables, we deleted
1,231 schools from the original frame due to missing strata
information. However, to represent these schools, we selected a
supplemental sample of 10 schools from the 1231 schools—the
supplemental frame. The new sample was selected such that a school of
size
would
have the same selection probability if it were included in the
original frame. Under the previously implemented sample design, a
school of size
has a selection probability of
,
where
is
the total number of teachers in the original frame. Suppose that the
supplemental frame contains a total number of
teachers.
Then, we should select
teachers
from the supplemental frame using the same PPS sampling method. With
this sample size, a school of size
in the supplemental frame would have selection probability
,
which is the same probability if the school is included in the
original sample. Thus, this procedure ensures that a school with a
fixed size would have the same selection probability regardless of
which frame it is in.
The total estimated number of algebra 1 teachers over these 1,231 schools is 2169.99. Recall that the total estimated number of algebra 1 teachers over the 35,122 schools in the original frame is 65,865.88. This means that we should select (300*2169.99/65865.88)=9.88 schools from the supplemental frame. We selected 10 using the same PPS method used in the original draw. The frame is sorted by the other three stratification variables before selection.
The following tables show the distribution of the supplemental frame and the sample. Just over a third of the schools are in NY, but 7 of the 10 selected are in NY, reflecting their larger size as measured by the estimated number of algebra 1 teachers. Also, none of the 10 is in a rural area.
Supplemental Frame
Cumulative Cumulative
LSTATE04 Frequency Percent Frequency Percent
ƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒ
AK 78 6.34 78 6.34
AL 2 0.16 80 6.50
AZ 162 13.16 242 19.66
FL 2 0.16 244 19.82
GA 1 0.08 245 19.90
ID 12 0.97 257 20.88
IL 91 7.39 348 28.27
IN 16 1.30 364 29.57
KY 28 2.27 392 31.84
ME 16 1.30 408 33.14
MI 153 12.43 561 45.57
MN 1 0.08 562 45.65
MS 1 0.08 563 45.74
MT 43 3.49 606 49.23
NC 75 6.09 681 55.32
NE 3 0.24 684 55.56
NH 2 0.16 686 55.73
NJ 43 3.49 729 59.22
NY 441 35.82 1170 95.04
OH 3 0.24 1173 95.29
OK 4 0.32 1177 95.61
OR 1 0.08 1178 95.69
PA 1 0.08 1179 95.78
SD 1 0.08 1180 95.86
TX 3 0.24 1183 96.10
VA 32 2.60 1215 98.70
WA 1 0.08 1216 98.78
WI 14 1.14 1230 99.92
WY 1 0.08 1231 100.00
School is 40% or higher minority enrollment
Cumulative Cumulative
DMINOR Frequency Percent Frequency Percent
ƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒ
0 620 50.37 620 50.37
1 611 49.63 1231 100.00
School is high school/ middle school/ combined
Cumulative Cumulative
GRDRANGE Frequency Percent Frequency Percent
ƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒ
1 386 31.36 386 31.36
2 475 38.59 861 69.94
3 370 30.06 1231 100.00
School is urban/ suburban/ rural
Cumulative Cumulative
URBANICITY Frequency Percent Frequency Percent
ƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒ
1 644 52.32 644 52.32
2 268 21.77 912 74.09
3 319 25.91 1231 100.00
Supplemental Sample
Cumulative Cumulative
LSTATE04 Frequency Percent Frequency Percent
ƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒ
IL 1 10.00 1 10.00
MI 1 10.00 2 20.00
NY 7 70.00 9 90.00
VA 1 10.00 10 100.00
School is 40% or higher minority enrollment
Cumulative Cumulative
DMINOR Frequency Percent Frequency Percent
ƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒ
0 4 40.00 4 40.00
1 6 60.00 10 100.00
School is high school/ middle school/ combined
Cumulative Cumulative
GRDRANGE Frequency Percent Frequency Percent
ƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒ
1 7 70.00 7 70.00
2 2 20.00 9 90.00
3 1 10.00 10 100.00
School is urban/ suburban/ rural
Cumulative Cumulative
URBANICITY Frequency Percent Frequency Percent
ƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒƒ
1 8 80.00 8 80.00
2 2 20.00 10 100.00
The narrative follows up on the questions and requests from the March 27, 2007 phone conference between OMB, USED, and NORC on the National Survey of Algebra Teachers (NSAT). Brian requested three additions to the protocol; our responses to each are included below.
Question: Please provide minimum levels of precision for direct estimates from the survey taking into account the design effect. For example, for percentage estimates from single items an estimate of 50% would have a 95% confidence interval of plus or minus 3 percentage points. For a composite scale (drawn from in previous research), the expected mean based on prior research will be XX with a 95% confidence interval of x to X.
We have estimated a design effect of 2.5 for the overall sample. The design effect is substantial for some groups, but it’s about 2 or lower for most subgroups defined by the stratification variables. With minimum weight trimming, the overall design effect drops to 2, and the subgroup design effects are reduced to between 1.4 and 2. Note that these estimates are based on unequal weighting alone, i.e., they do not reflect potential clustering effect. The clustering effect should be small, however, given that we do not plan to sample more than 5 teachers from any school and the average cluster size is likely to be around 3. Depending on the magnitude of the intracluster correlation, clustering may increase the design effect by up to 10%, although it’s hard to estimate its impact with confidence. For the following precision and power calculations, we have assumed a design effect of 2.0 at the subgroup level.
To estimate the minimum levels of precision using a design effect of 2.0, the following tables show the 95% confidence intervals for (a) a composite scale with mean of 3.0 and standard deviation of 1.5, and (b) proportion estimates of 0.25, 0.50, and 0.75 estimated with effective sample sizes ranging from 50 to 400.
Estimates of 95% Confidence Intervals for a Composite Scale
Actual sample size |
Effective sample size (assuming design effect=2.0) |
Standard Error of the Mean |
95% CI for Estimated Composite Scale with mean=3.0 and SD=1.5 |
|
100 |
50 |
0.21 |
2.58 |
3.42 |
200 |
100 |
0.15 |
2.70 |
3.30 |
300 |
150 |
0.12 |
2.76 |
3.24 |
400 |
200 |
0.11 |
2.79 |
3.21 |
500 |
250 |
0.10 |
2.81 |
3.19 |
600 |
300 |
0.09 |
2.83 |
3.17 |
700 |
350 |
0.08 |
2.84 |
3.16 |
800 |
400 |
0.08 |
2.85 |
3.15 |
Estimates of 95% Confidence Intervals for a Proportion
Actual sample size |
Effective sample size (assuming design effect=2.0) |
95% CI for Estimated Proportion |
|||||
0.25 |
0.50 |
0.75 |
|||||
100 |
50 |
0.13 |
0.37 |
0.36 |
0.64 |
0.63 |
0.87 |
200 |
100 |
0.17 |
0.33 |
0.40 |
0.60 |
0.67 |
0.83 |
300 |
150 |
0.18 |
0.32 |
0.42 |
0.58 |
0.68 |
0.82 |
400 |
200 |
0.19 |
0.31 |
0.43 |
0.57 |
0.69 |
0.81 |
500 |
250 |
0.20 |
0.30 |
0.44 |
0.56 |
0.70 |
0.80 |
600 |
300 |
0.20 |
0.30 |
0.44 |
0.56 |
0.70 |
0.80 |
700 |
350 |
0.20 |
0.30 |
0.45 |
0.55 |
0.70 |
0.80 |
800 |
400 |
0.21 |
0.29 |
0.45 |
0.55 |
0.71 |
0.79 |
Question: Please provide additional power analyses. It would be helpful to see a table with a few different sample sizes for the subgroups (50, 100, 150, 200, 300) and the effect sizes that you will be able to detect in the comparisons with an alpha of .05 and power of 80% (taking into account the design effect). However, instead of the table, you could just report the minimum sample size that you will require for subgroup analyses and the associated effect size if you will set a hard rule on this. So, in this case, if you will not be conducting any subgroup comparisons with less than 200 teachers per cell, then you don’t need to do anything further since that is the power analysis you have already done [but please check the power analysis to ensure it was done correctly].
The following table was developed by Michael Yang for general guidance, and is applicable to the NSAT with the caveat that the NSAT sample size will be 800 or lower. The sample sizes here should be interpreted as “effective sample sizes”, that is, as the actual sample size divided by the design effect (provisionally assumed to be 2.0 for the NSAT).
Sample Size for Detecting a Pre-specified Difference Between Two Proportions (=.05, 1-=.80)
|
Effect Size |
|
|
Sample Size |
Effect Size |
|
|
Sample Size |
Effect Size |
|
|
Sample Size |
Two- Tailed Test |
.05 |
.95 |
.90 |
430 |
.10 |
.95 |
.85 |
140 |
.15
|
.95 |
.80 |
80 |
.90 |
.85 |
680 |
.90 |
.80 |
200 |
.90 |
.75 |
100 |
||||
.85 |
.80 |
900 |
.85 |
.75 |
250 |
.85 |
.70 |
120 |
||||
.80 |
.75 |
1,080 |
.80 |
.70 |
290 |
.80 |
.65 |
140 |
||||
.75 |
.70 |
1,240 |
.75 |
.65 |
330 |
.75 |
.60 |
150 |
||||
.70 |
.65 |
1,360 |
.70 |
.60 |
360 |
.70 |
.55 |
165 |
||||
.65 |
.60 |
1,460 |
.65 |
.55 |
380 |
.65 |
.50 |
170 |
||||
.60 |
.55 |
1,520 |
.60 |
.50 |
390 |
.60 |
.45 |
175 |
||||
.55 |
.50 |
1,550 |
.55 |
.45 |
390 |
.55 |
.40 |
175 |
||||
.50 |
.45 |
1,550 |
.50 |
.40 |
390 |
.50 |
.35 |
170 |
||||
.45 |
.40 |
1,520 |
.45 |
.35 |
380 |
.45 |
.30 |
165 |
||||
.40 |
.35 |
1,460 |
.40 |
.30 |
360 |
.40 |
.25 |
150 |
||||
.35 |
.30 |
1,360 |
.35 |
.25 |
330 |
.35 |
.20 |
140 |
||||
.30 |
.25 |
1,240 |
.30 |
.20 |
290 |
.30 |
.15 |
120 |
||||
.25 |
.20 |
1,080 |
.25 |
.15 |
250 |
.25 |
.10 |
100 |
||||
.20 |
.15 |
900 |
.20 |
.10 |
200 |
.20 |
.05 |
80 |
||||
.15 |
.10 |
680 |
.15 |
.05 |
140 |
|
|
|
||||
.10 |
.05 |
430 |
|
|
|
|
|
|
||||
One-Tailed Test |
.05 |
.95 |
.90 |
340 |
.10 |
.95 |
.85 |
110 |
.15
|
.95 |
.80 |
60 |
.90 |
.85 |
540 |
.90 |
.80 |
160 |
.90 |
.75 |
80 |
||||
.85 |
.80 |
710 |
.85 |
.75 |
195 |
.85 |
.70 |
95 |
||||
.80 |
.75 |
850 |
.80 |
.70 |
230 |
.80 |
.65 |
110 |
||||
.75 |
.70 |
980 |
.75 |
.65 |
260 |
.75 |
.60 |
120 |
||||
.70 |
.65 |
1,070 |
.70 |
.60 |
280 |
.70 |
.55 |
130 |
||||
.65 |
.60 |
1,170 |
.65 |
.55 |
295 |
.65 |
.50 |
135 |
||||
.60 |
.55 |
1,200 |
.60 |
.50 |
305 |
.60 |
.45 |
140 |
||||
.55 |
.50 |
1,220 |
.55 |
.45 |
310 |
.55 |
.40 |
140 |
||||
.50 |
.45 |
1,220 |
.50 |
.40 |
305 |
.50 |
.35 |
135 |
||||
.45 |
.40 |
1,200 |
.45 |
.35 |
295 |
.45 |
.30 |
130 |
||||
.40 |
.35 |
1,170 |
.40 |
.30 |
280 |
.40 |
.25 |
120 |
||||
.35 |
.30 |
1,070 |
.35 |
.25 |
260 |
.35 |
.20 |
110 |
||||
.30 |
.25 |
980 |
.30 |
.20 |
230 |
.30 |
.15 |
95 |
||||
.25 |
.20 |
850 |
.25 |
.15 |
195 |
.25 |
.10 |
80 |
||||
.20 |
.15 |
710 |
.20 |
.10 |
160 |
.20 |
.05 |
60 |
||||
.15 |
.10 |
540 |
.15 |
.05 |
110 |
|
|
|
||||
.10 |
.05 |
340 |
|
|
|
|
|
|
As the table shows, an effective minimum sample size of about 80, which is equivalent to an actual sample size of 160 under our design, will be needed to detect differences in proportions of .15 at the high and low extremes of proportion estimates (i.e,, .95-.80 and .20-.05) under a two-tailed test with alpha=.05 and power=.80. In the middle range of proportion estimates (e.g., .60-.45) the effective minimum sample size needed increases to 140, implying actual sample sizes of about 280 cases. The actual subgroup sample sizes range from 50 to roughly 500 in our sample. So, we are likely to be able to detect differences of 15% in most cases. For estimates that are below 40% or above 60%, we are likely to be able to detect differences of 10% among the largest subgroups. While we have not extended this exercise to composite scales, it appears that the NSAT will generally be suited to detect moderate to large effect sizes.
The
tabulations will all be constructed using weighted counts of cases.
All
sampled teachers will have a base weight that reflects the selection
probability under the sample design. The base weight will be adjusted
for non-response through weighting class adjustments where the
sampling strata or collapsed strata serve as the weighting classes.
The final analysis weight for a teacher
who
completes the survey can be expressed as
where
is
the base weight,
is
a 0-1 response indicator, and
denote
the weighting class.
Two types of response rates are germane to the NSTA: the rate at which sampled schools agree to participate by providing the names of their algebra 1 teachers (“school response rate”), and the rate at which the teachers sampled within the participating schools (“the teacher response rate”).
Maximizing the school response rate. In the experience of NORC (the data collection contractor for the NSTA), district superintendents should always be notified of studies involving their schools in advance of contacting the school principals. Furthermore, larger districts will often require survey projects to submit a formal application prior to any contact with schools within those districts. Using a template developed by NORC in other school studies, NORC will prepare a “district research clearance package” to supply to any district which requires the study information in writing. This package will include information on the study’s purpose and funding, respondent burden, and number of respondents involved. Taking into account the short period of performance and the limited burden the study will have on schools, our overall approach to district gaining cooperation will include a notification letter to all districts with phone follow up in all larger school districts (defined provisionally as those with more than 10 schools with grades 8 or higher). This letter will briefly explain the study purpose and scope and will be sent via Fed Ex to either the superintendent or the district office responsible for reviewing research requests as soon as the districts are identified for inclusion in the study. The follow up phone calls to the larger districts will seek to determine whether a formal application is required and, if so, how to proceed with it as expeditiously as possible.
A week following the district letter mailing, except in those districts which will require additional treatment, NORC will send via Priority mail a letter to the principals of the sampled schools. The letter will include information on the study, including teacher burden, and assure the principals that we sent a letter to their district superintendent as well. In addition, the letter will notify the principals that someone from NORC will be calling within a week to answer questions and collect the names of their algebra 1 teachers.
Maximizing teacher response rate. The teacher response rate to the survey is calculated as the proportion of sampled teachers who return a completed questionnaire. NORC anticipates a 80% response rate to the teachers survey. However, to maximize this rate, NORC will implement a multimodal (mail, phone, e-mail) contact strategy that includes multiple attempts to reach each individual teacher sampled. This approach is built on the principles developed by Don A. Dillman in his Mail and Internet Surveys (2nd edition) (John Wiley & Sons, 2000) and is designed to increase response rates by giving sampled teachers many opportunities to participate in the survey. Five to seven attempts to contact the respondent will be undertaken. The contractor (NORC) is also providing incentive payments of $20 per sampled teacher (included along with the first questionnaire mailing).
The first contact will be through the mail in early April, and will consist of a “prenotice” letter briefly explaining that the teacher has been selected to participate in the study and that the questionnaire will arrive in the next few days. The first questionnaire mailing will include the instrument, a cover letter explaining the purpose of the study, a check for $20 as a token of our appreciation, and a prepaid return envelope for the completed questionnaire. One week after the first mailing, NORC will send a postcard to each teacher, thanking them if they have already completed or reminding them to do so if not. The postcard will also provide them with a phone number and e-mail to contact in case they need another survey mailed out to them. Three weeks after the first mailing, NORC will send a replacement questionnaire and a follow-up letter requesting that they complete and mail back the survey. Four weeks after the first mailing, NORC will begin phone prompts to remaining non-respondents, and use e-mail if available and efforts to contact by phone were unsuccessful. Six weeks after the initial mailing, NORC will mail out a third and final copy of the questionnaire to those who have not yet responded along with a cover letter and postage-paid return envelope.
The NSTA has undergone an accelerated period of testing of the items in the questionnaire, and of obtaining expert advice on the revised items. The proposed questionnaire was tested as follows:
NORC project staff conducted cognitive interviews with fewer than 9 algebra I teachers in Chicago;
NORC project staff convened a panel of questionnaire design experts from the NORC staff to review the draft questionnaire item by item and provide feedback on how to improve the instrument.
NORC obtained feedback from 4 experts on the National Mathematics Advisory Panel.
B.4. Reviewing Statisticians
The statistical experts associated with this survey are Dr. Michael Yang, Senior Statistician on the project at NORC (202-223-9160). At USED, Grover J. (Russ) Whitehurst, Director of the Institute for Educational Sciences will provide statistical oversight. (Point of Contact in the Department of Education for this survey is Marian Banfield, National Math Panel, Office Planning, Evaluation and Policy Development (202-401-7767).
File Type | application/msword |
File Title | SF-83-1 SUPPORTING STATEMENT |
Author | marian.banfield |
Last Modified By | kate.mullan |
File Modified | 2007-03-30 |
File Created | 2007-03-30 |