CRCP SS Part B 111711 rev_030512 comments_clean

CRCP SS Part B 111711 rev_030512 comments_clean.docx

Socioeconomics of Coral Reef Conservation

OMB: 0648-0646

Document [docx]
Download: docx | pdf

SUPPORTING STATEMENT


SOCIOECONOMICS OF CORAL REEF CONSERVATION


OMB CONTROL NO. 0648-xxxx



U.S. Department of Commerce

National Oceanic and Atmospheric Administration

Coastal Services Center

Coral Reef Conservation Program



B. COLLECTIONS OF INFORMATION EMPLOYING STATISTICAL METHODS


1. Describe (including a numerical estimate) the potential respondent universe and any sampling or other respondent selection method to be used. Data on the number of entities (e.g. establishments, State and local governmental units, households, or persons) in the universe and the corresponding sample are to be provided in tabular form. The tabulation must also include expected response rates for the collection as a whole. If the collection has been conducted before, provide the actual response rate achieved.


The potential respondent universe for this study is adults, eighteen years or older, who live near, and may use, coral reefs affected by activities related to the NOAA’s Coral Reef Conservation Program. The total population (all individuals) of the potentially impacted area is 11,244,759. Respondents will be classified into seven geographical jurisdictions and 22 reporting units, as defined in Table 5. In American Samoa, face-to-face interviews will be conducted and in the remaining jurisdictions, a combination of internet and phone based interview will occur. Each of the geographical jurisdictions is expected to be surveyed once every three to four years. Respondents will be randomly selected from the target audiences. Based on previous NOAA surveys of the target populations, we anticipate that the response rate will be 50 percent for mail surveys, 40% for internet surveys and 80 percent for in-person surveys.1 Dillman et al. (2009) considers a response rate above 50% a high response rate for mail surveys.2




Table 5: Study Jurisdictions and Reporting Units

Jurisdiction

Reporting units

Population

Puerto Rico

  • Western Islands

  • Eastern Islands

  • South & West PR

  • North & East PR

3,725,7893


Florida


  • Monroe County

  • Miami-Dade County

  • Martin County

  • Broward County

  • Palm Beach County

5,784,0434

U.S. Virgin Islands

  • St. Thomas

  • St. Croix

  • St. John

106,4055

Guam

  • 7 Municipalities

159,3586

American Samoa


  • Tutuila Island

  • Ofu County

  • Olosega County

  • Tpau County

55,0707

Main Hawaiian Islands

  • Hawaii County

  • Honolulu County

  • Kauai County

Maui County

1,360,2118

Commonwealth of the Northern Mariana Islands (CNMI)

  • Rota Municipality

  • Tinian Municipality

Saian Municipality

53,8839

Total

11,244,759


2. Describe the procedures for the collection, including: the statistical methodology for stratification and sample selection; the estimation procedure; the degree of accuracy needed for the purpose described in the justification; any unusual problems requiring specialized sampling procedures; and any use of periodic (less frequent than annual) data collection cycles to reduce burden.


For each of the jurisdictional populations, we intend to select a random sample of individuals over the age of eighteen, stratified geographically as described in Table 6. The random sample will be obtained from the selected survey firm using standard sample selection tools. The sample frame will be developed from telephone directories, mailing lists obtained and maintained by the survey firms and other sources as needed, depending onr the coverage of these sources. These strata have been designed to account for the differing sizes of the populations in the areas close



to coral reefs. We have used the standard approach to estimating sample size for a stratified population:

[t2 N p(1-p)] / [t2 p(1-p) + a2 (N-1)]

Where N is the size of the total number of cases, n is the sample size, a is the expected error, t is the value taken from the t distribution corresponding to a certain confidence interval, and p is the probability of an event. The final sampling size will be based on available resources.

Table 6: Sampling Requirements by Geographical Jurisdictions



Jurisdiction

Total Sample

Sample Size by Strata

1. American Samoa

550

400

Tutuila

Island

50

Ta’u

County

50

Olosega

County

50

Ofu

County

2. CNMI

500

400

Saipan

Municipality

50

Tinian

Municipality

50

Rota

Municipality

3. Guam

1,000

100

Agat

Municipality

100

Piti

Municipality

100

Asana

Municipality

100

Talafolo

Municipality

100

Merizo

Municipality

250

Tamuning

Municipality

250

Mangilao

Municipality

4. Hawaii

1,500

350

Hawaii

County

750

Honolulu

County

100

Honolulu

County

300

Maui

County

5. Florida

4,000

250

Monroe

County

1,250

Miami-Dade

County

500

Martin

County

1,000

Broward

County

1,000

Palm Beach

County

6. Puerto Rico

2,500

1,000

South & West

PR

1,000

North & East

PR

250

Western Islands

PR

250

Eastern Islands

PR

7. US Virgin Islands

750

350

St. Croix

Island

350

St. Thomas

Island

50

St. John

Island

Total

10,800


In addition to asking the questions regarding the impact of the Coral Reef Conservation Programs, the surveys will collect information on socioeconomics and demographics. This additional information will be used to sort and categorize the survey results in order to control for as many variables as possible. This approach will ensure a large enough respondent pool (particularly in more populated jurisdictions) to make comparisons between strata where required.


In each of the jurisdictions, we intend to hire local surveying contractors with databases of contact information in order to allow for the greatest possible randomization of survey participants. NOAA will also work with these contractors to select the most cost-effective survey methodology which will resonate with the population measured. Survey participants will be selected in American Samoa for face-to-face interviews due to the very low incidence of either cellular phones or land lines; in other locations, local opinion poll contractors will select participants at random using a combination of internet and telephone polling. The methodology to be employed by jurisdiction can be found in Table 7.


Table 7: Survey Methodology by Geographical Jurisdictions

Jurisdiction

Geographic scope

Estimated Response Rate (based on previous NOAA surveys)

1. American Samoa

Face-to-face (and mail as back-up)

50-80%

2. CNMI

Telephone

50-80%

3. Guam

Telephone

50-80%

4. Hawaii

Telephone or Internet

50-80%

5. Florida

Telephone or Internet

50-80%

6. Puerto Rico

Telephone or Internet

50-80%

7. US Virgin Islands

Telephone or Face-to-face

50-80%


We do not intend to compare survey results between jurisdictions (though comparisons between the larger regional strata are possible), so there is no concern about comparability issues between methodologies.

Survey Specific Challenges

As can be seen from Table 7, we have selected a number of different methods to collect data from different jurisdictions. Table 8 highlights the percent of population classified as internet users for the seven jurisdictions. In general we will attempt to collect data using a mixture of internet and telephone methods. The one exception is American Samoa where an in-person household survey backed-up by mail surveys will be conducted due to the extremely low level of internet usage in this jurisdiction (i.e., approximately 6 percent). In addition, the average internet use in CNMI, Guam, Puerto Rico and the US Virgin Islands is 39 percent as compared to 79 percent for Hawaii and Florida. As a result we will support the internet survey in these jurisdictions with lower internet usage with a telephone survey, or in the case of USVI with face-to-face interviews, to capture non-internet users.




Table 8: Internet Usage in Survey Jurisdictions

Jurisdiction

Population

Percent of Population Classified as Internet Users

1. American Samoa

55,070

6%

2. CNMI

53,883

30%

3. Guam

159,358

56%

4. Hawaii

1,360,211

79%

5. Florida

5,784,043

80%

6. Puerto Rico

3,725,789

40%

7. US Virgin Islands

106,405

28%

Source: Data from Hawaii and Florida US Census 2010. Other data “Internet World Statistics”, American Samoa data March 2011, CNMI data from August 2010, Guam data from June 2010, Puerto Rico data from, June 2011, and US Virgin Islands from December 2002 (see http://www.internetworldstats.com/).


We expect that there will be some language issues. Table 9 shows there are several major languages spoken beyond English by the populations of each jurisdiction.


Table 9: Languages Spoken in Survey Jurisdictions

Jurisdiction

Major Languages Spoken

1. American Samoa

English, Samoan

2. CNMI

English, Chamorro, Carolinian, Tagalog, Chinese, Korean, Japanese

3. Guam

English,Chamorro, Tagalog, Chinese, Korean, Japanese

4. Hawaii

English, Hawaiian pigin

5. Florida

English, Spanish

6. Puerto Rico

English, Spanish

7. US Virgin Islands

English, Negerhollands, Virgin Islands Creole


This language issue will be ameliorated by the use of polling specialists who speak the local language. These contractors will also be used to ensure that the questions posed in the survey are translated into the proper cultural contexts. Responses will be tracked to see if there are statistically significant differences in the survey results between those who speak English at home and those who do not. In addition, mail and internet surveys will be translated into local languages.


We also expect that there is some risk of sample selection bias towards those of higher incomes, particularly for the telephone and internet surveys. In areas where access to phone and internet services are not widely available, this bias may be more than minimal. To the greatest extent possible, we hope that this can be corrected through the use of telephone surveys. If responses appear to favor high-income groups we will use various weighting procedures in the post-survey analysis to adjust for bias. Specifically, we will overweight the underrepresented groups if expected responses are not obtained. We will identify ‘control totals’ for the population that the survey is aiming to reach and calculate weights to adjust the sample totals to the control totals. For example, suppose the distribution of income groups in the population is as follows:

Income

Percent of Population

$0-$25K

28.00

$25K-$50K

27.00

$50-$75K

18.00

$75K-$100K

11.00

$100K plus

16.00

Total

100.00

However, the response distribution is:

Income

Percent of Population

$0-$25K

5.00

$25K-$50K

27.00

$50-$75K

18.00

$75K-$100K

11.00

$100K plus

39.00

Total

100.00

In response to this disparity, we would weigh the sample for the “$0-$25K” and the “$100K plus” groups to bring it in to line with the proportion of that group in the population as a whole. That is, we would apply a weight of 5.6 to the results for the “$0-25K” group (i.e., 28 percent divided by 5 percent) and a weight of 0.41 to “$100K plus” group to deal with their overrepresentation.

Periodicity

This survey will be conducted every three to four years to minimize the cost burden.

3. Describe the methods used to maximize response rates and to deal with nonresponse. The accuracy and reliability of the information collected must be shown to be adequate for the intended uses. For collections based on sampling, a special justification must be provided if they will not yield "reliable" data that can be generalized to the universe studied.


While the surveys conducted in person are expected to yield standard rates of response (80% based on previous NOAA surveys of targeted population),10 there is some concern about the potential for non-response in the telephone/internet surveys. While response rates for many surveys have been declining in the United States for years, previous studies have indicated that the low response rates commonly associated with internet polling can be somewhat improved with the use of pre-poll telephone calls. To accomplish this, polling representatives ask respondents whether they are willing to participate in the online study and then direct them via a secure link or email. In addition we will conduct extensive online advertising to encourage response. Research has shown that under these conditions internet and telephone surveys can reach similar response rates as those found in mail surveys.11


A variety of techniques have been incorporated into this study to maximize response rates. The surveys are user-friendly, with clear, easy to comprehend questions. Each questionnaire is short and can be completed in a short period of time (see Part A). The survey topic and related questions were developed to be interesting to respondents. Each survey makes use of listing options to allow the respondent to answer questions by checking appropriate boxes, which may aid in recall and analysis.


In person surveys will be conducted at respondent’s homes and participants will be given the opportunity to receive and /or return the survey by mail if they are unable to complete the surveys at the time of interview. These individuals who complete the survey by mail will receive a pre-addressed stamped return envelope.


The implementation of the mail surveys is based on the Dillman Tailored Design Method.12 This approach includes multiple steps and points of contact. The initial mailing will include the questionnaire, a pre-addressed stamped envelope and a detailed cover letter. The cover letter will explain the project, why a response is important, a statement indicating that all personal information will be kept confidential, and instructions for completing and returning the completed survey (via mail/fax/email). Addresses on envelopes will be handwritten, and colored envelopes will be used to make them stand out. Surveys will be tracked using individual identification numbers. A follow-up thank you postcard will be sent seven to nine days after the questionnaire. The postcard will express appreciation for participating and will indicate that if the completed questionnaire has not yet been mailed, it is hoped that it will be returned soon. Three weeks after the initial mailing, a second mailing will be sent to all who have not returned the survey. This follow-up will consist of a different cover letter, another copy of the questionnaire, and another pre-addressed stamped envelope.


For internet surveys we will use a number of techniques13 to increase response including:

  • Subject lines on contact emails will clearly indicate the purpose of the survey and will explicitly avoid SPAM language in the subject line or body of the message (I.e. title all caps)

  • Information on how the respondents name was obtained, the survey intention, the use of the data, guarantees of anonymity

  • Personalized messages

  • Use of a .gov reply email address

  • Indication of how long the survey takes to complete and the cutoff date.

  • Use of only clean and updated email lists

  • Scheduled regular reminders and follow-ups.


Cross-cultural research faces additional methodological challenges that, if not properly addressed, may considerably increase the risk of inferential errors during the administration of surveys.14 Specifically, concepts may entail culture-specific attributes and meanings which need to be explicitly taken into account to ensure sound interpretation of cross-cultural data.15 As discussed above (see Question 2), we will address this cross cultural issue by using polling specialists who speak the local language to conduct in-person and phone surveys. These polling specialists’ knowledge of local culture and idioms are anticipated to have a positive impact on survey response rates.


In terms of increasing response rates for telephone surveys, we will use a number of techniques. First, we will work with survey firms to ensure that we have an accurate up-to-date list of phone numbers from which to draw potential respondents. Second, we will use a combination of proven approaches to increase surveys response, including conducting interviewer training so interviewers are sensitive to cultural issues and know how to administer the survey, setting clear establishment of researcher credentials in the introduction, and increasing call attempts and targeting call times. These methods have proven to be an effective approach in increasing response rates for telephone surveys. 16 In addition, we will ensure that we have multilingual telephone staff available for specific calls where language may be an issue.

4. Describe any tests of procedures or methods to be undertaken. Tests are encouraged as effective means to refine collections, but if ten or more test respondents are involved OMB must give prior approval.


Contracted polling groups will be asked to demonstrate their survey administration techniques on nine participants prior to execution of the full survey. This will include participants interviewed by those speaking their languages. This sample test will allow for the refinement and correction of any methodological issues that are identified.


5. Provide the name and telephone number of individuals consulted on the statistical aspects of the design, and the name of the agency unit, contractor(s), grantee(s), or other person(s) who will actually collect and/or analyze the information for the agency.


Christy Loper, Ph.D.
Coral Reef Conservation Program
US National Oceanic and Atmospheric Administration

on detail to:
Office of the Assistant Secretary for Conservation and Management
1401 Constitution Avenue, Suite 6224
Washington, DC 20230
202-482-5143 (office)
240-429-7044 (cell)
[email protected]



Individuals consulted on the statistical aspects of the design:

Victoria Adams, Ph.D.

Economist

Booz Allen Hamilton

8283 Greensboro Drive

McLean, VA, 22102

Telephone: (703)-377-4942


The individuals and firms that will collect and analyze the data have not yet been identified.


References


Cantor, D. and Cunningham, P. (2002) “Methods for Obtaining High Response Rates in Telephone Surveys” in “Studies of Welfare Populations: Data Collection and Research Issues Panel on Data and Methods for Measuring the Effects of Changes in Social Welfare Programs”, Eds.Ver Ploeg, M, Moffitt, R.A. and Citro, C.F. , Committee on National Statistics Division of Behavioral and Social Sciences and Education National Research Council.


Dillman, D., J. Smyth and L. Christian. (2009) Internet, Mail and MixedMode Surveys: The Tailored Design Method. New York: John Wiley & Sons.


Dillman, D. A. (2000). Mail and Internet surveys: The total design method (2nd ed.). New York:

Wiley. Division of Instructional Innovation and Assessment, The University of Texas at Austin. “Guidelines for Maximizing Response Rates.” Instructional Assessment Resources. 2007. http://www.utexas.edu/academic/diia/assessment/iar/teaching/gather/method/surveyResponse.php


Loomis, David K., “Beach Users Perceptions Concerning Zuma Beach Restoration”, University of Massachusetts Amherst, 2009.


NOAA, “2011 National Marine Recreational Fishing Expenditure Survey” 2011.


Peng, T. K., Peterson, M. F., & Shyi, Y.-P. (1991). Quantitative Methods in Cross-National Management Research: Trends and Equivalence Issues. Journal of Organizational Behavior, 12(2), 87-107.


Public Perception and Attitudes about the Hawaiian Monk Seal, Survey Results Report”, Sustainable Resources Group International, Inc., prepared for NOAA Fisheries Service Pacific Islands Regional Office, April 2011.


Singh, J. (1995). Measurement Issues in Cross-Cultural Research. Journal of International Business Studies, 26(3), 597-619.


Washington-Oregon-California Purse Seine Survey”, NOAA, 2007, OMB Control #: 0648-0369, Gulf States Marine Fisheries Commission (GSMFC).




1 See “Public Perception and Attitudes about the Hawaiian Monk Seal, Survey Results Report”, Sustainable Resources Group International, Inc., prepared for NOAA Fisheries Service Pacific Islands Regional Office, April 2011. “Beach Users Perceptions Concerning Zuma Beach Restoration”, David K. Loomis, University of Massachusetts Amherst, 2009.

“Washington-Oregon-California Purse Seine Survey”, NOAA, 2007, OMB Control #: 0648-0369, Gulf States Marine Fisheries Commission (GSMFC). NOAA,“2011 National Marine Recreational Fishing Expenditure Survey” 2011.

2 Dillman, D., J. Smyth and L. Christian. (2009) Internet, Mail and Mixed‐Mode Surveys: The Tailored Design Method. New York: John Wiley & Sons.

4 http://factfinder2.census.gov/faces/nav/jsf/pages/index.xhtml [Census Total Population figures for 5 counties, 2010]

7 http://2010.census.gov/news/xls/cb11cn177_as.xls [Eastern & Western Districts; Ofu, Olosega, Tau Counties, 2010]

8 http://factfinder2.census.gov/faces/nav/jsf/pages/index.xhtml [Census Total Population figures for Hawai’i, Honolulu, Kauai, and Maui Counties, 2010]

10 See “Public Perception and Attitudes about the Hawaiian Monk Seal, Survey Results Report”, Sustainable Resources Group International, Inc., prepared for NOAA Fisheries Service Pacific Islands Regional Office, April 2011, “Washington-Oregon-California Purse Seine Survey”, NOAA, 2007, OMB Control #: 0648-0369,  Gulf States Marine Fisheries Commission (GSMFC) and NOAA, “2011 National Marine Recreational Fishing Expenditure Survey” 2011.

11 See, 2000, “Complementary Methodologies: Internet versus Mail Surveys”, DSS Research, Inc.

12 Dillman, D., J. Smyth and L. Christian. (2009) Internet, Mail and Mixed Mode Surveys: The Tailored Design Method. New York: John Wiley & Sons.

13 See Dillman, D. A. (2000). Mail and Internet surveys: The total design method (2nd ed.). New York:

Wiley. Division of Instructional Innovation and Assessment, The University of Texas at Austin. “Guidelines for Maximizing Response Rates.” Instructional Assessment Resources. 2007. http://www.utexas.edu/academic/diia/assessment/iar/teaching/gather/method/surveyResponse.php

14 Singh, J. (1995). Measurement Issues in Cross-Cultural Research. Journal of International Business Studies, 26(3), 597-619.

15 Peng, T. K., Peterson, M. F., & Shyi, Y.-P. (1991). Quantitative Methods in Cross-National Management Research: Trends and Equivalence Issues. Journal of Organizational Behavior, 12(2), 87-107.

16 This approaches have been shown to have a positive impact on response rates, see Cantor, D. and Cunningham, P. (2002) “Methods for Obtaining High Response Rates in Telephone Surveys” in “Studies of Welfare Populations:

Data Collection and Research Issues Panel on Data and Methods for Measuring the Effects of Changes in Social Welfare Programs”, Eds.Ver Ploeg, M, Moffitt, R.A. and Citro, C.F. , Committee on National Statistics Division of Behavioral and Social Sciences and Education National Research Council.

File Typeapplication/vnd.openxmlformats-officedocument.wordprocessingml.document
Authorsarah.brabson
File Modified0000-00-00
File Created2021-01-31

© 2024 OMB.report | Privacy Policy