PRA Part B Speed Survey

PRA Part B Speed Survey.docx

National Survey of Speeding Attitudes and Behaviors

OMB: 2127-0613

Document [docx]
Download: docx | pdf

Information Collection Request Supporting Statement: Section B

National Survey of Speeding Attitudes and Behaviors

OMB Control Number 2127-0613


Abstract:1 The National Highway Traffic Safety Administration (NHTSA) of the U.S. Department of Transportation is seeking approval to reinstate a previously approved information collection to collect information from licensed and non-licensed adult drivers (18 years old and older) who drive a motor vehicle for a voluntary study on speeding attitudes and behaviors. NHTSA proposes to conduct the National Survey of Speeding Attitudes and Behaviors (NSSAB) by contacting an estimated 21,800 households by mail for participation. The push-to-web with mail supplement survey will be completed by a national probability sample of at least 7,000 adult drivers, age 18 and older. Participation by respondents would be voluntary. This collection only asks respondents to report their answers; there are no record-keeping costs to the respondents. NHTSA will use the information to produce a technical report that presents the results of the survey. The technical report will provide aggregate (summary) statistics and tables as well as the results of statistical analysis of the information, but it will not include any personally identifiable information. The purpose of the survey is to obtain up-to-date information from drivers about their speeding and their attitudes, perceptions, and motivations. NHTSA will use this data to identify countermeasures that are most likely to reduce speeding behavior. The technical report will be shared with State highway offices, local governments, and those who develop traffic safety communications that aim to reduce speed-related crashes. The total estimated burden for contacting 13,379 potential participant non-responders (1,273 hours) and 288 potential pilot participant non-responders (29 hours), contacting and screening out 208 ineligible participants (13 hours) and 4 ineligible pilot participants (2 hours), and contacting and recruiting at least 7,000 participants (2,457 hours) and 152 pilot participants (56 hours) to complete the study is 3,830 total hours. All estimates were rounded up to the nearest whole hour. When NHTSA last received approval of this information collection, the estimated burden was 2,010 hours. The increase in burden of 1,820 hours is a result of using a larger sample and including burden not just for the estimated number of completed surveys, but also for the estimated number of contacts of potential respondents. NHTSA has conducted the National Survey of Speeding Attitudes and Behaviors on three previous occasions—first in 1997,2 again in 2002,3 and most recently in 2011.4 (The final report for the 2011 administration of the survey is included as a Supplemental Document.) Up-to-date information is needed to identify trends across time and develop appropriate countermeasures for speeding-related traffic safety issues. Study results should produce useful information to traffic safety stakeholders. The legacy study is being redesigned to sample respondents using address data from the most recent U.S. Postal Service (USPS) computerized Delivery Sequence File (DSF) of residential addresses, and administer the survey via web and mail (replacing the former random-digit dial computer-assisted telephone interview design).


B.1. Describe the potential respondent universe and any sampling or other respondent selection method to be used.


The purpose of the survey is to obtain up-to-date information from drivers about their speeding; their attitudes; perceptions; and motivations, as well as identify countermeasures that are most likely to reduce speeding behavior. The study will collect information from at least 7,000 licensed and non-licensed adult drivers (18 years old and older) who drive a motor vehicle. Both licensed and non-licensed drivers will be surveyed as it is important to receive feedback about behaviors and attitudes from all drivers regardless of license status as they are all part of the driving environment. The proposed study will employ statistical sampling methods to collect information from the target population and draw inferences from the sample to the target population. The technical report will be shared with State highway offices, local governments, and those who develop traffic safety communications that aim to reduce speed-related crashes.


B.1.a. Respondent Universe

The 2020 NSSAB will be conducted with a national sample of at least 7,000 drivers ages 18 years and older, residing in the 50 States and the District of Columbia. A driver is defined as a person who has driven within the past year. Approximately 3% of adults in the United States are not drivers. Consequently, the design selects a probability sample of adults from sampled households. The interview screens for driving status of the selected adult and collects demographic data on non-drivers before they screen out while collecting the full survey data on drivers.5

We will conduct a web and mail multi-mode survey with households randomly selected from an address-based sampling (ABS) frame. We will stratify the ABS frame based on the 10 NHTSA regions and proportionally allocate the total sample of addresses to the total number of residential addresses in each region. Two oversamples will be incorporated into the national sample, one to increase the number of Spanish speaking households and one to increase the number of drivers aged 18-34 years old.

B.1.b. Respondent Sampling

The survey will use an ABS (Address-Based Sampling) approach to sample selection. The sampling frame will be based on address data from the U.S. Postal Service (USPS) computerized Delivery Sequence File (DSF) of residential addresses. The DSF is derived from mailing addresses maintained and updated by USPS and available from commercial vendors.6,7 With 146-million residential addresses nationally, the DSF provides a comprehensive frame that will reach the entire population living at an address that receives mail delivery.

B.1.b.1 Sampling Frame

The sampling frame will be based on address data from the U.S. Postal Service (USPS) computerized Delivery Sequence File (DSF) of residential addresses. The DSF is a computerized file that contains all delivery point addresses serviced by the USPS with the exception of general delivery. Each delivery point is a separate record that conforms to all USPS-addressing standards. The initial studies of the DSF estimated that it provided coverage of approximately 97-98% of the household population.8,9 The DSF coverage in rural areas tends to be lower than in urban areas10 but is increasing as more rural areas are converted to city-style addresses for 911 services.11 Nonetheless, the DSF address frame provides a near complete sampling frame for household population surveys in the United States. With 129-million residential addresses nationally, the DSF provides a comprehensive frame that will reach the entire population living at an address that receives mail delivery.

The DSF cannot be obtained directly from the USPS. It must be purchased through a licensing agreement with private vendors. These vendors are responsible for updating the address listing from the USPS, and augmenting the addresses with information (e.g., name, telephone number) from other data sources. ICF, the Contractor that will implement the NSSAB for NHTSA, will obtain the DSF augmented sample from Marketing Systems Group (MSG). By geocoding an address to a Census block, the MSG file augments the DSF by merging Census and other auxiliary information from the Census data files and other external data sources. MSG appends household, geographic, and demographic data to the frame.

MSG maintains a monthly updated, internal installation of the DSF from the Postal Service. By applying a series of enhancements to the DSF, MSG evolves this database of mail delivery into a sampling frame capable of accommodating multiple layers of stratification or clustering when selecting probability-based samples. Address enhancements provided by MSG include amelioration of some of the known coverage problems associated with the DSF, particularly in rural areas where more households rely on P.O. Boxes and inconsistent address formats.

There were approximately 146 million residential addresses in the DSF as of November 2019. This excludes business addresses. It also excludes addresses labeled as “No Stat” which are generally addresses where there is no mail delivery, such as buildings for which building permits have been obtained but mail delivery has not commenced.

The sampling frame for the NSSAB will include all residential addresses, including city-style addresses (89.2%), P.O. boxes (10.8%), rural routes (<0.1%), and highway contracts (<0.1%). The frame will exclude P.O. boxes where the household also receives home delivery. The DSF classifies P.O. Boxes as Only Way to Get Mail (1.4 million) or traditional Post Office Box where the household also receives delivery at a street address (14.3 million). The NSSAB will only include the Only Way to Get Mail (OWGM) P.O. Boxes since people having the latter are represented in the sampling frame based on their home delivery address.

The DSF includes flags identifying the address as seasonal (<0.1%) or vacant (7.0%). To maximize coverage of the population, the NSSAB frame will include these addresses.

Drop points are building addresses with multiple deliveries and no separate addresses within the building (i.e., apartment numbers). Drop units --- the number of delivery units within drop points --- represent about 2% of all residential addresses. In actual mail delivery, the drop units have names attached so that mail can be appropriately routed within the building by tenant or landlord. However, the commercial DSF file only provides the number of drop units within a drop point address. The most common approaches to handling drop points in address based samples are to either exclude the drop points (or those with more than a few drop units) or include all drop units for any selected drop point since there is no basis for selection within the drop unit. NSSAB will include drop points in the sampling frame. The drop points will be expanded based on the number of units at that location. If a drop point is selected, research on the units will be conducted to determine the unit identifiers for the building.

Some addresses are classified as educational (<0.1%), which represents student housing. They are effectively a special type of drop point since there are not individual unit addresses within the buildings. They will be included in the sample, like drop points, particularly given the importance of the young adult sample to this survey and its under-representation in most population surveys.

B.1.b.2 Sample Sizes

The NSSAB frame will be stratified into the 10 NHTSA regions and the overall sample of 7,000 will be allocated based on the total number of residential addresses in each region with a minimum of 500 per region. Assuming a 1.75 design effect due to weighting, we expect national estimates to have a margin of error of +/–1.5 percentage points at the 95% confidence level and error margins for regional estimates ranging from +/-4.8% to +/-5.8%.12 Table 1 contains the total number of occupied housing units, total addresses on the DSF, and the target sample size by region.


Table 1. Regional Frame Counts and Sample Size

NHTSA Region/States

2017 ACS Occupied HUs (household units)

Addresses

Selected Addresses

Expected completes

+/–95% CI

United States

118,825,921

132,068,579

21,800

7,000

1.5%

1

Maine, Massachusetts, New Hampshire, Rhode Island,

Vermont

4,337,049

4,666,766

1,555

500

5.8%

2

Connecticut, New Jersey, New York, Pennsylvania

16,871,018

18,017,814

2,585

830

4.5%

3

Delaware, District of Columbia, Kentucky, Maryland, North Carolina, Virginia, West Virginia

12,253,602

13,752,019

1,960

630

5.2%

4

Alabama, Florida, Georgia, South Carolina, Tennessee

17,449,182

20,869,071

2,995

960

4.2%

5

Illinois, Indiana, Michigan, Minnesota, Ohio, Wisconsin

20,359,388

22,313,285

3,195

1,025

4.1%

6

Louisiana, New Mexico, Mississippi, Oklahoma, Texas

14,510,984

16,775,644

2,400

770

4.7%

7

Arkansas, Iowa, Kansas, Missouri, Nebraska

6,655,429

7,241,118

1,555

500

5.8%

8

Colorado, Nevada, North Dakota, South Dakota, Utah, Wyoming

4,954,365

5,415,274

1,555

500

5.8%

9

Arizona, California, Hawaii

15,825,941

17,045,873

2445

785

4.6%

10

Alaska, Idaho, Montana, Oregon, Washington

5,608,963

5,971,715

1555

500

5.8%


Addresses from the DSF are drawn based on a 1-in-k systematic sample so that each address has an equal probability of selection within each stratum. The ABS database is sorted by ZIP+4 within state to ensure a geographically proportional allocation.

To oversample Spanish-speaking households, we will geographically stratify block groups with high percentages of Spanish-speaking households as identified in the 2013–2017 American Community Survey (ACS).

To oversample young adults, we will use a two-phase sample, called double sampling, for stratification. We will first select a national sample of addresses from the DSF as described above. We will then append a model-based age indicator (provided by MSG), identifying addresses where the head of household is likely to be aged between 18 and 34 years. We will stratify the sample based on the 18–34 indicator and select the second-phase sample by oversampling addresses in the 18–34 stratum relative to the non-18–34 stratum. While agreement rates tend to be low on telephones,13 they tend to be higher on addresses.14 Thus, we expect that stratifying by the “likely 18–34” indicator will increase the number of young adult respondents in the sample. We will correct for the oversampling of addresses in this stratum in the weighting.

B.1.b.3 Within-Household Selection

A number of respondent selection methods have been tested for ABS mail surveys, including for the Behavioral Risk Factor Surveillance System (BRFSS).15 Although past studies have indicated a tendency for the wrong person to complete the survey when applying birthday methods of within-household selection16 a recent evaluation of birthday selection methods for ABS surveys found a small degree of self-selection in larger households; however, the impact on the substantive estimates was small.17 Considering the low impact of the overall estimates and the simplicity of implementing the birthday methods, we will select the adult within the household who has the next birthday to complete the survey (as opposed to the last birthday or a split next/last sample). The within-household selection instructions will be included in all contacts with the household.


B.1.c. Response Rate

Table 2 details our assumptions for sample size and response rate by data collection wave. These assumptions are based on similar contact waves and response rates achieved for the 2016 Motor Vehicle Occupant Safety Survey (MVOSS-Version A), a national survey about driving behaviors and attitudes similar in length to the NSSAB. We expect to draw an initial sample of 21,800 addresses and that approximately 5.5% of these records will be non-deliverable. We expect the 20,600 remaining valid records—following the five-contact protocol—will result in an estimated 7,221 returned surveys for a response rate of 33% based on the American Association of Public Opinion Research (AAPOR) response rate formula #1 (RR1). (Previous NSSAB surveys were administered using random digit dialing telephone sampling instead of the current mode of mail with push to web. As a comparison to the current study, the 1997 survey achieved a 73.5% response rate; the 2002 survey had a 30.5% response rate; the 2011 survey had a 40.3% landline, 18.6% cell phone, and 44.6% oversample response rate.) Assuming 97% of the returned surveys qualify as drivers,18 we will obtain at least 7,000 completed surveys.


Table 2. Expected Data Collection Quantities and Response Rates

Contact Wave

Number of Records

Expected Return Rate

Completed Surveys

Eligible drivers

Web letter invitation

21,800




Undeliverable

1,200




Valid addresses

20,600

12%

2,470

2,398

Reminder postcard #1

18,130

9%

1,632

1,585

First survey mailing

16,498

12%

1,980

1,923

Reminder postcard #2

14,518

5%

726

705

Second survey mailing

13,793

3%

414

402

TOTAL RESPONDENTS



7,221

7,013



B.2. Describe the procedures for the collection of information.

B.2.a Data Collection Protocol

The Contractor, ICF, will select a national, stratified random sample of households from the DSF, as described in the previous section. Each household will be mailed an initial letter requesting participation in the survey. The survey will employ the next birthday method for random selection of one respondent aged 18 or over from the household.

Web response is NHTSA’s preferred method for the survey. Therefore, the survey will initially offer only a web response mode, where the letter requests the selected household member to go to a designated website to take the survey. Each letter/address will contain a unique Master ID that will be used to access the website and will help track whether someone from a household completed the survey. For those that do not respond, there will be a series of additional contact waves that will add alternative modes of responding. The contact waves are presented in Table 3. Households that respond or refuse the survey will be removed from subsequent contacts.






Table 3. NSSAB Contact Protocol

Wave

Step and Mode

Contents

Schedule

1

An invitation letter offering web response

Cover letter with PIN, hyperlink to web survey, instructions, $1 pre-incentive

Day 1

2

Reminder postcard #1

Postcard, black and white

Day 7

3

A mailed package offering mail response

Cover letter, 16-page printed questionnaire, prepaid return envelope

Day 28

4

Reminder postcard #2

Postcard, black and white

Day 35

5

A mailed replacement package offering mail response

Cover letter, 16-page printed questionnaire, prepaid return envelope

Day 49

Close data collection

Day 91



B.2.b Spanish-Language Data Collection

We will send materials in both Spanish and English to households highly likely to speak Spanish. Bilingual materials will be sent to households in block groups where the percentage of limited English-speaking households19 is at least 15% of the total households in the block group. We estimate that these block groups represent over 50% of the Spanish language isolate population. Areas outside these block groups will receive an English-language letter that contains information at the bottom, in Spanish, on how to access and complete the survey in Spanish. This approach is based on evidence from the Health Information National Trends Survey (HINTS) indicating that sending English and Spanish materials to everyone depresses response rates compared to sending Spanish- and English-language materials only to households most likely to speak Spanish (Westat, 2014). The web survey will have the option to complete the survey in English or Spanish.

B.2.c Precision of Sample Estimates

The objective of the sampling procedures described above is to produce a random sample of the target population. This means that with a randomly drawn sample, one can make inferences about population characteristics within certain specified limits of certainty and sampling variability.


The margin of error, d, of the sample estimate of a population proportion, P, equals:



Where tα equals 1.96 for 1-α = 0.95, and the standard error of P equals:


Where:

= design effect arising from the combined impact of the random selection of one eligible individual from a sample household, and unequal weights from other aspects of the sample design and weighting methodology, and

= the size of the sample (i.e., number of interviews)


Using these formulas, the margin of error for a sample size of at least 7,000 interviews is d = 0.015, using an average of 1.75. and setting equal to 0.50. We expect the design effect for the NSSAB to be similar to the design effect for MVOSS, which was 1.73 and 1.76 for the two survey versions.


The total sample size for the survey is also large enough to permit estimates for subgroup analysis including driver type, age, gender, and race/ethnicity and other demographics. A subgroup that represents at least 10% of the total sample will have a 95% confidence interval of +/-5%. Table 4 includes expected 95% error margins for demographic groups assuming the sample size is proportional to the population.


Table 4. Expected 95% Error Margins for Subgroups


Population Percentage

+/–95% CIs

Gender



Male

49%

2.2%

Female

51%

2.2%

Age Group



18-34

30%

2.8%

35-54

33%

2.7%

55+

37%

2.6%

Race/Ethnicity



Hispanic

16%

3.9%

Non-Hispanic White

64%

1.9%

Non-Hispanic Black

12%

4.5%

Non-Hispanic Other

9%

5.3%

Driver Type†



Non-Speeders

30%

2.8%

Sometime Speeders

40%

2.5%

Speeders

30%

2.8%

†Driver type distribution based on the 2011 NSSAB.


B.2.d Sample Weighting

The NSSAB will be weighted to reduce any potential bias related to differential selection probabilities and non-response. Since benchmark data is not available for drivers, we will collect demographic data for drivers and non-drivers. This will be the data used for weighting. The weighting process will compute:

  • Sampling weights that incorporate the probability of selection for households and the probability of selection of a respondent within a sample household;

  • Weight adjustments for non-response; and

  • Population calibration.

Sampling weights are the products of the reciprocals of the probabilities of selections associated with two sampling stages: (1) the selection of households from the ABS frame and (2) the selection of respondents within a household. The first-stage probabilities will be unequal to the extent that the sample allocation oversamples some of the regional strata based on age and Spanish-speaking likelihood. The address probability of selection is multiplied by the within-household probability of selection based on the number of adults in the household as reported in the survey.

We will apply weighting class adjustments designed to minimize the potential for non-response bias. These adjustments will be informed by the non-response analysis described previously. Specifically, we will select the variables used to define weight adjustment classes (or cells) using the propensity models in that analysis, as those predictors are most significant in the models. In general, adjustment classes will be homogeneous in terms of response behavior.

The weights will be calibrated based on known population totals for key demographics such as gender, age categories, education, marital status and race/ethnicity. The calibration will be based on raking, an iterative ratio adjustment of the sample to the population based on multiple dimensions. Because this step will align the survey respondents with the population, we will include non-drivers and drivers in the adjustment. Therefore, it is critical to collect basic demographic information from the non-drivers selected in the sample.


B.3. Describe methods to maximize response rates.

B.3.1 Maximizing Response

NHTSA is taking a number of steps to boost the NSSAB response rate. Foremost will be NHTSA’s use of the multi-mode approach, where different options for responding are sequentially presented to prospective respondents (web and mail). This offers greater opportunity for people to use a response mode that they prefer and with which they are comfortable, which should enhance participation.

The protocol includes up to five mailings to non-response households. The first contact will include a $1 pre-incentive and a $5 promised incentive upon completion. An incentive experiment done for MVOSS found that a $1 pre-incentive and $5 promised incentive provided the most cost-efficient way to increase response when compared to $2 pre-incentive and $10 post incentive.20

In contacting respondents, NHTSA will use official government envelopes for the mailings. People will often open government envelopes out of curiosity as to why they are being contacted by the government. As stated in the previous section, the invitation to participate in the survey will include wording in Spanish for those who are entirely or predominantly Spanish speaking so that they are not excluded from the survey.

In adapting the questionnaires to multi-mode administration, the project team will apply principles of heuristics that people follow in interpreting visual cues in visually laying out the questions.

Another facilitator of response will be adaption of the web-based questionnaires for mobile platforms (e.g., smartphones, tablets) so that prospective respondents who wish to use such devices when taking the survey are not deterred. Once a questionnaire is programmed, the platform will automatically adapt the presentation to optimize completion on a mobile device.

The survey will include assistance devices for respondents so that they do not become frustrated and terminate their participation prior to submission of a completed questionnaire. For the web response mode, this will include easy navigation from page to page, and the capability for respondents to pause and leave the system and then re-enter at the departure point without losing any previously inserted information.

During administration, ICF will maintain support for the respondents via an e-mail help desk and a toll-free phone number. Clear instructions for accessing this support will be provided on paper materials and the web survey.

B.3.2 Non-response Analysis

Non-response bias will occur if there are differential response rates for certain subgroups of the sample and these subgroups differ with respect to the substantive survey data.  Differential response occurs when one subgroup responds to the survey at a higher rate than another subgroup (e.g., males vs. females). Therefore, the non-response analysis will focus on the distribution of respondents as compared to the expected distribution based on the population.

The analysis of non-response bias for the NSSAB will follow three tracks.

  1. Bivariate analyses. First, the analysis will compare the distribution of survey respondents with known population distributions. This comparison will focus on key demographic variables such as race/ethnicity, gender, age groups, and education. Because many of these same factors will be used during post-stratification in the survey weighting process, the analysis will consider un-weighted data and data that are weighted prior to the post-stratification step, as well as using the final adjusted weights. Note that these analyses will capitalize on the augmented frame data (e.g., age flags used in the oversampling of young adults) as well as on Census data.

  2. Multivariate analyses. The demographic variables found to be significant in these bivariate analyses (or sub group analyses) will then be included in multivariate logistic models. In these logistic models, usually called propensity models, the dependent variable is a dichotomous (0-1) indicator for response so the logistic model may be expressed in terms of the probability of a response. The variables that turn out to be significant in these propensity models will be considered for weight adjustments for non-response (i.e., will be candidates for defining weight adjustment classes). This approach will ensure that weight adjustments minimize the potential for non-response bias.

  3. Comparisons across waves of respondents. The second set of analyses will compare responses obtained using different levels of effort. This approach typically compares early respondents to the initial survey waves (Waves 1 to 3) to respondents to the later waves (Waves 4 and 5). The idea is that the late respondents—a group of reluctant or perhaps recalcitrant respondents—resemble non-respondents statistically.

The non-response analysis will inform the weighting adjustments to correct for a sample that is disproportionate from the population. These weighting adjustments will mitigate the risk on non-response bias to the extent that the substantive survey data is correlated with the observed differences in respondents and non-respondents.


B.4. Describe any tests of procedures or methods to be undertaken.

No methodological tests are planned for the survey. The questions on the NSSAB have been cognitively tested and the web and paper survey instruments were subjected to usability testing. A total of six participants per mode completed the usability testing.

B.4.1 Web-based Questionnaire Usability Testing

During an in-person session, each participant was asked to follow the instructions in the invitation letter as if he or she were at home, starting with going to the website and accessing the survey. The individual was then asked to complete specific survey portions while thinking aloud. The facilitator noted errors and watched for hesitation, confusion, or frustration. Web-based questionnaire testing will include both desktop and mobile devices. Tests were recorded for review and identified:

  • Problems with following invitation letter instructions and/or accessing the survey;

  • Problems with navigating screens, sections, and questions;

  • Confusion about where and when responses are saved and returning to the survey later; and

  • Interface elements (e.g., icons, menus, buttons, forms, messages, warnings, alerts) that participants did not notice or understand or that do not behave as participants expect.

Adjustments were made to the web instrument based on the findings of this usability testing to correct for the above issues.


B.4.2 Paper Questionnaire Usability Testing

During these in-person sessions, the participant was given a copy of the appropriate invitation/reminder letter and the paper survey packet and was asked to complete survey items while thinking aloud. Tests were recorded for review and identified:

  • Not marking answers in the correct location or answers not fitting in the space provided;

  • Missing or misunderstanding instructions (e.g., choosing multiple responses in a case where only one response is allowed); and

  • Difficulty following skip patterns or answering questions as “not applicable.”

Adjustments were made to the paper instrument based on the findings of this usability testing to correct for the above issues.


B.5. Provide the name and telephone number of individuals consulted on statistical aspects of the design.


The following individuals have reviewed technical and statistical aspects of procedures that will be used to conduct the 2020 NSSAB:


Kristie Johnson

NHTSA Project Manager / COR(TO)

Research Psychologist

1200 New Jersey Avenue, SE

Washington, DC 20590

202-366-2755

E-mail: kristie.johnson@dot.gov


Andy Dyer

ICF, Project Director (Contractor)

126 College St

Burlington, VT 05401

802-264-3726

andrew.dyer@icf.com


Kisha Bailly

ICF, Administrative Manager (Contractor)

530 Gaither Road, Suite 500

Rockville, MD 20850

703-225-2160

kisha.bailly@icf.com


Randy ZuWallack

ICF, Senior Statistician (Contractor)

126 College St

Burlington, VT 05401

802-264-3724

randy.zuwallack@icf.com






1 The Abstract must include the following information: (1) whether responding to the collection is mandatory, voluntary, or required to obtain or retain a benefit; (2) a description of the entities who must respond; (3) whether the collection is reporting (indicate if a survey), recordkeeping, and/or disclosure; (4) the frequency of the collection (e.g., bi-annual, annual, monthly, weekly, as needed); (5) a description of the information that would be reported, maintained in records, or disclosed; (6) a description of who would receive the information; (7) the purpose of the collection; and (8) if a revision, a description of the revision and the change in burden.


2 Boyle, J., Dienstfrey, S., & Sothoron, A. (1998). National Survey of Speeding and Other Unsafe Driving Actions – Volume I: Methodology (Report No. DOT HS 808 748). National Highway Traffic Safety Administration. https://rosap.ntl.bts.gov/view/dot/1652

Boyle, J., Dienstfrey, S., & Sothoron, A. (1998). National Survey of Speeding and Other Unsafe Driving Actions – Volume II: Driver attitudes and behavior (Report No. DOT HS 808 749). National Highway Traffic Safety Administration. https://rosap.ntl.bts.gov/view/dot/1651

Boyle, J., Dienstfrey, S., & Sothoron, A. (1998). National Survey of Speeding and Other Unsafe Driving Actions – Volume III: Countermeasures (Report No. DOT HS 808 750). National Highway Traffic Safety Administration. https://rosap.ntl.bts.gov/view/dot/1673

3 Royal, D. (2003). National Survey of Speeding and Unsafe Driving Attitudes and Behavior: 2002 - Volume II: Findings (Report No. DOT HS 809 688). National Highway Traffic Safety Administration. https://rosap.ntl.bts.gov/view/dot/1719

Royal, D. (2003). National Survey of Distracted and Drowsy Driving Attitudes and Behavior: 2002. National Survey of Speeding and Unsafe Driving Attitudes and Behavior: 2002 - Volume III: Methods (Report No. DOT HS 809 568). National Highway Traffic Safety Administration. https://one.nhtsa.gov/people/injury/drowsy_driving1/distracted03/VolumeIII03/DD%20Volume%20III%20032803/Default.htm

4 Schroeder, P., Kostyniuk, L., & Mack, M. (2013, December). 2011 National Survey of Speeding Attitudes and Behaviors (Report No. DOT HS 811 865). National Highway Traffic Safety Administration. https://rosap.ntl.bts.gov/view/dot/1960

5 Data will be collected in English and Spanish; those who speak neither language will be excluded.

6 Link, M. W., Battaglia, M. P., Frankel, M. R., Osborn, L., & Mokdad, A. H. (2008). A comparison of address-based sampling (ABS) versus random-digit dialing (RDD) for general population surveys. Public Opinion Quarterly, 72, 6-27.

7 Iannacchione, V. G. (2011). The changing role of address-based sampling in survey research. Public Opinion Quarterly, 75(3), 556-575.

8 Iannacchione, V. G., Staab, J. M., & Redden, D. T. (2003). Evaluating the use of residential mailing addresses in a metropolitan household survey. Public Opinion Quarterly, 67(2), 202-210.

9 Link, M. W., Battaglia, M. P., Frankel, M. R., Osborn, L., & Mokdad, A. H. (2008). A comparison of address-based sampling (ABS) versus random-digit dialing (RDD) for general population surveys. Public Opinion Quarterly, 72, 6-27.

10 Iannacchione, V. G. (2011). The changing role of address-based sampling in survey research. Public Opinion Quarterly, 75(3), 556-575.

11 American Association of Public Opinion Research (AAPOR), Task Force on Address-based Sampling. (2016) Address-based sampling. Retrieved from www.aapor.org/Education-Resources/Reports/Address-based-Sampling.aspx

12 The calculation for the maximum possible error, achieved for an estimated percentage of 50%, is also premised on a design effect of 1.75 due to weighting.

13 Boyle, J., Weiss, A., Schroeder, P., Meyers, M., & Johnson, K. (2013). Reliability of auxiliary data in RDD surveys: NHTSA 2012 Distracted Driving Survey. Presented at the 68th annual conference of the AAPOR, Boston, MA.

14 DiSogra, C., Dennis, J. M., & Fahimi, M. (2010). On the quality of ancillary data available for address-based sampling. In Proceedings of the Survey Research Methods Section of the Joint Statistical Meetings, 417483.

15 Battaglia, M. P., Link, M. W., Frankel, M. R., Osborn, L., & Mokdad, A. H. (2008). An evaluation of respondent selection methods for household mail survey. Public Opinion Quarterly, 72(3), 459–469.

16 Olson, K., Stange, M., & Smyth, J. (2014). Assessing within-household selection methods in household mail surveys. Public Opinion Quarterly, 78(3), 656–678.

17 Boyle, J., Tortora, R., Higgins, B., & Freedner-Maguire, N (2017). Mode effects within the same individual between web and mail administration. AAPOR 72nd annual conference, May 18-21, 2017.

18 MVOSS found that almost all (97%) respondents aged 18 years or older drive a motor vehicle.

19 The U.S. Census Bureau considers this to be households where no residents aged 14 years or older speak English well.

20 Bailly, K., Higgins, B., Freedner-Maguire, N., & Boyle, J. (2017). Impact of pre- and post-incentives on response rates to a web and mail survey using an address-based sample frame. AAPOR 72nd annual conference, May 18-21, 2017.

File Typeapplication/vnd.openxmlformats-officedocument.wordprocessingml.document
Authorrandolph.atkins
File Modified0000-00-00
File Created2021-02-04

© 2024 OMB.report | Privacy Policy