American Community Survey Research and Evaluation Program
May 5, 2021
ACS Research & Evaluation Analysis Plan (REAP)
2021 Strategic Framework
Mail Materials Test
Version |
Date |
Description |
1 |
12/16/20 |
Initial Draft for Feedback |
|
4/27/21 |
Second Draft for Approval |
|
5/5/21 |
Final REAP |
1. INTRODUCTION 1
2. BACKGROUND 1
2.1 Strategic Framework Project 1
2.1.1 Phase 1—Research Best Practices 2
2.1.2 Phase 2—Assess Current Materials 2
2.1.3 Phase 3—Develop New Materials 2
2.1.4 Phase 4—Cognitively Test New Materials 3
2.2 2021 American Community Survey Mail Contact Strategy 3
3. RESEARCH-BASED DEVELOPMENT OF MATERIALS 4
3.1 Themes of Messages in Each Mailing 5
3.1.1 First Mailing – Establish Legitimacy and Trust 5
3.1.2 Second Mailing – Convey Local, Tangible Benefits 5
3.1.3 Third Mailing – Reduce the Sense of Burden of Responding 5
3.1.4 Fourth Mailing – Restate Appeals and Express Gratitude 6
3.1.5 Fifth Mailing – Heighten the Sense of Urgency 6
3.2 Plain Language Principles Used in the Materials 7
3.2.1 Elements that Improve Readability 7
3.2.2 Elements that Improve Visual Appeal 8
3.3 Other Researched-Based Design Elements 8
4. METHODOLOGY AND RESEARCH QUESTIONS 10
4.1 Sample Design 10
4.2 Experimental Design 10
4.2.1 Treatment 1 (Icon Treatment): 11
4.2.2 Treatment 2 (Column and Header Treatment): 11
4.2.3 Treatment 3 (Sidebar Treatment): 11
4.2.4 Treatment 4 (Minimalist Treatment): 11
4.2.5 Treatment 5 (Production, Sorted Separately): 12
4.3 Research Questions 12
4.4 Analysis Metrics 12
4.4.1 Unit Response Analysis 12
4.4.2 Item Response Analysis 14
4.4.3 Cost Analysis 16
4.5 Research Question Analysis 16
4.5.1 Questions Involving Unit-level Response Analysis 16
4.5.2 Questions Involving Item-level Response Analysis 18
4.5.3 Question involving Cost Analysis 18
5. DECISION CRITERIA FOR PICKING A “WINNING” TREATMENT 19
6. ASSUMPTIONS AND LIMITATIONS 20
6.1 Assumptions 20
6.2 Limitations 20
7. TABLE SHELLS 21
7.1 Self-Response Returns Rates 21
7.2 Final Response Rates 22
7.3 Item and Section Nonresponse 22
7.4 Form Completeness 22
7.5 Response Distributions of Respondent Demographics 23
8. POTENTIAL CHANGES TO ACS 23
9. REFERENCES 23
Appendix A. Materials for the Experiment 28
Appendix B. 2021 Mailing Descriptions and Schedule for the 2021 September Production Panel 79
Appendix C. Future Testing Ideas 80
In 2017, the U.S. Census Bureau began the Strategic Framework Project—a long-term, multi-phase project to update the messaging in the American Community Survey (ACS) mail materials. The goals of the project are to improve communication with potential respondents, increase self-response to the survey, reduce program costs, and reduce respondent burden. The project includes research of best practices in messaging to gain survey cooperation, development of new materials based on the research, and testing (qualitative and quantitative) of the new materials.
After conducting research on best practices in communications in a variety of disciplines, the Census Bureau’s Strategic Framework Project team made recommendations for messaging in ACS mail materials in two reports (Oliver, Heimel, and Schreiner, 2017; Schreiner, Oliver, and Poehler, 2020). Following the recommendations, new materials were designed holistically resulting in four sets of materials. The Census Bureau’s Center for Behavioral Science Methods (CBSM) tested the materials in three rounds of cognitive testing. The materials are now ready to be field-tested.
The purpose of the 2021 Strategic Framework Mail Materials Test is to identify which treatment is most successful at increasing self-response. This test will not be able to isolate the effectiveness of individual messages or ideas incorporated into the overall design of any given treatment.
The Census Bureau began the Strategic Framework Project following a 2016 workshop where a panel of experts offered advice on how to improve the ACS mail materials (NAS, 2016). Panelist Nancy Mathiowetz suggested that the Census Bureau develop a strategic plan—grounded in communication theory as well as theories from survey methodology—for messaging in the ACS materials (Mathiowetz, 2016). Mathiowetz thought that having a strategic plan would allow the Census Bureau to judge how expert recommendations fit with the strategic plan. Her suggestion led to the creation of the multi-phase Strategic Framework Project that began in 2017. The five phases of this project are depicted in Figure 1.
Figure 1. Phases of the Strategic Framework Project
The first phase involved research of best practices for messaging to increase survey cooperation. To guide the research, we focused on answering the following questions: (1) What are the demographics of the target audience of the ACS mail materials? (2) What are the best practices in survey messaging to use in the materials to obtain a survey response? (3) What are the best ways to convey those messages? The research and the resulting recommendations for ACS mail messaging are recorded in “Strategic Framework for Messaging in the American Community Survey Mail Materials,” referred to as the Strategic Framework Report in this document (Oliver, Heimel, and Schreiner, 2017).
The second phase of the project involved reviewing and assessing messaging in the ACS mail materials considering the best practices outlined in the Strategic Framework Report. For this phase, we sought to answer the following questions: (1) Do the messages in the current materials meet the recommendations? (2) If so, in what ways? (3) If not, how can the messaging be improved? The findings from the current messaging assessment are recorded in the report, “Assessment of Messaging in the 2018 American Community Survey Mail Contact Materials,” referred to as the Messaging Assessment Report in this document (Schreiner, Oliver, and Poehler, 2020).1
For the third phase, we developed mail materials, incorporating the recommended messaging and design ideas from the first two phases of the project. While the designs of the materials are new, they adhere to the current ACS mail contact strategy (see Section 2.2), which includes the type of mailers (package, pressure seal, or postcard) and the number and timing of mailings that are sent. Each set of materials was designed holistically so that the messaging and look-and-feel within and across the five mailings are interconnected. The Census Bureau designed three sets of updated ACS mail materials and a team of researchers outside of the Census Bureau designed a fourth set.
The fourth phase involved cognitive testing and an expert review of the newly designed materials. The materials were cognitively tested in three iterative rounds by Census Bureau researchers in CBSM (Martinez et al., forthcoming). We also received feedback from a panel of survey methodology experts at a National Academy of Sciences (NAS) meeting in 2019 before the third round of testing. Suggestions for improvements from the NAS meeting experts were incorporated into the materials for the third round of cognitive testing. After the third round, the materials were modified according to suggestions from CBSM researchers. The resulting materials are now ready for field testing.
The current ACS mail contact strategy is detailed below to provide context for the field test. The test materials were designed using the types of mailers and the timing of the mailings in the current strategy. We will compare the test materials to production materials.
The first two mailings are sent to all mailable addresses in the monthly sample. The first mailing is a package that includes a letter, a multilingual brochure, and a card with instructions on how to respond via the internet. The letter contains an invitation to participate in the ACS online and more information in a frequently asked questions format on the back of the letter. A week later, the same addresses are sent a second mailing (reminder letter in a pressure seal mailer).
Responding addresses are removed from the address file after the second mailing to create a new mailing universe of nonrespondents; these addresses are sent the third and fourth mailings.2 The third mailing is a package that includes a letter, a paper questionnaire, and a business reply envelope. Four days later, these addresses are sent a fourth mailing (reminder postcard) which encourages them to respond.
After the fourth mailing, responding addresses are again removed from the address file to create a new mailing universe of nonrespondents. The remaining sample addresses are sent the fifth mailing (a more urgent final reminder letter with a due date in a pressure seal mailer).
Two to three weeks later, responding addresses are removed and the unmailable and undeliverable addresses (from the initial sample) are added to create the universe of addresses eligible for the Computer-Assisted Personal Interview (CAPI) nonresponse followup operation.3 Of this universe, a subsample is chosen to be included in the CAPI operation. Census Bureau field representatives (FR) first attempt to interview those selected for CAPI by phone.4 If the FR is unable to complete a phone interview, they visit the address to conduct an in-person interview.
Additional information can be found in the ACS Design and Methodology Report (U.S. Census Bureau 2014).
The complete literature review for the concepts used to design the mail materials in this test is contained in two reports: “Strategic Framework for Messaging in the American Community Survey Mail Materials” and “Assessment of Messaging in the 2018 American Community Survey Mail Contact Materials (Oliver, Heimel, and Schreiner, 2017; Schreiner, Oliver, and Poehler, 2020).
High-level recommendations from the two reports were to limit the number of messages in each mailing; reduce repetitious messaging; use new appeals; use messages justified by research; and make a clear connection to the well-known Census Bureau brand, in a more prominent way.5 However, the reports also recognized that, for those recommendations to be successful, we had to use plain language writing to lower the reading level of the letters and plain language design principles (such as white space, organization of letter text, graphics, and color) to make the letters easier to read. (See Section 6 of the Messaging Assessment Report for more details.)
While the test materials strongly adhere to the high-level recommendations in general, this section highlights some specific ways those recommendations were implemented.
The Strategic Framework Report (Section 5) recommended using messages designed to resonate with the cynical and distrusting segments of the population, as an increase in their self-response has the most potential to increase overall self-response (Oliver, Heimel, and Schreiner, 2017).6 The report recommended thematic messages for the first four mailings only. The elements for the fifth mailing were developed after we conducted subsequent research to learn about the demographics of ACS nonrespondents receiving the fifth mailing (Berkley, 2018).
The first mailing focuses on building trust with the respondent through messages that legitimize the survey and connect the survey to the Census Bureau, a known and trusted organization. In social exchange theory, building trust is the most important aspect of survey messaging (Groves et al., 2012; Dillman, Smyth, and Christian, 2014). With increased trust, subsequent statements such as the benefits of survey participation are more likely to be believed.
The second mailing focuses on communicating how ACS data has tangible benefits to communities (Reingold, 2014b). This mailing communicates local-level survey benefits because research has shown that prospective respondents are more interested in potential benefits for his or her own neighborhood than for the nation, state, or city (Reingold, 2014b).
This mailing also uses some benefits to show that responding to the survey may directly help other people.7 Some people feel a sense of accomplishment when completing a task for someone else and generally feel a sense of reward when they feel they have helped others. For some, this sense of accomplishment is heightened when the action provides no personal benefit aside from helping someone else (Homans, 1961; Blau, 1964; Dillman, Smyth, and Christian, 2014).
The third mailing focuses on messages that reduce the sense of burden associated with responding to the ACS. The three main burden-reducing messages used are:
Providing a choice in response mode—This mailing reduces the burden for respondents who are unwilling or unable to respond by internet. Offering a choice of response mode can have a positive effect on response (Gentry and Good, 2008; Smyth et al., 2010; Millar and Dillman, 2011; Olson, Smyth, and Wood, 2012).
Explaining that response to the ACS is a normal activity regularly completed by others in the community—Knowing others have responded may help make it more comfortable to respond for those who are hesitant to do so (Cialdini, 1984; Hallsworth et al., 2014).
Linking ACS response to civic duty or responsibility—Some feel a sense of pride as they fulfill their civic obligations and feel a sense of reward when they fulfill a patriotic duty that helps their country (Groves, Singer, and Corning, 2000; Reingold, 2014b).
The fourth mailing primarily summarizes messages from the first three mailings by restating the appeals to trust, benefits, and burden reduction in a different way. We did not want to introduce new concepts in this mailing and overwhelm the recipients of our mailings. So, we repackaged the previous important messages in one mailing using different wording, to avoid repetition (Dillman, Smyth, and Christian, 2014).
This mailing also prominently includes a “thank you” statement which is good communication practice when you ask someone to complete a task. It is natural to thank people for their time and effort.
The fifth mailing is the last opportunity to obtain a self-response through mail contact before the start of the CAPI nonresponse followup operation. Thus, we used the following strategies to heighten the sense of urgency and to make it even easier to self-respond:
The tone of the letter is more formal and urgent than the previous mailings. The opening salutation “An important message from the U.S. Census Bureau:” replaces the salutation used in the previous mailings, “Dear Resident.” The opening sentence “Time is running out.” is short and direct. The letter contains only one paragraph, and then presents the response options—much shorter and to the point than previous mailings.
A due date is prominently displayed three times in the mailing to convey a sense of urgency. Providing a deadline or a due date reduces burden on the respondent by giving clear instructions on when a task is due, which fits into a respondent’s mail prioritization process (Dillman, 2016). A recent ACS mail messaging test showed an increase in self-response when a due date was used on the outside and the inside of the fifth mailing pressure seal mailer (Risley and Oliver, forthcoming in 2021).
A new response option is provided (phone response) to increase the likelihood of a participant being able to respond in their preferred mode. We know that a substantial portion of individuals who have not yet responded have lower reading levels. These individuals may be more likely to respond via a telephone interview. We also know that a good portion of individuals who receive the fifth mailing do not speak English proficiently. So, we also added Spanish text to the bottom of the letter to highlight that respondents have the option to complete the survey by phone in Spanish.
A commitment device is used in this mailing to get recipients to commit to responding. This commitment device asks the potential respondent which response option he or she will use to respond to the survey by the due date. Asking for a potential respondent’s commitment to an action can increase the chance that an action is taken (Milkman et al., 2011; Feygina, Foster, and Hopkins, 2015; Shephard and Bowers, 2016).
In addition to the above recommended messaging themes and strategies, the Messaging Assessment Report also recommended applying plain language principles when developing the new mail materials. Plain language involves wording choices and visual design techniques to make things easier to read.
Section 6.5 of the Messaging Assessment Report outlines the reading level assessment of the current ACS materials. Both the choice of text and the amount of text on each mailing piece caused the materials to rank at high reading levels. The new designs reduce text and use words that are easier to read and understand. The visual design elements mentioned in this section were ideas that developed during the design process through cognitive testing observations, further research, and expert recommendations for improving the materials.
According to the Plain Writing Act of 2010, all government documents issued to the public must be written clearly so that people can find the information they need, understand what they find, and use the information they find to meet their needs. In general, the text and layout in the new materials adhere to the official writing guidelines proposed on plainlanguage.gov. Here are a few other specific ways that we implemented plain language principles in the experimental treatments:
Instruction Card: We redesigned the front of the instruction card in the first mailing with a simpler design and more precise text about how to respond to the survey online. We also changed the font size and style of the user ID (embedded in the address label) to make it easier to locate on the card (Martinez et al., forthcoming).
All mail pieces, where applicable: We eliminated the “https://” from the URL for the online response website (https://respond.census.gov/acs) to ease the burden of typing too many characters. A simplified URL is also more visually pleasing and created more white space.8
Questionnaire: We redesigned the front cover of the questionnaire by reducing text and updating the icons for a cleaner appearance and ease of reading. (See Attachment A for an image of the front cover.)
Letters: We moved the Census Bureau address to the upper right-hand corner, instead of beneath the logo, to allow more space for the body of the letter.
Research has shown that, for government surveys, some color and graphic elements can be used in the letters to catch the eye and draw attention to important information, as long as the letters still appear official and “governmental” (Dillman et al., 1996; Leslie, 1997; Whitcomb and Porter, 2004; Hagedorn, Panek, and Green, 2014; Reingold, 2014a). Where possible, all treatments developed by the Census Bureau incorporated the color of the ACS stateside housing unit questionnaire (“ACS green”) into the redesigned mail pieces.
The treatments use different visual design elements to test which design best resonates with mail recipients. Since we do not know which visual appeal will work the best at increasing overall response, we are experimenting with different designs.
The Icon Treatment uses icons, or symbolic pictures, to quickly convey messages through pictures rather than with text. Icons, which are signs or images that represent the objects that they signify, are known to draw attention and increase readability (Mertz, 2012). Icons could be beneficial for visual learners, readers with low literacy levels, or for those who are not fluent in English.
The Column and Header Treatment organizes text in a way that makes the letters easier to navigate and thus easier to read. Plain language principles suggest using headers, columns, and short sentences and paragraphs.9 For people who prefer scanning text to obtain information quickly, this may be the most beneficial letter design.
The Sidebar Treatment uses a graphic element to draw attention to details not found in the letter. Research has shown that we remember visual images much easier and better than words (Kouyoumdjian, 2012). This design style may appeal to readers who desire more details before responding to the survey.
Within each treatment, the same visual “look and feel” of the design is used throughout the mailings to maintain the cohesive nature of the continuing conversation concept. Research suggests that messages sent across multiple mail contacts, as well as the overall design of graphics, need to look and feel as if they came from the same place and should feel like a continuous conversation (Whitcomb and Porter, 2004; Dillman, Smyth, and Christian, 2014; Hagedorn, Panek, and Green, 2014; Reingold, 2014a).
This section describes some design features that came about through recommendations from the two reports and some design features that came about from ideas that surfaced during the design phase.
Multi-Lingual Brochure: We eliminated the multilingual brochure in the first mailing to reduce the volume of messages as suggested in the Messaging Assessment Report.
The multilingual brochure is written in six languages: English, Spanish, Chinese, Vietnamese, Korean, and Russian. The brochure informs people that they can respond to the ACS via telephone in the languages other than English and Spanish.10 A 2010 study showed that including the brochure in a mailing significantly increased response for the five non-English languages in the brochures (Joshipura, 2010).
In the past five years, an FAQ brochure and an Instruction Booklet and an instruction card have been removed from ACS mailing packages, due to positive results from field tests (Clark, 2015a; Clark, 2015b; Risley and Berkley, forthcoming). Removing the materials reduced costs and reduced respondent burden and did not negatively affect survey response.
Because of past test results with brochures, we feel confident that removing the multilingual brochure will not decrease overall survey response. However, to mitigate the possibility of losing response in non-English languages, we included a sentence in the non-English languages on the back of the Instruction Card sent in the first mailing. The sentence instructs the household that the survey can be answered by phone in Chinese, Vietnamese, Korean, and Russian. A different phone number for each language is given.11 Spanish speakers can also respond to the survey by internet, so the Spanish text on the instruction card mentions responding by internet or calling a number to speak to a representative in Spanish.
Letters: We included Spanish language text on the back of the third letter and the bottom of the fifth letter to remove language barriers that may impede response. One of the recommendations from previous focus group testing of ACS materials was to tailor the materials to acknowledge cultural nuances and make response options readily apparent in Spanish and other languages (Reingold, Inc., 2014b).
In past ACS production materials, the paper questionnaire mailing included more Spanish text to help people respond in Spanish. We have eliminated the instruction card in that mailing that had Spanish text on the back of it. Also, an older design of the front cover of the paper questionnaire included more Spanish text than is currently found on the cover.
We hope that including Spanish on the letters, something new for ACS materials, will increase response for Spanish speakers.
Letters: Wherever text in the mail materials refers to responding online, we added the words “using your computer, smartphone, or tablet.” In some of the materials, we used an icon with the three computing devices as a visual reminder.
One of the researchers at the NAS conference said that his study found that if people start a survey on a smartphone, they are more likely to finish it, whether it be on the same device or on a different one. In short, once you get started you are more likely to finish, and most people have their smartphones with them all the time.
Participants in cognitive testing noticed that they could respond on a smartphone, and they acknowledged that this was helpful and good to know (Martinez et. al, forthcoming).
Envelopes: To increase the likelihood that the recipient would recognize that the letter came from the government and to increase the likelihood of the enveloped being opened, we included the phrase “Official U.S. Government Mail” on the outside of the first and third mailing envelopes. A study mentioned in the Messaging Analysis Report (6.4.2) showed that people first look at their own address on an envelope. Thus, we placed “Official U.S. Government Mail” directly above the recipient’s mailing address, so that it could be readily seen. We were also building on the fact that government-sponsored survey requests receive the highest response rates among all types of surveys (Presser, Blair, and Triplett, 1992).
The 2021 Strategic Framework Mail Materials Test will be conducted using the September 2021 ACS production sample. The monthly ACS production sample consists of approximately 295,000 housing unit addresses and is divided into 24 nationally representative groups (referred to as methods panel groups) of approximately 12,000 addresses each. The sample for each of the four experimental treatments in this test will consist of two randomly assigned methods panel groups (approximately 24,000 mailing addresses per treatment). The sample for the control treatment will also consist of two randomly assigned methods panel groups. The control treatment will receive production ACS materials, but will be sorted and mailed separately from production.12 All remaining methods panel groups will receive production ACS materials.
The treatments will adhere to the current ACS mailing strategy (the number of mailings, types of mailings, and timing of mailings) detailed in Section 2.2 of this report. Images of each mailing piece can be found in Appendix A.
The Strategic Framework Report did not make recommendations of a specific messaging theme for the fifth mailing. [We conducted research for the fifth mailing universe to identify characteristics of nonrespondents with the current mailing strategy, to use messaging appropriate for the target audience; however, those characteristics may change when the first four mailings are changed (Berkley, 2018).13] For this reason, while each treatment employs a distinctive design for the first four mailings, the designs and messaging of the fifth mailing letters vary only slightly among the four experimental treatments.
A brief description of each of the five treatments in this test are provided below.
Treatment 1 uses icons as its distinctive design feature. Icons are symbols used to replace words or to draw attention to key text. The icons break the monotony of text, segment the letter content into different parts, and make the content more interesting to read. Effective use of icons improves content readability. The letters in this treatment are written in a traditional letter format but incorporate icons in the body of the letter. The “ACS green” color, found on the paper questionnaire, is used throughout the five mailings to create visual appeal and to build cohesiveness among the mail pieces and mailings.
Treatment 2 uses columns and headers in green font as distinctive design features. Using columns and headers to segment the content makes it easier to read and navigate to the most important information on the page. The letters in this treatment minimize text and get to the point in a more direct way than a traditional letter format. Like the Icon Treatment, the “ACS green” color is used throughout this treatment.
Treatment 3 uses images as its distinctive design feature. The images are imbedded in a sidebar that is either green or grey in color, depending on the mailing.14 Response option icons are used in some of the letters. The sidebar look is common among flyers and infographics, so it is a recognized look. The sidebar also provides a unique space to add more information about the survey that is not found in the letter. The “ACS green” color is also used throughout this treatment.
This treatment was designed with a minimalist approach, using as few words as possible to convey the most important information needed to respond to the survey. This treatment maintains more of a “governmental” look and feel than the other treatments; no color or graphics are used in the letters. While this treatment does not include the recommended thematic messaging from the Strategic Framework Report, it does employ some plain language principles to make the letters easier to read.
Treatment 5 will have materials identical to production, but the mailings will be sorted separately from production. Previous ACS testing found that smaller volumes of mail arrive later than larger volumes of mail. Thus, we sort the control treatment separately to ensure mail delivery timing consistency with the experimental treatments (Heimel, 2016).
How do the treatments affect self-response to the survey before CAPI?
If a treatment affects self-response before CAPI, how does it affect overall response to the survey?
How do the treatments affect Spanish language self-responses?
How do the treatments affect hard-to-count areas?
How does adding visual design elements and messaging affect self-response to the survey before CAPI, compared to the minimalist approach?
How does the redesigned front cover of the questionnaire affect item nonresponse for the questions on the front cover?
How do the treatments affect overall form completeness?
How do the treatments affect the demographics of early respondents? Late respondents? Overall respondents before CAPI?
How would the treatments affect the costs of data collection if implemented in production ACS?
All self-response analyses, except for the cost analysis, will be weighted using the ACS base sampling weight (the inverse of the probability of selection). Cases in the CAPI subsample will have their weight multiplied by a CAPI subsampling factor unless they are self-responses. The sample size will be able to detect differences of approximately 1.25 percentage points between the self-response return rates of the experimental treatments (with 80 percent power and α=0.1). Detectable differences for the analysis of item-level data (such as item nonresponse rates) vary depending on the item, with housing-level items having minimum detectable differences up to 1.6 percentage points. We will use a significance level of α=0.1 when determining significant differences between treatments. For analysis that involves multiple comparisons, we will adjust for the Type I familywise error rate using the Hochberg method (Hochberg, 1988).
To determine the effect of each treatment on self-response, we will calculate the self-response return rates at selected points in time in the data collection cycle. The selected points in time reflect the dates of additional mailings or the end of the self-response data collection period. An increase in self-response presents a cost savings for each subsequent phase of the mailing process by decreasing the number of mailing pieces that need to be sent out. A significant increase in self-response before CAPI decreases the number of costly interviews that need to be conducted. Calculating the self-response return rates at different points in the data collection cycle gives us an idea of how the experimental treatments would affect operational and mailing costs if they were implemented into a full ACS production year.
To determine whether the experimental changes affect the final self-response and CAPI response by the end of the data collection period, we will calculate final response rates and how each response mode contributes to the total final response.
Self-response return rates will be calculated for total self-response combined and separately for internet, mail, and TQA responses. If no significant differences in TQA rates are detected, we may combine mail and TQA rates.
The return rates will be calculated using the following formula:
Self-Response Return Rate |
= |
Number of mailable and deliverable sample addresses that provided a response by mail, Telephone Questionnaire Assistance (TQA), or internet |
* 100 |
Total number of mailable and deliverable sample addresses13 |
To determine the effect of the experimental treatments on overall response to the survey, we will calculate final overall response rates and how each response mode contributes to the overall final response rate. The final response rates will be calculated using the following formula:
Final Response Rate |
= |
Number of mailable and deliverable sample addresses that provided a response by mail, Telephone Questionnaire Assistance (TQA), internet, or CAPI |
*100 |
Total number of mailable and deliverable sample addresses15 in the universe that were eligible to respond to the survey |
Although the questionnaire cover redesign proposed in this test does not differ substantially from the production design, we would still like to see if the redesign influences response to the questions on the cover.
We will calculate item nonresponse rates for each item on the front cover individually and for all the items combined to determine the effect of the redesigned questionnaire cover on the survey questions that appear on the front cover of the paper questionnaire. We want to determine if the new design of the cover affects response to the questions for Last Name, First Name, Middle Initial, Phone Number, and the number of people living or staying at the address.
The formula for the item nonresponse rate is:
Item Nonresponse Rate |
= |
Number of nonresponses to the questions on the front cover |
*100 |
Number of nonblank mail responses |
While we have no reason to believe that any part of this experiment will affect overall form completeness, and we assume that there will be no effect, we would like to verify this assumption.
To determine the effect of the treatments on the quantity of survey questions answered by each household, we will calculate and compare form completeness rates. Form completeness measures the number of questions on the form that were answered among those that should have been answered, based on questionnaire skip patterns and respondent answers.
We will only calculate and compare form completeness rates for mail and internet responses, because with phone or in-person interviews form completeness can vary depending on the interviewer. For internet responses, the “form” is the internet instrument; for mail responses, the “form” is the paper questionnaire. The following formula will be used:
Overall Form Completeness Rate |
= |
Number of questions answered |
*100 |
Number of questions that should have been answered |
This analysis will not be conducted for treatments that have lower self-response return rates or lower final response rates than the production treatment. Any treatment that performs worse than current production will not be considered as a candidate for replacing production materials. This analysis is only concerned with the success of any treatment that may be used in production to increase self-response with populations that are considered hard-to count.
The letters being tested were designed to have a broad appeal, but they were also designed with a lower reading level and with added Spanish text. Part of the reason these features were used was to gain more self-response from demographic groups that typically respond by personal interview in the CAPI phase of data collection. We hope to convert late CAPI respondents to earlier self-respondents. To see if we were successful in reaching the target audience, we will compare respondent demographics before the third and fifth mailings and before CAPI.
We will calculate and compare the distributions of all non-blank self-responses for the following demographic and housing categories: age, educational attainment, Hispanic origin, race, sex, building type, and tenure.
Proportion estimates will be calculated using the following formula:
Valid in-scope responses will be included in the analysis. The demographic characteristics will be for the respondent, or Person 1, in the survey. We will use uncoded data for the race and Hispanic origin analysis.
This analysis will only use self-responses. In our calculations, we will calculate combined self-response and separate the distributions by mode: mail, internet, and TQA.16 We will use Rao-Scott chi-squared tests of independence to determine whether the response distributions are statistically different at the α=0.1 level (Rao & Scott, 1987). If the distributions are significantly different, we will perform t-tests on the differences for each subcategory.
To control for the overall Type I error rate for a set of hypotheses tested simultaneously, we will perform multiple-comparison procedures using the Hochberg method (Hochberg, 1988). The overall Type I error rate is called the familywise error rate and is the probability of making one or more Type I errors among all hypotheses tested simultaneously. A family for our analysis will be the list of p-values obtained from comparisons of the overall characteristic categories (age, educational attainment, Hispanic origin, race, sex, building type, and tenure). If the response distributions differ significantly for a specific topic, the list of p-values for subcategory comparisons will be used as a family for multiple comparisons.
Excluding the multilingual brochure from the first mailing presents a cost savings. Aside from that savings, we will use estimated workloads to determine any other effects on the cost of implementing any of the experimental treatments into a full production. We will only perform cost analysis for treatments that show significant increases or decreases in self-response before CAPI. We will compare each experimental treatment to Treatment 5 (production materials) before the file creation for the second mailout phase (third and fourth mailings), before the file creation of the third mailout phase (the fifth mailing), and before the creation of the CAPI sample.
In addition to changes in workloads for each mailing, the cost analysis will consider any differences associated with the mailing materials including printing, assembly, and postage costs for potential cost savings if we reduce the mailing workload for the paper questionnaire package mailing. Cost differences associated with CAPI will account for any significant changes in the CAPI workload due to a significant increase or decrease in self-response before CAPI for each experimental treatment compared to production.
RQ1. How do the treatments affect self-response to the survey before CAPI? This analysis will evaluate the effect of each treatment on self-response to the survey. We will calculate and compare self-response return rates of the initial mailing universe for all treatments vs. production materials (Treatment 5). Since an increase in self-response will decrease the cost of subsequent phases of the data collection cycle targeting nonresponders, we will compare self-response return rates just before the third mailing, before the fifth mailing, and before the start of CAPI. We will compare return rates by response mode and overall (modes combined). We will make each comparison using a two-tailed hypothesis test, for a total of four comparisons. Each null hypothesis will be H0: T5 = Ti and each alternative hypothesis will be HA: T5 ≠ Ti
where i = 1, 2, 3, and 4 and T5 = control treatment.
RQ2. If a treatment affects self-response before CAPI, how does it affect overall response to the survey? This analysis will be performed if a treatment is statistically different from the production treatment before the start of CAPI. To evaluate the effect of an experimental treatment on overall response to the survey, we will calculate final overall response rates and how each response mode contributed to the overall final response rate. These rates will be compared with the production treatment. We will make each comparison using a two-tailed hypothesis test, for up to four comparisons. Each null hypothesis will be H0: T5 = Ti and each alternative hypothesis will be HA: T5 ≠ Ti where i = 1, 2, 3, and 4 and T5 = control treatment.
RQ3. How do the treatments affect Spanish language self-response? We will calculate and compare final self-response rates for Spanish self-responses.17 We will calculate a combined rate for all experimental treatments and compare the rate with production using a two-tailed hypothesis test. The null hypothesis will be H0: T5 + Production = T1 + T2 + T3 + T4 and each alternative hypothesis will be HA: T5 + Production ≠ T1 + T2 + T3 + T4.
RQ4. How do the treatments affect areas with hard-to-count populations? We will calculate and compare final self-response rates and final overall response rates by designated high and low response areas. The areas will be defined at the tract level using the low response score (LRS) on the Census Bureau’s planning database.18 The LRS is a modeled variable derived from ACS data for “hard to count” populations and 2010 Census mail response. Some characteristics of “hard to count” populations used to create the LRS are the following: ages 5 and under, ages 18-24, don’t speak English very well, foreign born, renters, low education level, below poverty level, and no internet access.
Defining high and low response areas: A low LRS means a tract is in a high response tract. The 75 percent lowest LRS scores will determine the high response area in our analysis. The remaining tracts will be the low response areas.19
The rates will be compared with the production treatment. We will make each comparison using a two-tailed hypothesis test, for up to four comparisons. Each null hypothesis will be H0: T5 = Ti and each alternative hypothesis will be HA: T5 ≠ Ti where i = 1, 2, 3, and 4 and T5 = control treatment.
RQ5. How does adding visual design elements and strategic messaging affect self-response before CAPI, compared to the minimalist approach? The visual design elements refer to the layout or how things are arranged on a page and the colors and imagery. Strategic messaging refers to the words chosen to convey the thematic messages used in the letters suggested by the Strategic Framework Report. Treatment 4 did not use color, graphics, or any of the recommended thematic messaging from the Strategic Framework Report. However, most other elements found in Treatments 1, 2, and 3 are also in Treatment 4. By comparing the first three treatments separately to the minimalist treatment, we can see how the layout, visual design elements, and strategic messaging of the Icon, Column and Header, and Sidebar Treatments affect self-response as compared to a treatment that does not use the targeted themes, color, or images and only uses minimal wording throughout. We will calculate and compare self-response return rates for treatments 1, 2, and 3 versus treatment 4. We will make each comparison using a two-tailed hypothesis test, for a total of three comparisons. Each null hypothesis will be H0: T4 = Ti and each alternative hypothesis will be HA: T4 ≠ Ti where i = 1, 2, and 3 and T4 = Minimalist Treatment.
Item-level response analysis is for occupied housing units only.
RQ6. How does the redesigned front cover of the questionnaire affect item nonresponse for the questions on the front cover? We will calculate and compare item nonresponse rates for each question individually and for the front cover overall. We will make each comparison using a two-tailed hypothesis test, for a total of four comparisons. Each null hypothesis will be
H0: T5 = Ti and each alternative hypothesis will be HA: T5 ≠ Ti where i = 1, 2, 3, and 4 and
T5 = control treatment. The universe for calculations will be all mail responses to the survey.
RQ7. How do the treatments affect overall form completeness? We will calculate and compare form completeness rates. We will make each comparison using a two-tailed hypothesis test, for a total of four comparisons. Each null hypothesis will be H0: T5 = Ti and each alternative hypothesis will be HA: T5 ≠ Ti where i = 1, 2, 3, and 4 and T5 = control treatment.
The universe for calculations will be all internet and mail responses.
RQ8. How do the treatments affect the demographics of early respondents? Late respondents? Overall respondents before CAPI? We will calculate and compare distributions of responses to questions about the following demographic and housing categories: age, educational attainment, Hispanic origin, race, sex, building type, and tenure. We will compare each treatment to the production treatment (T1, T2, T3, T4 vs. T5). For early respondents we will use internet, mail, and TQA responses received before the third mailing is sent. For late respondents we will use internet, mail, and TQA responses received before the fifth mailing is sent. For the overall calculations, we will use all self-responses received before the CAPI sample is created.
RQ9. How would the treatments affect the costs of data collection? Apart from the cost savings from the removal of the multilingual brochure, we will only assess cost impacts on treatments that perform better than the production treatment. To assess impacts on costs we will calculate the annual expected cost of implementing each experimental treatment into a full ACS production year and compare it to the production costs. A confidence interval for the cost that accounts for sampling error in the workload estimates will also be calculated. Cost differences will be calculated as described in Section 4.4.3.
An abbreviated version of the cost analysis will appear in the final report. We will prepare a separate, more detailed, cost report for internal use, if necessary.
Each treatment in this test exhibits clear benefits when compared to current production materials. Those benefits are: (1) reduction in costs from eliminating the multilingual brochure and (2) reduction in respondent burden by improving the visual appeal and readability of the letters. Having said that, the goal of this test is to choose one experimental treatment as the best among all treatments. To do that we will adhere to the following decision criteria, listed in Table 1 by order of importance.
Table 1. Decision Criteria for “Winning” Treatment
Priority Order |
Decision Criteria |
1 |
Self-Response Return Rates |
2 |
Response in Low Response vs High Response Areas |
3 |
Response Distributions |
4 |
Form Completeness Rates |
5 |
Final Response Rates |
As stated in the introduction, the purpose of this field test is to identify which experimental treatment is most successful at increasing self-response when compared to the control treatment (production materials). Self-response is the most important metric, because increasing self-response reduces data collection costs. We will look at self-response throughout the data collection period. The earlier we receive self-response, the more money we save. If we do not see a “winner” in the self-response category, or if some of the treatments are not statistically different, we will continue with the list of decision criteria to determine a “winner”.
We have also stated that a very important goal of the new materials is to convert households with characteristics of those that typically respond in CAPI to be households that self-respond. We have implemented many new features in the materials that we hope will reach this goal. We expect this strategy to increase overall self-response, but if it does not, the next best metric to look at will be response in Low Response Areas. If we don’t see any differences in this category, we will look for changes in the demographic makeup of early responders, which we hope to see with our response distribution analysis.
Not that we will consider magnitude of differences as we make comparisons. For example, if Treatment 1 performs nominally better than Treatment 2 in self-response but significantly worse than Treatment 2 in response in low response areas, we will take that into consideration.
We hope to come to a decision without having to use #4 and #5 on the list. These are extremely low priority. We don’t expect these categories to be significantly different for any of the treatments, but if two treatments are tied for #1 - #3 and one of them performs extremely well for #4 and #5, we will take that into consideration when making a decision on the winning treatment.
If, using the above criteria, a single treatment cannot be identified as the “best” among all treatments, then additional field testing will be conducted. The additional field testing may include a subset of treatments, updated materials, or a combination of the current treatments. (See Appendix C for some ideas for future testing.)
A single ACS monthly sample is representative of an entire year (twelve panels) and the entire frame sample, with respect to both response rates and cost, as designed.
A single methods panel group (1/24 of the full monthly sample) is representative of the full monthly sample, as designed.
We assume that there is no difference between treatments in mail delivery timing or subsequent response time. The treatments had the same sample size and used the same postal sort and mailout procedures. Previous research indicated that postal procedures alone could cause a difference in response rates at a given point in time between experimental treatments of different sizes, with response for the smaller treatments lagging (Heimel, 2016).
Group quarters and sample housing unit addresses from remote Alaska and Puerto Rico are not included in the sample for the test. Any conclusions from the test can only be made for housing units like those in the test sample.
The cost analysis uses estimates to make cost projections. These estimates do not account for monthly variability in production costs such as changes in staffing, production rates, or printing price adjustments.
Each treatment was designed holistically and, as such, if differences in response are detected we will not be able to identify the specific elements in each treatment that caused differences to occur.
Table 1. Sample Table for Total Self-Response Return Rates: Comparison of Treatment vs. Control
Point in Data Collection Cycle |
Treatment |
Control |
Difference |
P-Value |
Before the Third Mailing All Self-response Modes |
NN.N (N.N) |
NN.N (N.N) |
N.N (N.N) |
N.NN |
Before the Fifth Mailing All Self-response Modes |
NN.N (N.N) |
NN.N (N.N) |
N.N (N.N) |
N.NN |
Before CAPI All Self-response Modes |
NN.N (N.N) |
NN.N (N.N) |
N.N (N.N) |
N.NN |
Source: U.S. Census Bureau, American Community Survey, <Insert Test Name>; DRB Approval Number: <insert approval #>
Note: Minor additive discrepancies are due to rounding. Standard errors are in parentheses. An asterisk (*) indicates a statistically significant result. Significance was tested based on a two tailed t-test at the α=0.1 level.
Table 2. Sample Table for Mail and TQA Return Rates: Comparison of Treatment vs. Control
Point in Data Collection Cycle |
Treatment |
Control |
Difference |
P-Value |
Before the Third Mailing All Self-response Modes |
NN.N (N.N) |
NN.N (N.N) |
N.N (N.N) |
N.NN |
Before the Fifth Mailing All Self-response Modes |
NN.N (N.N) |
NN.N (N.N) |
N.N (N.N) |
N.NN |
Before CAPI All Self-response Modes |
NN.N (N.N) |
NN.N (N.N) |
N.N (N.N) |
N.NN |
Source: U.S. Census Bureau, American Community Survey, <Insert Test Name>; DRB Approval Number: <insert approval #>
Note: Minor additive discrepancies are due to rounding. Standard errors are in parentheses. An asterisk (*) indicates a statistically significant result. Significance was tested based on a two tailed t-test at the α=0.1 level.
NOTE 1: If significant differences in TQA rates are detected we will report Mail and TQA rates separately.
NOTE 2: There is hidden text (in white 1-point font) in the tables that were there for 508 compliance when the tables by mode were combined into one table. Fine to leave it for now, but that’s what that strange underline is.
Table 3. Sample Table for Internet Return Rates: Comparison of Treatment vs. Control
Point in Data Collection Cycle |
Treatment |
Control |
Difference |
P-Value |
Before the Third Mailing All Self-response Modes |
NN.N (N.N) |
NN.N (N.N) |
N.N (N.N) |
N.NN |
Before the Fifth Mailing All Self-response Modes |
NN.N (N.N) |
NN.N (N.N) |
N.N (N.N) |
N.NN |
Before CAPI All Self-response Modes |
NN.N (N.N) |
NN.N (N.N) |
N.N (N.N) |
N.NN |
Source: U.S. Census Bureau, American Community Survey, <Insert Test Name>; DRB Approval Number: <insert approval #>
Note: Minor additive discrepancies are due to rounding. Standard errors are in parentheses. An asterisk (*) indicates a statistically significant result. Significance was tested based on a two tailed t-test at the α=0.1 level.
Table 4. Sample Table for Final Response Rates and Response Distributions by Mode
Response Mode |
Treatment |
Control |
Difference |
P-Value |
Overall Self-Response |
NN.N (N.N) |
NN.N (N.N) |
N.N (N.N) |
N.NN |
Internet |
NN.N (N.N) |
NN.N (N.N) |
N.N (N.N) |
N.NN |
Mail and TQA |
NN.N (N.N) |
NN.N (N.N) |
N.N (N.N) |
N.NN |
CAPI |
NN.N (N.N) |
NN.N (N.N) |
N.N (N.N) |
N.NN |
Source: U.S. Census Bureau, American Community Survey, <Insert Test Name>; DRB Approval Number: <insert approval #>
Note: Minor additive discrepancies are due to rounding. Standard errors are in parentheses. An asterisk (*) indicates a statistically significant result. Significance was tested based on a two tailed t-test at the α=0.1 level.
Table 6. Sample Table for Final Overall Response Rates by Designated High and Low Response Areas
Response Area |
Test |
Control |
Test Minus Control |
P-Value |
High Response Area |
|
|
|
|
Low Response Area |
|
|
|
|
Difference |
|
|
|
|
Source: U.S. Census Bureau, American Community Survey, <Insert Test Name>; DRB Approval Number: <insert approval #>
Note: Minor additive discrepancies are due to rounding. Standard errors are in parentheses. An asterisk (*) indicates a statistically significant result. Significance was tested based on a two tailed t-test at the α=0.1 level.
Table 7. Sample Table for Item and Section Nonresponse Rates for Mail Responses –
Treatment X vs. Treatment Y
Item |
Treatment X |
Treatment Y |
Difference |
P-Value |
Section Nonresponse |
NN.N (N.N) |
NN.N (N.N) |
N.N (N.N) |
N.NN |
Name |
NN.N (N.N) |
NN.N (N.N) |
N.N (N.N) |
N.NN |
Telephone Number |
NN.N (N.N) |
NN.N (N.N) |
N.N (N.N) |
N.NN |
Number of Persons in Household |
NN.N (N.N) |
NN.N (N.N) |
N.N (N.N) |
N.NN |
Source: U.S. Census Bureau, American Community Survey, <Insert Test Name>; DRB Approval Number: <insert approval #>
Note: Minor additive discrepancies are due to rounding. Standard errors are in parentheses. An asterisk (*) indicates a statistically significant result. Significance was tested based on a two tailed t-test at the α=0.1 level.
Table 8. Sample Table for Form Completeness by Response Mode, Treatment vs. Control
Response Mode |
Treatment |
Control |
Difference |
P-Value |
All Self-Response |
NN.N (N.N) |
NN.N (N.N) |
N.N (N.N) |
N.NN |
NN.N (N.N) |
NN.N (N.N) |
N.N (N.N) |
N.NN |
|
Internet |
NN.N (N.N) |
NN.N (N.N) |
N.N (N.N) |
N.NN |
Source: U.S. Census Bureau, American Community Survey, <Insert Test Name>; DRB Approval Number: <insert approval #>
Note: Minor additive discrepancies are due to rounding. Standard errors are in parentheses. An asterisk (*) indicates a statistically significant result. Significance was tested based on a two tailed t-test at the α=0.1 level.
Table 9. Sample Table for Response Distributions for <insert characteristic> by Response Mode and Treatment, Treatment vs. Control
Characteristic Group and Response Mode |
Treatment |
Control |
P-Value |
Total Self-Response |
100.0 |
100.0 |
N.NN |
All Modes: Category 1 |
NN.N (N.N) |
NN.N (N.N) |
|
All Modes: Category 2 |
NN.N (N.N) |
NN.N (N.N) |
|
All Modes: Category 3 |
NN.N (N.N) |
NN.N (N.N) |
|
100.0 |
100.0 |
N.NN |
|
Mail: Category 1 |
NN.N (N.N) |
NN.N (N.N) |
|
Mail: Category 2 |
NN.N (N.N) |
NN.N (N.N) |
|
Mail: Category 3 |
NN.N (N.N) |
NN.N (N.N) |
|
Internet |
100.0 |
100.0 |
N.NN |
Internet and TQA: Category 1 |
NN.N (N.N) |
NN.N (N.N) |
|
Internet and TQA: Category 2 |
NN.N (N.N) |
NN.N (N.N) |
|
Internet and TQA: Category 3 |
NN.N (N.N) |
NN.N (N.N) |
|
Source: U.S. Census Bureau, American Community Survey, <Insert Test Name>; DRB Approval Number: <insert approval #>
Note: Minor additive discrepancies are due to rounding. Standard errors are in parentheses. An asterisk (*) indicates a statistically significant result. Statistical significance was determined by comparing distributions using a Rao-Scott chi-square test at the α=0.1 level.
Note: To be 508 Compliant the Categories in this table are preceded by words not seen, in white text. Before each category group you will have to type “All Modes”, “Mail”, “Internet”, etc.
Based on the results of this research, the Census Bureau may consider revising the ACS production mail materials.
Berkley, J. (2018). Respondent Demographics Presentation and Report. Internal U.S. Census Bureau Report: unpublished.
Blau, P. (1964). Exchange Power in Social Life. New York, NY: Wiley.
Cialdini, R. (1984) Influence: The New Psychology of Modern Persuasion. New York, NY: Quill.
Clark, S. (2015a). “2015 Replacement Mail Questionnaire Package Test,” 2015 American Community Survey Research and Evaluation Report Memorandum Series #ACS15-RER-18. Retrieved on May 12, 2020 from https://www.census.gov/library/working‑papers/2015/acs/2015_Clark_02.html
Clark, S. (2015b). “2015 Mail Contact Strategy Modification Test,” 2015 American Community Survey Research and Evaluation Report Memorandum Series #ACS15-RER-19. Retrieved on May 12, 2020 from https://www.census.gov/library/working‑papers/2015/acs/2015_Clark_03.html
Dillman, D., Jenkins, C., Martin, B. and DeMaio, T. (1996). “Cognitive and Motivational Properties of Three Proposed Decennial Census Forms.” Center for Survey Methods Research Memorandum Series #2006-4, U.S. Bureau of the Census, Washington, D.C. Retrieved on May 27, 2020 from https://www.census.gov/srd/papers/pdf/ssm2006‑04.pdf
Dillman, D. (2016). “Part 2: What Theory Can Tell Us About Why People Do or Do Not Answer Survey Requests.” U.S. Census Bureau Summer at Census Workshop, July 11, 2016. Suitland, MD.
Dillman, D., Smyth, J., and Christian, L. (2014). Internet, Phone, Mail, and Mixed-Mode Surveys: The Tailored Design Method (4th ed.). Hoboken, NJ: Wiley.
Dillman, D. and Greenberg, P. (2017). “Making Mail Communications More Effective: A test of Presuasion vs. Social Exchange Theories.” Presented to the 7th Conference of the European Survey Research Association, July 21, 2017. Lisbon, Portugal.
Feygina, I., Foster, L., and Hopkins, D. (2015). “Re: Potential Pilot Interventions to Increase Response Rates to Census Surveys.” Received by Ted Johnson, September 10, 2015.
Fishbein, M. and Ajzen, I. (2011). Predicting and Changing Behavior: The Reasoned Action Approach. Oxford, United Kingdom: Taylor & Francis.
Gentry, R. and Good, C. (2008). “Offering Respondents a Choice of Survey Mode: Use Patterns of an Internet Response Option in a Mail Survey”. Presented to the Annual Conference of The American Association for Public Opinion Research, May 5, 2008. New Orleans, LA.
Groves, R., Singer E., and Corning, A. (2000). “Leverage-Saliency Theory of Survey Participation.” Public Opinion Quarterly, 64(3): 209–308.
Groves, R., Presser, S., Tourangeau, R., West, B.T., Couper, M.P., Singer, E. and Toppe, C. (2012). “Support for the Survey Sponsor and Nonresponse Bias.” Public Opinion Quarterly, 76(3): 512–524.
Hagedorn, S., Panek, M., and Green, R. (2014). “American Community Survey Mail Package Research: Online Visual Testing.” Washington, D.C.: U.S. Census Bureau. Retrieved on May 12, 2020 from https://www.census.gov/content/dam/Census/library/working-papers/2014/acs/2014_Hagedorn_04.pdf
Hallsworth, M., List, J. A., Metcalf, R. D., and Vlaev, I. (2014). “The Behavioralist as Tax Collector: Using Natural Field Experiments to Enhance Tax Compliance.” NBER Working Paper No. 20007.
Heimel, S. (2016). “Postal Tracking Research on the May 2015 ACS Panel.” 2016 American Community Survey Research and Evaluation Report Memorandum Series #ACS16-RER-01, April 1, 2016.
Hochberg, Y. (1988). “A Sharper Bonferroni Procedure for Multiple Tests of Significance,” Biometrika, 75 (4), 800-802. Retrieved on July 23, 2020 from http://www.jstor.org/stable/2336325?seq=1#page_scan_tab_contents
Homans, C. G. (1961). Social Behavior: Its Elementary Forms. New York, NY: Harcourt, Brace & World.
Joshipura, M. (2010). “Evaluating the Effects of a Multilingual Brochure in the American Community Survey.” Washington, D.C.: U.S. Census Bureau. Retreived on May 26, 2020 from https://www.census.gov/content/dam/Census/library/working-papers/2010/acs/2010_Joshipura_01.pdf
Kouyoumdjian, Haig. (2012) “Learning Through Visuals.” Retrieved on June 4, 2020 at https://www.psychologytoday.com/us/blog/get-psyched/201207/learning-through-visuals
Leslie, T.F. (1997) “Comparing Two Approaches to Questionnaire Design: Official Government versus Public Information Design.” Proceedings of the Survey Research Section of the American Statistical Association. Retrieved on May 27, 2020 from http://citeseerx.ist.psu.edu/viewdoc/download;jsessionid=C5DDD16163DDF8AB4A36DB7B236E588B?doi=10.1.1.631.6688&rep=rep1&type=pdf
Martinez, M., Katz, J., Luck, J., and Kephart, K. (forthcoming) “Cognitive Testing of 2020 American Community Survey Strategic Framework Mailing Materials.” Washington, D.C.: U.S. Census Bureau.
Mathiowetz, Nancy (2016). Post meeting thoughts. Received by Todd Hughes, June 3, 2016.
Mertz, Nikolaj. (2012) “How and Why Icons Improve Your Web Design.” Retrieved on June 4, 2020 from https://usabilla.com/blog/how-and-why-icons-improve-you-web-design/
Milkman, K. L., Beshears, J., Choi, J. J., Laibson, D., and Madrian, B. (2011). “Using Implementation Intentions Prompts to Enhance Influenza Vaccination Rates.” Proceedings of the National Academy of Sciences, 108(26): 10415–10420.
Millar, M. and Dillman, D. (2011). “Improving Response to Web and Mixed-Mode Surveys.” Public Opinion Quarterly, 75(2): 249–269.
National Academies of Sciences. (2016). Reducing Response Burden in the American Community Survey: Proceedings of a Workshop. Washington, DC: The National Academies Press. Retrieved on May 6, 2020 from https://doi.org/10.17226/23639
National Academies of Sciences. (2019). Improving the American Community Survey: Proceedings of a Workshop. Washington, DC: The National Academies Press. Retrieved on May 12, 2020 from https://doi.org/10.17226/25387
Olson, K., Smyth, J., and Wood, H. (2012). “Does Giving People their Preferred Survey Mode Actually Increase Survey Participation?” Public Opinion Quarterly, 74(4), 611–635.
Oliver, B., Heimel, S., and Schreiner, J. (2017). “Strategic Framework for Messaging in the American Community Survey Mail Materials.” Washington, DC: U.S. Census Bureau. Retrieved on May 6, 2020 from https://www.census.gov/content/dam/Census/library/working-papers/2017/acs/2017_Oliver_01.pdf
Plain Writing Act of 2010. (2010). Pub. L. No. 111-274, Stat 2861, 2862, and 2863.
Presser, S., Blair, J. and Triplett, T. (1992). “Survey Sponsorship, Response Rates, and Response
Effects.” Social Science Quarterly, 73(3), 699-702.
Reingold, Inc. (2014a). “ACS Mail Package Research: Focus Groups and Interviews.” 2014 American Community Survey Research And Evaluation Report Memorandum Series #ACS 14-RER-25.
Reingold, Inc. (2014b). “ACS Messaging Research: Cumulative Findings.” Retrieved on May 11, 2020 from https://www.census.gov/content/dam/Census/library/working‑papers/2014/acs/2014_Walker_01.pdf
Risley, M, and Berkley, J. (2020). “2018 ACS Mail Materials Test,” 2021 American Community Survey Research and Evaluation Report Memorandum Series #ACS21-RER-01. Retrieved on January 25, 2021 from https://www.census.gov/library/working‑papers/2020/acs/2021_Risley_01.html
Risley, M, and Oliver, B. (forthcoming). “2019 ACS Due Dates Test,” 2021 American Community Survey Research and Evaluation Report Memorandum Series #ACS41-RER-##.
Schreiner, J., Oliver, B., and Poehler, E. (2020). “Assessment of Messaging in the 2018 American Community Survey Mail Contact Materials.” Washington, DC: U.S. Census Bureau. Retrieved on May 6, 2020 from https://www.census.gov/content/dam/Census/library/working‑papers/2020/acs/2020_Schreiner_01.pdf
Shephard, D. and Bowers, J. (2016). Suggestions for ACS 2016 Softened Revised Design (v1). Received by Ted Johnson, May 6, 2016.
Smyth, J., Dillman, D., Christian, L., and O’Neill, A. (2010). “Using the Internet to Survey Small Towns and Communities: Limitations and Possibilities in the Early 21st Century.” American Behavioral Scientist, 53: 1423–1448.
U.S. Census Bureau (2014). “American Community Survey Design and Methodology,” Retrieved on May 11, 2020 from https://www.census.gov/programs-surveys/acs/methodology/design-and-methodology.htmlWhitcomb, M. and Porter, S. (2004). “E-mail Contacts: A Test of Complex Graphical Designs in Survey Research.” Social Science Computer Review, 22(3), 370–376.
Materials for the Experiment
Materials in Each Mailing
First Mailing
|
Letter: ACS-13(L) Instruction Card: ACS-34IM (same for all experimental treatments) Envelope: ACS-46IM (same for all experimental treatments) Multilingual Brochure: ACS-9 (production only) |
Second Mailing |
Pressure Seal Mailer: ACS-20 |
Third Mailing |
Letter: ACS-14(L) (back of letter same for all experimental treatments) Return Envelope: 6385-47 (same for all experimental treatments) Questionnaire: ACS-1 (same for all experimental treatments) Envelope: ACS-46 (same for all experimental treatments) |
Fourth Mailing |
Postcard: ACS-29 |
Fifth Mailing |
Pressure Seal Mailer: ACS-23 |
Icon Treatment: ACS-13(L) (Front of Letter)
Icon and Sidebar Treatments: ACS-13(L) (Back of Letter)
All Treatments: ACS-34IM (Front and Back of Instruction Card)
Translation of languages below:
Spanish: HOW TO RESPOND: Go to respond.census.gov/acs to complete the survey in Spanish by computer, smartphone, or laptop. Need help or have questions? Call 1-877-833-5625.
Chinese (simplified), Vietnamese, Russian, and Korean: If you have questions about the survey call [number] to speak to one of our employees who speaks [language]. The employee can answer your questions, or you can complete the survey over the phone.
All Treatments: ACS-46IM (Initial Mailing Envelope)
Icon Treatment: ACS-20 (Inside of Mailer)
All Treatments: ACS-20 (Outside of Mailer) [The only difference between all treatments is the form number found in the lower right-hand corner.]
Icon Treatment: ACS-14(L) (Front of Letter)
All Treatments: ACS-14(L) (Back of Letter)
All Treatments: 6385-47 (Front of Postage Paid Return Envelope for Questionnaire)
All Treatments: ACS-1 (Front Cover of Questionnaire)
All Treatments: ACS-1 (Back Cover of Questionnaire)
All Treatments: ACS-46(EX) (Front of Envelope – Paper Questionnaire Mailing)
All Treatments: ACS-46(EX) (Back of Envelope - Paper Questionnaire Mailing)
Icon Treatment: ACS-29 (Front and Back of Postcard)
Icon Treatment: ACS-23 (Inside Pressure Seal Mailer) [The sample variable data: the date, the phrase “Wednesday, September 30, 2020?”, and the user ID.]
All Treatments: ACS-23 (Outside Pressure Seal Mailer) [The only difference between all treatments is the form number found in the lower right-hand corner. Sample variable data: the due date.]
Sidebar Treatment: ACS-13(L) (Front of Letter)
Sidebar Treatment: ACS-20 (Inside of Pressure Seal Mailer)
Sidebar Treatment: ACS-14(L) (Front of Letter)
Sidebar Treatment: ACS-29 (Front and Back of Postcard)
Sidebar Treatment: ACS-23 (Inside Pressure Seal Mailer) [The sample variable data: the date, the phrase “Wednesday, September 30, 2020?”, and the user ID.]
Column and Header Treatment: ACS-13(L) (Front of Letter)
C olumn and Header Treatment: ACS-13(L) (Back of Letter)
Column and Header Treatment: ACS-20 (Inside of Pressure Seal Mailer)
Column and Header Treatment: ACS-14(L) (Front of Letter)
Column and Header Treatment: ACS-29 (Front and Back of Postcard)
Column and Header Treatment: ACS-23 (Inside of Pressure Seal Mailer) [The sample variable data: the date, the phrase “Wednesday, September 30, 2020?”, and the user ID.]
Minimalist Treatment: ACS-13(L) (Front of Letter)
M inimalist Treatment: ACS-13(L) (Back of Letter)
Minimalist Treatment: ACS-20 (Inside of Pressure Seal Mailer)
Minimalist Treatment: ACS-14(L) (Front of Letter)
Minimalist Treatment: ACS-29 (Front and Back of Postcard)
Mailing
address printed here
Minimalist Treatment: ACS-23 (Inside of Pressure Seal Mailer) [The sample variable data: the date, the phrase “Wednesday, September 30, 2020?”, and the user ID.]
Production: ACS-13(L) (Front of Letter)
Production: ACS-13(L) (Back of Letter)
Production: ACS-34IM (Front and Back of Instruction Card)
Production: ACS-46IM (Front and Back of Envelope)
Production: ACS-9 (Multilingual Brochure)
Production: ACS-20 (Inside of Pressure Seal Mailer)
Production: ACS-20 (Outside of Pressure Seal Mailer)
P roduction: ACS-14(L) (Front of Letter)
Production: ACS-14(L) (Back of Letter)
Production: 6385-47 (Front of Postage Paid Return Envelope for the Questionniare)
Production: ACS-1 (Front Page of the Questionnaire)
Production: ACS-1 (Back Page of the Questionnaire)
Production: ACS-46 (Front of the Envelope)
Production: ACS-46 (Back of the Envelope)
Production: ACS-29 (Front and Back of Postcard)
Production: ACS-23 (Inside of Pressure Seal Mailer) [Sample variable data: The due date, the user ID, and the sentence “Respond by <insert date> to be removed from our schedule for a visit.”]
Production: ACS-23 (Outside of Pressure Seal Mailer) [Sample variable data: the date]
Mailing
address printed here
2021 Mailing Descriptions and Schedule for the 2021 September Production Panel
Mailing |
Description of Materials |
Mailout Date |
Initial Mailing Package |
A package of materials containing the following: Introduction Letter, Frequently Asked Questions (FAQ) Brochure, Multi-Lingual Informational Brochure, and Internet Instruction Card. This mailing urges housing units to respond via the internet. |
8/26/2021 |
Reminder Letter |
A reminder letter sent to all addresses that were sent the Initial Mailing Package, reiterating the request to respond. [Pressure seal mailer] |
9/2/2021 |
Paper Questionnaire Package |
A package of materials sent to addresses that have not responded. Contains the following: Introduction Letter, Paper Questionnaire, Return Envelope, Internet Instruction Card, and FAQ Brochure. |
9/16/2021
|
Reminder Postcard |
A reminder postcard sent to all addresses that were also sent the Paper Questionnaire Package, reiterating the request to respond. |
9/20/2021
|
Additional Postcard |
An additional reminder postcard sent to addresses that have not yet responded and are ineligible for telephone follow-up. [Pressure seal mailer] |
10/13/2021
|
CAPI Letter† |
A letter sent to all mailable addresses in the CAPI sample that includes an internet user ID. This letter informs households that a Census Bureau interviewer will contact them soon if they don’t respond to the survey. |
11/3/2021 |
†NPC mails the CAPI Internet Letter on a flow basis. The 'Mailout Date' is the date all CAPI Internet Letters have been mailed from NPC.
Future Testing Ideas
Some ideas we have for possible future testing are listed below.
Testing a due date in the second mailing (“Responding by __________ avoids additional mailings.”)
Using a date on the letters
Testing different types of benefit messages
Testing to understand the effect of specific messaging or design elements of the winning treatment
Test ideas for the fifth mailing:
Different messaging on the outside of the pressure seal mailer
Pressure seal mailer with tailored messaging targeted for specific areas
An infographic type letter in place of a pressure seal mailer
A letter with a card stock graphical insert in place of a pressure seal mailer
Bilingual mailer or letter for targeted geographical areas
1 The 2018 production materials were used for the assessment. For this test, the experimental treatments will be tested against new materials that went into production starting in 2020.
2 Addresses deemed “undeliverable as addressed” (UAA) by the United States Postal Service are also removed from the address files for subsequent mailings.
3 CAPI interviews start at the beginning of the month following the fifth mailing.
4 A pressure seal reminder letter, which includes a user ID needed for an internet response, is also sent to all mailable addresses sampled in CAPI at the start of the interviewing month. This began in October of 2020.
5 The recommendations were at the forefront of the design process for all test treatments materials, except the Minimalist Treatment. This treatment was designed to keep text to a minimum while maintaining a “governmental feel” to the materials.
6 One of the treatments, the Minimalist Treatment, does not use the recommended themes for each mailing.
7 For example, one benefit mentioned is “services to help the elderly, veterans, and the disabled.”
8 Participants in cognitive testing were able to successfully navigate to the ACS response landing page using the abbreviated URL.
9 See plainlanguage.gov for more ideas on visual design and text layout to improve readability.
10 In English and Spanish, a number is given to call if a respondent has questions about the survey, but the text does not directly state that the survey can be completed over the phone.
11 We acknowledge the risk we are taking with respect to survey response in non-English languages, and we will monitor the situation, possibly isolating the experimental change in a future test. We may, however, see an increase in response since the text is easier to see on the instruction card than inside of the trifold brochure.
12 See Appendix B for dates of the mailout schedule for the September 2021 panel.
13 We are hopeful that the results from this test will give us insight into the best design to use for the first four mailings and insight into the likely (perhaps new) target audience for the fifth mailing.
14 The pressure seal mailers are only printed with black and grey ink.
15 We will remove addresses deemed to be Undeliverable as Addressed by the U.S. Postal Service if no response is received.
16 We may combine mail and TQA responses if there is no significant difference in TQA responses.
17 We do not expect there to be enough phone responses in Chinese, Korean, and Russian to detect significant differences, but we will monitor the number of responses in those languages to see if the text on the new Instruction Card appears to have an effect on response in those languages.
18 The Census Bureau Planning Database can be found at https://www.census.gov/topics/research/guidance/planning-databases.html
19 For the ACS tracts that do not appear on the Census Planning Database, we will create county level LRS scores based on the weighted averages of the low response scores at the tract-levels.
File Type | application/vnd.openxmlformats-officedocument.wordprocessingml.document |
Subject | American Community Survey |
Author | U.S. Census Bureau |
File Modified | 0000-00-00 |
File Created | 2021-06-25 |