Alumni Supporting Statement B_04_02_2018Final

Alumni Supporting Statement B_04_02_2018Final.docx

Alumni Outcomes Survey

OMB: 3045-0171

Document [docx]
Download: docx | pdf

AMERICORPS ALUMNI OUTCOMES SURVEY

Shape1

SUPPORTING STATEMENT FOR PAPERWORK REDUCTION ACT SUBMISSIONS

B. COLLECTIONS OF INFORMATION EMPLOYING STATISTICAL METHODS

B1. Describe (including numerical estimate) the potential respondent universe and any sampling or other respondent selection method to be used. Data on the number of entities (e.g., establishments, State and local government units, households, or persons) in the universe covered by the collection and in the corresponding sample are to be provided in tabular form for the universe as a whole and for each of the strata in the proposed sample. Indicate expected response rates for the collection as a whole. If the collection had been conducted previously, include the actual response rate achieved during the last collection.

There is no current plan for sampling as this project currently being designed. The potential project will replicate the process and analysis that was employed for the 2015 project. Sample estimates and burden are based on highly stable rates and figures from the previous study. The following methodology and description from the previous study will be replicated however the time frame for each new cohort and strata will be between 2008-2010, 2013-2015, 2016-2018:

The potential respondent universe for the AmeriCorps Alumni Outcomes Survey consists of alumni of AmeriCorps programs who meet the following criteria: Sample size calculation are based on members whose last term of service ended between 2005, 2010, or 2013, corresponding to 10, 5, and 2 years before the original survey administration;

  • For whom a corporate code indicating the program they participated in their last term of service is available. These codes match the following AmeriCorps programs: AmeriCorps State and National (ASN), including America’s Promise, Education Award Program (EAP), Homeland Security, and Tribes and Territories; National Civilian Community Corps (NCCC); and Volunteers in Service to America (VISTA);

  • For whom total length of service across all terms of service is greater than or equal to 84 days full time equivalency (FTE), which is equivalent to completing at least one reduced half time term of service;

  • For whom length of service in most recent term was greater than zero days;

  • For whom total length of service did not exceed ((exit year of last term - year of birth - x) × 365.25), where x = minimum age of service - 1. For individuals who served multiple terms or who served one term with ASN, minimum age of service of 17 was used. For individuals who served one term with NCCC or VISTA, minimum age of service of 18 was used. One was subtracted from minimum age of service to account for rounding error due to exit year only being available without month and date;

  • For whom length of service of most recent term did not exceed ((exit year of last term - year of birth - x) × 365.25), where x is as defined above and minimum date of service is 17 for people whose last term of service was in ASN and 18 for people whose last term of service was in NCCC or VISTA.

The restriction to participants with a total service time greater than or equal to 84 days FTE is to ensure that participants have sufficient exposure to AmeriCorps program for it to have an impact on them.

The restriction to participants with a service time greater than zero days is to ensure that at least one term of service was with an eligible program. Information besides total length of service is only available regarding the most recent program; which programs the participant served in during earlier terms are not known.

The maximum length of service is designed to remove participants with data entry errors.

The estimated universe is stratified by time since last service (10, 5, and 2 years; or between 2008-10, 2013-15, and 2016-18 alumni) and program (ASN, NCCC, and VISTA). Exhibit B1 shows the size of the universe.

Exhibit B1. Estimated Universe Size for AmeriCorps Alumni Outcomes Survey based on Previous Study


Program


Cohort

ASN

NCCC

VISTA

Total

2005

17,628

883

4,493

23,004

2010

24,954

730

6,294

31,978

2013

18,483

1,307

4,574

24,364

Total

61,065

2,920

15,361

79,346

The sample drawn will be limited to participants for whom an email address and one of either a mailing address or phone number is available. The estimated size of the population from which the sample will be drawn is shown in Exhibit B2 below. These estimates are based in the population and sample from the 2014 study. The focus on participants for whom contact information is available represents a trade-off between coverage error on the one hand and nonresponse and sampling error on the other. The population from which the sample will be drawn is smaller than the population as a whole, creating potential for coverage error: systematic differences in estimands of interest between the population covered and the population not covered, leading to misestimates. The risk of coverage error is, however, offset by reduced risk of nonresponse error and a reduction in sampling error. We anticipate that response rates will be higher for a sample for which additional information is available, reducing the risk of nonresponse error: systematic differences in estimands of interest between respondents and nonrespondents to the survey, leading to misestimates. Although increases in response rates do not guarantee a reduction in nonresponse error (cf. Lin and Schaeffer 1995), it is nevertheless true that as response rates increase, the potential impact on estimates from systematic differences between respondents and nonrespondents decreases. With respect to sampling error, nonsystematic differences in estimands of interest from between the specific realization of the sampling scheme and underlying population, smaller sample sizes will be associated with increased sampling error. The inclusion of AmeriCorps participants with poorer quality contact information would reduce sample size for a given budget, increasing sampling error.

Exhibit B2. Estimated Sampling Frame Size for AmeriCorps Alumni Survey Base on Previous Study


Program


Cohort

ASN

NCCC

VISTA

Total

2005

7,566

527

2,137

10,230

2010

24,805

729

6,277

31,811

2013

18,479

1,304

4,573

24,356

Total

50,850

2,560

12,987

66,397

The focus on participants for whom contact information is available represents a trade-off between coverage error on the one hand and nonresponse and sampling error on the other. Coverage error is discussed first below, followed by nonresponse and sampling error.

Due to the exclusion of participants without sufficient contact information, the population from which the sample will be drawn is smaller than the population as a whole, creating potential for coverage error (i.e., systematic differences in estimands between the population covered and the population not covered, leading to misestimates).

The risk of coverage error is, however, offset by reduced risk of nonresponse error and a reduction in sampling error, which we discuss in turn. We anticipate that response rates will be higher for a sample for which additional information is available, reducing the risk of nonresponse error (i.e., systematic differences in estimates of interest between respondents and nonrespondents to the survey, leading to misestimates). Although increases in response rates do not guarantee a reduction in nonresponse error (cf. Lin and Schaeffer 1995), it is nevertheless true that as response rates increase, the potential bias from systematic differences between respondents and nonrespondents decreases. With respect to sampling error, the smaller sample sizes that would be achieved were individuals with less contact information to be included in the sampling frame would increase sampling error (i.e., nonsystematic differences in estimates of interest from between the specific realization of the sampling scheme and underlying population).

We recognize that restricting the sampling frame to participants with an email address and either a phone number or mailing address is not a complete solution to the problem of contact information. Much of the contact information in the sampling frame was collected when the participant applied to AmeriCorps and has not been updated since. As we describe in section B3, we have included multiple steps to improve the quality of contact information and increase the likelihood of successfully reaching a participant.

The desired sample sizes for the nine program by cohort strata are shown in Exhibit B3, below.

Exhibit B3. Desired Sample Size for AmeriCorps Alumni Outcomes Survey based on previous study


Program


Cohort

ASN

NCCC

VISTA

Total

2005

491

153

182

826

2010

613

176

356

1,145

2013

551

357

271

1,179

Total

1,655

686

809

3,150

Note: Completed surveys.

Expected Response Rate

We aim to achieve a response rate of approximately 50%. A higher response is anticipated this year compared to the previous year based on the following factors:

  • Sample will be released in a single replicate.

  • The survey makes additional use of telephone interviews. Telephone interviewers will be instructed to complete the survey over the phone where possible with the option of sending an email containing a link to the survey or reading the survey link over the phone to be used only where the respondent is unable or unwilling to complete the interview over the phone. The protocol encourages completion of the survey over the phone because only 39% of people emailed the link by an interviewer completed the survey online in 2015.

Actual Response Rate in Previous Data Collection

A 20% response rate was achieved in the 2015 AmeriCorps Alumni Outcomes Survey (CNCS 2015). In 2015, sample was released in three replicates, termed “phases” by CNCS (2015). The final two replicates did not have sufficient time to be worked fully. This is reflected in response rates by replicate. The first replicate achieved a 26% response rate; the second achieved a 19% response rate; the third achieved a 9% response rate.

B2. Describe the procedures for the collection of information, including: Statistical methodology for stratification and sample selection; Estimation procedure; Degree of accuracy needed for the purpose described in the justification; Unusual problems requiring specialized sampling procedures; and any use of periodic (less frequent than annual) data collection cycles to reduce burden.

Statistical Methodology for Stratification and Sample Selection

The desired sample sizes shown in Exhibit B3 were calculated using optimal allocation procedures. Sampling fractions ( ) are allocated to the nine strata formed from the three cohorts (2005, 2010, 2013) and three programs (ASN, NCCC, and VISTA), together with the subsampling rate for cases with bad email addresses, to:

subject to:

;

;

;

where:

where is the sample size of completed interviews and the population size for each of the strata;

pairs of programs (ASN-NCCC, ASN-VISTA, and NCCC-VISTA);

pairs of cohorts (2005-2010, 2005-2013, 2010-2013);

(i.e., the minimum detectable difference of two proportions);

(i.e., calculated around a proportion of 0.5 where variance will be at a maximum);

is defined similarly to above for the subsamples of each pair of cohorts within and programs within . is the finite population correction for the th subsample;

is the effective sample size of completed interviews the th subsample, calculated using the approximation, where is the coefficient of variation of the weights in the th subsample;

is the th quantile of the distribution for a sample size of . is used in preference to the normal distribution due to the small size of some subsamples;

is the probability of Type I error, set at 0.05;

is the probability of Type II error, set at 0.20;

are importance weights for cohorts ( ) and programs ( ) used to weight the relative importance of precision for each (here, and ); and,

is a penalty function where is the mean of within program or cohort, as appropriate, and is a constant (here, ) used to scale the impact of the penalty function. Where is higher, inequality between minimum detectable differences will more heavily penalize the objective function.

Within strata, sample will be selected using a systematic random sample with implicit stratification. The sampling frame will be sorted by available demographic and programmatic data, with missing data included (e.g., sex would consist of female, male, and missing) and, at the lowest level, by a randomly generated number. The demographic and programmatic fields to be used for implicit stratification are CRPP code (a more detailed measure of the specific program the participant served in for ASN participants), age upon start of last service term, number of terms served, sex, length of last term of service, term type (e.g., full-time, half-time, etc.), whether the education award was used, education, and whether the most recent term was completed. Starting from a random point, every 1 in cases will be selected, where . This procedure reduces design effects compared to a simple random sample by ensuring proportional representation of groups within the stratum.

Exhibit B4, below, shows the proposed sample drawn and expected number of completed surveys.

Exhibit B4. Estimated Universe and Sample Sizes for AmeriCorps Alumni Outcomes Survey based on previous study


Program


Cohort

ASN

NCCC

VISTA

Total

Population

2005

17,628

883

4,493

23,004

2010

24,954

730

6,294

31,978

2013

18,483

1,307

4,574

24,364

Total

61,065

2,920

15,361

79,346

Sampling Frame

2005

7,566

527

2,137

10,230

2010

24,805

729

6,277

31,811

2013

18,479

1,304

4,573

24,356

Total

50,850

2,560

12,987

66,397

Completed Surveys Expected

2005

491

153

182

826

2010

613

176

356

1,145

2013

551

357

271

1,179

Total

1,655

686

809

3,150

Estimation Procedures

In order to obtain valid survey estimates, estimation will be done using properly weighted survey data. The weight to be applied to each respondent is a function of the overall probability of selection, and appropriate nonresponse and post-stratification ratio adjustments.

Base weights ( ) are calculated as the inverse of the selection probability based on the sample design:

.

Subsampling weights ( ) for participants with email addresses are calculated as:

for participants with good email addresses and,

for participants with bad email addresses.

Note, however, that the optimal allocation procedures generated a subsampling rate of 1 (i.e., there will be no subsampling of bad email addresses and for participants with bad email addresses).

There will inevitably be nonrespondents to the survey and weighting adjustments will be used to compensate for them. The nonresponse-adjusted weight ( ) for the th weighting class will be computed as:

where S is the set of completed surveys in the th weighting class and the weighting classes will be based on a propensity score model created with the goal to minimize the bias due to nonresponse. The propensity score model will estimate the probability of response using logistic regression. The dependent variables will be based on frame data. The propensity scores will be grouped into quintiles. Within each quintile class, we will ratio adjust the respondents to reflect the nonrespondents, as described above.

To help reduce possible under-coverage errors in the sampling frame and reduce possible nonresponse bias, the final estimation weights will also be calibrated to the population sizes of the strata, as well as other frame data. Candidate variables for use in calibration include sex, education, age, CRPP code, term type, term length, education, whether the education award was used and number of terms served. Race/ethnicity is not available in the sampling frame and cannot be used for calibration.

Degree of Accuracy

Analyses of interest to CNCS are comparisons between cohorts and comparisons between programs. Expected design effects and margins of error for for a proportion of are shown in Exhibit B5. Minimum detectable differences for pairs of cohorts and programs are shown in Exhibit B6 for and statistical power of (i.e., ). Design effects are calculated using the approximation and have been increased by a factor of to account for the effect of calibration.

Exhibit B5. Estimated Expected Design Effects and 95% Confidence Intervals

Estimand

Design Effect

Margin of Error

ASN

1.34

2.7%

NCCC

1.13

4.0%

VISTA

1.15

3.7%

2005

1.27

3.7%

2010

1.65

3.4%

2013

1.63

3.5%

Total

1.83

2.1%

Note: Margins of error for a statistic of 50% shown; margins of error will be lower for statistics above or below 50%.

Exhibit B6. Estimated Expected Minimum Detectable Differences for Comparisons of Interest

Comparison

Minimum Detectable Difference

2010-2013

6.9%

2005-2013

7.2%

2005-2010

7.2%

ASN-NCCC

6.3%

ASN-VISTA

6.2%

NCCC-VISTA

7.0%

Notes: For a statistic of 50%; margins of error will be lower for statistics above or below 50%. ; .

Standard errors will be computed using statistical software that accounts for the complex survey such as Stata svy and SAS SURVEY procedures.

B3. Describe methods to maximize response rates and to deal with issues of non-response. The accuracy and reliability of information collected must be shown to be adequate for intended uses. For collections based on sampling, a special justification must be provided for any collection that will not yield “reliable” data that can be generalized to the universe studied.

The AmeriCorps Alumni Survey employs a number of strategies to maximize response rates while maintaining cost control, which are detailed below.

The availability and quality of contact information for AmeriCorps alumni is a major driver of nonresponse. Among the eligible population, email addresses are available for 84% participants (44% of 2005 participants, 99% of 2010 participants, and 100% of 2013 participants), addresses are available for 99% participants (97% of 2005 participants and 100% of 2010 and 2015 participants), and phone numbers are available for 56% of participants (93% of 2005 participants, 43% of 2010 participants, and 40% of 2013 participants). As is described in section B1, an initial step in maximizing response rates is to restrict the sampling frame to cases with an email address and either a telephone number or a mailing address on file. However, much of the contact information available will be out-of-date. Additional steps are therefore taken to update contact information.

The contact information available for the sample will be updated in order to maximize the probability of successfully reaching sample members. All cases will be updated using LexisNexis’s Accurint service. The availability of Social Security Numbers for all participants improves match rates compared to not having this information available.

Prior to data collection, efforts will be made to verify contact information. This outreach will occur prior to fielding the survey and will be used to select the sample. These procedures are shown in Exhibit B7.

Shape30
Shape29
Shape28
Shape27
Shape26
Shape25
Shape24
Shape23
Shape22
Shape21
Shape20
Shape19
Shape18
Shape17
Shape16
Shape15
Shape14
Shape13
Shape12
Shape11
Shape10
Shape9
Shape8
Shape7
Shape6
Shape5
Shape4
Shape2
Shape3

Cases can be on multiple tracks where different types of contact information

Undeliverable email identified

Undeliverable identified; update from forwarding notice

Bad numbers identified

Inactive, unknown numbers identified

Landlines dialed on autodialer

Cell WINS check if cell number active

LNPA lookup to identify cell phones

Send study description postcard

Send study description email

Phone number on file

Mailing address on file

Email address on file

Accurint batch process update

Sample drawn

Exhibit B7. Contact Information Verification


Participants will be sent an email describing the study; as described above, only participants with email addresses are eligible for selection. The purpose of these emails is primarily to identify undeliverable (“bounced”) email addresses. A contact line will be provided for finding more out about the study as well as a link to the CNCS website describing the study. See Appendix B for the text of all planned participant communications.

Alumni with mailing addresses will be sent a study description postcard describing the study with return service requested. A contact line will be provided for finding more out about the study as well as a link to the CNCS website describing the study. Address forwarding notifications will be updated in the study records and undeliverable addresses will be recorded.

Phone numbers on file will be validated. Numbers will be classified as landline or cell phones based on telephone exchange and the Local Number Portability database in order to comply with the recent Federal Communications Commission ruling regulating dialing cell phones. For numbers classified as belonging to cell phones, numbers will be checked using Marketing Systems Group’s Cell WINS database to identify whether the number is active. Landline numbers will be dialed by autodialer to determine whether the number is working or not. No interviews will take place; the autodialer will only determine whether the number is working and immediately terminate the call.

Contact procedures for the survey itself are shown in Exhibit B8. These are separate from the procedures for verifying contact information shown in Exhibit B7.

Shape31

No working email

Exhibit B8. Contact Procedures


Shape64
Shape63
Shape62
Shape61
Shape60
Shape59
Shape58
Shape57
Shape56
Shape55
Shape54
Shape53
Shape52
Shape51
Shape50
Shape49
Shape48
Shape47
Shape46
Shape45
Shape44
Shape32
Shape43
Shape42
Shape41
Shape40
Shape39
Shape38
Shape37
Shape36
Shape35
Shape34
Shape33

Send final reminder letter if email not available & no phone

No phone

Email

Phone

No phone

Email

Phone

No phone

No address

Send final email if email available

Reminder email x 3

Phone calls if number available

Send advance postcard if address available

Send reminder postcard with URL if no phone

Invitation letter with URL and QR code if working email not available

Send reminder letter if address available

Invitation email if email available

The initial contact for the survey will be an advance postcard. (This is distinct from the postcard used for contact information verification.) Postcards have been shown to be effective in increasing response rates in two experiments. An advance postcard mailing increased response rates by four percentage points in an experiment conducted on the Pew Research Center’s American Trends Panel (Bertoni et al. 2015), while a nine percentage point increase was observed in a similar experiment with advance postcards by Kaplowitz, Westin, and Levin (2005). These findings are consistent with the literature on the use of advance letters for telephone surveys (c.f. De Leeuw et al. 2007; Goldstein and Jennings 2002; Link and Mokdad 2005). Our use of mailed advance notification is analogous to the procedures used for the first replicate in 2015, where advance letters were mailed to alumni with postal addresses.

Alumni with working email addresses will be sent an invitation email from CNCS’s GovDelivery email system approximately one week after the advance postcard is mailed. The use of a CNCS.gov email address is intended to emphasize the legitimacy of the data collection request. Three email reminders will be sent to nonrespondents at approximately one week intervals thereafter. These procedures are similar to those used in 2015, except that the length of time between reminders is increased from four days to a week and, in 2015, the invitations were sent from the contractor (see CNCS 2015 for further details; an advance email was sent via GovDelivery last year, but not a survey invitation). The longer period between email contacts is intended to reduce potential frustration on the part of alumni due to the frequency of email contacts.

An invitation letter will be sent to alumni for whom no working email address is available. The invitation letter will be mailed on AmeriCorps stationery and will be signed by AmeriCorps directors. The letter will contain a unique URL for the survey and will include a QR code that can be used by respondents to go to their unique URL without having to manually enter the address. The use of a mailed letter in 2015 was limited to alumni who could not be reached via email or phone and did not include a QR code (CNCS 2015). The letter is to be mailed in advance of calls in order to maximize cost-efficiency: surveys completed via web are less expensive than surveys completed via phone. A reminder postcard also containing the survey link will be sent to nonrespondents following the invitation letter.

To enhance response rates, a robust telephone effort will be made for alumni for whom a telephone number is available. Interviewers place calls and will be trained to attempt to complete the survey on the telephone except where the alumnus or alumna cannot or will not do so. Although the interviewer will be able to send an email containing the alumna or alumnus’ unique URL to an email address provided by an informant and will also be able to read out the URL upon request, completing the survey via phone is emphasized due to the relatively low response rate of 39% in 2015 for alumni who were reached by phone and requested to be emailed the study link. Interviewers will be able to update the telephone number on file if they receive new information (e.g., if the phone number is that of the alumnus or alumnus’ parents).

The final contacts made for nonrespondents will be a final email reminder for alumni with working email addresses and a final reminder letter for alumni without working email addresses. The letter will be mailed on AmeriCorps stationery.

As an experiment, 50% of all respondents will be offered a $2 incentive paid to the National Park Foundation on their behalf as a thank you for completing the survey. We recommend using the National Park Foundation as a recognizable, federally-affiliated, politically neutral recipient of the charitable donation. The National Park Foundation, in partnership with the National Park Service, enriches America’s national parks and programs through private support, safeguarding our heritage and inspiring generations of national park enthusiasts. The National Park Foundation has a well-developed framework for applying for permission to use their name and was able to supply permission quickly.

The prior alumni survey achieved only a 20% response rate, so this experiment is proposed as both an attempt to increase the response rate to this year’s alumni survey and to advance the body of knowledge on the effectiveness of a charitable donation incentive. Although the experimental evidence for the effectiveness of charitable donations has been predominantly negative, AmeriCorps alumni may be an exception due to altruistic orientation.1 AmeriCorps alumni scored higher on sense of connection to the community, a sense of civic obligations, being active in community affairs, and volunteering than did a control group, and were more likely to have donated to Hurricane Katrina relief than were controls (Corporation for National and Community Service, 2008a). Given these findings, it is not unreasonable to expect that the offer of a charitable donation may be more effective for AmeriCorps alumni than for the population at large. After the survey data collection period is complete, the response rates will be compared for the groups that were offered and not offered the charitable donation incentive, and any statistically significant effect will be reported in the final survey technical report.

Due to the limited field time available for the survey, the entire sample will be released at once rather than by replicates in order to avoid the difficulties experienced in 2015 when the second and third replicates could not be worked fully due to time limitations.

To assess the impact of nonresponse bias in our study, we will conduct statistical analysis to identify any characteristics of respondents that are correlated with response. Propensity scores from the logistic regression of survey response on explanatory variables will be used to create nonresponse weighting classes for statistical adjustment. These nonresponse weights will be combined with sampling weights based on our stratification plan to create the final weights for our analysis.

B4. Describe any tests of procedures or methods to be undertaken.

The survey instrument included in this justification is based upon the survey used in 2015. Cognitive interviews were conducted in addition to a pilot survey with 14 responses. In addition it was reported that “Responses showed good variability on almost all items, and no unexpected answer patterns that would indicate a question was not working as intended” (CNCS 2015:4). In turn, the 2015 instrument was based on largely on the AmeriCorps exit survey, with wording changes kept to a minimum.

To test the instrument to be used in 2016, a sample of 150 alumni was drawn from the sampling frame (see Exhibit B9 below). Due to expectations that all or almost all NCCC alumni would be sampled for the survey, the pilot survey sample was drawn solely from ASN and VISTA. More recent cohorts were oversampled in an effort to obtain more responses, and the sample was further restricted to alumni with an email address modification date in 2015 in order to further increase the number of responses. In addition, only alumni with full contact information on file were included in the pilot sample frame, as the 2015 survey found that alumni with both pieces of contact information were more likely to respond to the survey even before telephone reminders.


Exhibit B9. Pilot Sample for AmeriCorps Alumni Outcomes Survey


Program


Cohort

ASN

NCCC

VISTA

Total

2005

6

0

6

12

2010

30

0

30

60

2013

39

0

39

78

Total

75

0

75

150

A total of 15 complete responses were obtained. The median length of response to the survey was 27.5 minutes, excluding the pilot-only questions. Several pilot respondents had very long completion times, suggesting that they started the survey, put it aside, and completed it much later. Therefore, the pilot response burden estimate of 25 minutes seems reasonable.

All pilot respondents indicated that the survey instructions were clear and easy to follow and that the order of the questions made sense. Pilot respondents suggested changes to several items, and while some items were clarified based on this feedback it was determined that some changes should not be made in order to maintain direct comparability with the AmeriCorps Exit Survey. Eight of the fifteen pilot respondents felt that the survey was too long. In response, 12 questions were dropped from the survey and one question was added, reducing the estimated response burden by three minutes, to 22 minutes.

B5. Provide the name and telephone number of individuals consulted on statistical aspects of the design, and the name of the agency unit, contractor(s), grantee(s), or other person(s) who will actually collect and/or analyze the information for the agency.

  • CONTRACT HAS NOT BEEN AWARDED AND WILL NOT BE COMPLETED UNTIL FICAL YEAR 2019

The project officer at CNCS for this project is:

  • NOT ASSIGNED



References

Bertoni, Nick, Andrew Burkey, Molly Caldaro, Scott Keeter, Charles DiSogra, and Kyley McGeeney. 2015. “Advance Postcard Mailing Improves Web Panel Survey Participation.” Paper presented at the annual conference of the American Association for Public Opinion Research, Hollywood, FL.

Corporation for National and Community Service. 2015. “AmeriCorps Alumni Outcomes: Technical Memo.” Corporation for National and Community Service, Washington, D.C.

De Leeuw, Edith, Mario Callegaro, Joop Hox, Ellen Korendijk, and Gerty Lensvelt-Mulders. 2007. “The Influence of Advance Letters on Response in Telephone Surveys: A Meta-Analysis.” Public Opinion Quarterly 71:413-43.

Goldstein, Kenneth M. and M. Kent Jennings. 2002. “The Effect of Advance Letters on Cooperation in a List Sample Telephone Survey.” Public Opinion Quarterly 66:608-17.

Kaplowitz, M.D., T.D. Hadlock, and R. Levine. 2004. “A Comparison of Web and Mail Survey Response Rates.” Public Opinion Quarterly 68:94-101.

Lin, I-Fen and Nora-Cate Schaeffer. 1995. “Using Survey Participants to Estimate the Effect of Nonparticipation.” Public Opinion Quarterly 59:236-58.

Link, Michael W. and Ali Mokdad. 2005. “Advance Letters as a Means of Improving Respondent Cooperation in Random Digit Dial Studies: A Multistate Experiment.” Public Opinion Quarterly 69:572-87.



Appendix B

Study Description Postcard Sent for Contact Information Verification

In the coming months, you will receive a request to complete a survey of AmeriCorps alumni. You were selected because you served in [program]. The Corporation for National and Community Service, the federal agency that administers AmeriCorps programs, is seeking your input to help us better understand your service experience and how your participation affected your life and career. The survey will be conducted by [NAME OF CONTRACTOR], a survey research firm. Please contact them at [CONTACT INFORMATION] and mention ID [qkey] to update your contact information to make sure you receive this important survey or if you have any questions about this study. For more information about the study, please visit www.nationalservice.gov/impact-our-nation/evidence-exchange/americorps-alumni-survey.

[signature]

[name and title]



Study Description Email Sent for Email Address Verification

Dear [First Name],

In the coming months, you will receive a request to complete a survey of AmeriCorps alumni. You were selected because you served in [program]. The Corporation for National and Community Service, the federal agency that administers AmeriCorps programs, is seeking your input to help us better understand your service experience and how your participation affected your life and career. The survey will be conducted by [CONTRACTOR NAME], a survey research firm. Please contact us at [CONTACT INFORMATION] and mention ID [qkey] to update your contact information to make sure you receive this important survey or if you have any questions about this study. For more information about the study, please visit the [DEDICATED WESITE LINK].

Sincerely,

NAME:

PROJECT DIRECTOR



Advance Postcard (sent to all)

A few days from now, you will receive [email: an email / no email: a mailed] request to complete a survey of AmeriCorps alumni. You were selected because you served in [program]. The Corporation for National and Community Service, the federal agency that administers AmeriCorps programs, is seeking your input to help us better understand your service experience and how your participation affected your life and career. The results of the study will be shared on the CNCS website. Please contact [CONTACT INFORMATION] if you have any questions. For more information about the study, please visit www.nationalservice.gov/impact-our-nation/evidence-exchange/americorps-alumni-survey.

[SIGNATURE]

[NAME AND TITLE]



Invitation Email Sent from CNCS

Subject: Survey of AmeriCorps Alumni

Dear [firstname],

I am writing to ask you to complete a survey of AmeriCorps alumni. Through the survey, we hope to learn about your AmeriCorps experience and how it has affected you in the years since you served. The information we gather will be used to improve the quality of the service experience for all AmeriCorps members and ensure that the programs enable AmeriCorps members to excel, both in service and beyond. You were selected because you served in [program].

To take the survey, please click the link below:

[URL]

The survey is on behalf of the Corporation for National and Community Service, the federal agency that administers AmeriCorps programs. The findings will be used to help improve AmeriCorps programs and the support AmeriCorps can offer to its members. It will take approximately 20 minutes to complete.

For more information about the study, please visit the [DEDICATED WEBSITE LINK]. If you encounter any technical problems with the survey, please contact NAME at CONTRACTOR CONTACT INFO and mention your study ID, [qkey]. If you have questions about this project, please contact CNCS PROJECT STAFF NAME at CONTACT INFORMATION.

Thank you very much for your time.

Sincerely,

[CNCS signatory]

To opt out of receiving future email about this survey, please click here. [where “here” is opt out link]



Invitation Letter for Bad Email Cases

Dear [firstname],

I am writing to ask you to complete a survey of AmeriCorps alumni. Through the survey, we hope to learn about your AmeriCorps experience and how it has affected you in the years since you served. The information we gather will be used to improve the quality of the service experience for all AmeriCorps members and ensure that the programs enable AmeriCorps members to excel, both in service and beyond. You were selected because you served in [program].

To take the survey, please go to the following link:

[URL]

Or you may scan this QR code with your smartphone:

[QR]

The survey is on behalf of the Corporation for National and Community Service, the federal agency that administers AmeriCorps programs. The findings will be used to help improve AmeriCorps programs and the support AmeriCorps can offer to its members. It will take approximately 20 minutes to complete.

For more information about the study, please visit www.nationalservice.gov/impact-our-nation/evidence-exchange/americorps-alumni-survey. If you encounter any technical problems with the survey, please contact [NAME] at [CONTRACTOR CONTACT INOFRMATION] and mention your study ID, [qkey]. If you have questions about this project, please contact CNCS PROJECT STAFF LEAD NAME AND CONTACT INFORMATION].

Thank you very much for your time.

Sincerely,

[CNCS signatory]





Email Reminder 1 Sent from [CONTRACTOR NAME] to Nonrespondents with Email Addresses

Subject: Reminder: Survey of AmeriCorps Alumni


Dear [First Name],


You were recently invited to take part in an online survey of AmeriCorps alumni by the Corporation for National and Community Service. We really value your experiences and opinions and hope that you will be able to take part in this important survey.

To take the survey, please click the link below:

[URL]

The survey asks about your AmeriCorps experience and how it has affected you in the years since you served. The information gathered will be used to improve the quality of the service experience for AmeriCorps members and ensure that the programs enable AmeriCorps members to excel, both in service and beyond. You were selected because you served in [program].

The survey is on behalf of the Corporation for National and Community Service, the federal agency that administers AmeriCorps programs. The findings will be used to help improve AmeriCorps programs and the support AmeriCorps can offer to its members. It will take approximately 20 minutes to complete.

If you encounter any technical problems with the survey, please contact [NAME] at [CONTRACTOR CONTACT INOFRMATION] and mention your study ID, [qkey]. If you have questions about this project, please contact CNCS PROJECT STAFF LEAD NAME AND CONTACT INFORMATION].

Thank you very much for your time.

Sincerely,

NAME.

CONTACTOR NAME

POSITION

To opt out of receiving future email about this survey, please click here. [where “here” is opt out link]



Email Reminder 2 Sent from [CONTRACTOR NAME] to Nonrespondents with Email Addresses

Subject: Please complete the AmeriCorps Alumni Survey


Dear [First Name],


I am writing again to ask for your participation in the Survey of AmeriCorps Alumni. We recently sent you an email about the survey. If you have already completed the survey, thank you very much. If you have not completed the survey yet, please take the time now to do so by clicking the link below:

[URL]

As a reminder, the results of this research will provide critical information that will be used to improve the quality of the service experience for all AmeriCorps. You were selected because you served in [program].

The survey is on behalf of the Corporation for National and Community Service, the federal agency that administers AmeriCorps programs. The findings will be used to help improve AmeriCorps programs and the support AmeriCorps can offer to its members. It will take approximately 20 minutes to complete.

If you encounter any technical problems with the survey, please contact [NAME] at [CONTRACTOR CONTACT INOFRMATION] and mention your study ID, [qkey]. If you have questions about this project, please contact CNCS PROJECT STAFF LEAD NAME AND CONTACT INFORMATION].Thank you very much for your time.

Sincerely,

NAME.

CONTACTOR NAME

POSITION

To opt out of receiving future email about this survey, please click here. [where “here” is opt out link]



Email Reminder 3 Sent from [CONTRACTOR NAME] to Nonrespondents with Email Addresses

Subject: It’s not too late to complete the Survey of AmeriCorps Alumni


Dear [First Name],


It’s not too late for you to complete the Survey of AmeriCorps alumni. [IF INCENTIVE: As a thank you for your time, a donation of $2 will be made to the National Park Foundation for every survey completed.] We will [IF INCENTIVE: also] share the results of the survey on the CNCS website. According to my records, I have not yet received your completed survey. To take the survey, please click the link below:

[URL]

The survey is on behalf of the Corporation for National and Community Service, the federal agency that administers AmeriCorps programs. The findings will be used to help improve AmeriCorps programs and the support AmeriCorps can offer to its members. It will take approximately 20 minutes to complete.

For more information about the study, please visit the [DEDICATED WEBSITE LINK]. If you encounter any technical problems with the survey, please contact [NAME] at [CONTRACTOR CONTACT INOFRMATION] and mention your study ID, [qkey]. If you have questions about this project, please contact CNCS PROJECT STAFF LEAD NAME AND CONTACT INFORMATION].Thank you very much for your time.

.

Thank you very much for your time and consideration.

Sincerely,

NAME.

CONTACTOR NAME

POSITION

To opt out of receiving future email about this survey, please click here. [where “here” is opt out link]



Reminder Postcard for Bad Email or Phone Nonrespondents

I am writing to ask for your participation in the Survey of AmeriCorps Alumni. I recently sent you a letter about the survey and, according to my records, I have not yet received your completed survey. We really value your experiences and opinions and hope that you will be able to take part in this important survey. You were selected because you served in [program]. The Corporation for National and Community Service, the federal agency that administers AmeriCorps programs, is seeking your input to help us better understand your service experience and how your participation affected your life and career. This online survey will take about 20 minutes. We] will share the results on the CNCS website. Contact AmeriCorps@srbi.com if you have any questions and mention study ID [qkey]. Please complete the survey now by going to:

[URL]

[signature]

[name and title]



Final Reminder Letter for All Nonrespondents with Good Mailing Address

Dear [firstname],

During the past two months, we contacted you several times about an important study of AmeriCorps alumni. We really value your experiences and opinions and hope that you will be able to take part in this important survey, which will take approximately 20 minutes to complete.

Its purpose is to learn about your AmeriCorps experience and how it has affected you in the years since you served. Information from the survey will be used to improve the quality of the service experience for all AmeriCorps members and ensure that the programs enable AmeriCorps members to excel, both in service and beyond. You were selected because you served in [program].

[IF NO EMAIL: The study is drawing to a close.] To take the survey, please go to the following link:

[URL]

Alternately, you may scan this QR code with your smartphone:

[QR]

The survey is on behalf of the Corporation for National and Community Service, the federal agency that administers AmeriCorps programs. The findings will be used to help improve AmeriCorps programs and the support AmeriCorps can offer to its members.

If you encounter any technical problems with the survey, please contact [NAME] at [CONTRACTOR CONTACT INOFRMATION] and mention your study ID, [qkey]. If you have questions about this project, please contact CNCS PROJECT STAFF LEAD NAME AND CONTACT INFORMATION].Thank you very much for your time.

Thank you very much for your wiliness to consider our request as we conclude this effort to better understand the lives of AmeriCorps alumni.

Sincerely,

[CNCS signatory]



Telephone Interview Script

INTRO1 Hello, can I please speak to [firstname] [lastname]? [GO TO TIPRESP]

  1. Named Respondent speaking [GO TO INTRO2]

  2. Wrong number (haven’t heard of named respondent) [GO TO SORRY1] (FINAL OUTCOME) TRY ALTERNATIVE NUMBER IF AVAILABLE

  3. Named respondent is temporarily unavailable [IDENTIFY BEST TIME TO CALL BACK, MAKE SOFT APPOINTMENT]

  4. Named respondent cannot be reached by phone at any time/uncontactable by phone [GO TO SORRY1] (FINAL OUTCOME) TRY ALTERNATIVE NUMBER IF AVAILABLE

  5. Named respondent is unavailable during fieldwork [GO TO SMS RECORD OUTCOME] (FINAL OUTCOME) TRY ALTERNATIVE NUMBER IF AVAILABLE

  6. Language difficulty [MAKE SOFT APPOINTMENT TO TRY AGAIN ANOTHER TIME]

  7. Call back on another number [UPDATE NUMBER AND ARRANGE TO CALL NEW NUMBER]

  8. Number unobtainable/disconnected/out of order TRY ALTERNATIVE NUMBER IF AVAILABLE

  9. Computer/modem/fax line TRY ALTERNATIVE NUMBER IF AVAILABLE

  10. Answer machine/voicemail IF CONFIRMED, LEAVE VOICEMAIL AND DO NOT TRY AGAIN

  11. No answer TRY AGAIN/TRY ALTERNATIVE NUMBER IF AVAILABLE

  12. Line or extension busy TRY AGAIN/TRY ALTERNATIVE NUMBER IF AVAILABLE

  13. Refusal by gatekeeper (PA/assistant etc), unable to be connected to named respondent [GO TO SORRY 1] (FINAL OUTCOME – HARD REFUSAL)

  14. Refused – organisation policy [GO TO SORRY 1] (FINAL OUTCOME – HARD REFUSAL)

  15. Gatekeeper requested information to pass on to respondent [GO TO GATEKEEPER1]

(+Dialer outcomes not visible to interviewer) TRY AGAIN IF APPROPRIATE OR TRY ALTERNATIVE NUMBER IF AVAILABLE

ASK IF CODE 1 AT INTRO1:

INTRO2 Hello, my name is _______ and I am calling from [CONTRACTOR NAME], a national survey research firm, on behalf of the Corporation for National and Community Service, a U.S. government agency. The purpose of the call is to learn about your AmeriCorps service experience. This is a nationwide survey, developed by AmeriCorps, and your valuable input will help us make AmeriCorps more effective for its members. The survey will take approximately 20 minutes. Will you be able to take the survey? (IF NECESSARY: Would you prefer to complete it online?)

  1. Yes – online – GO TO INTRO3

  2. Yes – phone interview – GO TO INTRO7

  3. Not heard about survey before/not received letter or emails – GO TO INTRO5

  4. Already completed interview – GO TO SORRY1 (FINAL OUTCOME - COMPLETE)

  5. Language difficulty, cannot continue GO TO SORRY2 (FINAL OUTCOME- SCREENOUT)

  6. Call back on another number NOTE NUMBER AND ARRANGE TO CALL NEW NUMBER – RESUME AT INTRO2

  7. Hard callback on same number to go through reminder script GO TO SMS AND RECORD DETAILS OF APPOINTMENT – RESUME AT INTRO2

  8. Refused – GO TO WHYREFUSED

  9. Refused – remove from further research GO TO SORRY 1 (FINAL OUTCOME– HARD REFUSAL)

  10. Refused – organization policy GO TO SORRY 1 (FINAL OUTCOME– HARD REFUSAL)

  11. Has question for project staff before will complete - NOTE QUESTION IN OPEN-END FOLLOW-UP AND GO TO INTRO5

ASK IF CODE 1 AT INTRO2:

INTRO3 Would you like me to send an email containing the link to the survey to you?

  1. Yes – IF EMAIL ADDRESS ON FILE ASK INTRO4; IF NO EMAIL ADDRESS ON FILE SKIP TO INTRO6

  2. No - DO NOT SEND THE EMAIL. GO TO THANKS1 (FINAL OUTCOME - COMPLETE)

ASK IF CODE 1 AT INTRO3 & EMAIL ADDRESS ON FILE:

INTRO4 Can I check the e-mail address we have is correct? {SHOW EMAIL ADDRESS FROM SAMPLE FILE HERE} READ OUT EMAIL ADDRESS

  1. Yes correct – AUTOMATICALLY SEND THE EMAIL. THEN GO TO THANKS1 (FINAL OUTCOME - COMPLETE)

  2. No not correct – ASK INTRO6

ASK IF CODE 3 OR 11 AT INTRO2:

INTRO5 [IF INTRO2=3: If you haven’t received the letter or the e-mails, it’s probably best if we send you some information so you can see what the survey is about. / IF INTRO2=11: I will request that someone from the project team contact you to answer your question.] Can I check the email address we have is correct, and then we’ll email you a letter with some background information and a link to the survey? {SHOW EMAIL ADDRESS FROM SAMPLE FILE HERE} READ OUT EMAIL ADDRESS. Is this correct?

  1. Yes correct – AUTOMATICALLY SEND THE EMAIL. THEN GO TO THANKS1 (FINAL OUTCOME - COMPLETE)

  2. No not correct – ASK INTRO6

  3. No, doesn’t want to provide email address/doesn’t want to take part - GO TO WHYREFUSED

ASK IF CODE 2 AT INTRO4 OR CODE 2 AT INTRO5:

INTRO6 Can I take down the [IF NOT MISSING EMAIL ADDRESS: correct] email address? ENTER EMAIL ADDRESS AND SEND AUTOMATICALLY. ALLOW REFUSED. THEN GO TO THANKS1 (FINAL OUTCOME)

ASK IF CODE 2 AT INTRO2:

INTRO7 Is now a good time to take the survey? It will take about 20 minutes.

  1. Yes – GO TO PHONE_INTRO

  2. No – ARRANGE APPOINTMENT (RESUME AT INTRO 7)

ASK IF CODE 1 AT INTRO7:

PHONE_INTRO Thank you. Please allow me a moment to pull up your survey.

Thank you for your commitment to service. To support current and future members and improve AmeriCorps programming, we are asking you to answer some brief questions so that we can better understand your AmeriCorps service experience. This survey is being conducted on behalf of the Corporation for National and Community Service (CNCS) by [CONTRACTOR NAME], a survey research firm, and is entirely voluntary. Your responses will remain private to the extent permitted by law, as is provided for in the Privacy Act of 1974. Your survey responses will be summarized in reports along with the responses of thousands of other AmeriCorps alumni in aggregate form only. CNCS will receive your responses with your name attached. However, all identifiable information will be removed from the data before they are made available to other researchers. The OMB control number for this collection is #3045-0163, which expires on 10/31/17. For questions/concerns about your rights as a study participant, please call the [CONTRACTOR] Institutional Review Board at [PHONE NUMBER].



INTERVIEWER: GO TO opinionport.com/CNCS_CATI AND ENTER [USERID] TO BEGIN SURVEY. STAY ON THIS SCREEN UNTIL DONE WITH PHONE SURVEY.

WHEN DONE WITH PHONE INTERVIEW PRESS 1 TO CONTINUE TO PHONE_END.

PHONE_END INTERVIEWER RECORD STATUS OF PHONE SURVEY:

  1. COMPLETE [END]

  2. PARTIAL - CALLBACK

  3. PARTIAL - REFUSAL [GO TO WHYREFUSED]


ASK IF CODE 3 AT INTRO2 OR CODE 3 AT INTRO5 OR CODE 3 AT PHONE_END:

WHYREFUSED While participation in this survey is completely voluntary, may I ask why you don’t want to take this survey? Your response to this question may help the Corporation for National and Community Service to better understand people’s different reasons for not taking the survey. DO NOT READ OUT PRECODES. PROBE SENSITIVELY IF REQUIRED.

  1. Too busy/survey takes too long [GOTO DECLINE] (FINAL OUTCOME – HARD REFUSAL)

  2. Don’t want to answer personal questions [GOTO CONFIDENTIALITY] (FINAL OUTCOME – HARD REFUSAL)

  3. Concerned about identification/confidentiality [GOTO CONFIDENTIALITY] (FINAL OUTCOME – HARD REFUSAL)

  4. Don’t view study as important or relevant [GOTO DECLINE] (FINAL OUTCOME – HARD REFUSAL)

  5. Don’t respond to unsolicited requests [GOTO RANDOMLY SELECTED] (FINAL OUTCOME – HARD REFUSAL)

  6. Don’t want to give a reason/refused to answer [GOTO SORRY1] (FINAL OUTCOME – HARD REFUSAL)

  7. Some other reason (PLEASE SPECIFY) [GOTO DECLINE] (FINAL OUTCOME – HARD REFUSAL)

ASK IF CODE 1 OR CODE 4 OR CODE 10 AT WHYREFUSED:

DECLINE I hope that you will reconsider. Your perspectives are important to us and your participation is essential to understanding the lives of AmeriCorps participants. If you change your mind, the link to the survey will be active until [date TBA]. Thank you very much for your time.

ASK IF CODE 2 OR 3 AT WHYREFUSED:

CONFIDENTIALITY I would like to reassure you that we have extensive protocols in place to protect your privacy and confidentiality. Your survey responses will be summarized in reports along with the responses of thousands of other AmeriCorps alumni in aggregate form only. While CNCS will receive your responses with your name attached, all identifiable information will be removed from the data before they are made available to other researchers. I hope you will reconsider as your opinions are very important to the study. If you change your mind, the link to the survey will be active until [date TBA]. Thank you very much for your time.

ASK IF CODE 5 AT WHYREFUSED:

RANDOMLY SELECTED The research team identified AmeriCorps participants and randomly selected you to participate. I hope you will reconsider as your opinions are very important to the study. If you change your mind, the link to the survey will be active until [DATE TBA]. Thank you very much for your time.

ASK IF CODE 17 AT INTRO1:

GATEKEEPER1 This message is for [firstname] [lastname]. I am calling from [CONTRACTOR NAME] on behalf of the Corporation for National and Community Service about an important survey of AmeriCorps participants. The Corporation for National and Community Service is the federal agency that administers AmeriCorps programs. [firstname] [lastname] was randomly selected from AmeriCorps participants. We would like to interview [firstname]. [firstname] call us toll-free at [NUMBER] and reference study number 30370 and ID number [ID], or email [DEDICATED EMAIL ADDRESS] and reference ID number [ID]. Thank you.

THEN GO TO GATEKEEPER 2

GATEKEEPER2 Can I also just check we have the right email address for {NAME OF RESPONDENT}? [SCRIPTWRITER PLEASE SHOW EMAIL ADDRESS HERE AND “CORRECT” YES/NO OPTIONS. ALSO ALLOW REFUSED HERE. REFUSALS SHOULD GO TO THANKS2] (FINAL OUTCOME – COMPLETE)

INTERVIEWER PLEASE READ OUT E-MAIL ADDRESS/ES TO GATEKEEPER AND CODE WHETHER CORRECT OR NOT. IF NOT CORRECT, PLEASE OBTAIN CORRECT DETAILS FROM GATEKEEPER.

ASK IF CODE 7 OR 8 OR 9 AT WHYREFUSED:

SORRY1 In that case I am sorry to have disturbed you. Thank you very much for your time.

ASK IF CODE 2 AT INTRO3 OR CODE 1 AT INTRO4 OR CODE 1 AT INTRO5 OR AFTER INTRO6:

THANKS1 OK then, thanks very much for speaking to me, and I hope you will be able to complete the survey soon.

ASK IF CODE 5 AT INTRO 2:

SORRY2 I am sorry it is too difficult for us to speak on the phone. I hope you will be able to take part in the survey online. Thank you very much for your time.

ASK IF CODE 17 AT INTRO1:

THANKS2 Thank you very much for your help today.

ASK IF CODE 6 AT WHYREFUSED:

ELIGIBLE You are still eligible to participate and we hope that you will be able to take part in the survey online. Thank you very much for your time.

IF VOICEMAIL (LEAVE ON 2nd and 4th ATTEMPTS):

VM Hello, I am calling from [CONTRACTOR NAME] on behalf of the Corporation for National and Community Service, a U.S. government agency, for [firstname] [lastname]. This is a follow-up to an invitation to participate in a survey of AmeriCorps alumni. If you have received the invitation I encourage you to complete the survey at your earliest convenience. Participation is voluntary.

If you have not received an invitation or would like it resent, please call toll free at 1-XXX or by email at [DEDICATED WEBSITE LINK] (INTERVIEWER SPELL OUT EMAIL ADDRESS AFTER READING IT) and mention study 30370 and ID number [ID]. Thank you.







Final Email Reminder

SUBJECT: Last chance to take the Survey of AmeriCorps Alumni


Dear [First Name],


I am extending a final invitation for you to complete the Survey of AmeriCorps alumni. The findings will be used to help improve AmeriCorps programs and the support AmeriCorps can offer to its members.

To take the survey, please click the link below:

[URL]

The survey is on behalf of the Corporation for National and Community Service, the federal agency that administers AmeriCorps programs. It will take approximately 20 minutes to complete.

For more information about the study, please visit the DEDICATED WEBSITE LINK]. If you encounter any technical problems with the survey, please contact [NAME] at [CONTRACTOR CONTACT INOFRMATION] and mention your study ID, [qkey]. If you have questions about this project, please contact CNCS PROJECT STAFF LEAD NAME AND CONTACT INFORMATION].Thank you very much for your time.

Thank you very much for your time.

Sincerely,

[CNCS signatory]

To opt out of receiving future email about this survey, please click here. [where “here” is opt out link]


1 Brennan, Seymour, and Gendall (1993) and Robertson and Bellenger (1978) found a positive effects for promised donations to charity, although the Brennan et al. result held only for the first of three waves of the survey. Other studies found no effect for a charitable donation (Boyle et al. 2012 both for a meta-analysis of existing literature and an experiment; Furse and Stewart 1982; Hubbard and Little 1988; Olson, Schneiderman, and Armstrong 1993; Pedersen and Nielsen 2014; Skinner, Ferrell, and Pride 1984; Warriner et al. 1996).

File Typeapplication/vnd.openxmlformats-officedocument.wordprocessingml.document
AuthorDiTommaso, Adrienne (Guest)
File Modified0000-00-00
File Created2021-01-21

© 2024 OMB.report | Privacy Policy