December 2011
Education
Longitudinal Study: 2002
(ELS:2002)
Third Follow-up, 2012 Full-scale Study
OMB Supporting Statement
Part A
OMB# 1850-0652 v.8
National Center for Education Statistics
Institute of Education Sciences
U.S. Department of Education
TABLE OF CONTENTS
Section Page
A. Justification of the Study A-1
A.1 Circumstances Making Collection of Information Necessary A-1
A.1.a Purpose of This Submission A-1
A.1.b Legislative Authorization A-3
A.1.c Prior and Related Studies A-3
A.2 Purposes and Use of ELS:2002 A-3
A.2.a Content Justifications A-4
A.3 Improved Information Technology A-5
A.4 Efforts to Identify Duplication A-5
A.5 Methods Used to Minimize Burden on Small Businesses A-6
A.6 Frequency of Data Collection A-6
A.7 Special Circumstances of Data Collection A-8
A.8 Consultants Outside the Agency A-8
A.9 Provision of Payments or Gifts to Respondents A-9
A.10 Assurance of Confidentiality A-12
A.12 Estimates of Hour Burden for Information Collection for the Full-scale Study A-14
A.15 Reasons for Changes in Response Burden and Costs A-16
A.16 Publication Plans and Time Schedule A-16
A.17 Approval to Not Display Expiration Date for OMB Approval A-17
A.18 Exception to Certification for Paperwork Reduction Act Submissions A-17
Exhibits
Number Page
Exhibit A-1. Third Follow-up Technical Review Panel (Research and Policy Community Members) A-8
Exhibit A-2. ELS:2002 Third Follow-up Data Security Plan Outline A-13
Exhibit A-3. Estimated Burden for ELS:2002 Third Follow-up Full-scale Study A-14
Exhibit A-4. Total Costs to NCES A-15
Exhibit A-5. Operational Schedule for ELS:2002/12 Field Test and Full-Scale Activities A-16
LIST OF Appendixes
Appendix 1. Main Study Questionnaire
Appendix 2: Data Collection Materials (Brochure, Lead Letters)
Appendix 3: Cognitive Labs Report Summary
This request concerns the third follow-up of the Education Longitudinal Study: 2002 (ELS:2002), an ongoing longitudinal study with a completed field test in 2011 and a forthcoming full-scale data collection in 2012. This document requests clearance for data collection activities and supplements earlier requests concerned with the 2011 field test, direct locating and contacting of individual respondents or their parents, and two generic clearances for cognitive interviews. Per the field test approval (OMB # 1850-0652 v.7), this submission is subject to a 60-day Federal Register notice waiver. ELS:2002 is being conducted by the National Center for Education Statistics, part of the Institute of Education Sciences, within the U.S. Department of Education. The primary contractor for this study is RTI International (Contract number ED-04-CO-0036/0004).
The materials in this document support a request for clearance to conduct the third follow-up of ELS:2002, which has the following basic components and key design features:
baseline survey of high school sophomores, in spring term 2002;
assessments in reading and mathematics;
surveys of parents, English and math teachers, media center specialists, and school administrators, plus a facilities checklist;
samples sizes of about 750 schools and approximately 17,600 students (15,300 base-year respondents; schools are first-stage unit of selection, with sophomores randomly selected within schools);
oversampling of Asian Americans, private schools;
design linkages (test concordances) with other assessment programs: Program for International Student Assessment (PISA), National Assessment for Educational Progress (NAEP), and test score reporting linkages to the prior longitudinal studies.
follow-up in spring 2004, when most sample members were seniors, but some were dropouts or enrolled in other grades;
student questionnaires, dropout questionnaires, in-school math assessments, and school administrator questionnaires;
returned to the same schools, but separately followed transfer students and those no longer in school by telephone (computer-assisted telephone interview; CATI) or field (computer-assisted personal interview; CAPI);
freshening for a nationally representative senior cohort; and
high school transcript component in fall/winter, 2004–05.
follow-up in 2006 using web-based self-administered instrument, telephone (CATI) and field (CAPI) data collection;
follow-up in 2012 using web-based self-administered instrument, telephone (CATI), and field (CAPI) data collection; and
collection of postsecondary transcripts.
The third follow-up study will provide data to map and understand the outcomes of the high school cohorts’ transition to adult roles and statuses at about age 26. For the cohort as a whole, the third follow-up will obtain information that will permit researchers and policymakers to better understand issues of postsecondary persistence and attainment, as well as sub-baccalaureate (and to a more limited degree, baccalaureate) rate of economic and noneconomic return on investments in education. The third follow-up will also provide information about high school completion (for students who dropped out or were held back) and the status of high school dropouts, late completers, and students who have obtained an alternative credential, such as the GED. Finally, for both college-bound and non–college-bound students, the third follow-up will map their labor market activities and family formation.
For many cohort members, complex pathways, with alternative timings and durations for work and postsecondary enrollment, will be followed at this point of transition. In the 6-year period since the previous round, a sample member may both have worked and attended school, either serially or simultaneously; a cohort member may have attended school part-time or full-time and combined education and work spells with marriage and family formation. The singular strength of longitudinal studies is their power to provide data on transitions that are both complex and of some duration. The transition from adolescence to adult roles—and in particular, the transition to and through postsecondary education, and to labor force activity, and family formation—is of this very type. The timing of the ELS:2002/12 data collection will facilitate capturing all of this complexity at a time when such roles are becoming the norm for many young adults.
ELS:2002 is sponsored by the National Center for Education Statistics (NCES) of the Institute of Education Sciences (IES), in close consultation with other offices and organizations within and outside the U.S. Department of Education (ED).ELS:2002 is authorized under Section 9543 of the Education Sciences Reform Act of 2002 (20 U.S.C.).
In 1970, NCES initiated a program of longitudinal high school studies. Its purpose was to gather time-series data on nationally representative samples of high school students that would be pertinent to the formulation and evaluation of education policies.
Starting in 1972, with the National Longitudinal Study of the High School Class of 1972 (NLS:72), NCES began providing education policymakers and researchers with longitudinal data that linked education experiences with later outcomes, such as early labor market experiences and postsecondary education enrollment and attainment.
Almost 10 years later, in 1980, the second in a series of NCES longitudinal surveys was launched, High School and Beyond (HS&B), which included one cohort of high school seniors comparable to the seniors in NLS:72 as well as a sophomore cohort.
The third longitudinal study of students sponsored by NCES was the National Education Longitudinal Study of 1988 (NELS:88), a cohort of eighth-graders.
The High School Longitudinal Study of 2009, the successor study to ELS:2002, follows a cohort of fall 2009 ninth-graders.
ELS:2002 is designed to monitor the transition of a national sample of young people as they progress from tenth grade through high school and on to postsecondary education and/or the world of work. ELS:2002 has collected data on young people in high school from multiple perspectives; previous waves surveyed parents, teachers, and school administrators. This study follows young adults on many pathways, including dropouts from high school, early high school graduates, college bound graduates, and non-college bound graduates. Because it draws on respondent survey information as well as administrative records such as transcripts, ELS:2002 is able to provide information on the many possible outcomes of secondary education.
ELS:2002 supports both longitudinal and descriptive cross-cohort analyses, although the study is first and foremost a longitudinal study. Survey items are chosen for their usefulness as outcome measures, particularly in the context of previously collected predictor items. ELS:2002 content will be kept comparable to that of the prior NCES high school studies, to facilitate cross-cohort comparisons (for example, trends over time can be examined by comparing 1980, 1990, and 2002 high school sophomores; or 1972, 1980, 1982, 1992, and 2004 high school seniors). The 2012 (third follow-up) round of ELS:2002 can be compared to the year 2000 round of NELS:88, when cohorts from both studies will be, typically, 8 years beyond high school graduation.
The third follow-up interview will focus on postsecondary education, work experiences, family formation, community involvement, and other life course outcomes. It will also include a range of new issues concerning students’ attainment in postsecondary education, the amount of student aid received, and, from college transcripts from all colleges attended, a complete record of all the courses they enrolled in, and the grades they received. New data will also be collected through jobs summary measures on the dynamics of the employment they have entered into and their progress in finding and forming a promising career. The data will have as an important context the special economic circumstances that this cohort has encountered. In addition, special attention will be given to high school dropouts’ progress toward a high school diploma, GED, or other equivalency, including GED test score information. Because some sample members will have chosen not to continue their education in the 8 years following high school, a series of questions will focus on experiences in the workforce. Yet, because another group of respondents will have been going to school and working, work and educational summaries must be collected, covering the 6 years since last interview. In addition to collecting factual information about educational enrollments and work experiences, the interview will collect information on respondents’ basic life goals. As sample members turn 26 years of age, the modal age of the participants at the time of the interview, marriage and parenthood become more common. Therefore, the third follow-up is the appropriate time to determine which participants have started forming families. With regard to community involvement, participation in volunteer work, and the political process will be examined. All outcomes must be collected in this round, in the compass of a relatively brief (35-minute) interview.
The questionnaire is provided in Appendix 1 and a change justifications grid for survey items in Part C. While the content of the field test questionnaires was justified in the approved field test OMB submission, there are some changes in content based on the findings of the field test, cognitive testing, and the deliberations and recommendations of the Technical Review Panel. These changes are of three kinds: some field test items have been deleted, some items have been added, and some field test items have been revised. The grid in Part C documents and justifies these changes.
The same technologically innovative, web-based data collection technology employed in the ELS:2002 second follow-up will be used again in the third follow-up. With this web technology, the resulting survey instruments have been carefully designed to be virtually indistinguishable from each other in terms of screen text and skip patterns across all three modes of data collection: self-administered web, CATI, and CAPI. Expectations are that in the third follow-up, over 40 percent of the responses will be web self-administered. The advantages of a web-based instrument include real-time data capture and access, including data editing in parallel with data collection, and increased efficiencies in effecting timely delivery.
Additional features of the system include (1) online help for selected screens to assist in question administration (in all three modes); (2) full documentation of all instrument components, including variable ranges, formats, record layouts, labels, question wording, and flow logic; (3) capability for creating and processing hierarchical data structures to eliminate data redundancy and conserve computer resources; (4) a scheduler system to manage the flow and assignment of cases to interviewers by time zone, case status, appointment information, and prior cases disposition; (5) an integrated case-level control system to track the status of each sample member across the various data collection activities; (6) automatic audit file creation and timed backup to ensure that, if an interview is terminated prematurely and later restarted, all data entered during the earlier portion of the interview can be retrieved; and (7) a screen library containing the survey instrument as displayed to the respondent (or interviewer).
Since the inception of its secondary education longitudinal program in 1970, NCES has consulted with other federal offices to ensure that the data collected in the series do not duplicate other national data sources. The inclusion on the Technical Review Panels for ELS:2002 both of members of the research community and of other government agencies helps to focus study and instrument design on features of youth transition that ELS:2002 uniquely can illuminate.
ELS:2002 does not duplicate, but temporally extends, the prior NCES longitudinal studies—NLS:72, HS&B, and NELS:88.
Other NCES studies involve assessments of similar age groups to ELS:2002 (PISA 15-year-olds, NAEP eighth-graders and high school seniors), but are not longitudinal, and do not collect data from parents. By the time of the second follow-up (2006, when most sample members were out of high school for 2 years), there is some similarity in sample to the NCES Beginning Postsecondary Students (BPS). However, the BPS longitudinal study focuses only on beginning postsecondary students, including late entrants into the system. In contrast, ELS:2002 includes both cohort members who go on to postsecondary education and those who do not—but misses many late entrants to the system. Thus BPS and ELS:2002 are fundamentally complementary, not duplicative.
The only non-NCES federal study that would appear to be comparable to ELS:2002 is the BLS National Longitudinal Survey of Youth (NLSY)—the NLSY79 and, sampling respondents closer to ELS:2002 in age, the NLSY97 shares with ELS:2002 (and the prior NCES high school cohorts) the goal of studying the transition of adolescents into adult roles. There are also important design differences between NLSY79/ NLSY97 and ELS:2002 that render them more complementary than duplicative. NLSY is a household-based longitudinal survey; ELS:2002 is school-based. NLSY is an age cohort while ELS:2002 is a grade cohort. For both NLSY cohorts, base year Armed Service Vocational Aptitude Battery (ASVAB) test data are available, but there is no longitudinal high school achievement measure. Although NLSY97 also gathers information from schools (including principal and teacher reports and high school transcripts), it cannot study school processes in the same way as ELS:2002, given its household sampling basis. Any given school contains only one to a handful of NLSY97 sample members, a number that constitutes neither a representative sample of students in the school nor a sufficient number to provide within-school estimates. Additionally, ELS:2002 puts more emphasis on postsecondary education, while NLSY stresses labor market outcomes and collects detailed employment event histories. Thus, although both studies provide important information for understanding the transition from high school to the labor market, ELS:2002 is uniquely able to provide information about education processes and within-school dynamics and how these affect both academic achievement and ultimate labor market outcomes.
This section has limited applicability to the proposed data collection effort. Target respondents for ELS:2002 are individuals, and direct data collection activities via web-based self-administration, CATI, and CAPI will involve no burden to small businesses or entities.
The rationale for conducting ELS:2002 is based on a historical national need for information on academic and social growth, school and work transitions, and family formation. In particular, structural changes in the economy; sector changes such as the growth of community colleges; changing youth demography; the continuing need to monitor postsecondary educational access, choice, persistence and attainment; and changes in federal policy concerning postsecondary student support and other interventions necessitate frequent studies. By following the same students over time, longitudinal studies provide better measures of the effects of program, policy, and environmental changes than would multiple cross-sectional studies.
To address this need, NCES began the National Longitudinal Studies Program approximately 40 years ago with NLS:72. This study collected a wide variety of data on students’ family background, schools attended, labor force participation, family formation, and job satisfaction at five data collection points through 1986. NLS:72 was followed approximately 10 years later by HS&B, a longitudinal study of two high school cohorts (10th- and 12th-grade students). NELS:88 followed an eighth-grade cohort, which now, with a modal age of 26 years, represents the final data collection point. With the addition of ELS:2002, a 32-year trend line will be available. Taken together, these studies provide much better measures of the effects of social, environmental, and program and policy changes than would a single longitudinal study or multiple cross-sectional studies.
It could be argued that more frequent data collection would be desirable; that is, there would be a gain in having a program of testing and questionnaire administration that is annual throughout the high school years. However, the 2-year interval was employed with both the HS&B sophomore cohort and NELS:88, and proved sufficient to the realization of both studies’ primary objectives. Although there would be benefits to more frequent data collection in the high school years, it must also be considered that the effect would be to greatly increase the burden on schools and individuals, and that costs would be greatly increased. Probably the most cost-efficient and least burdensome method for obtaining continuous data on student careers through the high school years comes through the avenue of collecting school records. High school transcripts were collected for a subsample of the HS&B sophomore cohort, as well as for the entire NELS:88 cohort retained in the study after eighth grade. A similar academic transcript data collection (covering grades 9 through 12) was conducted for the first follow-up of ELS:2002.
Periodicity of the survey after the high school years (at the very terminus of the study) may also be questioned—there is a 6-year gap between the 2006 round (2 years out of high school) and the final round in 2012 (8 years out of high school). Undoubtedly, more process and postsecondary education context information could be obtained if there were surveys in the intervening years (say at age 22, which would optimally capture the college experience). However, the strategy of waiting until about age 26 for the third follow-up interview is cost-effective, in that the information collected at that time includes both final outcomes and statuses, and provides a basis for identifying the postsecondary institutions that individual sample members have attended. In turn, postsecondary transcripts are then obtained that provide continuous enrollment histories for specific courses taken, and provide records of course grades and other information needed to analyze postsecondary persistence and attainment.
All data collection guidelines in 5 CFR 1320.5 are being followed. No special circumstances of data collection are anticipated.
In recognition of the significance of ELS:2002, several strategies have been incorporated into the project’s work plan that allow for the critical review and acquisition of comments regarding project activities, interim and final products, and projected and actual outcomes. These strategies include consultations with persons and organizations both internal and external to the National Center for Education Statistics, the U.S. Department of Education, and the federal government.
ELS:2002 project staff have established a Technical Review Panel (TRP) to review study plans and procedures. The third follow-up TRP includes some of the earlier ELS:2002 panelists for continuity with prior phases of the study. However, the membership has been reconstituted to reflect the shift in focus from high school experiences to postsecondary and labor market transitions that mark the final outcomes of the study. See Exhibit A-1 for a list of the TRP membership and their affiliations. The TRP met in October of 2010 and in November of 2011, and its recommendations, based on field test results presented at the November 2011 session, have been taken into consideration in revising the instrument for the full-scale study.
Exhibit A-1. Third Follow-up Technical Review Panel (Research and Policy Community Members)
Sara Goldrick-Rab
University of Wisconsin-Madison
1025 West Johnson Street, 575K
Madison, WI 53706
Phone: (608)265-2141
E-mail: SRab@education.wisc.edu
Robert Gonyea
Indiana University
Center for Postsecondary Research
107 S. Indiana Avenue, Eigenmann 443
Bloomington, IN 47405
Phone: (812)856-5824
E-mail: rgonyea@indiana.edu
Donald Heller
The Pennsylvania State University
406 Rackley Building
University Park, PA 16802
Phone: (814) 865-9756
Email: deh29@psu.edu
Robert Lent
University of Maryland
RM 3214D Benjamin Building
College Park, MD 20742
Phone: (301)774-6390
E-mail: boblent@umd.edu
Amaury Nora
The University of Texas at San Antonio
College of Education and Human Development
One UTSA Circle
San Antonio, TX 78249
Phone: (210)458-4370
E-mail: Amaury.Nora@utsa.edu
Randall Olsen
The Ohio State University
921 Chatham Lane, Suite 100
Columbus, OH 43221
Phone: (614)442-7348
E-mail: olsen.6@osu.edu
Aaron Pallas
Columbia University, Teachers College
464 Grace Dodge Hall
New York, NY 10027
Phone: (212)678-8119
E-mail: Amp155@colmbia.edu
Kent Phillippe
American Association of Community Colleges
One Dupont Circle, NW, Suite 410
Washington, DC 20036
Phone: (202)728-0200
E-mail: kphillippe@aacc.nche.edu
Michael Shanahan
University of North Carolina at Chapel Hill
Department of Sociology
CB#3210, Hamilton Hall
Chapel Hill, NC 27599
Phone: (919)843-9865
E-mail: mjshan@e-mail.unc.edu
Marvin Titus
University of Maryland
EDHI
Room 2200 Benjamin
College Park, MD 20742
Phone: (301)405-2220
E-mail: mtitus@umd.edu
Incentive payments to respondents, ranging from $20 to $60, were a major feature of the data collection plan for the ELS:2002 second follow-up study in 2006. About 90 percent of first follow-up respondents and 67 percent of non-respondents were respondents in the second follow-up, for an overall second follow-up weighted response rate of 88 percent (89 percent unweighted). The results of the 2003 field test experiments and the success of the 2004 round of data collection provided evidence of the value of respondent incentives in achieving high response rates (Education Longitudinal Study of 2002: Base-Year to First Follow-up Data File Documentation, NCES 2006-344, Appendix J, Section J3, 2005). In the second follow-up plan, all respondents were given an incentive, with larger incentives for early response and for hard-to-get groups such as past-round nonrespondents and high school dropouts.
In the recently completed third follow-up field test, use of incentives was tied to a response propensity experiment. The details and results of the experiment are presented in part E of this package. In summary, the response propensity model successfully predicted response outcome. The inclusion of low-propensity cases showed an apparent reduction in unit-level biases. Including more low-propensity cases in the data may reduce bias and may help improve final estimates since low-propensity cases appear to be different in terms of their survey responses. A higher incentive amount for low-propensity experiment cases, $45 versus $25 which increased to $55 and $35 in week 10 of data collection, produced an observed – but not statistically significant – 6% higher response rate for the low-propensity experiment cases as compared with the low-propensity control cases. It is not known how the experiment would have concluded if the field test data collection had continued two additional months to its scheduled end. The data collection ended two months early because the interview yield goal was met. Please refer to B.3 for the full set of strategies recommended for full-scale study data collection.
For the third follow-up full-scale study, we propose a base incentive of $25 and strategically targeting certain groups of cases (described below) with modified treatments, including a higher incentive level of $55. The first group identified for modified treatment in the third follow-up is the high school dropouts. Sample members who have ever dropped out of high school are an important analytic group that was targeted for an increased incentive in past rounds (an additional $20 in the second follow-up). In the third follow-up field test, the response rate for those cases who have ever dropped out was 41.1 percent in contrast to the sample members who never dropped out, whose response rate was 59.7 percent (p<.001). Given the interest in this group of cases who have ever dropped out of high school, targeting these sample members is indicated.
In the third follow-up field test, we identified groups of cases for modified treatment by using prior-round information to model response propensities with a goal that response bias be minimized. Because the propensity-modeling plan was able to consider respondent information (such as a full range of response and sociodemographic characteristics) more inclusively and broadly, it was posited that it would also be able to determine which cases would contribute most to bias in estimates, and ensure that these cases receive priority. The details and results of the third follow-up field test experiment are presented in part E of this package.
In light of the field test results and recent discussions between OMB and NCES, ELS:2002 will move toward implementing a survey design with more responsive and adaptive features. ELS:2002 will move beyond response rates and toward a metric that better indicates bias reduction and stability in key estimates. The goal of this new approach will be to produce, when the data are weighted, more precise estimates of less biased key population parameters and population characteristics. Section B.3 contains further detail on the new approach, targeting cases based on a case level Mahalanobis distance function describing a nonrespondent case’s distance from the mean responding case.
ELS:2002 contains three key data collection points at which the Mahalanobis function should be assessed. The three points are: 1) at the conclusion of the early web self-administration phase before CATI begins, 2) when cases are being targeted for CAPI, and 3) prior to sending out a proposed express shipment mailing containing a $5 prepaid incentive near the end of data collection. At each of these points in the data collection, NCES and its contractor will evaluate progress such as the importance of the remaining cases and how different they are from the ones already interviewed. In order to reduce the risk of bias, some of the remaining cases will be designated as cases to specially target. The team will then institute appropriate strategies listed below for targeting special cases. As part of the responsive design being developed for ELS:2002/12, paradata (e.g., call counts, refusal history, contact status, etc.) will be regularly analyzed to help determine where design changes may be beneficial for reducing nonresponse.
Specifically, plans include classifying all cases who have ever dropped out of high school as a special group to target prior to data collection, and developing customized approaches for those cases as well as the special target cases identified at the three points during data collection to reduce the risk of bias. The strategies include:
Pre-data collection intensive tracing
Field tracing/CAPI field interviewing
Overnight express shipment near the end of data collection with $5 prepaid incentive
Increased incentive amount for targeted groups: $55 for targeted groups; $25 for other sample members
Phone interviewing from the start of data collection (i.e., no web-only period). This strategy is proposed for ever-dropouts only. They will be identified prior to data collection.
We recommend a base incentive of $25 for completing the third follow-up interview, which would increase to $55 if the case was targeted for modified treatment as a result of one or more of the three evaluations during data collection. Cases will also be targeted for the additional $5 prepaid incentive mailing near the end of data collection based on the evaluation performed prior to the mailing. Cases who have ever dropped out of high school would receive $55 throughout the data collection as well as the additional $5 prepaid incentive mailing, provided they have not responded to the interview prior to that mailing. The evaluations may result in adjustments to the selection of cases for certain treatments as data collection progresses.
Increased incentives are also indicated for achieving higher response from special groups from a cost perspective. In the second follow-up, sample members who had ever dropped out of high school presented considerable data collection challenges. We offered up to $60 and even with that increased incentive over the base level, that group’s response rate was still proportionately low (83% unweighted response rate for ever-dropouts vs. 89% overall unweighted response rate). The costs of working the case longer in CATI, or moving it on to field work are far greater than increasing the incentive level above the base level. The field test average data-collector labor cost for a completed case in RTI’s Call Center Services was $58 (not including the cost for incentive, mailings, labor other than for data-collectors, or other costs). If incentives yield more self-administered web interviews before the CATI effort begins, then only a small cost per case is incurred in Help Desk labor. Similarly, if incentives enable CATI interviewers to secure an interview earlier in the data collection period than would have been achieved otherwise, then the average dollar amount per complete will decrease. Further, if a case is not interviewed during CATI and the sample member must be contacted and interviewed by field staff, the cost of that additional effort is expensive. In the ELS:2002 second follow-up, the cost for a CAPI completed case was an average of $530 on top of the already-incurred costs. The ELS:2002/12 data collection approach places particular emphasis on tracing including multiple pre-data-collection panel maintenance updates, intensive tracing, field follow-up, and targeted mailings. These efforts are augmented with the incentive plan described above.
A data security plan (DSP) was developed and approved by the computer security review board for the ELS:2002 third follow-up. The ELS:2002 DSP represents best-practice survey systems and procedures for protecting respondent confidentiality and securing survey data. An outline of this plan is provided in Exhibit A-2. The ELS:2002 DSP:
establishes clear responsibility and accountability for data security and the protection of respondent confidentiality with corporate oversight to ensure adequate investment of resources;
details a structured approach for considering and addressing risk at each step in the survey process and establish mechanisms for monitoring performance and adapting to new security concerns;
includes technological and procedural solutions that mitigate risk and emphasize the necessary training to capitalize on these approaches; and
is supported by the implementation of data security controls recommended by the National Institute of Standards and Technology for protecting federal information systems.
Exhibit A-2. ELS:2002 Third Follow-up Data Security Plan Outline
ELS:2002 Data Security Plan Summary Maintaining the Data Security Plan Information Collection Request Our Promise to Secure Data and Protect Confidentiality Personally Identifying Information That We Collect and/or Manage Institutional Review Board Human Subject Protection Requirements Process for Addressing Survey Participant Concerns Computing System Summary General Description of the RTI Networks General Description of the Data Management, Data Collection, and Data Processing Systems Integrated Monitoring System Receipt Control System Instrument Development and Documentation System Data Collection System Document Archive and Data Library Employee-Level Controls Security Clearance Procedures Nondisclosure Affidavit Collection and Storage Security Awareness Training Staff Termination/Transfer Procedures Subcontractor Procedures |
Physical Environment Protections System Access Controls Survey Data Collection/Management Procedures Protecting Electronic Media Encryption Data Transmission Storage/Archival/Destruction Protecting Hard-Copy Media Internal Hard-Copy Communications External Communications to Respondents Handling of Mail Returns, Hard-Copy Student Lists, and Parental Consent Forms Handling and Transfer of Data Collection Materials Tracing Operations Software Security Controls Data File Development: Disclosure Avoidance Plan Data Security Monitoring Survey Protocol Monitoring System/Data Access Monitoring Protocol for Reporting Potential Breaches of Confidentiality Specific Procedures for Field Staff |
Under this plan, the ELS:2002 third follow-up data collection will conform fully to federal privacy legislation, including the Privacy Act of 1974 (5 U.S.C. 552a) and Section 9543 of Education Sciences Reform Act of 2002 (20 U.S.C.). ELS:2002 will also conform to the NCES Restricted Use Data Procedures Manual and NCES Standards and Policies. The plan for maintaining confidentiality includes obtaining signed confidentiality agreements and notarized nondisclosure affidavits from all personnel who will have access to individual identifiers. Each individual working on ELS:2002 will also complete the e-QIP clearance process. The plan includes annual personnel training regarding the meaning of confidentiality and the procedures associated with maintaining confidentiality, particularly as it relates to handling requests for information and providing assurance to respondents about the protection of their responses. The training will also cover controlled and protected access to computer files, built-in safeguards concerning status monitoring and receipt control systems, and a secured and operator-manned in-house computing facility.
Communication materials provided or sent to sample members will include a statement about the voluntary nature of the survey and of the confidentiality provision, stating that their responses may be used for statistical purposes only and may not be disclosed, or used, in identifiable form for any other purpose except as required by law [Education Sciences Reform Act of 2002 (ESRA 2002) 20 U.S.C., § 9573].
The interview contains items about earnings, assets, and debts. Federal regulations governing the administration of these questions, which might be viewed as “sensitive” due to personal or private information, require (a) clear documentation of the need for such information as it relates to the primary purpose of the study, and (b) provisions to respondents which clearly inform them of the voluntary nature of participation in the study, and (c) assurances of confidential treatment of responses. Information about earnings and assets are vital labor force variables and provide important indicators of the rate of return of educational experiences to the respondent.
If a sample member’s SSN is unknown despite the prior rounds of data collection, it will be collected in the interview. This information is needed to obtain data from a variety of extant data sources including student financial aid data from Central Processing System (CPS), data from the National Student Loan Data System (NSLDS) Pell loan and grant files, and GED test results. A description of matching procedures and the security measures in place for the linkages to extant data sources is provided in Part D, while the wording of the SSN question can be found in appendix 1.
Estimates of response burden for the ELS:2002 third follow-up full-scale study interview are shown in Exhibit A-3.The field test interview took on average 37.4 minutes to complete. Based on field test results and Technical Review Panel input, the interview has been trimmed and streamlined (see Appendix 1 for the questionnaire and Part C for a summary of changes), resulting in an estimated overall average interview length of 35 minutes for the full-scale instrument.
Exhibit A-3. Estimated Burden for ELS:2002 Third Follow-up Full-scale Study
|
Sample |
Expected response rate |
Number of respondents |
Average
burden/ |
Range of response times (minutes) |
Total burden (hours) |
Spring 2012 panel maintenance |
16,200 |
20% |
3,240 |
5 |
---- |
270 |
ELS:2002/12 interview |
16,200 |
90% |
14,580 |
35 |
25 to 45 |
8,505 |
NOTE: Table does not include transcript collection which will take place in 2013-14 and will be submitted in a separate package. The table includes the pre-data collection panel maintenance planned for May/June 2012. The table does not include the already-approved full-scale panel maintenance activities (OMB# 1850-0652 v.8) conducted in the fall of 2011.
Assuming a $20 hourly wage, the cost to ELS:2002/12 respondents for completing the survey is estimated at $170,100. Combined with the panel maintenance estimated cost of $5,400, the overall estimate is $175,500.
Included in the notification letter and on the entry page to the online survey will be the following burden statement:
“According to the Paperwork Reduction Act of 1995, no persons are required to respond to a collection of information unless it displays a valid OMB control number. The valid OMB control number of this voluntary information collection is 1850-0652. The time required to complete this information collection is estimated to average around 35 minutes per response. If you have any comments concerning the accuracy of the time estimate or suggestions for improving the interview, please write to: U.S. Department of Education, Washington, DC 20202-4537. If you have comments or concerns regarding the status of your individual interview, write directly to: Education Longitudinal Study (ELS), National Center for Education Statistics, 1990 K Street NW, 9th floor, Washington, DC 20006.”
There are no capital, startup, or operating costs to respondents for participation in the project. No equipment, printing, or postage charges will be incurred.
A.14 Costs to Federal Government
Estimated costs to the federal government for ELS:2002 are shown in Exhibit A-4. Included in the contract estimates are all staff time, reproduction, postage, and telephone costs associated with the management, data collection, analysis, and reporting for which clearance is requested.
Exhibit A-4. Total Costs to NCES
Costs to NCES |
Amount (in $) |
Total ELS:2002/12 costs |
$10,397,075 |
Salaries and expenses |
$750,000 |
Contract costs |
$9,647,075 |
NOTE: All costs quoted are exclusive of award fee. Table does not include transcript collections.
The apparent increase in respondent burden is due to the fact that the last OMB approval was for the ELS:2002 third follow-up field test and field test and full scale panel maintenance activities, while this request is for the ELS:2002 third follow-up full-scale study data collection plus pre-collection full scale panel maintenance activities.
The ELS:2002 contract requires the following reports and other public information releases:
a detailed methodological report (in the form of a comprehensive Data File Documentation Report covering the base year through the third follow-up, with an appendix for the field test) describing all aspects of the data collection effort; and
complete restricted-use and public-use longitudinal data files and documentation for research data users and a First Look Report, presenting initial descriptive findings for dissemination to a broad audience.
Final deliverables for the third follow-up are scheduled for completion in 2013. (Final deliverables for the transcript study are scheduled for completion in 2015.) The operational schedule for the ELS:2002 third follow-up field test and full-scale study is presented in Exhibit A-5.
Exhibit A-5. Operational Schedule for ELS:2002/12 Field Test and Full-Scale Activities
Activity |
Start |
End |
Field test |
|
|
Panel maintenance: contact updates for sample* |
10/2010 |
6/2011 |
First round of cognitive testing of items* |
8/2010 |
9/2010 |
Data collection* |
7/2011 |
9/2011 |
Second round of cognitive testing* |
9/2011 |
10/2011 |
|
|
|
Full-scale study |
|
|
Panel maintenance: contact updates for sample* |
10/2010 |
6/2012 |
Data collection |
7/2012 |
1/2013 |
|
|
|
Transcript collection |
|
|
Pilot testing of operations |
2/2013 |
8/2013 |
Transcript data collection |
8/2013 |
3/2014 |
Transcript keying and coding |
11/2013 |
8/2014 |
* Denotes activities already approved by OMB.
Note: The current request for OMB review includes only data collection activities for the full-scale study.
The expiration date for OMB approval of the information collection will be displayed on data collection instruments and materials. No special exception to this requirement is requested.
No exceptions are requested to the certification statement identified in the Certification for Paperwork Reduction Act Submissions of OMB Form 83-I.
File Type | application/vnd.openxmlformats-officedocument.wordprocessingml.document |
File Title | December 2010 |
Author | cannada |
File Modified | 0000-00-00 |
File Created | 2021-01-31 |