OMB generic clearance request revision 05 12 10

OMB generic clearance request revision 05 12 10.docx

SRS-Generic Clearance of Survey Improvement Projects for the Division of Science Resources Statistics

OMB: 3145-0174

Document [docx]
Download: docx | pdf

REQUEST FOR GENERIC CLEARANCE OF SURVEY IMPROVEMENT PROJECTS

FROM THE DIVISION OF SCIENCE RESOURCES STATISTICS (SRS)




The Division of Science Resources Statistics (SRS) of the National Science Foundation (NSF) requests a three-year extension of the Office of Management and Budget’s (OMB’s) generic clearance that will allow SRS to continue to rigorously develop, test, and evaluate its survey instruments and methodologies. NSF has a mandate to “provide a central clearinghouse for the collection, interpretation, and analysis of data on scientific and engineering resources and to provide a source of information for policy formulation by other agencies of the Federal Government.” This request is part of an ongoing initiative to improve SRS surveys as recommended by both its own guidelines and those of OMB.0


In the last decade, state-of-the art techniques have been increasingly instituted by NSF and other federal agencies, and are now routinely used to improve the quality and timeliness of survey data and analyses, while simultaneously reducing respondents’ cognitive workload and burden. The purpose of this generic clearance is to allow SRS to continue to adopt and use these techniques to improve its current data collections on science, engineering, and technology inputs and outputs. They will be used to improve the content of existing surveys, to aid in the development of new data collections to capture changes in the U.S. science and engineering (S&E) enterprise, and to fill gaps in coverage of the S&E enterprise in the existing SRS portfolio.


Following standard OMB requirements, SRS will submit to OMB an individual request for each survey improvement project it undertakes under this generic clearance. SRS will request OMB approval in advance and provide OMB with a copy of the questionnaire (if one is used) and materials describing the project.


SRS envisions using a variety of survey improvement techniques, as appropriate to the individual projects, such as focus groups, cognitive and usability laboratory and field techniques, exploratory interviews, behavior coding, respondent debriefing, pilot studies, pretests and split-panel tests. SRS has used such techniques in previous activities conducted under generic clearance.


  1. Focus Groups. A qualitative methodology that brings together a small number of relatively homogenous subjects to discuss pre-identified topics. A protocol containing questions or topics focused on a particular issue or issues is used to guide these sessions, and is administered by a trained monitor. Focus groups are useful for exploring and bringing issues to the surface with either respondents or stakeholders. Focus groups are a good choice during the development of a survey or survey topic, when a pre-existing questionnaire or survey questions on the topic do not yet exist. SRS has used focus groups for several projects under generic clearance to assist with redesign of surveys when it became evident that the content of a survey was outdated and did not reflect current issues or the context that respondents were facing.


  1. Cognitive and Usability Laboratory and Field Techniques. A qualitative methodology that refers to a set of tools employed to study and identify errors that are introduced during the survey process. These techniques are generally conducted one-on-one with respondents. Cognitive techniques are generally used to understand the question-response process, whereas usability is generally used to understand the physical features of a survey, for instance, its display and navigational features. In concurrent interviews, respondents are asked to think aloud as they actually answer the survey. In retrospective interviews, respondents answer the survey as they would normally, then ‘think aloud’ afterwards. Other techniques, which are described in the literature and which will be employed as appropriate include: followup probing, memory cue tasks, paraphrasing, confidence rating, response latency measurements, free and dimensional sort classification tasks, and vignette classifications. The objective of all of these techniques is to aid in the development of surveys that work with respondents’ thought processes, thus reducing response error and burden. These techniques are generally very useful for studying and revising a pre-existing questionnaire. SRS has used cognitive and usability testing in previous generic clearance projects to improve existing survey items, to develop new content on existing surveys, and to explore content for new surveys.


  1. Exploratory Interviews. A technique where interviews are conducted with individuals to gather information about a topical area. These may be used in the very early stages of developing a new survey. They may cover discussions related to administrative records, subject matter, definitions, etc. Exploratory interviews may also be used to investigate whether there are sufficient issues related to an existing data collection to consider a redesign. SRS has used such interviews extensively in recordkeeping studies with respondents to several of its establishment surveys to determine both what types of records institutions keep (and therefore what types of information they can supply) as well as where and in what format such records are kept by the institution.


  1. Respondent Debriefing. A technique in which individuals are queried about how they have responded to a particular survey. The purpose of the debriefing is to determine if the original survey questions are understood as intended, to learn about respondents’ form filling behavior and recordkeeping systems, or to elicit respondents’ satisfaction with the survey. This information can then be used (especially if it is triangulated with other information) to improve the survey. This technique can be used as a qualitative or quantitative measurement, depending on how it is administered. This technique has been employed in SRS generic clearance projects to identify potential problems with existing survey items both quantitatively (response behavior study, or RBS, using web survey questions with GSS respondents) and qualitatively (interviews using semi-structured protocols with Academic R&D Survey respondents).


  1. Pilot Studies/Pretests. These methodologies are used to test a preliminary version of the data collection instrument, as was done with the Postdoc Data Project. Pretests are used to gather data to refine questionnaire items and scales and assess reliability or validity. Pilot studies are also used to test aspects of implementation procedures in addition to testing survey measurement issues. The sample may be purposive in nature, or limited to particular groups for whom the information is most needed. Alternatively, small samples can be selected to statistically represent at least some aspect of the survey population.


  1. Split Panel Tests. A technique for controlled experimental testing of alternatives. Thus, they allow one to choose from among competing questions, questionnaires, definitions, error messages, surveys, or survey improvement methodologies with greater confidence than other methods alone. Split panel tests conducted during the actual fielding of the survey are superior in that they support both internal validity (controlled comparisons of variable under investigation) and external validity (represent the population under study). Nearly any of the previously mentioned survey improvement methods can be strengthened when teamed with this method.


  1. Behavior Coding. A quantitative technique in which a standard set of codes is systematically applied to respondent/interviewer interactions in interviewer-administered surveys or respondent/questionnaire interactions in self-administered surveys. The advantage of this technique is that it can quantifiably identify problems with the wording of questions, but the disadvantage is that it does not necessarily illuminate the underlying causes.


SECTION A. JUSTIFICATION

A1. Legal Authority and Circumstances Requiring the Collection of Information

The NSF Division of Science Resources Statistics (SRS) is responsible for collecting, analyzing, evaluating and disseminating information on science, engineering and technology employment, work force, and education, as well as research and development (R&D) funding and performance. In accordance with Sec. 3(a)(6) of the National Science Foundation Act of 1950, as amended, the National Science Foundation (NSF) is directed to “provide a central clearinghouse for the collection, interpretation, and analysis of data on scientific and engineering resources and to provide a source of information for policy formulation by other agencies of the Federal Government.” SRS publishes data in individual reports and in such general reports as Science and Engineering Indicators and Women, Minorities and Persons with Disabilities in Science and Engineering.


An extension to SRS’ previously granted generic clearance is requested for several reasons. As a federal statistical agency, SRS is engaged in a process of continuous improvement in the data collections it conducts. Critical to the improvement in existing surveys is the ability to engage in small scale projects to test alternatives to current approaches being utilized in the surveys. Generic clearance authority substantially enhances SRS’ ability to engage in such testing and exploration. Furthermore, as the world continues to change, SRS must continuously evaluate its surveys in light of these changes. Respondent behaviors will change (e.g., response rates decrease over time); technology will change (e.g., the web quickly became a data collection option); and the S&E enterprise will change (e.g., today’s students increasingly pursue multi/interdisciplinary studies rather than a single discipline). Similarly, the understanding of how to improve surveys continues to evolve (e.g., today’s research continues to update the interpretation of the best implementation for web surveys).


In addition, SRS has one project in progress under the existing generic clearance (3145-0174) that has not been completed. Below is the task under the current clearance that needs renewal.

Project SRS Contact

SRS Generic Clearance; Postdoc Data Project Jeri Mulrow


Thus, SRS needs an OMB generic clearance structure to continue improving the overall quality of its statistical surveys, reduce the burden on respondents to SRS surveys, and shorten the time required for SRS to update and improve its data collections.


A2. Purposes and Use of the Information

The information obtained from these efforts will be used to develop new NSF surveys and improve current ones. Specifically, the information will be used to reduce respondent burden and to improve the quality of the data collected in these surveys. These objectives are met when respondents are presented with plain, coherent, and unambiguous questionnaires asking for data compatible with respondents’ memory and/or current reporting and recordkeeping practices. The purpose of the survey improvement projects will be to ensure that SRS surveys are continuously attempting to meet these standards of excellence.


Improved NSF surveys will help policy makers in decisions on R&D funding, graduate education, scientific and technical workforce, innovation, as well as contributing to increased agency efficiency and reduced survey costs. In addition, methodological findings have broader implications for survey research and may be presented in technical papers at conferences or published in the proceedings of conferences or in journals.


A3. Use of Information Technology to Reduce Burden

SRS will employ information technology, as appropriate, to reduce the burden of respondents who agree to participate in its survey improvement projects. Many respondents of current SRS surveys supply email addresses that can be used to recruit respondents for survey improvement projects. This allows respondents to communicate with SRS at their convenience. Respondents to current SRS surveys of academic institutions can often provide addresses for websites with additional information (e.g., about their schools), once again reducing their workload. SRS will continue to explore state-of-the-art technology to find ways to reduce burden on respondents to both individual and establishment surveys. For example, SRS used desktop sharing and teleconferencing software to conduct usability testing in remote locations for the redesign of the Academic R&D Survey. By using this software to digitally record both comments and web screen interactions, SRS had a complete record of each session that made it unlikely that there would be a need to call respondents back to clarify notes of the sessions.


Web surveys facilitate accurate data by providing respondents with automated tabulations and feedback on inconsistent answers. These features potentially reduce the need for followup contact with respondents. However, the success of these features resides in their being well designed to ensure that respondents are aware of these features. Thus, one focus of SRS improvement activities is improving the usability of SRS web surveys. SRS improvement projects help ensure that respondents are presented with the most user friendly and least burdensome survey instruments possible. In addition, SRS continues to explore the adoption of innovative methods that could reduce respondent burden.


A4. Efforts to Identify Duplication

Survey improvement projects will be conducted both to improve existing surveys as well as to develop new SRS data collections. The SRS data collections themselves are subject to great scrutiny to ensure there is no duplication of other efforts. Likewise, the projects conducted under the generic clearance authority will be structured in order not to duplicate other efforts. Furthermore, generic clearance activities help avoid repetitious efforts to hone implementation for the full-scale surveys.



A5. Provisions for Reducing Burden on Small Organizations

One goal of SRS’ efforts to improve its surveys is to minimize the burden on the small organizations that respond to SRS surveys. By learning about organizational and recordkeeping practices of small, medium, and large organizations, SRS is in a better position to design surveys and procedures that minimize the survey burden for various types of respondents—especially, small and very small entities. For example, SRS is investigating the best methods to survey companies with fewer than five employees about their R&D activities.


In the case of pilot studies or split-panel tests, if probability samples are utilized, sampling rates proportional to size are often used to make sure that a large institution has a higher probability of being selected than a small institution. This ensures that a high proportion of the attribute of interest—U.S. S&E funding, performance, employment, or education—is captured while minimizing the burden on small entities.


A6. Consequences of Not Collecting the Information

Numerous consequences would result if SRS could not conduct the survey improvement projects requested in this document. The quality of the data collected would decrease because the current surveys would not be systematically evaluated and updated to better reflect the current state of S&E. Over time, surveys that are now well designed would eventually become obsolete. New survey items and procedures would be implemented without adequate testing and refinement. Advances in understanding of how organizations/individuals answer surveys and how SRS can better serve respondents would be curtailed. And, SRS’ ability to develop timely, new, and well-designed surveys would be diminished.


A7. Special Circumstances for Collection

Under this clearance, SRS will explain any circumstances that would result in respondents being required to:


  • Report information to the agency more often than quarterly;


  • Prepare a written response to a collection of information in fewer than 30 days after receipt of it;


  • Submit more than an original and two copies of any document;


  • Retain records, other than health, medical, government contract, grant-in-aid, or tax records for more than three years;


  • Respond to a statistical survey in a manner that is not designed to produce valid and reliable results, i.e., results that can be generalized to the universe of study;


  • Use a statistical data classification that differs from one approved by OMB;


  • Respond in a manner that includes a pledge of confidentiality that is not supported by authority established in statute or regulation, that is not supported by disclosure and data security policies that are consistent with the pledge, or which unnecessarily impedes sharing of compatible data with other agencies for confidential use;


  • Submit proprietary trade secret or other confidential information unless the agency can demonstrate that it has instituted procedures to protect the information’s confidentiality to the extent permitted by law.



A8. Federal Register Notice and Consultation Outside the Agency

Comments on this data collection effort were solicited in the Federal Register, Vol. 74, No. 247, on Monday, December 28, 2009 (see Attachment A). SRS received no comments that were deemed relevant to this request.


The primary objectives of the survey improvement projects include involving respondents in the development of new survey content, soliciting respondent reactions to current surveys, observing respondent navigation of survey forms, and exploring how respondents’ recordkeeping systems work. These objectives focus on consultation with respondents to reach the goals of understanding (1) how to minimize the time and effort to complete survey tasks, (2) how to reduce other aspects of burden such as concerns about the use of the survey data, and (3) how to motivate respondents to provide survey answers that have the highest quality and most accuracy.


A9. Remuneration to Respondents

SRS and its contractors sometime provide compensation to participants in survey improvements projects. In some cases, the compensation covers travel costs only. In other cases, compensation is offered in the range of $25-$75 for activities such as focus groups or cognitive interviews. This practice has proven necessary and effective in recruiting some types of subjects to participate in this small-scale research, and is also employed by other federal agencies.  Focus groups and cognitive interviews are sometimes conducted in contractors’ cognitive laboratories or other comparable facilities.


Unless otherwise specified and with approval granted by OMB, such incentives will be limited to no more than $40 for participation in a cognitive interview and no more than $75 for participation in a focus group.  Respondents for field test activities such as split sample tests, behavior coding of interviewer/respondent interaction, and respondent debriefings will receive payment only when there are extenuating circumstances that warrant it. Generic clearance packages for projects offering participant compensation will explain the rationale and describe the amounts of compensation.


A10. Assurance of Confidentiality Provided to Respondents

Respondents in the survey improvement projects will be advised that their participation is voluntary. In focus groups, interviews, and other respondent activities, SRS may ask respondents for permission to record sessions via audio or video recording. Such recordings are conducted to provide project staff, including those not present at the activity, with a complete and accurate record to supplement note taking. Recording the session also allows staff to focus more on what is taking place during the session rather than on the completeness of their notes. In some cases, recordings may be used to train others to conduct this type of research or for illustrative purposes in presentations to professional audiences. For sessions that are recorded, respondents will be asked for their consent to the audio or video recording. They will be notified if there is any chance that a session may be played for audiences for research purposes.

Often activities conducted under the generic clearance authority will not involve pledging confidentiality to participants. However, there will be instances where confidentiality will be pledged, such as where the information being requested, either about an individual or an organization, may be sensitive. The pledge of confidentiality will be made under the Privacy Act (where applicable) and the National Science Foundation Act of 1950 (as amended). Specifically, when confidentiality is pledged to individuals, the pledge used will be the following:

The information is solicited under authority of the National Science Foundation Act of 1950 as amended. All information you provide is protected under the NSF Act as amended and the Privacy Act of 1974 as amended and will only be used for research or statistical purposes. Any information publically released such as statistical summaries will be in a form that does not personally identify you.

When confidentiality is pledged to organizations, such as businesses, colleges and universities, and other non-profit organizations, the pledge used will be the following:

The information is solicited under authority of the National Science Foundation Act of 1950 as amended. All information you provide is protected under the NSF Act as amended and will only be used for research or statistical purposes. Any information publically released such as statistical summaries will be in a form that does not personally identify you or your organization.

A11. Questions of a Sensitive Nature

No questions of a sensitive nature are anticipated in work conducted under this generic clearance. However, the nature of the exploration of survey content may include asking respondents whether items might be considered sensitive in the context of data collection.

A12. Estimate of Response Burden

SRS estimates that a total reporting burden of 14,280 hours over the three years of the requested generic clearance will result from working to evaluate/improve existing surveys and to develop new ones. This includes both the burden placed on respondents participating in each activity as well as burden imposed on potential respondents during screening activities. Table 1 provides a list of potential surveys for which generic clearance activities might be conducted, along with estimates of the number of respondents and burden hours that might be involved in each.


Table 1. Potential surveys for improvement projects, with the number of respondents and burden hours




Survey Name

Number of Respondents0


Hours

Graduate Student Survey


1,5000

2,500


SESTAT Surveys

4,000

2,000


Postdoc Project

2,000

2,500

New and Redesigned R&D Surveys

Higher Education R&D

Government R&D

Nonprofit R&D

Business R&D

Microbusiness R&D




400

60

100

50

150



1,200

180

300

150

450

Survey of Scientific & Engineering Facilities

300


300

Public Understanding of S&E Surveys


200

50

Survey of Earned Doctorates

700

450

Additional surveys not specified

1,600

4,200


Total

11,060

14,280

A13. Estimate of Total Cost to Respondents

The cost to respondents generated by the list of potential projects is estimated to be $585,480 over the three years of the clearance. No one year’s cost would exceed $585,480. In other words, if all work were done in one year, costs in that one year would be $585,480 and the costs in each of the other 2 years would be zero. As in previous requests for generic clearance authority, the total cost was estimated by summing all the hours that might be used on all projects over the three years (14,280) and multiplying that figure by the hourly wage ($41.00) of the level of employee who typically answers SRS’ questionnaires or attends SRS workshops. This wage amount is the May 2008 national cross-industry estimate of the mean hourly wage for a financial analyst, or Job Category 13-2051, by the Bureau of Statistics. http://www.bls.gov/oes/oes_dl.htm#2003_m. The total hours are based on similar SRS projects over the past few years.


There are no capital, startup, operation or maintenance costs to the respondents. The costs generated by future data collections will be described in the clearance request for each specific data collection. SRS does not anticipate any capital, startup, operation, or maintenance costs for future surveys.



A14. Estimates of Annualized Costs to the Federal Government

The 3-year cost to the Federal government generated by the survey improvement projects is estimated to be approximately $2,000,000. The main components of these costs are contractor costs and staff time. There are no startup, equipment, operations or maintenance costs. Bidders on the SRS contracts are required to have all software, licenses, and hardware needed to complete the survey improvement projects. The costs generated by future data collections will be described in the clearance request for each specific data collection.


A15. Changes in Burden

The request for burden hours (14,280) is slightly less than the current generic clearance, which requested 14,950.


A16. Plans for Publication

Data will be collected to develop new surveys or improve the content or methodology of current surveys. Methodological findings from survey improvement projects may be referenced in the technical notes for published survey data, in methodology reports, in technical papers presented at conferences, in the proceedings of conferences, or in journals. Generic clearance activities will not be used to calculate substantive results/estimates that will be released.


A17. OMB Approval Expiration Date

SRS will display the expiration date for OMB approval of the information collection on survey instruments.


A18. Exceptions to the Certification Statement

No exceptions to the Certification Statement should be required. If so, OMB approval will be requested in advance of conducting the survey.




0 NSF Information Quality Guidelines are available on http://www.nsf.gov/policies/infoqual.jsp. OMB Information Quality Guidelines are available on http://www.whitehouse.gov/omb/inforeg/infopoltech.html. OMB standards and guidelines for statistical surveys are available on http://www.whitehouse.gov/omb/inforeg/statpolicy/standards_stat_surveys.pdf.


0 Number of respondents listed for any individual survey may represent several methodological improvement projects.

0This number refers to the science, engineering, and health-related departments within the academic institutions of the United States (not the academic institutions themselves).

7


File Typeapplication/vnd.openxmlformats-officedocument.wordprocessingml.document
File TitleREQUEST FOR GENERIC CLEARANCE OF
Authormvburke
File Modified0000-00-00
File Created2021-01-31

© 2024 OMB.report | Privacy Policy