Development of an Evaluation Protocol for Assessing Impacts of OMH State Initiatives
Request for Modifications to OMB Clearance Previously Issued for the OMH Uniform Data Set (UDS)
May 25, 2007
Submitted by:
Development Services Group, Inc.
7315 Wisconsin Ave., Suite 800E
Bethesda, MD 20814
(301) 951‑0056
TABLE OF CONTENTS
Background
A. Justification
A.1 Explanation of the Circumstances that Make the
Data Collection Necessary
A.2 Use of the Data
A.3 Consideration of the Use of Improved Information Technology
A.4 Efforts to Identify Duplication
A.5 Minimizing Burden on Small Businesses and Entities
A.6 Consequences of Less Frequent Data Collection
A.7 Special Circumstances of Data Collection
A.8 Consultation with Persons Outside the Agency
A.9 Payment to Respondents
A.10 Assurances of Confidentiality
A.11 Sensitive Questions
A.12 Estimates of Respondent Burden
A.13 Annual Cost Burden to Respondents
A.14 Cost Estimates
A.15 Changes in Burden
A.16 Tabulation and Publication Plans
A.17 Display of Expiration Date
A.18 Exception to Certification Statement
B. Collection of Information Employing Statistical Methods
APPENDIX
A. Inventory of Data Items in Uniform Data Set
Background
Since its inception in 1985, OMH has been the unit of the U.S. Department of Health and Human Services (HHS) that coordinates Federal efforts to improve the health status of racial and ethnic minority populations. The agency was established with the passage of the Disadvantaged Minority Health Improvement Act (Pub. L. 101‑527) and given a broad mandate to advance efforts to improve minority health and address racial/ethnic disparities in health (recently reauthorized). In order to achieve this broad mission, OMH supports research, demonstrations and evaluations of new and innovative programs, and strategies and interventions that increase understanding of ways to improve the health of minority communities and reduce the burden of disease, disability, and premature death that disparately impacts them.
As part of HHS, OMH works under an overall policy and strategic framework that includes key goals set forth in Healthy People 2010 and the 1993 Government Results and Performance Act or GPRA (Public Law 103‑62). Specifically, GPRA introduced requirements for performance measurement and benchmarking for Federal agencies. While performance measurement has been traditionally associated with cost factors and efficiency, e.g., the cost or number of hours per unit of output (Hatry, 1989), GPRA applied performance measurement to nonfinancial sectors of public management (Osborne and Gaebler, 1993), thus the increasing reliance by Federal agencies on strategic plans with goals and standards (such as the HHS Strategic Plan/Performance Goals). GPRA emphasizes public accountability and requires the development of strategic plans, performance goals, and an annual report filed with Congress on actual measured performance compared with pre‑established goals. More recently, OMB has implemented another accountability process, known as the Program Assessment Rating Tool (PART), designed to assess strategic planning, evaluation, and evidence of results for Federal agencies and their programs.
As part of efforts to improve program monitoring and performance measurement with regard to its grant programs, OMH worked with the contractor to develop a system to collect standardized data on project activities from all of its grantees and cooperative partners. This work encompassed several efforts. The first was the Study of the Bilingual/Bicultural Service Demonstration Grant Program, focusing on grantees funded by OMH in fiscal years 1993B95 (hereinafter referred to as the Bilingual/Bicultural Study). This program was designed to improve bilingual/bicultural assistance in providing health services to designated racial and ethnic minority populations. The methodology involved: the empanelling of an expert, multicultural Advisory Committee; reviewing all grantee records; developing and administering a mail survey to FY 1993B94 grantees (including a telephone follow‑up interview); and conducting site visits to FY 1995 grantees. A key issue in conducting this study was the lack of standardized data across the spectrum of these projects, making it difficult to measure impact. Thus the resulting report included 12 recommendations, one of which was to develop the Uniform Data Set (AUDS Development@ project).
Following that report, the contractor undertook development of the UDS through an extensive process of consultation with agency staff, grantees and cooperative agreement partners, and an Advisory Group composed of representatives from both the public (Federal agencies) and private (associations and foundations) sectors. A final paper-based UDS instrument and a limited, demonstration Web-based UDS prototype were developed as the outcome of those efforts. This UDS system was subsequently developed, under the recent UDS Implementation project (called “Implementation of an Internet and Paper-Based Uniform Data Set/UDS for all OMH Grants and Cooperative Agreements”), into a full Web-based system.
It is important to note that the UDS Development project has received significant recognition for excellence. Out of 26 evaluation projects submitted for consideration, the UDS Development project was one of only eight HHS evaluation projects nominated for and selected to be highlighted in the 2003 annual evaluation report to Congress. The project was summarized in the Highlights/Excellent Evaluation chapter of Performance Improvement 2003, HHS=report to Congress on FY2002 ongoing and completed evaluation efforts. The project was deemed excellent by non-Federal evaluation experts on the basis that it was well designed, the developed product included data to measure the project=s qualitative impact, and that it contributed significantly to the goals of GPRA.
The UDS is now being (or has been) used for reporting by the following types of OMH grants: Bilingual/Bicultural Service Demonstration Grant Program; Community Programs to Improve Minority Health; HIV/AIDS Health Promotion and Education Program; State and Territorial Minority HIV/AIDS Demonstration Grant Program; Youth Empowerment Demonstration Grant Program; Minority Community Health Partnership HIV/AIDS Demonstration Grant Program; and Technical Assistance and Capacity Development Demonstration Grant Program for HIV/AIDS-Related Services in Minority Communities. The proposed modifications discussed herein are intended to adapt the UDS so that it can be used to report program and performance data for activities undertaken by States under the OMH State Initiative and grantees funded under the UCA program.
Current UDS Data Elements
The UDS includes routinely reported project/program uniform data, including: 1) core data, consisting of organizational and program information; and 2) activity data, consisting of project data pertinent to the 19 activity modules, which are standardized categories of project activity (e.g., health education and outreach) for which systematized, standard data are collected. Any project that conducts a given activity will report data pertinent to that activity. Apart from the core data, the UDS asks grantees only for data on the activities they conduct.
Core Data Elements: This refers to a limited, minimum data set across all OMH-funded programs and activity types, and includes very basic, largely descriptive and organizational data collectible across projects:
grant type;
grantee organization type;
project location and environment (region, urban, rural, etc.);
project services/activity types;
OMH‑specific funding and staffing information;
OMH-specific partnerships and collaborations; and
other similar information.
B. Activity Type Data: As noted, data categorization by activity type is necessary because each different program funds a number of activities per project, such that comparison of even two projects under the same program may not be possible, if done solely by project. Reporting data by activity type allows activities across projects to be evaluated, and for activity types to be assessed against each other. Thus, a comprehensive typology of project activities was developed. Table 1 shows the kinds of indicators/data elements that will serve to measure activities and the current list of “activity categories” and their definitions is included in Tables 2a and 2b.
Data Element for Activity |
Description |
Demographics |
Characteristics (e.g., age, gender, race, and ethnicity) of the persons served under the specific activity. |
Number served |
A count of the number of people/organizations served by the project while conducting the activity. |
Process |
The number of times a particular activity (e.g., training) was conducted, the number of materials developed and disseminated, the number of referrals made, organizations contacted, etc... |
Short term outcome |
Short‑term gains made as a result of conducting the activity (e.g., gains in knowledge, awareness or skills, short‑term changes in diet, referrals made after health screenings). |
Long‑term outcome |
Data on long-term outcome will not be available for many activity categories. Where possible, however, such data will be collected. Examples include actual changes over the project period in specific health indicators (e.g., blood pressure, body fat), or changes in health risk behaviors (e.g., diet, smoking) that occurred as a result of the activity. |
System changes |
Changes in health systems or delivery of health for the target population that came about because of the activity (e.g., local task forces formed, new policies put in place at local health providers, new units formed). |
Qualitative data |
This type of data is intended to allow projects to record or describe (in a brief format) how the particular activity impacts clients in a way that is not necessarily reflected in the other data that is reported. It is a place for each project to "tell the story" of this activity as they see it. |
The current set of 19 activity categories is divided into services impacting individuals and services impacting organizations or systems. The entire set of categories is displayed in Tables 2a and 2b.
Activity Category |
Description |
1) Training health care providers |
Structured instruction or education sessions administered to health providers (doctors, nurses, staff) on issues related to racial/ethnic minority health (this includes development/adaptation of a training curriculum). |
2) Interpretation/ Translation |
Interpreting for minority clients who do not speak English (or do not speak it well) while clients are at appointments with health providers, case managers, social services providers, and other related services or in connection with (or translation of) written materials that clients need to understand and fill out. In addition, this involves interpreting what the client is saying for the health provider. |
3) Target population health education and outreach |
Any kind of educational activity (whether individual or group) concerning health information, promotion, and prevention directed to the minority target population(s), including workshops, home health parties, health fairs, and outreach. Violence, substance abuse and delinquency prevention should be included under this category. (NOTE: Materials development and dissemination are counted separately under Activity Category #4.) |
4) Target population health educational materials development/ dissemination |
Original development or adaptation of educational materials (concerning health issues, risks, prevention, treatments, local health providers) to be provided to minority and/or under‑served populations, and dissemination of those materials, via outreach, presentations, in doctor's offices, etc. This category is divided into print, audio and TV/video materials. |
5) Screening and referral |
Any diagnostic screening procedure to detect health risk or presence of a health problem (e.g., cholesterol screening, mammograms, blood pressure, and others). Referrals means that, as a result of the screening, the client was referred to actual medical services. |
6) Case management |
Planning, coordination and monitoring of specific prevention or treatment protocols for individual clients (e.g., through the use of individual treatment plans or other tools), and coordinating referrals and access to services based on the individual protocol. |
7) Wellness and exercise |
For the purpose of this database (to distinguish wellness from education), any structured physical/mental activities designed to address a health risk, such as exercise classes, stretching, meditation, yoga, and others. |
8) Academic support/career preparation |
Any structured activity designed to improve client educational performance (e.g., remedial sessions, tutorial, academic enrichment) and improve skills/knowledge related to jobs/careers. (NOTE: Organizational internships and staff development programs are to be included under Activity Category #16, Technical Assistance and Organizational Capacity Building.) Related to prevention of health risk behaviors related to violence and substance abuse. |
9) Mentoring |
Structured, ongoing adult‑youth relationships for a specified period of time in which the adult spends regularly scheduled time with the youth and provides support, information, encouragement, and role‑modeling to help prevent a range of health risk behaviors. |
10) Parent skills training/family counseling |
Any structured activities (e.g., workshops, regular meetings, classes, family counseling, and family education) with parents as participants, in which parents receive instruction and/or practice in parenting and family management skills. |
11) Self‑esteem building |
Any structured activities (workshops, sessions, curriculum segments) designed to increase client self‑esteem. |
12) Cultural activities |
Any structured activities including workshops, curriculum segments, sessions, ceremonies/rituals or field trips whose primary purpose is to increase client youth awareness of their cultural background, and increase bonding to cultural background as part of their sense of identity. |
13) Recreational/ sports |
Any structured athletic or physical activities, outdoor or indoor. |
14) Crisis intervention |
Unplanned activities, directed to clients and/or the target community, that are deemed necessary in order to resolve a crisis, or solve a problem that poses a barrier to client participation in project activities, or to help solve a problem that is diverting attention from the health issues of importance to the project. |
15) Conference planning and management |
The planning and conduct of conferences, meetings, trainings and other events that require logistical support activities, materials development (e.g., agendas, notebooks), and on-site management and follow-up. |
Activity Category |
Description |
16) Linkage‑building or community coordination |
Linkage‑building is a broad category that includes the formation and maintenance of coalitions, the formation of referral arrangements, and any other activity that is specifically part of grant activities and is intended to expand the impact of the grantee by developing and maintaining linkages with other organizations and/or health providers. |
17) Technical assistance and organizational capacity building |
Activities that increase the capacity of the grantee organization and/or any of its linkage partners to provide improved services, improved health promotion, and improved access to health care for its target minority population(s). This could mean staff training, hiring bilingual staff, instituting an internal training program or training materials, implementing internal policies designed to ensure the growth of cultural competence, mentoring client organizations, leadership development, adding administrative resources, technology, staff, or new systems, and other similar actions. |
18) Resource coordination |
The identification, coordination, and facilitation of resources to aid minority/community health organizations in expanding access to prevention and health care for target minority populations. |
19) Needs assessment |
The conduct of a community or target population needs assessment as part of project activities (not as part of preparing a grant application). The needs assessment may include surveys of community attitudes and practices, available services, and other factors which would inform the targeting of project activities. |
The full instrument (data elements only) is included in Appendix A. As noted in the original application for clearance, the UDS has been pilot-tested, in both paper and Internet forms. According to the original pilot test data, participants (a selection of OMH grantees) typically reported data in two or three activity modules, for an average reporting time of 4.5 hours, as indicated in Section A.12. The participants were also asked for feedback on the reporting instrument. The majority of participants felt that the UDS was easy to use, and that they were also able to use the UDS for their own organizational needs and to help them organize their own data.
OMH uses this information for project management and performance monitoring. Aggregate data from the UDS will be included in reports to OMH leadership and other policy/decision makers. Individual project data will be available to that project=s key staff as well, who will be able to use it to improve the overall management of their projects, to enhance their service delivery capacity, and to assess project impacts.
References
Hatry, Harry. 1999. Performance Measurement: Getting Results. Washington, D.C.: Urban Institute Press.
Osborne, D., and T. Gaebler. 1992. Reinventing Government: How the Entrepreneurial Spirit is Transforming the Public Sector. Reading, MA: Addison-Wesley.
A. Justification
A.1 Explanation of the Circumstances that Make Additional Data Collection Necessary
The Office of Minority Health (OMH), OPHS, U.S. Department of Health and Human Services (HHS) is submitting a Request for OMB Review in support of modifications to the OMB-approved UDS for a project entitled “Development of an Evaluation Protocol for Assessing Impacts of OMH State Initiatives,” hereinafter referred to as the “State Initiative.” The UDS received OMB clearance in March 2004 (OMB No. 0990-0275).
The modifications to the UDS presented in this request are intended to:
Continue the process of integrating all OMH-funded activities under a systematic, uniform reporting mechanism to maximize efficiency; and
Continue the development of the UDS as a reporting system that will capture the types of data needed to identify best practices and assess the progress of OMH-funded activities towards the achievement of OMH-specific goals and Healthy People 2010 goals.
The UDS was developed based on OMH’s portfolio of grant programs current at the time of development. Two grant programs not included in the research to develop the UDS need to begin reporting through the system. These are the State Initiatives to Eliminate Racial/Ethnic Disparities in Health program and the National Umbrella Cooperative Agreement (UCA) program. Once the modifications are implemented, the UDS will serve as the regular system for reporting of program management and performance data for all active OMH-funded grant and cooperative agreement programs.
Additional modifications are requested as part of OMH’s efforts to improve the agency’s evaluation and planning capacities and compliance with Federal reporting requirements. These modifications include the addition of selected impact measures. Reporting and analyzing such data will allow the identification of best practices and evaluation of effective approaches. The ability to monitor and evaluate performance in this manner and to work towards continuous program improvement are basic functions that OMH must be able to accomplish in order to carry out its mandate with the most effective and appropriate use of resources.
Requested modifications therefore include both additions/changes to the data reported in order to increase the number of programs using the UDS (including the State Initiative and UCA grantees), and additions/changes intended to increase the reporting of program impacts. These changes will improve OMH evaluation and planning capacities and support program accountability.
The steps to adapt the UDS are:
Review program documents and determine whether program activities undertaken by State and UCA grantees can be reported within the categories of data in the current UDS.
If they cannot, identify the modifications that need to be made.
Where possible, add impact data items to the existing set of data to be reported. The State Initiative (under which this request is submitted) includes activities related to the adaptation of the UDS as well as activities designed to develop an Evaluation Protocol for OMH-funded State efforts to eliminate racial/ethnic health disparities. Impact data were selected in collaboration with this effort.
Make the necessary modifications to the UDS (text and programming)
Pilot test the modified UDS with a small sample of grantees
Upon OMB approval, train new grantees and implement the system.
The requested modifications to the UDS are detailed in Table 3. [The complete set of data items in the UDS, including the proposed modifications, is included as Appendix A.] These modifications are based only on the review of State Initiative grantee program documents; no revisions were deemed necessary to adapt the UDS for UCA grantees.
Module |
Recommended Modification(s) |
Core Project Data |
Below Question 10, add the following question:
If new staff were hired, were they:
|
Training Health Care Providers (Module 1) |
Change name of module to “Training and education for health professionals and community stakeholders” to accommodate educational activities for individuals such as community leaders who are not health care providers, and individuals who are not receiving training as part of an organizational capacity-building effort (covered under Module 17). |
|
Add the following response choices to the “type of training” in Sections II and IV:
|
|
In Section III, add the following question:
Who attended your training/education sessions (e.g., health care providers, community leaders, CBO staff member, etc…)? |
Language Interpretation (Module 2) |
In Section II, after Question 1, add the following question:
How many clients accessed services as a result of your language interpretation services? |
Target population health education and outreach (Module 3) |
In Section II, Question 2, change the follow-up question to read:
If yes, please complete the following: [add Table 3-3 to capture, by activity type (individual or group education),
|
|
Add new Section IV: Short-Term Outcomes
For those education sessions where trainee outcome was evaluated: Was it with (check one)?
What was evaluated (check all that apply)?
If Pre and Post Tests: Add Table 3-6 to capture, by type of training (individual or group):
|
Target population health educational materials development and dissemination (Module 4) |
Change title of module to “Materials development and dissemination.” |
|
Revise the choices for type of media (print, audio, TV/video, web) to reflect a blend of type of media and material. We recommend the following list of types:
|
|
After Table 4-2, add the following questions:
If you developed a Web site or disseminated materials on the Web: How many Web site hits did you have? How many materials were downloaded from your Web site? |
Screening and referral (Module 5) |
Add column to Table 5-2:
Screening Site:
|
|
Add column to Table 5-3:
Number of Successful Referrals |
Case Management (Module 6) |
Add Table 6-3: Number of Clients Receiving Services Through Case Management By Type of Service
Type of Service
|
Wellness and Exercise (Module 7) |
In Section II, Question 1, add blood glucose test and weight/BMI change as evaluation criteria. |
Academic support and career preparation (Module 8) |
In Section I, Table 8-2 and Section II, Question 1, change types of activity to:
|
|
In Section I, add Table 8.3 to capture, by activity type:
Program Issue Addressed:
Education Level of Participants
Number of Participants Number of New Participants Recruited in this Reporting Period |
|
In Section II, add the following questions:
Did any participants apply to or gain acceptance into medical school, other health service training programs, or programs in the health sciences? If yes, how many individuals submitted applications? How many applications were accepted? |
Mentoring (Module 9) |
Add Section III Short-Term Outcomes
For those sessions where participant outcome was evaluated: Was it with (check one)?
What was evaluated (check all that apply)?
Add Table 9-2: If Pre- and Post-Test
|
Parent skills training (Module 10) |
In Section II, Question 1, change the topics of evaluation to:
|
Self-esteem building (Module 11) |
In Section II, Question 1, change the topics of evaluation to:
|
Cultural activities (Module 12) |
In Section II, Question 1, change the topics of evaluation to:
|
Recreational/sports (Module 13) |
Add Section II: Short-term Outcomes
For those activities where participant outcome was evaluated: Was it with:
What was evaluated (check all that apply)?
If Pre- and Post-Tests Add Table 13-2 to capture by type of activity (sports, other recreational):
If standardized tests were used, please list the names of the test(s) |
Linkage-building and community coordination (Module 16) |
In Section II, add the following questions:
Did you form any new coalitions or collaborations in the past reporting period?
Add Table 16-3 to capture for each collaboration:
For those coalitions or collaborations you formed or participated in, how many times did they meet?
Were any of these collaborations part of ongoing task forces or committees? If yes, how many times did they meet? If yes, are there plans for this partnership to continue meeting? If no, did the partnership complete its goals? |
|
In Sections I and II, add the following to the list of response choices for “role in grant activity”:
|
Technical assistance and organizational capacity building (Module 17) |
Add “strategic planning for internal improvement” to list of types of TA provided. |
|
In Table 17-2, add “grantee organization” to list of response choices for “organization type” |
|
Change follow-up question in Section II, Question 3 to table to include information on the:
|
Resource coordination (Module 18) |
Add the following questions:
Did you provide mini-grants to organizations as a project activity? If Yes, please describe.
|
Needs assessment (Module 19) |
Change title of module to “Planning and Evaluation.” |
|
Change Question 1 to read: Which of the following methodologies were employed in your planning and evaluation activities (check all that apply)? |
|
Change Question 2 to read: Did your planning and evaluation activities address specific health conditions? If yes, which health conditions were addressed? |
|
Add the following question:
Did your planning and evaluation activities address specific populations? If yes, which populations were addressed? |
|
Change Question 3 to read: Which of the following areas were addressed in your planning and evaluation activities? |
|
Change Question 4 to read: What were the main findings or results of your planning and evaluation activities? |
|
Add the following question:
Were data collected for planning purposes or to target resources? If yes, please describe. |
|
Add the following question:
Did you implement any changes in the data collection (such as collecting new kinds of data, enhancing data technology) to improve internal data systems? If yes, please describe. |
|
Change Question 5 to read: Does your project address gaps or problems identified through your planning and evaluation activities? If yes, please describe. |
|
Add the following questions:
Did you evaluate efforts funded under your grant? If yes, please describe. Were your evaluation criteria related to goals or other targets in your strategic plan? If yes, please describe. |
Pilot-Test Summary
Protocol. After the modifications to the Web-based system were completed, a pilot-test of the modified UDS was conducted with a small sample of State grantees. The contractor worked with OMH to select grantee sites to contact for the pretest. The sample included 3 State Initiative grantees. UCA grantees were not included in the pilot test as no specific changes were needed to accommodate this grant program in the UDS.
Each grantee in the sample was contacted by email and phone to introduce the UDS effort and invite participation in the pretest. The grantees participating were:
Delaware Office of Minority Health, Division of Public Health, Department of Health and Social Services
Michigan Office of Minority Health, Department of Community Health
New Mexico Office of Policy and Multicultural Health, Department of Health
Since the current UDS has already been extensively tested with OMH grantees, a complete test of the system was not necessary. The pilot test focused primarily on modules that include new, untested questions. These were:
Module 4 – Materials development and dissemination
Module 8 – Academic support/career preparation
Module 16 – Linkage-building and community coordination
Module 18 – Resource coordination
Module 19 – Planning and evaluation
An account was set up in the UDS for each participant. Three activities were selected for each grantee based on our need to test activity modules and the specific activities conducted by the grantee project. An evaluation form was provided for feedback on several aspects of the UDS. Participants were asked to assess:
The ease of use of the UDS
The time needed to complete a report through the UDS
Whether it would take more or less time to report through the UDS
Whether a report could be completed using data that are already collected, or if new data would be needed
The applicability of the data items/questions in the UDS to their project activities
A conference call was held on May 16, 2007 to train the grantees on the basic use and functions of the UDS. During this session, grantees were given a brief discussion of the background and structure of the UDS, training on data entry, and instruction on the protocol for the pilot test. Participants were instructed to:
Enter provided project data into activity module(s) and submit a report
Utilize the online help/technical assistance functions
Notify the Deputy Project Director upon completion of the report
Complete and return the pilot test evaluation form within 1 week of the training conference call
Follow-up instructions, sample data, and the evaluation form were sent to participants following the training session. Phone and online technical assistance was offered as necessary during the pilot test period. Data submissions were monitored and checked for completeness and accuracy. Evaluation forms were received and reviewed.
Data accuracy: Data submissions from each participant were accurate and complete.
Ease of use: Two participants said the UDS was “easy” to use and the remaining participant said it was “very easy” to use. [Note: the response choices were: very easy, easy, somewhat easy, somewhat difficult, difficult, and very difficult.]
Time to complete a report: The participants said it would take an average of 2 hours to complete a report using the UDS. This is significantly less than the average of 4.5-hour estimate given by participants during the original pilot test of the system.
UDS reporting vs. standard reporting: Two participants said that it would take “about the same time” to complete a report and the third said it would take “a bit less time.” [Note: the response choices were: significantly less time, a bit less time, about the same time, a bit more time, and significantly more time.]
Data needs: All of the participants said that they already collect the data required in the UDS and would not have to collect additional data to complete a report.
Applicability of data items: Generally-speaking, the participants felt that the questions were applicable to their project activities. They did, however, say that they would not be able to report information on community meetings, advisory council meetings, and speaker’s bureau presentations using the selected activity modules.
Recommendations. No additional changes are recommended based on the results of the pilot-test. Activity Module 15 (Conference/meeting planning) is designed to capture information on community meetings and advisory council meetings. This activity module was not tested during the pilot test because it would not change under this request for modification. Since the UDS is already able to collect the information identified by grantees above, no further modifications are needed.
A.2 Use of the Data
The overall purpose of the UDS – for the State Initiative as well as current UDS applications -- is to enable the OMH-funded grantees, cooperative agreement partners and others to routinely report uniform data to a central coordinating center where the uniform data will be routinely received, analyzed and coordinated into reports to 1) monitor the project=s status, and 2) generate information regarding best practices, program inputs, outcomes and return on investment. As noted, the UDS has already received OMB approval. This application requests approval to modify the UDS so that OMH-funded State and UCA grantees can utilize the system.
This is the second reporting year of full implementation of the OMB-approved UDS with OMH grantees and cooperative agreement partners. The UDS has been used to generate reports at two levels. First, OMH program officers use the system to review individual grantee reports and aggregate reports on projects in their grant stream and to improve the overall management of their projects. Secondly, The UDS has generated program data on populations served, health issues addressed, and funding that has been used to respond to inquiries made to OMH leadership and other policy/decision makers.
Tables 4 and 5 display the State and UCA grantees currently funded by OMH who will be using the UDS upon approval of this request.
State/Territory |
Grantee |
ALABAMA |
Department of Public Health, Minority Health Section |
ARIZONA |
Department of Health Services, Center for Minority Health |
CALIFORNIA |
Department of Health Services, Office of Multicultural Health |
COLORADO |
Department of Public Health & Environment, Office of Health Disparities |
DELAWARE |
Department of Health & Social Services, Division of Public Health |
FLORIDA |
Department of Health, Office of Minority Health |
GEORGIA |
Department of Community Health, Office of Minority Health |
ILLINOIS |
Department of Public Health, Center for Minority Health Services |
INDIANA |
State Department of Health, Office of Minority Health |
MAINE |
Department of Health & Human Services, Office of Minority Health |
MARYLAND |
Department of Health & Mental Hygiene, Office of Minority Health/Health Disparities |
MASSACHUSETTS |
Department of Public Health, Office of Multicultural Health |
MICHIGAN |
Department of Community Health, Health Disparities Reduction & Minority Health Program |
MINNESOTA |
Department of Health, Office of Minority & Multicultural Health |
MISSISSIPPI |
Department of Health, Office of Health Disparity |
MISSOURI |
Department of Health & Senior Services, Office of Minority Health |
NEBRASKA |
Health & Human Services System, Office of Minority Health |
NEVADA |
Department of Health & Human Services, Office of Minority Health |
NEW HAMPSHIRE |
Department of Health & Human Services, Office of Minority Health |
NEW JERSEY |
Department of Health & Senior Services, Office of Minority & Multicultural Health |
NEW MEXICO |
Department of Health, Office of Policy & Multicultural Health |
NEW YORK |
Department of Health, Office of Minority Health |
NORTH CAROLINA |
Department of Health & Human Services, Office of Minority Health & Health Disparities |
OHIO |
Commission on Minority Health |
OREGON |
Department of Human Services, Office of Multicultural Health |
PUERTO RICO |
Department of Health, Office of External Affairs/Office of Minority Health |
RHODE ISLAND |
Department of Health, Office of Minority Health |
SOUTH CAROLINA |
Department of Health & Environmental Control, Office of Minority Health |
TENNESSEE |
Department of Health, Office of Minority Health |
TEXAS |
Department of State Health Services, Office for the Elimination of Health Disparities |
UTAH |
Department of Health, Center for Multicultural Health |
WISCONSIN |
Department of Health and Family Services, Minority Health Program |
Grantee |
University of Notre Dame du Lac, Inter-University Program for Latino Research |
Association of American Indian Physicians |
National Minority AIDS Council |
Association of Asian Pacific Community Health Organizations |
Auxiliary to the National Medical Association, Inc. |
Hispanic Association of Colleges and Universities |
Quality Education for Minorities |
The President and Fellows of Harvard College |
Interamerican College of Physicians and Surgeons |
These projects represent a wide variety of health conditions, target populations, State systems, community situations, and project modalities. As noted, the UDS has been configured as an activity‑based, core and module data set. To minimize respondent burden, a given project will provide “core data” together with data only for those “activity categories” that are applicable to the project. These data are discussed in detail above under “UDS Data Elements.” Thus cross‑site comparisons of project performance and implementation can be made in the only way feasible given project variety, by comparing data by activity category. The activity categories were developed in close consultation with OMH and OMH-funded grantees and cooperative agreement partners, building upon already developed tools. They are an attempt to obtain reasonable documentation of process and outcome (intermediate outcome) for each category, through the standardized reporting of both quantitative and qualitative data.
A.3 Consideration of the Use of Improved Information Technology
The UDS is specifically designed as an Internet application. There will always be paper-based forms available; however, there have been no requests for the UDS in this format from grantees. The Internet system was selected for implementation as a result of pilot-testing the paper-based UDS. The participants= sole and unanimous suggestion on improving the UDS was to provide it as a computer-based system. A wide range of grantees preferred the Internet over the paper format for the UDS as the tool to minimize burden and maximize utility. Virtually all grantees had the technical capability to access the Internet for this purpose.
Many community grantees have limited staff time and capacity to collect and report their project data, and the Web-based UDS provides a range of features that augment their capacity and thus improve the quality and regularity of reported data. For example, the Web-based format offers:
Well-defined, easily understandable data items for reporting
Explanation boxes for every data item in the system
On-line technical assistance
Downloadable reported data records and forms to help with ongoing data reporting
A “Frequently Asked Questions” (FAQs) function
Automatic edit checks to “flag” possible errors or inconsistencies
Usability by OMH-partners
A.4 Efforts to Identify Duplication
Reporting data under the UDS system does not duplicate other data reporting, but instead replaces the previous data reporting process. It is the first such system implemented at OMH. Prior to the UDS, a small amount of project data was reported on a voluntary basis using a limited, though standard, form and each project provided evaluation data based on its project-specific evaluation format. The latter data were not, however, standardized across projects. The UDS incorporated the data items included on the voluntary form, and standardized evaluation/performance data for all projects B thus creating an entirely new data system that supersedes that which previously existed.
In
addition, during the UDS Development project, an extensive effort was
made to identify any other uniform data systems
among Federal agencies or private foundations that were duplicative,
and found none. In fact, that effort was also intended to identify
any similar data items included in these data sets or methodology so
that the OMH data could conform to cross-agency standards where
applicable. In our review, only a few such data items were
identified. The definitions utilized for these data items were
adopted in the initial development of the UDS in order to maximize
the utility of UDS data for cross-agency comparison where possible,
as well as to avoid re-inventing the wheel.
A.5 Minimizing Burden on Small Businesses and Entities
Grantees and cooperative agreement partners funded by OMH are, for the most part, public or private non-profit minority community based organizations, as well as State agencies (State Offices of Minority Health or other appropriate agency). Whether or not the UDS existed, these projects would have to provide some project and evaluation data. The modifications recommended here represent the minimum data needed to be useful for project reporting, program monitoring, and performance measurement by OMH and its partners.
The UDS was specifically designed to provide additional support for OMH partners in order to facilitate their efforts to report data, and to standardize and simplify the nature of reported data. Moreover, the UDS was designed to provide additional capacity to grantees and cooperative agreement partners in ways that they do not now have available, by providing online technical assistance, online guidance in filling out data forms, and downloadable sample forms and worksheets to help each project report its data. In addition, the UDS system has built-in edit checks to identify inconsistencies and errors in the data entered into the system, thus bypassing the need to use valuable staff time to accomplish such tasks. In short, the UDS system includes many features that both minimize respondent burden and increase respondent capacity.
A.6 Consequences of Less Frequent Data Collection
UDS data is reported every six months, as it was under the previous reporting system. UDS reporting is required for all new OMH grantee and cooperative agreement partners. The grantees these modifications are for would have to report every six months, regardless of whether the UDS existed. The modifications proposed here do not change the frequency of project reporting by OMH partners.
There are no legal obstacles to reduce the burden of this data collection.
A.7 Special Circumstances of Data Collection
This request fully complies with the regulation.
A.8 Consultation with Persons Outside the Agency
The agency’s 60-day notice appeared in the Federal Register Thursday, May 24, 2007, Vol.72, No. 100, pp. 29166 as required by 5 CFR 1320. (d). No public comments were received in response to the notice. There was no outside consultation on the UDS modifications presented in this request; however, the original OMB-approved UDS was developed with extensive input from industry specialists in public and private sectors, as well as OMH grantees and cooperative agreement partners.
A.9 Payment to Respondents
This study does not involve payment or gifts as incentives for respondents.
A.10 Assurances of Confidentiality
Only aggregate, periodic project data from each project or other agency activity is reported. Nevertheless, these data are password-protected and each project manager selects a unique ID once they begin using the UDS. That ID is necessary in order to log on to the Internet system, and it allows access only to that project=s own records.
All data is maintained in aggregate form at a Central Coordinating Center supported under contract to OMH. This Center manages operations of the UDS and provides reports to OMH as requested.
A.11 Sensitive Questions
The UDS requests data on program structure and characteristics, program operation, program implementation, services provided numbers/types of clients served, and short-term or intermediate outcomes. These data do not include any items of a personal or sensitive nature.
A.12 Estimates of Respondent Burden
It is estimated that the hour burden for this project will be approximately 9 hours per OMH partner per year, which is the sum of the hour burden for the OMH partners to report the data via UDS (regular gathering of data is not included in this estimate because that is already a customary part of project activities). While reporting such data is a regular part of project activities, it will be systematized under the new UDS and automated at the Central Coordinating Center. Depending upon the number of activities a given project conducts, and the number of clients served, we estimate the following time ranges for completing regular UDS requirements:
In terms of time burden, the day-to-day gathering/recording of data is not different under the UDS system from what it has been under the previous data management system. Projects already gather data; it is just not systematized as it will be for the UDS.
For routine data reporting through the UDS, we estimate that each OMH-funded entity will spend approximately 4.5 hours to report the data each period, estimating the OMH partner has two to three activity modules to complete. This estimate is based on the results of a small pilot test during the UDS development effort where grantees were trained to use the system and used it to submit a routine report. With semiannual reporting frequency, this calculates to approximately 9 hours total time burden for the task over a year.
An adjustment was reported on OMB Form 83-I. This adjustment represents an adjustment from the original PRA submission in the estimated number of responses per year. The original number of responses per year per respondent was estimated to be 4. In implementation, the number of responses per year per respondent was changed to 2, reducing the total burden hours by approximately one half.
Table 6. Estimated Annualized Burden Hours
Type of Respondent |
Form Name |
No. of Respondents |
No. Responses per Respondent |
Average Burden per Response (in hours) |
Total Burden Hours |
OMH Grantee |
UDS |
150 |
2 |
270/60 |
1350 |
Table 7. Estimated Annualized Cost to Respondents
Type of Respondent |
Total Burden Hours |
Hourly Wage Rate |
Total Respondent Costs |
Project or Evaluation Manager |
1350 |
$30.00 |
$40, 500.00 |
A.13 Capital Costs
As with the OMB-approved UDS, the modified UDS described in this request does not constitute an additional effort for respondents beyond regular project duties/obligations. No additional staff time or cost is anticipated other than the time/cost allocated for regular project administrative requirements. No additional materials or equipment are needed to use the UDS.
A.14 Cost Estimates
The overall cost to the Federal government for modifying the UDS for use with OMH-funded State grantees will be $71,582, distributed according to the following categories:
Category |
Cost |
|
|
Personnel |
$69,887 |
Other Direct Costs (including travel, consultants, computer equipment, etc.) |
$1,695 |
Total |
$71,582* |
*G&A and fee included in total.
This includes the completed modification and testing of the UDS system, training of and technical assistance for State and UCA grantees in its use, and ongoing operation of the system (e.g., processing data, monitoring, generating reports).
A.15 Changes in Burden
This is a request for a modification to an existing, OMB-approved data collection. The modifications include additional questions intended to capture the nature and extent of activities conducted by grantees under funding programs that were not active during the development of the UDS. There will be a slight increase in burden as the system is implemented among these new grantees - the burden of learning to use the system. Following that, no additional burden or changes in burden are anticipated.
The modifications covered under this request represent minimal additional burden on existing users of the UDS. The activity modules that would be changed most with this revision apply to a minority of existing users. Additional impact measures were included in this revision above and beyond what was needed to accommodate State and UCA grantees. As a result, the number of data items to respond to in any activity module may increase by one or two for some grantees.
A.16 Tabulation and Publication Plans
The purpose of the UDS, as described herein, is to serve as the regular, ongoing system of data reporting for all grants and cooperative agreements funded by OMH. Data reported as part of this system will be used for project management and monitoring, assessment of project implementation and performance, and to identify best practices and approaches in support of OMH goals and the goals of Healthy People 2010 and HealthierUS. In addition, aggregate program data will also be used in periodic reports to OMH leadership and other HHS policymakers and decision makers as needed and appropriate.
No specific plans to publish results from the UDS system are underway at this time.
A.17 Display of Expiration Date
This section does not apply to this submission.
A.18 Exception to Certification Statement
There are no exceptions to the certification.
B. Collection of Information Employing Statistical Methods
This section does not apply to the UDS. The project does not involve sampling. All OMH grantees/cooperative agreement partners will be reporting uniform data on their activities using this system to analyze performance and identify best practices.
Appendix A
Data Items in Modified Uniform Data Set
DATA ITEM |
RESPONSE |
Organization Name |
|
Grant Award Year |
|
Address |
|
Phone/Fax |
|
Contact Person/ Phone/Email |
|
Organization Key Code |
|
Organization Type |
|
Select if reporting for multiple programs |
|
Project Name |
|
Project Director/ Email |
|
Contact Person |
|
Number of Positions (FTE's) Filled Using OMH Funding |
|
Number of OMH-Funded Staff |
|
Number of Consultants |
|
Number of Individuals Paid on a Fee-For-Service Basis (e.g., interpreters paid per interpretation) |
|
Number of New Staff Hired |
|
If new staff were hired, were they: |
|
Number of Volunteers |
|
Current Grant Year |
|
Grant Number |
|
Grant Type |
|
Total Annual Budget of Grantee Organization |
|
OMH Funding |
|
What additional funding did you receive to conduct your OMH-funded activities? Federal Funding (amount) State Funding (amount) Local Funding (amount) Private Funding (amount) In-Kind Contributions (amount) |
|
How were your OMH funds distributed across health issues, activities, and demographic categories? |
TABLE (for each category, enter) |
Health Issues |
Select Health Issue/ Enter Percent of Funding Used |
Activities |
Select Activity Modules/ Enter Percent of Funding Used |
Race |
Select Race/ Enter Percent of Funding Used |
Ethnicity |
Select Ethnicity/ Enter Percent of Funding Used |
Gender |
Select Gender/ Enter Percent of Funding Used |
Age |
Select Age/ Enter Percent of Funding Used |
What other activities does your organization do that are not funded by OMH? (Note: This question only applies to grantees receiving funding through the State Partnership Initiative) |
Enter Other Activities Funded and Funding Source |
Were you involved with any partnerships or collaborating organizations as an essential part of the project? |
TABLE (for each partnership, enter) |
Name of Organization |
|
Type of Agreement |
Select:
|
Type of Organization |
Select |
Role in Project Activity |
Select:
|
Postal zip codes where your project conducts its activities |
|
Project Environment |
Check all that apply:
|
Report Information |
|
Project Name |
|
Reporting Period |
|
Report Narrative |
TEXT /Attach Document |
Activities Conducted |
Select Activity Modules (checkbox) |
DATA ITEM |
RESPONSE |
Section I: Number of Individuals Trained and Sessions Conducted |
|
Table 1-1: Number of Individuals Served and Demographics |
Demographic Characteristics of Individuals Served |
Section II: Number of Sessions Conducted |
|
Type of Training |
|
Number of Sessions |
|
Total Served in All Sessions |
|
Length of Each Session in Hours |
|
Evaluated? |
Yes/No |
Section III: Additional Training Information |
|
What were the training topics? |
|
Who attended your training/education sessions? (e.g., health care providers, community leaders, CBO staff member, etc…) |
|
Section IV: Short-term Outcomes of Training and Education |
|
For those trainings where trainee outcome was evaluated |
|
Was it with |
|
What was evaluated (check all that apply)? |
|
If Pre and Post Tests |
TABLE (for each type of training, enter) |
Type of Training |
|
Number of People who took Pre Tests |
|
Number of People who took Post Tests |
|
Number with Increase In Score from Pre- to Post-Test |
|
Section V: Qualitative Impacts |
|
Please describe how your trainings have impacted on three sample trainees. To fill out this section, you can draw from evaluation responses, conversations with or observations of trainees, your own notes, or your experience with trainees |
|
DATA ITEM |
RESPONSE |
Section I: Number of Individuals Served |
|
Table 2-1: Number of Individuals Served and Demographics |
Demographic Characteristics of Individuals Served |
Section II: Sessions Conducted and Short-term Outcomes |
|
1. Please enter the total number of interpretations provided by language and the percentage of clients that received a follow-up health/medical referral or assessment as a result of language interpretation. |
TABLE (for each language, enter) |
Language |
Select from list |
Total Interpretations |
|
Total Clients Served |
|
Total Providers Served |
|
Number Receiving Referral/Assessment |
|
2. How many clients accessed services as a result of your language interpretation services? |
|
3. What was the average duration of each session of language interpretation? |
hour(s) |
4. What was the average amount of preparation or other additional time (e.g., transportation time, waiting room time, etc.) per session? |
hour(s) |
5. Did you translate any materials as part of the service you provided? |
Yes/No |
For each language, enter total number of materials |
|
6. Please list the kinds of materials you translated |
|
7. Did you provide any simultaneous translation for group sessions or meetings? |
Yes/No |
If yes, for each language, enter: |
|
Number of Sessions |
|
Approximate Number of People Per Session |
|
Section III: Qualitative Impacts |
|
1. Please describe how the interpretations you provide have impacted on three sample clients. To fill out this section, you can draw from client evaluation responses, conversations with or observations of clients, notes, or your experience with clients. |
|
DATA ITEM |
RESPONSE |
|
|
Section I: Number of Individuals Served and Sessions Conducted |
|
Table 3-1: Number of Individuals Served and Demographics |
Demographic Characteristics of Individuals Served |
Table 3-2: Number of Sessions Conducted |
TABLE (For each type of session, enter) |
Type of Session |
|
Number Of Sessions |
|
Number Of Sessions Per Course |
|
Number Of Courses Conducted |
|
Evaluated? |
Yes/No |
Section II: Additional Information |
|
1. What were the education session topics? |
|
For Individual Education |
|
For Group Education |
|
2. During the course of your health education and outreach activities, were any clients given referrals to medical, mental health, or other services? |
Yes/No |
If yes, how many referrals were given? |
|
How many of these clients accessed services as a result of referrals? |
|
Section III: Health Fairs and Other Events |
|
1. Did you conduct or participate in any health fairs during this reporting period? |
Yes/No |
If YES, what is the total number of health fairs conducted/participated in? |
|
|
TABLE (for each health fair enter) |
Target Population |
|
Health Issue(s) |
|
Approximate Number Served |
|
Date: (MM/DD/YYYY) |
|
2. Did you conduct or participate in any type of educational event other than those reported above (examples, performing arts, rallies, walks/runs, benefit events)? |
Yes/No |
If YES, what is the total number of other events conducted/participated in? |
|
|
TABLE (for each other event enter) |
Event Type |
|
Target Population |
|
Health Issue(s) |
|
Approximate Number Served |
|
Date: (MM/DD/YYYY) |
|
Section IV: Short-term Outcomes of Health Education and Outreach |
|
For those education sessions where trainee outcome was evaluated |
|
Was it with |
|
What was evaluated (check all that apply)? |
|
If Pre and Post Tests |
TABLE (for each type of training, enter) |
Type of Education |
|
Number of People who took Pre-Tests |
|
Number of People who took Post-Tests |
|
Number with Increase In Score from Pre- to Post-Test |
|
Section V: Qualitative Impacts |
|
1. Please describe how your health education and outreach activities have impacted on three sample clients. To fill out this section, you can draw from evaluation responses, conversations with or observations of clients or members of the target population, your own notes, or your experience with clients. |
|
DATA ITEM |
RESPONSE |
|
|
Section I: Number of Individuals Trained and Sessions Conducted |
|
Table 4-1: Number of Individuals Served and Demographics |
Demographic Characteristics of Individuals Served |
Table 4-2: Materials Development |
TABLE (For each material developed, enter)
|
Type of Material |
|
Source |
|
Target Audience |
|
Health Issue |
|
Language |
|
Number Developed |
|
If you developed a Web site or disseminated materials on the Web: |
|
How many Web site hits did you have? |
|
How many materials were downloaded from your Web site? |
|
Section II: Qualitative Impacts |
|
1. For each type of material you developed/adapted, please describe how the language and graphics are appropriate for the intended targeted audience and how you determined this. |
|
2. What kinds of organizations and/or individuals received, heard or saw the materials you developed? |
|
DATA ITEM |
RESPONSE |
|
|
Section I: Number of Individuals Served and Sessions Conducted |
|
Table 5-1: Number of Individuals Served and Demographics |
Demographic Characteristics of Individuals Served |
Table 5-2: Number of Screenings Conducted |
TABLE (for each type of screening, enter) |
Type of Screening |
|
Number of Screenings |
|
Screening Site |
|
Table 5-3: Number of Referrals Given |
TABLE (for each type of referral, enter) |
Type of Referral |
|
Number of Referrals |
|
Number of Successful Referrals |
|
Section II: Qualitative Impacts |
|
1. Please describe how your work providing screenings and referrals has impacted on three sample clients. To fill out this section, you can draw from client evaluation responses, conversations with or observations of project clients, notes, or your experience with clients. |
|
DATA ITEM |
RESPONSE |
|
|
Section I: Number of Individuals Served |
|
Table 6-1: Demographic Characteristics of Individuals |
Demographic Characteristics of Individuals Served |
Table 6-2: Number and Type of Case Management Contacts (With Clients) |
|
Total Number of Case Management Contacts: In-Person |
|
Total Number of Case Management Contacts: By Telephone |
|
Table 6-3: Number of Clients Receiving Services Through Case Management By Type of Service |
|
Type of Service |
|
Number of Clients Receiving Services |
|
Section II: Qualitative Impacts |
|
1. Please describe how your case management activities have impacted on three sample clients. To fill out this section, you can draw from (non-confidential) case notes, client evaluation responses, conversations with or observations of clients, other notes, or your experience with clients. |
|
DATA ITEM |
RESPONSE |
|
|
Section I: Number of Individuals Served |
|
Table 7-1: Demographic Characteristics of Individuals Served |
Demographic Characteristics of Individuals Served |
Table 7-2: Number of Sessions Conducted |
|
Type of Class |
|
Total Number of Sessions |
|
Average Number of Participants Per Session |
|
How many individuals received individual physical/wellness training? |
|
Section II: Short-term Impacts |
|
1. Were the wellness/exercise participants evaluated using pre-post tests or screenings? |
Yes/No |
If Yes |
TABLE (for each activity enter) |
Type of Wellness Activity |
|
Evaluation Method |
|
Number of People Taking Pre-Test |
|
Number of People Taking Post-Test |
|
Number of People with Improved Score From Pre- to Post-Tests |
|
Section III: Qualitative Impacts |
|
1. Please describe how your wellness activities have impacted on three sample clients. To fill out this section, you can draw from client evaluation responses, conversations with or observations of clients, notes, or your experience with clients. |
|
DATA ITEM |
RESPONSE |
|
|
Section I: Number of Individuals Served |
|
Table 8-1: Demographic Characteristics of Individuals Served |
Demographic Characteristics of Individuals Served |
Table 8-2: Number of Sessions Conducted |
TABLE (for each type of session, enter) |
Type of Activity |
|
Total Number of Sessions |
|
Average Number of Participants Per Session |
|
Evaluated? |
Yes/No |
Table 8-3: Program Information |
|
Type of Activity |
|
Program Issue Addressed |
|
Education Level of Participants |
|
Number of Participants |
|
Number of New Participants Recruited in this Reporting Period |
|
Section II: Short-term Outcomes |
|
Did any participants apply to or gain acceptance into medical school, other health service training programs, or programs in the health sciences? |
Yes/No |
If yes, how many individuals submitted applications? |
|
How many applicants were accepted? |
|
For those sessions where participant outcome was evaluated: |
|
Was it with |
|
What was evaluated (check all that apply)? |
|
If Pre and Post Tests |
TABLE (for each type of session, enter) |
Type of Activity |
|
Number of People Who Took Pre-Tests |
|
Number of People Who Took Post-Tests |
|
Number of People Who Took STANDARDIZED Pre-Tests |
|
Number of People Who Took STANDARDIZED Post-Tests |
|
Number of People with Increase in Score From Pre- to Post-Tests |
|
If standardized tests were used, please list the names of the test(s) |
|
Section III: Qualitative Impacts |
|
1. Please describe how your work in academic support/career preparation has impacted on three sample clients. To fill out this section, you can draw from client evaluation responses, feedback from teachers/school personnel, conversations with or observations of clients, notes, or your experience with clients. |
|
DATA ITEM |
RESPONSE |
Section I: Number of Individuals Served |
|
Table 9-1: Demographic Characteristics of Individuals Served |
Demographic Characteristics of Individuals Served |
Section II: Additional Information on Mentoring |
|
1. What was the average length of the mentoring relationship (months)? |
|
2. Typically, what was the frequency of face-to-face contact between mentors and mentees? times per week times per month |
|
3. Typically, what was the frequency of telephone contact between mentors and mentees? times per week times per month |
|
4. How many mentors were involved in your project activities? |
|
Section III: Short-term Outcomes Mentoring |
|
For those sessions where participant outcome was evaluated: (If no sessions were evaluated, skip to section III) |
|
Was it with: |
|
What was evaluated (check all that apply)? |
|
If Pre- and Post-Test |
|
Number of People Who Took Pre- Tests |
|
Number of People Who Took Post- Tests |
|
Number of People Who Took STANDARDIZED Pre-Tests |
|
Number of People Who Took STANDARDIZED Post-Tests |
|
Number of People with Increase in Score From Pre- to Post-Tests |
|
If standardized tests were used, please list the names of the test(s) |
|
Section IV: Qualitative Impacts |
|
1. Please describe how your work providing mentoring has impacted on three sample clients. To fill out this section, you can draw from client evaluation responses, feedback from teachers/school personnel, conversations with or observations of clients, notes, or your experience with clients. |
|
DATA ITEM |
RESPONSE |
Section I: Number of Individuals Served |
|
Table 10-1: Number of Individuals Served and Demographics |
Demographic Characteristics of Individuals Served |
Section II: Number of Sessions Conducted and Other Information |
|
Total Number of Sessions Conducted: Individual Counseling |
|
Total Number of Sessions Conducted: Group Session or Class |
|
1. What was the average duration of the individual counseling? hours per session total sessions per person |
|
2. What was the average duration of the group sessions? hours per session total sessions per person |
|
Section III: Short-term Outcomes of Parent Skills Training/Family Counseling |
|
For those sessions where participant outcome was evaluated: (If no sessions were evaluated, skip to section III) |
|
Was it with: |
|
What was evaluated (check all that apply)? |
|
If Pre- and Post-Test |
TABLE (for each type of activity, enter) |
Type of Activity |
|
Number of People Who Took Pre- Tests |
|
Number of People Who Took Post- Tests |
|
Number of People Who Took STANDARDIZED Pre |
|
Number of People Who Took STANDARDIZED Post |
|
Number of People with Increase in Score From Pre |
|
If standardized tests were used, please list the names of the test(s) |
|
Section IV: Qualitative Impacts |
|
1. Please describe how your parenting skills training/family counseling activities have impacted on three sample clients. To fill out this section, you can draw from client evaluation responses, (non-confidential) case notes, conversations with or observations of training clients, other notes, or your general experience with clients. |
|
DATA ITEM |
RESPONSE |
Section I: Number of Individuals Served and Sessions Conducted |
|
Table 11-1: Demographic Characteristics of Individuals Served |
Demographic Characteristics of Individuals Served |
Table 11-2: Total Number of Sessions Conducted by Type of Activity |
|
Individual Sessions (Total) |
|
Group Sessions or Classes (Total) |
|
Evaluated? |
Yes/No |
1. What (self esteem) curricula were used (if curriculum was developed by project, write "self developed")? |
|
Section II: Short-term Outcomes |
|
For those sessions where participant outcome was evaluated: (If no sessions were evaluated, skip to section III) |
|
Was it with: |
|
What was evaluated (check all that apply)? |
|
|
TABLE (for each type of activity, enter) |
Type of Activity |
|
Number of People Who Took Pre- Tests |
|
Number of People Who Took Post- Tests |
|
Number of People Who Took STANDARDIZED Pre- Tests |
|
Number of People Who Took STANDARDIZED POST- Tests |
|
Number of People with Increase in Score From Pre- to Post- Tests |
|
1. If standardized tests are used, please list the name(s) of the test(s)? |
|
Section III: Qualitative Impacts |
|
1. Please describe how your work in self-esteem building has impacted on three sample clients. To fill out this section, you can draw from client evaluation responses, conversations with or observations of project clients, notes, or your experience with clients. |
|
DATA ITEM |
RESPONSE |
Section I: Number of Individuals Served |
|
Table 12-1: Demographic Characteristics of Individuals Served |
Demographic Characteristics of Individuals Served |
Table 12-2: Number of Individuals Served and Type of Activity |
TABLE (for each type of activity, enter) |
Type of Activity |
|
Total Number Served |
|
Total Number of Events |
|
Section II: Short-term Outcomes |
|
For those activities where participant outcome was evaluated: (If no sessions were evaluated, skip to section III) |
|
Was it with |
|
What was evaluated (check all that apply)? |
|
|
TABLE (for each type of activity, enter) |
Type of Activity |
|
Number of People Who Took Pre- Tests |
|
Number of People Who Took Post- Tests |
|
Number of People Who Took STANDARDIZED Pre- Tests |
|
Number of People Who Took STANDARDIZED POST- Tests |
|
Number of People with Increase in Score From Pre- to Post- Tests |
|
If standardized tests were used, please list the names of the test(s) |
|
Section III: Qualitative Impacts |
|
1. Please describe how your cultural activities have impacted on three sample clients. To fill out this section, you can draw from client evaluation responses, conversations with or observations of project clients, notes, or your experience with clients. |
|
DATA ITEM |
RESPONSE |
Section I: Number of Individuals Served |
|
Table 13-1: Demographic Characteristics of Individuals Served |
Demographic Characteristics of Individuals Served |
Total Number of Sessions Conducted by Type |
|
Sports |
|
Other Recreational |
|
Section II: Short-term Outcomes |
|
For those activities where participant outcome was evaluated: (If no sessions were evaluated, skip to section III) |
|
Was it with |
|
What was evaluated (check all that apply)? |
|
If Pre- and Post-Tests, |
TABLE (for each type of activity, enter) |
Type of Activity |
|
Number of People Who Took Pre- Tests |
|
Number of People Who Took Post- Tests |
|
Number of People Who Took STANDARDIZED Pre- Tests |
|
Number of People Who Took STANDARDIZED POST- Tests |
|
Number of People with Increase in Score From Pre- to Post- Tests |
|
If standardized tests were used, please list the names of the test(s) |
|
DATA ITEM |
RESPONSE |
Section I: Number of Individuals Served |
|
Table 14-1: Demographic Characteristics of Individuals Served |
Demographic Characteristics of Individuals Served |
Table 14-2: Number of Interventions |
TABLE (for each type of intervention, enter) |
Type of Intervention |
|
Total Number of Interventions by Type |
|
Average Number of Participants Per Intervention |
|
Section II: Short Term Outcomes |
TABLE (for each type of intervention, enter) |
Type of Intervention |
|
Number of Situations Resolved |
|
Number of Situations Unresolved |
|
Section III: Qualitative Impacts |
|
1. Please describe how your work in crisis intervention has impacted on three sample clients. To fill out this section, you can draw from project client responses, conversations with or observations of clients, incident reports or notes, or your general experience with clients. |
|
DATA ITEM |
RESPONSE |
|
|
If you have more than one contract for this activity, a separate module should be filled out for each contract. Are you are filling out more than one Module 15? |
Yes/No |
IF YES: Which one is this? |
1 2 3 4 5 |
If Other, please list number: |
|
What is the role of conferences/meetings with respect to your OMH contract, cooperative agreement, or grant? |
|
For your OMH project, were you supposed to conduct (check one): |
|
As your only task, or as part of other project activities? |
|
Please describe: |
|
Section I: Number of Individuals Served and Sessions Conducted |
|
Table 15-1: Number of Individuals Served and Demographics |
Demographic Characteristics of Individuals Served |
Table 15-2: Number Served by Type of Event |
TABLE (for each type of event, enter) |
Type of Event |
|
Total Number Attending all Events |
|
Section II: Additional Conferences/Meetings Information |
|
Table 15-3: Conference/Meeting Chronology and Type of Event |
|
1. Please complete the following table for all conferences/meetings conducted (as part of your OMH contract, cooperative agreement or grant) during this reporting period |
TABLE (for each event, enter) |
Duration in Days |
|
Conference Name |
|
Date |
|
Target Population |
|
Health Issues |
|
Type of Event |
|
Table 15-4: Number of Materials Developed/Disseminated (at Conferences/Meetings) |
TABLE (for each material, enter) |
Conference Name
|
|
Date |
|
Type of Material |
|
Number Developed |
|
Total Number Distributed |
|
1. Conference/Meeting Purpose and Topics Please identify the primary purpose of each conference/meeting and list the major topics presented by event |
TABLE (for each event, enter) |
Conference Name |
|
Primary Purpose |
|
Topic |
|
2. Conference/Meeting Collaborations Please complete the following table for the same events listed above. On this table, we are asking for information concerning partners or collaborators you may have had in conducting the conferences/meetings. |
TABLE (for each event, enter) |
Table 15-5: Conference/Meeting Collaborations |
|
Conference Name |
|
Collaboration? |
Yes/No |
Number of Partners |
|
Type of Organizations |
Select An Organization Type |
Nature of Collaborations |
|
Section III: Evaluation of Conferences/Meetings |
|
1. Please complete the following table for the same events listed above. On this table, we are asking for information concerning evaluations you conducted for each conference/meeting |
TABLE (for each event, enter) |
Table 15-6: Evaluation |
|
Conference Name |
|
Evaluated? |
Yes/No |
Type of Evaluation |
|
Conduct Follow up? |
Yes/No |
Type of Follow up |
|
Follow up Method |
|
Section IV: Qualitative Impacts |
|
1. Please describe how your work in conference planning and management has impacted on three sample clients. To fill out this section, you can draw from client evaluation responses, feedback from event attendees, conversations with or observations of clients, notes, or your experience with clients. |
|
DATA ITEM |
RESPONSE |
|
|
Section I: Process Information |
|
1. Were you involved with any partnerships or collaborating organizations as an essential part of your OMH project? |
Yes/No |
If Yes, please describe |
TABLE (for each partnership, enter) |
Name of Organization |
|
Type of Agreement |
|
Type of Organization |
Select An Organization Type |
Role in Grant Activity |
|
Total Number of Meetings Conducted with that Organization |
|
Total Number of Activities conducted with that Organization |
|
Section II: Short-term Outcomes of Linkage-building and Community Coordination |
|
1. How many NEW organizations have you formed linkages with over the past reporting period? Please list |
TABLE (for each new linkage, enter) |
Name of Organization |
|
Type of Agreement |
|
Type of Organization |
Select type of organization |
Role in Grant Activity |
|
2. Did you form any new coalitions or collaborations in the past reporting period? Please list |
TABLE (for each new coalition, enter) |
Name of Organization |
|
Type of Agreement |
|
Type of Organization |
Select type of organization |
Role in Grant Activity |
|
For those coalitions or collaborations you formed or participated in, how many times did they meet? |
|
Were any of these collaborations part of ongoing task forces or committees? |
Yes/No |
If Yes, How many times did they meet? |
|
Are there plans for this partnership to continue meeting? |
Yes/No |
If No, did the partnership complete its goals? |
Yes/No |
Section III: System Change Data |
|
1. As a result of your work on linkage-building/community coordination, were any new polices or procedures implemented at the linked organizations? |
Yes/No/N/A |
If YES, please describe: |
|
2. As a result of your work on linkage-building/community coordination, has the grantee or partner organization (or their staff) become part of a local/regional coalition, committee, or other policy-related body? |
Yes/No |
If Yes, please describe |
TABLE (for each coalition, enter) |
Name of Committee |
|
Description of Task Force/Committee/Coalition |
|
Types of Members |
|
Other Information (IF APPLICABLE) |
|
|
|
3. As a result of your work on linkage-building/community coordination did any local providers form task forces, committees, coalitions, or other groups in order to address health services provided to the target population(s)? |
Yes/No |
If YES, please describe |
TABLE (for each task force, enter) |
Name of Provider |
|
Description of Task Force/Committee/Coalition |
|
Types of Members |
|
Other Information (IF APPLICABLE) |
|
4. As a result of your work on linkage-building/community coordination, did any community organizations collaborate to increase services, obtain funds, or engage in other collaborative activities? |
Yes/No/N/A |
If YES, please describe |
|
5. As a result of your work on linkage-building/community coordination, did the city, county or state initiate any changes in legislation or regulations regarding access to health care by your target community/ies? |
Yes/No/N/A |
If YES, please describe |
|
6. As a result of your work on linkage-building/community coordination, did the city, county or state draft any policy statements or guidelines regarding access to health care by your target community/ies? |
Yes/No/N/A |
If YES, please describe: |
|
Section IV: Qualitative Impacts |
|
Please describe how your work in linkage building/community coordination has impacted on three sample clients (either individuals or organizations). To fill out this section you can draw from project client evaluation responses, conversations with or observations of clients, notes, or your general experience with clients |
|
DATA ITEM |
RESPONSE |
Section I: Number of Individuals Served and Sessions Conducted |
|
Table 17-1: Number of Individuals Served |
Demographic Characteristics of Individuals Served |
Table 17-2: Organizations Served and TA Provided |
TABLE (for each organization, enter) |
Name of Organization |
|
Type of Organization |
Select An Organization Type |
New / Existing |
|
TA Provided |
|
Target Population |
|
Table 17-3: Number of Activities Conducted |
TABLE (for each organization, enter) |
Type of Activity (TA) |
|
Number of Times Activity Provided |
|
Total Number Served |
|
Section II: Short-Term Outcomes |
|
1) As a result of your work on organizational capacity building: Were any new polices or procedures developed at client organizations? |
Yes/No N/A |
If YES, please describe |
|
2. As a result of your work in this activity, were any new programs (e.g., HIV/AIDS education) implemented? |
Yes/No |
If YES, please describe |
|
3. As a result of your work in this activity, were any new funding applications submitted (by client organizations)? |
Yes/No N/A |
If YES, please describe? |
TABLE (for each funding source, enter) |
Funding Source |
|
Number of Applications Submitted |
|
Number of Applications Funded |
|
4. As a result of your work in this activity, were any new technologies or systems implemented? |
Yes/No N/A |
If YES, please describe. |
|
Section III: Qualitative Impacts |
|
Please describe three case examples of how your work in technical assistance and organizational capacity building has impacted on different, sample organizations, noting their situation and capacity before and after your assistance. To fill out this section you can draw from project client evaluation responses, conversations with or observations of clients, notes, or your general experience with clients. |
|
DATA ITEM |
RESPONSE |
|
Section I: Resources Provided to Organizations |
|
|
Table 18-1: Resources Provided to Organizations |
TABLE (for each activity, enter) |
|
Organization Name |
|
|
Organization Type |
Select An Organization Type |
|
Funding |
Yes/No |
|
Materials |
Yes/No |
|
Technology or Equipment |
Yes/No |
|
People |
Yes/No |
|
Other |
Yes/No |
|
1. Did you provide mini-grants to organizations as a project activity? |
Yes/No |
|
If Yes, please describe the recipient organization and the purpose of the grant in the space below. |
|
|
2. Did you develop/maintain a Web site for the purpose of making information available to community organizations? |
Yes/No |
|
If Yes, please describe the Web site in the space below. |
|
DATA ITEM |
RESPONSE |
Section I: Basic Information on Planning and Evaluation |
|
1. Which of the following methodologies were employed in your planning and evaluation activities (check all that apply)? |
|
2. Did your planning and evaluation activities address specific health conditions? |
Yes/No |
If YES, which health conditions were addressed? |
|
3. Did your planning and evaluation activities address specific populations? |
Yes/No |
If YES, which populations were addressed? |
|
4. Which of the following areas were covered in your planning and evaluation activities? (Check all that apply)? |
|
4. What were the main findings or results of your planning and evaluation activities? Please summarize, but include all key findings. |
|
5. Were data collected for planning purposes or to target resources? |
Yes/No |
If yes, please describe. |
|
6. Did you implement any changes in the data collection (such as collecting new kinds of data or enhancing data technology) to improve internal data systems? |
Yes/No |
If yes, please describe. |
|
7. Does your project address gaps or problems identified through your planning and evaluation activities? |
Yes/No |
If yes, please describe |
|
8. Did you evaluate efforts funded under your grant? |
Yes/No |
If yes, please describe. |
|
8a. Were your evaluation criteria related to goals or other targets in your strategic plan? |
Yes/No |
If yes, please describe. |
|
File Type | application/msword |
File Title | File: R:\OMH UDS IMP\OMB\New OMB UDS\OMBDraft1e |
Author | MEdberg |
Last Modified By | DHHS |
File Modified | 2007-06-07 |
File Created | 2007-06-07 |