Supporting Statement A-MCHB- ADAPT HV Active Engagement-6.5.24

Supporting Statement A-MCHB- ADAPT HV Active Engagement-6.5.24.docx

Assessing Strategies To Promote Children’s Engagement and Active Participation in Virtual Visits

OMB: 0906-0092

Document [docx]
Download: docx | pdf

Supporting Statement A



Assessing Strategies to Promote Children’s Engagement and Active Participation in Virtual Home Visits



OMB Control No. 0906-XXXX



Terms of Clearance: None

A. Justification

1. Circumstances Making the Collection of Information Necessary

This is a new Information Collection Request (ICR). The Health Resources and Services Administration (HRSA) requests Office of Management and Budget (OMB) approval to initiate information collection as part of the Assessing and Describing Practice Transitions Among Evidence-Based Home Visiting Programs in Response to the COVID-19 Public Health Emergency (ADAPT-HV) study to assess strategies to promote children’s engagement and active participation in virtual home visits.

The Maternal, Infant, and Early Childhood Home Visiting (MIECHV) Program is authorized by the Social Security Act, Title V, § 511(c) (42 USC 711(c)), as amended by Section 6101 of the Consolidated Appropriations Act, 2023 (P.L. 117-328). Through the MIECHV Program, HRSA provides grants to States, jurisdictions, and Tribal entities, who then have flexibility to develop, implement, and tailor their home visiting programs based on community needs, capacity, and resources within the parameters of statutory and programmatic requirements. State and jurisdiction awardees often contract with local implementing agencies to provide home visiting services in the communities. Home visiting is a service delivery strategy that matches expectant parents and caregivers of young children with a designated support person—typically a trained nurse, social worker, or early childhood specialist—who supports healthy pregnancy practices, encourages early language development and early learning at home, teaches positive parenting skills, connects families to other resources in their community, and provides information to support family health and well-being.1, 2 Services are voluntary and provided in the family’s home or another location of the family’s choice.

The COVID-19 Public Health Emergency (PHE) drastically impacted the needs of families and societal norms for in-person interactions. In response, many home visiting programs adjusted their practices to better serve families, such as shifting to virtual home visits and utilizing practice changes to support virtual delivery. These changes impacted the broader landscape in which home visiting programs operate to address families’ health, education, and other needs, and present opportunities to innovate how home visiting programs engage and deliver services to families. After the onset of the PHE, home visiting programs had to find new ways to reach families, especially in the early months of the PHE when many organizations suspended in-person services, and later to adhere to requirements and guidelines for social distancing. During this time, virtual delivery of home visits was common and required major shifts in how home visitors approached components of service delivery, including trust building with families, modeling and observations of child-parent interactions, the delivery of content, screenings, and assessments. The goals of the ADAPT-HV study are to identify such practice changes implemented in response to the COVID-19 PHE and build evidence on how the practice changes might help advance future home visiting service quality and delivery.

The switch from primarily in-person services to primarily virtual services introduced new considerations on how to adapt strategies for maintaining and promoting children’s engagement and active participation during visits.3 For example, virtual settings sometimes prevented regular or quality relationship building between families and home visitors, which made engaging caregivers and children participating in the home visiting program challenging. Home visitors also described difficulty engaging directly with infants and toddlers to model interactions or conduct assessments in a virtual setting.

While the PHE has ended and many programs have transitioned back to delivering home visits in person, MIECHV reauthorization allows programs to conduct virtual home visits within certain guidelines.4 Limited research exists about how current home visiting programs have implemented virtual engagement strategies to address these challenges. Understanding virtual engagement strategies that home visitors used or are using to promote children’s active engagement and participation in virtual visits, how these strategies can be implemented, and how children and families perceive these strategies can inform home visiting programming moving forward.

While there are no legal or administrative requirements that necessitate this data collection, subsection 511(h)(3) of the Social Security Act (42 USC 711(h)(3)(A)) authorizes the Secretary to “carry out a continuous program of research and evaluation activities in order to increase knowledge about the implementation and effectiveness of home visiting programs.”

2. Purposes and Use of the Information Collection

Limited research exists about how current early childhood home visiting (ECHV) programs have and continue to address the challenges of virtual engagement. The purpose of this information collection is to understand strategies implemented by home visitors to promote children’s engagement and active participation in virtual home visits. The information collected in this study will help inform ECHV programs about how virtual engagement implementation strategies can be refined to improve services. The information collection uses rapid cycle learning to identify, refine and test various virtual engagement strategies between home visitors and families. It will answer the following research questions:

  1. What strategies do home visitors use to promote children’s active engagement and participation in virtual home visits?

  2. How can the implementation of virtual engagement strategies be improved?

  3. How can virtual engagement strategies improve service delivery and promote family engagement and satisfaction with home visiting programs?


Research question 1 will be addressed through focus groups with home visiting program staff and families participating in home visiting and administrative information provided by program staff. The research question will explore the following topics:

  • Types of virtual engagement strategies home visitors use and how they vary from strategies used during in-person service delivery.

  • The variance of virtual engagement strategies by child age.

  • Home visitor application of virtual engagement strategies to in-person services.

  • Guidance that supports the implementation of virtual engagement strategies.

  • Programmatic, community, and family contexts that serve as facilitators and barriers to the implementation of virtual engagement strategies.


Research question 2 will be addressed through a brief questionnaire completed by home visitors and a focus group with program staff (including home visitors). The research question will explore the following topics:

  • Facilitators and barriers to implementation of virtual engagement strategies.

  • Refinements to virtual engagement strategies that have the potential to improve or expand implementation.

  • If refinements improve implementation and whether further refinements are needed.


Research question 3 will be addressed through questionnaires with families and home visitors, as well as focus groups with program staff and families. The research question will explore the following topics:

  • Program staff perspectives on improvements to service delivery and family satisfaction influenced by the implementation of virtual engagement strategies.

  • How home visiting modality (virtual, hybrid) influences how home visitors use virtual engagement strategies.


Study design

The study will use a formative rapid cycle learning framework to identify virtual engagement strategies, then improve and deliver them as defined strategies that can be scaled across other programs and contexts. ECHV programs (agencies that deliver home visiting services to families) will be the unit of analysis. The ADAPT-HV study will use a program eligibility protocol (Instrument 1) to purposely select four programs (referred to as program sites) that use virtual engagement strategies.

Once the program sites are selected, the study will include four phases designed to define, test, and refine virtual engagement strategies:

  • The first phase—co-definition—uses semi-structured program staff and family focus groups (Instruments 2-4) to identify, prioritize, define, and select virtual engagement strategies that will be used at each site.

  • During the second phase—installation and initial pilot—sites will implement selected virtual child engagement strategies. The study team will gather implementation data through a brief weekly home visitor questionnaire (Instrument 6) and program staff feedback on the implemented strategies through semi-structured focus groups (Instrument 5). This information will be used to identify refinements to the strategies that may improve implementation.

  • The third phase—refinement—involves up to two cycles of testing refinements to the strategies. In addition to gathering implementation data using semi-structured staff focus groups and home visitor questionnaires (Instruments 5 and 6, respectively), family post-visit questionnaires (Instrument 7) will be used to assess perceived improvements in the implementation of virtual engagement strategies during a cycle and identify further refinements to test.

  • The fourth phase—summary— focuses on reviewing and assessing overall implementation and process outcomes collected across the duration of the study. Using semi-structured family and staff focus groups (Instruments 8 and 4), the study team will assess the perceived potential of virtual child engagement strategies to improve service delivery and promote family engagement and family satisfaction with home visiting programs; and summarize the practice change strategy in its most useful form based on the iterative testing and the perceived potential of the virtual engagement strategies to improve home visiting services.

  • All respondents that participate in a semi-structured focus group will complete a participant characteristics form – All Phases (Instrument 9), to capture the demographic and contextual characteristics of the respondents of the study.

Universe of data collection efforts

Under this clearance, the study will use a variety of approaches. The exact data collection methods and the samples for each instrument will depend on the site. Data collection instruments (summarized in Table 1) include placeholders in sections that will refer directly to each individual site’s strategy that it is testing. Respondents will include MIECHV-funded ECHV program staff (including program directors, managers, supervisors, and home visitors), and families served by participating program sites. The samples will vary based on the content of the collection and the programs of interest. The study will collect data using well-established methodologies, including:

  • Semi-structured focus groups: This method involves group sessions guided by a moderator who follows a topical outline containing questions or topics focused on a particular issue, rather than adhering to a standardized questionnaire. Focus groups can be more efficient than individual interviews, since multiple individuals participate at one time. In addition, the group dynamics can yield richer responses than individual interviews for some types of topics. To encourage a high level of engagement, we estimate a maximum of six participants will participate in each focus group.

  • Questionnaires/Surveys: Questionnaires are common and popular tools to gather data from multiple people. Information from a questionnaire can inform research and evaluation planning as well as program support. Questionnaires may be used to gather information about specific implementation strategies or perceptions of those strategies.


Table 1 includes additional information about each data collection activity across all study phases, by instrument.

Table 1. Summary of Data Collection Activities

Instrument

Respondent, Content, Purpose of Collection

Mode

Duration and Frequency

Instrument 1. Program Eligibility Protocol

Respondents: Program director

Estimated number of respondents: 16

(16 sites x 1 respondent)

Content: Program eligibility, interest, and capacity to participate in study

Purpose: To prepare the program to sign an MOU and develop plans for formative evaluation

Phone, or virtual meeting platform

60 minutes (once)

Instrument 2. Program Staff Focus Group Protocol 1 -Co-definition Phase

Respondent: Program staff (e.g., director, managers, supervisors, and home visitors)

Estimated number of respondents: 24

(4 sites x 6 respondents)

Content: Guidance surrounding in-person and virtual child engagement strategies, facilitators of and challenges to implementation, family and community contexts, virtual child engagement strategies, mode of content delivery during visit

Purpose: Understand the strategies that have been implemented to promote virtual child engagement and implementation successes and challenges

Phone or virtual meeting platform

90 minutes (once)

Instrument 3. Program Staff Focus Group Protocol 2 - Co-definition Phase

Respondent: Program staff (e.g., director, managers, supervisors, and home visitors)

Estimated number of respondents: 24

(4 sites x 6 respondents)

Content: Review of information gathered in first co-definition staff focus group, group activity to prioritize strategies of interest, selection of strategy to refine and test

Purpose: Prioritize and select a practice to test in subsequent phases

Phone or virtual meeting platform

90 minutes (once)

Instrument 4. Family Focus Group Protocol -Co-definition & Summary Phases

Respondent: Families

Estimated number of respondents: 48

(4 sites x 6 respondents, per collection)

Content: Awareness of, satisfaction with, and perceived utility of child engagement and active participation during the visit

Purpose: Understand impressions of the services they received over the study, perceptions of how they work, and suggestions for improvement

Phone or virtual meeting platform

60 minutes (twice)

Instrument 5. Program Staff Focus Group Protocol - Installation & Refinement Phases

Respondent: Program staff and home visitors

Estimated number of respondents: 72

(4 sites x 6 respondents, per collection)

Content: Virtual child engagement strategies used, perception of family caregiver–child interactions, perception of child active participation

Purpose: Understand how staff implement a strategy, their perceptions of how it is working, and suggestions for improvement

Phone or virtual meeting platform

60 minutes (3 times)

Instrument 6. Home Visitor Questionnaire -Installation & Refinement Phases

Respondent: Home visitors

Estimated number of respondents: 40

(4 sites x 10 respondents)

Content: Virtual engagement strategies used, mode of content delivery during visits, perception of family caregiver–child interactions, perception of child engagement and active participation, rapport with family

Purpose: Understand staff use of engagement strategies over each learning cycle

Web-based

10 minutes, weekly (9 times)

Instrument 7. Family Post-Visit Questionnaire -Refinement Phase

Respondent: Families

Estimated number of respondents: 48

(4 sites x 6 respondents, per collection)

Content: Virtual engagement strategies used, mode of content delivery during visits, perception of family caregiver–child interactions, perception of child engagement and active participation

Purpose: Understand impressions of the home visit and perceptions of their child’s level of engagement during the visit

Web-based

5 minutes (6 times during each of 2 collection periods)

Instrument 8. Program Staff Focus Summative Group Protocol - Summary Phase

Respondent: Program staff and home visitors

Estimated number of respondents: 24

(4 sites x 6 respondents)

Content: Lessons learned and reflection surrounding virtual child engagement strategies used throughout learning cycles, home visitor self-efficacy, perceptions of family caregiver–child interactions and child engagement in relation to strategies

Purpose: Understand type, frequency, and purpose of implementation practices used, perceived potential to improve services

Phone or virtual meeting platform

60 minutes (once)

Instrument 9. Focus Group Participant Characteristics Form -All Phases

Respondent: All focus group respondents: program staff, families, home visitors

Estimated number of respondents: 120

(4 sites x 6 focus groups per site x 5 respondents per group; assumes some individuals participate in more than one focus group)

Content: Program staff: race/ethnicity, tenure in position, tenure with agency, enrollment capacity, actual enrollment staff capacity, race/ethnicity of families served. Home visitors: race/ethnicity, tenure in position, tenure with agency. Families: parent and child gender identity, ages, race/ethnicity of parent, participation (number of home visiting sessions attended), how long they have been receiving services.

Purpose: Describe focus group sample

Web-based

5 minutes (once per focus group)



3. Use of Information Technology and Burden Reduction

Each phase of the study will collect multiple sources of information to assess implementation and the success of the tested active participation strategies for children during virtual home visiting that are identified by the program. This information will be shared with program staff throughout the study. The planned information collection includes the use of technological data collection techniques. Specifically, the study includes online questionnaires and web-based forms.

Online data collection allows for efficiencies and reductions in respondent burden. Web-based surveys provide efficient ways of using skip logic to quickly move to the next relevant question depending upon a respondent’s answer selection. Web-based surveys also provide ways to limit invalid responses so that they cannot be entered and can prompt the respondent to enter a valid response. Web-based surveys reduce burden related to completing and mailing (or otherwise submitting) paper forms.

All online data collection instruments will be offered through Qualtrics, a web-based survey tool. Qualtrics uses Transport Layer Security (TLS) encryption for all transmitted data. Instruments and data collected through Qualtrics can only be accessed by authorized users with appropriate login credentials. Data exported from the platform for analysis will be stored on restricted, encrypted folders on PRG and Mathematica’s networks, which are only accessible to the study team. Data will be deleted from the platform at the end of the study.

To ensure consistency in data collection, the study team will develop appropriate processes for each data collection activity. For all questionnaires (Home Visitor Questionnaire [Installation & Refinement Phases], Family Post-Visit Questionnaire [Refinement Phase]), the study team will develop web-based survey response criteria. For example, numeric range restrictions on questions about caseload, age, and program start and end dates, among others. The study team will also use skip patterns to ask respondents only the most relevant questions.

The home visitors will provide the questionnaire to families in the form of a self-administered web survey, developed by the study team, via a tablet or laptop. If they are conducting a virtual home visit, the link to the questionnaire will be shared with the family member through the video platform chat or by email or text message.

To minimize burden on program staff and participants, the study design incorporates small-scale and iterative testing with small sample sizes and short data collection periods. The study team will tailor procedures to reduce burden on program staff. The study team will be cognizant of the burden participation in research can place on programs and participants and will work closely with programs to ensure methods build on existing processes and data.



4. Efforts to Identify Duplication and Use of Similar Information

This research will not duplicate any other information collections. Based on a survey of existing literature and expert consultation, none of the instruments ask for information that can be reliably obtained through other sources and used or modified for the purposes described in section 2 above. To the maximum extent possible, we, in collaboration with HRSA staff, will make use of existing data sources before we attempt to utilize the additional fieldwork sought under this clearance and will avoid any duplication of information collection efforts. HRSA is conducting two other related studies exploring adaptations to home visiting practices during COVID-19. Each study focuses on a different set of practices and will together provide a more comprehensive picture of how home visiting programs adjusted service delivery during the PHE. While home visiting programs are permitted to participation in multiple studies, the four programs who participate in this study will not participate in the other two studies as they will be asked to prioritize one set of practice changes during recruitment.

5. Impact on Small Businesses or Other Small Entities

This information collection will not have a significant economic impact on a substantial number of small businesses or other small entities. The home visiting program sites in the study may be small, non-profit organizations. The study team will only request information required for the intended use. The burden for respondents will be minimized by restricting the focus group and questionnaire length to the minimum required, by conducting focus groups at times convenient for the respondents, and by not requiring recordkeeping on the part of the programs.

6. Consequences of Collecting the Information Less Frequently

A key goal of the study is to be able to test strategies for virtual engagement and then make refinements to improve implementation, with an ultimate aim of strengthening programs’ virtual engagement practices. The iterative nature of rapid cycle evaluation necessitates repeat data collection over multiple cycles. Some of the study’s data collection will occur over two to three cycles: once during the Installation Phase, and one to two times during the Refinement Phase. Without repeat data collection, it would be difficult to assess the feasibility and effectiveness of the strategies that are implemented. The approach attempts to limit the scope of data collection to just the information needed to assess the success of the strategy being implemented. Over the rapid cycle evaluation, each home visitor respondent will complete the 10-minute Home Visitor Questionnaire – Installation and Refinement Phases (Instrument 6) up to nine times, and each family respondent will complete the 5-minute Post-Visit Questionnaire – Refinement Phase (Instrument 7) up to six times, depending on how frequently this evaluation activity is deemed useful to the program while piloting a new strategy or process. The short time frame and multiple responses facilitate rapid adaptation and refinement when program changes do not appear to be working as intended. Administering the survey less frequently would yield less actionable data for programs to use to refine their strategies.


The program eligibility protocol (Instrument 1), program staff interview protocol 1- Co-definition Phase (Instrument 2), program staff interview protocol 2- Co-definition Phase (Instrument 3) program staff summative focus group protocol -Summary Phase (Instrument 8), and focus group participant characteristics form – All Phases (Instrument 9) are one-time data collection activities. The program staff focus group protocol – Installation & Refinement Phases (Instrument 5) will be administered at the end of each cycle, up to three times. Each administration will be used to assess progress in the rapid cycle evaluations and adjust the approach. The family focus group protocol – Co-definition & Summary Phases (Instrument 4) will be administered twice. The first administration will be used to select virtual engagement strategies and define the evaluation project’s strategy. The second administration will contribute to an overall assessment of the success of the virtual engagement strategies tested and provide information about potential improvements to those strategies.


7. Special Circumstances Relating to the Guidelines of 5 CFR 1320.5

This request fully complies with the regulation.

8. Comments in Response to the Federal Register Notice and Efforts to Consult Outside the Agency



Section 8A:

A 60-day notice published in the Federal Register on 12-05-2023, vol. 88, No. 232; pp. 84342-43 (Appendix A). There were no public comments. A 30-day notice published in the Federal Register on 4-03-2024, vol. 89, No. 65; pp. 23023-24.



Section 8B:

Consultation with subject matter experts external to study team was conducted in 2023 as part of the design of this study (Table 2). The subject matter experts provided feedback on the design, specifically the research questions.

Table 2. Individuals consulted with about this information collection

Subject Matter Expert Role

Director of Early Childhood and Family Support Research School of Social Work

Senior Research Fellow Chapin Hall

Assistant Professor, Department of Population, Family and Reproductive Health

Professor, Suzanne Dworak-Peck School of Social Work 

Institution

Portland State University 

The University of Chicago

Johns Hopkins University 

University of Southern California



The study team also consulted with three advisory boards developed through this study, each consisting of five individuals in three different types of roles: program staff, home visitors, and families participating in home visiting. Each advisory board was asked a unique set of questions designed to draw on their specific expertise. These individuals provided feedback on language used to describe specific concepts in information collection instruments and strategies to best implement information collection procedures.

9. Explanation of Any Payment or Gift to Respondents

Focus group data are not intended to be representative of the experiences of all participant experiences in ECHV programs. However, it is important to recruit participants with a range of background characteristics to capture a range of possible program experiences. Without offsetting the direct costs of participating in the focus groups, such as arranging childcare, the research team increases the risk that only individuals able to overcome financial barriers to attend will participate in the study. A relatively consistent finding is that the tokens of appreciation can increase response rates,5,6 and in some cases can double response rates.7 Providing tokens of appreciation to respondents can also be a way of acknowledging the value of the time and knowledge being shared with the study team. They also reinforce the notion that the local community knowledge being shared by respondents is valued, respected, and honored. In this collection, the study team will offer $40 gift cards to families that participate in focus groups to encourage participants to engage in information sharing. The amount is based on estimated burden and evidence from previous studies with similar respondents and types of data collection requests. For example, results from a study conducted in 2013 on the effectiveness of tokens of appreciation for engaging participants in qualitative research found that any monetary token of appreciation was more likely to result in willingness to engage in a qualitative interview than a nonmonetary one or none.8 In the same study, among those participants who reported some willingness to engage in a 90-minute qualitative interview, those who were offered $75 were more willing to engage than those who were offered $25. As such, the study team prorated to a $40 token of appreciation for participation in 60-minute focus groups.

10. Assurance of Confidentiality Provided to Respondents

This data collection effort will collect personally identifiable information (PII) from program staff (names, work email addresses and telephone numbers) and families (names, phone numbers, and email addresses) for the purposes of arranging data collection (including scheduling and sending invitations to virtual data collection activities) and, for families, for sending tokens of appreciation. Any data collected will be de-identified, and transmitted via a password protected, secure file share system. Information will be saved in a secure folder accessible only to the study team. The study team will destroy data at the end of the study.

Information collected will be kept private to the extent permitted by law. Issues of privacy will be discussed during training sessions with staff working on the study. All study team members receive training on data security policies and procedures consistent with their organization’s documented policies. Program staff will transfer records using a secure file transfer protocol site in case the files contain PII.

Participation in all data collection activities is voluntary. All respondents will be informed that their responses will be kept private to the extent permitted by law. At the beginning of each focus group session, the facilitator will explain and ensure all participants understand the purpose of the study, their privacy rights, and that their participation in the study is voluntary. The focus group facilitators will ask that respondents do not share any information or personal experiences that they hear from others during the group. However, given the nature of a focus group (i.e., multiple respondents sharing information together), all respondents will hear responses from the group and confidentiality cannot be fully guaranteed. Focus groups will be recorded. Recordings, notes, and transcriptions will be saved to a secure drive and only the study team will have access.

For questionnaires, privacy means that only the study team will have access to their responses and that individual responses will not be shared with any home visiting program or local or state agencies or identified in any report. To ensure privacy for questionnaires administered after an in-person visit, home visitors will instruct families to return the tablet or laptop only after they have pressed the “submit” button on the web survey. The web questionnaire will include a consent statement at the beginning of the form explaining that home visitors will not have access to any families’ responses.

No information will be given to anyone outside of the study team and HRSA. All PII, typed notes, and audio recordings will be stored on restricted, encrypted folders on PRG and Mathematica’s networks, which are only accessible to the study team. The data will be stored for the duration of the study and destroyed at the end of the study in accordance with the National Archives and Records Administration (NARA) requirements.

The project team will submit all final data collection protocols for Institutional Review Board (IRB) review and approval prior to the commencement of data collection.

11. Justification for Sensitive Questions

The study will collect information about respondents’ race and ethnicity in Instrument 9. Focus Group Participant Characteristics Form - All Phases which will be used to describe the study sample. Race and ethnicity questions will align with Figure 2. Race and Ethnicity Questions with Minimum Categories Only and Examples from the revised SPD-15 standards (effective 3/28/24; see 89 FR 22182). The study will use minimum categories and examples, rather than the more detailed categories, to protect participant confidentiality. Given the small respondent sample sizes (N=120), the potential risks to confidentiality in situations where there are a very small number of individuals from a given racial or ethnic group outweigh the potential benefits of having more detailed demographic information. These demographic questions are essential to understand the contexts that strategies are being implemented, and enable strategies to be tailored in response to the needs of respondents. Additionally, to address variability, the study team will explicitly define the adaptations and contextual factors that the study responded to when framing our findings so that future programs can examine their own contexts and pursue appropriate strategies.

No other sensitive questions are included in the instruments.

12.A Estimated Annualized Burden Hours

The estimated burden per respondent varies (as shown in Table 3). The total burden for this information collection is 325.8 hours. There may be variation in the number of respondents in each program site (e.g., some selected program sites may have fewer home visitors). The total burden hours presented here assumes the maximum number of respondents across all 4 program sites. This burden estimate includes the time expended by persons to generate, maintain, retain, disclose, or provide the information requested. The general equation used to calculate burden for each form is: # participants * # responses per participant * number of program sites * average estimated burden per response = total burden hours for each instrument. Below, we describe how the total number of responses were calculated for forms with multiple responses per participant:

  • Program Staff Focus Group Protocol – Installation & Refinement Phases (Instrument 5): Program participants will participate in semi-structured focus groups. There will be up to three focus groups in each of the 4 sites, each with up to 6 participants, for a total of 72 observations (3 focus groups * 4 sites * 6 participants).

  • Home Visitor Questionnaire – Installation & Refinement Phases (Instrument 6): The study team will administer a 10-minute web survey to up to 10 home visitors in each of the 4 sites, for a total of 40 respondents. The team will administer these surveys to home visitors multiple times over the study. The study team assumes they will administer questionnaires surveys weekly for 3 weeks during each of the 3 cycles within the Installation & Refinement Phases for a total of 9 questionnaires per home visitor (3 cycles * 3 weeks/cycle), resulting in a total of 360 observations (9 responses per respondent * 40 respondents).

  • Family Post-Visit Questionnaire – Refinement Phase (Instrument 7): The study team will administer a 5-minute web survey to up to 12 family members in each of the 4 sites, for a total of 48 respondents. The study team assumes they will administer questionnaires surveys after each visit during a 6–12-week period, up to a maximum of 6 times per family, resulting in a total of 288 observations (6 responses per respondent * 48 respondents).



The instruments as written and submitted deliberately include more topics/questions than there is time for in the length allocated to the data collection activities in the burden table. Individual study team members working with the program sites will select the questions most relevant to what their sites are working on and drop irrelevant questions.

Table 3. Estimated Annualized Burden Hours

Type of Respondent

Form Name

Number of Respondents

Number of Responses per Respondent

Total Responses

Average Burden per Response (in hours)

Total Burden Hours

Program staff

Program Eligibility Protocol

16

1

16

1.00

16.0

Program staff

Program Staff Focus Group Protocol 1 (Co-definition Phase)

24

1

24

1.50

36.0

Program staff

Program Staff Focus Group Protocol 2 (Co-definition Phase)

24

1

24

1.50

36.0

Program staff

Program Staff Focus Group Protocol (Installation & Refinement Phases)

24

3

72

1.00

72.0

Program staff

Program Staff Focus Group Protocol (Summary Phase)

24

1

24

1.00

24.0

Families who participate in home visiting

Family Focus Group Protocol (Co-definition & Summary Phases)

48

1

48

1.00

48.0

Home visitor

Home Visitor Questionnaire (Installation & Refinement Phases)

40

9

360

0.17

61.2

Families who participate in home visiting

Family Post-Visit Questionnaire (Refinement Phase)

48

6

288

0.08

23.04

Program staff and families who participate in home visiting

Focus Group Participant Characteristics Form (All Phases)

120

1

120

0.08

9.6

Total


368


976


325.8




12.B Estimated Annualized Cost to Respondents

The estimated total cost to respondents is approximately $19,314.70 (as shown in Table 4). These estimates are based on our experience with collecting information, interviewing professional staff, and conducting focus groups. The study team based the average hourly wage estimates for deriving total annual costs on data from the Bureau of Labor Statistics, Usual Weekly Earnings of Wage and Salary Workers (2024 first quarter). For each instrument in Table 4, the team calculated the total annual cost by multiplying the annual burden hours by the average hourly wage. The study team used the median hourly wage of $32.689 for women in professional and related occupations for program staff (including home visitors), as the team expects many of the staff working in these positions to be women.10 The team used the median hourly wage of $20.1011 for women high school graduates with no college education for families participating in the focus groups.12

For the Focus Group Participant Characteristics Form, the cost to respondents is based on two types of respondents: families and program staff (including home visitors). Based on the planned data collection, a maximum of 96 respondents will be families and 24 will be program staff. The study team used the mean hourly wages for families and program staff for this instrument as outlined above.

The total respondent cost presented here assumes the maximum number of respondents in each program site. Since award recipients are spread across the country, the median hourly rate is used, as opposed to adjusting for locality. Additionally, wage has been doubled to account for overhead costs.



Table 4. Estimated Annualized Cost to Respondents

Type of Respondent

Total Burden

(hours)

Average Hourly Wage

Total Respondent Cost ($)

Families who participate in home visiting

78.68

$40.20

$3,162.94

Program staff

247.12

$65.36

$16,151.76

Total

325.8


$19,314.70



13. Estimates of Other Total Annual Cost Burden to Respondents or Record Keepers

The study team will provide an honorarium to participating programs in recognition of the administrative burden associated with their participation in the study. The study team will provide an honorarium of $100 to any program that participates in the initial eligibility discussion, regardless of whether they are eligible and/or are selected as a study site. For sites that participate in the study, the honoraria will be $1,500 per program site. As demonstrated in Table 3, which provides burden estimates for just the data collection activities associated with the study, this amount is less than the anticipated cost of staff time needed to complete individual-level recruitment, coordinating focus groups with program staff and families, and data collection throughout the cycles. The study team will ask each study site to designate a liaison to coordinate data collection activities. There are no other total annual cost burdens to respondents or record keepers resulting from the collection of information.

14. Annualized Cost to the Federal Government

The total cost of this information collection to the Federal Government is $243,895, which includes the cost of the contract to PRG to perform the study as well as the cost of federal employees supporting the study.


The annualized contractual cost for this information collection is $204,375, which includes designing data collection instruments, collecting all data and analyzing data. Estimates are based on the study team’s budget for each task (design, implementation, and analysis) and include labor hours, other direct costs, and subcontractor costs. The cost of the incentives is included in the cost of supplies paid to PRG.


In addition, costs to the Federal Government include the cost of federal staff time for project oversight and development. This includes approximately 10% FTE of a federal public health analyst at Grade 13, Step 4 ($95 per hour for 416 hours) for a total cost of $39,520. Wage has been multiplied by 1.5 to account for overhead costs.


A.15 Explanation for Program Changes or Adjustments

This is a new information collection request.

16. Plans for Tabulation and Publication and Project Time Schedule

The information collected under this request will be used to test the success of various strategies implemented at selected ECHV programs and inform HRSA activities.

Project Time Schedule

Beginning 1-2 months after OMB approval, the study team will conduct study recruitment activities, including collecting data through the program eligibility protocol. 2-4 months after OMB approval, the study team will work with selected programs to implement a collaboratively identified strategy and begin information collection activities. Timeline of collection may vary for individual program sites.

Plans for Tabulation

For this information collection, quantitative questionnaire data, such as from a Home Visitor Questionnaire, will include descriptive statistics and cross tabulations to assess sample size, characteristics, response rates, and data quality. To analyze qualitative data, such as focus groups, the study team will develop and apply a coding scheme to identify and summarize common themes across topics or respondent types. Following each learning cycle, the study team will share findings with program staff and HRSA, such as summary statistics from a learning cycle form, or initial themes identified during focus groups.

Plans for Publication

At the end of the study, the study team will develop a summative report that shares information with HRSA describing work with the sites and lessons learned about how implementation strategies address challenges ECHV programs face around virtual engagement. These findings, such as themes and recommendations identified through thematic coding of focus groups, will be disseminated in publications (briefs, manuscripts) and other public facing products. The main audience for these publications will vary, but may include HRSA leadership and staff, ECHV practitioners, researchers, advocates, and other interested parties in the broader field of ECHV. In sharing findings, we will describe the study methods and limitations with regard to generalizability and as a basis for policy.

17. Reason(s) Display of OMB Expiration Date Is Inappropriate

The OMB number and Expiration date will be displayed on every page of every form/instrument.

18. Exceptions to Certification for Paperwork Reduction Act Submissions

There are no exceptions to the certification.


1 National Home Visiting Resource Center. (2022). What is Home Visiting? https://nhvrc.org/what-is-home-visiting/

3 Marshall, J., Kihlstrӧm, L., Buro, A., Chandran, V., Prieto, C., Stein-Elger, R., Koeut-Futch, K., Parish, A., & Hood, K. (2020). Statewide implementation of virtual perinatal home visiting during COVID-19. Maternal and Child Health Journal, 24(10), 1224–1230. DOI: 10.1007/s10995-020-02982-8; Chazan-Cohen, R., Fisk, E., Ginsberg, I., Gordon, A., Green, B. L., Kappesser, K., Lau, S., Ordonez-Rojas, D., Perry, D. F., Reid, D., Rodriguez, L., & Tomkunas, A. (2021, September). Parents’ experiences with remote home visiting and infant mental health programs during COVID-19: Important lessons for future service delivery. Perigee Fund. Retrieved April 21, 2023, from https://perigeefund.org/wp-content/uploads/2021/10/ParentVoices-FullReport-English.pdf; Korfmacher, J., Molloy, P., & Frese, M. (2021, October). Virtually the same? Virtual home visits in response to COVID-19. Chicago, IL. Erikson Institute. Retrieved April 24, 2023, from https://www.erikson.edu/wp-content/uploads/2021/10/Research-Brief-1-HV-COVID-Survey.pdf; Korfmacher, J., Molloy, P., & Frese, M. (2021, October). “But it’s not the same”: What happens in virtual home visits? Chicago, IL. Erikson Institute. Retrieved April 21, 2023, from https://www.erikson.edu/wp-content/uploads/2021/10/Research-Brief-2-HV-COVID-Obs-Int.pdf

4 Social Security Act, Title V, § 511(e)(10)

5 Singer, E., and C. Ye. “The Use and Effects of Incentives in Surveys.” The ANNALS of the American Academy of Political and Social Science, vol. 645, no. 1, 2013, pp. 112–141. https://doi.org/10.1177/0002716212458082

6 Sundstrom, E. D., Hardin, E. E., & Shaffer, M. J. (2016). Extra credit micro-incentives and response rates for online course evaluations: Two quasi-experiments. Teaching of Psychology, 43(4), 276-284.

7 Edwards, P., Roberts, I., Clarke, M., DiGuiseppi, C., Pratap, S., Wentz, R., & Kwan, I. (2002).

Increasing response rates to postal questionnaires: Systematic review. BMJ, 324(7347),

1183.

8 Kelly, Bridget, Marjorie Margolis, Lauren McCormack, Patricia A. LeBaron, and Dhuly Chowdhury. “What Affects People’s Willingness to Participate in Qualitative Research? An Experimental Comparison of Five Incentives.” Field Methods, vol. 29, no. 4, November 2017, pp. 333–350. https://doi.org/10.1177/1525822X17698958.

10 Sandstrom, Heather, Sarah Benatar, Rebecca Peters, Devon Genua, Amelia Coffey, Cary Lou, Shirley Adelstein, and Erica Greenberg. 2020. Home Visiting Career Trajectories: Final Report. OPRE Report #2020-11, Washington, DC: Office of Planning, Research, and Evaluation, Administration for Children and Families, U.S. Department of Health and Human Services. https://www.acf.hhs.gov/sites/default/files/documents/opre/hvct_final_report_feb_2020.pdf

12 https://mchb.hrsa.gov/sites/default/files/mchb/about-us/home-visiting-infographic.pdf

8


File Typeapplication/vnd.openxmlformats-officedocument.wordprocessingml.document
AuthorJulius Anastasio
File Modified0000-00-00
File Created2024-08-01

© 2024 OMB.report | Privacy Policy