Protocol
Abstract
Background: Although substantial progress has been made in establishing evidence-based psychosocial clinical interventions and implementation strategies for mental health, translating research into practice—particularly in more accessible, community settings—has been slow.
Objective: This protocol outlines the renewal of the National Institute of Mental Health–funded University of Washington Advanced Laboratories for Accelerating the Reach and Impact of Treatments for Youth and Adults with Mental Illness Center, which draws from human-centered design (HCD) and implementation science to improve clinical interventions and implementation strategies. The Center’s second round of funding (2023-2028) focuses on using the Discover, Design and Build, and Test (DDBT) framework to address 3 priority clinical intervention and implementation strategy mechanisms (ie, usability, engagement, and appropriateness), which we identified as challenges to implementation and scalability during the first iteration of the center. Local redesign teams work collaboratively and share decision-making to carry out DDBT.
Methods: All 4 core studies received institutional review board approval by June 2024, and each pilot project will pursue institutional review board approval when awarded. We will provide research infrastructure to 1 large effectiveness study and 3 exploratory pilot studies as part of the center grant. At least 4 additional small pilot studies will be solicited and funded by the center. All studies will explore the use of DDBT for clinical interventions and implementation strategies to identify modification targets to improve usability, engagement, and appropriateness in accessible nonspecialty settings (Discover phase); develop redesign solutions with local teams to address modification targets (Design and Build phase); and determine if redesign improves usability, engagement, and appropriateness (Test phase), as well as implementation outcomes. Center staff will collaborate with local redesign teams to develop and test clinical interventions and implementation strategies for community settings. We will collaborate with teams to use methods and centerwide measures that facilitate cross-project analysis of the effects of DDBT-driven redesign on outcomes of interest.
Results: As of January 2025, three of the 4 core studies are underway. We will generate additional evidence on the robustness of DDBT and whether combining HCD and implementation science is an asset for improving clinical interventions and implementation strategies.
Conclusions: During the first round of the center, we established that DDBT is a useful approach to systematically identify and address chronic challenges of implementing clinical interventions and implementation strategies. In this subsequent grant, we expect to increase evidence of DDBT’s impact on clinical interventions and implementation strategies by expanding a list of common challenges that could benefit from modification, a list of exemplary solutions to address these challenges, and guidance on using the DDBT framework. These resources will contribute to broader discourse on how to enhance implementation of clinical interventions and implementation strategies that integrate HCD and implementation science.
International Registered Report Identifier (IRRID): PRR1-10.2196/65446
doi:10.2196/65446
Keywords
Introduction
Background
Psychosocial clinical interventions such as psychotherapy, counseling, and case management are a preferred mode of treatment by most people seeking care for mental health problems [
- ]. Access to evidence-based clinical interventions remains variable among diverse groups, leading to mental health disparities across racial and ethnic, geographic, and socioeconomic status [ - ]. Furthermore, implementing clinical interventions in nontraditional and integrated settings (eg, primary care, telehealth platforms, and schools) has shown mixed success. These settings can serve as a safety net for accessing mental health treatment when traditional options are inaccessible [ - ]. Addressing barriers to implementing clinical interventions in these settings is vital to promote more equitable access to mental health services for all.Poor availability of clinical interventions is often due to intervention complexity and suboptimal fit with many settings where clinical interventions are deployed [
]. Implementation strategies—“systematic intervention process(es) to promote the uptake of evidence-based health innovations into usual care”—have often taken the form of complex tools and processes (eg, train the trainer, booster training, incentive models, and decision supports) [ , ]. However, these often fall short because they can be excessively costly and cumbersome [ , ]. Different needs of recipients and settings can lead to high rates of reactive adaptations of clinical interventions and implementation strategies by their intended users in many settings where they are deployed. Reactive adaptations are unplanned or improvised changes during an implementation process in response to unanticipated challenges [ ]. While reactive adaptations can compromise clinical potency, those that are proactively tailored to different care settings can improve sustainability and impact [ - ]. A systematic review of cultural adaptations to health and mental health services highlighted how adaptations motivated by cultural sensitivity are not guaranteed to demonstrate increased efficacy [ ]. Instead, patient-centered approaches that account for individualized needs and barriers to service are recommended to guide adaptations [ , ].Human-Centered Design and Implementation Science
Human-centered design (HCD) and the closely related discipline user-centered design offer a suite of methods to develop useful, compelling, intuitive, and enjoyable products, services, and tools based on people’s needs [
, ]. HCD relates to the evolution of human-computer interaction (HCI), a multidisciplinary field that incorporates computer science, cognitive science, and human factors engineering as a response to personal computing, collaborative work, and interconnected technologies in everyday life [ ]. While HCD’s origins are rooted in technology, it has been used beyond the context of digital technologies to address therapeutic elements and implementation supports [ ]. HCD has been applied to improve usability, reduce burden, and increase the contextual appropriateness of clinical interventions and implementation strategies [ - ].The fields of HCD and implementation science share common objectives and offer complementary methods that can support clinical interventions and implementation strategies innovation and redesign [
- ]. HCD techniques are particularly well positioned to help with redesign, which we define as adaptations to clinical interventions and implementation strategies while preserving effective components (ie, fidelity-consistent adaptation) [ ]. HCD’s traditions of situating problem discovery and solutions in user needs, usability, engagement, innovation, and rapid exploration are core strengths that align with implementation science’s goal of improved adoption, fidelity, reach, and adaptation of clinical interventions and implementation strategies [ ]. Combining HCD and implementation science traditions for clinical intervention and implementation strategy redesign grounds novelty in empirical evidence.Integrating HCD and Implementation Science Through the Discover, Design and Build, and Test Framework
The University of Washington Advanced Laboratories for Accelerating the Reach and Impact of Treatments for Youth and Adults with Mental Illness Center (UWAC), which is funded by a grant from the National Institute of Mental Health (NIMH), is a multidisciplinary team of experts from the fields of mental health, implementation science, and HCD focused on improving usability, engagement, and appropriateness of clinical interventions and implementation strategies in diverse and nonspecialty settings (eg, rural, urban, low-income, primary care, and schools). Drawing on strengths from different disciplines, UWAC developed the Discover, Design and Build, and Test (DDBT) framework (
) at the start of the center (ie, “UWAC 1.0”). The current DDBT framework guides teams in redesigning clinical interventions and implementation strategies to improve usability, engagement, appropriateness, and implementation outcomes while preserving clinical interventions’ core components [ ]. Key principles underlying this model include the following: (1) not all clinical interventions and implementation strategies are designed for all settings; (2) “there is no implementation without adaptation” [ ]; (3) unchecked, reactive adaptations have the potential to exclude essential active ingredients [ , ]; and (4) reactively adapted clinical interventions and implementation strategies can negatively impact implementation and clinical outcomes.
The DDBT framework is modeled after existing HCD frameworks [
] and is an iterative stepped approach to systematically (1) understand usability constraints of existing clinical interventions and implementation strategies, (2) iteratively design solutions for usability challenges with redesign teams of direct and indirect users, and (3) test and refine prototypes. We define direct users (also known as “primary users”) as people who directly interact with the clinical intervention and implementation strategy and indirect users (also known as “secondary users”) as people affected by the clinical intervention and implementation strategy. HCD places a strong emphasis on explicitly identifying relevant community collaborators and users to ensure that new products effectively meet their needs [ , ]. DDBT starts by identifying multilevel factors that drive clinical intervention and implementation strategy usability problems, engagement challenges, and problems with contextual appropriateness (Discover phase). Once problems and challenges are identified, modifications are iteratively created between the design team and practitioners and clients, until a new version of the clinical interventions and implementation strategies is developed to address crucial issues and enhance usability, engagement, and appropriateness ( ; Design and Build phase). Early prototypes of clinical interventions and implementation strategies are assessed with small samples (eg, 5-25 participants) to answer design questions using paper or other “low-fidelity” (ie, sufficient to communicate a concept but potentially lacking functionality, some content, and look and feel of final materials) versions of modifications, which reduces waste of unnecessary investment in programming and development until as late in the process as possible. Findings from the Design and Build phase are incorporated to develop high-fidelity prototypes, which are tested against the unadapted version to ascertain if the modified clinical interventions and implementation strategies result in improved implementation (eg, increased adoption, fidelity, reach, and reduced reactive adaptations) and equivalent or better mental and behavioral health outcomes because of the changes to usability, engagement, and appropriateness (Test phase). Additional details on the DDBT framework are outlined in our UWAC 1.0 protocol paper [ ].All UWAC research uses the DDBT framework, which is applied flexibly based on project needs and allows us to evaluate the extent to which incorporating HCD and implementation science methods impacts clinical interventions and implementation strategies. Since 2018, DDBT has been used in >18 UWAC studies and 16 National Institutes of Health (NIH)–funded awards external to UWAC. During UWAC 1.0, we originally assessed impact of DDBT on 3 mechanisms: learnability (ie, extent to which users can understand or facilitate use) [
], usability (ie, extent to which users can achieve specified goals of effectiveness, efficiency, and satisfaction) [ ], and sustained quality of care (ie, extent of treatment fidelity and impact on target outcomes) [ ]. Analysis of UWAC 1.0 projects resulted in (1) identification of common usability issues in clinical interventions and implementation strategies that could benefit from modification (ie, “typology of modification targets”) and corresponding heuristics to guide their design [ ], (2) reflections on potential exemplary solutions to these challenges (ie, library of clinical interventions and implementation strategies redesign solutions), and (3) guidelines for using the DDBT framework [ ].
UWAC Theory of Change for Clinical Intervention and Implementation Strategy Redesign
Our UWAC 1.0 findings and the implementation literature [
] outline how limited availability and the use of clinical interventions and implementation strategies are persistently attributable to organizational and system characteristics (eg, readiness to adopt, resources and culture, and leadership), clinician and adopter characteristics (eg, appropriateness and perceived efficacy of clinical interventions and implementation strategies for patients) [ , , ], and incentives to engage in clinical interventions and implementation strategies [ , ].As a result, we updated DDBT’s underlying theory of change to highlight how adoption of clinical interventions and implementation strategies are largely due to usability (eg, extent to which clinical interventions and implementation strategies can be used to achieve specified goals with effectiveness, efficiency, and satisfaction) [
]; engagement (eg, degree of user participation and enthusiasm for the aspects of clinical interventions and implementation strategies that require user involvement) [ ]; and appropriateness (eg, perceived fit, relevance, or compatibility of clinical interventions and implementation strategies for a given practice setting, practitioner, or consumer) [ ] ( ). These mechanisms are direct targets of DDBT-driven redesign and the focus of the new iteration of funding between 2023 and 2028 (ie, “UWAC 2.0”). We ultimately expect DDBT to result in changes to proximal implementation outcomes (adoption, fidelity, reach, and reactive adaptation) and clinical outcomes.DDBT Mechanisms
Usability
Usability is an underlying outcome at all stages of the HCD process. Understanding the extent to which designs are unusable and opportunities to increase an existing solution’s usability can inspire innovation and adoption [
]. Deployment of clinical interventions and implementation strategies will continue to be subpar unless usability can be addressed, the historically unidirectional relationship between developers and users can be overcome, and insufficient incorporation of user perspectives can be remedied [ ]. Usability is assessed through usability evaluations and usability testing, where prototypes are evaluated using established heuristics [ ] and observing users complete critical tasks [ ]. Usability assessment methods stem from evaluating technologies; however, these techniques have been used to improve usability, decrease burden, and increase contextual appropriateness of nontechnological mental health clinical interventions and implementation strategies [ , , , ]. Approaches to assessing usability frequently couple interview-type questions with surveys and observation [ ]. As a form of usability evaluation, metrics such as the System Usability Scale (SUS) [ ] are questionnaires that assess perceived usability. The SUS is a widely used instrument to measure usability of technologies by industry. Task-based usability testing involves asking participants to complete tasks while using a prototype or product. This method can be used to gather baseline usability data for an existing clinical intervention and implementation strategy and assess usability of clinical intervention and implementation strategy prototypes [ ]. The think-aloud protocol (TAP) involves participants verbalizing thought processes as they use clinical interventions and implementation strategies [ ] to complete assigned tasks, including actions they consider taking and reactions to materials as they encounter them. Similarly, the Cognitive Walkthrough for Implementation Strategies (CWIS) is a 6-step method for evaluating clinical interventions and implementation strategies usability, which can include interviews as part of task-based usability testing [ ].UWAC 1.0 projects assessed usability through usability questionnaires, CWIS [
], and TAP [ , ]. We developed the Intervention Usability Scale (IUS) [ ] and Implementation Strategy Usability Scale (ISUS) [ ] to better measure clinical intervention and implementation strategy usability; these scales are closely aligned with SUS. SUS scores of ≥70 out of 100 are considered adequate usability, and we anticipate a similar threshold for IUS and ISUS. Cross-project usability data informed the typology of modification targets: 12 unique categories of clinical intervention and implementation strategy usability issues of varying severity. These categories help researchers understand common barriers to clinical intervention and implementation strategy use that can be prevented and addressed during clinical intervention and implementation strategy redesign [ ].Engagement
Engagement, and adaptations to improve engagement, relates to clinical interventions fidelity and clinical outcomes [
- ] and is a defining feature of quality of care [ - ]. Engagement describes user connection to clinical interventions and implementation strategies and their capacity to sustain a connection [ ]. Engagement is distinct from common health concepts of compliance, adherence, and coverage because it incorporates a dimension of quality and welcomes the possibility that different people may engage differently with different parts of clinical interventions and implementation strategies based on varying needs. Typical users include the practitioners who deliver them (eg, clinicians, implementation practitioners, and intermediaries) and the individuals who receive them (eg, clients, practitioners, and service system administrators). Our work has found that engagement can be negatively impacted by insufficient buy-in, components that are inaccessible to different users, little support for communication or rapport building, and requirements or constraints that inadvertently shift one user’s responsibilities to another (eg, practitioners completing client tasks) [ ].Engagement is a common focus in HCI literature [
, - ], where there is an active conversation around how to best assess the quality of interactions rather than quantity of interactions. Multidisciplinary UWAC project teams present an opportunity to incorporate different approaches to improving engagement in translational research [ - ]. We conceptualize engagement as a multifaceted construct focused on interaction quality (ie, participation and enthusiasm) that is enhanced by clinical interventions and implementation strategies that are well designed and result in improved adoption, fidelity, reach, and adaptations. There are subjectivity-oriented and objectivity-oriented approaches to measuring engagement [ ]. Subjectivity-oriented measures are self-reported and include questionnaires, observation, perceived task outcomes, and interviews. Objectivity-oriented measures minimize researcher involvement and can include behavior logging, psychophysiological measurements, or telemetry. Within HCI, objectivity-oriented measures such as user data—logs, time, number of interactions, and frequency of log-ins—are commonly used.Appropriateness
Appropriateness captures perceived fit between clinical interventions and implementation strategies and the settings to which they are deployed [
]. Client needs and therapeutic style drive how mental health practitioners modify clinical interventions and implementation strategies in practice [ ]. During UWAC 1.0, we observed how practitioners and recipients felt that exciting innovations can be inappropriate for specific contexts (eg, schools) or for users facing challenges with identifying and selecting goals or implementation plans. Challenges included clinical interventions and implementation strategies’ excessive time demands in their delivery, incompatibility with existing workflows or roles, unavailable system infrastructure requirements, overreliance on digital technology, and, importantly, practitioner perceptions of the fit of the clinical interventions and implementation strategies to specific client problem types. These issues are well documented as multifaceted factors that influence clinical intervention and implementation strategy adaptations [ , ]. During UWAC 1.0, we used the Intervention Appropriateness Measure (IAM) [ ] across projects to assess appropriateness.Study Purpose
Center Aims and Structure
Our goal is to overcome obstacles that prevent quality mental health interventions from reaching historically marginalized groups through addressing critical clinical intervention and implementation strategy problems with the DDBT framework. Building on learnings during the first iteration of the center, UWAC 2.0 focuses on addressing longstanding problems with usability, engagement, and appropriateness of clinical interventions and implementation strategies that result in high rates of reactive adaptations in settings where they are deployed. Within UWAC, the Methods Core team provides methodological and technical support to all projects and maintains cross-cutting project data on UWAC outcomes to determine the impact of DDBT on clinical interventions and implementation strategies. These data will be used to refine and expand UWAC’s typology of modification targets and library of redesign solutions. This iteration of the center emphasizes increased leadership and application of DDBT methods by local project redesign teams (eg, administrators or champions) that receive methodological training and supports from the Methods Core team to work alongside investigators, increasing their decision-making at all stages of the design process. This is a shift from the previous centralized model, where projects engaged users, but project principal investigators (PIs) and UWAC Methods Core team members often led the design process. The Methods Core team will address the subsequent aims during UWAC 2.0.
Aim 1: Identify Clinical Intervention and Implementation Strategy Modification Targets to Improve Usability, Engagement, and Appropriateness in Accessible Nonspecialty Settings (Discover Phase)
Known determinants of successful clinical intervention and implementation strategy use exist at several levels, including clinical interventions (eg, complexity); practitioner and client (eg, training, attitudes, and intention to use); and organizational (eg, climate, leadership, resources, and supervision). Historically, adaptations of clinical interventions and implementation strategies have been driven by academics rather than the lived experiences of recipients. Using the Consolidated Framework for Implementation Research (CFIR) [
] as our guide in the Discover phase of each UWAC project, we support project redesign teams to use HCD methods to identify targets within our typology of modification targets [ ]. Qualitative analyses will allow us to compare targets identified via local DDBT, characterized by user involvement in redesign teams and shared decision-making about target prioritization and solutions, with those derived from our original, centralized application of DDBT [ ]. Aim 1 outcomes will inform typology revisions and allow for comparisons between the original, centralized DDBT and the local DDBT. An updated typology will be broadly disseminated to inform future research.Aim 2: Develop Redesign Solutions With Local Teams to Address Clinical Intervention and Implementation Strategy Modification Targets (Design and Build Phase)
Using rapid, iterative design principles, we are supporting research project teams in redesigning clinical interventions and implementation strategies to enhance usability, engagement, and appropriateness. We will systematically catalog these design solutions using the Framework for Modifications and Adaptations of Evidence-Based Interventions (FRAME) or FRAME for Implementation Strategies (FRAME-IS) [
, , ]. We will examine solutions and the populations, organization types or structures, practitioner types, and clinical interventions and implementation strategies in which they work, identifying solutions that transfer across different contexts or are uniquely suited to specific contexts. Aim 2 outcomes will be compared with centralized DDBT outputs, resulting in an updated library of redesign solutions organized by target and redesign method (ie, localized vs centralized), and shared with the interested community.Aim 3: Determine If Redesign Affects Changes in Usability, Engagement, and Appropriateness (Test Phase)
Each project conducting a Test phase will include a hybrid effectiveness-implementation trial [
, ] with a primary comparison between the original clinical intervention and implementation strategy and the DDBT-adapted version on theorized mechanisms (usability, engagement, and appropriateness), implementation outcomes (adoption, fidelity, reach, and adaptations), and client outcomes. Projects will apply FRAME or FRAME-IS [ , ] to examine the extent to which DDBT decreases reactive adaptations to the clinical intervention and implementation strategy (ie, unplanned or due to unanticipated obstacles) during implementation. We hypothesize that DDBT-informed, prospective, and planned adaptations will reduce the number and extent of reactive adaptations. The Methods Core team will systematically integrate new data from the UWAC projects with existing data in an integrated cross-project analysis of how redesign affects theorized mechanisms, implementation outcomes, and patient outcomes. Aim 3 outcomes will be disseminated to the field and inform new projects designed to test which redesign strategies best improve DDBT mechanisms and outcomes.Methods
Overview
Composed of an interdisciplinary team and advisory board with experience in HCD, implementation science, psychosocial clinical interventions, and research methods and data, the Methods Core team supports one large hybrid effectiveness-implementation study (NIH R01), 3 exploratory pilot studies (NIH R34s;
), and at least 4 pilot projects (NIH R03s) during UWAC 2.0. These projects aim to improve clinical intervention and implementation strategy access and scale in diverse settings. The R01 Problem Solving Treatment-Aid (PST-Aid; NCT06494384; PIs: IB, PR, and SAM) will test a DDBT-designed decision support tool for Problem Solving Treatment (PST) in a large network of primary care clinics. R34 Research Units on Behavioral Intervention in Educational Settings (RUBIES; NCT06508515; PIs: KB and JL) will create a novel implementation strategy to support the delivery of evidence-based classroom supports for students with autism. R34 Trauma-Focused Cognitive Behavioral Therapy (TF-CBT; PIs: ARL and DW) will redesign and test an evidence-based clinical intervention for youth trauma for use in education settings. R34 Brief Intervention for School Clinicians (BRISC; PIs: EB, JVD, and Elizabeth McCauley) will adapt an existing set of effective implementation strategies to enhance delivery of a school-based engagement, assessment, brief mental health intervention, and triage strategy. Funding for all core projects began in 2023. outlines key details (eg, study design and sample) of each study. PST-Aid and RUBIES were part of UWAC 1.0 as R34 and R03 projects, respectively, highlighting how UWAC 2.0 activities build on previous accomplishments. UWAC funds pilot projects through a competitive solicitation process, with a particular focus on supporting and mentoring investigators from historically marginalized groups. All projects use the DDBT framework to address clinical intervention and implementation strategy usability, engagement, and appropriateness in partnership with local community collaborators. UWAC provides projects direct support for integrating methods and measurement approaches and professional development that centers diversity, equity, and inclusion (DEI) values.DDBT Constructs
Projects collect common data on DDBT mechanisms and constructs to determine the impact of modifying clinical intervention and implementation strategy targets. The Methods Core team maintains a list of recommended and required measures for projects to gather at each DDBT phase (“Center Measures and Guidance”;
). We developed Center Measures and Guidance to facilitate DDBT hypothesis testing and data management across UWAC, help teams select methods based on project design objectives, and satisfy NIMH reporting requirements. Center Measures and Guidance include 15 constructs with 26 quantitative and qualitative measures across DDBT phases. The Methods Core team provides project teams support with integrating and adapting these measures for projects through a consultation model. This is a shift from our approach in UWAC 1.0, where the Methods Core team provided more personalized measurement support to project teams. [ ] outlines each construct with a description of related measures and activities, and relevance to each DDBT phase. The Methods Core team provides data management for all projects, maintains survey instruments in REDCap (Research Electronic Data Capture; Vanderbilt University), and conducts cross-project analyses. Each project is responsible for conducting its own analyses. The Methods Core team provides guidance on ensuring recruitment, data analysis, and dissemination practices incorporate diverse perspectives and accurately represent the lived experiences of participants to inform clinical intervention and implementation strategy redesign.DDBT Theory of Change Mechanisms
Usability
All projects are expected to report usability issues on existing or redesigned clinical interventions and implementation strategies and standardized usability metrics to the Methods Core team. Because interviews alone can be limiting for identifying usability issues because of issues with recall or challenges with describing behavior, UWAC projects are encouraged to combine interviews with other methods such as CWIS and TAP. This helps projects learn what a participant is considering doing next and why, better understand their in-the-moment goals, and identify misconceptions. For example, R34 BRISC will use TAP and a cognitive walk-through methodology with users to identify opportunities for redesign and improve implementation based on user needs [
]. The Methods Core team supports projects with adapting surveys (eg, SUS, IUS, and SUS) and implementing cognitive walkthroughs and usability testing for projects.Engagement
We expect all projects to assess engagement quantitatively using the User Responsiveness Scale and qualitatively (ie, thematic findings from observation or other chosen methods) during the Discover and Test phases of existing and redesigned clinical intervention and implementation strategy. The Methods Core team developed the User Responsiveness Scale based on the Patient Responsiveness Scale [
]. The User Responsiveness Scale has 10 statements that cover participation and enthusiasm for a clinical intervention and implementation strategy that participants rate on a Likert scale. The original Patient Responsiveness Scale has demonstrated strong internal consistency (Cronbach α=0.86) and construct validity.Appropriateness
R01 and R34s are expected to administer the IAM [
] or revised goodness-of-fit interview [ ] during the Discover and Test phases of existing and redesigned clinical interventions and implementation strategies to probe areas of alignment and misalignment on goals and expectations, roles, etc. IAM is a 4-item survey and the leading instrument for measuring clinical intervention and implementation strategy contextual fit with good internal consistency (Cronbach α=0.87) and adequate test-retest reliability (r=0.73). The goodness-of-fit interview is particularly well-suited to probe on clinical intervention and implementation strategy appropriateness issues identified through IAM. We will use content analysis to analyze goodness-of-fit interview data. For example, R34 RUBIES will conduct goodness-of-fit interviews to explore the appropriateness of the RUBIES implementation strategy (“RUBIES-Team”) for the school environment using the CFIR domains to drive questioning.Proximal Implementation Outcomes
Adoption and Reach
Adoption and reach are implementation outcomes specified in the CFIR [
] and Reach, Effectiveness, Adoption, Implementation, and Maintenance [ ] frameworks. We expect projects to report on adoption and reach as part of the Design (if feasible) and Test phases. All projects will report adoption and reach of the intervention. For redesigned implementation strategies, the approach to measuring adoption and reach depends on the project. For example, R34 RUBIES defines adoption as educators’ first use of the RUBIES-Team at any point during the study and will measure reach for both RUBIES-Individual (ie, an original implementation strategy) or RUBIES-Team in three ways: (1) the number of other educators with whom trained educators share RUBIES strategies; (2) the number of other students with whom trained educators share RUBIES strategies; and (3) the number of other contexts and settings in which they applied RUBIES strategies.Intervention and Implementation Strategy Fidelity
Fidelity is a core implementation outcome [
]. We expect projects to report on fidelity as part of the Design (for existing clinical interventions and implementation strategies) and Test phases. Teams choose an approach to measuring fidelity based on redesign goals. For example, R34 RUBIES rates paraeducator fidelity to treatment (eg, weekly ratings of the paraeducator’s homework completion and behavior support plan implementation). R34 TF-CBT will code session recordings using the Therapy Process Observational Coding Scale [ ] and the Therapy Process Observational Coding Scale–Self-Reported Therapist Intervention Fidelity for Youth [ ] at baseline, 3 months, and 6 months. Finally, R01 PST-Aid will measure initial and sustained fidelity using a PST Fidelity Scale. Initial fidelity will be measured as the number of sessions it takes providers to get certified; providers must have 2 sessions rated as “satisfactory” on the PST Fidelity Scale to receive certification. Sustained fidelity will be measured as the number of “satisfactory” sessions during the 6 months after certification, as measured by the PST Fidelity Scale.Planned Adaptations
Characterizing adaptations, or redesign solutions, is key to all 3 Methods Core team aims to better understand and address challenges to clinical interventions and implementation strategies. We expect projects to characterize adaptations with FRAME or FRAME-IS [
] as part of the Design and Build phase, where planned and proactive changes will be made as a part of the redesign process. For example, R34 BRISC will analyze recorded intervention sessions using FRAME or FRAME-IS. The R01 PST-Aid will code randomly selected session audio recordings per client for fidelity and adaptations using FRAME or FRAME-IS. Projects will share adaptations made with the Methods Core team and describe whether these adaptions were made proactively (eg, as part of the redesign process in the design and build phases) or reactively (eg, unplanned or due to unanticipated obstacles in the test phase). Across projects, the Methods Core team will systematically categorize adaptations to examine solutions and the populations, organization types or structures, practitioner types, and clinical interventions and implementation strategies in which they work. Our objective is to identify transferable and unique solutions to different contexts and clinical interventions and implementation strategies. This information will ultimately inform an updated version of the library of redesign solutions.Unplanned or Reactive Modifications
Reactive or unplanned modifications during the Test phase will be measured using a 17-item Center-developed measure based on the FRAME or FRAME-IS to assess the nature of modifications. Providers will self-report any changes they made while they administer the intervention.
Distal Service Recipient Outcomes
During the Test phase, projects are expected to collect clinical and functional outcomes. Teams will administer standardized assessments (eg, Patient Health Questionnaire [
], Quality of Life in Neurological Disorders [Neuro-QOL] [ ], Satisfaction with Social Roles for adults [ ], and the Neuro-QOL Social Relations Scale for youth [ ]). The Neuro-QOL measures are widely used to assess functioning in usual social roles, activities, and responsibilities. The scales have been evaluated with thousands of participants from the general population of the United States and in clinical inpatient and outpatient settings who have a wide variety of presenting problems [ - ]. All projects will also use idiographic (ie, individualized) client outcome monitoring based on the Top Problems Assessment [ ], an approach informed by goal attainment scaling [ , ] that has been found to be highly sensitive for monitoring clinical intervention outcomes and thus is preferred over standardized or nomothetic assessments by both practitioners and clients [ ].Demographic and Process Measures
Demographics
All projects are expected to collect participant demographic data mandated by the NIMH at all phases. Teams collect additional demographic data on the basis of the project needs. For example, the R34 RUBIES collects required educator and student demographics (eg, age, gender, and race) and additional data on school characteristics (eg, school size, percentage eligible for free or reduced lunch, racial and ethnic composition, percentage of English language learners, percentage in special education, annual funding for external resources, and per capita number of community-based organizations). Data will be tabulated to satisfy federal demographic and data reporting requirements as well as cross-project meta-analyses and comparisons.
User Needs and Experience
All projects are expected to clearly identify direct and indirect users and incorporate methods that address user needs. Explicit user identification produces more usable products and ensures that the design team does not underestimate user diversity [
] or create designs based on the team’s own needs [ , , ]. clinical intervention and implementation strategy users should include the deliverers (eg, providers of clinical interventions) and recipients (eg, clients and implementation strategy targets such as administrators or clinicians). Identification of users for a clinical intervention and implementation strategy includes (1) generating a preliminary user list, (2) articulating the most relevant characteristics that reflect anticipated users, (3) describing and prioritizing main user groups, and (4) selecting typical and representative users [ ]. For example, in R34 TF-CBT, direct users are school counselors and social workers who provide mental health services, as well as public school students with histories of traumatic stress. Potential indirect users in this project include caregivers of students. We included users who are diverse with respect to characteristics such as age (eg, students), race and ethnicity (eg, students and practitioners), culture (eg, students and practitioners), and clinical domain experience (eg, practitioners), which are features known to impact experiences of usability, engagement, or appropriateness [ , - ].Projects will use interviews to identify key challenges that users might face when applying clinical interventions and implementation strategies. Interviews consist of questions derived from HCD principles such as organizational and community culture, values, and challenges in applying clinical interventions and implementation strategies. For example, in R34 RUBIES, the team will interview educators about their existing opportunities to learn behavioral management strategies for students with autism who exhibit challenging behavior. Interviews will identify promising professional development approaches and areas to improve the existing RUBIES multifaceted implementation strategy. Additional interviews with school administrators or lead special educators are likely to surface critical organizational factors that can serve as design constraints for any subsequent redesign solution [
, ].As described earlier, interviews can be supplemented through observation methods to better understand interactions in real-world settings. Projects can use an adapted form of TAP, where participants (eg, clinicians or clients) and researchers watch recordings of sessions while the participant explains what they were thinking in the moment. This approach can offer additional suggestions for improvement on the design of a clinical intervention or offer ideas for tools that could support the clinician during implementation. Interviews can also supplement comparative testing (eg, A/B testing) to explore and evaluate a broader landscape of design options and reach more robust solutions. A/B testing is an evaluation method in which ≥2 versions of a prototype are compared sequentially or in parallel to determine which version is easier to use and better meets user needs [
]. For example, R34 BRISC will build prototypes of digital asynchronous learning modules for novel users as well as posttraining support tools; initial prototypes will undergo comparative testing to finalize solutions to be evaluated in the Test phase. The pragmatic applicability and match of potential designs to the targeted service environments and resource constraints will be systematically addressed.Participant Research Burden, Incentive Appropriateness, and Research Satisfaction
At the end of the Test phase, projects are expected to measure the burden of participation in the study. This instrument includes 6 questions to understand participants’ perceived burden of participating in the study, appropriateness of the level of compensation offered, and overall satisfaction with the study experience. This information will be used to help improve future protocols. Response frequencies will be tabulated for the 4 close-ended responses and themes will be summarized from open-ended responses.
Adherence to DDBT Process (DDBT Fidelity and Cost Measure)
All projects are expected to complete a Fidelity and Cost survey in REDCap about their application of HCD techniques at the end of each DDBT phase. We are developing the survey to systematically collect data on how the DDBT framework guides clinical intervention and implementation strategy redesign and link design activities with project goals. To facilitate teams in drawing on a range of methods, the Fidelity and Cost survey focuses on understanding which goals of each DDBT process teams completed (
) and the methods they used to support each goal. We will conduct descriptive statistical analyses and content analysis of data to understand the frequency of goals completed and HCD strategies used, links between strategies and goals, and modifications made to strategies during their use.We will measure the costs of applying DDBT to help understand the resource requirements involved in its use, which can be a major challenge of HCD and coproduction methods [
]. Projects are expected to report total costs of redesign, reported through a Fidelity and Cost Survey in REDCap at the end of each phase (this is optional for R03 projects). We will aggregate activity-based costs (eg, time to create, complete, and analyze each activity; participant payments or time) across individuals, use budgets or other institutional records to assign hourly costs by role, and then add in any fixed costs (eg, materials and activity-specific software). We will calculate total DDBT costs, as well as phase-specific and activity-specific DDBT costs. Analyses will follow best practices by placing all US dollar values onto the same metric, including an index year to account for inflation; local or national average cost-of-living values to account for geographic variation in prices; and discounting of costs from different years due to preferences for delayed over immediate costs. We will conduct sensitivity analyses to examine the robustness of our cost estimates [ , ] by identifying areas of uncertainty in measuring units and prices for our ingredients, and then calculating costs across a range of plausible values (eg, we can substitute limits of 95% CIs for uncertain prices).Team Collaboration, Trust, and Respect
At the end of each DDBT phase, R01 and R34s team members are expected to complete a survey that assesses satisfaction with the collaboration, impact of collaboration, trust, and respect. This survey is modeled after the Transdisciplinary Tobacco Use Research Center’s measure of Team Collaboration and Transdisciplinary Integration, which assesses satisfaction with the collaboration, impact of collaboration, trust, and respect [
]. Continuous review of outcomes will allow for critical assessment and course correction as needed and recommended by these bodies. Participation is confidential, and teams will receive an aggregate report of the number of team member participants and average scores for each item. Any free-response comments are additionally summarized. We will encourage teams to discuss results to improve their projects.Community Participation in Research
The collaboration survey questions described earlier will be administered with additional questions to characterize the extent to which redesign teams engage users in a localized DDBT process. This instrument is based on an existing measure of community participation in research [
, ], which has been modified to target the design of clinical interventions and implementation strategies across 6 dimensions: identification of design issues, design activities, use of resources, design methods, indicators of success, and sustainability. Redesign teams will complete the measure at or near the end of each DDBT phase and then discuss ratings in an interview.Investigator Satisfaction With the Support They Receive From the Center
At the conclusion of projects, we will ask investigators to share their level of satisfaction with support from UWAC through a 5-item Likert survey adapted from a survey used by the University of Washington IMPACT Center [
]. We intend to use these data to improve how the Methods Core team provides projects technical support.Data Analysis
The Methods Core team provides data management and guidance on all DDBT constructs. For Aim 3, fundamental comparisons are the differences in DDBT mechanisms (usability, engagement, and appropriateness) and implementation outcomes (adoption, reach, adaptation, and fidelity [and sustainment for the R01 PST]) for the original (unadapted) clinical intervention and implementation strategy versus the DDBT-informed (localized) clinical intervention and implementation strategy. We will conduct a (1) qualitative multiple case study analysis and (2) quantitative meta-analysis across projects. Case studies will examine each project’s context, implementation, mechanisms, and outcomes.
For each project, we will also develop analytical summaries to facilitate between-project comparisons. Using the constant comparative method [
], we will compare projects to group common and divergent themes. The meta-analytic synthesis will increase our inferential ability by combining results from the underpowered R34s. For the meta-analysis, each project’s mechanism and outcome will be summarized as a Cohen d effect size comparing localized DDBT with original clinical intervention and implementation strategy and corresponding 95% CI, using random effects weighting by the inverse of the within- and between-studies variance. Standard data screening and adjustments will be made to the data (eg, to limit the effect of outliers, they will be winsorized). Each project will be additionally advised on how to address possible confounders in analysis and reporting. For instance, projects will be encouraged to use naive participants in the Design and Build and Test phases, recognizing that adaptations of smaller elements of complex clinical interventions may require the participation of experienced participants during the Design and Build phase. In addition, randomization will occur at appropriate levels to avoid contamination by intralevel communication. Nonmonotonic missing data will be addressed via inverse probability weighting or multiple imputation, as appropriate [ ]. The Methods Core will aggregate these data across projects into a series of working meta-analyses of the effectiveness of DDBT on each mechanism and outcome.Finally, determining whether a DDBT-modified clinical intervention and implementation strategy leads to better implementation and clinical outcomes is ultimately a question of mediation. Although the initial R34 and R03 studies are not likely to yield large sufficient sample sizes to meaningfully test such an implementation mechanism question, the R01 PST will provide a direct test of the DDBT theory of change (
). Aggregating project data over time will allow us to eventually test a range of mediation-focused hypotheses via multivariate network meta-analyses.Incorporating DEI
During UWAC 2.0, we are improving the integration of DEI initiatives throughout Center activities. Projects selected for UWAC 2.0 and pilots must demonstrate potential substantial impact on clinical or public health outcomes, especially for historically marginalized communities. The Methods Core team provides mentorship on incorporating and adapting methods so that teams are positioned to conduct research that respects diverse populations and maximizes community benefits. Project teams will be provided training and consultation on the Adapting strategies to promote implementation reach and equity method, a 3-step process for adapting implementation science to promote equity, and expertise in methods for explicitly incorporating equity into the measurement of implementation outcomes [
]. We will also collaborate with projects to ensure diverse representation and decision-making during the DDBT phases, crucial stages where diverse viewpoints and demographically representative samples are essential. To facilitate diverse engagement, we will offer resources on building equitable research-practice partnerships, contextualizing implementation science to specific communities, and enhancing community collaborator capacities for community-engaged research. Consultation on quantitative critical research [ ] will be provided to examine the treatment of race within quantitative methods and support equity testing through disaggregation, moderator exploration, and mixed methods triangulation.UWAC additionally supports faculty and staff as part of its DEI work. The center team engages with historically marginalized investigators in planning and conducting center activities so that DEI efforts are integrated throughout center mentoring, pilot funding, methods support, and support for investigators planning future proposals. These measures include enhancing communication strategies based on team science [
] and avoiding a “minority tax,” which refers to assigning additional responsibilities to marginalized or underrepresented team members to promote diversity [ , ]. The Methods Core team also advises investigators on using patient-centered and nonstigmatizing language when reporting findings.Ethical Considerations
The University of Washington Institutional Review Board (IRB) approved materials and procedures for all 4 core projects and deemed projects minimal risk by June 2024 (PST-Aid: STUDY00017272; RUBIES: STUDY00017261; TF-CBT: STUDY00017262 and STUDY00019451; and BRISC: STUDY00017263 and STUDY00019682). All studies follow best practices across studies: review and collect informed consent and Health Insurance Portability and Accountability Act authorizations (when not waived) from participants; collect parental assent for individuals <18 years of age; compensate participants financially for their time and, when appropriate, with continuing education credits; and ensure participants are aware that they may opt out or leave the study at any time. When personal health information (eg, name, date of birth, and contact information) is collected, we preserve participant privacy and confidentiality by storing those identifiers separately from the study data and only linking them to study data via a code. The Methods Core team will provide support to pilot project teams on institutional review board applications after studies are funded.
Results
Overview
UWAC 2.0, including the 4 core projects detailed in
, received funding in June 2023. We provide a brief synopsis of each study’s progress as of January 2025 in the subsequent sections. Across studies, as well as the to-be-funded R03 pilots, we anticipate DDBT will result in changes to the clinical intervention and implementation strategy mechanisms, proximal implementation outcomes (ie, adoption, fidelity, reach, and adaptation), and clinical outcomes.R01 PST-Aid
This study began in the Design and Build phase, as the Discover phase was completed during an R34 in UWAC 1.0. For the Discover phase, the study team has completed a codesign workshop and user testing with patients and providers from a nonprofit, health informatics network with independent community-based health centers. The study is beginning the Test phase, and the team is recruiting the first cohort of providers. Providers will be randomized to receive training in either PST implementation as usual (ie, training and sessions supported by paper worksheets for practitioners and clients to use as they complete PST) or PST-Aid (ie, a web-based app that promotes practitioner-client collaboration in the use of PST for goal setting and action planning).
R34 TF-CBT
At the time of submission, the study team has completed a task analysis of unadapted TF-CBT to prioritize components to develop scenarios for testing with students and school-based mental health practitioners. The study has finalized TF-CBT user-testing scenarios and are actively recruiting students and practitioners to begin the first of 3 waves of testing for the study. These data will then inform the redesign of TF-CBT, which aims to be more usable in school settings.
R34 RUBIES
Discover and Design and Build phase study activities, including focus groups, cognitive walkthroughs, and user testing with paraeducators and other school personnel (eg, principals and teachers), are complete. These data are being used to inform the Test phase where the study team is recruiting paraeducators and students and their caregivers to begin RUBIES training. Once recruited, this Test phase will include a 2-year randomized controlled trial enrolling paraeducators and students who will then be randomized to 1 of 2 implementation strategies: RUBIES-Individual or RUBIES-Team.
R34 BRISC
This study is scheduled to begin in April 2025, and the study team has received institutional review board approval.
Discussion
Charting New Research Directions
Our vision is to address persistent issues with usability, engagement, and appropriateness that are barriers to clinical intervention and implementation strategy use by drawing from the fields of HCD and implementation science. The first iteration of the center advanced our understanding of how DDBT can guide clinical intervention and implementation strategy adaptations for uptake in historically marginalized communities. During UWAC 2.0, we aim to continue serving as a multidisciplinary incubator to find viable solutions for improving implementation of clinical interventions and implementation strategies using DDBT through the R01, R34s, and pilot projects.
summarizes potential outputs and future directions by aim. Focusing on usability, engagement, and appropriateness and providing and testing ways to measure engagement in a clinical intervention and implementation strategy context is particularly novel. UWAC 2.0 will further test the robustness of the DDBT theory of change, expand the potential evidence base for its utility in combining HCD and implementation science for clinical intervention and implementation strategy redesign, and add to the field’s understanding of how to apply DDBT to a variety of clinical intervention and implementation strategies and contexts.Our experiences underscore a benefit of developing additional resources for DDBT, HCD, and implementation science. UWAC 1.0 outputs contributed to foundational conversations on the intersection of HCD and implementation science and produced potential pathways to address conceptual overlap and distinctions [
] and terminology [ ]. We additionally developed resources such as CWIS [ ], IUS [ ], the ISUS [ ], and a typology of modification targets and usability issues [ ]. We have accumulated substantial experience adapting common HCD methods for different contexts and communities, and we hope these adaptations can support future teams in their use these methods. We plan to build on these methodological advancements, which have been used beyond the UWAC team by a broader research community interested in HCD and implementation science methods and measurement. As we learn with UWAC 2.0 project teams, we will continue to identify and develop additional resources for UWAC project teams and the broader research community on specific methods, team science, equity-oriented design practices, and grant writing. Developing additional resources aligns with the Methods Core team shifting to a consultation model on using the DDBT framework during UWAC 2.0. UWAC projects receive technical support from the Methods Core team with greater emphasis on building local capacity to apply DDBT rather than having the Methods Core team members conduct some of the DDBT activities.Aim | Expected outputs | Potential future work |
Aim 1: identify clinical intervention and implementation strategy modification targets to improve usability, engagement, and appropriateness in accessible nonspecialty settings (Discover phase) |
|
|
Aim 2: develop redesign solutions with local teams to address clinical intervention and implementation strategy modification targets (Design and Build phase) |
|
|
Aim 3: determine if redesign affects changes in usability, engagement, and appropriateness (Test phase) |
|
|
aUWAC: University of Washington Advanced Laboratories for Accelerating the Reach and Impact of Treatments for Youth and Adults with Mental Illness Center.
bDDBT: Discover, Design and Build, and Test.
Limitations
UWAC 2.0 and the DDBT framework represent a robust effort to integrate HCD with implementation science, although several methodological limitations merit consideration. The standardized measures we are using to assess usability, appropriateness, and engagement have evidence of psychometric soundness, but these measures have not yet been evaluated in all the contexts in which they will be applied. Redesign solutions developed by local teams may rely primarily on context-specific adaptations, which may not generalize to other settings and may not address systemic barriers that impact usability or accessibility. Therefore, reactive modifications (ie, primary outcome) needed to address additional barriers may not be reduced. In this work, there is a tension between our desire to standardize DDBT phases and measures, to facilitate center-wide learning, with promoting the adaptability and flexibility to adapt to specific goals, contexts, and populations that is necessary for good design projects. This, combined with the variability of research project contexts, the small sample sizes of redesign teams and subjective nature of the proposed mechanisms may impede our ability to make inferences across projects.
Conclusions and Impact
There is a pressing need to ensure that clinical interventions and implementation strategies are easily implementable and meet the needs of the communities they aim to help. Integrating HCD and implementation science offers promising approaches to tackle this challenge. UWAC 2.0 expands and strengthens our efforts to ensure that accessible community service settings and marginalized communities see the benefit of decades of research on effective clinical interventions and implementation strategies.
Acknowledgments
This project is funded by the National Institute of Mental Health (2 P50MH115837; principal investigators: ARL and SAM). The funding body had no role in the study design, writing of the manuscript, or decision to submit the paper for publication.
Data Availability
Clinical outcome data for the R01s and R34s will be made publicly available through the National Institute of Mental Health National Data Archive. The quantitative datasets generated and analyzed during this study are available from the corresponding author on reasonable request.
Authors' Contributions
All authors contributed to conceptualizing the study. ARL, SAM, and MDP led funding acquisition and created the protocol with support from the team on methodology. MDP led plans for data curation and formal analysis. ARL, SAM, BM, and TA wrote the original draft with review and editing support from all authors. BM and KPO oversaw project administration.
Conflicts of Interest
KB reports royalties from Oxford University Press.
Peer-review report by the National Institute of Mental Health Special Emphasis Panel - National Institute of Mental Health - Advanced Laboratories for Accelerating the Reach and Impact of Treatments for Youth and Adults with Mental Illness (National Institutes of Health, USA).
PDF File (Adobe PDF File), 211 KBSummary of projects.
DOCX File , 32 KBCenter measures and guidance.
DOCX File , 29 KBReferences
- Arean PA, Raue PJ, Sirey JA, Snowden M. Implementing evidence-based psychotherapies in settings serving older adults: challenges and solutions. Psychiatr Serv. Jun 2012;63(6):605-607. [FREE Full text] [CrossRef] [Medline]
- Houle J, Villaggi B, Beaulieu MD, Lespérance F, Rondeau G, Lambert J. Treatment preferences in patients with first episode depression. J Affect Disord. May 2013;147(1-3):94-100. [FREE Full text] [CrossRef] [Medline]
- McHugh RK, Whitton SW, Peckham AD, Welge JA, Otto MW. Patient preference for psychological vs pharmacologic treatment of psychiatric disorders. J Clin Psychiatry. Jun 15, 2013;74(06):595-602. [CrossRef]
- Quiñones AR, Thielke SM, Beaver KA, Trivedi RB, Williams EC, Fan VS. Racial and ethnic differences in receipt of antidepressants and psychotherapy by veterans with chronic depression. Psychiatr Serv. Feb 01, 2014;65(2):193-200. [FREE Full text] [CrossRef] [Medline]
- Raue PJ, Weinberger MI, Sirey JA, Meyers BS, Bruce ML. Preferences for depression treatment among elderly home health care patients. Psychiatr Serv. May 2011;62(5):532-537. [FREE Full text] [CrossRef] [Medline]
- Alegría M. Disparities in child and adolescent mental health and mental health services in the U.S. William T. Grant Foundation. 2015. URL: http://cfs.cbcs.usf.edu/projects-research/_docs/Disparities_in_child_and_adolescent_health.pdf [accessed 2024-04-29]
- Chandra A, Scott MM, Jaycox LH, Meredith LS, Tanielian T, Burnam A. Racial/ethnic differences in teen and parent perspectives toward depression treatment. J Adolesc Health. Jun 2009;44(6):546-553. [CrossRef] [Medline]
- Dwight Johnson M, Apesoa-Varano C, Hay J, Unutzer J, Hinton L. Depression treatment preferences of older white and Mexican origin men. Gen Hosp Psychiatry. Jan 2013;35(1):59-65. [FREE Full text] [CrossRef] [Medline]
- Fernandez Y Garcia E, Franks P, Jerant A, Bell RA, Kravitz RL. Depression treatment preferences of Hispanic individuals: exploring the influence of ethnicity, language, and explanatory models. J Am Board Fam Med. 2011;24(1):39-50. [FREE Full text] [CrossRef] [Medline]
- Givens JL, Houston TK, van Voorhees BW, Ford DE, Cooper LA. Ethnicity and preferences for depression treatment. Gen Hosp Psychiatry. May 2007;29(3):182-191. [CrossRef] [Medline]
- Lu W, Todhunter-Reid A, Mitsdarffer ML, Muñoz-Laboy M, Yoon AS, Xu L. Barriers and facilitators for mental health service use among racial/ethnic minority adolescents: a systematic review of literature. Front Public Health. 2021;9:641605. [FREE Full text] [CrossRef] [Medline]
- Mossakowski KN. Dissecting the influence of race, ethnicity, and socioeconomic status on mental health in young adulthood. Res Aging. Sep 10, 2008;30(6):649-671. [CrossRef]
- Nadeem E, Lange JM, Miranda J. Mental health care preferences among low-income and minority women. Arch Womens Ment Health. Jun 8, 2008;11(2):93-102. [FREE Full text] [CrossRef] [Medline]
- Sleath B, West S, Tudor G, Perreira K, King V, Morrissey J. Ethnicity and depression treatment preferences of pregnant women. J Psychosom Obstet Gynaecol. Jun 2005;26(2):135-140. [FREE Full text] [CrossRef] [Medline]
- Villatoro AP, Mays VM, Ponce NA, Aneshensel CS. Perceived need for mental health care: the intersection of race, ethnicity, gender, and socioeconomic status. Soc Ment Health. Mar 01, 2018;8(1):1-24. [FREE Full text] [CrossRef] [Medline]
- Wagner JA, Perkins DW, Piette JD, Lipton B, Aikens JE. Racial differences in the discussion and treatment of depressive symptoms accompanying type 2 diabetes. Diabetes Res Clin Pract. Nov 2009;86(2):111-116. [FREE Full text] [CrossRef] [Medline]
- Duong MT, Bruns EJ, Lee K, Cox S, Coifman J, Mayworm A, et al. Rates of mental health service utilization by children and adolescents in schools and other common service settings: a systematic review and meta-analysis. Adm Policy Ment Health. May 2021;48(3):420-439. [CrossRef] [Medline]
- Olfson M, Kroenke K, Wang S, Blanco C. Trends in office-based mental health care provided by psychiatrists and primary care physicians. J Clin Psychiatry. Mar 15, 2014;75(03):247-253. [CrossRef]
- Owens JS, Lyon AR, Brandt NE, Warner CM, Nadeem E, Spiel C, et al. Implementation science in school mental health: key constructs in a developing research agenda. School Ment Health. May 01, 2014;6(2):99-111. [FREE Full text] [CrossRef] [Medline]
- Rones M, Hoagwood K. School-based mental health services: a research review. Clin Child Fam Psychol Rev. 2000;3(4):223-241. [CrossRef]
- Jane M, Butler AS, Gonzalez ML. Psychosocial Interventions for Mental and Substance Use: Disorders A Framework for Establishing Evidence-Based Standards. Washington, DC. National Academies Press; 2015.
- Powell BJ, McMillen JC, Proctor EK, Carpenter CR, Griffey RT, Bunger AC, et al. A compilation of strategies for implementing clinical innovations in health and mental health. Med Care Res Rev. Apr 26, 2012;69(2):123-157. [FREE Full text] [CrossRef] [Medline]
- Cook CR, Lyon AR, Locke J, Waltz T, Powell BJ. Adapting a compilation of implementation strategies to advance school-based implementation research and practice. Prev Sci. Aug 31, 2019;20(6):914-935. [FREE Full text] [CrossRef] [Medline]
- Powell BJ, Proctor EK, Glass JE. A systematic review of strategies for implementing empirically supported mental health interventions. Res Soc Work Pract. Mar 01, 2014;24(2):192-212. [FREE Full text] [CrossRef] [Medline]
- Weiner BJ, Lewis CC, Sherr K. Practical Implementation Science: Moving Evidence into Action. Cham, Switzerland. Springer; 2022.
- Stirman SW, Miller CJ, Toder K, Calloway A. Development of a framework and coding system for modifications and adaptations of evidence-based interventions. Implement Sci. Jun 10, 2013;8(1):65. [FREE Full text] [CrossRef] [Medline]
- Collins LM, Murphy SA, Bierman KL. A conceptual framework for adaptive preventive interventions. Prev Sci. Sep 2004;5(3):185-196. [FREE Full text] [CrossRef] [Medline]
- Chambers DA, Glasgow RE, Stange KC. The dynamic sustainability framework: addressing the paradox of sustainment amid ongoing change. Implement Sci. Oct 02, 2013;8(1):117. [FREE Full text] [CrossRef] [Medline]
- Miller CJ, Barnett ML, Baumann AA, Gutner CA, Wiltsey-Stirman S. The FRAME-IS: a framework for documenting modifications to implementation strategies in healthcare. Implement Sci. Apr 07, 2021;16(1):36. [FREE Full text] [CrossRef] [Medline]
- Healey P, Stager ML, Woodmass K, Dettlaff AJ, Vergara A, Janke R, et al. Cultural adaptations to augment health and mental health services: a systematic review. BMC Health Serv Res. Jan 05, 2017;17(1):8. [FREE Full text] [CrossRef] [Medline]
- Saha S, Beach MC, Cooper LA. Patient centeredness, cultural competence and healthcare quality. J Natl Med Assoc. Nov 2008;100(11):1275-1285. [FREE Full text] [CrossRef] [Medline]
- Giacomin J. What is human centred design? Design J. Dec 01, 2014;17(4):606-623. [CrossRef]
- What is human-centered design (HCD)? Interaction Design Foundation. URL: https://www.interaction-design.org/literature/topics/human-centered-design [accessed 2024-03-10]
- Rogers Y. HCI Theory: Classical, Modern, and Contemporary. New York, NY. Morgan & Claypool Publishers; 2012.
- Lyon AR, Dopp AR, Brewer SK, Kientz JA, Munson SA. Designing the future of children's mental health services. Adm Policy Ment Health. Sep 2020;47(5):735-751. [FREE Full text] [CrossRef] [Medline]
- Lyon AR, Koerner K. User-centered design for psychosocial intervention development and implementation. Clin Psychol (New York). Jun 17, 2016;23(2):180-200. [FREE Full text] [CrossRef] [Medline]
- Lyon A, Chung J, Koerner K. How implementable is that evidence-based practice?: a methodology for assessing complex innovation usability. AcademyHealth. 2018. URL: https://academyhealth.confex.com/academyhealth/2018di/meetingapp.cgi/Paper/28235 [accessed 2024-03-09]
- Alexopoulos GS, Raue PJ, Gunning F, Kiosses DN, Kanellopoulos D, Pollari C, et al. "Engage" therapy: behavioral activation and improvement of late-life major depression. Am J Geriatr Psychiatry. Apr 2016;24(4):320-326. [FREE Full text] [CrossRef] [Medline]
- Lyon AR, Coifman J, Cook H, McRee E, Liu FF, Ludwig K, et al. The Cognitive Walkthrough for Implementation Strategies (CWIS): a pragmatic method for assessing implementation strategy usability. Implement Sci Commun. Jul 17, 2021;2(1):78. [FREE Full text] [CrossRef] [Medline]
- Mohr DC, Lyon AR, Lattie EG, Reddy M, Schueller SM. Accelerating digital mental health research from early design and creation to successful implementation and sustainment. J Med Internet Res. May 10, 2017;19(5):e153. [FREE Full text] [CrossRef] [Medline]
- Lyon A, Munson SA, Reddy M, Schueller SM, Agapie E, Yarosh S, et al. Bridging HCI and implementation science for innovation adoption and public health impact. Ext Abstr Hum Factors Computing Syst. Apr 2023;2023:56. [FREE Full text] [CrossRef] [Medline]
- Dopp AR, Parisi KE, Munson SA, Lyon AR. A glossary of user-centered design strategies for implementation experts. Transl Behav Med. Nov 25, 2019;9(6):1057-1064. [CrossRef] [Medline]
- Dopp AR, Parisi KE, Munson SA, Lyon AR. Integrating implementation and user-centred design strategies to enhance the impact of health services: protocol from a concept mapping study. Health Res Policy Syst. Jan 08, 2019;17(1):1. [FREE Full text] [CrossRef] [Medline]
- Chen E, Neta G, Roberts MC. Complementary approaches to problem solving in healthcare and public health: implementation science and human-centered design. Transl Behav Med. May 25, 2021;11(5):1115-1121. [FREE Full text] [CrossRef] [Medline]
- Wiltsey Stirman S, Baumann AA, Miller CJ. The FRAME: an expanded framework for reporting adaptations and modifications to evidence-based interventions. Implement Sci. Jun 06, 2019;14(1):58. [FREE Full text] [CrossRef] [Medline]
- Lyon AR, Munson SA, Renn BN, Atkins DC, Pullmann MD, Friedman E, et al. Use of human-centered design to improve implementation of evidence-based psychotherapies in low-resource communities: protocol for studies applying a framework to assess usability. JMIR Res Protoc. Oct 09, 2019;8(10):e14990. [FREE Full text] [CrossRef] [Medline]
- Lyon AR, Bruns EJ. User-centered redesign of evidence-based psychosocial interventions to enhance implementation-hospitable soil or better seeds? JAMA Psychiatry. Jan 01, 2019;76(1):3-4. [CrossRef] [Medline]
- Areán PA, Renn B, Ratzliff A. Making psychotherapy available in the United States: implementation challenges and solutions. Psychiatr Serv. Feb 01, 2021;72(2):222-224. [CrossRef] [Medline]
- Bearss K, Tagavi D, Lyon AR, Locke J. Iterative redesign of a caregiver-mediated intervention for use in educational settings. Autism. Apr 2022;26(3):666-677. [FREE Full text] [CrossRef] [Medline]
- ISO 9241-210:2019(en): ergonomics of human-system interaction — part 210: human-centred design for interactive systems. International Organization for Standardization. URL: https://www.iso.org/obp/ui/en/#iso:std:iso:9241:-210:ed-2:v1:en [accessed 2023-12-14]
- Cooper A, Reimann R, Cronin D. About Face 3: The Essentials of Interaction Design. 3rd edition. Hoboken, NJ. John Wiley & Sons; 2007.
- Grudin J, Pruitt J. Personas, participatory design and product development: an infrastructure for engagement. In: Proceedings of the 7th Biennial Participatory Design Conference. 2002. Presented at: PDC '02; June 23-25, 2002:144-152; Malmø, Sweden. URL: https://ojs.ruc.dk/index.php/pdc/article/view/249 [CrossRef]
- Munson SA, Friedman EC, Osterhage K, Allred R, Pullmann MD, Areán PA, et al. Usability issues in evidence-based psychosocial interventions and implementation strategies: cross-project analysis. J Med Internet Res. Jun 14, 2022;24(6):e37585. [FREE Full text] [CrossRef] [Medline]
- Pérez D, van der Stuyft P, Zabala M del C, Castro M, Lefèvre P. A modified theoretical framework to assess implementation fidelity of adaptive public health interventions. Implement Sci. Jul 08, 2016;11(1):91. [FREE Full text] [CrossRef] [Medline]
- Kumpfer KL, Scheier LM, Brown J. Strategies to avoid replication failure with evidence-based prevention interventions: case examples from the strengthening families program. Eval Health Prof. Jun 2020;43(2):75-89. [FREE Full text] [CrossRef] [Medline]
- Barrera Jr M, Castro FG. A heuristic framework for the cultural adaptation of interventions. Clin Psychol Sci Pract. 2006;13(4):311-316. [CrossRef]
- Doherty K, Doherty G. Engagement in HCI: conception, theory and measurement. ACM Comput Surv. Nov 19, 2018;51(5):1-39. [CrossRef]
- Proctor E, Silmere H, Raghavan R, Hovmand P, Aarons G, Bunger A, et al. Outcomes for implementation research: conceptual distinctions, measurement challenges, and research agenda. Adm Policy Ment Health. Mar 2011;38(2):65-76. [FREE Full text] [CrossRef] [Medline]
- Westfall JM, Mold J, Fagnan L. Practice-based research--"Blue Highways" on the NIH roadmap. JAMA. Jan 24, 2007;297(4):403-406. [CrossRef] [Medline]
- 10 usability heuristics for user interface design. Nielsen Norman Group. URL: https://www.nngroup.com/articles/ten-usability-heuristics/ [accessed 2024-11-15]
- Rubin J, Chisnell D, Spool J. Handbook of Usability Testing: How to Plan, Design, and Conduct Effective Tests. Hoboken, NJ. John Wiley & Sons; 2008.
- Lyon AR, Aung T, Bruzios K, Munson SA. Human-centered design to enhance implementation and impact in health. Annu Rev Public Health. Dec 10, 2024:46. [CrossRef] [Medline]
- Lewis JR. The system usability scale: past, present, and future. Int J Hum Comput Interact. Mar 30, 2018;34(7):577-590. [CrossRef]
- Boren T, Ramey J. Thinking aloud: reconciling theory and practice. IEEE Trans Profess Commun. 2000;43(3):261-278. [CrossRef]
- Jääskeläinen R. Think-aloud protocol. In: Gambier Y, van Doorslaer L, editors. Handbook of Translation Studies. New York, NY. John Benjamins Publishing; 2010:371-373.
- Moullin JC, Sabater-Hernández D, García-Corpas JP, Kenny P, Benrimoj SI. Development and testing of two implementation tools to measure components of professional pharmacy service fidelity. J Eval Clin Pract. Jun 2016;22(3):369-377. [CrossRef] [Medline]
- Berkel C, Mauricio AM, Schoenfelder E, Sandler IN. Putting the pieces together: an integrated model of program implementation. Prev Sci. Mar 2011;12(1):23-33. [FREE Full text] [CrossRef] [Medline]
- Longtin Y, Sax H, Leape LL, Sheridan SE, Donaldson L, Pittet D. Patient participation: current knowledge and applicability to patient safety. Mayo Clin Proc. Jan 2010;85(1):53-62. [FREE Full text] [CrossRef] [Medline]
- Doherty C, Stavropoulou C. Patients' willingness and ability to participate actively in the reduction of clinical errors: a systematic literature review. Soc Sci Med. Jul 2012;75(2):257-263. [FREE Full text] [CrossRef] [Medline]
- Kruk ME, Gage AD, Arsenault C, Jordan K, Leslie HH, Roder-DeWan S, et al. High-quality health systems in the sustainable development goals era: time for a revolution. Lancet Glob Health. Nov 2018;6(11):e1196-e1252. [FREE Full text] [CrossRef] [Medline]
- 2019 national healthcare quality and disparities report. Agency for Healthcare Research and Quality. 2020. URL: https://www.ncbi.nlm.nih.gov/books/NBK579353/ [accessed 2024-06-23]
- Delivering quality health services: a global imperative for universal health coverage. World Health Organization. 2018. URL: https://www.who.int/publications/i/item/9789241513906 [accessed 2024-04-29]
- Garett R, Chiu J, Zhang L, Young SD. A literature review: website design and user engagement. Online J Commun Media Technol. Jul 2016;6(3):1-14. [FREE Full text] [Medline]
- Wei Y, Zheng P, Deng H, Wang X, Li X, Fu H. Design features for improving mobile health intervention user engagement: systematic review and thematic analysis. J Med Internet Res. Dec 09, 2020;22(12):e21687. [FREE Full text] [CrossRef] [Medline]
- Zanjani N. The important elements of LMS design that affect user engagement with e-learning tools within LMSs in the higher education sector. Australas J Educ. Nov 09, 2016;33(1):17. [FREE Full text] [CrossRef]
- Meissner P, Cottler LB, Eder MM, Michener JL. Engagement science: the core of dissemination, implementation, and translational research science. J Clin Transl Sci. Jan 20, 2020;4(3):216-218. [FREE Full text] [CrossRef] [Medline]
- National Research Council. Enhancing the Effectiveness of Team Science. New York, NY. National Academies Press; 2015.
- Agapie E, Haldar S, Poblete SG. Using HCI in cross-disciplinary teams: a case study of academic collaboration in HCI-health teams in the US using a team science perspective. Proc ACM Hum Comput Interact. Nov 11, 2022;6(CSCW2):1-35. [CrossRef]
- Weiner BJ, Lewis CC, Stanick C, Powell BJ, Dorsey CN, Clary AS, et al. Psychometric assessment of three newly developed implementation outcome measures. Implement Sci. Aug 29, 2017;12(1):108. [FREE Full text] [CrossRef] [Medline]
- Damschroder LJ, Aron DC, Keith RE, Kirsh SR, Alexander JA, Lowery JC. Fostering implementation of health services research findings into practice: a consolidated framework for advancing implementation science. Implement Sci. Aug 07, 2009;4:50. [FREE Full text] [CrossRef] [Medline]
- Curran GM, Bauer M, Mittman B, Pyne JM, Stetler C. Effectiveness-implementation hybrid designs: combining elements of clinical effectiveness and implementation research to enhance public health impact. Med Care. Mar 2012;50(3):217-226. [FREE Full text] [CrossRef] [Medline]
- Landes SJ, McBain SA, Curran GM. An introduction to effectiveness-implementation hybrid designs. Psychiatry Res. Oct 2019;280:112513. [FREE Full text] [CrossRef] [Medline]
- Arean P, Hull D, Pullmann MD, Heagerty PJ. Protocol for a sequential, multiple assignment, randomised trial to test the effectiveness of message-based psychotherapy for depression compared with telepsychotherapy. BMJ Open. Nov 02, 2021;11(11):e046958. [FREE Full text] [CrossRef] [Medline]
- Albin RW, Lucyshyn JM, Horner RH, Flannery KB. Contextual fit for behavioral support plans: a model for "goodness of fit. In: Koegel LK, Koegel RL, Dunlap G, editors. Positive Behavioral Support: Including People with Difficult Behavior in the Community. New York, NY. Brookes Publishlnig; 1996:81-98.
- Glasgow RE, Vogt TM, Boles SM. Evaluating the public health impact of health promotion interventions: the RE-AIM framework. Am J Public Health. Sep 1999;89(9):1322-1327. [CrossRef] [Medline]
- Dorsey S, Pullmann MD, Deblinger E, Berliner L, Kerns SE, Thompson K, et al. Improving practice in community-based settings: a randomized trial of supervision - study protocol. Implement Sci. Aug 10, 2013;8:89. [FREE Full text] [CrossRef] [Medline]
- Becker-Haimes EM, Klein MR, McLeod BD, Schoenwald SK, Dorsey S, Hogue A, et al. The TPOCS-self-reported therapist intervention fidelity for youth (TPOCS-SeRTIFY): a case study of pragmatic measure development. Implement Res Pract. 2021;2:2633489521992553. [FREE Full text] [CrossRef] [Medline]
- Kroenke K, Spitzer RL, Williams JB. The PHQ-9: validity of a brief depression severity measure. J Gen Intern Med. Sep 2001;16(9):606-613. [FREE Full text] [CrossRef] [Medline]
- Carlozzi NE, Goodnight R, Kratz AL, Stout JC, McCormack MK, Paulsen JS, et al. Validation of neuro-QoL and PROMIS mental health patient reported outcome measures in persons with Huntington disease. J Huntington Dis. 2019;8(4):467-482. [FREE Full text]
- Gershon RC, Lai JS, Bode R, Choi S, Moy C, Bleck T, et al. Neuro-QOL: quality of life item banks for adults with neurological disorders: item development and calibrations based upon clinical and general population testing. Qual Life Res. Apr 2012;21(3):475-486. [FREE Full text] [CrossRef] [Medline]
- Lai JS, Nowinski C, Victorson D, Bode R, Podrabsky T, McKinney N, et al. Quality-of-life measures in children with neurological conditions: pediatric Neuro-QOL. Neurorehabil Neural Repair. Jan 2012;26(1):36-47. [FREE Full text] [CrossRef] [Medline]
- Weisz JR, Chorpita BF, Frye A, Ng MY, Lau N, Bearman SK, et al. Youth Top Problems: using idiographic, consumer-guided assessment to identify treatment needs and to track change during psychotherapy. J Consult Clin Psychol. Jun 2011;79(3):369-380. [CrossRef] [Medline]
- Kiresuk TJ, Sherman RE. Goal attainment scaling: a general method for evaluating comprehensive community mental health programs. Community Ment Health J. Dec 17, 1968;4(6):443-453. [CrossRef] [Medline]
- Kiresuk TJ, Smith A, Cardillo JE. Goal Attainment Scaling: Applications, Theory, and Measurement. Berlin, Germany. Lawrence Erlbaum Associates; 1994.
- Lindhiem O, Bennett CB, Orimoto TE, Kolko DJ. A meta-analysis of personalized treatment goals in psychotherapy: a preliminary report and call for more studies. Clin Psychol (New York). Jun 2016;23(2):165-176. [FREE Full text] [CrossRef] [Medline]
- Kujala S, Kauppinen M. Identifying and selecting users for user-centered design. In: Proceedings of the 3rd Nordic conference on Human-computer interaction. 2004. Presented at: NordiCHI '04; October 23-27, 2004:297-303; Tampere, Finland. URL: https://dl.acm.org/doi/10.1145/1028014.1028060
- Cooper A. The Inmates Are Running the Asylum: Why High Tech Products Drive Us Crazy and How to Restore the Sanity. Carmel, IN. Sams Publishing; 2004.
- Kujala S, Mäntylä M. How effective are user studies? In: McDonald S, Waern Y, Cockton G, editors. People and Computers XIV — Usability or Else! Cham, Switzerland. Springer; 2000:61-71.
- Lyon AR, Koerner K, Chung J. Usability evaluation for evidence-based psychosocial interventions (USE-EBPI): a methodology for assessing complex intervention implementability. Implement Res Pract. 2020;1:2633489520932924. [FREE Full text] [CrossRef] [Medline]
- Sonderegger A, Schmutz S, Sauer J. The influence of age in usability testing. Appl Ergon. Jan 2016;52:291-300. [CrossRef] [Medline]
- Clemmensen T, Shi Q, Kumar J, Li H, Sun X, Yammiyavar P. Cultural usability tests – how usability tests are not the same all over the world. In: Proceedings of the 2nd International Conference on Usability and Internationalization, UI-HCII 2007, held as Part of HCI International 2007. 2007. Presented at: UI-HCII '07; July 22-27, 2007:281-290; Beijing, China. URL: https://link.springer.com/chapter/10.1007/978-3-540-73287-7_35 [CrossRef]
- Locke J, Lee K, Cook CR, Frederick L, Vázquez-Colón C, Ehrhart MG, et al. Understanding the organizational implementation context of schools: a qualitative study of school district administrators, principals, and teachers. School Ment Health. Sep 2019;11(3):379-399. [FREE Full text] [CrossRef] [Medline]
- Locke J, Lawson GM, Beidas RS, Aarons GA, Xie M, Lyon AR, et al. Individual and organizational factors that affect implementation of evidence-based practices for children with autism in public schools: a cross-sectional observational study. Implement Sci. Mar 13, 2019;14(1):29. [FREE Full text] [CrossRef] [Medline]
- Albert B, Tullis T. Measuring the User Experience: Collecting, Analyzing, and Presenting Usability Metrics. 2nd edition. New York, NY. Morgan Kaufmann Publishers; 2013.
- Oliver K, Kothari A, Mays N. The dark side of coproduction: do the costs outweigh the benefits for health research? Health Res Policy Syst. Mar 28, 2019;17(1):33. [FREE Full text] [CrossRef] [Medline]
- Briggs AH, Gray AM. Handling uncertainty in economic evaluations of healthcare interventions. BMJ. Sep 04, 1999;319(7210):635-638. [FREE Full text] [CrossRef] [Medline]
- Levin HM, Belfield CR. Cost-effectiveness and educational efficiency. In: Johnes G, Johnes J, Agasisti T, López-Torres L, editors. Handbook of Contemporary Education Economics. New York, NY. Edward Elgar Publishing; 2017:338-356.
- Mâsse LC, Moser RP, Stokols D, Taylor BK, Marcus SE, Morgan GD, et al. Measuring collaboration and transdisciplinary integration in team science. Am J Prev Med. Aug 2008;35(2 Suppl):S151-S160. [FREE Full text] [CrossRef] [Medline]
- Blevins D, Morton B, McGovern R. Evaluating a community-based participatory research project for elderly mental healthcare in rural America. Clin Interv Aging. Sep 2008;3(3):535-545. [FREE Full text] [CrossRef] [Medline]
- Naylor PJ, Wharf-Higgins J, Blair L, Green L, O'Connor B. Evaluating the participatory process in a community-based heart health project. Soc Sci Med. Oct 2002;55(7):1173-1187. [CrossRef] [Medline]
- Optimizing evidence-based practice implementation for clinical impact. IMPACT. URL: https://www.mhimpact.org/ [accessed 2024-04-10]
- Eisman AB, Kilbourne AM, Dopp AR, Saldana L, Eisenberg D. Economic evaluation in implementation science: making the business case for implementation strategies. Psychiatry Res. Jan 2020;283:112433. [FREE Full text] [CrossRef] [Medline]
- Mansournia MA, Altman DG. Inverse probability weighting. BMJ. Jan 15, 2016;352:i189. [CrossRef] [Medline]
- Gaias LM, Arnold KT, Liu FF, Pullmann MD, Duong MT, Lyon AR. Adapting strategies to promote implementation reach and equity (ASPIRE) in school mental health services. Psychol Sch. Apr 08, 2021;59(12):2471-2485. [CrossRef]
- Garcia NM, López N, Vélez VN. QuantCrit: rectifying quantitative methods through critical race theory. Race Ethn Educ. Sep 27, 2017;21(2):149-157. [CrossRef]
- Rodriguez NM, Burleson G, Linnes JC, Sienko KH. Thinking beyond the device: an overview of human- and equity-centered approaches for health technology design. Annu Rev Biomed Eng. Jun 08, 2023;25(1):257-280. [FREE Full text] [CrossRef] [Medline]
- Gewin V. The time tax put on scientists of colour. Nature. Jul 2020;583(7816):479-481. [CrossRef] [Medline]
Abbreviations
BRISC: Brief Intervention for School Clinicians |
CFIR: Consolidated Framework for Implementation Research |
CWIS: Cognitive Walkthrough for Implementation Strategies |
DDBT: Discover, Design and Build, and Test |
DEI: Diversity, Equity, and Inclusion |
FRAME: Framework for Modifications and Adaptations of Evidence-Based Interventions |
FRAME-IS: Framework for Modifications and Adaptations of Evidence-Based Interventions for Implementation Strategies |
HCD: human-centered design |
HCI: human-computer interaction |
IAM: Intervention Appropriateness Measure |
ISUS: Implementation Strategy Usability Scale |
IUS: Intervention Usability Scale |
Neuro-QOL: Quality of Life in Neurological Disorders |
NIH: National Institutes of Health |
NIMH: National Institute of Mental Health |
PI: principal investigator |
PST: Problem Solving Treatment |
PST-Aid: Problem Solving Treatment-Aid |
REDCap: Research Electronic Data Capture |
RUBIES: Research Units on Behavioral Intervention in Educational Settings |
SUS: System Usability Scale |
TAP: think-aloud protocol |
TF-CBT: Trauma-Focused Cognitive Behavioral Therapy |
UWAC: University of Washington Advanced Laboratories for Accelerating the Reach and Impact of Treatments for Youth and Adults with Mental Illness Center |
Edited by A Schwartz; The proposal for this study was peer reviewed by the National Institute of Mental Health Special Emphasis Panel - National Institute of Mental Health - Advanced Laboratories for Accelerating the Reach and Impact of Treatments for Youth and Adults with Mental Illness (National Institutes of Health, USA). See the Multimedia Appendices for the peer-review report; submitted 19.08.24; accepted 02.12.24; published 29.01.25.
Copyright©Aaron R Lyon, Sean A Munson, Michael D Pullmann, Brittany Mosser, Tricia Aung, John Fortney, Alex Dopp, Katie P Osterhage, Helen G Haile, Kathryn E Bruzios, Brittany E Blanchard, Ryan Allred, Macey R Fuller, Patrick J Raue, Ian Bennett, Jill Locke, Karen Bearss, Denise Walker, Elizabeth Connors, Eric Bruns, Jenna Van Draanen, Doyanne Darnell, Patricia A Areán. Originally published in JMIR Research Protocols (https://www.researchprotocols.org), 29.01.2025.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Research Protocols, is properly cited. The complete bibliographic information, a link to the original publication on https://www.researchprotocols.org, as well as this copyright and license information must be included.