Hybrid Effectiveness-Implementation Trials: Testing Strategies, Assessing Outcomes, and Planning for Sustainability
December 15, 2023Speaker: Rani Elwy, Ph.D
November 17th, 2023
This seminar discusses the process of implementing evidence-based treatments covers three stages. this talk will provide guidance on designing and conducting Hybrid Type II and III effectiveness-implementation trials, focusing specifically on theoretical guidance, selecting, and operationalizing implementation strategies, and assessing implementation and effectiveness outcomes.
Information
- ID
- 11097
- To Cite
- DCA Citation Guide
Transcript
- 00:01<v ->Okay. Well, hello, everybody.</v>
- 00:04Welcome to our CMIPS seminar.
- 00:07CMIPS is the acronym for our Center for Methods
- 00:10in Implementation and Prevention Science.
- 00:13I'm Donna Spiegelman, the Director of the Center,
- 00:16and this seminar is being co-sponsored
- 00:20by the Dissemination and Implementation
- 00:22Science Methods Core, the NIH T32 training grant,
- 00:26Implementation Science Research in Methods,
- 00:32and R3EDI, the Rigorous, Rapid,
- 00:34and Relevant Evidence Adaptation and Implementation
- 00:38to Ending the HIV Epidemic Implementation Science Hub.
- 00:42We're very pleased to welcome our guest,
- 00:45Dr. Rani Elwy,
- 00:46who's Professor of Psychiatry and Human Behavior,
- 00:50and Professor of Behavioral and Social Sciences
- 00:53at Brown University.
- 00:57She is a health psychologist, health services researcher,
- 01:01and an implementation scientist
- 01:03who examines patients' access to
- 01:06and uptake of mental health care,
- 01:08the effectiveness and implementation
- 01:10of complementary and integrative health services
- 01:13for treating mental health disorders and pain,
- 01:17and she works on crisis and risk communication
- 01:20between patients, families, providers, and health systems.
- 01:25Dr. Elwy is the Founding Director of the Brown Research
- 01:28on Implementation and Dissemination
- 01:30to Guide Evidence Use BRIDGE Program,
- 01:33which sounds like a sister program to our own here,
- 01:37Co-Director of Implementation Science in Advance,
- 01:40Rhode Island Clinical Translational Research,
- 01:43and Implementation Scientist
- 01:45in the Biobehavioral Sciences Core
- 01:47of the Providence/Boston Center for AIDS Research.
- 01:51Dr. Elwy is a VA
- 01:53or Veterans Administration Implementation Scientist
- 01:56serving as the multiple principal investigator
- 01:59of two program grants:
- 02:01The Bridging the Care Continuum
- 02:03for Vulnerable Veterans Bridge QUERI Program,
- 02:07which consists of three hybrid
- 02:09type three effectiveness implementation trials
- 02:12to increase the uptake of mental health
- 02:14and substance use services and treatments
- 02:16among veterans in 18 sites.
- 02:21And two, the Evidence Policy and Implementation Center,
- 02:25a QUERI program
- 02:26dedicated to building capacity in implementation science
- 02:29across the entire VA healthcare system.
- 02:32Additionally, Dr. Elwy is a Fellow
- 02:34of the Society of Behavioral Medicine
- 02:37and the recipient of a VA Research Center Scientist Award.
- 02:41So, clearly, she's got her bona fide
- 02:43in implementation science all across the spectrum.
- 02:48Today, she's gonna talk about
- 02:49Hybrid Effectiveness Implementation Trials:
- 02:52Testing Strategies, Assessing Outcomes,
- 02:55and Planning for Sustainability.
- 02:57And, Dr. Elwy, would you prefer to give your talk
- 03:00and then take questions,
- 03:02or would you like to have people pop in questions
- 03:04into the chat as you go along?
- 03:06(indistinct)
- 03:20<v ->Thank you for the invitation and that introduction</v>
- 03:22and apologies that it was so long.
- 03:25But anyway, I'm really happy to be here.
- 03:27And as you see,
- 03:28I actually took out the word trials and put in studies
- 03:31and I'll explain why in a minute.
- 03:36So I have some funding grants
- 03:39and just wanted to provide those disclosures,
- 03:41and also that my views are mine alone
- 03:44and not that of the VA or the federal government.
- 03:48So when I think about why I feel so passionate
- 03:52about implementation science...
- 03:54Actually I just realized that just so you know,
- 03:56I can't actually see the chat here in the system that...
- 04:02So yeah, so yeah, thank you very much.
- 04:06<v Donna>You know, I'll take care of it. (indistinct)</v>
- 04:09<v ->Thank you.</v>
- 04:10So when I think about why I'm so passionate
- 04:13about implementation science
- 04:14or how I really just fell into it, two things come to mind,
- 04:17and I'm very well aware
- 04:19that I'm talking to a very quantitatively strong group,
- 04:22so at least here are some statistics for you.
- 04:25So the first is that we all know that on average
- 04:27it takes 17 years for 14% of research
- 04:31to make its way into practice.
- 04:32I recently read that the statistic
- 04:34is now approximately 15 years,
- 04:37but that's really not enough of a change,
- 04:40we still have a long way to go.
- 04:42And then the other statistic is that 85% of federal research
- 04:47is wasted every year due to the fact
- 04:50that we do not move our evidence into practice,
- 04:54and this equates to about $70 billion per year,
- 04:58which is kind of shocking.
- 05:00And it was actually those two statistics
- 05:03that I said to my dean
- 05:07who made him really sort of pay attention
- 05:09to implementation science
- 05:10and decide to invest in our program at Brown.
- 05:14I feel very fortunate to have grown up
- 05:16in the VA healthcare system as an implementation scientist.
- 05:20Our national implementation science program is called QUERI,
- 05:23which is the Quality Enhancement Research Initiative.
- 05:26And under the leadership of Amy Kilborn, Dr. Amy Kilborn,
- 05:30we now have over 50 centers
- 05:32who partner with up to 70 or more different program offices
- 05:37and regional partners around the country.
- 05:39One of the programs that I am PI of,
- 05:41with three other colleagues,
- 05:42is the BRIDGE program, which is in Bedford, Massachusetts.
- 05:46And it's one of these five year
- 05:50five plus million dollar grants
- 05:52where we have three hybrid type three trials
- 05:55that we are doing all at once,
- 05:57implementing three different evidence-based practices,
- 05:59and so that's what I'm gonna be talking about today.
- 06:02I brought my implementation science knowledge and training
- 06:06from the VA into Brown five and a half years ago,
- 06:09and we recently rebranded our implementation science core
- 06:12as the Brown Research on Implementation and Dissemination
- 06:17to Guide Evidence Use BRIDGE Program.
- 06:19And as you all know here,
- 06:21we just don't have enough training programs,
- 06:23we don't have enough capacity building programs
- 06:25for implementation science,
- 06:26and so a lot of us
- 06:27are just starting to like implement our own as a result.
- 06:32So I just wanted to talk about hybrid designs and trials,
- 06:38but I'm actually gonna try to convince us all
- 06:42to use the language of hybrid studies from here on out
- 06:46and I'll explain why.
- 06:47So we all know the original paper that came out in 2012
- 06:50by Jeff Curran, Mark Bauer, Brian Mittman,
- 06:54Jeff Pyne, and Cheryl Stettler, that was really seminal.
- 06:58I was just starting my implementation science training
- 07:00in this year at the Implementation Research Institute,
- 07:04and everyone was so excited
- 07:06about these different hybrid designs,
- 07:07the one, the two, the three.
- 07:08I know that many of you already know what these are.
- 07:11I'm just gonna be focusing today on the three
- 07:13where really the primary aim is on testing the effectiveness
- 07:17of the implementation strategies
- 07:18to increase the uptake of the evidence-based practice.
- 07:24When this paper came out,
- 07:25Brian Mittman was one of my mentors and he said, you know,
- 07:28they really should have said right from the beginning
- 07:30that every randomized trial is always a hybrid one,
- 07:35and so they've actually started
- 07:36to really change that language.
- 07:39And so just recently they published,
- 07:41this was from last December,
- 07:43reflections on 10 years of using hybrid designs,
- 07:47hybrid trials,
- 07:48and they tried to make a good case
- 07:50for why it's really much more important
- 07:53to be calling these hybrid studies.
- 07:55And I'm really happy about this because this...
- 08:00And I know that, again,
- 08:01I'm talking to a very quantitatively savvy group.
- 08:04Not all things in the real world can be randomized,
- 08:07and so we have had to do studies
- 08:09that I would consider a hybrid
- 08:11but didn't fit under the definition
- 08:13because it wasn't a randomized control trial.
- 08:16So in these updated recommendations,
- 08:18they really have three things
- 08:20that they want people to take away.
- 08:21Replacing the term design in favor of the word study
- 08:26because, as I just said,
- 08:27many people are applying hybrids and non trial designs
- 08:30and it is possible to conduct a hybrid study
- 08:35to answer questions about intervention effectiveness
- 08:37and implementation in a wide range of study design.
- 08:40(indistinct)
- 08:43They offer in this paper,
- 08:44four questions to help people decide
- 08:49which of the hybrid studies that they should be conducting,
- 08:53and I'll tell that to you in a minute,
- 08:54and they've also really emphasized how you can build cost.
- 08:57And I know that we have some cost people in the audience,
- 08:59how to really bring in cost into hybrid studies,
- 09:02because when our ultimate goal from all of this work
- 09:06is to implement and sustain our evidence-based practice,
- 09:09cost is such a huge and driving factor for that.
- 09:13So these are the four questions that are asked in the paper.
- 09:17What is the nature of your effectiveness data?
- 09:19How much do you expect
- 09:21the intervention will need to be adapted?
- 09:23How much do you know about implementation determinants?
- 09:26And how ready are you to evaluate
- 09:28real world implementation strategies?
- 09:31If you know a lot about your effectiveness data,
- 09:33if you feel that there needs to be some components
- 09:37of adaptation built into your actual aims,
- 09:42if you already have a good sense
- 09:44of what your implementation determinants,
- 09:46your barriers and facilitators are,
- 09:48and if you feel that you can develop and evaluate
- 09:52those real world implementation strategies
- 09:54to address those determinants,
- 09:56then you're probably ready for a hybrid three
- 09:58or if not at least a hybrid two.
- 10:00But if you're more on the end of I don't know,
- 10:03then you probably wanna go to more of a hybrid one.
- 10:06And so this paper really helps people
- 10:08think that through more than the original one.
- 10:13So this is gonna be the crux
- 10:14of what I'm gonna be talking about today.
- 10:15When I think about what we want to achieve
- 10:18in implementation science,
- 10:19I think of these three big buckets,
- 10:21and I'm sure that there are other ones
- 10:22and we can definitely have a conversation
- 10:24about why I haven't included others.
- 10:26But I think of testing strategies
- 10:28because strategies are what is going to make us
- 10:31be able to get things implemented uptake in the real world.
- 10:37But it has a lot of different factors to it.
- 10:39You need to specify and operationalize your strategies,
- 10:42you have to randomize on strategies,
- 10:44and for some people that's a very new thing,
- 10:46you're not randomizing on the intervention,
- 10:48you're randomizing on the strategies, and tracking.
- 10:52You know, I often say that as implementation scientists,
- 10:55our job is just to track,
- 10:56track everything, track adaptations,
- 10:59track whether they were fidelity consistent,
- 11:01track your strategies,
- 11:02and especially if you're doing something
- 11:03across a lot of sites, that's a lot of tracking.
- 11:05So we'll talk about that.
- 11:07I also think that we really need to be focusing
- 11:09on assessing outcomes,
- 11:10and these always need to be guided
- 11:12by a theory, model or framework.
- 11:14You need to do this over multiple time points.
- 11:16One time is not enough
- 11:17in the scope of an implementation trial,
- 11:20things change over time.
- 11:22So one strategy might be leading to a better outcome,
- 11:27but when you use other strategies,
- 11:28maybe those outcomes aren't so good,
- 11:30and so we really need to know those differences.
- 11:33And one of the most important things,
- 11:35and this has been a message I've had to give
- 11:37to a lot of my clinical trialist efficacy researchers,
- 11:40our measures need to be pragmatic.
- 11:42Please don't use a 60 item measure
- 11:45in your implementation study,
- 11:47it's not gonna be used in the real world.
- 11:48Please don't include a two hour
- 11:51clinical structured interview to assess outcomes,
- 11:54again, not gonna be used in the real world,
- 11:56really focusing on pragmatic, easy to use,
- 11:59transferable measures.
- 12:01And then finally planning for sustainability.
- 12:04If we don't plan, it will not happen,
- 12:07and that's also a very hard message for people to hear.
- 12:10How do I know, if I'm doing a hybrid type one,
- 12:13that whatever I'm implementing is gonna be effective?
- 12:15And I say, let's just pretend that it will be.
- 12:18Let's find out what is happening during that trial
- 12:21that actually can inform our next steps
- 12:24and help us think through,
- 12:26you know, who is gonna own this eventually.
- 12:28And obviously this involves
- 12:30a lot of engagement with the partners, community based,
- 12:32health system based that we're gonna be needing
- 12:35to eventually sustain our efforts.
- 12:39When I think about the theories,
- 12:40models and frameworks I use, I always have a process model,
- 12:43and so in the VA we have this QUERI implementation roadmap
- 12:47that everyone can use.
- 12:49It's very foundational, but it works in every case,
- 12:55so we have a pre-implementation phase,
- 12:57an implementation phase, a sustainment phase.
- 13:00It looks like you always go in one way around this,
- 13:02but you can go back and forth,
- 13:04especially in the pre-implementation
- 13:05and implementation phases.
- 13:07So we use this to guide our work
- 13:09and this is really applicable
- 13:10and we'll talk more about these details,
- 13:12but it's something that you should consider
- 13:15to have a process model
- 13:16to drive what your steps are of your work.
- 13:21In our Bridge QUERI Program,
- 13:23which is testing the uptake and hopefully sustainability
- 13:28of three evidence-based practices,
- 13:30these are each in a hybrid type three trial.
- 13:32So they're simultaneously happening
- 13:34led by three different people and their teams.
- 13:37We're working with veterans who have opioid use disorders,
- 13:41we're working with veterans who have been recently released
- 13:43from an incarceration setting,
- 13:45and then we have veterans
- 13:47who are engaging in criminal activities
- 13:50and are going through the veteran treatment court.
- 13:53Veteran treatment courts
- 13:53are actually based in the community,
- 13:55but we work with our veteran justice outreach group
- 13:57within the VA to work with that.
- 13:59And so, as you can see,
- 14:00these are not easy to solve problems,
- 14:03this is a vast amount of effort on these people.
- 14:07So the Homeless Overdose Prevention Expansion Project, HOPE,
- 14:12is really trying to implement an opioid overdose education
- 14:16and naloxone distribution to reduce overdoses.
- 14:19So that's the evidence-based practice there,
- 14:21and that is led by Dr. Amanda Midboe at Palo Alto.
- 14:26The PIE Project is a peer support initiative.
- 14:30It's a Post-Incarceration Engagement Project
- 14:32where we're really helping to work with veterans
- 14:35when they come out of jails for social support,
- 14:37skill building, linkage to care.
- 14:40And then we have MISSION-CJ,
- 14:41which is very long acronym I'll tell about it in a second,
- 14:44where we are aiding veterans in case management,
- 14:46assertive outreach, hybrid treatments, linkage support.
- 14:52Also really trying to make sure
- 14:53that we are examining the health equity needs
- 14:57of our veterans
- 14:58as well as how we can help them stay housed
- 15:04if they're at risk for homelessness.
- 15:06My job in all of this is I run our implementation core,
- 15:10and here we are trying across all these three projects
- 15:13to have similar phases, similar measures, similar designs.
- 15:18So that's been a real learning experience for me
- 15:22to simultaneously work with three different trials
- 15:24at the same time to really make sure
- 15:26that we are capturing data on a widespread basis.
- 15:30Here are my three MPI colleagues.
- 15:33Keith McInnes is running the PIE Project with his team,
- 15:36David Smelson is running the MISSION-CJ Project,
- 15:39and Amanda Midboe is running the HOPE Project.
- 15:41It's definitely a village that's doing this.
- 15:45And what's really exciting
- 15:46is when your work is totally aligned with policy,
- 15:49and I really want people to think about that
- 15:51with the types of work that you do,
- 15:52because policy is actually going to help you
- 15:54with your sustainability.
- 15:56So this is from the February State of the Union,
- 15:59and there were three parts of this that completely aligned
- 16:03with the work that we are doing.
- 16:04So expanding peer support services in the VA.
- 16:08Two of our three evidence-based practices
- 16:10that I'll tell you about, the PIE and MISSION-CJ,
- 16:14are being implemented by peer support specialists.
- 16:16So we need more of them in the VA,
- 16:18and so the federal government is funding this.
- 16:21We're trying to reduce homelessness.
- 16:24So this is also a focus
- 16:26of the Biden Administration for veterans.
- 16:29And then we're also expanding outreach
- 16:31to justice involved veterans,
- 16:32which is a very big part of the MISSION-CJ Project
- 16:35as well as PIE.
- 16:36So we can see that we have
- 16:38a lot of policies support behind this
- 16:40and we just also need the funding too.
- 16:43So just a brief thing about HOPE,
- 16:45and just to sort of maybe state the obvious to people,
- 16:49I am not the content experts with any of these,
- 16:51I'm the implementation scientist.
- 16:53My colleagues who run these projects
- 16:55also have implementation science expertise,
- 16:56but we needed a central core
- 16:59to sort of oversee all of these.
- 17:00And so Amanda is working with HOPE in five sites
- 17:05that are in California, Nevada, and Hawaii,
- 17:08with veterans who have a diagnosis
- 17:11of an opioid use disorder,
- 17:12a stimulant use disorder or are being prescribed opioids.
- 17:17The PIE Project, the Post-Incarceration Project,
- 17:20really works intensely with veterans
- 17:22when they're coming outta incarceration
- 17:23and coordinates with our healthcare
- 17:25for reentry a veteran office and also our housing,
- 17:30and the HUD-VASH is taking housing vouchers
- 17:34from the federal government
- 17:35and then pairing that with veteran supported housing
- 17:39and other support services.
- 17:41There are four components to PIE,
- 17:43linkage and referral, skill building and goal setting,
- 17:46community reintegration, and social and emotional support.
- 17:49So right now PIE is being implemented in six sites,
- 17:52other sites have either previously implemented it
- 17:54or are no longer implementing it.
- 17:58And then MISSION-CJ, which is our most complex intervention,
- 18:00is Maintaining Independence and Sobriety
- 18:03through System Integration Outreach
- 18:05and Networking-Criminal Justice.
- 18:07MISSION was developed 25 years ago by David Smelson
- 18:10and has had a whole bunch of evidence behind it,
- 18:14but it's pretty complex.
- 18:15They build it into the criminology concept
- 18:19of risk, need, responsivity,
- 18:22where they're trying to identify
- 18:24which person in the criminal justice system,
- 18:27how can we tailor what they need to support them.
- 18:29And so they have core services
- 18:30of critical time intervention, empowering pro-social change,
- 18:34dual recovery therapy, peer support,
- 18:36and then they also offer
- 18:38some vocational and educational support
- 18:40and trauma-informed care,
- 18:42and they are implementing this across eight sites.
- 18:45David actually has a really massive $12 million HEAL grant
- 18:50where he's doing this in even greater sites,
- 18:52so I really don't know how he's managing,
- 18:54but there's a lot of MISSION implementation
- 18:56happening around the country right now.
- 18:59So HOPE has five sites, PIE has six sites,
- 19:02and MISSION has eight sites.
- 19:06My job is to say how are we gonna do this all.
- 19:09How can we keep similar methods
- 19:12across everything that we're doing?
- 19:13And so again, we've used the roadmap model, process model,
- 19:17and we have a really strong pre-implementation phase.
- 19:21And I will just also say, I don't know if anyone saw it,
- 19:23but maybe four months ago, Lisa Saldana,
- 19:28came out with a paper with some colleagues
- 19:29that showed across a large swath of papers
- 19:34that they reviewed that studies
- 19:36that had a very in-depth pre-implementation period
- 19:41actually were more successful
- 19:42in their implementation efforts later.
- 19:44So a lot of people, you know,
- 19:45they wanna just dive in and I say embrace this period.
- 19:50Even when you have an effective evidence-based practice
- 19:53that you wanna implement,
- 19:54you still need to know about a new site,
- 19:57you still need to know what the clinical workflow is,
- 19:59you still need to know what are the anticipated barriers
- 20:03and facilitators implementing something.
- 20:04And so from there, we're not gonna change
- 20:07what our evidence-based practices are,
- 20:08but what we're gonna do is we're gonna think
- 20:10how can our strategies address these
- 20:12and how can we create any adaptations
- 20:15but without changing those core components.
- 20:17So really, you know, let yourself be immersed in this phase.
- 20:23Yeah, of course.
- 20:25(audience conversing indistinctly)
- 20:42(audience member speaking indistinctly)
- 20:58(audience member speaking indistinctly)
- 21:29Yeah, I don't know if the Zoom audience can hear,
- 21:33but there's some conversation around
- 21:38how difficult it is to get RO1 funding
- 21:41to have a substantial pre-implementation phase
- 21:43even when you already have an evidence-based practice
- 21:47and whether the VA is different.
- 21:49I do think the VA is different,
- 21:53but we have built this into RO1 grants,
- 21:55and in fact, I will say that what was really interesting
- 21:58for my colleagues who...
- 22:00So what I do at Brown
- 22:01is work with a lot of efficacy researchers
- 22:03who are building an implementation science
- 22:05into their grants.
- 22:06So I have several colleagues who do suicide prevention work,
- 22:09and they were taking a suicide prevention
- 22:11an intervention called STEP into an inpatient setting.
- 22:16And it has a lot of evidence-based,
- 22:18they're just moving it to a new setting and it's an RO1,
- 22:22and the program, the POs from NIMH came back and said,
- 22:28you need to do, as part of this,
- 22:30a bunch of formative evaluation
- 22:32with a health system leadership
- 22:34before you can do the rest of the aims.
- 22:38They weren't saying we're not gonna fund your project,
- 22:40they said you need to build in a pre-aim
- 22:43before your aim one.
- 22:44And so I worked with the team
- 22:46to build a formative evaluation phase
- 22:48because the STEP had never been in an inpatient,
- 22:51it had always been used outpatient,
- 22:52and the program office said,
- 22:55we're not gonna fund it
- 22:56until you know more about what you're gonna go into,
- 22:58and so that was very positive.
- 23:00It didn't come from the reviewers,
- 23:02it came from the program office,
- 23:03so you know, that's a positive thing.
- 23:05So yes, I will say you kind of have to limit yourself.
- 23:10You could go crazy and spend a lot of time,
- 23:12but like we lot a lot ourselves.
- 23:14And unfortunately we started this in October of 2020
- 23:19when the VA was still dealing with a lot of challenges
- 23:22with COVID and virtual work
- 23:24and also about a month and a half
- 23:27before the vaccines were implemented.
- 23:30So this was not a good time to do a lot of informative work,
- 23:34I'll tell you, but anyway, we did it.
- 23:36And then our implementation is a lot of the training of...
- 23:40You know, because in a hybrid three,
- 23:42the goal is to get the people at the sites
- 23:44to do the implementation.
- 23:45You are helping to direct that,
- 23:47you're providing them with support through your strategies,
- 23:49but you should not be implementing that in the hybrid three
- 23:52because that's not real world.
- 23:53And so we have, in the HOPE project, social workers,
- 23:56and in the other team, PIE and MISSION-CJ,
- 23:58we have peer support specialists
- 24:00and we need to train them to do this,
- 24:02and so we spend a lot of time on that and implementation.
- 24:06And then obviously as we go through,
- 24:08we're also assessing our outcomes
- 24:09but they are a secondary aspect.
- 24:12And then sustainability.
- 24:13So we have just finished three years of our trial,
- 24:17see I slipped, three years of our study,
- 24:21and we've launched year four,
- 24:24and so some of our sites
- 24:25have gone through the implementation
- 24:27and are moving towards sustainment assessment
- 24:29and some are just starting.
- 24:30So there's kind of like a...
- 24:33Well, it's a stepped wedge design,
- 24:34so we're not there with all of them right now.
- 24:38So again, I'm not the statistician on the project,
- 24:43but I just wanted to let you know
- 24:45that we're doing cluster randomized step wedge trials,
- 24:49in parentenses, studies,
- 24:51as I try to transition to this language.
- 24:54But our overall goal is really to estimate the effect
- 24:57of what it's like to transition
- 24:58to a higher intensity implementation strategy package
- 25:02from a baseline lower intensity strategy package.
- 25:05On each of the effectiveness outcomes that we're using,
- 25:07we're gonna use mixed effects regression models,
- 25:10we'll have a fixed effect
- 25:11before the implementation strategy package.
- 25:14I keep saying package but we do know
- 25:17because we're tracking these things really well
- 25:19that not all of our implementation strategies are being used
- 25:23despite all our best efforts.
- 25:24So we're trying to track individual strategies
- 25:27within each of these higher intensity
- 25:29and lower intensity states.
- 25:32And then also we will,
- 25:34we are trying to do some balancing
- 25:36and, you know, we did that prior, you know,
- 25:38to look at the different site characteristics
- 25:40of our different wedges.
- 25:42And we have had sites drop out,
- 25:43we've had sites that we've added in,
- 25:45it has not been textbook perfect by any means.
- 25:50<v Donna>I want to ask a few questions.</v>
- 25:51<v ->Yeah, but I may not be able to answer. (laughs)</v>
- 25:56<v Donna>In terms of your primary outcome of analysis,</v>
- 26:00when you write implementation strategy,
- 26:02you mean the whole bundle or your putting in variables,
- 26:06each one of the components individually assess the impact
- 26:10of each one of the components of the bundle?
- 26:13<v ->So when we wrote the proposal,</v>
- 26:14we thought bundle slash package.
- 26:17But as we are doing a lot of tracking of those strategies,
- 26:20I think that our analysis is gonna be by the individual
- 26:24so they don't overlap.
- 26:25We tried really hard to make sure our bundle of strategies
- 26:28in the higher intensity do not overlap
- 26:30with the bundle of strategies in the lower intensity.
- 26:32So if only a few
- 26:34of the higher intensity strategies get used,
- 26:36we'll know that those are higher intensity,
- 26:38but they may not be the whole package.
- 26:42That is definitely a difficult thing
- 26:44to get people to use all of them.
- 26:47And I will tell you about those strategies in just a second.
- 26:50So now I'm gonna move into the testing strategy.
- 26:51So as you remember,
- 26:52I said really important to achieve what our goals,
- 26:55in implementation science,
- 26:56we need to randomize some strategies,
- 26:58we need to do a lot of tracking,
- 26:59we need to do a lot of specifying and operationalizing.
- 27:02And one of the things I really wanna get across is this,
- 27:04nothing about me without me motto
- 27:07that was developed in the late 90s, early 2000s,
- 27:10about patient-centered care,
- 27:11and I would argue that this is absolutely necessary
- 27:15for doing an implementation study.
- 27:17Because as you're developing your implementation strategies,
- 27:20you are doing this in concert with your sites,
- 27:23with your partners, with the champions,
- 27:26with everyone that you're gonna be working with.
- 27:29And I think the reason this is so important is because,
- 27:32in the end,
- 27:33successful implementation is going to be because of trust.
- 27:38And Alison Metz wrote a paper
- 27:40in "Implementation Science Communications" last year
- 27:43on sort of thinking of trust in a conceptual way
- 27:45across implementation studies,
- 27:47and she talks about intrapersonal trust
- 27:50and interpersonal trust.
- 27:51But I believe that in any shape or form,
- 27:55this is why the pre-implementation work is so important,
- 27:58you are building trust through those efforts.
- 28:00And so people realize
- 28:02that you are not just doing a research study
- 28:04'cause you got funding,
- 28:05you're not just trying to write a publication,
- 28:06that you're actually trying to change care
- 28:09and improve care and make lives better.
- 28:11And so if people can see that
- 28:13as you're doing your formative work,
- 28:15then I think that that is the basis
- 28:17for your successful implementation.
- 28:20I like to show this slide
- 28:22about the nine buckets of implementation strategies.
- 28:25Even though it's the earlier paper
- 28:26on implementation strategies by Byron Powell and colleagues,
- 28:30it's really hard for people to conceptualize
- 28:3173 implementation strategies,
- 28:33but when you think about them in nine buckets,
- 28:35I think that's much more helpful.
- 28:37And so when you look at them,
- 28:39you can imagine trust being part of all of this.
- 28:41You know, when you're assessing people for readiness,
- 28:44when you're trying to do interactive assistance,
- 28:46when you're doing adaptations,
- 28:48when you're trying to develop relationships
- 28:50and training people, supporting clinicians,
- 28:53engaging consumers, et cetera,
- 28:54trust is such an important part of that.
- 28:56People wanna know like,
- 28:57why are you interested in doing this?
- 28:58You know, what is in it for me and what is in it for you?
- 29:03So I think relationship building is so critical
- 29:06and the trust piece comes naturally with that.
- 29:10So in addition to our process model
- 29:12of the QUERI implementation roadmap,
- 29:13we also have a model that's helping us examine
- 29:16the various determinants
- 29:18that we're going to be needing to consider throughout.
- 29:21And we chose, because it's a hybrid type three
- 29:23and we're focused on sustainability,
- 29:25The Dynamic Sustainability Framework
- 29:26that was developed by David Chambers,
- 29:28Russ Glasgow, and Kurt Stingy.
- 29:30And this is saying that there are three main components
- 29:34that we need to consider, intervention,
- 29:36the evidence based practice, the practice setting,
- 29:38the context of what we're implementing,
- 29:40and then the wider ecological system,
- 29:42which is very much a thing to think about
- 29:44from a sustainability perspective.
- 29:46But what's different about their suggestions is that,
- 29:48again, it's not just a one-time assessment.
- 29:51How does this change over time?
- 29:53So at the pre-implementation phase,
- 29:54it may look like one thing,
- 29:56at the implementation phase,
- 29:57it may have a different feeling about it,
- 29:59and then at the sustainment phase, we might see things.
- 30:02So we need to have a constant process
- 30:05by which we're examining that.
- 30:07And in fact, Enola Proctor recently published a paper
- 30:10sort of looking at, you know,
- 30:1110 years of implementation outcomes
- 30:13according to her implementation outcome framework.
- 30:16The critiques that she has of the literature
- 30:18is that people are just doing one-time assessments
- 30:20of implementation analysis, that's just not enough.
- 30:24So this is just an example from HOPE
- 30:26about how we're sort of doing this
- 30:27across the three phases of the roadmap
- 30:29and then guided by The Dynamic Sustainability Framework.
- 30:32So in phase one,
- 30:34they did 52 interviews of various people at the sites,
- 30:38and HOPE is still in the phase two and phase three stages,
- 30:41as are all the other projects.
- 30:42So there's been 21 interviews
- 30:43so far in implementation phase,
- 30:4521 interviews so far in sustainment phase.
- 30:48But really doing interviews with the housing,
- 30:50supportive housing staff, the prescribers for the Naloxone,
- 30:55other key staff, pharmacists,
- 30:58social workers and veteran patients.
- 31:02And we're using a rapid directed content analysis approach
- 31:06with really guided by
- 31:08The Dynamic Sustainability Framework construct.
- 31:12So when we decided on this proposal,
- 31:15we had years of research building up to this
- 31:18and so we decided that we were going to use facilitation
- 31:21as our implementation strategy.
- 31:24But when I say that, it sounds so funny,
- 31:26because facilitation is like literally like 10 things.
- 31:28It's a bundle in itself, so it's a natural bundle of things.
- 31:32And so we're trying to use, you know, engagement,
- 31:35identifying champions, action planning, staff training,
- 31:38problem solving, technical support,
- 31:41which is different from technical assistance, I'll just say.
- 31:44So technical support is a much more hands-on process
- 31:47and audit and feedback process.
- 31:48So lots of things go into facilitation,
- 31:50it's a naturally existing high intensity bundle.
- 31:55And then we start with the more lower intensity bundle,
- 31:59which is either education outreach or academic detailing.
- 32:04They're very similar.
- 32:05HOPE uses academic detailing,
- 32:06the other projects use education outreach.
- 32:08But this is more
- 32:09to really have these targeted structured visits,
- 32:12we're delivering tailored training
- 32:14and we're doing technical assistance
- 32:15as in contact us if you have a problem
- 32:17as opposed to us contacting you.
- 32:20So it's much more lower intensity.
- 32:22I would love for our results to be really strong
- 32:26in the lower intensity
- 32:26because that's gonna be much more sustainable.
- 32:28But that's an empirical question so we will...
- 32:33It really comes from the world of pharmaceuticals, I think,
- 32:37people would show up and have like a one-on-one and say,
- 32:40"Dr. Spiegelman, let me tell you
- 32:42about this medicine that I have
- 32:45that can help patients with diabetes,"
- 32:47and they'll have like a one-on-one conversation
- 32:48and really just tell them about it.
- 32:49And so we've taken that and made it into, I mean,
- 32:53and we're not dealing,
- 32:54but other people have made that
- 32:55into more of a one-on-one strategy just to inform.
- 32:58It's more of an educational activity.
- 33:03So the good news is there's lots of tracking supports
- 33:06available for us out there that are getting published.
- 33:09The bad news is they're a lot of work.
- 33:13So in 2020 our colleagues in Little Rock Arkansas
- 33:17who are really have been the group
- 33:19that have been defining what facilitation is,
- 33:22doing lots of trainings, they have a manual,
- 33:24if you need that let me know,
- 33:25but it's probably listed in that paper.
- 33:28But in this paper they actually, as an appendix,
- 33:31gave a Excel tracking sheet.
- 33:33This is how you can track facilitation and we use this,
- 33:36we've adapted it a little bit
- 33:37because we're also tracking the stage of implementation
- 33:40in which people are at, but we can see what type of event.
- 33:43So when you have facilitation,
- 33:44you have an external person who's part of the team,
- 33:46you have an internal facilitator at the site
- 33:48and you're working one-on-one,
- 33:50but the internal facilitator
- 33:51is the one who's doing the work.
- 33:53So our peer support specialists,
- 33:54our social workers are the ones doing the work.
- 33:56And so we are tracking all type of communication they have,
- 33:59we're tracking the type of personnel involved at the site,
- 34:03we're tracking the facilitation activity codes,
- 34:05which of the various things of facilitation are happening.
- 34:09Really importantly,
- 34:10we're tracking how many hours and minutes
- 34:12each facilitation activity takes.
- 34:15And so clearly we know what we do on the external side,
- 34:20what our research staff does
- 34:21when they reach out to the peer support specialist
- 34:22or the social worker,
- 34:24what we don't know is what happens on their side.
- 34:25And so we have these check-in calls with them,
- 34:28it might just be 15 minutes just to say like,
- 34:30what did you do this week? Who did you talk to?
- 34:33And that's really essential
- 34:34because to ask people to complete this type of tracker
- 34:37would be really difficult to do.
- 34:40We actually also adapted this
- 34:42so that we could add in some education
- 34:44and academic detailing outreach to this
- 34:47so that we didn't have to have more than one tracker.
- 34:50(audience member speaking indistinctly)
- 34:52I'll come to that. (laughs)
- 34:55(audience member speaking indistinctly)
- 34:57Yeah, no, this is really, thank you for...
- 34:59'Cause I was gonna say this
- 35:00and what has been so essential about this
- 35:02is because we know what type of personnel it is,
- 35:04we know how much time they spend,
- 35:05we can actually estimate their salary
- 35:08and we know how much every facilitation activity took.
- 35:11Yes, we're doing that.
- 35:12It was actually a requirement of the project to do that.
- 35:15So we've created our own sort of Excel dashboard,
- 35:18it's not really a dashboard
- 35:19because it's not updated automatically,
- 35:22but we're sort of tracking every project.
- 35:24This is the PIE project with its six sites
- 35:27to see whether people are in pre-implementation
- 35:28and implementation.
- 35:29So we're sort of using these trackers
- 35:31to find out where those are.
- 35:33And then here is a snapshot
- 35:35of how we can sort of examine
- 35:37those different facilitation activities.
- 35:39The pie chart on the left
- 35:40shows us the different support staff
- 35:43who are involved in the facilitation activities.
- 35:45So we are working with either the social worker
- 35:48or the peer support specialist,
- 35:49but then they're going on and working with other people too.
- 35:52And then we see on the right side,
- 35:55all the different activities
- 35:56that's happening in facilitation,
- 35:57and you can see that some are more popular than others,
- 36:00and so this is why we know that not all are getting used.
- 36:04And then the bottom left shows us
- 36:05how much time in minutes is being spent.
- 36:09And so obviously site one and four are doing great,
- 36:13like they're really spending a lot of time on this
- 36:15and the other sites are spending less time,
- 36:17that doesn't necessarily mean that they're worse or better,
- 36:19it's just there's so many different dynamics
- 36:22that go into any organization in any site.
- 36:24And so, you know, this'll be something
- 36:27that we'll have to examine when we do our analysis,
- 36:29but it could be really important to know,
- 36:31does more time lead to better outcomes? Who knows?
- 36:37So how are we gonna assess these outcomes?
- 36:39So again, remember guided by a theory model framework,
- 36:42multiple time points as outcomes change over time
- 36:45and involving pragmatic measures is really key.
- 36:48So on the left column is all the pragmatic measures
- 36:51that we said and/or how we are going to conceptualize
- 36:54some of these things.
- 36:55So at the top we have an organizational readiness
- 36:57for implementing change scale,
- 36:59we have the three quantitative assessments
- 37:02of acceptability, appropriateness and feasibility,
- 37:05which align with the Proctor model.
- 37:07We are looking at four of the re-aim outcomes
- 37:11in terms of implementation outcomes,
- 37:13reach, adoption, fidelity to the implementation,
- 37:16and then sustainability.
- 37:18And so each of these is being assessed in different ways,
- 37:21slightly,
- 37:22the measures stay the same across all three studies,
- 37:25but the re-aim ones
- 37:27are a little bit different depending on the project.
- 37:30So these are the two scales that we're using,
- 37:34scale packages,
- 37:35organizational readiness for implementing change,
- 37:38and then the implementation outcome measures.
- 37:41Yes, oh.
- 37:44(audience member speaking indistinctly)
- 37:51They're really kind of basic.
- 37:53(audience member speaking indistinctly)
- 38:00Oh, interesting. Yeah, I haven't heard that section.
- 38:05I did reach out to him at one point,
- 38:06because I work with a lot of clinical trials,
- 38:10they were like, is there a cutoff point for these scales,
- 38:12and Brian was like,
- 38:13"Oh no, we're nowhere near having that kind of data."
- 38:16But I agree that they're really basic
- 38:19and we do use them at three time points,
- 38:22luckily they are not exactly next to each other,
- 38:26like they're like a six months to 12 months apart.
- 38:29(audience member speaking indistinctly)
- 38:30Well, each of those has four questions.
- 38:33Yeah, this has 12.
- 38:35(audience member speaking indistinctly)
- 38:51We've also, I think,
- 38:53only given the feasibility to some people,
- 38:55I don't think the feasibility
- 38:56has been relevant for everybody.
- 38:57Oh, there's something in the chat.
- 38:59(audience member speaking indistinctly)
- 39:07ORCHA's big, this is 12 item.
- 39:10Yeah, there's some statement that they're having,
- 39:13people are having trouble hearing the questions.
- 39:14So the questions are about just the relevance
- 39:18and the usefulness of these implementation outcome measures
- 39:21because people find them very repetitive
- 39:24and not really informative,
- 39:28and I think there's more work to be done
- 39:30in this space for sure.
- 39:32(audience member speaking indistinctly)
- 39:44Oh, yeah, that's great to think about.
- 39:48<v Speaker>You have an international audience,</v>
- 39:49people participating from all over the world.
- 39:52<v ->Oh, wonderful. I'm glad this time works. Thank you.</v>
- 39:55So in our dashboard we are collecting the data,
- 39:59so this we can do it automatically.
- 40:01So we send out a red cap survey to people, they complete it,
- 40:05and it transitions into our Excel spreadsheet automatically.
- 40:10So this is actually some very good part
- 40:12of having these three projects
- 40:14and being able to collect the data that way.
- 40:16So we have sort of some assessments,
- 40:18this is by site of where people are with their mean scores,
- 40:22and then we can sort of compare the scores
- 40:25across the different projects as well.
- 40:29The challenge here is that
- 40:31everybody's not doing the same measurement at the same time,
- 40:33so we try to stay on top of this and sort of remind people
- 40:37that it's a little bit more challenging
- 40:39than I hoped it would be.
- 40:40Just looking at reach for HOPE, you know,
- 40:44maybe the bottom is the best to look at.
- 40:46We just had a technical expert panel meeting
- 40:48on November 7th,
- 40:49so a lot of these more recent information
- 40:52on reach and adoption, implementation,
- 40:54just came from that November 7th meeting.
- 40:56So you can see that the percent
- 40:58of eligible veterans offered
- 41:00versus six months post implementation
- 41:03has definitely gone up in the opioid education
- 41:08and naloxone distribution.
- 41:09So this all very positive, no statistics here,
- 41:13we're just tracking it at the moment.
- 41:15We're also looking at how many people were offered it,
- 41:17how many case managers trained,
- 41:19and one site has just started implementation in this case.
- 41:23<v ->Go back a second?</v> <v ->Yeah.</v>
- 41:24<v Speaker>There's a very interesting issue</v>
- 41:26that arises in looking this data,
- 41:29which is, you know,
- 41:30we always think about privacy of individual,
- 41:32but actually in some of the work I've done now,
- 41:36we found that there are privacy issues
- 41:38all concerns by facilities
- 41:41where like it could be embarrassing to say Palo Alto
- 41:44that they were only at 19%
- 41:47and they could even get in trouble or lose their funding.
- 41:50And so I'm wondering-
- 41:51<v ->I shoulda probably. (laughs)</v>
- 41:55Yeah, we probably should've DM'd.
- 41:56Maybe I'll go on, so no one's looking at that anymore.
- 42:01Yeah, well, I mean I probably should have done it here too.
- 42:04So yeah, I mean I think it's like, you know...
- 42:08I think everyone knows that people are working really hard.
- 42:13We've done a lot of qualitative work
- 42:15to show that the reasons why people don't offer it,
- 42:18like veterans get offended that you're offering it to them,
- 42:22like, is one, you know, like other social workers said
- 42:26that it was not within their scope of work,
- 42:28you know, like scope of practice.
- 42:30So we have a lot of barriers that we've identified
- 42:33that we're trying to address,
- 42:34and obviously some of those things
- 42:35we need to raise to a higher level
- 42:36to say like we go to the National Social Work Agency
- 42:39and say, "How can you help us?"
- 42:42Because we want social workers to be able to do this,
- 42:44but they don't think it's in their scope of practice.
- 42:45So yeah, so those are the types things
- 42:47that we are working on.
- 42:49Then we look at how many veterans
- 42:51have been released from jail in the PIE Project,
- 42:54the Post-Incarceration Engagement,
- 42:56and we can see how many we actually served in our project.
- 42:59You know, there's six sites on the bottom,
- 43:01so this is more just to give you an idea,
- 43:03you don't have to look at the details,
- 43:04but just to how we're trying to track things.
- 43:06And then we also look at
- 43:09the different total encounters post release.
- 43:12So it's just, you know, some projects are just starting,
- 43:15like we know, so we don't have very much data in them there.
- 43:20And then in MISSION,
- 43:22we are tracking who is trained at each of the sites.
- 43:25And this is actually a really difficult one
- 43:27because we're working with a community organization,
- 43:29which is the veteran treatment court,
- 43:30and so we're now going completely outside of the VA
- 43:33to do this, so it's really challenging.
- 43:36Who's trained, who's implementing
- 43:38of those who have been trained,
- 43:39how many veterans have been served,
- 43:41and how many mission encounters have happened,
- 43:43and mission encounters are pretty complex.
- 43:45So the fact that there are over a thousand already
- 43:49after year three is really great news.
- 43:51<v Speaker>Can you remind us what a mission encounter is?</v>
- 43:53<v ->Yeah, mission encounter is,</v>
- 43:55I'll just quickly scroll back
- 43:57'cause I won't be able to remember it all.
- 43:58it's several different evidence-based practices.
- 44:04There's a critical time intervention, pro-social change,
- 44:08dual recovery therapy, peer support sessions.
- 44:11So lots of things happening,
- 44:13and this is to keep veterans out of the jail basically.
- 44:19Sorry, close your eyes as I scroll back down.
- 44:26<v Donna>It's all so interesting,</v>
- 44:27but we have 10 minutes left.
- 44:28<v ->Okay, yeah,</v>
- 44:29so here are some of our effectiveness outcomes,
- 44:31which we're not assessing yet,
- 44:33but we're gonna look at linkage to care, overdose rates,
- 44:35criminal recidivism, et cetera.
- 44:39So from planning for sustainability,
- 44:41I'll just go and just give you a high level overview
- 44:44of what we're trying to do to do this.
- 44:45So we just started year four,
- 44:47we've decided this is the right time
- 44:49to really start engaging our partners.
- 44:50We talked about this at the November 7th meeting.
- 44:53And so just as a paper,
- 44:55this is a really useful paper to look at
- 44:57in terms of thinking about how do you design
- 44:59post dissemination and sustainability.
- 45:01I really have learned a lot from this group.
- 45:05And we're collecting a lot of qualitative data, as I said,
- 45:08we're actually putting that into our dashboard
- 45:12to sort of see what qualitative data emerges
- 45:14from each of the three phases of pre-implementation,
- 45:18implementation and sustainment,
- 45:19all guided by The Dynamic Sustainability Framework.
- 45:24This was a poster presented last year at the DNI conference.
- 45:27And we're using this tool,
- 45:29the Program Sustainability Assessment Tool,
- 45:31which is freely available online,
- 45:33developed by Doug Luke at Wash U.
- 45:35We do not have people fill out this 40 item survey,
- 45:38and this is like for our program partners.
- 45:40What we do instead is when we have our conversations
- 45:43like we just had with the technical expert panel meeting,
- 45:46we'll choose a few of these concepts to talk about.
- 45:49So what are we gonna need
- 45:51in terms of organizational capacity to keep this running?
- 45:54What is the funding going to be like?
- 45:56How do we adapt this to continue to make it useful?
- 45:59What information do you still need?
- 46:02So we are using this more in a conceptual way,
- 46:04and I do this with a lot of my NIH funded projects too.
- 46:08This is a very short and sweet pragmatic measure
- 46:11called PRESS to get at sustainment use.
- 46:14So these three questions, Donna, are being asked to people.
- 46:19This is a fairly new measure.
- 46:22So we are trying to see,
- 46:24when we're done with the implementation effort,
- 46:26are people using PIE, are people using HOPE,
- 46:28are people using MISSION
- 46:30now that we're no longer actively implementing?
- 46:34And then just as a cost piece that you brought up,
- 46:36we are using the coins,
- 46:38which is built on the sticks of coins
- 46:40as for cost of implementing new strategies.
- 46:42The stick is the stages of implementation completion,
- 46:45both have been developed by Lisa Saldana.
- 46:47So we are actually taking these eight steps
- 46:51of the stick through our tracker,
- 46:54we added them to our facilitation tracker,
- 46:56and we're deciding which of our activities
- 46:58are in pre-implementation, implementation and sustainment,
- 47:02and then we already have that data
- 47:04on how many hours and minutes, the personnel involved,
- 47:07and we are capturing those costs.
- 47:09I'll just say that we use the same exact approach
- 47:11in a Cory funded paper.
- 47:12This just came for Cory funded project,
- 47:14this paper just came out.
- 47:16I only leave it here just to show
- 47:18like a completed cost of implementation.
- 47:22You know, I'm not an economist, I led this,
- 47:25it took me out of my comfort zone,
- 47:27but I've decided that if I can do it, anyone can do it.
- 47:31And so really just wanna wrap up
- 47:33and say this is a gigantic village project,
- 47:37I think team science is everything.
- 47:39Everyone on here has something to contribute
- 47:41and it's absolutely not me,
- 47:45it's all of these people that we have made this possible,
- 47:49and we'll have our final results in two years.
- 47:52So we'll stay tuned
- 47:52to see how effective everything has been.
- 47:55So questions.
- 47:57(audience member speaking indistinctly)
- 48:05<v ->The question here is,</v>
- 48:07can we hear about power consideration
- 48:09for effectiveness versus implementation outcomes
- 48:11with hybrid studies one through three.
- 48:14Specifically, how much should we consider power
- 48:16for implementation outcomes in hybrid two and three studies?
- 48:19We did actually, not me,
- 48:24we did do power analysis on our hybrid threes.
- 48:27We estimated how many veterans
- 48:31we would need to include across all sites,
- 48:35so not just one site because we're aggregating data,
- 48:40and I think we have met that bar,
- 48:43but I would be very happy
- 48:46to put you more in touch with our statistician if you...
- 48:50I'm gonna capture this person's name
- 48:53and I'll put them in touch with our statistician.
- 48:55But yes, it's necessary to power
- 48:59because we are going to be using these regression analysis
- 49:02to determine whether which types of strategies
- 49:05actually led to increased uptake,
- 49:07but also the uptake is one implementation outcome,
- 49:10but we wanna look at the other implementation outcomes too.
- 49:15(audience member speaking indistinctly)
- 49:23Yeah, I think I actually have it in my proposal
- 49:27in my laptop somewhere that I could look up,
- 49:29but I think we had to do a milestone report
- 49:35of how many providers or people we thought we would train,
- 49:39peer support specialists, social workers,
- 49:41so we have that for every project,
- 49:43how many veterans that they would then serve.
- 49:46So we have that and we kind of have it by time,
- 49:50so after year one, year two, year three,
- 49:52of course when we wrote this proposal,
- 49:54COVID had not happened.
- 49:54We just submitted this in December of 2019,
- 49:58so we absolutely got behind on that.
- 50:01So we didn't follow the milestones
- 50:04despite best efforts as best as we can.
- 50:07Yes, so the provider piece is important,
- 50:09but then we also are learning all about their challenges
- 50:13with talking to veterans.
- 50:14Like the fact that we could get a provider
- 50:17really bought into this, trained, willing to implement it,
- 50:20but if their initial conversations
- 50:22with veterans are not positive, then that's a challenge too.
- 50:26So we're trying to interview veterans to learn more too.
- 50:33(audience member speaking indistinctly)
- 50:41Oh, I'll repeat.
- 50:43I'll repeat the question.
- 50:46(audience member speaking indistinctly)
- 51:07So the question is about...
- 51:09Thank you, Jen.
- 51:11The question is about how do you build trust by showing up
- 51:15when you have real world challenges of,
- 51:19you know, you aren't funded on a project at that point.
- 51:22So in this project were funded to do this,
- 51:25but we also had...
- 51:29Often, by the time you get to a hybrid three,
- 51:30we've already built relationships.
- 51:33These sites are new to us,
- 51:34but our program partners are not new,
- 51:36and so we had their backing to help us.
- 51:38But for people who are just starting out doing this,
- 51:42I often have had calls from, you know,
- 51:44a full professor saying,
- 51:46"I've never done this before. How do I do it?"
- 51:48And I literally say to them,
- 51:50"Please go have coffee with someone."
- 51:52Like, they'll say, "The only thing I've ever done
- 51:55is I've gone to a clinic
- 51:56and I've hung up flyers for my project."
- 51:58That's the extent of their engagement, and I get it,
- 52:03like all they needed is to recruit people from that site.
- 52:05That site had to say,
- 52:06"Sure, you can recruit people, but we're not gonna help you.
- 52:09This is something you do on your own."
- 52:10But I say, you know, yes, it takes time and effort,
- 52:15but try to meet with a clinic head, somebody,
- 52:19and just do not tell them what you wanna do.
- 52:22Go meet with them and say,
- 52:23"I would love to learn more about what matters to you.
- 52:27What are you trying to work on? What are your priorities?
- 52:30What keeps you up at night? What would you love to address?"
- 52:32And they'll tell you seven things right off the bat.
- 52:35One of those things might already align
- 52:37with what you wanna do.
- 52:39And at that point though, they feel like you're listening,
- 52:42you are, and then you can say, well, I have this idea,
- 52:44and you can start that conversation.
- 52:46If, for example, they don't say anything
- 52:48related to what you wanna do,
- 52:50then I say, then step back and go,
- 52:53well, clearly there's a mismatch.
- 52:55You know, and they're the ones who are living,
- 52:57breathing this day to day,
- 52:58and maybe your idea needs to change a little bit.
- 53:01But I love the idea of starting with asking questions
- 53:06and showing up as opposed to
- 53:08coming in with a fully developed specific aims page
- 53:10and saying, I really wanna do this, yeah.
- 53:13Is that what you meant? Yeah.
- 53:17And someone says,
- 53:19it'll be interesting to have an in-depth session
- 53:21on the methodologies you...
- 53:23Sorry, I can't read this thing, there's a little thing.
- 53:28Oh, that you'll apply for analysis
- 53:31of the stepped wedge design.
- 53:33Yes, luckily that isn't me,
- 53:37but so in our implementation core,
- 53:39we have a qualitative core and we have a quantitative core,
- 53:42and I meet with them.
- 53:44I'm a stronger qualitative person,
- 53:47that doesn't mean I haven't done quantitative analyses,
- 53:49but our quantitative statistician is Dr. Tom Berry,
- 53:54he's at BU,
- 53:56also our head economist Dr. Laura Saban is at BU,
- 54:00and so we meet with them regularly
- 54:03to talk through the different issues.
- 54:04But Tom has the homeless opioid use
- 54:10incarceration perspective
- 54:12as well as incredible statistical knowledge.
- 54:16So it's a great partnership.
- 54:20(indistinct)
- 54:22<v Donna>And so maybe we have to end,</v>
- 54:23but I can say that Rani has provided her email address
- 54:26as you can see here
- 54:27so I'm sure you'll welcome further comments and questions.
- 54:31<v ->Yes, absolutely, I'd be very happy to.</v>
- 54:34<v ->Thank you so much.</v> <v ->Thank you.</v>
- 54:36Thanks, everyone.
- 54:37I really appreciate it, great to see you all.