Trial design
The trial is an unrestricted, 2x2, clustered, sequential, multiple-assignment, randomized trial (SMART) [25-27] (see Figure 2). Because our goal is to test the effectiveness of the components of an implementation strategy while gathering clinical data related to patient outcomes, the study can be viewed as a hybrid type 3 effectiveness-implementation trial design [30]. All clinics will receive EM/AF to start. At the end of 3 months, half of clinics will be randomly assigned to receive PF for 18 months. At the end of 9 months, a second randomization occurs where half of clinics will receive PPC for 12 months in addition to previously assigned strategies. Clinics will have equal probability of being assigned to one of the four implementation sequences represented by the boxes (A, B, C, or D) on the right side of Figure 2.
Setting
We are recruiting primary care clinics from two health systems in the Midwestern U.S. One health system operates in the northeast region of Wisconsin and the southern part of the Upper Peninsula of Michigan, which are predominantly rural areas that have been particularly hard hit by prescription opioids. The second health system operates in the Madison, Wisconsin, metropolitan area and the surrounding, predominantly rural region of southcentral Wisconsin. Efforts to interest and involve clinics in the trial took place in the fall of 2019, and a run-in period started with the first educational/engagement meeting on February 13, 2020. The run-in period is planned to end and recruitment to close on May 13, 2020, when randomization takes place.
Clinic inclusion/exclusion criteria
Only primary care clinics (non-pediatric primary care, internal medicine, and family medicine) will be approached to participate. Clinics that explicitly prohibit initiating opioid therapy will be excluded (e.g., some clinics require that opioids be initiated by a specialty pain clinic). At baseline (before the start of EM/AF), a clinic will be considered ineligible if it already shows exemplary performance on key measures of guideline concordance and would thus receive no benefit from the implementation support we would provide. Specifically, we define a clinic as ineligible if it meets these criteria: (1) 80% or more of a clinic’s long-term opioid patients have treatment agreements and a urine drug screen in the last 12 months and (2) fewer than 10% of the clinic’s patients on long-term opioid therapy have doses above 90 MME.
Prescriber inclusion/exclusion criteria
Prescribers must be primary care physicians or other providers with prescribing privileges (e.g., nurse practitioners, physician assistants). We will exclude “float” providers (temporary physicians who do not manage stable panels of patients).
Patient inclusion/exclusion criteria
Patients included in the calculation of the prescriber-level outcome will have three consecutive months with an opioid prescription in the most recent three months documented in the electronic health record, indicating long-term opioid use. We will exclude patients from the calculation who have a cancer diagnosis or are receiving hospice care.
Randomization and stratification
After three months of EM/AF, eligible clinics will be stratified by health system, average number of patients at the clinic prescribed opioids over the first three months (above or below the median), and average MME over the first three months (above or below the median); clinics will then be randomly assigned with equal probability to PF or no PF arms within each of the eight stratum. At nine months, clinics will again be stratified by health system, by average number of patients at the clinic prescribed opioids over the past three months (median cut), and by average MME over the past three months; clinics within each of the resulting strata will be randomly assigned with equal probability to PPC or no-PPC arms. The project statistician will generate the random allocation sequence using a random number generator to perform block randomization with blocks of two and four. The study coordinator will enroll clinics and assign them to their randomized group. Consents will be obtained from prescribers and other clinic staff who participate in study activities.
Implementation strategies
Educational/engagement meetings
Educational/engagement meetings (EM) are conceived as a low-intensity, system-level implementation strategy. EM involves a broadcast model of communication, involving one or two experts imparting information to many clinicians at once. Educational/engagement
meetings will take place at the beginning of the study and then quarterly; a total of six will take place during the 21-month intervention period. The first will be a regionally hosted, in-person training session for each health system; attendees will have the option of participating via webinar. Before the first educational/engagement meeting, the implementation team will ask each clinic’s medical director to identify a change team leader to work with a group of three to seven clinic staff members on improving workflows related to opioid prescribing, and ask the medical director and change team leader to identify other members of the clinic’s change team. We will ask each clinic’s medical director, clinic manager, and change team leader to attend the educational/engagement meeting, minimally, but all change team members and others involved in clinic workflows related to opioid prescribing—prescribers (physicians, nurse practitioners, physician assistants), nurses, medical assistants, lab techs, and so on—will be invited. The educational/engagement meeting will be led by physicians from the University of Wisconsin-Madison with expertise in primary care and addiction medicine and extensive experience managing the care of long-term opioid patients. The session will be designed both to impart information and elicit early engagement and enthusiasm from clinic staff.
As part of educational/engagement meetings, the implementation team will ask system leaders and the medical directors of participating clinics about the challenges they face in opioid prescribing for patients with chronic non-cancer pain. Their responses will be used to focus each presentation on the issues most salient to each system. In addition to covering the current status of opioid prescribing in the system, the sessions will explain how implementation efforts address key issues identified by participants and include time for questions and answers. Subsequent educational/engagement meetings will take place as webinars, be led by physician experts, and cover such topics as effective tapering, opioid rotation, and balancing goals for quality of life, pain, and opioid dose.
Audit and feedback
The AF implementation strategy involves system-generated performance feedback reports being sent to participating clinicians. System- and clinic-level feedback reports will be introduced at the first educational/engagement meeting to initiate AF and referred to in all subsequent meetings. After the first meeting, a data coordinator at each clinic (a change team member asked to assume this role) will access and distribute monthly feedback reports to other change team members, prescribers, and the clinic’s medical director and manager.
The combination of EM and AF represents a blended, system-level implementation strategy. While system-level implementation strategies are relatively inexpensive and easy to scale, such strategies have limited evidence of effectiveness [17]. Nonetheless, simply learning and being convinced about what to do with respect to opioid prescribing—and having access to performance data that can be used to guide changes—may suffice to improve prescribing in some clinics. Moreover, the provision of subsequent strategies (e.g., PF and PPC) is expected to build synergistically on the foundational knowledge and reports generated by EM and AF, respectively.
Practice facilitation
Practice facilitation (PF) is a clinic-level strategy that targets clinic processes and workflows. In general, practice facilitation focuses on local customization and has a stronger evidence base than educational/engagement meetings and audit and feedback reports, which are usually not tailored to specific clinics [31]. Practice facilitation is also more labor intensive than educational/engagement meetings and audit and feedback reports. Solving workflow problems (e.g., streamlining processes around opioid prescription refills), which practice facilitation addresses, may be the key to improvement in some clinics.
In a clinic randomized to receive practice facilitation, an external change agent trained in practice facilitation (the facilitator) will visit the clinic in-person, and then follow up over the course of five monthly and four quarterly videoconferences or teleconferences to help clinics improve processes related to opioid prescribing, such as (1) ensuring that treatment agreements are initiated and regularly updated, and (2) integrating urine drug testing into clinic workflows. Clinic change teams consisting of a change team leader, a data coordinator responsible for distributing AF reports, at least one prescriber, and up to four other staff members will form the change team for opioid prescribing. The facilitator will work with the change team to use systems engineering tools (e.g., walkthrough exercises, flowcharting, and nominal group technique [32, 33]) to make changes in clinic workflows. The facilitator will reinforce the content of the EMs and guide teams in using their clinic and prescriber level AF reports to monitor progress towards goals.
Prescriber peer consulting
Prescriber peer consulting (PPC) is a prescriber-level strategy that aims to help prescribers manage their patients on long-term opioids by providing the opportunity to consult with a physician experienced in opioid management. PPC will be available to all prescribers at clinics randomized to receive this strategy; hence, it is conceptualized in this trial as a clinic-level strategy in terms of its delivery. Peer consultants will be physicians or pharmacists with relevant experience in opioid prescribing and addiction medicine nominated by health system leaders to help their peers manage patients on long-term opioid therapy (e.g., how to manage the tapering of opioid doses for long-term opioid patients with clinically indicated dose reductions). Participating prescribers in clinics randomized to PPC (including nurse practitioners and physician assistants) will receive up to four quarterly consulting sessions over 12 months. Consultations will be delivered via videoconference or teleconference. Prescribers may choose to include other staff at their clinic (e.g., RNs, MAs) in these consultations as well. Prescriber peer consulting is highly resource intensive, but our preliminary research suggests that physician-to-prescriber interaction may be the most effective way to change prescribing behavior.
Measures and outcomes
The study uses the RE-AIM model as an evaluation framework [34]. RE-AIM is a comprehensive evaluation framework that assesses five dimensions: Reach, Effectiveness, Adoption, Implementation, and Maintenance. Specific measures for each RE-AIM dimension are presented in Table 1. Evaluation data will come primarily from electronic health records (EHRs). Both health systems use Epic Systems’ EHR, which will facilitate the extraction of EHR data. One system was the site of our pilot research. Detailed specifications were developed during the pilot that will be used to ensure consistent data definitions across both systems. The primary outcome is prescriber average of MME dose per day per opioid patient, calculated over a three- month period.
Primary aim analyses
All clinics randomized at the end of month 3 will be included in the intent-to-treat sample for all aims. The primary research outcome, MME, will be available for all prescribers within all clinics that consent to be in the study (approximately 6 per clinic). Table 2 shows the sequences of implementation strategies that will be employed in the trial.
For the primary aim, we will determine the effect of strategy sequence D (the most intensive sequence of strategies) vs. strategy sequence A (the least intensive strategy) on change in MME from intervention months 3 to 21. Strategy sequence D offers EM/AF during months 3-21 months, augments with PF during months 4-21, and then further augments with PPC during months 10-21. By contrast, strategy sequence A offers EM/AF but never offers PF or PPC. This analysis is a mean comparison of change in MME between strategy sequence D versus A. The analysis will use a longitudinal (repeated-measures) analysis. Time will be coded such that t=0 denotes the end of month 3 of the intervention period (the initial randomization); in the following text, data collected prior to t=0 is considered baseline data. The primary outcome (MME) is a continuous measure and is collected at intervention month 3 (t=0, immediately prior to randomization) and every month up to intervention month 21 (t=18). The primary outcome is an average over 3 months; thus, there are a total of 7 measurement occasions. This is a 3-level analysis: repeated measures of MME, within prescribers, within clinics.
A piecewise-linear model with a knot at intervention month 9 (t=6, MME collected immediately before the second randomization) will be used to model the temporal trajectories over the course of intervention months 4-21. Figure 3 displays the planned longitudinal model we will use to model the mean MME over time, and test the primary aim.
Figure 3. Longitudinal model for mean MME
X are mean-centered baseline covariates (clinic aggregate MME at t=0 and a dummy indicator for health system); A1 is the indicator for the first randomization (PF=1 vs. no PF=-1); and A2 denotes the second randomization (PPC=1 vs. no PPC=-1). The model has a linear trend from t=0 to t=6 for prescribers in PF and no PF clinics, and a linear trend from t=6 to month t=18 for each of the four sequences of strategies (A-D). We allow for changes in the mean trajectory (i.e., deflections) at intervention month 9 (t=6) since this is the point at which clinician prescribers may begin receiving PPC. γ0 is the mean outcome at intervention month 3 (t=0), averaged across all four strategy sequences; γ1 is the average change in MME from month 3 (t=0) to month 9 (t=6), averaged across all four strategy sequences; 2*γ2 is the causal effect of PF vs. no PF on change in MME from intervention month 3 (t=0) to intervention month 9 (t=6); γ3 represents the average change in MME from intervention month 9 (t=6) to intervention month 21 (t=18), averaged across all four strategy sequences; 2*γ4 is the main causal effect of PF vs. no PF on change in MME from month 9 (t=6) to month 21 (t=18), averaged over PPC vs no PPC; 2*γ5 is the main causal effect of PPC vs. no PPC on change in MME from month 9 (t=6) to month 21 (t=18), averaged over PF vs no PF; γ6 is the interaction term to quantify whether and how PF and PPC work together to impact change in MME from month 9 (t=6) to month 21 (t=18).
The planned statistical test associated with the primary aim (for which we power the study) is a test of the null hypothesis that 12γ2 + 24γ4 + 24γ5 = 0; that is, that there is no difference on change in MME from month 3 (t=0) to month 21 (t=18) between implementation sequence A vs. implementation sequence D. We will report estimates of each coefficient in the model with their corresponding 95% confidence intervals.
Since the covariates X and the primary outcome data are available and passively collected from the EHR, except in rare cases (e.g., clinician turnover, clinician retirement, or an error leading to data loss in the EHR), we expect to have little missing data.
Additional file 1 describes the planned analysis for the primary outcome in more detail.
Sample size and power
The total sample size for this study is based on the primary aim: a comparison on average difference on change in MME from intervention month 3 (t=0) to intervention month 21 (t=18) between implementation sequence D vs. implementation sequence A. This is a comparison between two of the four groups embedded in the trial (see Table 2). The sample size calculator for this comparison is a straightforward adjustment to the sample size calculator for a standard two-sample hypothesis test. The adjustment accounts for the clustering of prescribers within clinics through a variance inflation factor (VIF) of 1+(m-1) ρ, where m is the (average) number of prescribers per clinic and ρ is inter-clinic correlation coefficient (ICC) for MME at month 21 (t=18). Based on intervention clinics in the R34 pilot data, the ICC was estimated to be ρ = 0.14. Assuming an average of m = 6 prescribers per clinic (based on information from the new health systems that have agreed to participate), a Type-1 error rate of α = 5%, and ρ = 0.14, a minimum of 64 prescribers in each group (11 clinics per group) will provide at least 80% power to detect a moderate effect size of d=2/3 between the two implementation sequences on change in MME. Because we have four groups in this trial, the minimum total study sample size is 256 clinician prescribers, corresponding to roughly N=40-45 clinics (depending on prescriber count).
Based on the pilot data that found a standard deviation of 35 for MME, an effect size of d=2/3 corresponds to detecting an average difference of at least 23 on the MME between the two implementation sequences after 21 months. The above calculation is expected to be conservative because it does not account for within-prescriber correlation in MME, which is accounted for in the longitudinal analyses and could permit detection of smaller differences in MME.
Exploratory aim
Q-learning [35]—a generalization of moderated regression analysis to multiple stages of implementation—will be used to test the moderators and construct a candidate adaptive implementation strategy.
Secondary aim
In the secondary aim, we will estimate the cost of delivering four different sequences and combinations of strategies (EM/AF, EM/AF+PF, EM/AF+PPC, and EM/AF+PF+PPC), including the incremental cost effectiveness of adding facilitation and prescriber peer consulting. Results will help decision-makers weigh the costs and effects of using different sequences of implementation strategies.
In line with the pragmatism that underlies this research, we will employ an operational cost analysis based on tenets of engineering economics. Traditional health economic approaches incorporate concepts from welfare economics and take a societal perspective towards decision analysis [36, 37]. Engineering economic analysis tends to have a narrower scope. Whereas health economic evaluation provides information primarily for policymakers, engineering economic analysis produces information primarily for the organizational leaders who ultimately make decisions about the adoption of evidence-based practices in their organizations. We adopt the perspective of the healthcare system (rather than society at large) in considering the incremental costs and effects associated with ratcheting up the implementation strategy. This perspective deemphasizes some societal costs (e.g., patient travel time) and effects (e.g., crime related to addiction) that are often considered in traditional cost-effectiveness analysis [36]. However, the health system perspective aligns with updated guidelines for cost-effectiveness that were re-issued in 2016 [37] that acknowledge the importance of the health care perspective for pragmatic purposes. The health systems we will work with, like many health systems, are Accountable Care Organizations, which means that they are responsible for their patients’ total cost of care. The pragmatic optimization approach featured in this aim was designed in close partnership with our research collaborators to model the considerations healthcare decision makers told us they actually use when making decisions about adopting and sustaining evidence-based practices.
We developed an approach to costing the systems consultation strategy in our pilot research [18]. Detailed logs were kept of all contacts between members of the research team and the clinic change teams to estimate the number of hours spent delivering systems consultation. These estimates were multiplied by hourly wage rates for physician consultants and the facilitators. Costs for clinic incentives (continuing education credits, clinic stipends) and expenses associated with site visits were also included in the cost assessment. The total cost of delivering the entire systems consultation implementation strategy for six months (i.e., the full package of strategies corresponding to box D in Figure 2) was estimated in the pilot research. The log-based costing approach is sufficiently fine-grained to construct detailed breakdowns of unit costs associated with each component (EM/AF, PF, and PPC) of the full systems consultation intervention [18].
We will use incremental cost effectiveness ratios (ICERs) to quantify the tradeoff between the additional effectiveness achieved through scaling up the intensity of implementation strategies. The primary ICER will be the incremental cost per unit reduction in MME. We will use a 21-month timeline for the cost analysis. Implementation costs will be organized using the Cost of Implementing New Strategies framework [38], and categorized under the “Implementation” domain of RE-AIM (see Table 1). Secondary ICERs will include incremental cost per unit change in prevalence of opioid / benzodiazepine co-prescribing, completion of treatment agreements, urine drug screens, and so on, as shown in Table 1. Cost-effectiveness acceptability curves will be generated (using Monte Carlo simulation techniques) for all primary and secondary ICERs to model uncertainty in our estimates of cost-effectiveness [39].
Trial status
Sites were identified and participation confirmed by January 31, 2020. Site training began with the first educational/engagement meetings, which were held in February and March 2020. The trial was temporarily suspended on March 25, 2020, because the coronavirus pandemic took priority in both health systems.