Optimisation of complex health interventions prior to a randomised controlled trial: a scoping review of strategies used
Pilot and Feasibility Studies volume 2, Article number: 17 (2016)
Many complex intervention trials fail to show an intervention effect. Although this may be due to genuine ineffectiveness, it may also be the result of sub-optimal intervention design, implementation failure or a combination of these. Given current financial constraints and the pressure to reduce waste and increase value in health services research, pre-trial strategies are needed to reduce the likelihood of design or implementation failure and to maximise the intervention’s potential for effectiveness. In this scoping review, we aimed to identify and synthesise the available evidence relating to the strategies and methods used to ‘optimise’ complex interventions at the pre-trial stage.
We searched MEDLINE, CINAHL, AMED, PsycINFO and ProQuest Nursing & Allied Health Source for papers published between January 2000 and March 2015. We included intervention development and optimisation studies that explored potential intervention weaknesses and limitations before moving to a definitive randomised controlled trial (RCT). Two reviewers independently applied selection criteria and systematically extracted information relating to the following: study characteristics; intervention under development; framework used to guide the development process; areas of focus of the optimisation process, methods used and outcomes of the optimisation process. Data were tabulated and summarised in a narrative format.
We screened 3968 titles and 231 abstracts for eligibility. Eighty-nine full-text papers were retrieved; 27 studies met our selection criteria. Optimisation strategies were used for a range of reasons: to explore the feasibility and acceptability of the intervention to patients and healthcare professionals; to estimate the effectiveness and cost-effectiveness of different combinations of intervention components; and to identify potential barriers to implementation. Methods varied widely across studies, from interviews and focus groups to economic modelling and probability analysis.
The review identifies a range of optimisation strategies currently used. Although a preliminary classification of these strategies can be proposed, a series of questions remain as to which methods to use for different interventions and how to determine when the intervention is ready or ‘optimised enough’ to be tested in a RCT. Future research should explore potential answers to the questions raised, to guide researchers in the development and evaluation of more effective interventions.
Complex health interventions (CHIs) are defined as multicomponent interventions in which individual, collective and organisational elements act both independently and interdependently . Interactions between intervention components and their effects on outcomes are not always linear or obvious, and they are influenced by several factors . These include, for example, the number of interacting components, the intensity of behaviour change required by those delivering or receiving the intervention, the number of groups or organisational levels targeted by the intervention and the complexity of outcomes, as well as the context in which interventions are implemented [3–5]. This results in considerable challenges to the evaluation of CHIs, which in turn requires substantial resources.
Randomised controlled trials (RCTs) are historically recognised as the ‘gold standard’ methodology in the evaluation of interventions and they have a long record of successful application in determining a causal relationship between an intervention and its putative outcomes . However, in the case of an intervention that does not influence the outcomes as expected, trials often fail to detect or report whether the lack of intervention effect is due to sub-optimal intervention design, implementation failure or genuine ineffectiveness [7, 8].
As Sermeus states, the increasing number of components that characterise interventions leads to them being even more complex, less understood and much harder to implement . This raises two specific questions; the first one, how to understand if the intervention works as predicted, and the second when it is time to move to the evaluation phase and test the intervention in a full-scale RCT. Methods to improve intervention design, reduce implementation failure and enhance trial processes have developed considerably over the past 15 years and several frameworks and practical guidelines have been issued. In this paper, we focus on probably the least explored and understood process related to the development of complex interventions: the optimisation of the intervention under development prior to a full-scale RCT. In the late 2000s, Collins and colleagues introduced the multiphase optimisation strategy (MOST) framework—a strategy for developing and optimising behaviour interventions. The element of focus of this framework is the role of the different intervention components and their contribution to the overall success of the intervention, as complex interventions may contain inactive components . The framework proposes to adopt a programmatic and sequenced experimental approach that can efficiently and systematically identify the most promising components, in order to assemble these in an optimised version of the intervention, which is finally tested in a RCT. While the MOST framework has some conceptual roots in the phased approach to intervention development and evaluation proposed by the Medical Research Council (MRC), it draws attention to the importance of optimising complex interventions—where optimised interventions are defined as ‘the most effective intervention given certain constraints’, such as for example the resources available for the intervention or the time available for the delivery (e.g. intervention delivered for ≤$500 or for a maximum of 10 h/week per healthcare professional).
Alternative approaches to CHI optimisation include strategies proposed by the MRC framework 2000  and 2008 , the normalisation process theory (NPT) [11, 12] and the process modelling in implementation research (PRIME) approach . The key stages of each of these frameworks and guidelines are shown in Fig. 1. These guidelines and frameworks acknowledge the need to limit sub-optimal intervention design and implementation failure. As such, they all emphasise the importance of testing the intervention’s potential effect and evaluating how interventions work before embarking on a full-scale RCT (Fig. 1). However, researchers could be forgiven for not knowing which methods to use and when. Existing frameworks differ in the language and terminology used, and there is a lack of clarity over the specific purpose and scope of each proposed stage of work to be conducted before the full-scale RCT . In addition, the different guidelines and frameworks propose a range of methods, from computer simulations and factorial experiments to qualitative studies involving key stakeholders. But evidence to support the use of these methods for particular purposes is lacking, and there is limited guidance on the specific detail of how to plan and design optimisation studies. This leads to confusion about which guideline or framework to follow and which optimisation strategy is likely to be most suitable for the different types of intervention under evaluation.
For the purpose of this review, we have defined optimisation as a process aimed to evaluate or test intervention components and/or drafted interventions in order to identify what works and what does not work within the intervention under design. Thus, the ultimate aim of optimisation processes is to isolate those interventions or intervention components that are more likely to be effective if implemented in a full-scale trial setting. For this reason, we use the term ‘optimisation strategies’ as an umbrella term to encompass a wide range of approaches, such as those referred to above, used to optimise the intervention itself before moving to a full-scale RCT. This means that we are not exploring optimisation strategies aimed to optimise trial parameters, such as recruitment and randomisation processes, in preparation for the main trial. Furthermore, this review focuses exclusively on those strategies adopted before moving to the full-scale RCT stage. As such, optimisation processes may represent a separate stage or can be integrated within the development or pilot and feasibility phase.
This scoping review aimed to explore the strategies and methods currently used by researchers to optimise CHIs before the definitive trial stage so as to understand how, when and why certain strategies might be most usefully applied.
The rapid increase in the amount of primary research available has led to the development of different and new strategies for synthesising evidence in a more effective and rigorous way . Scoping reviews represent a useful and increasingly popular method of collecting and organising important background information on a topic and are described as a process for mapping the existing literature. In 2005, Arksey and O’Malley proposed a framework for conducting scoping reviews, which included the following five iterative stages: (1) identifying the research question(s); (2) identifying relevant studies; (3) study selection; (4) charting the data; (5) collating, summarising and reporting the results . Scoping reviews can be conducted for several reasons, such as to map fields of studies where it is difficult to anticipate the range of material that might be available, to determine the value of undertaking a systematic review and define more precise questions and suitable inclusion criteria, to identify research gaps in the existing literature or to clarify working definitions and/or the conceptual boundaries of a topic. Typically, scoping reviews differ from systematic reviews in several ways, as outlined in Table 1. Scoping reviews, in particular, identify a broader ‘scope’ and research questions with less restrictive inclusion and exclusion criteria, which are determined in an iterative way on the basis of familiarity with the literature [17, 18]. Another important distinction between scoping reviews and systematic reviews is that, unless otherwise specified, a quality assessment of the included studies is generally not performed .
The purpose of this scoping review was to map the literature available on the optimisation of CHIs before moving to an RCT and to identify potential gaps in the current literature. The review followed the iterative stages proposed by Arksey and O’Malley, with each feeding into the next stage (Fig. 2). To add rigour to the review process, a systematic team approach was adopted. Team meetings included iterative discussions surrounding keywords to be searched, inclusion/exclusion criteria and study selection at different stages of the review process. Prior to conducting this scoping review, the literature was searched in order to identify any frameworks and guidelines for researchers that had been published on the development and evaluation of complex health interventions. This enabled us to familiarise ourselves with the different recommended stages of intervention development and the terminology used to refer to optimisation processes. The recent guidance issued by Joanna Briggs Institute  for reporting scoping reviews is used here to describe the different criteria and processes adopted in our scoping review.
For the purpose of this review, we included any type of pre-trial study, such as intervention development studies, pilot and feasibility studies that either explicitly referred to optimisation or described processes that fitted with our definition of optimisation studies. The review used an exploratory approach and it encompassed any type of CHI, including those targeting patients and those designed to address healthcare professionals’ practice. Final selection criteria agreed by the review team at multiple consensus meetings are shown in Table 2.
To conduct this scoping review, we searched the following electronic databases, MEDLINE, CINAHL, AMED, The Cochrane Methodology Register and PsycINFO and ProQuest Nursing & Allied Health Source for peer-reviewed publications. We limited our searches to publications in the English language only and those published between 1 January 2000 to 31 March 2015. The search strategy used a combination of key terms related to complex health interventions, keywords related to the overall process of optimising complex interventions, together with framework or guideline-specific keywords used to describe an optimisation process (e.g. ‘modelling’ for the MRC framework, ‘intervention modelling experiment’ (IME) for the PRIME approach). A multi-database searching strategy was adopted. Boolean operators were used in order to maximise the penetration of terms searched, while appropriate ‘wild cards’ were adopted to account for plurals, variations in databases and spelling. Reference lists of relevant publications and key journals were hand-searched.
Because of the broad nature of scoping reviews, databases were searched at different points in time, and results were used to inform the following searches in order to get a balance between breadth of the search and relevant hits. Appendix 1 shows examples of search strings used together with the decision-making process that led to the constant refinement of the search strategy, in order to ensure the coverage of the most relevant literature.
The variation of terminology used in the field and the identification of the most appropriate keywords represent one of the main challenges we faced in this review. It transpired that different terms could have been used to refer to an optimisation study and thus we conclude that there is a substantial lack of consensus on the terminology and core concepts related to ‘optimisation’. Thus, for the purpose of this scoping review, not only did we search for papers that specifically used the word optimisation, as this would have led to the identification of only a subset of all possible studies. Moreover, this review used a search strategy developed through the different stages of the review process, which combined keywords related to the overall process of optimising complex interventions together with framework or guideline-specific keywords used to describe an optimisation process (Appendix 1). In this way, we aimed to identify a heterogeneous group of studies that optimised the intervention under development prior to the full-scale RCT by following different frameworks and guidelines.
One author (SL) screened all of the titles and removed any obviously irrelevant records, such as studies from other fields (e.g. biomedical and pharmaceutical). One reviewer (SL) assessed all abstracts for relevance. To check for rater reliability, a second review author (RF) independently assessed a randomised subset of 15 % of abstracts and the full text of all papers for which there was uncertainty about inclusion. Disagreement was resolved by consensus between SL and RF, with input from a third reviewer (SH) where necessary.
Extraction of results
Data were extracted from the included studies using a structured form. We systematically extracted information relating to study characteristics (author, date of publication and country of the study), intervention under development, framework used to guide the development stage, areas of focus of the optimisation process and the methods used. Although there is strong emphasis in the literature on the need for clear objectives for any study, such as pilot and feasibility studies , not all the included studies clearly reported the objectives of the optimisation studies conducted. However, it was possible to identify an overall area of interest for each reported use of an optimisation process. Thus, for the purpose of this paper, the term areas of focus was used to identify the different objectives and more generally the areas addressed by each optimisation process included . In the case of missing data, attempts to contact the corresponding author of the study were made by SL. The data extraction process and form were initially piloted by two review authors (SL and RF) on five papers. Each author then independently extracted data from the remainder of the studies.
We used the hierarchical task analysis (HTA) approach to explore the mechanisms and the structure that characterised optimisation processes within each individual study. Hierarchical task analysis, developed in 1971 by Annett, is an engineering and decision analysis-based process for decomposing and describing an activity, which can be used to analyse any type of task in any domain . A key feature of HTA is that tasks—what people are seeking to achieve—are defined by goals. Thus, complex tasks, such as optimisation studies, can be analysed by deconstructing a hierarchy of goals, sub-goals and activities with a parent–child relationship at each level in the hierarchy . In particular, each individual study was decomposed according to (1) the aim and the area of focus of the optimisation process, (2) the methods adopted and (3) the outcome of the optimisation process.
Graphical representations of each study flow were subsequently analysed and compared, in order to compare the tasks involved and the structure of different optimisation studies. Following this strategy, we were able to identify and explore similarities and differences across various optimisation processes for all included studies.
After removal of duplicates, we screened the titles of 3968 papers. We identified 231 potentially relevant studies and, after abstract screening, 89 full papers were considered for inclusion in the review. Twenty-seven studies were finally included. Results of the search are displayed in Fig. 3. Table 3 provides an overview of the intervention, geographical location, framework, methods and area of focus of the optimisation process for each included study.
Most of these studies were conducted in Europe and specifically in the UK (n = 19, 70.4 %), the Netherlands (n = 2, 7.4 %), Ireland (n = 1, 3.7 %) and Norway (n = 1, 3.7 %). Three studies were conducted in the USA (11.1 %), and one study was conducted in New Zealand (3.7 %). The majority of included studies were published from 2011 onwards (n = 17, 63.0 %) (Table 3).
Guidelines or frameworks used to guide the intervention development process
A range of different guidelines or frameworks for the development and evaluation of CHIs were employed by authors of the included studies. In particular, 17 of the 27 studies included in this review used the MRC 2000 framework, the updated 2008 version or a combination of the two [24–40]. Two studies adopted the intervention mapping framework  developed by Bartholomew and colleagues in 1998 [42, 43] and two applied the MOST framework [10, 44]. In addition, two studies followed the intervention modelling process [45, 46] and one study conducted in New Zealand introduced and applied new guidelines specifically for the development of mobile health interventions . The remaining studies did not specify the guidelines or frameworks adopted to develop the intervention of interest [48–50].
Types of intervention
The interventions reported across the review varied widely and included a few targeting behavioural change at the individual patient level, such as weight reduction  and smoking cessation programs [10, 44, 49], or at the level of healthcare professionals, such as interventions targeting general practitioners to reduce inappropriate prescribing behaviours [26, 45, 46, 48]. Interventions were delivered in a variety of settings (e.g. inpatient, outpatient clinics and home-based settings) and targeted a wide range of conditions, such as mental health conditions [28, 30, 35, 40, 47], stroke [31, 36, 39], cancer [29, 32, 33, 37, 42] and other chronic illnesses [25, 34, 38]. Two studies reported on preventive interventions targeting older people [24, 50] and one on a programme to empower patients undergoing gynaecological surgery during the perioperative period .
Figure 4 shows a representative example of how we applied the HTA approach to one of the included studies . Results emerging from the comparisons of the different tasks are described under the following main conceptual categories: areas of focus, methods used and outcome of the optimisation processes.
Areas of focus of optimisation processes
The common aim of all included studies was to evaluate or test health interventions or their components before moving to a definitive trial. However, closer inspection using the HTA approach showed that they focused on a series of different aspects of the intervention under development.
Fourteen studies explored the acceptability and/or feasibility of the intervention to different stakeholders, including providers, recipients or policy makers and service commissioners [27, 29–32, 34, 36, 37, 39, 40, 42, 43, 47, 48]. Seven studies were a preliminary investigation of the effect of the intervention or a combination of its components on proxy or final outcomes and intervention long-term impact [10, 24, 44–46, 49, 50]. Six studies evaluated more than one element; in particular, they explored the acceptability to different stakeholders together with either the intervention’s effectiveness [25, 26, 28, 35, 38] or cost-effectiveness .
A number of different methodological approaches were used to undertake the optimisation process (Table 3). Several studies employed well-established qualitative methods including interviews or focus groups with providers and recipients or relevant stakeholders [26, 30, 32, 34, 39, 47, 48] and observations and consensus processes including researchers, patients and clinical experts . A number of studies adopted quantitative methods such as fractional factorial experiments [10, 44], economic modelling , small uncontrolled pilot studies (with no control group) [25, 29, 38], experimental 2 × 2 randomised controlled trials [45, 46], probability models [49, 50] and evaluation questionnaires . Several studies used mixed methods, combining qualitative strategies to explore stakeholders’ perspectives and quantitative analysis to estimate the intervention effect [24, 27, 28, 33, 35–37, 40, 42]. The level of detail reported on the methodology used varied widely across the studies (Table 3).
Overall, the use of specific methods appears to be informed by the research questions and the areas of focus of the single optimisation study. For example, factorial experiments, uncontrolled pilot studies and probability models tended to be used to test specific components and to identify an intervention’s potential effect [10, 25, 29, 38, 44, 49, 50]. Optimisation strategies applying qualitative methods tended to explore acceptability and feasibility issues [26, 30, 32, 34, 39, 47, 48]. On close inspection, it was possible to classify optimisation strategies in relation to when the optimisation process takes place with reference to the pilot stage. In particular, prospective strategies tend to look at theoretical feasibility and acceptability issues and are completed before the pilot stage; as in the optimisation studies conducted by Barley  and Byng and Jones , where focus groups with patients and healthcare professionals were run to explore the potential acceptability and feasibility of the proposed intervention to patients and practitioners. In vivo strategies, such as the real-time re-modelling and optimisation approach applied by Palmer and colleagues , explore actual implementation issues, by allowing researchers to immediately respond to problems and deviations arising in practice in a pilot setting. The last category, the retrospective strategies tend to explore stakeholders’ feedback once they have piloted the drafted intervention; thus, as in the optimisation study conducted by Lewis , a group of patients affected by post-traumatic stress disorder who took part in the piloting of the drafted intervention were interviewed to explore their perspectives on the intervention and its components, in order to identify potential required changes.
The included studies reported on the involvement of different stakeholders, who can be grouped as follows: service users, such as patients, informal carers and family members; service providers, including GPs, nurses and other healthcare professionals; and ‘other’ stakeholders, including academics and researchers, organisation representatives and policy makers. Of note, there were conflicting reports about the added value of stakeholders in optimisation studies. The involvement of stakeholders was identified as a strength by several studies, as stakeholders play a key role in the implementation of interventions into practice [30, 32, 33, 42, 43, 45]. However, other studies suggested that the process of identifying stakeholders and the potential lack of representativeness of the stakeholders involved were a limitation for optimisation studies [27, 37].
Outcome of the optimisation processes
Overall, the optimisation processes implemented across the studies included in this review were viewed positively by the authors; providing researchers with important information about the potential effectiveness of the intervention and informing decisions on how (or whether) to proceed to the next stage. The majority of studies (n = 16) subsequently refined the intervention to then test the optimised version of the intervention within pilot studies or full-scale RCTs [24–32, 34, 38–40, 43, 47, 48]. In two studies, findings from optimisation processes discouraged researchers from moving to the full-scale RCT stage, as the interventions did not show the potential for effectiveness that was expected [35, 50]. A few studies reported that the optimisation processes allowed the research team to identify the ‘best ingredients’ out of several candidate components [33, 35–37, 42]. Despite attempts to contact the leading authors, data on the outcome of two of the included optimisation studies were not available [10, 44].
To the best of our knowledge, this is the first review to synthesise the different strategies used to optimise CHIs prior to full-scale RCTs. Empirical studies were identified to map current practice and identify gaps in the literature on intervention optimisation. Interestingly, most studies identified were published within the last five years, suggesting a recent surge in interest from the research community in maximising, up-front, the potential effectiveness of CHIs and reporting pre-trial evaluation processes. This is in line with the consensus that pre-trial stages are increasingly seen as critical steps in the development of new interventions, as sub-optimal intervention design may result in weak effects .
The novel use of the HTA method to facilitate decomposition of optimisation studies allowed us to isolate (a) factors that are currently prioritised and tested prior to evaluation by RCT, (b) the methods used in optimisation and (c) the outcome of each optimisation process. The areas of focus of the different optimisation studies included the acceptability and feasibility of the intervention to key stakeholders but also the exploration of the potential effects and cost-effectiveness of the intervention. This suggests that the acceptability of the intervention to those directly involved in the delivery and receipt of the final intervention, together with the anticipated effect of the intervention, are important elements to take into account as early as possible in the pre-trial stage. Results also indicate that the area of focus of the optimisation process informed decisions on which methods to use. For example, economic modelling and probability models were used to explore costs and the potential effectiveness of interventions, whereas interviews and questionnaires were generally used to evaluate acceptability to service clients and healthcare professionals.
From an in-depth analysis of the included studies, we were able to classify optimisation strategies as prospective, in vivo and retrospective. It is possible to argue that prospective strategies tend to explore theoretical perspectives and hypothetical intentions from different stakeholders in relation to the proposed intervention, presented through oral or video presentations and informative material. On the other hand, retrospective strategies look for stakeholders’ feedback and involve people (such as service clients and providers) who have piloted the intervention in a small, often uncontrolled, study. In vivo strategies, instead, look at implementation issues, in order to identify and apply potential changes to the draft intervention. When retrospective and in vivo optimisation strategies are applied, stakeholders are involved in the piloting of the drafted intervention; thus, their feedback on feasibility and acceptability is informed by a ‘real’ experience of the intervention. Whereas, prospective strategies allow stakeholders to influence from the outset rather than once the piloting is underway. This suggests that different strategies can be applied in different situations and for different purposes; for example, prospective strategies might be more appropriate in the design and development of the intervention, to help researchers identifying those components that increase the feasibility and acceptability of the intervention to the groups of people directly involved. Retrospective strategies might be helpful to gain confirmation of the potential effect of the intervention and its potential feasibility. Finally, in vivo strategies might be used in those situations where researchers are looking to implement changes during the pilot process to immediately verify how these influence the intervention effect.
Given the current financial constraints upon health services research and the large number of trials that fail to show effectiveness, it has been suggested that it is increasingly important to define strategies that support researchers in the development of more effective interventions [3, 50, 51]. This review evidences the emerging role of optimisation studies in developing interventions that are potentially more likely to be effective and highlights that a range of strategies are used for a range of different purposes and that greater clarity in both the terminology and the selection of different methods in order to develop and improve interventions would be helpful.
Some of the studies included suggested that optimisation has the potential to support researchers in identifying interventions or components that are likely to fail or show little effect if implemented in a full-scale RCT, but questions as to when the intervention is ready to be evaluated in a realistic setting and how researchers decide whether to move to a full-scale RCT still need to be answered. As Sermeus  suggests, complex interventions indeed generate some effect, but ‘the real question is how to establish when this is enough’ and when the intervention has been optimised enough to be the best intervention possible.
We believe the findings and the questions emerging from this scoping review should inform future research exploring the mechanisms of actions and the benefits and challenges of conducting optimisation studies. This review could also be used to generate much needed discussion amongst healthcare researchers undertaking complex intervention trials on when to apply different methods and which are most useful in relation to different circumstances and types of intervention. Furthermore, findings from this review could help researchers in thinking about and planning future optimisation studies, which could ultimately lead to the design of more successful RCTs.
Strengths and limitations
To our knowledge, this is the first review to document the literature available on pre-trial strategies for the optimisation of CHIs. The HTA analytical approach enabled a detailed analysis of the different tasks and tools involved in different optimisation strategies, which led to the development of a preliminary classification of optimisation strategies. Furthermore, the iterative nature of the scoping review allowed us to extensively explore the literature available on the topic of enquiry; however, by its nature and the issues related to the terminology currently in use, it does not aim to be exhaustive. It is therefore possible that other optimisation strategies exist, which have not been captured by this review.
Implication for research
Future research should explore, in-depth, the decision-making process behind optimisation studies, the benefits and challenges of optimising CHIs and those related to specific optimisation strategies, by gaining researchers’ accounts of the process of optimising complex interventions. Furthermore, examples of RCTs of complex interventions should be investigated in order to explore the impact of optimisation processes on the effectiveness of the intervention implemented in a real-life setting.
In summary, our review explored strategies and methods that are currently used prior to a definitive RCT to assess situations of sub-optimal intervention design and to anticipate potential implementation failure. Findings from this scoping review represent the first step towards helping healthcare researchers to plan and conduct studies aimed at identifying what works and what does not work within the intervention under design, in order to ensure that those interventions and intervention components which proceed to full-scale RCT are those most likely to be effective. The review suggests a classification of optimisation strategies which is of help in understanding which methods to use in different situations, but it also raises a series of questions in relation to how researchers know when the intervention is ready or optimised enough to move to the full-scale trial stage. In the current economic climate, the answers to these questions are deemed invaluable in fostering a wiser use of public funding for the development and evaluation of more effective interventions.
complex health interventions
hierarchy task analysis
multiphase optimisation strategy
Medical Research Council
normalisation process theory
process modelling in implementation research
randomised controlled trial
Campbell M, Fitzpatrick R, Haines A, Kinmonth AL, Sandercock P, Spiegelhalter D, et al. Frameworks for design and evaluation of complex interventions to improve health. BMJ. 2000;321:694–6.
Shepperd S, Lewin S, Straus S, Clarke M, Eccles MP, Fitzpatrick R, et al. Can we systematically review studies that evaluate complex interventions? PLoS Med [Internet]. 2009;6(8):e1000086.
Craig P, Dieppe P, Macintyre S, Health P, Unit S, Michie S, et al. Developing and evaluating complex interventions: new guidance. BMJ [Internet]. 2008;337(October):979–83.
Byford S, Sefton T. Economic evaluation of complex health and social care interventions. Natl Inst Econ Rev [Internet]. 2003;186(1):98–108.
Wells M, Williams B, Treweek S, Coyle J, Taylor J. Intervention description is not enough: evidence from an in-depth multiple case study on the untold role and impact of context in randomised controlled trials of seven complex interventions. Trials [Internet]. Trials. 2012;13(1):95.
Rychetnik L, Frommer M, Hawe P, Shiell A. Criteria for evaluating evidence on public health interventions. 2006. p. 119–27.
Treweek S, Ricketts IW, Francis J, Eccles M, Bonetti D, Pitts NB, et al. Developing and evaluating interventions to reduce inappropriate prescribing by general practitioners of antibiotics for upper respiratory tract infections: a randomised controlled trial to compare paper-based and web-based modelling experiments. Implement Sci [Internet]. 2011;6:16.
Campbell NC. Designing and evaluating complex interventions to improve health care. BMJ. 2007;334:455–9.
Sermeus W. Modelling process and outcomes in complex interventions. In: Richards D, Hallberg IR, editors. Complex interventions in health: an overview of research methods. 1st ed. Oxon: Routledge; 2015. p. 408.
Collins LM, Murphy SA, Nair VN, Strecher VJ. A strategy for optimizing and evaluating behavioral interventions. Ann Behav Med A Publ Soc Behav Med [Internet]. 2005;30(1):65–73.
May C, Finch T, Mair F, Ballini L, Dowrick C, Eccles M, et al. Understanding the implementation of complex interventions in health care: the normalization process model. BMC Health Serv Res [Internet]. 2007;7(148):1–7.
Murray E, Treweek S, Pope C, MacFarlane A, Ballini L, Dowrick C, et al. Normalisation process theory: a framework for developing, evaluating and implementing complex interventions. BMC Med [Internet]. 2010;8:63.
Walker AE, Grimshaw J, Johnston M, Pitts N, Steen N, Eccles M. PRIME—PRocess modelling in ImpleMEntation research: selecting a theoretical basis for interventions to change clinical practice. BMC Health Serv Res [Internet]. 2003;3(1):22.
Arain M, Campbell MJ, Cooper CL, Lancaster GA. What is a pilot or feasibility study? A review of current practice and editorial policy. BMC Med Res Methodol [Internet]. 2010;10:67.
The Joanna Briggs Institute. Joanna Briggs Institute Reviewers’ Manual: 2015 edition / Supplement. Methodology for JBI Scoping Reviews. Adelaide: The Joanna Briggs Institute; 2015.
Arksey H, O’Malley L. Scoping studies: towards a methodological framework. Int J Soc Res Methodol [Internet]. 2005;8(1):19–32.
Daudt HML, van Mossel C, Scott SJ. Enhancing the scoping study methodology: a large, inter-professional team’s experience with Arksey and O’Malley’s framework. BMC Med Res Methodol [Internet]. BMC Med Res Methodol. 2013;13(1):48.
Levac D, Colquhoun H, O’Brien KK. Scoping studies: advancing the methodology. Implement Sci [Internet]. 2010;5(69):1–9.
Armstrong R, Hall BJ, Doyle J, Waters E. Cochrane update. “Scoping the scope” of a Cochrane review. J Public Health (Oxf) [Internet]. 2011;33(1):147–50.
Lancaster GA, Dodd S, Williamson PR. Design and analysis of pilot studies: recommendations for good practice. J Eval Clin Pract. 2001;10(2):307–12.
Bowen DJ, Kreuter M, Spring B, Cofta-Woerpel L, Linnan L, Weiner D, et al. How we design feasibility studies. Am J Prev Med [Internet]. 2009;36(5):452–7.
Annett J, Stanton NA. Task analysis. London: Taylor & Francis; 2000.
Czaja SJ, Schulz R, Lee CC, Belle SH. A methodology for describing and decomposing complex psychosocial and behavioral interventions. Psychol Aging [Internet]. 2003;18(3):385–95.
Ettema RG a, Hoogendoorn ME, Kalkman CJ, Schuurmans MJ. Development of a nursing intervention to prepare frail older patients for cardiac surgery (the PREDOCS programme), following phase one of the guidelines of the Medical Research Council. Eur J Cardiovasc Nurs [Internet]. 2014;13(6):494–505.
Carnes D, Homer K, Underwood M, Pincus T, Rahman A, Taylor SJC. Pain management for chronic musculoskeletal conditions: the development of an evidence-based and theory-informed pain self-management course. BMJ Open [Internet]. 2013;3(11):e003534.
Clyne B, Bradley MC, Hughes CM, Clear D, McDonnell R, Williams D, et al. Addressing potentially inappropriate prescribing in older patients: development and pilot study of an intervention in primary care (the OPTI-SCRIPT study). BMC Health Serv Res [Internet]. 2013;13(1):307.
Gray CM, Hunt K, Mutrie N, Anderson AS, Leishman J, Dalgarno L, et al. Football Fans in Training: the development and optimization of an intervention delivered through professional sports clubs to help men lose weight, become more active and adopt healthier eating habits. BMC Public Health [Internet]. 2013;13(1):232.
Lewis C, Roberts N, Vick T, Bisson JI. Development of a guided self-help (GSH) program for the treatment of mild-to-moderate posttraumatic stress disorder (PTSD). Depress Anxiety [Internet]. 2013;30(11):1121–8.
Palmer J, Bozas G, Stephens A, Johnson M, Avery G, O Toole L, et al. Developing a complex intervention for the outpatient management of incidentally diagnosed pulmonary embolism in cancer patients. BMC Health Serv Res [Internet]. 2013;13(1):235.
Barley EA, Haddad M, Simmonds R, Fortune Z, Walters P, Murray J, et al. The UPBEAT depression and coronary heart disease programme: using the UK Medical Research Council framework to design a nurse-led complex intervention for use in primary care. BMC Fam Pract [Internet]. 2012;13:119.
Kirkevold M, Bronken AB, Martinsen R, Kvigne K. Promoting psychosocial well-being following a stroke: Developing a theoretically and empirically sound complex intervention. Int J Nurs Stud [Internet]. 2012;49(4):386–97.
Smith SM, Murchie P, Devereux G, Johnston M, Lee AJ, Macleod U, et al. Developing a complex intervention to reduce time to presentation with symptoms of lung cancer. Br J Gen Pract J R Coll Gen Pract [Internet]. 2012;62(602):e605–15.
Burr JM, Campbell MK, Campbell SE, Francis JJ, Greene A, Hernández R, et al. Developing the clinical components of a complex intervention for a glaucoma screening trial: a mixed methods study. BMC Med Res Methodol [Internet]. BioMed Central Ltd; 2011; 11(1):54.
Farquhar MC, Ewing G, Booth S. Using mixed methods to develop and evaluate complex interventions in palliative care research. Palliat Med [Internet]. 2011;25(8):748–57.
Lovell K, Bower P, Richards D, Barkham M, Sibbald B, Roberts C, et al. Developing guided self-help for depression using the Medical Research Council complex interventions framework: a description of the modelling phase and results of an exploratory randomised controlled trial. BMC Psychiatry [Internet]. 2008;8:91.
Redfern J, Rudd AD, Wolfe CD, McKevitt C. Stop stroke: development of an innovative intervention to improve risk factor management after stroke. Patient Educ Couns [Internet]. 2008;72(2):201–9.
Murchie P, Hannaford PC, Wyke S, Nicolson MC, Campbell NC. Designing an integrated follow-up programme for people treated for cutaneous malignant melanoma: a practical application of the MRC framework for the design and evaluation of complex interventions to improve health. Fam Pract [Internet]. 2007;24(3):283–92.
Sturt J, Whitlock S, Hearnshaw H. Complex intervention development for diabetes self-management. J Adv Nurs [Internet]. 2006;54(3):293–303.
Robinson L, Francis J, James P, Tindle N, Greenwell K, Rodgers H. Caring for carers of people with stroke: developing a complex intervention following the Medical Research Council framework. Clin Rehabil [Internet]. 2005;19(5):560–71.
Byng R, Jones R. Mental Health Link: the development and formative evaluation of a complex intervention to improve shared care for patients with long-term mental illness. J Eval Clin Pract [Internet]. 2004;10(1):27–36.
Bartholomew LK, Parcel GS, Kok G. Intervention mapping: a process for developing theory and evidence-based health education programs. Heal Educ Behav [Internet]. 1998;25(5):545–63.
Munir F, Kalawsky K, Wallis DJ, Donaldson-Feilder E. Using intervention mapping to develop a work-related guidance tool for those affected by cancer. BMC Public Health [Internet]. 2013;13:6.
Vonk Noordegraaf A, Huirne JAF, Pittens CA, van Mechelen W, Broerse JEW, Brölmann HAM, et al. eHealth program to empower patients in returning to normal activities and work after gynecological surgery: intervention mapping as a useful method for development. J Med Internet Res [Internet]. 2012;14(5):e124.
Collins LM, Baker TB, Mermelstein RJ, Piper ME, Jorenby DE, Smith SS, et al. The multiphase optimization strategy for engineering effective tobacco use interventions. Ann Behav Med [Internet]. 2011;41(2):208–26.
Treweek S, Bonetti D, Maclennan G, Barnett K, Eccles MP, Jones C, et al. Paper-based and web-based intervention modeling experiments identified the same predictors of general practitioners’ antibiotic-prescribing behavior. J Clin Epidemiol [Internet]. Elsevier Inc. 2014;67(3):296–304.
Hrisos S, Eccles M, Johnston M, Francis J, Kaner EFS, Steen N, et al. An intervention modelling experiment to change GPs’ intentions to implement evidence-based practice: using theory-based interventions to promote GP management of upper respiratory tract infection without prescribing antibiotics #2. BMC Health Serv Res [Internet]. 2008;8:10.
Whittaker R, Merry S, Dorey E, Maddison R. A development and evaluation process for mHealth interventions: examples from New Zealand. J Health Commun [Internet]. 2012;17 Suppl 1:11–21.
Grant AM, Guthrie B, Dreischulte T. Developing a complex intervention to improve prescribing safety in primary care: mixed methods feasibility and optimisation pilot study. BMJ Open [Internet]. 2014;4(1):e004153.
Schaefer DR, Adams J, Haas SA. Social networks and smoking: exploring the effects of peer influence and smoker popularity through simulations. Health Educ Behav [Internet]. 2013;40(1 Suppl):24S–32.
Eldridge S, Spencer A, Cryer C, Parsons S, Underwood M, Feder G. Why modelling a complex intervention is an important precursor to trial design: lessons from studying an intervention to reduce falls-related injuries in older people. J Health Serv Res Policy [Internet]. 2005;10(3):133–42.
Treweek S, Sullivan F. How much does pre-trial testing influence complex intervention trials and would more testing make any difference? An email survey. BMC Med Res Methodol [Internet]. 2006;6:28.
Booth S, Farquhar M, Gysels M, Bausewein C, Higginson IJ. The impact of a breathlessness intervention service (BIS) on the lives of patients with intractable dyspnea: A qualitative phase 1 study. Palliat Support Care. 2006;4:287–93.
The authors declare that they have no competing interests.
SL, ND and SH were involved in the development of the study. SL and RF were responsible for the data collection, overseen by ND and SH. Data analysis was undertaken by SL guided by ND and SH. The manuscript was prepared by SL, PC, MW and CD. All authors contributed to the final version of the manuscript. SL is the principal investigator and responsible for this study. All authors read and approved the final manuscript.
About this article
Cite this article
Levati, S., Campbell, P., Frost, R. et al. Optimisation of complex health interventions prior to a randomised controlled trial: a scoping review of strategies used. Pilot Feasibility Stud 2, 17 (2016). https://doi.org/10.1186/s40814-016-0058-y