‘We know what we have to do, but we don't know how to do it’ has been a recurring comment among global health actors for a long time. In 2010, for example, the United Nations affirmed that ‘we know what works’1 in taking care of the health of women and children. The WHO Commission on the Social Determinants of Health (2008) has highlighted effective interventions to improve the health of populations and to establish health equity.2 However, while the content of interventions, which are theoretically effective, are relatively well known, their level of coverage is weak.3 Furthermore, the conditions of their implementation are less understood. An old meta-analysis shows that the potential effectiveness of interventions is reduced by 50% because of multiple contextual factors which act against the implementation.4 Therefore, it is not enough to know if a health intervention is effective; it is also necessary to understand why the intervention works, how, for whom and in which contexts. It is here where implementation science is an undeniable aid.
In this editorial, the focus will not be on the controversies concerning the definition of implementation science or the academic arguments made in order to appropriate or better sell the training of implementation science. Essentially, what is of interest is to call on the community of students, researchers, implementers and donors to commit themselves to further and a better quality research in order to have a greater understanding of how to implement health interventions. To quote Joseph Durlak, an important author in this field, ‘studying programme implementation is not easy but it is essential’.5 Implementation is comprised of one or several processes organised in a particular context so as to bring about the desired changes of an intervention (whether policy, programme or project) through the means necessary to deploy it. Implementation science is about mobilising theories, concepts and methods to better ‘understand what, why, and how interventions work in ‘real world’ settings’.6 There is a movement away from implementation research, which is centred on analysing the way interventions consider evidence, a field that is close to knowledge transfer, which is also not fully developed in low and middle income countries (LMICs).7 However, we do agree with the fact that ‘research in both fields deals with the challenges of translating intentions into desired changes’.8 A meeting of major journal editors has been organized by the Canadian Institutes of Health Research Institute of Population and Public Health and the Canadian Journal of Public Health in April 2016 (http://sparkingsolutions.ca). They will soon launch the Ottawa Statement to promote publications in the field of population health intervention research.
BMJ Global Health wishes to participate in the development of implementation science but with a focus on equity and on a better adaptation and/or creation of theoretical, conceptual and methodological approaches in the context of LMICs. In fact, a review of writings (1933–2003) concerning research on the implementation of public policies shows that only 4% concerned Africa, 2% Latin America and 15% was on health.9 The author of this review clearly highlighted ‘the ethnocentric bias in implementation studies’.9 This observation was confirmed in another analysis (1986–2006) of research in public policies in the field of health promotion: ‘all the most authoritative conceptualizations mentioned here were modelled on Western-style democratic governance systems’.10 Two rapid bibliographic searches using Pubmed database show an important increase of papers about global health and implementation since 1970, but implementation still concern just around 5% (figure 1). Therefore, there is an urgency to act, since both analyses confirm that we are far from having a body of theories, frameworks and approaches which is sufficient for the in-depth study of the implementation of interventions;9 ,11 ,12 interventions which still need to be largely ‘tested and operationalised in real-world settings’.13
This urgency should not be taken lightly. It is important that implementation studies in LMICs adhere to what is often called the third generation, which uses more rigorous research design.14 While global health actors appear to have discovered implementation science recently,15 it was actually mobilised at least more than 30 years ago by political science researchers.9 It is not necessary in this editorial to discuss the history of the analysis of intervention implementation8 ,9 ,11 which should be better understood (and thus better taught16) and used by those who study implementation. However, research on global health implementation should better exploit the theories, conceptual frameworks and approaches of the social sciences.12 As essential as the inductive and empirical approaches may be, recourse or contribution to theoretical and conceptual development is as important, if not even more.17 As a reminder, among the health promotion studies which analysed public policy, only 18% made reference to a theoretical framework.18 Studies showed that Kingdon's stream theory19 along with Lipsky's street-level bureaucrats20 and Rogers’ innovation theory,21 which were all developed in the USA, could also have been adapted in the context of certain LMICs in order to better understand implementation. Additionally, the role of ideas in the implementation (rather than the emergence) of interventions in LMICs have not really been tackled,22 whereas the writings on high-income countries in this regard have been abundant.23
Implementation science clearly cannot be developed alone without considering the effects of interventions (or the principles of effectiveness, to borrow from Patton's words24), because it risks falling into ‘type 3 error’, where an intervention is evaluated even though it has not yet been implemented as anticipated.25 Thus it is important to better describe the content of interventions using available tools26 ,27 and to report on the (classic) fidelity and intensity of implementation. It is also necessary to update the adjustments of interventions—inevitable in a natural context—and the fidelity of the theory.28 ,29 It is in this contextual and holistic research approach that we should conduct and report implementation science. The recurrence of several types of interventions in many settings, the links between the effects (expected or not), the processes, the actors and their context should be brought to light, notably with the theoretical approach of critical realism, which is still not well-tested in LMICs.30 Critical theory could be used to better understand power relations. Individual relations and the role that context plays in the matter should be studied to understand, for example, the heterogeneity of effects in the implementation of interventions.31 When not forgotten, contextual analysis specifically associated with interventions is often the poor parent.32
The use and presentation of methods should also be considered in developing this field by making it more credible to those who think that qualitative research or the flexible approach to research design,33 ,34 which are integral parts of implementation science, are not rigorous enough. For this, it is necessary that authors are more precise in the description of their methods of qualitative sampling and of their analysis procedures. Beyond these interdisciplinary, boundary-spanning approaches, which are necessary in global health,35 ,36 the recourse to mixed methods and to multiple case studies, if longitudinally possible, would be a major benefit for implementation science.14
It is certainly time to seriously consider the charge which was made almost 10 years ago on policy research in LMICs: ‘more work on implementation, and specifically, the challenges of implementing equity-oriented policies, as well as more examination of successful policy change experiences’.37 We invite authors to participate in this by proposing and submitting implementation science articles to BMJ Global Health.