18
GUIDELINE Open Access RAMESES II reporting standards for realist evaluations Geoff Wong 1* , Gill Westhorp 2 , Ana Manzano 3 , Joanne Greenhalgh 3 , Justin Jagosh 4 and Trish Greenhalgh 1 Abstract Background: Realist evaluation is increasingly used in health services and other fields of research and evaluation. No previous standards exist for reporting realist evaluations. This standard was developed as part of the RAMESES II project. The projects aim is to produce initial reporting standards for realist evaluations. Methods: We purposively recruited a maximum variety sample of an international group of experts in realist evaluation to our online Delphi panel. Panel members came from a variety of disciplines, sectors and policy fields. We prepared the briefing materials for our Delphi panel by summarising the most recent literature on realist evaluations to identify how and why rigour had been demonstrated and where gaps in expertise and rigour were evident. We also drew on our collective experience as realist evaluators, in training and supporting realist evaluations, and on the RAMESES email list to help us develop the briefing materials. Through discussion within the project team, we developed a list of issues related to quality that needed to be addressed when carrying out realist evaluations. These were then shared with the panel members and their feedback was sought. Once the panel members had provided their feedback on our briefing materials, we constructed a set of items for potential inclusion in the reporting standards and circulated these online to panel members. Panel members were asked to rank each potential item twice on a 7-point Likert scale, once for relevance and once for validity. They were also encouraged to provide free text comments. Results: We recruited 35 panel members from 27 organisations across six countries from nine different disciplines. Within three rounds our Delphi panel was able to reach consensus on 20 items that should be included in the reporting standards for realist evaluations. The overall response rates for all items for rounds 1, 2 and 3 were 94 %, 76 % and 80 %, respectively. Conclusion: These reporting standards for realist evaluations have been developed by drawing on a range of sources. We hope that these standards will lead to greater consistency and rigour of reporting and make realist evaluation reports more accessible, usable and helpful to different stakeholders. Keywords: Realist evaluation, Delphi approach, Reporting guidelines Background Realist evaluation is a form of theory-driven evaluation, based on a realist philosophy of science [1, 2] that addresses the questions, what works, for whom, under what circumstances, and how. The increased use of real- ist evaluation in the assessment of complex interven- tions [3] is due to the realisation by many evaluators and commissioners that coming up with solutions to complex problems is challenging and requires deeper insights into the nature of programmes and implementa- tion contexts. These problems have multiple causes op- erating at both individual and societal levels, and the interventions or programmes designed to tackle such problems are themselves complex. They often have mul- tiple, interconnected components delivered individually or targeted at communities or populations, with success dependent both on individualsresponses and on the wider context. What works for one family, or organisa- tion, or city may not work in another. Effective complex interventions or programmes are difficult to design and * Correspondence: [email protected] 1 Nuffield Department of Primary Care Health Sciences, University of Oxford, Radcliffe Primary Care Building, Radcliffe Observatory Quarter, Woodstock Road, Oxford OX2 6GG, UK Full list of author information is available at the end of the article © 2016 The Author(s). Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated. Wong et al. BMC Medicine (2016) 14:96 DOI 10.1186/s12916-016-0643-1

RAMESES II reporting standards for realist evaluations

  • Upload
    others

  • View
    2

  • Download
    0

Embed Size (px)

Citation preview

Page 1: RAMESES II reporting standards for realist evaluations

GUIDELINE Open Access

RAMESES II reporting standards for realistevaluationsGeoff Wong1*, Gill Westhorp2, Ana Manzano3, Joanne Greenhalgh3, Justin Jagosh4 and Trish Greenhalgh1

Abstract

Background: Realist evaluation is increasingly used in health services and other fields of research and evaluation.No previous standards exist for reporting realist evaluations. This standard was developed as part of the RAMESES IIproject. The project’s aim is to produce initial reporting standards for realist evaluations.

Methods: We purposively recruited a maximum variety sample of an international group of experts in realistevaluation to our online Delphi panel. Panel members came from a variety of disciplines, sectors and policy fields.We prepared the briefing materials for our Delphi panel by summarising the most recent literature on realistevaluations to identify how and why rigour had been demonstrated and where gaps in expertise and rigour wereevident. We also drew on our collective experience as realist evaluators, in training and supporting realist evaluations,and on the RAMESES email list to help us develop the briefing materials.Through discussion within the project team, we developed a list of issues related to quality that needed to beaddressed when carrying out realist evaluations. These were then shared with the panel members and their feedbackwas sought. Once the panel members had provided their feedback on our briefing materials, we constructed a set ofitems for potential inclusion in the reporting standards and circulated these online to panel members. Panel memberswere asked to rank each potential item twice on a 7-point Likert scale, once for relevance and once for validity. Theywere also encouraged to provide free text comments.

Results: We recruited 35 panel members from 27 organisations across six countries from nine differentdisciplines. Within three rounds our Delphi panel was able to reach consensus on 20 items that should beincluded in the reporting standards for realist evaluations. The overall response rates for all items for rounds1, 2 and 3 were 94 %, 76 % and 80 %, respectively.

Conclusion: These reporting standards for realist evaluations have been developed by drawing on a range ofsources. We hope that these standards will lead to greater consistency and rigour of reporting and makerealist evaluation reports more accessible, usable and helpful to different stakeholders.

Keywords: Realist evaluation, Delphi approach, Reporting guidelines

BackgroundRealist evaluation is a form of theory-driven evaluation,based on a realist philosophy of science [1, 2] thataddresses the questions, ‘what works, for whom, underwhat circumstances, and how’. The increased use of real-ist evaluation in the assessment of complex interven-tions [3] is due to the realisation by many evaluators andcommissioners that coming up with solutions to

complex problems is challenging and requires deeperinsights into the nature of programmes and implementa-tion contexts. These problems have multiple causes op-erating at both individual and societal levels, and theinterventions or programmes designed to tackle suchproblems are themselves complex. They often have mul-tiple, interconnected components delivered individuallyor targeted at communities or populations, with successdependent both on individuals’ responses and on thewider context. What works for one family, or organisa-tion, or city may not work in another. Effective complexinterventions or programmes are difficult to design and

* Correspondence: [email protected] Department of Primary Care Health Sciences, University of Oxford,Radcliffe Primary Care Building, Radcliffe Observatory Quarter, WoodstockRoad, Oxford OX2 6GG, UKFull list of author information is available at the end of the article

© 2016 The Author(s). Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, andreproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link tothe Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver(http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.

Wong et al. BMC Medicine (2016) 14:96 DOI 10.1186/s12916-016-0643-1

Page 2: RAMESES II reporting standards for realist evaluations

evaluate. Effective evaluations need to be able to con-sider how, why, for whom, to what extent, and in whatcontext complex interventions work. Realist evaluationscan address these challenges and have indeed addressednumerous topics of central relevance in health servicesresearch [4–6] and other fields [7, 8].

What is realist evaluation?The methodology of realist evaluation was originally de-veloped in the 1990’s by Pawson and Tilley to addressthe question ‘what works, for whom, in what circum-stances, and how?’ in a broad range of interventions [1].Realist evaluations, in contrast to other forms of theorydriven evaluations [9], must be underpinned by a realistphilosophy of science. In realist evaluations it is assumedthat social systems and structures are ‘real’ (because theyhave real effects) and also that human actors responddifferently to interventions in different circumstances.To understand how an intervention might generatedifferent outcomes in different circumstances, a realistevaluation examines how different programme mecha-nisms, namely underlying changes in the reasoning andbehaviour of participants, are triggered in particular con-texts. Thus, programmes are believed to ‘work’ in differ-ent ways for different people in different situations [10]:

� Social programmes (or interventions) attempt tocreate change by offering (or taking away) resourcesto participants or by changing contexts withinwhich decisions are made (for example, changinglaws or regulations);

� Programmes ‘work’ by enabling or motivatingparticipants to make different choices;

� Making and sustaining different choices requires achange in a participant’s reasoning and/or theresources available to them;

� The contexts in which programmes operate make adifference to and thus shape the mechanismsthrough which they work and thus the outcomesthey achieve;

� Some factors in the context may enable particularmechanisms to operate or prevent them fromoperating;

� There is always an interaction between contextand mechanism, and that interaction is whatcreates the programme’s impacts or outcomes(Context +Mechanism = Outcome);

� Since programmes work differently in differentcontexts and through different mechanisms,programmes cannot simply be replicated from onecontext to another and automatically achieve thesame outcomes. Theory-based understandings about‘what works, for whom, in what contexts, and how’are, however, transferable;

� One of the tasks of evaluation is to learn moreabout ‘what works for whom’, ‘in which contextsparticular programmes do and don’t work’ and ‘whatmechanisms are triggered by what programmes inwhat contexts’.

In a realist evaluation the assumption is that pro-grammes are ‘theories incarnate’ [1]. That is, whenever aprogramme is designed and implemented, it is under-pinned by one or more theories about what ‘might causechange’, even though that theory or theories may not beexplicit. Undertaking a realist evaluation requires thatthe theories within a programme are made explicit, bydeveloping clear hypotheses about how, and for whom,to what extent, and in what contexts a programmemight ‘work’. The evaluation of the programme testsand refines those hypotheses. The data collected in arealist evaluation needs to enable such testing and soshould include collecting data about: programme im-pacts and the processes of programme implementation,the specific aspects of programme context that mightimpact on programme outcomes, and how these con-texts shape the specific mechanisms that might be creat-ing change. This understanding of how a particularaspects of the context shapes the mechanism whichleads to outcomes can be expressed as a context-mechanism-outcome (CMO) configuration and is theanalytical unit on which realist evaluation is built. Elicit-ing, refining and testing CMO configurations allows adeeper and more detailed understanding of for whom, inwhat circumstances and why the programme works. Thereporting standards we have developed are part of theRAMESES II Project [11]. In addition, we are developingtraining materials to support realist evaluators and thesewill be made freely available on: www.ramesesproject.org.A realist approach has particular implications for the

design of an evaluation and the roles of participants. Forexample, rather than comparing the outcomes for partic-ipants who have and have not taken part in aprogramme (as is done in a randomised controlled orquasi-experimental designs), a realist evaluation com-pares CMO configurations within programmes or acrosssites of implementation. It asks, for example, whether aprogramme works more or less well in different localities(and if so, how and why), or for different participants.These participants may be programme designers, imple-menters and recipients.When seeking input from participants, it is assumed

that different participants have different perspectives, in-formation and understandings about how programmesare supposed to work and whether they in fact do. Assuch, data collection processes (interviews, focus groups,questionnaires and so on) need to be constructed so thatthey are able to identify the particular information that

Wong et al. BMC Medicine (2016) 14:96 Page 2 of 18

Page 3: RAMESES II reporting standards for realist evaluations

those stakeholder groups will have in a realist way[12, 13]. These data may then be used to confirm, re-fute or refine theories about the programme. More indepth discussions about the philosophical underpin-nings of realist evaluation and its origins may befound in the training materials we are developing andother sources [1, 2, 11].

Why are reporting standards needed?Within health services research, reporting standardsare common and increasingly expected (for example,[14–16]). Within realist research, reporting standardshave already been developed for realist syntheses (alsoknown as reviews) [17, 18]. The rationale behind ourreporting standards is that they will guide evaluatorsabout what needs to be reported in a realist evalu-ation. This should serve two purposes. It will helpreaders to understand the programme under evalu-ation itself and the findings from the evaluation andit will provide readers with relevant and necessary in-formation to enable them to assess the quality andrigour of the evaluation.Such reporting guidance is important for realist evalu-

ations. Since its development almost two decades ago,realist evaluation has gained increasing interest and ap-plication in many fields of research, including health ser-vices research. Published literature [19, 20], postings inthe RAMESES email list we have run since 2011 [21],and our experience as trainers and mentors in realistmethodologies all suggest that there is confusion andmisunderstanding among evaluators, researchers, journaleditors, peer reviewers and funders about what countsas a high quality realist evaluation and what, conversely,as substandard. Even though experts still differ on de-tailed conceptual methodological issues, the increasingpopularity of realist evaluation has prompted us to de-velop baseline reporting (and later) quality standardswhich, we anticipate, will advance the application ofrealist theory and methodology. We anticipate that boththe reporting standards and the quality standards willthemselves evolve over time, as the theory and method-ology evolve.The present study aims to produce initial reporting

standards for realist evaluations.

MethodsThe methods we used to develop these reporting stan-dards have been published elsewhere [11]. In summary,we purposively recruited an international group of ex-perts to our online Delphi panel. We aimed to achievemaximum variety and sought out panel members whouse realist evaluation in a variety of disciplines, sectorsand policy fields. To prepare the briefing materials forour Delphi panel, we collated and summarised the most

recent literature on realist evaluations, seeking to iden-tify how and why rigour had been demonstrated andwhere gaps in expertise and rigour were evident. Wealso drew on our collective experience as realist evalua-tors, in training and supporting realist evaluations andon the RAMESES email list to help us develop the brief-ing materials.Through discussion within the project team, we con-

sidered the findings of our review of recent examples ofrealist evaluations and developed a list of issues relatedto quality that needed to be addressed when carryingout realist evaluations. These were then shared with thepanel members and additional issues were sought. Oncethe panel members had provided their feedback on ourbriefing materials we constructed a set of items forpotential inclusion in the reporting standards and circu-lated these online to panel members. Panel memberswere asked to rank each potential item twice on a 7-point Likert scale (1 = strongly disagree to 7 = stronglyagree), once for relevance (i.e. should an item on thistheme/topic be included at all in the standards?) andonce for validity (i.e. to what extent do you agreewith this item as currently worded?). They were alsoencouraged to provide free text comments. We ranthe Delphi panel over three rounds between June2015 and January 2016.

Description of panel and itemsWe recruited 35 panel members from 27 organisationsacross six countries. They comprised of evaluators ofhealth services (23), public policy (9), nursing (6), crim-inal justice (6), international development (2), contractevaluators (3), policy and decision makers (2), funders ofevaluations (2) and publishing (2) (note that some indi-viduals had more than one role).In round 1 of the Delphi panel, 33 members provided

suggestions for items that should be included in thereporting standards and/or comments on the nature ofthe standards themselves. In rounds 2 and 3, panelmembers ranked the items for relevance and validity.For round 2, the panel was presented with 22 items torank. The overall response rate across all items for thisround was 76 %. Based on the rankings and free textcomments our analysis indicated that two items neededto be merged and one item removed. Minor revisionswere made to the text of the other items based on therankings and free text comments. After discussionwithin the project team we judged that only one item(the newly created merged item) needed to be returnedto round 3 of the Delphi panel. The response rate forround 3 was 80 %. Consensus was reached within threerounds on both the content and wording of a 20 itemreporting standard. Table 1 provides an overview ofthese items. Those using the list of Items in Table 1 to

Wong et al. BMC Medicine (2016) 14:96 Page 3 of 18

Page 4: RAMESES II reporting standards for realist evaluations

Table 1 List of items to be included when reporting realist evaluations

TITLE Reported indocumentY/N/Unclear

Page(s) indocument

1 In the title, identify the document as a realist evaluation

SUMMARY OR ABSTRACT

2 Journal articles will usually require an abstract, while reports andother forms of publication will usually benefit from a shortsummary. The abstract or summary should include brief detailson: the policy, programme or initiative under evaluation;programme setting; purpose of the evaluation; evaluationquestion(s) and/or objective(s); evaluation strategy; datacollection, documentation and analysis methods; key findingsand conclusionsWhere journals require it and the nature of the study isappropriate, brief details of respondents to the evaluation andrecruitment and sampling processes may also be includedSufficient detail should be provided to identify that a realistapproach was used and that realist programme theory wasdeveloped and/or refined

INTRODUCTION

3 Rationale for evaluation Explain the purpose of the evaluation and the implications forits focus and design

4 Programme theory Describe the initial programme theory (or theories) thatunderpin the programme, policy or initiative

5 Evaluation questions, objectives and focus State the evaluation question(s) and specify the objectives forthe evaluation. Describe whether and how the programmetheory was used to define the scope and focus of the evaluation

6 Ethical approval State whether the realist evaluation required and has gainedethical approval from the relevant authorities, providing detailsas appropriate. If ethical approval was deemed unnecessary,explain why

METHODS

7 Rationale for using realist evaluation Explain why a realist evaluation approach was chosen and(if relevant) adapted

8 Environment surrounding the evaluation Describe the environment in which the evaluation took place

9 Describe the programme policy, initiative orproduct evaluated

Provide relevant details on the programme, policy or initiativeevaluated

10 Describe and justify the evaluation design A description and justification of the evaluation design (i.e. theaccount of what was planned, done and why) should beincluded, at least in summary form or as an appendix, in thedocument which presents the main findings. If this is not done,the omission should be justified and a reference or link to theevaluation design given. It may also be useful to publish ormake freely available (e.g. online on a website) any originalevaluation design document or protocol, where they exist

11 Data collection methods Describe and justify the data collection methods – which oneswere used, why and how they fed into developing, supporting,refuting or refining programme theoryProvide details of the steps taken to enhance thetrustworthiness of data collection and documentation

12 Recruitment process and sampling strategy Describe how respondents to the evaluation were recruited orengaged and how the sample contributed to the development,support, refutation or refinement of programme theory

13 Data analysis Describe in detail how data were analysed. This section shouldinclude information on the constructs that were identified, theprocess of analysis, how the programme theory was furtherdeveloped, supported, refuted and refined, and (where relevant)how analysis changed as the evaluation unfolded

Wong et al. BMC Medicine (2016) 14:96 Page 4 of 18

Page 5: RAMESES II reporting standards for realist evaluations

help guide the reporting of their realist evaluations mayfind the last two columns (‘Reported in document’ and‘Page(s) in document’) as a useful way to indicate toothers where in the document each item has beenreported.

Scope of the reporting standardsThese reporting standards are intended to help evalua-tors, researchers, authors, journal editors, and policy-and decision-makers to know and understand whatshould be reported when writing up a realist evaluation.They are not intended to provide detailed guidance onhow to conduct a realist evaluation; for this, we wouldsuggest that interested readers access summary arti-cles or publications on methods [1, 19, 20, 22, 23].These reporting standards apply only to realist evalu-ation. A list of publication or reporting guidelines forother evaluation methods can be found on theEQUATOR Network’s website [24], but at presentnone of these relate specifically to realist evaluations.As part of the RAMESES II project we are also devel-oping quality standards which will be available as aseparate publication and training materials for realistevaluations [11].

How to use these reporting standardsThe layout of this document is based on the RAMESESpublication standards: realist syntheses [17, 18], whichitself was based on previous methodological publications(in particular, on the ‘Explanations and Elaborations’document of the PRISMA statement [25]. After eachitem there is an exemplar drawn from publically avail-able evaluations followed by a rationale for its inclusion.Within these standards, we have drawn our exemplartexts mainly from realist evaluations that have been pub-lished in peer review journals, as these were easy to ac-cess and publically available. Our choice of exemplartexts should not be taken to imply that the standard ofreporting of realist evaluations that have not been pub-lished in peer review journals is in any way substandard.The exemplar text is provided to illustrate how an

item might be written up in a report. However, each ex-emplar has been extracted out of a larger document andso important contextual information has been omitted.It may thus be necessary to consult the original docu-ment from which the exemplar text was drawn to fullyunderstand the evaluation it refers to.What might be expected for each item has been set

out within these reporting standards, but authors will

Table 1 List of items to be included when reporting realist evaluations (Continued)

RESULTS

14 Details of participants Report (if applicable) who took part in the evaluation, thedetails of the data they provided and how the data was usedto develop, support, refute or refine programme theory

15 Main findings Present the key findings, linking them to contexts, mechanismsand outcome configurations. Show how they were used tofurther develop, test or refine the programme theory

DISCUSSION

16 Summary of findings Summarise the main findings with attention to the evaluationquestions, purpose of the evaluation, programme theory andintended audience

17 Strengths, limitations and future directions Discuss both the strengths of the evaluation and its limitations.These should include (but need not be limited to):(1) consideration of all the steps in the evaluation processes;and (2) comment on the adequacy, trustworthiness andvalue of the explanatory insights which emergedIn many evaluations, there will be an expectation to provideguidance on future directions for the programme, policy orinitiative, its implementation and/or design. The particularimplications arising from the realist nature of the findingsshould be reflected in these discussions

18 Comparison with existing literature Where appropriate, compare and contrast the evaluation’sfindings with the existing literature on similar programmes,policies or initiatives

19 Conclusion and recommendations List the main conclusions that are justified by the analyses ofthe data. If appropriate, offer recommendations consistentwith a realist approach

20 Funding and conflict of interest State the funding source (if any) for the evaluation, the roleplayed by the funder (if any) and any conflicts of interests ofthe evaluators

Wong et al. BMC Medicine (2016) 14:96 Page 5 of 18

Page 6: RAMESES II reporting standards for realist evaluations

still need to exercise judgement about how much infor-mation to include. The information reported should besufficient to enable readers to judge that a realist evalu-ation has been planned, executed, analysed and reportedin a coherent, trustworthy and plausible fashion, bothagainst the guidance set out within an item and for theoverall purposes of the evaluation itself.Evaluations are carried out for different purposes and

audiences. Realist evaluations in the academic literatureare usually framed as research, but many realist evalua-tions in the grey literature were commissioned as evalua-tions, not research. Hence, the items listed in Table 1should be interpreted flexibly depending on the purposeof the evaluation and the needs of the audience. Thismeans that not all evaluation reports need necessarily bereported in an identical way. It would be reasonable toexpect that the order in which items are reported mayvary and not all items will be required for every type ofevaluation report. As a general rule, if an item withinthese reporting standards has been excluded from thewrite-up of a realist evaluation, a justification should beprovided.

The RAMESES reporting standards for realist evaluationsItem 1: TitleIn the title, identify the document as a realist evaluation.

ExampleHow do you modernize a health service? A realist evalu-ation of whole-scale transformation in London [26].

ExplanationOur background searching has shown that some realistevaluations are not flagged as such in the title and mayalso be inconsistently indexed, and hence are more diffi-cult to locate. Realist evaluation is a specific theoreticaland methodological approach, and should be carefullydistinguished from evaluations that use different ap-proaches (e.g. such as other theory-based approaches)[27]. Researchers, policy staff, decision-makers and otherknowledge users may wish to be able to locate reportsusing realist approaches. Adding the term ‘realistevaluation’ as a keyword may also aid searching andidentification.

Item 2: Summary or abstractJournal articles will usually require an abstract while re-ports and other forms of publication will usually benefitfrom a short summary. The abstract or summary shouldinclude brief details on the policy, programme or initia-tive under evaluation; programme setting; purpose of theevaluation; evaluation question(s) and/or objective(s);evaluation strategy; data collection, documentation andanalysis methods; key findings and conclusions.

Where journals require it and the nature of the studyis appropriate, brief details of respondents to the evalu-ation and recruitment and sampling processes may alsobe included.Sufficient detail should be provided to identify that a

realist approach was used and that realist programmetheory was developed and/or refined.

ExampleThe current project conducted an evaluation of acommunity-based addiction program in Ontario, Canada,using a realist approach. Client-targeted focus groups andstaff questionnaires were conducted to develop preliminarytheories regarding how, for whom, and under what circum-stances the program helps or does not help clients. Individ-ual interviews were then conducted with clients andcaseworkers to refine these theories. Psychological mecha-nisms through which clients achieved their goals were re-lated to client needs, trust, cultural beliefs, willingness,self-awareness and self-efficacy. Client, staff and settingcharacteristics were found to affect the development ofmechanisms and outcomes [28].

ExplanationEvaluators will need to provide either a summary or ab-stract depending on the type of document they wish toproduce. Abstracts are often needed for formal academicpublications and summaries for project reports.Apart from the title, a summary or abstract is often

the only source of information accessible to searchersunless the full text document is obtained. Many busyknowledge users will often not have the time to read anentire evaluation report or publication in order to deter-mine its relevance and initially only access the summaryor abstract. The information in it must allow the readerto decide whether the evaluation is a realist evaluationand relevant to their needs.The brief summary we refer to here does not replace

the Executive Summary that many evaluation reportswill provide. In a realist evaluation, the ExecutiveSummary should include concise information aboutprogramme theory and CMO configurations, as well asthe other items described above.

Introduction sectionItem 3: Rationale for evaluationExplain the purpose of the evaluation and the implica-tions for its focus and design.

ExampleIn this paper, we aim to demonstrate how the realistevaluation approach helps in advancing complex sys-tems thinking in healthcare evaluation. We do this bycomparing the outcomes of cases which received a

Wong et al. BMC Medicine (2016) 14:96 Page 6 of 18

Page 7: RAMESES II reporting standards for realist evaluations

capacity-building intervention for health managersand explore how individual, institutional and context-ual factors interact and contribute to the observedoutcomes [29].

ExplanationRealist evaluations are used for many types of pro-grammes, projects, policies and initiatives (all referred tohere as ‘programmes’ or ‘evaluands’ – ‘that which isevaluated’ – for ease of reference). They are also con-ducted for multiple purposes (e.g. to develop or improvedesign and planning; understand whether and how newprogrammes work; improve effectiveness overall or forparticular populations; improve efficiency; inform deci-sions about scaling programmes out to other contexts; orto understand what is causing variations in implementa-tion or outcomes). The purpose has significant implica-tions for the focus of the work, the nature of questions,the design, and the choice of methods and analysis [30].These issues should be clearly described in the ‘back-ground’ section of a journal article or the introduction toan evaluation report. Where relevant, this should also de-scribe what is already known about the subject matter andthe ‘knowledge gaps’ that the evaluation sought to fill.

Item 4: Programme theoryDescribe the initial programme theory (or theories) thatunderpin the programme, policy or initiative.

ExampleThe programme theory (highlighting the underlying psy-chosocial mechanisms providing the active ingredients tofacilitate change), was:

– By facilitating critical discussion of ‘what works’(academic and practice-based evidence, in writtenand verbal form), across academic and field expertsand amongst peers, understanding of the practicalapplication and potential benefits of evidence usewould be increased, uptake of the evidence would befacilitated and changes to practice would follow.

A secondary programme theory was that:

– Allowing policy, practice and academic partners tocome together to consider their common interests,share the challenges and opportunities of working onpublic health issues, trusting relationship would beinitiated and these would be followed-up byfuture contact and collaborative work [31].

ExplanationRealist evaluations set out to develop, support, refute orrefine aspects of realist programme theory (or theories).

All programmes or initiatives will (implicitly or expli-citly) have a programme theory or theories [32] – ideasabout how the programme is expected to cause itsintended outcomes – and these should be articulatedhere. Initial programme theories may or may not berealist in nature. As an evaluation progresses, aprogramme theory that was not initially realist in naturewill need to be developed and refined so that it becomesa realist programme theory (that is, addressing all ofcontext, mechanism and outcome) [33].Programmes are theories incarnate. Within a realist

evaluation, programme theory can serve many functions.One of its functions is to describe and explain (some of)how and why, in the ‘real world’, a programme ‘works’,for whom, to what extent and in which contexts (the as-sumption that the programmes will only work in somecontexts and for some people, and that it may fire differ-ent mechanisms in different circumstances, thus gener-ating different outcomes, is one of the factors thatdistinguishes realist programme theory from other typesof programme theory). Other functions include focusingan evaluation, identifying questions, and determiningwhat types of data need to be collected and from whomand where, in order to best support, refute or refine theprogramme theory. The refined programme theory canthen serve many purposes for evaluation commissionersand end users [34]. It may support planning anddecision-making about programme refinement, scalingout and so on.Different processes can be used for ‘uncovering’ or

developing initial programme theory, including litera-ture review, programme documentation review, andinterviews and/or focus groups with key informants.The processes used to develop the programme theoryare usually different from those used later to refineit. The programme theory development processesneeds to be clearly reported for the sake of transpar-ency. The processes used for programme theory de-velopment may be reported here or in ‘Item 13: Dataanalysis’.A figure or diagram may aid in the description of the

programme theory. It may be presented early in the re-port or as an appendix.Sometimes, the focus of the evaluation will not be ‘the

programme itself ’ but a particular aspect of, or questionabout, a programme (for example, how the programmeaffects other aspects of the system in which it operates).The theory that is developed or used in any realist evalu-ation will be relevant to the particular questions underinvestigation.

Item 5: Evaluation questions, objectives and focusState the evaluation question(s) and specify the objec-tives for the evaluation. Describe whether and how the

Wong et al. BMC Medicine (2016) 14:96 Page 7 of 18

Page 8: RAMESES II reporting standards for realist evaluations

programme theory was used to define the scope andfocus of the evaluation.

ExampleThe objective of the study was to analyse the manage-ment approach at CRH [Central Regional Hospital]. Weformulated the following research questions: (1) What isthe management team's vision on its role? (2) Whichmanagement practices are being carried out? (3) What isthe organisational climate? … ; (4) What are the results?(5) What are the underlying mechanisms explaining theeffect of the management practices? [6].

ExplanationRealist evaluation questions contain some or all of theelements of ‘what works, how, why, for whom, to whatextent and in what circumstances, in what respect?’Specifically, realist evaluation questions need to reflectthe underlying purpose of realist evaluation – that is toexplain (how and why) rather than only describe out-come patterns.Note that the term ‘outcome’ will mean different

things in different kinds of evaluations. For example, itmay refer to ‘patterns of implementation’ in processevaluations; ‘patterns of efficiency or cost effectivenessfor different populations’ in economic evaluation; as wellas outcomes and impacts in the normal uses of the term.Moreover, realist evaluation questions may deal withother traditional evaluation questions of value, relevance,effectiveness, efficiency, and sustainability, as well asquestions of ‘outcomes’ [35]. However, they will applythe same kinds of question structure to those issues (forexample, effectiveness for whom, how and why; effi-ciency in what circumstances, how and why, and so on).Because a particular evaluation will never be able to

address all potential questions or issues, the scope of theevaluation has to be clarified. This may involve discus-sion and negotiation with (for example) commissionersof the evaluation, context experts, research funders and/or users. The processes used to establish purposes,scope, questions, and/or objectives should be described.Whether and how the programme theory was used indetermining the scope of the evaluation should beclearly articulated.In the real world, the programme being evaluated does

not sit in a vacuum. Instead, it is thrust into a messyworld of pre-existing programmes, a complex policy en-vironment, multiple stakeholders and so on [2, 36]. Allof these may have a bearing on (for example) the re-search questions, focus and constraints of the evaluation.Reporting should provide information to the readerabout the policy and other circumstances that may haveinfluenced the purposes, scope, questions and/or objec-tives of the evaluation.

Given the iterative nature of realist evaluation, ifthe purposes, scope, questions, objectives, programmetheory and/or protocol changed over the course ofthe evaluation, it should either be reported here or in‘Item 15: Main findings’.

Item 6: Ethical approvalState whether the realist evaluation required and hasgained ethical approval from the relevant authorities,providing details as appropriate. If ethical approval wasdeemed unnecessary, explain why.

ExampleThe study was reported to the Danish Data ProtectionAgency (2008-41-2322), the Ethics Committee of theCapital Region of Denmark (REC; reference number0903054, document number 230436) and Trials Regis-tration (ISTCTN54243636) and performed in accord-ance with the ethical recommendations of the HelsinkiDeclaration [37].

ExplanationRealist evaluation is a form of primary research and willusually involve human participants. It is important thatevaluations are conducted ethically. Evaluators comefrom a range of different professional backgrounds andwork in diverse fields. This means that different profes-sional ethical standards and local ethics regulatory re-quirements will apply. Evaluators should ensure thatthey aware of and comply with their professional obliga-tions and local ethics requirements throughout theevaluation.Specifically, one challenge that realist evaluations may

face is that legitimate changes may be required to themethods used and participants recruited as the evalu-ation evolves. Anticipating that such changes may beneeded is important when seeking ethical approval.Flexibility may need to be built into the project to allowfor updating ethics approvals and be explained to thosewho provide ethics approvals.

Methods sectionItem 7: Rationale for using realist evaluationExplain why a realist evaluation approach was chosenand (if relevant) adapted.

ExampleThis study used realist evaluation methodology… A cen-tral tenet of realist methodology is that programs workdifferently in different contexts–hence a community part-nership that achieves ‘success’ in one setting may ‘fail’ (oronly partially succeed) in another setting, because themechanisms needed for success are triggered to differentdegrees in different contexts. A second tenet is that for

Wong et al. BMC Medicine (2016) 14:96 Page 8 of 18

Page 9: RAMESES II reporting standards for realist evaluations

social programs, mechanisms are the cognitive or affectiveresponses of participants to resources offered [Reference].Thus the realist methodology is well suited to the studyof CBPR [Community-Based Participatory Research],which can be understood, from an ecological perspective,as a multiple intervention strategies implemented in di-verse community contexts [References] dependant on thedynamics of relationships among all stakeholders [38].

ExplanationRealist evaluation is firmly rooted in a realist philosophyof science. It places particular emphasis on under-standing causation (in this case, understanding howprogrammes generate outcomes) and how causalmechanisms are shaped and constrained by social,political, economic (and so on) contexts. This makesit particularly suitable for evaluations of certain topicsand questions – for example, complex social pro-grammes that involve human decisions and actions[1]. It also makes realist evaluation less suitable thanother evaluation approaches for certain topics andquestions – for example those which seek primarily to de-termine the average effect size of a simpler interventionadministered in a limited range of conditions [39]. The in-tent of this item in the reporting standard is not that thephilosophical principles should be described in everyevaluation but that the relevance of the approach to thetopic should be made explicit.Published evaluations demonstrate that some evalua-

tors have deliberately adapted or been ‘inspired’ by therealist evaluation approach as first described by Pawsonand Tilley. The description and rationale for any adapta-tions made or what aspects of the evaluations have been‘inspired’ by realist evaluation should be provided.Where evaluation approaches have been combined, thisshould be described and the implications for methodsmade explicit. Such information will allow criticism anddebate amongst users and evaluators on suitability ofthose adaptations for the particular purposes of theevaluation.

Item 8: Environment surrounding the evaluationDescribe the environment in which the evaluation tookplace.

ExampleThe … SMART [Self-Management Supported by Assistive,Rehabilitation and Telecare Technologies] Rehabilitationresearch programme(www.thesmartconsortium.org) [Refer-ence], began in 2003 to develop and test a prototype tele-rehabilitation device (The SMART RehabilitationTechnology System) for therapeutically prescribed strokerehabilitation for the upper limb. The aim was to enablethe user to adopt theories and principles underpinning

post-stroke rehabilitation and self-management. Thisincluded the development and initial testing of theSMART Rehabilitation Technology System (SMART 1)[References] and then from 2007, a Personalised Self-Management System for chronic disease (SMART 2)[References] [40].

ExplanationExplain and describe the environment in which theprogramme was evaluated. This may (for example) in-clude details about the locations, policy landscape, stake-holders, service configuration and availability, funding,time period and so on. Such information enables thereader to make sense of significant factors affecting theevaluand at differing levels (e.g. micro, meso and macro)and time points. This item may be reported here or earl-ier (e.g. in the Introduction section).This description does not substitute for ‘context’ in

realist explanation, which refers to the specific aspects ofcontext which affect how specific mechanisms operate(or do not). Rather, this overall description serves toorient the reader to the evaluand and the evaluation ingeneral.

Item 9: Describe the programme, policy, initiative orproduct evaluatedProvide relevant details on the programme, policy orinitiative evaluated.

ExampleIn a semirural locality in the North East of England, 14general practitioner (GP) practices covering a populationof 78,000 implemented an integrated care pathway (ICP)in order to improve palliative and end-of-life care in2009/2010. The ICP is still in place and is coordinatedby a multidisciplinary, multi-organisational steeringgroup with service user involvement. It is delivered in linewith national strategies on advance care planning (ACP)and end-of-life care [References] and aims to providehigh-quality care for all conditions regardless of diagno-sis. It requires each GP practice to develop an accurateelectronic register of patients through team discussionusing an agreed palliative care code, emphasising the im-portance of early identification and registration of thosewith any life-limiting illness. This enables registered pa-tients to have access to a range of interventions such asACP, anticipatory medication and the Liverpool CarePathway for the Dying Patient (LCP) [Reference]. In orderto register patients, health care professionals use the ‘sur-prise question’ which aims to identify patients ap-proaching the last year of their life [Reference]. Theregister is a tool for the management of primary care pa-tients; it does not involve family and patient notification.Practice teams then use a palliative rather than a

Wong et al. BMC Medicine (2016) 14:96 Page 9 of 18

Page 10: RAMESES II reporting standards for realist evaluations

curative ethos in consultation by determining patients’future wishes. Descriptive GP practice data analysis hadindicated that palliative care registrations had increasedsince the implementation of the ICP but more detailedanalysis was required [41].

ExplanationRealist evaluation may be used in a wide range of sectors(e.g. health, education, natural resource management,education, climate change), by a wide range of evaluators(academics, consultants, in-house evaluators, contentexperts) and on diverse evaluands (programmes, policies,initiatives and products), focusing on different stages oftheir implementation (development, pilot, process, out-come or impact) and considering outcomes of differenttypes (quality, effectiveness, efficiency, sustainability) fordifferent levels (whole systems, organisations, providers,end participants).It should not be assumed that the reader will be famil-

iar with the nature of the evaluand. The evaluand shouldbe adequately described: what does it consist of, whodoes it target, who provides it, over what geographicalreach, what is it supposed to achieve, and so on. Wherethe evaluation focuses on a specific aspect of the eva-luand (e.g. a particular point in its implementationchain, a particular hypothesised mechanism, a particulareffect on its environment), this should be identified –either here or in ‘Item 5: Evaluation questions, objec-tives and focus’ above.

Item 10: Describe and justify the evaluation designA description and justification of the evaluation design(i.e. the account of what was planned, done and why)should be included, at least in summary form or as anappendix, in the document which presents the mainfindings. If this is not done, the omission should be justi-fied and a reference or link to the evaluation designgiven. It may also be useful to publish or make freelyavailable (e.g. online on a website) any original evalu-ation design document or protocol, where they exist.

ExampleAn example of a detailed realist evaluation study proto-col is: Implementing health research through academicand clinical partnerships: a realistic evaluation of theCollaborations for Leadership in Applied Health Re-search and Care (CLAHRC) [42].

ExplanationThe design for a realist evaluation may differ signifi-cantly from some other evaluation approaches. Themethods, analytic techniques and so on will follow fromthe purpose, evaluation questions and focus. Further, asnoted in Item 5 above, the evaluation question(s), scope

and design may evolve over the course of the evaluation[19]. An accessible summary of what was planned in theprotocol or evaluation design, in what order, and why itis useful for interpreting the evaluation. Where an evalu-ation design changes over time, description of the natureof and rationale for the changes can aid transparency.Sometimes evaluations can involve a large number of

steps and processes. Providing a diagram or figure of theoverall structure of the evaluation may help to orient thereader [43].Significant factors affecting the design (e.g. time or

funding constraints) may also usefully be described.

Item 11: Data collection methodsDescribe and justify the data collection methods – whichones were used, why and how they fed into developing,supporting, refuting or refining programme theory.Provide details of the steps taken to enhance the trust-

worthiness of data collection and documentation.

ExampleAn important tenet of realist evaluation is making expli-cit the assumptions of the programme developers. Overtwo dozen local stakeholders attended three ‘hypothesisgeneration’ workshops before fieldwork started, wheremany CMO configurations were proposed. These werethen tested through data collection of observations, inter-views and documentation.Fifteen formal observations of all Delivering Choice ser-

vices occurred at two time points: 1) baseline (August2011) and 2) mid-study (Nov–Dec 2011). These consistedof a researcher sitting in on training sessions facilitatedby the End of Life Care facilitators with care home staffand shadowing Delivering Choice staff on shifts at theOut of Hours advice line, Discharge in Reach service andcoordination centres. Researchers also accompanied thepersonal care team on home visits on two occasions.Notes were taken while conducting formal observationsand these were typed up and fed into the analysis [44].

ExplanationBecause of the nature of realist evaluation, a broad rangeof data may be required and a range of methods may benecessary to collect them. Data will be required for all ofCMOs. Data collection methods should be adequate tocapture intended and unintended outcomes, and thecontext-mechanism interactions that generated them.Data about outcomes should be obtained. Where pos-sible, data about outcomes should be triangulated (atleast using different sources, if not different types, of in-formation) [20].Commonly, realist evaluations use more than one data

method to gather data. Administrative and monitoringdata for the programme or policy, existing data sets (e.g.

Wong et al. BMC Medicine (2016) 14:96 Page 10 of 18

Page 11: RAMESES II reporting standards for realist evaluations

census data, health systems data), field notes, photo-graphs, videos or sound recordings, as well as data col-lected specifically for the evaluation may all be required.The only constraints are that the data should allow ana-lysis of contexts, mechanisms and outcomes relevant tothe programme theory and to the purposes of and thequestions for the evaluation.Data collection tools and processes may need to be

adapted to suit realist evaluation. The specific tech-niques used or adaptations made to instruments or pro-cesses should be described in detail. Judgements canthen be made on whether the approaches chosen, instru-ments used and adaptations made are capable of captur-ing the necessary data, in formats that will be suitablefor realist analysis [45].For example, if interviews are used, the nature of the

data collected must change from only accessing respon-dents’ interpretations of events, or ‘meanings’ (as is oftendone in constructivist approaches) to identifying causalprocesses (i.e. mechanisms) or relevant elements ofcontext – which may or may not have anything to dowith respondents’ interpretations [13].Methods for collecting and documenting data (for

example, translation and transcription of qualitativedata; choices between video or oral recording; and thestructuring of quantitative data systems) are all theorydriven. The rationale for the methods used and their im-plications for data analysis should be explained.It is important that it is possible to judge whether the

processes used to collect and document the data used ina realist evaluation are rational and applied consistently.For example, a realist evaluation might report that alldata from interviews were audio taped and transcribedverbatim and numerical data were entered into a spread-sheet, or collected using particular software.

Item 12: Recruitment process and sampling strategyDescribe how respondents to the evaluation were re-cruited or engaged and how the sample contributed tothe development, support, refutation or refinement ofprogramme theory.

ExampleTo develop an understanding of how the ISP [InterpersonalSkills Profile] was actually used in practice, threegroups were approached for interviews, including prac-titioners from each field of practice. Practice educationfacilitators (PEFs) –whose role was to support allhealth care professionals involved with teaching andassessing students in the practical setting and whohad wide experience of supporting mentors – wereapproached to gain a broad overview. Educationchampions (ECs) – who were lecturers responsible forsupporting mentors and students in particular areas,

like a hospital or area in the community – providedan HEI perspective. Mentors were invited to contributeinterviews and students were invited to lend theirpractice assessment booklets so that copies of the ISPassessments could be taken. Seven PEFs, four ECs, 15mentors and 20 students contributed to the study (seetable). Most participants were from the Adult field ofnursing. The practice assessment booklets containedexamples of ISP use and comments by around 100mentors [46].

ExplanationSpecific kinds of information are required for realistevaluations – some of which comes from respondentsor key informants. Data are used to develop and re-fine theory about how, for whom, and in what cir-cumstances programmes generate their outcomes.This implies that any processes used to invite or re-cruit individuals need to find those who are able toprovide information about contexts, mechanisms and/or outcomes, and that the sample needs to be struc-tured appropriately to support, refute or refine theprogramme theory [23]. Describing the recruitmentprocess enables judgements to be made aboutwhether the process used is likely to recruit individ-uals who were likely to have the information neededto support, refute or refine the programme theory.

Item 13: Data analysisDescribe in detail how data were analysed. This sectionshould include information on the constructs that wereidentified, the process of analysis, how the programmetheory was further developed, supported, refuted andrefined, and (where relevant) how analysis changed asthe evaluation unfolded.

ExampleThe initial programme theory to be tested against ourdata and refined was therefore that if the student has atrusted assessor (external context) and a learning goalapproach (internal context), he or she will find thatgrades (external context) clarify (mechanism) and ener-gise (mechanism) his or her efforts to find strategies toimprove (outcome). …The transcripts were therefore examined for CMO

configurations:what effect (outcome) did the feedbackwith and without grades have (contexts)? What causedthese effects (mechanisms)? In what internal and ex-ternal learning environments (contexts) did theseoccur?The first two transcripts were analysed by all authors

to develop our joint understanding of what constitutes acontext, mechanism and outcome in formative WBA[workplace-based assessment]. A table was produced

Wong et al. BMC Medicine (2016) 14:96 Page 11 of 18

Page 12: RAMESES II reporting standards for realist evaluations

for each transcript listing the CMO configurationsidentified, with columns for student comments aboutfeedback with and without grades (the manipulatedvariable in the context). Subsequent transcripts werecoded separately by JL and AH, who compared theiranalyses. Where interpretation was difficult, one ortwo of the other researchers also analysed the tran-script to reach a consensus.The authors then compared CMO configurations con-

taining cognitive, self-regulatory and other explanationsof the effects of feedback with and without grades, seekingevidence to corroborate and refine the initial programmetheory. Where it was not corroborated, alternative expla-nations that referred to different mechanisms that mayhave been operating in different contexts were sought [7].

ExplanationIn a realist evaluation, the analysis process occurs itera-tively. Realist evaluation is usually multi-method ormixed-method. The strategies used to analyse the dataand integrate them should be explained. How these dataare then used to further develop, support, refute and re-fine programme theory should also be explained [13].For example, if interviews were used, how were the in-terviews analysed? If a survey was also conducted, howwas the survey analysed? In addition, how were thesetwo sets of data integrated? The data analyses may be se-quential or in parallel – i.e. one set of data may be ana-lysed first and then another or they might be analysed atthe same time.Specifically, at the centre of any realist analysis is the

application of a realist philosophical ‘lens’ to data. Arealist analysis of data seeks to analyse data using realistconcepts. Specifically, realism adheres to a generativeexplanation for causation – i.e. an outcome (O) ofinterest was generated by relevant mechanism(s) (M)which were triggered by, or could only operate in,context (C). Within or across the data sources theanalysis will need to identify recurrent patterns ofCMO configurations [47].During analysis, the data gathered is used to itera-

tively develop and refine any initial programme theory(or theories) into one or more realist programme the-ories for the whole programme. This purpose has im-plications for the type of data that needs to begathered – i.e. the data that needs to be gatheredmust be capable of being used for further programmetheory development. The analysis process requires (orresults in) inferences about whether a data item isfunctioning as a context, mechanism or outcome inthe particular analysis, but also about the relationshipsbetween the contexts, mechanisms and outcomes. Inother words the data gathered needs to contain infor-mation that [20, 48]:

(1)May be used to develop, support, refute or refine theassignment of the conceptual label of C, M or Owithin a CMO configuration – i.e. in this aspect ofthe analysis, this item of data is functioning ascontext within this CMO configuration.

(2)Enables evaluators to make inferences about therelationship of contexts, mechanisms and outcomeswithin a particular CMO configuration.

(3)Enables inferences to be made about (and latersupport, refute or refine) the relationships acrossCMO configurations – i.e. the location andinteractions between CMO configurations within aprogramme theory.

Ideally, a description should be provided if the dataanalysis processes evolved as the evaluation took shape.

Results sectionItem 14: Details of participantsReport (if applicable) who took part in the evaluation,the details of the data they provided and how thedata was used to develop, support, refute or refineprogramme theory.

ExampleWe conducted a total of 23 interviews with members ofthe DHMT [District Health Management Team] (8), dis-trict hospital management (4), and sub-district manage-ment (7); 4 managers were lost to staff transfers (2 fromthe DHMT and 2 at the sub-district level). At theregional level, we interviewed 3 out of 4 members ofthe LDP facilitation team, and one development part-ner supporting the LDP [Leadership DevelopmentProgramme]; 17 respondents were women and 6 weremen; 3 respondents were in their current posting lessthan 1 year, 13 between 1–3 years, and 7 between 3–5 years. More than half the respondents (12) had noprior formalised management training [4].

ExplanationOne important source of data in a realist evaluationcomes from participants (e.g. clients, patients, serviceproviders, policymakers and so on), who may haveprovided data that were used to inform any initialprogramme theory or to further develop, support, refuteor refine it later in the evaluation.Item 12 above covers the planned recruitment process

and sampling strategy. This item asks evaluators to re-port the results of recruitment process and samplingstrategy – who took part in the evaluation, what datadid they provide and how were these data used?Such information should go some way to ensure trans-

parency and enable judgements about the probativevalue of the data provided. It is important to report

Wong et al. BMC Medicine (2016) 14:96 Page 12 of 18

Page 13: RAMESES II reporting standards for realist evaluations

details of who (anonymised if necessary) provided whattype of data and how it was used.This information can be provided here or in Item 12

above.

Item 15: Main findingsPresent the key findings, linking them to CMO configu-rations. Show how they were used to further develop,test or refine the programme theory.

ExampleBy using the ripple effect concept with context-mechanism-outcome configuration (CMOc), trust was attimes configured as an aspect of context (i.e. trust/mis-trust as a precondition or potential resource), othertimes as a mechanism (i.e. how stakeholdersresponded to partnership activities) and was also anoutcome (i.e. the result of partnership activities) in dy-namically changing partnerships over time. Trust ascontext, mechanism and outcome in partnerships gen-erated longer term outcomes related to sustainability,spin-off project and systemic transformations. Thefindings presented below are organized in two sections:(1) Dynamics of trust and (2) Longer-term outcomesincluding sustainability, spin-off projects and systemictransformations [38].

ExplanationThe defining feature of a realist evaluation is that it isexplanatory rather than simply descriptive, and that theexplanation is consistent with a realist philosophy ofscience.A major focus of any realist evaluation is to use the data

to support, refute and refine the programme theory –gradually turning it into a realist programme theory.Ideally, in realist evaluations, these processes used to sup-port, refute or refine the programme theory should be ex-plicitly reported.It is common practice in evaluation reports to align

findings against the specific questions for the evaluation.The specific aspects of programme theory that are rele-vant to the particular question should be identified, thefindings for those aspects of programme theory pro-vided, and modifications to those aspects of programmetheory made explicit.The findings in a realist evaluation necessarily include

inferences about the links between context, mechanismand outcome and the explanation that accounts for theselinks. The explanation may draw on formal theoriesand/or programme theory, or may simply comprise in-ferences drawn by the evaluators on the basis of the dataavailable. It is important that, where inferences are made,this is clearly articulated. It is also important to include asmuch detailed data as possible to show how these

inferences were arrived at. These data provided may (forexample) support inferences about a factor operating as acontext within a particular CMO configuration. The the-ories developed within a realist evaluation often have tobe built up from multiple inferences made on data col-lected from different sources. Providing the details of howand why these inferences were made may require that(where possible) additional files are provided, either onlineor at request from the evaluation team.When reporting CMO configurations, evaluators

should clearly label what they have categorised as con-text, what as mechanism and what as outcome withinthe configuration.Outcomes will of course need to be reported. Realist

evaluations can provide ‘aggregate’ or ‘average’ out-comes. However, a defining feature of a realist evaluationis that outcomes will be disaggregated (for whom, inwhat contexts) and the differences explained. That is,findings aligned against the realist programme theoryare used to explain how and why patterns of outcomesoccur for different groups or in different contexts. In otherwords, the explanation should include a description andexplanation of the behaviour of key mechanisms underdifferent contexts in generating outcomes [1, 20].Transparency of the evaluation processes can be dem-

onstrated, for example, by including such things as ex-tracts from evaluators’ notes, a detailed worked example,verbatim quotes from primary sources, or an explorationof what initially appeared to be disconfirming data (i.e.findings which appeared to refute the programme theorybut which, on closer analysis, could be explained byother contextual influences).Multiple sources of data might be needed to support

an evaluative conclusion. It is sometimes appropriate tobuild the argument for a conclusion as an unfoldingnarrative in which successive data sources increase thestrength of the inferences made and the conclusionsdrawn.Where relevant, disagreements or challenges faced by

the evaluators in making any inferences should be re-ported here.

Discussion sectionItem 16: Summary of findingsSummarise the main findings with attention to theevaluation questions, purpose of the evaluation,programme theory and intended audience.

ExampleThe Family by Family Program is young and there is, asyet, relatively little outcomes data available. The findingsshould, therefore, be treated as tentative and open torevision as further data becomes available. Nevertheless,the outcomes to date appear very positive. The model

Wong et al. BMC Medicine (2016) 14:96 Page 13 of 18

Page 14: RAMESES II reporting standards for realist evaluations

appears to engage families in genuine need of support, in-cluding those who may be considered ‘difficult’ in trad-itional services and including those with child protectionconcerns. It appears to enable change and to enable dif-ferent kinds of families to achieve different kinds of out-comes. It also appears to enable families to start withimmediate goals and move on to address more funda-mental concerns. The changes that families make appearto generate positive outcomes for both adults and chil-dren, the latter including some that are potentially verysignificant for longer term child development outcomes[49].

ExplanationThis section should be succinct and balanced. Specific-ally for realist evaluations, it should summarise and ex-plain the main findings and their relationships to thepurpose of the evaluation, questions and the ‘final’ re-fined realist programme theory. It should also highlightthe strength of evidence [50] for the main conclusions.This should be done with careful attention to the needsof the main users of the evaluation. Some evaluatorsmay choose to combine the summary and conclusionsinto one section.

Item 17: Strengths, limitations and future directionsDiscuss both the strengths of the evaluation and itslimitations. These should include (but need not belimited to): (1) consideration of all the steps in the evalu-ation processes and (2) comment on the adequacy, trust-worthiness and value of the explanatory insights whichemerged.In many evaluations, there will be an expectation to pro-

vide guidance on future directions for the programme,policy or initiative, its implementation and/or design. Theparticular implications arising from the realist nature ofthe findings should be reflected in these discussions.

ExampleAnother is regarding the rare disagreement to what wasproposed in the initial model as well as participants in-frequently stating what about the AADWP [AboriginalAlcohol Drug Worker Program] does not work for clients.The fact that participants rarely disagreed with our pro-posed model may be due to acquiescence to the inter-viewer, as it is often easier to agree than to disagree. Aswell, because clients had difficulty generating ways inwhich the program was not helpful, this may speak to apotential sampling bias, as those clients that have hadpositive experiences in the program may have been morelikely to be recruited and participate in the interviews.This potential issue was addressed by conducting an ex-ploratory phase (Theory Development Phase) and then aconfirmatory phase (Evaluation Phase) to try to ensure

the most accurate responses were retrieved from a widervariety of participants [28].

ExplanationThe strengths and limitations in relation to realist meth-odology, its application and utility and analysis shouldbe discussed. Any evaluation may be constrained by timeand resources, by the skill mix and collective experienceof the evaluators, and/or by anticipated or unanticipatedchallenges in gathering the data or the data itself. Forrealist evaluations, there may be particular challengescollecting information about mechanisms (which cannotusually be directly observed), or challenges evidencingthe relationships between context, mechanism and out-come [19, 20, 22]. Both general and realist-specific limi-tations should be made explicit so that readers caninterpret the findings in light of them. Strengths (e.g. be-ing able to build on emergent findings by iterating theevaluation design) or limitations imposed by any modifi-cations made to the evaluation processes should also bereported and described.A discussion about strengths and limitations may need

to be provided earlier in some evaluation reports (e.g. aspart of the methods section).Realist evaluations are intended to inform policy or

programme design and/or to enable refinements totheir design. Specific implications for the policy orprogramme should follow from the nature of realistfindings (e.g. for whom, in what contexts, and how).These may be reported here or in Item 19.

Item 18: Comparison with existing literatureWhere appropriate, compare and contrast the evalua-tion’s findings with the existing literature on similar pro-grammes, policies or initiatives.

ExampleA significant finding in our study is that the main motiv-ating factor is ‘individual’ rather than ‘altruistic’ whereasthere is no difference between ‘external’ and ‘internal’motivation. Previously, the focus has been on internaland external motivators as the key drivers. Knowles et al.(1998), and Merriam and Caffarella (1999), suggestedthat while adults are responsive to some external motiva-tors (better jobs, promotions, higher salaries, etc.) themost potent motivators are internal such as the desire forincreased job satisfaction, self-esteem, quality of life, etc.However, others have argued that to construe motivationas a simple internal or external phenomenon is to denythe very complexity of the human mind (Brissette andHowes 2010; Misch 2002). Our view is that motivation ismultifaceted, multidimensional, mutually interactive,and dynamic concept. Thus a person can move between

Wong et al. BMC Medicine (2016) 14:96 Page 14 of 18

Page 15: RAMESES II reporting standards for realist evaluations

different types of motivation depending on the situation[51].

ExplanationComparing and contrasting the findings from an evalu-ation with the existing literature may help readers to putthe findings into context. For example, this discussionmight cover questions such as how does this evaluationdesign compare to others (e.g. were they theory-driven?)? What does this evaluation add, and to whichbody of work does it add? Has this evaluation reachedthe same or different conclusion to previous evaluations?Has it increased our understanding of a topic previouslyidentified as important by leaders in the field?Referring back to previous literature can be of great

value in realist evaluations. Realist evaluations developand refine realist programme theory (or theories) to ex-plain observed outcome patterns. The focus on howmechanisms work (or do not) in different contexts po-tentially enables cumulative knowledge to be developedaround families of policies and programmes or acrossinitiatives in different sectors that rely on the sameunderlying mechanisms [37]. Consequently, reporting forthis item should focus on comparing and contrasting thebehaviour of key mechanisms under different contexts.Not all evaluations will be required (or able) to report

on this item, although peer-reviewed academic articleswill usually be expected to address it.

Item 19: Conclusion and recommendationsList the main conclusions that are justified by the ana-lyses of the data. If appropriate, offer recommendationsconsistent with a realist approach.

ExampleOur realist analysis (Figure 1) suggests that workforce de-velopment efforts for complex, inter-organizationalchange are likely to meet with greater success when thefollowing contextual features are found:

� There is an adequate pool of appropriately skilledand qualified individuals either already working inthe organization or available to be recruited.

� Provider organizations have good human resourcessupport and a culture that supports staff developmentand new roles/role design.

� Staff roles and identities are enhanced and extendedby proposed role changes, rather than undermined ordiminished.

� The policy context (both national and local) allowsnegotiation of local development goals, rather thanimposing a standard, inflexible set of requirements.

� The skills and responsibilities for achievingmodernization goals are embedded throughout the

workforce, rather than exclusively tied todesignated support posts.

In reality, however, the optimum conditions formodernization (the right hand side of Fig. 1) are rarelyfound. Some of the soil will be fertile, while other key pre-conditions will be absent [52].

ExplanationA clear line of reasoning is needed to link the conclu-sions drawn from the findings with the findings them-selves, as presented in the results section. If theevaluation is small or preliminary, or if the strength ofevidence behind the inferences is weak, firm implica-tions for practice and policy may be inappropriate. Someevaluators may prefer to present their conclusions along-side their data (i.e. in ‘Item 15: Main findings’).If recommendations are given, these should be consist-

ent with a realist approach. In particular, if recommen-dations are based on programme outcome(s), therecommendations themselves should take account ofcontext. For example, if an evaluation found that aprogramme worked for some people or in some contexts(as would be expected in a realist evaluation), it wouldbe inappropriate to recommend that it be run every-where for everyone [53]. Similarly, recommendations forprogramme improvement should be consistent withfindings about how the programme has been found towork (or not) – for example, to support the features ofimplementation that fire ‘positive’ mechanisms in par-ticular contexts, or to redress features that preventintended mechanisms from firing.

Item 20: Funding and conflict of interestState the funding source (if any) for the evaluation, therole played by the funder (if any) and any conflicts of in-terests of the evaluators.

ExampleThe authors thank and acknowledge the funding providedfor the Griffith Youth Forensic Service—NeighbourhoodsProject by the Department of the Prime Minister andCabinet. The views expressed are the responsibility of theauthors and do not necessarily reflect the views of theCommonwealth Government [54].

ExplanationThe source of funding for an evaluation and/or personalconflicts of interests may influence the evaluationquestions, methods, data analysis, conclusions and/orrecommendations. No evaluation is a ‘view from no-where’, and readers will be better able to interpret theevaluation if they know why it was done and for whichcommissioner.

Wong et al. BMC Medicine (2016) 14:96 Page 15 of 18

Page 16: RAMESES II reporting standards for realist evaluations

If an evaluation is published, the process for reportingfunding and conflicts of interest as set out by the pub-lisher should be followed.

DiscussionThese reporting standards for realist evaluation havebeen developed by drawing together a range ofsources – namely, existing published evaluations,methodological papers, a Delphi panel with feedbackcomments, discussion and further observations from amailing list, training sessions and workshops. Ourhope is that these standards will lead to greater trans-parency, consistency and rigour of reporting and,thereby, make realist evaluation reports more access-ible, usable and helpful to different stakeholders.These reporting standards are not a detailed guide on

how to undertake a realist evaluation. There are existingresources, published (see Background) and in prepar-ation, that are better suited for this purpose. These stan-dards have been developed to assist the quality ofreporting of realist evaluations and the work of pub-lishers, editors and reviewers.Realist evaluations are used for a broad range of topics

and questions by a wide range of evaluators from differ-ent backgrounds. When undertaking a realist evaluation,evaluators frequently have to make judgements and in-ferences. As such, it is impossible to be prescriptiveabout what exactly must be done in a realist evaluation.The guiding principle is that transparency is important,as this will help readers to decide for themselves if thearguments for the judgements and inferences made werecoherent, trustworthy and/or plausible, both for thechosen topic and from a methodological perspective.Within any realist evaluation report, we strongly encour-age authors to provide details on what was done, whyand how – in particular with respect to the analytic pro-cesses used. These standards are intended to supplementrather than replace the exercise of judgement by evalua-tors, editors, readers and users of realist evaluations.Within each item we have indicated where judgementneeds to be exercised.The explanatory and theory-driven focus of realist

evaluation means that detailed data often needs to be re-ported in order to provide enough support for inferencesand/or judgments made. In some cases, the word countand presentational format limitations required by com-missioners of evaluations and/or journals may not en-able evaluators to fully explain aspects of their worksuch as how judgments were made or inferences arrivedat. Ideally, alternative ways of providing the necessarydetails should be found such as by online appendices oradditional files available from authors on request.In developing these reporting standards, we were

acutely aware that realist evaluations are undertaken by

a range of evaluators from different backgrounds on di-verse topics. We have therefore tried to produce report-ing standards that are not discipline specific but suitablefor use by all. To achieve this end, we deliberately re-cruited panel members from different disciplinary back-grounds, crafted the wording in each item in such a wayas to be accessible to evaluators from diverse back-grounds and created flexibility in what needs to be re-ported and in what order. However, we have not had theopportunity to evaluate the usefulness and impact ofthese reporting standards. This is an avenue of workwhich we believe is worth pursuing in the future.

ConclusionsThese reporting standards for realist evaluations havebeen developed by drawing on a range of sources. Wehope that these standards will lead to greater consistencyand rigour of reporting and make realist evaluation re-ports more accessible, usable and helpful to differentstakeholders. Realist evaluation is a relatively new ap-proach to evaluation and with increasing use and meth-odological development changes are likely to be neededto these reporting standards. We hope to continue im-proving these reporting standards through our email list(www.jiscmail.ac.uk/RAMESES) [21], wider networks,and discussions with evaluators, researchers and thosewho commission, sponsor, publish and use realistevaluations.

AcknowledgementsThis project was funded by the National Institute for Health Research HealthServices and Delivery Research Programme (project number 14/19/19). Weare grateful to Nia Roberts from the Bodleian Library, Oxford, for her helpwith designing and executing the searches. We also wish to thank RayPawson and Nick Tilley for their advice, comments and suggestions whenwe were developing these reporting standards. Finally, we wish to thank thefollowing individuals for their participation in the RAMESES II Delphi panel:Brad Astbury, University of Melbourne (Melbourne, Australia); Paul Batalden,Dartmouth College (Hanover, USA); Annette Boaz, Kingston and St George’sUniversity (London, UK); Rick Brown, Australian Institute of Criminology(Canberra, Australia); Richard Byng, Plymouth University (Plymouth, UK);Margaret Cargo, University of South Australia (Adelaide, Australia); SimonCarroll, University of Victoria (Victoria, Canada); Sonia Dalkin, NorthumbriaUniversity (Newcastle, UK); Helen Dickinson, University of Melbourne(Melbourne, Australia); Dawn Dowding, Columbia University (New York, USA);Nick Emmel, University of Leeds (Leeds, UK); Andrew Hawkins, ARTDConsultants (Sydney, Australia); Gloria Laycock, University College London(London, UK); Frans Leeuw, Maastricht University (Maastricht, Netherlands);Mhairi Mackenzie, University of Glasgow (Glasgow, UK); Bruno Marchal,Institute of Tropical Medicine (Antwerp, Belgium); Roshanak Mehdipanah,University of Michigan (Ann Arbor, USA); David Naylor, Kings Fund (London,UK); Jane Nixon, University of Leeds (Leeds, UK); Peter O’Halloran, Queen’sUniversity Belfast (Belfast, UK); Ray Pawson, University of Leeds (Leeds, UK);Mark Pearson, Exeter University (Exeter, UK); Rebecca Randell, University ofLeeds (Leeds, UK); Jo Rycroft-Malone, Bangor University (Bangor, UK); RobertStreet, Youth Justice Board (London, UK); Nick Tilley, University CollegeLondon, (London, UK); Robin Vincent, Freelance consultant (Sheffield, UK);Kieran Walshe, University of Manchester (Manchester, UK); Emma Williams,Charles Darwin University (Darwin, Australia). All the authors were alsomembers of the Delphi panel.

Wong et al. BMC Medicine (2016) 14:96 Page 16 of 18

Page 17: RAMESES II reporting standards for realist evaluations

Authors’ contributionsGWo carried out the literature review. GWo, GWe, AM, JG, JJ and TG analysedthe findings from the review and produced the materials for the Delphipanel. They also analysed the results of the Delphi panel. TG conceived ofthe study and all the authors participated in its design. GWo coordinated thestudy, ran the Delphi panel and wrote the first draft of the manuscript. Allauthors read and contributed critically to the contents of the manuscript andapproved the final manuscript.

Competing interestsThe authors declare that they have no competing interests. The views andopinions expressed therein are those of the authors and do not necessarilyreflect those of the HS&DR program, NIHR, NHS or the Department of Health.

Author details1Nuffield Department of Primary Care Health Sciences, University of Oxford,Radcliffe Primary Care Building, Radcliffe Observatory Quarter, WoodstockRoad, Oxford OX2 6GG, UK. 2Community Matters, PO Box 443, MountTorrens, South Australia 5244, Australia. 3School of Sociology and SocialPolicy, University of Leeds, Leeds LS2 9JT, UK. 4Centre for Advancement inRealist Evaluation and Synthesis, Institute of Psychology, Health and Society,University of Liverpool, Waterhouse Building, Block B, Brownlow Street,Liverpool L69 3GL, UK.

Received: 29 March 2016 Accepted: 14 June 2016

References1. Pawson R, Tilley N. Realistic evaluation. London: Sage; 1997.2. Pawson R. The Science of Evaluation: A Realist Manifesto. London: Sage; 2013.3. Moore G, Audrey S, Barker M, Bond L, Bonell C, Hardeman W, et al. Process

evaluation of complex interventions: Medical Research Council guidance.BMJ. 2015;350:h1258.

4. Kwamie A, van Dijk H, Agyepong IA. Advancing the application of systemsthinking in health: realist evaluation of the Leadership DevelopmentProgramme for district manager decision-making in Ghana. Health ResPolicy Syst. 2014;12:29.

5. Manzano-Santaella A. A realistic evaluation of fines for hospital discharges:Incorporating the history of programme evaluations in the analysis.Evaluation. 2011;17:21–36.

6. Marchal B, Dedzo M, Kegels G. A realist evaluation of the managementof a well-performing regional hospital in Ghana. BMC Health Serv Res.2010;10:24.

7. Lefroy J, Hawarden A, Gay SP, McKinley RK, Cleland J. Grades in formativeworkplace-based assessment: a study of what works for whom and why.Med Educ. 2015;49:307–20.

8. Horrocks I, Budd L. Into the void: a realist evaluation of the eGovernmentfor You (EGOV4U) project. Evaluation. 2015;21:47–64.

9. Coryn C, Noakes L, Westine C, Schroter D. A systematic review of theory-driven evaluation practice from 1990 to 2009. Am J Eval. 2010;32:199–226.

10. Westhorp G. Brief Introduction to Realist Evaluation. http://www.communitymatters.com.au/gpage1.html. Accessed 29 March 2016.

11. Greenhalgh T, Wong G, Jagosh J, Greenhalgh J, Manzano A, Westhorp G, etal. Protocol—the RAMESES II study: developing guidance and reportingstandards for realist evaluation. BMJ Open. 2015;5:e008567.

12. Pawson R. Theorizing the interview. Br J Sociol. 1996;47:295–314.13. Manzano A. The craft of interviewing in realist evaluation. Evaluation. 2016.

doi:10.1177/1356389016638615.14. AGREE collaboration. Development and validation of an international

appraisal instrument for assessing the quality of clinical practice guidelines:the AGREE project. Qual Saf Health Care. 2003;12:18–23.

15. Moher D, Schulz KF, Altman DG. The CONSORT statement: revisedrecommendations for improving the quality of reports of parallel-grouprandomised trials. Lancet. 2001;357:1191–4.

16. Davidoff F, Batalden P, Stevens D, Ogrinc G, Mooney S, for the SQUIREDevelopment Group. Publication Guidelines for Improvement Studiesin Health Care: Evolution of the SQUIRE Project. Ann Intern Med.2008;149:670–6.

17. Wong G, Greenhalgh T, Westhorp G, Pawson R. RAMESES publicationstandards: realist syntheses. BMC Med. 2013;11:21.

18. Wong G, Greenhalgh T, Westhorp G, Pawson R. RAMESES publicationstandards: realist syntheses. J Adv Nurs. 2013;69:1005–22.

19. Marchal B, van Belle S, van Olmen J, Hoerée T, Kegels G. Is realist evaluationkeeping its promise? A review of published empirical studies in the field ofhealth systems research. Evaluation. 2012;18:192–212.

20. Pawson R, Manzano-Santaella A. A realist diagnostic workshop. Evaluation.2012;18:176–91.

21. RAMESES JISCM@il. www.jiscmail.ac.uk/RAMESES. Accessed 29 March 2016.22. Dalkin S, Greenhalgh J, Jones D, Cunningham B, Lhussier M. What's in a

mechanism? Development of a key concept in realist evaluation. ImplementSci. 2015;10:49.

23. Emmel N. Sampling and choosing cases in qualitative research: A realistapproach. London: Sage; 2013.

24. EQUATOR Network. http://www.equator-network.org/Accessed 18 June2016.

25. Liberati A, Altman D, Tetzlaff J, Mulrow C, Gotzsche P, Ioannidis J, et al. ThePRISMA statement for reporting systematic reviews and meta-analyses ofstudies that evaluate healthcare interventions: explanation and elaboration.BMJ. 2009;339:b2700.

26. Greenhalgh T, Humphrey C, Hughes J, Macfarlane F, Butler C, Pawson R.How do you modernize a health service? A realist evaluation of whole-scaletransformation in London. Milbank Q. 2009;87:391–416.

27. Blamey A, Mackenzie M. Theories of change and realistic evaluation: peas ina pod or apples and oranges? Evaluation. 2007;13:439–55.

28. Davey CJ, McShane KE, Pulver A, McPherson C, Firestone M. A realistevaluation of a community-based addiction program for urban aboriginalpeople. Alcohol Treat Q. 2014;32:33–57.

29. Prashanth NS, Marchal B, Narayanan D, Kegels G, Criel B. Advancing theapplication of systems thinking in health: a realist evaluation of a capacitybuilding programme for district managers in Tumkur, India. Health ResPolicy Syst. 2014;12:42.

30. Westhorp G. Realist impact evaluation: an introduction. London: OverseasDevelopment Institute; 2014.

31. Rushmer RK, Hunter DJ, Steven A. Using interactive workshops to promptknowledge exchange: a realist evaluation of a knowledge to actioninitiative. Public Health. 2014;128:552–60.

32. Owen J. Program evaluation: forms and approaches. 3rd ed. New York:The Guilford Press; 2007.

33. Pawson R, Sridharan S. Theory-driven evaluation of public healthprogrammes. In: Killoran A, Kelly M, editors. Evidence-based public health:effectiveness and efficacy. Oxford: Oxford University Press; 2010. p. 43–61.

34. Rogers P, Petrosino A, Huebner T, Hacsi T. Program theory evaluation:practice, promise, and problems. N Dir Eval. 2004;2000:5–13.

35. Mathison S. Encyclopedia of Evaluation. London: Sage; 2005.36. Manzano A, Pawson R. Evaluating deceased organ donation: a programme

theory approach. J Health Organ Manag. 2014;28:366–85.37. Husted GR, Esbensen BA, Hommel E, Thorsteinsson B, Zoffmann V.

Adolescents developing life skills for managing type 1 diabetes: aqualitative, realistic evaluation of a guided self-determination-youthintervention. J Adv Nurs. 2014;70:2634–50.

38. Jagosh J, Bush P, Salsberg J, Macaulay A, Greenhalgh T, Wong G, et al. Arealist evaluation of community-based participatory research: partnershipsynergy, trust building and related ripple effects. BMC Public Health.2015;15:725.

39. Hawkins A. Realist evaluation and randomised controlled trials for testingprogram theory in complex social systems. Evaluation. 2016. doi:10.1177/1356389016652744.

40. Parker J, Mawson S, Mountain G, Nasr N, Zheng H. Stroke patients'utilisation of extrinsic feedback from computer-based technology in thehome: a multiple case study realistic evaluation. BMC Med Inform DecisMak. 2014;14:46.

41. Dalkin S, Lhussier M, Philipson P, Jones D, Cunningham W. Reducinginequalities in care for patients with non-malignant diseases: Insights from arealist evaluation of an integrated palliative care pathway. Palliat Med. 2016.doi:10.1177/0269216315626352.

42. Rycroft-Malone J, Wilkinson J, Burton C, Andrews G, Ariss S, Baker R, et al.Implementing health research through academic and clinical partnerships: arealistic evaluation of the Collaborations for Leadership in Applied HealthResearch and Care (CLAHRC). Implement Sci. 2011;6:74.

43. Mirzoev T, Etiaba E, Ebenso B, Uzochukwu B, Manzano A, Onwujekwe O, etal. Study protocol: realist evaluation of effectiveness and sustainability of a

Wong et al. BMC Medicine (2016) 14:96 Page 17 of 18

Page 18: RAMESES II reporting standards for realist evaluations

community health workers programme in improving maternal and childhealth in Nigeria. Implement Sci. 2016;11:83.

44. Wye L, Lasseter G, Percival J, Duncan L, Simmonds B, Purdy S. What worksin 'real life' to facilitate home deaths and fewer hospital admissions forthose at end of life?: results from a realist evaluation of new palliative careservices in two English counties. BMC Palliative Care. 2014;13:37.

45. Maxwell J. A Realist Approach for Qualitative Research. London: Sage; 2012.46. Meier K, Parker P, Freeth D. Mechanisms that support the assessment of

interpersonal skills: A realistic evaluation of the interpersonal skills profile inpre-registration nursing students. J Pract Teach Learn. 2014;12:6–24.

47. Astbury B, Leeuw F. Unpacking black boxes: mechanisms and theorybuilding in evaluation. Am J Eval. 2010;31:363–81.

48. Wong G, Greenhalgh T, Westhorp G, Pawson R. Development ofmethodological guidance, publication standards and training materials forrealist and meta-narrative reviews: the RAMESES (Realist And Meta-narrativeEvidence Syntheses - Evolving Standards) project. Health Serv Deliv Res.Southampton (UK): NIHR Journals Library; 2014.

49. Community Matters Pty Ltd. Family by Family: Evaluation Report 2011-12.Adelaide: The Australian Centre for Social Innovation; 2012.

50. Haig B, Evers C. Realist Inquiry in Social Science. London: Sage; 2016.51. Sorinola O, Thistlethwaite J, Davies D, Peile E. Faculty development for

educators: a realist evaluation. Adv Health Sci Educ. 2015;20(2):385–401.doi:10.1007/s10459-014-9534-4.

52. Macfarlane F, Greenhalgh T, Humphrey C, Hughes J, Butler C, Pawson R. Anew workforce in the making? A case study of strategic human resourcemanagement in a whole-system change effort in healthcare. J Health OrganManag. 2009;25:55–72.

53. Tilley N. Demonstration, exemplification, duplication and replication inevaluation research. Evaluation. 1996;2:35–50.

54. Rayment-McHugh S, Adams S, Wortley R, Tilley N. ‘Think Global Act Local’: aplace-based approach to sexual abuse prevention. Crime Sci. 2015;4:1–9.

• We accept pre-submission inquiries

• Our selector tool helps you to find the most relevant journal

• We provide round the clock customer support

• Convenient online submission

• Thorough peer review

• Inclusion in PubMed and all major indexing services

• Maximum visibility for your research

Submit your manuscript atwww.biomedcentral.com/submit

Submit your next manuscript to BioMed Central and we will help you at every step:

Wong et al. BMC Medicine (2016) 14:96 Page 18 of 18