Skip to main content
Advertisement
  • Loading metrics

Twelve principles for transformation-focused evaluation

  • Sam J. Buckton ,

    Roles Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Project administration, Visualization, Writing – original draft, Writing – review & editing

    sam.buckton@york.ac.uk

    Affiliations Department of Environment and Geography, University of York, York, United Kingdom, International Evaluation Academy (IEAc),

  • Ioan Fazey,

    Roles Conceptualization, Funding acquisition, Methodology, Supervision, Writing – review & editing

    Affiliation Department of Environment and Geography, University of York, York, United Kingdom

  • Peter Ball,

    Roles Conceptualization, Funding acquisition, Methodology, Writing – review & editing

    Affiliation School for Business and Society, University of York, York, United Kingdom

  • Zenda Ofir,

    Roles Supervision, Writing – review & editing

    Affiliations International Evaluation Academy (IEAc),, Independent evaluation specialist,

  • John Colvin,

    Roles Writing – review & editing

    Affiliation Emerald Network Ltd, Horsley, United Kingdom

  • Matthew Darby,

    Roles Writing – review & editing

    Affiliation Department of Environment and Geography, University of York, York, United Kingdom

  • Adam Peter Hejnowicz,

    Roles Writing – review & editing

    Affiliation School of Geosciences, University of Edinburgh, Edinburgh, United Kingdom

  • Graham Leicester,

    Roles Writing – review & editing

    Affiliation International Futures Forum, Aberdour, United Kingdom

  • Rebecca Newman,

    Roles Writing – review & editing

    Affiliation Department of Environment and Geography, University of York, York, United Kingdom

  • Glenn G. Page,

    Roles Writing – review & editing

    Affiliation SustainaMetrix, Portland, Maine, United States of America

  • Kelly Parsons,

    Roles Writing – review & editing

    Affiliation MRC Epidemiology Unit, University of Cambridge School of Clinical Medicine, Cambridge, United Kingdom

  • Barbara van Mierlo

    Roles Writing – review & editing

    Affiliation Knowledge, Technology and Innovation, Wageningen University & Research, Wageningen, Netherlands

Abstract

There are growing calls for societies to transform towards regenerative futures that support the flourishing of life and transcend interlinked social and ecological crises. Many actors are now trying to identify whether desired transformations are occurring, and how different interventions are contributing to transformation. These are questions of evaluation. However, we lack a holistic appreciation of how evaluation can most effectively assist transformation efforts. We undertake a systematized scoping review of a large body of evaluation literature to identify principles and methodologies for transformation-focused evaluation. We identify twelve distinct but interdependent principles which, when applied by evaluators (and supported by evaluation commissioners, sponsors and funders), are considered to significantly enhance the effectiveness of evaluation in assessing and assisting transformations, although tensions exist between some of the principles. The principles were clustered into three overarching themes of Complexity Principles (how evaluation approaches complex systems), Power Principles (evaluators’ power relations), and Purpose Principles (evaluation’s deeper purpose and values). The Complexity Principles call for transformation-focused evaluators to appreciate evaluands (e.g., a transition initiative, sector, or region) as unique, entangled, nested, dynamic and uncertain systems, and employ more diverse, developmental, contextually adapted and future-sensitive evaluation approaches. The Power Principles urge evaluators to promote justice, embrace diverse and marginalized perspectives, and reduce evaluator-evaluand polarization by shifting power towards evaluation users, whilst also fostering a more autonomous evaluation profession and mutualistic partnerships of knowledge and action. Most fundamentally, the Purpose Principles advocate for a recentering of values, reflexivity and learning at the heart of evaluation practice, and ensure that the core purpose of evaluation is to support systemic, societal transformation towards regenerative futures. Relevant methodologies are suggested for operationalizing the principles. Transformation-focused evaluation is a radical shift from conventional practice but the urgency to address global crises makes the shift a crucial one.

Author summary

There are growing calls for societies to transform towards regenerative futures that support the flourishing of life and transcend current crises. Many actors are now trying to identify whether desired transformations are occurring, and how different interventions are contributing to transformation. These are questions of evaluation. However, we lack a holistic appreciation of how evaluation can most effectively assist transformation efforts. Here, we review a large body of evaluation literature to identify principles and methodologies for transformation-focused evaluation. We identify twelve distinct but interdependent principles which, when applied by evaluators (and supported by evaluation commissioners, sponsors and funders), are considered to significantly enhance the effectiveness of evaluation in assessing and assisting transformations, although tensions exist between some of the principles. The principles were grouped into three overarching themes of Complexity Principles (how evaluation approaches complex systems, like food or energy systems), Power Principles (evaluators’ power relations with different groups of people and nature more widely), and Purpose Principles (evaluation’s deeper purpose and values). Relevant methodologies are suggested for operationalizing the principles. Transformation-focused evaluation is a radical shift from conventional practice but the urgency to address global crises makes the shift a crucial one.

Introduction

The world faces deepening environmental, social and geopolitical crises, including climate change, biodiversity loss, wealth and health inequality, and an ‘infodemic’ of misinformation [14], driven strongly by the globalization of Western capitalism [5]. The complex nature of Earth’s social-ecological systems [6] (e.g., food, energy or transport systems, cities, economies, regions, and societies) – characterized by interconnectivity, feedbacks and other emergent and often unpredictable dynamics [7] – means that these crises are interlinked, often reinforcing, and ‘wicked problems’ to deal with [8,9]. There is increasing recognition of the need for transformative system-wide changes in societies to fundamentally address these crises [1,4,10,11], including within the framework of the United Nations’ Sustainable Development Goals [12]. Transformation is in fact inevitable, as failure to steward necessary transformations is likely to result in catastrophic disruption to human societies [13], which has already begun in earnest across many parts of the world [13,14]. It has been strongly argued that more desirable transformations should aim for radically different dynamics that support the flourishing of life – what many now refer to as ‘regenerative’ dynamics [15]. Regenerative framings go beyond conventional framings of sustainability that focus on minimizing anthropogenic harm, and instead focus on fostering conditions that result in mutual reinforcement or ‘spiraling up’ of human and wider ecological wellbeing [15]. Such transformations present a huge challenge, not least because they require fundamental shifts in our values and worldviews [15], various lock-ins and power dynamics make the status quo highly resistant to change [13,16], and we may have to act even whilst many of the familiar social and ecological systems around us are collapsing [13,17].

Accordingly, actors are now increasingly grappling with the challenge of answering questions like: Is our system transforming towards a desired future? What are the impacts (both desirable and undesirable) of our systems’ transformations? Is our initiative or intervention significantly contributing to transformation, and does it have the enabling conditions in place to achieve this? To what extent is our initiative or intervention aligned to transformative or regenerative principles? [4,18,19]. These are fundamentally questions of evaluation – the systematic assignment of quality, value or importance to something by people [20,21] – that enable learning about transformations and their stewardship [19]. While people apply evaluative thinking almost constantly in practically every activity, often subconsciously [22], evaluation has also become a formalized discipline and practice [23], and evaluation can be applied in more or less structured ways to particular evaluands, such as projects, programs, initiatives, policies, portfolios, strategies, interventions, communities, organizations, societies, regions, and social-ecological systems. However, many mainstream Western evaluation techniques – although valuable in the right circumstances – were not designed for contexts of transformation [24]. Accordingly, it has been stated that ‘Evaluating transformation means transforming evaluation’ [4], and there is increasing interest in, and application of, more transformation-focused evaluation approaches [24]. However, we lack a holistic appreciation of what ‘transforming evaluation’ entails for this purpose, particularly because the literature around this topic has received scant review to date.

In this paper we aim to provide comprehensive guidance which, when applied by evaluators (and supported by evaluation commissioners, sponsors and funders), is considered to fundamentally shift evaluation practice to significantly enhance its effectiveness in assessing and assisting transformation efforts. This guidance is based on a systematized scoping review and inductive thematic analysis of a large body of evaluation literature focused on transformation contexts. We structure the outputs of the review as a set of principles for transformation-focused evaluation along with suggested approaches, tools and other methodologies for operationalizing these principles. This is a valuable response to calls for greater coherence and alignment in ‘guiding principles, ethical guidelines and competencies for evaluators’ supporting transformation [25]. In the following sections, we first outline the historical context of the evaluation field and some of its key paradigmatic shifts over time to address issues inherent in transformations, and where our review adds value. We then describe our framing and methodological approach in carrying out the review. After presenting the resulting set of principles, we discuss them in a wider context and end with a call to action to take this work forward.

Background: transformations and evaluation

The need to support desirable transformations and navigate the transformations already happening around us raises new challenges for evaluation and evaluators [4,19,24,26,27]. Transformational change is broadly understood to involve some kind of major, fundamental change, as opposed to marginal or incremental change [28], such as adjustments and reforms [8,28,29]. As such, and when considered in relation to the many crises threatening societal systems – such as energy, transport, or food systems – transformation is often considered to go beyond merely technological, behavioral and policy changes, and involve shifts in underlying beliefs, values, intent, purpose, structures, institutions, power relations, paradigms or worldviews that in turn bring about system reconfiguration [3,5,3032]. These deeper shifts are generally more difficult to achieve [32]. From an evaluation perspective, this raises many issues, such as how to work with complexity, unpredictability, power inequalities and conflicting values, and ultimately how evaluation can help different actors learn how to adapt to, and support, systemic change and maintain ambition and transformational intent.

Many non-Western evaluation approaches have been working effectively with these issues for thousands of years. For instance, appreciation of complexity and relationality is deeply embedded in Eastern and African cultures and Indigenous thought and practice more generally [3335]. Western evaluation, on the other hand, has historically been rooted in a positivist paradigm that limits its utility in contexts of transforming complex social-ecological systems [36,37]. The positivist paradigm assumes that there is a single objective reality or truth best measured by a neutral, independent observer, and that change processes can be planned and controlled rationally [36,37]. Positivism lends itself to evaluation approaches that assess evidence for predetermined outcomes and causal relationships between interventions and these outcomes, focusing on quantitative and objective indicators of change [36]. However, positivist approaches are not well suited to contexts where there is more uncertainty in the goals of innovation projects, a need for ongoing adaptation and learning in the face of dynamically changing conditions, and a need to reconcile the diverse and often conflicting values and desires of different stakeholders [36,37]. These conditions are part and parcel of transformations in complex social-ecological systems.

Particularly since the 1960s, there have been efforts to align Western evaluation practice to addressing these kinds of issues. Several evaluation paradigms emerged out of critiques of positivism, including the post-positivist [38,39], constructivist [36,40], realist [4143], and pragmatic or utilization-focused paradigms [39]. Various other branches of evaluation have adopted more complexity-informed approaches and theories of change, rejecting the linear logic models that characterize positivist evaluation [44]. Collectively these paradigms bring greater attention to aspects such as observer biases [38,39], multiple subjective and co-created realities [36,40], the importance of the context (e.g., culture) in which interventions take place and its implications for (ir)replicability [41], the processes that lead to observed outcomes [43], and the practical use and effects of stakeholders’ knowledge, beliefs and values [39]. Associated evaluation approaches also adopt more qualitative, participatory and mixed methodologies [36,40], and typically choose methodologies to suit particular contexts rather than applying one-size-fits-all approaches [39]. Some of these paradigms are relatively mainstream in some sectors. For instance, realist, complexity-aware and participatory evaluation approaches are gaining prominence in the health sector [4547].

Some evaluation paradigms have gone further by encouraging stakeholders to challenge problematic values and institutions, which are seen as key barriers to reconfiguring social-ecological systems [36]. For example, the critical realist paradigm, related to critical systems thinking/heuristics [48,49], brings in greater awareness of the systemic structures that influence a system’s behavior, and reflexivity of evaluators about their own positionality and power entanglements [50,51]. The reflexive evaluation paradigm, meanwhile – grounded in Beck’s theory of reflexive modernization [52] – takes the stance of supporting networks of actors to challenge existing systems and develop radically different visions, practices and systems, and not merely reflecting on them [36]. It also differs from other critical theories in considering intertwined rules and norms, interdependencies, infrastructure and existing technologies, as well as the stabilizing mechanism of power relations [53]. Mertens’ transformative evaluation paradigm adopts a clear normative stance of leveling power inequalities and elevating marginalized voices [54]. These paradigms radically shift the traditional neutral, independent stance of evaluators towards more embedded, advocative and ethical stances, but also greater reflexivity about evaluators’ inevitable biases and their implications for how evaluations are conducted.

More recently (spurred particularly by the COVID-19 pandemic), evaluators have been urged to fundamentally shift their practices to support large-scale (even global-scale) systemic transformations towards more sustainable and resilient futures, in recognition of the urgency to address planetary crises [4,24,26,27]. Some have linked this shift to trends towards post-normal evaluation that recognizes and engages with complexity, unpredictability, incompleteness, instability, transformations, and a plurality of values [4,55]. Post-normal evaluation gives greater importance to practical or phronetic reasoning (including moral, experiential, emotional, social and political dimensions) rather than focusing only on scientific and economic rationality [55]. As well as attending to issues of power and complexity, this shift includes focusing evaluation criteria on aspects such as ‘transformation fidelity’ (the match between transformation rhetoric and change on the ground), the alignment and momentum amongst different actors and initiatives, and how ecosystem resilience is being manifested [4]. Fundamentally, this also means evaluators acknowledging that they have ‘skin in the game … the game being the future of humanity on Earth’ (i.e., a stake in successful transformations) and taking a moral position of supporting positive transformations [4]. Methodological approaches such as Blue Marble Evaluation or Dynamic Evaluation have been advocated for this transformation-focused or post-normal evaluation paradigm [56,57]. Examining the historical paradigm shifts in the evaluation field situates the recent turn to transformation-focused evaluation as another step in the field’s evolution towards accounting for the challenges of stewarding change in complex social-ecological systems.

Ideas around transformation-focused evaluation are gaining increasing recognition in the evaluation field [24,26]. Several recent conferences of prominent evaluation associations and societies have focused on evaluation for transformations [4], generating three books on transformational evaluation [5860] and the Prague Declaration on Evaluation for Transformational Change: a set of pledges for transformational evaluation agreed to by attending evaluators, commissioners, parliamentarians and other evaluation users [19]. An update of the Prague Declaration is currently under review; on 22 April 2024, the Earth Day Evaluation Declaration promoted an update to Point 6 of the Prague Declaration (‘Focus on sustainability’) that calls for ‘context-specific criteria for environmental, social, and economic sustainability and regeneration [in] all evaluations recognizing that these three domains are intertwined’ [61]. Criteria for evaluating transformations have also emerged from recent special issues of evaluation journals [4,62,63]. With the increasing number of organizations and networks working on transformations and transitions, the demand for transformation-focused evaluation is also growing, although examples of its application are still relatively few [19]. Several major cohering and funding organizations have started to embrace transformation-focused evaluation [34,62,6468], including the International Evaluation Academy [69].

Despite these developments, there is still no comprehensive, holistic guidance about what is needed for effective transformation-focused evaluation. Evaluation geared towards assisting transformations remains a niche practice and a major leap from how evaluation is most often undertaken [25]. Particularly for evaluators, commissioners, sponsors and funders with no prior experience of evaluations in contexts of transformation in complex social-ecological systems, who may naturally lean towards more positivist approaches, it can be challenging to find easily accessible guidance or navigate the many different evaluation approaches, methods and tools available (e.g., see [70]). We address this gap by conducting a rigorous review and stock-take across literature on evaluation in transformation contexts. Over 2300 documents were screened, and guidance was drawn out from 138 accepted documents to develop core principles for transformation-focused evaluation and identify methodologies to support operationalization of these principles.

Materials and methods

Overview

We conducted a systematized scoping review and inductive thematic analysis of academic and gray literature to identify: a) guiding principles and b) methodologies relevant to transformation-focused evaluation. Below we explain how the research was framed, followed by the specific methods used, including the process of literature searching, document screening for inclusion/exclusion, data extraction, coding, analysis and synthesis.

Framing

Our work was framed by our understanding of the concepts of social-ecological systems, transformation, regenerative systems or futures, and evaluation (see Introduction and Background), all of which fed into our understanding of transformation-focused evaluation. We saw transformation-focused evaluation as evaluation that acts in service of efforts to transform social-ecological systems [6] (e.g., food, energy or transport systems, cities, economies, regions, and societies) towards regenerative futures (broadly speaking, where human and non-human life can flourish) [15]. Such evaluation would: be a dynamic process of examining, questioning, deliberating and altering both means (e.g., decisions and interventions) and ends (e.g., goals, values and principles) [21] in contexts of actors in transformation-focused initiatives; focus on aspects such as identifying signals of transformational change and regenerative dynamics, understanding the impacts of transformation, determining the extent to which an initiative is contributing to transformation or has the potential to do so, assessing an initiative’s alignment to transformational principles, and evaluating whether initiatives’ principles, values, goals and theories of change are appropriate, effective and sufficiently ambitious or transformation-focused; and be supported by appropriate evaluation methodologies, approaches, methods, techniques, concepts, frameworks and tools. We also recognized transformation-focused evaluation as requiring transformation of evaluation itself: that evaluation practice needs to fundamentally change compared to how it is frequently undertaken today and conventions in the past, and not just tweaked, if it is to be fit for contexts of transformation and supporting transformation efforts [4,27]. This has also been strongly argued by others [19], e.g., in terms of the inadequacy of the revised Development Assistance Committee (DAC) criteria in communicating the urgency of transformation [4]. Some have used the term ‘transformational evaluation’ to describe transformation-focused evaluation [59], but we preferred ‘transformation-focused evaluation’ to avoid confusion with Mertens’ Transformative Evaluation paradigm [54]. Transformation-focused evaluation is not yet a cohered evaluation paradigm like Transformative Evaluation, and it was not our intention to suggest otherwise. Rather, our definition of transformation-focused evaluation, as well as the concepts underpinning it, were broad ‘sensitizing concepts’ [4] requiring contextual adaptation and high-inference interpretation, rather than universal and black-and-white [4].

Method

Taking our definition of transformation-focused evaluation as a starting point, we wished to understand: a) What are the different dimensions of transformation-focused evaluation that set it apart from other kinds of evaluation approaches, and are important for enhancing its effectiveness in assisting transformation? and b) What methodologies are relevant for transformation-focused evaluation? To answer these questions we used a systematized scoping review method [71]. Scoping reviews are appropriate for reviewing the ‘extent, range and nature of research activity’ around a poorly cohered topic, where breadth is more important than depth [71]. It was not our intention to carry out a fully replicable review, given the qualitative nature of document screening and analysis and the requirement for high-inference judgment from the researchers. Nonetheless, to increase methodological transparency and rigor, we drew on the recommendations of the PRISMA framework extension for scoping reviews (PRISMA-ScR) [72] (Fig 1; see S5 Table). The only deviations from the full PRISMA-ScR checklist (S5 Table) were: not critically appraising documents or identifying their funding sources before inclusion for coding (other than appraisal during screening), due to the inclusive nature of the review and the qualitative nature of the results sought; and for conciseness, not choosing to refer to the paper as a scoping review in the paper’s title, instead referring to this method in the Abstract.

thumbnail
Fig 1. Flow diagram for the scoping review, based on PRISMA-ScR [

72]. Image created by S.J.B. in Mural (https://www.mural.co/).

https://doi.org/10.1371/journal.pstr.0000164.g001

Search strategy.

We aimed to search for literature providing guidance for evaluation and evaluators (justified based on empirical evidence, other literature, or authors’ own opinions and experience) in contexts of transforming social-ecological systems towards regenerative futures, whilst also recognizing that in these contexts evaluation needs to fundamentally shift compared to how evaluation is frequently undertaken today and past evaluation conventions.

Literature searches were performed in Scopus, Google Scholar and Google (i.e., google.com) (Fig 1). Scopus was chosen over similar databases due to its combination of high overlap with other databases, high coverage of social sciences literature, and support of sophisticated Boolean and other search operators [73]. Although Google Scholar has more limited support of search operators, making it less suited to reproducible query-based searches [73], it was searched because it is the most comprehensive literature database known [73] and could therefore identify academic documents missed by Scopus, as well as moderate amounts of gray literature [74]. Google was searched mainly to obtain gray literature, of which it is an important source [75]. The search language was English, the dominant language of academia [76]. Searches in Google and Google Scholar were performed incognito in Google Chrome Version 110 (Official Build, 64-bit) with all cookies rejected, to avoid bias of search personalization. The global – i.e., no country redirect (NCR) – version of the search engines were used (https://scholar.google.com/ and https://www.google.com/). For explaining the relevance of particular evaluation approaches to transformation-focused evaluation, where information in the main literature sample was too scant, external texts located by hand-searching in Google were consulted.

Search strings (S1 Table) were created both inductively and deductively. Trial searches with initial search strings were tested, and additional terms then added based on terms used by documents found in the searches. All search strings included a term for evaluation (evaluat*). We focused on the term ‘evaluation’ for our searches because of its importance in describing the discipline of evaluation, and also its explicit reference to values, which are fundamental to reaching evaluative decisions [21,77]. We recognize that other terms may be used to describe evaluation (such as ‘assessment’, ‘monitoring’, or ‘appraisal’), and that our review may therefore omit knowledge from sources referring to evaluation only in this way (potentially more prevalent outside the main evaluation discipline). The strings also contained terms related to transformation or regenerative social-ecological systems (based on a recent review [15]).

As recommended for scoping reviews [71], we also included a ‘consultation exercise’ [71], in our case with a group of experienced evaluation and transformation researchers and practitioners, to supplement the findings of the literature review. This ‘advisory group’ was invited to suggest additional documents (academic or gray literature) not found in the database searches that offered substantively different perspectives on transformation-focused evaluation to those already identified, or presented other methodologies relevant to transformation-focused evaluation (Fig 1). They were also invited to draw on their own expertise to suggest additional guidance for transformation-focused evaluation, if this was not already evident from the literature sample and deemed vital for the effectiveness of transformation-focused evaluation. Any guidance based on the opinions of the advisory group were cited as such [78]. In practice the findings of the review were dominated by guidance from the literature sample, with only a handful of supplementary suggestions from the advisory group. Finally, the advisory group was asked to suggest case studies of real-world evaluations that illustrated the application of transformation-focused evaluation, two of which were chosen for the final paper. The advisory group also acted as co-authors in reviewing and editing the manuscript of this paper.

We acknowledge that while the search strategy includes diverse perspectives, it will also omit many voices. Academic knowledge is dominated by Western and Global North perspectives [79,80], and the cultural make-up of our advisory group means that we are unable to fully do justice to the principles and approaches of more marginalized groups, including a wide array of Indigenous and Eastern philosophies and methodologies [34,81], although some members of the advisory group had experience in working with such groups and knowledge.

Document screening for inclusion/exclusion.

Documents were then screened by S.J.B. for inclusion or exclusion in the review (Fig 1; S4 Table). Thus, no co-authors of this paper screened documents that they had themselves authored. Documents were first screened based on title (see S2 Table for full details of screening criteria) and excluded if they were clearly unrelated to desired transformation of social-ecological systems or regenerative social-ecological systems. In Google and Google Scholar, pages were screened until over half of the (ten) results on a page were excludable based on title, since this was typically an indication that relevance of results would decline significantly in subsequent pages; no further pages were screened past this point.

Remaining documents were screened based on abstract and keywords (or an equivalent section, such as a book overview or executive summary in a report). If these clearly did not suggest evaluation would be discussed in the context of desired transformation of social-ecological systems or regenerative social-ecological systems, the document was excluded.

Remaining documents had their full text screened. Reasons for exclusion at this stage included insufficient detail about evaluation (e.g., mentioned only in passing), evaluation insufficiently in the context of desired transformation of social-ecological systems or regenerative social-ecological systems, insufficient focus on how evaluation needs to fundamentally change, and documents neutrally observational of, rather than advocating, fundamental change in evaluation. Accepted documents advocated evaluation for contexts of desired transformation of social-ecological systems or regenerative social-ecological systems, and recognized that fundamental shifts were required in evaluation for it to be fit for these contexts. If a document lacked an abstract and keywords or equivalent section, it was immediately screened based on full text after the title screening. In books, if chapters had abstracts, all chapters were screened as if they were separate documents. Otherwise, the full book was screened.

Acceptable document types included journal articles, reports, books, book chapters, blogs, and formal statements. Editorials that were standalone articles were included, whereas those that primarily summarized articles in a special issue were excluded, given that these articles should in theory appear elsewhere in the literature sample if they are relevant. Documents in a language other than English were translated into English using Google Translate [82]. Documents whose full text could not be accessed were either requested from the University of York Library or the documents’ authors. Some documents in Google or Google Scholar were inaccessible because of a non-functional URL (Fig 1).

Data extraction, coding, analysis and synthesis.

From documents accepted post-screening, guidance for evaluators (and explanation of its importance) was extracted from the text (Fig 2; S4 Table). All data extraction, coding and analysis was undertaken by S.J.B., with the results critiqued by the advisory group. Inductive open-coding was used to identify initial distinct semantic themes [83], with quotes selected from the document to evidence any theme assigned (Fig 2). An iterative approach was used, whereby themes were identified as coding progressed, with earlier themes refined in light of later documents. In Mural (https://www.mural.co/), these first-level themes were then inductively grouped into second-level themes with a similar (either semantic or latent) topic (Fig 2) (for details see mural) [83]. In some cases, one first-level theme was incorporated into more than one second-level theme. A final round of thematic analysis investigated whether the second-level themes, when examined together, revealed any latent ‘meta-themes’ (Fig 2). Key components of the first-level themes were summarized based on the selected quotes, and these fed into the final text of this paper (Fig 2). The descriptions aimed to capture all unique substantive ideas associated with a theme. A saturating approach was used, such that the only themes and ideas recorded in a given document and incorporated into the analysis were those substantively different from ones already identified. This effectively treats each unique idea as equally valid regardless of how many documents mention it, rather than elevating the most commonly mentioned ideas. This was considered to help mitigate biases in the literature (e.g., towards Western perspectives).

thumbnail
Fig 2. Example of the thematic analysis used to identify principles for transformation-focused evaluation.

The example focuses on the Purpose Principles. Image created by S.J.B. in Mural (https://www.mural.co/).

https://doi.org/10.1371/journal.pstr.0000164.g002

We referred to the second-level themes as ‘principles’ because they provide broad advice and guidance and are hypotheses rather than truths: they may or may not be followed and lead to desired outcomes in a particular context [84]. The fidelity to, or effectiveness of, principles is the concern of Principles-focused Evaluation [84]. Principles can be useful ‘anchors’ in the dynamic, uncertain and complex environments that characterize transformations in social-ecological systems, and are sufficiently flexible to guide action even when contexts vary across space and time whereas more rigid, prescriptive criteria would risk losing relevance [84,85]. We maintained the advocative voice of the literature analyzed, e.g., wording the principles as ‘evaluators should…’ or ‘evaluators need to…’. The advisory group was asked to identify complementarities and tensions between the principles.

To identify methodologies for transformation-focused evaluation, we identified any specific named methodologies advocated or applied in accepted documents, with a clear link to transformation-focused evaluation. Under the broad umbrella of ‘methodologies’ we included any overarching evaluation approaches, complex system concepts and frameworks, theories of transformational change, regenerative system frameworks and tools, visioning methods, indicator frameworks, frameworks of evaluation criteria, and analysis methods (many of which imply a method of data collection). We interpreted overarching evaluation approaches as typically not prescribing specific evaluation methodologies, but rather as conceptual analytical perspectives based on principles which then have implications for the kinds of specific methodologies that might be relevant.

As with the principles, our interest was in uncovering the range of different methodologies advocated rather than understanding how many documents advocated a particular approach, so as coding progressed only methodologies not previously advocated were recorded. The relevance of any methodology to transformation-focused evaluation, and to particular principles of transformation-focused evaluation, was determined based on quotes from the document in the main literature sample, other literature, and/or opinions from the advisory group. We wish to emphasize that this is a quick and rudimentary form of Principles-focused Evaluation rather than an in-depth analysis.

Results

Twelve distinct guiding principles for transformation-focused evaluation were identified (Table 1). Broadly speaking, the principles addressed: issues of complexity (e.g., interconnectivity, irreplicability, uncertainty, rapidly changing conditions, and future-sensitivity); power (e.g., the power relations between evaluators and evaluation users, commissioners, sponsors and funders, as well as other actors, and the importance of justice in these relations); or the deeper purpose and values of evaluation (Table 1). The principles are nonetheless strongly interdependent and show many complementarities, as well as some tensions (Fig 3). The principles are explained in detail below. A diverse range of different evaluation methodologies were advocated for transformation contexts (Table 2). Some overarching evaluation approaches emphasize certain principles or groups of principles more than others, whilst a few (such as Blue Marble Evaluation and Reflexive Monitoring in Action) have relatively broad coverage across all the principles (Table 2).

thumbnail
Table 1. Principles for transformation-focused evaluation.

https://doi.org/10.1371/journal.pstr.0000164.t001

thumbnail
Fig 3. Examples of interdependencies between principles for transformation-focused evaluation.

Complementarities are shown in pale yellow boxes; tensions are shown in amber boxes. Image created by S.J.B. in Mural (https://www.mural.co/).

https://doi.org/10.1371/journal.pstr.0000164.g003

thumbnail
Table 2. Overarching evaluation approaches relevant to transformation-focused evaluation. The approaches listed are those advocated for transformation contexts in our literature sample or suggested by the advisory group. Principles are color-coded according to the three main groups of principles: green for Complexity Principles, purple for Power Principles, and blue for Purpose Principles. A tick in a given box indicates that the evaluation approach emphasizes or supports the corresponding principle, based on the literature examined and/or opinions of the advisory group. Details about each evaluation approach are provided in S3 Table.

https://doi.org/10.1371/journal.pstr.0000164.t002

Complexity principles

Five principles related to the systemic complexity inherent in transformations and its implications for evaluation. They include the need for evaluators to: treat evaluands as nested systems with complex interactions; recognize that evaluands are unique and need context-adapted evaluation; use mixed methods to understand contributions to change under conditions of uncertainty, imperfect information, and pluralist lived realities; use more agile and developmental approaches; and include foresight alongside retrospective evidence.

1. Complex systems principle: Treat evaluands as nested systems with complex interactions.

Evaluators should treat an evaluated system as having complex, interdependent relations with the surrounding systems it is nested within and interacts with during transformation processes [18,57].

This principle brings evaluation closer to the physical reality of systemic transformations [18]. As highlighted by global crises such as the COVID-19 pandemic, we live in an increasingly globalized and interconnected world, underscoring the need to remain aware of the interlinkages of system changes [56]. There are several key implications of this worldview for evaluation practice. Evaluators must be able to think and work integratively across spatial and temporal scales [113], taking into account ‘macro-issues’ rather than focusing solely on ‘micro-issues’ [57]. For instance, evaluators might adopt a ‘multi-level perspective’ that looks at niche, regime and landscape levels [134], a ‘glocal’ perspective that considers linkages between local and global systems [27], and consider how specific programs and projects contribute to broader strategies and policy [57]. Evaluators must also recognize that while some changes take place rapidly, others (particularly those at wider systemic scales) may take decades [78,113]. The Developmental and Foresight Principles detail further implications for evaluators regarding futures and timescales.

Evaluators must also understand how initiatives and interventions, and their effects, are influenced via complex interactions with their contexts [18]. This requires a more theory-driven approach of understanding mechanisms behind observed outcomes, taking context into account and mapping relationships and interdependencies between actors [43,123]. Examining a pre-existing system before intervening would include understanding its parts and their interconnections, the system as a whole, and the system’s dynamics over time, including its response to other interventions [102]. In contrast, evaluations with a less complexity-aware ‘project mentality’ may treat evaluated interventions ‘as if they existed in a vacuum’ [135], work from a positivist worldview grounded in Newtonian assumptions that favor predictability, replicability, and linear causal relationships between initiatives and measured outcomes [123,136], and adopt the general Western tendency for reductionism rather than appreciating interconnections [121]. Transformations can rarely be commanded and controlled in such linear and predictable ways [99], and innovation is messy and iterative rather than a smooth orderly process [137]. Moreover, many evaluations still focus on planned activities and intended outcomes and impacts rather than more emergent results [105], and although many give due consideration to side-effects on human systems, they neglect impacts on environmental systems, which means that evaluation can inadvertently contribute to the ongoing destruction of natural systems rather than support their protection and recovery [105].

Evaluators should instead bring into view the full range of social, economic and ecological impacts of an initiative beyond its immediate sphere, their interactions, and their modification by external forces [4,134,135,138] – looking not only at how interventions affect targeted feedback loops, but also how these then influence other feedback loops, subsystems, and the system as a whole [102], considering costs, benefits, spillover effects, trade-offs, synergies and dissergies [4,139]. System transformation is often influenced by contextual forces, such as the price of oil influencing renewable energy investment [113], or an initiative might start out as radical but because of environmental changes becomes more mainstream [18]. Other signs of complex system interactions that evaluators should be prepared for include outcomes that are emergent and unanticipated [18,67,99,139], self-organization [139], chaotic processes (where system outcomes are sensitive to initial conditions) [140], feedbacks [138], attractors (patterns that a system is ‘drawn’ to being in) [140], stigmergy (coordinated behavior of independent actors because of the shared environmental context) [140], rebound effects [113], path dependency [141], lock-in effects and system inertia [141,142], and tipping points [89]. A complex systems perspective of power relations may reveal apparently negative outcomes, like backlash from the establishment, to in fact be signs of program success [100]. The uncertainty inherent in complex systems also highlights the importance of mixed methods (see Detective Principle), and the need for constant adaptation and rapid feedback in evaluations [136] (see Developmental Principle).

The Complex Systems Principle has implications for chosen indicators and indicator frameworks. These should be holistic, assessing the multiple dimensions of sustainability in an integrated rather than siloed way [56,143]. For example: food system evaluators should recognize the cultural and social dimensions of how food links to health alongside nutritional dimensions [144]; the evaluation of regenerative buildings should explicitly recognize interconnections between human and ecological systems rather than just listing a limited set of performance requirements [145]; and more broadly, evaluations should assess how well integrated transformative initiatives are across sectors and silos, and the interactions between different transformative initiatives [27] (see Partnerships Principle). Reductionist and fragmented approaches to indicators, e.g., traditional pillars of sustainability (environmental, social, economic) – especially if they focus on easily quantified metrics – may omit important (especially qualitative) aspects, like justice, values, paradigms, worldviews, and human-nature connections [146148]. Evaluators may need to incorporate harder-to-measure compound indicators to capture systemic impacts [149]. Indicators should in any case be developed and altered dynamically, especially post-intervention when the system is better understood [102] (see also Developmental Principle).

Operationalizing this principle means moving away from reliance on linear logic models and key progress indicators, SMART goals, randomized controlled designs, and standardized, replicable, scalable models that are geared towards evaluating isolated projects and programs in stable environments rather than complex systems [27,123,150]. Evaluators should instead become fluent in systems thinking and complexity science [64,140]. Many concepts and frameworks can encourage systemic perspectives, including boundary critique [106], complexity theory [18], critical systems thinking/heuristics [106], living systems theory [147], relational systems thinking [106], complex adaptive systems [102,147], socio-technical systems [113], social-ecological systems [89], value-creating systems [78,151], and structuration theory [152]. Indicator frameworks from regenerative design and development fields could also help to consider system contexts in a more holistic way [153] (see Transformation Principle). Various analysis methods can then help to visualize, explore or otherwise experience the particular systems in question, such as causal loop diagramming [154], Participatory Impact Pathways Analysis [155], social network analysis [154], agent-based modeling [18], soft systems methodology [154], the World Mandala [78,156,157], Digital Twins [139], digital bioregional story maps [106], systems sensing [26] and systemic constellation [26]. These system concepts and tools can be used to develop more complexity-appropriate theories of change [26], although more traditional positivist approaches can also help in clarifying assumptions in theories of change [36]. Evaluators should also familiarize themselves with methods for conceptualizing and analyzing transformations and transitions in complex systems (see Transformation Principle).

Numerous overarching evaluation approaches are based on complexity and systems thinking principles, including Blue Marble, Design-driven, Developmental, Dynamic, Feminist, Footprint, Systemic, and Visionary Evaluation, amongst many others [67,113,123,91,92,94,107] (Table 2). Footprint Evaluation (aka Sustainability-inclusive Evaluation) provides guidance for evaluations to embed criteria of environmental sustainability and contribute to the protection and recovery of natural systems, encouraging appreciation of the interconnections between human and natural systems [105]. Reflexive Evaluation builds on the multi-level perspective that explains how interactions between niches, socio-technological regimes and landscape may lead to system change [53,125]. Indigenous Evaluation approaches are commonly based on more holistic and relational worldviews [33], where ‘worlds are made up of extended networks of kin to whom we have reciprocal responsibilities’ [109], whilst many Eastern philosophies still in use today also carry a deep appreciation of complexity [34]. The African Evaluation Principles are one of the first and few evaluation association principle sets to so explicitly embed a complex systems approach [88]. Other evaluation approaches are not necessarily specified for complex systems but are nonetheless valuable in such contexts; for instance, Theory-based Evaluation [158] (e.g., Realist Evaluation [18,41]) works from endogenously generated theories of change to unpack the mechanisms behind expected or observed impacts [158] (Table 2).

Capturing all outcomes (both positive and negative) of system interventions – the general principle of ‘true cost accounting’ [56,159] – would be supported by methods such as outcome harvesting/mapping and ripple effects mapping [65] (Box 1). These encourage brainstorming of observed outcomes and impacts regardless of whether they were expected. A form of Adversary Evaluation has additionally been advocated for Dynamic Evaluations in weighing up both positive and negative impacts of system interventions [57].

Part of an evaluator’s role is also to help evaluation users to unpack and navigate complexity, to facilitate understanding and agency [4]. Frameworks such as Three Horizons and the prioritization process in the Most Significant Change method, which are underpinned by (inductive) thematic analysis, can assist in this task [160162]; other recommendations include for evaluators to report specific impact stories alongside big-picture considerations [4], and orient to principles, as in Principles-focused Evaluation [163] (see Developmental Principle).

Box 1. Method – Ripple effects mapping

Ripple effects mapping (REM) is a participatory method that is useful for exploring diverse, emergent, unpredicted and unintended impacts (both positive and negative) of an initiative in a complex system, facilitating more agile, adaptive decision-making [164]. It moves away from questions about attribution (did A cause B?) towards questions about contribution (how did an intervention contribute to changing an outcome or a system?) [164]. Drawing on the principles of appreciative inquiry, REM can help to identify the ‘added value’ or spillover effects of an intervention [164]. These characteristics give REM an affinity to the Complexity, Context and Developmental Principles. REM is typically done in workshop settings using group interviews and reflections to brainstorm how the effects of an intervention have ‘rippled out’ across a system (visualizing these using mind maps or similar), without worrying whether these effects were intended as part of an initiative’s official plans or theory of change [164].

2. Context principle: Recognize that evaluands are unique and need context-adapted evaluation.

Evaluators should recognize that every evaluand and its context, culture and transformation have unique characteristics, necessitating context-adapted rather than one-size-fits-all evaluation approaches [4,57,136,140].

Transformation is inevitably context-specific [136]. This is a shift away from Newtonian assumptions of replicability that have characterized many evaluations [136]. As such, evaluators should be wary of using standardized methods and criteria without thought for their appropriateness to the situation [4]. They should instead be ‘bricoleurs’ [56], choosing among the diverse array of available approaches and tools (potentially exploring new and unfamiliar techniques) to select and creatively adapt those best-suited to their particular context [57,136,140] (Box 2). The urgency of societal transformation (see Transformation Principle) means that evaluators may have to rapidly develop credible methods for the situation in hand in limited timeframes (as well as with limited budget and expertise) [67]. Ideally, these methods would be co-developed with the evaluation’s intended users, treating them as authorities on whether and how the methods should be applied (see Power Shift Principle) [4]. This principle is important for ensuring that the effectiveness of evaluations is optimized for particular situations of transformation, and for honoring and celebrating evaluands’ unique qualities [78]. This is likely to make the evaluation more useful to its intended users [131].

This principle requires evaluators and stakeholders to build up a rich understanding of their particular locale, culture and context [123,165]. Evaluators may need to investigate the cultural, political and economic history of a place (e.g., visualized on a timeline of key events), as well as its norms and values [123,165] (see Values Principle). The shifting and contextual nature of LGBTQ+ identity is a good example of how evaluation should not rely on rigid standardized methods but remain more flexible and responsive [111].

The Context Principle has implications for selecting indicators of success. Transformation-focused evaluation has been contrasted against evaluations with market-driven, homogenized approaches that impose uniformity on a multicultural reality [166], and use prescriptive, one-size-fits-all indicators [167,168]. For instance, there has been criticism of education evaluations with a technical-positivist focus on students attaining set goals, and relatively superficial psychometric measurement of learning [169]. Recognizing evaluands as unique instead suggests that indicators should be collaboratively developed and tailored to particular places so that they are meaningful and culturally appropriate [167,168,170], or that evaluators should identify ‘best principles in practice’ rather than looking for adherence to predefined specific ‘best practice’ [123], and give a ‘tentative best judgment’ rather than recommending ‘best practice’ when identifying or extrapolating insights from evaluation results [171]. In education contexts, evaluation would encourage the development of more reflexive, conscientious, autonomous and democratically active citizens, appreciating that students cannot be slotted into predefined boxes [166,169].

Operationalizing this principle requires awareness of the many approaches, frameworks and tools outlined in the other principles. Furthermore, Blue Marble Evaluation makes bricolage explicit as one of its operating principles, whilst many approaches emphasize the importance of designing evaluations to fit their particular contexts (Table 2). Various methods have been advocated for deepening understanding of specific contexts, including examination of publicly available data (e.g., on demographics), media and social media analysis, interviews of people with long-term experience, and design labs [123]. To support stakeholders in exploring their context, building on strengths and focusing on what people already love and are motivated to protect – using approaches including asset-based community development, appreciative inquiry and positive psychology – could be more successful in engaging people in dialogue and encouraging sustained collective action than problem-focused approaches [172].

Box 2. Case study – Biodiverse Landscapes Fund evaluation

The Biodiverse Landscapes Fund (BLF) is a UK government-funded £100 million project spanning 2022-29 that funds initiatives in six biodiversity hotspots around the world that safeguard biodiversity, enhance natural carbon sinks, and develop economic opportunities for local people [173]. It aims to center the voices and rights of Indigenous peoples and other marginalized communities. There is potential for the fund to contribute to regenerative landscapes by decelerating and in some cases reversing current patterns of landscape-level ecocide and cognitive violence, and contribute to local to global learning and co-innovation networks dedicated to regenerative futures.

An independent evaluator team for the BLF is being hosted by Itad and Oxford Policy Management [173]. The team has developed a complexity and systems-informed evaluation framework (cf. Complex Systems Principle) which blends Developmental Evaluation and impact evaluations in 6-monthly, annual and expansive adaptive learning cycles (cf. Developmental Principle). Overall it exemplifies the use of bricolage to adapt to context, blending Principles-based Evaluation, Systemic Evaluation, contribution analysis (Box 4), and multiple learning traditions (cf. Context and Detective Principles). It also aims to contribute to systemic transformation towards regenerative futures (cf. Foresight and Transformation Principles). A complexity lens is necessary not least because the BLF engages with mosaic and often transboundary landscapes, human-wildlife interconnections, 18 host governments in the Global South plus the sponsoring government (UK) and large numbers of conservation and development partners, and has a multi-level focus including engagement with enabling factors outside landscapes (government policy, legal and illegal markets, and long-term finance).

3. Detective principle: Be a mixed-methods detective in an uncertain, multi-reality and hard-to-control world.

Given the uncertainty and pluralist lived realities in complex systems change, evaluators should draw on multiple, mixed methods to improve understanding about transformations and how interventions contribute to them [102], accepting that there are many equally valid ways of knowing [102], clear-cut conclusions about causality are usually unattainable [62,123,135,136] and most large-scale change is outside the control of individual initiatives [78].

Evaluators should be prepared to seek narratives, correlations, patterns and explanations rather than well-defined causal connections in complex systems change [18,136], and maximize opportunities to understand the system and enhance its learning capacity [123] through using a wide variety of tools, drawing on both qualitative and quantitative data sources [62,135], and improving access to information [123]. Evaluators may even have to be guided by hunches and principles in place of hard data about impact in systems [67]. In this way, conducting a transformation-focused evaluation is similar to being a detective or a court judge weighing up multiple sources of evidence and possible scenarios of change, and using abductive reasoning to identify the most likely chain of events [78].

This principle is important because available information about a complex system is never perfect or ideal [62], with clear causal relationships (e.g., between an intervention in a complex system and the system’s outcomes) and experimentally controlled counterfactuals rarely possible to establish [18,67]. Changes in one system may trigger changes in other systems, leading to transformations that seem distant in space and time from original interventions [174]. Moreover, a system or transformation is never a single reality, given that all actors experience it differently, highlighting the need for epistemological pluralism in evaluations [102]. Mixed methods are also important for engaging appropriately with diverse cultural groups in evaluations [175] (see Justice Principle).

A key shift in this principle is from the language of ‘attribution’ to ‘contribution’, which is more transparent about uncertainty [146]. Evaluators will need to consider multiple contributions from different initiatives and forces rather than attributions to single interventions [64], and assess whether a given contribution is aligning or cooperating with (or at least not obstructing) other efforts, which is important for enabling wider system change [78] (see Partnerships and Transformation Principles). Rather than attempting to causally link an initiative to specific outputs and outcomes, the initiative may also have to shift focus to its influences on wider system conditions and the directionality or trajectory of system change [64,86]. The difficulty of linking interventions to impacts also highlights the importance of examining the designs of strategies, policies, etc. prior to implementation to assess their likelihood of achieving desired impacts (e.g., whether they are aligned to transformation principles) [78] (Box 3).

Adversary Evaluation is the clearest embodiment of the ‘detective’ or ‘judicial’ approach to evaluations [87]. However, this principle has implications for data collection and analysis in evaluations more generally. Real-time data sources (e.g., Big Data) may be key, especially for more developmental evaluation approaches with rapid feedback [139] (see Developmental Principle), in addition to literature reviews, surveys, stakeholder opinions, modeling results, digital storytelling, blogs, social media/network analysis, etc. [4,123]. Combining more interpretivist epistemology and associated qualitative and relatively subjective data (e.g., from focus groups and open-ended interviews), with more positivist epistemology and associated quantitative and relatively objective data (e.g., quantitative indicators and systems modeling), could help to gain a more comprehensive picture of transitions and the diverse perspectives, needs and values of people in the system [102,142,176], whilst information from futures and foresight methods (e.g., scenarios) can complement information from the past and present [177] (see Foresight Principle). It is also important for evaluators to ‘catch the wave’ of recent advances in Artificial Intelligence, Machine Learning and Digital Twins for enriching system understanding [139]. In general, while quantitative modeling may be valuable, evaluators should bear in mind that its role is not so much to find ‘truths’, but more to facilitate the ‘probing’ of the system [134].

Various analysis methods enable structured exploration of the influence of interventions on systemic change in an abductive way, including contribution analysis [18] (Box 4), general elimination method [65], ripple effects mapping [164] (Box 1), and process tracing [18,65]. Comparative case studies and qualitative comparative analysis provide additional analysis methods for assessing complex forms of causation in situations with limited and imperfect data [65]. The Most Significant Change method [65], meanwhile, adopts the approach of collecting personal stories about the changes people have experienced, which can enhance impact evaluations [161] (Box 5). In general, evaluators may have to rely more on grounded-theory or inductive analysis rather than hypothesis-deductive approaches [78].

Box 3. Method – The importance of principles, criteria and questions for evaluating transformations

While there are many specific evaluation approaches, methods, frameworks and tools that can support transformation-focused evaluation, it is ultimately dependent on appropriate principles, criteria and evaluation questions, ideally developed collaboratively with evaluation users to fit particular contexts [4,78] (see Context and Power Shift Principles).

This paper presents a set of principles that can guide transformation-focused evaluations (Table 1), and the methodologies advocated emphasize or support some or many of these principles (Table 2). Principles are important not only for how evaluations are done, but also for the transformation-focused initiatives being evaluated; these then become criteria against which the initiative can be evaluated, and inform what evaluation questions should be asked [4,63]. For example, the principle of transformation fidelity (i.e., that action should back up transformation rhetoric) leads to an equivalent criterion: the extent to which the realities of transformational change initiatives match transformational aspirations and rhetoric [4]. An associated evaluation question would be: ‘What is being envisioned as transformational and how [well] is that vision being realized through action?’ [4]. As a further illustration, Realist Evaluation is built around the principle that the system context (including cultural and normative dimensions) influences the impact of interventions [41]. This leads to more nuanced evaluation questions than merely ‘How well does the intervention work?’ – rather, ‘What works for whom, in what circumstances, in what respects, and how?’ [42]. Principles can be useful ‘anchors’ in dynamic, uncertain and complex environments, and are sufficiently flexible to guide action even when contexts vary across space and time whereas more rigid, prescriptive criteria would risk losing relevance [84,85]. Given the difficulty of determining an initiative’s contributions to systemic transformation, evaluations may often have to shift focus back to the initiative’s alignment to transformational principles in its design and implementation (see Detective Principle).

Changing evaluation principles and criteria is a key aspect of transforming evaluation practice to make it fit for evaluating transformation [4]. Conventionally, many evaluations follow the Development Assistance Committee (DAC) evaluation criteria: namely, assessing evaluands’ relevance, coherence, effectiveness, efficiency, impact, and sustainability [4]. However, these criteria have been criticized for giving a ‘business as usual message’ that fails to sufficiently address the complexity, magnitude and urgency of transformation [4].

Box 4. Method – Contribution analysis

Contribution analysis uses abductive reasoning (deciding which of various possible explanations for a phenomenon is the most compelling) to evaluate the contribution made by an intervention to an observed change [18]. Although originally designed for evaluating simple interventions, it has since been used in contexts of transformation in complex systems [18]. Broadly speaking, contribution analysis involves an ongoing iterative and participatory process of creating a theory of change or ‘contribution story’ (potentially involving multiple ‘sub-theories of change’ in complex settings), assessing evidence in support of it and possible alternative explanations for observed outcomes, and then refining the contribution story based on this analysis [18]. Rather than answering black-and-white questions about attribution and causation, contribution analysis helps to answer more nuanced questions about an intervention’s contribution to change, recognizing that it will be just one of many influencing factors [18]. Contribution analysis is thus allied to principles such as the Complexity, Detective and Developmental Principles.

4. Developmental principle: Remain agile and developmental.

Evaluators should avoid treating evaluands as static, and instead recognize the dynamic, unsettled nature of transformations and therefore the need for continual reflection, feedback, experimentation, adaptation and development [99,141,101].

This principle is in contrast to evaluations based on a linear process of setting predetermined objectives, implementing them, and then summatively evaluating the impact of an evaluand (a project, program, initiative, etc.) in light of those objectives [178]. Indeed, the principle represents a fundamentally different perspective on sustainability that recognizes the need for resilience and continual adaptation, rather than the linear continuation of positive results [27]. It is also a shift from aiming towards fixed, ideal states (which may not exist), instead focusing more on evaluating an ongoing process with emergent outcomes [26,147].

Rather, evaluations should be rapid, agile and adaptive [146], use a monitoring system with real-time feedback [102], and avoid setting evaluation logic (e.g., criteria, approaches and theories of change) in stone from the outset, but rather allow it to evolve over time [18,99]. For instance, theories of change may need frequent revision [18]. Evaluations may also need to expand their time horizons to start seeing meaningful effects and system ‘tipping points’ resulting from interventions [134,102], and ensure capacity for post-implementation evaluation when designing initiatives [179]; in many evaluations, questions about longer-term impact and sustainability are only briefly explored or dismissed as outside the evaluation’s timeframe [24]. Similarly, a shift may be required from assessing outputs, cost-effectiveness, efficacy and impact, to assessing an initiative’s capacity to deal with ongoing global challenges, including reflection and learning processes that can contribute to longer-lasting changes [94] (see Values and Learning Principles). These shifts clash with what many people expect from an evaluation, which could otherwise result in forced premature adoption of models not because they are appropriate, but because evaluators, funders or other stakeholders demand them to fit their preconceived notions of what good evaluation is [180]. Funders will need to recognize the need for more developmental and agile evaluation, with potentially longer funding cycles, rather than requiring identifiable outcomes and concrete strategies from the outset [136,163].

In developmental evaluations it is more empowering to evaluands to have a hopeful and solutions-oriented approach [181] (e.g., appreciative inquiry [162]) that focuses on their positives, strengths, competence and potential, and how these can be developed further, rather than problems, weaknesses, and punishment for poor practice, which can reinforce power inequalities and marginalization [101,182] (see Context Principle). While risks, disappointments, failures and weaknesses should not be ignored, they should be acknowledged as valuable opportunities for learning [62,182] (see Learning Principle).

This principle is important for enabling adaptation to dynamically changing conditions [183]. In reality there is rarely a position of stability from which to retrospectively evaluate a project [18], and systems may revert back to old patterns after an intervention, highlighting the need for ongoing monitoring, trajectory analysis, and a focus on long-term results [56,67,136]. Experimental approaches that require more fixed conditions are rarely feasible in complex systems [18]. Demanding specificity and measurability up front, as in linear logic models, can actually be damaging in complex system contexts because it constrains exploration, experimentation and adaptation [180]. Furthermore, the urgency of addressing global crises (see Transformation Principle) means that there is insufficient time to go through traditional sequences of ‘in-depth situation analysis, comprehensive needs assessment, planning, design, implementation and evaluation’ [56]. Instead, these activities must all happen simultaneously, continually and rapidly [56].

Evaluation approaches such as Developmental Evaluation and Reflexive Evaluation take this principle to heart, supporting more rapid, continual feedback for innovation initiatives adapting to dynamic conditions and moving away from the traditional cycles of formative and summative evaluation that are typically ill-suited to such instability [53,85,89] (Table 2). Methods such as rapid cycle testing [65], snapshot surveys [123], the What? So What? Now What? framework [184], actual-ideal comparison [184], Most Significant Change [184] (Box 5), outcome mapping [184] and Williams’ pattern-spotting framework for qualitative data analysis [185] may also assist rapid, dynamic, developmental evaluation approaches. These methods may be complemented by data science technologies such as Big Data to support more real-time and quasi-real-time evaluations [57]. Principles, which underpin Principles-focused Evaluation, can also be vital ‘anchors’ in environments of dynamism, uncertainty and complexity, and are sufficiently flexible to guide action even when contexts vary across space and time [84] (Box 3). The GUIDE framework offers recommendations for formulating high-quality principles that are Guiding, Useful, Inspiring, Developmental and Evaluable [84]. Rapid feedback or learning memos, debriefs, critical incident reviews, and after-action reviews, meanwhile, are useful methods for sharing emerging insights [123], which may require shorter, more user-friendly report formats [179]. In general, a ‘learning by doing’ approach is advocated, using pilot projects, living labs, policy labs and hackathons to encourage experimentation [141,94].

Box 5. Method – Most Significant Change

Most Significant Change (MSC) is a participatory, dialogic method of monitoring and evaluation [161,162]. It involves researchers/evaluators collecting stories (usually via interviews) from those involved in or affected by an intervention, followed by participatory analysis and discussion of these stories to identify the most significant ones for demonstrating the impact of the intervention [161]. Typical MSC questions might be: ‘During the last month, in your opinion, what was the most significant change that took place for participants in the program?’, followed by ‘From among all these significant changes, what do you think was the most significant change of all?’ [161]. MSC questions can also be directed more strongly towards personal experience: for example, an evaluation of youth workers’ experiences in a voluntary sector youth organization asked participants: ‘Looking back over the last month or so, what do you think was the most significant change that occurred for you as a result of coming here?’ [162].

MSC is useful for capturing a rich, qualitative picture of impact and people’s experience, including people’s values, rather than over-simplified quantitative indicators [161] (cf. Values Principle). In this way it can be a valuable complement to more quantitative impact evaluations [161]. It can also be useful for identifying unexpected outcomes and as an iterative, dynamic, developmental approach [161,184] (cf. Complex Systems and Developmental Principles). It aims to support a culture of organizational learning rather than just accountability [161] (cf. Learning Principle), and encourages incorporation of a diversity of views – indeed, the simple approach of collecting stories facilitates communication across different cultures [161] (cf. Justice Principle).

5. Foresight principle: Include foresight alongside retrospective evidence.

Evaluators should become future-sensitive, using and evaluating information from futures and foresight methods alongside evidence from the past and present to guide transformative action [177].

Many evaluations use a retrospective approach – investigating past events to generate recommendations for future action – with the assumption that what works (or not) in the past will work (or not) in the future [177]. However, in VUCA contexts (characterized by volatility, uncertainty, complexity and ambiguity), and to support transformational change, evaluations also need to draw on futures and foresight methods, integrate future-sensitivity into their terms of reference, and judge interventions against the future [177], rather than relying on generalizations from hindsight [4]. For example, a more future-oriented formulation of the evaluation question ‘To what extent did the intervention respond to beneficiaries’ needs as defined in the program documents?’ would be ‘To what extent is the program geared to ensuring that it meets the needs of the stakeholders in the next ten years, even if circumstances change?’ [177]. Without such future-sensitivity, there is a danger that initiatives will be continually caught out by changing contexts, fail to sustain desirable outcomes, and fall back on marginal, incremental change rather than systemic transformation [113,177].

An important role for evaluators is then to expand foresight evaluation capacity [4], extend attention to mid- and long-term futures [177], and support stakeholders in: envisioning desired futures and innovations [113,177]; exploring possible future scenarios, global shifts and megatrends, including how current paths of action will respond to wider system dynamics [78]; and questioning assumptions about the future (see also Values Principle) [177]. At the least, a simple exercise would be to identify existing strategic plans of relevant organizations that interact with the intervention at hand, or major milestones that are due to occur, such as Agenda 2030 expiring in 2030 [177]. To encourage transformational change, it is important for people to envision radically different desired futures [167,170,177]. These futures can then inform theories of change (including expected times for outcomes to occur), and generate rich, contextually adapted and collaboratively developed indicators of change [170,177]. Evaluators might also evaluate the quality (e.g., transformation-orientation) of envisioned futures [63] (see Transformation Principle), and initiatives’ ‘futures literacy’, future preparedness and resilience, making recommendations for how these could be improved [177].

Future-sensitivity is embedded in a number of overarching evaluation approaches, including Culturally Responsive Indigenous Evaluation (CRIE), Gates et al.‘s ethics of evaluation for socio-ecological transformation, Mickwitz et al.’s theory-based framework for evaluating transitions, the Orders of Outcomes framework, Patton’s Evaluation Criteria for Evaluating Transformation, Patton & Felcis’ Principles for Evaluating Transformation, Reflexive Evaluation and Reflexive Monitoring in Action (RMA), Sitra’s Evaluation Framework, and Blue Marble, Design-driven, Feminist, Indigenous Feminist, and Visionary Evaluation (Table 2). This principle also has an affinity to Theory-based Evaluation, in the sense of futures methods results being used to enrich theories of change [113,177]. Various methods and tools help to structure the envisioning or exploration of desired or possible futures. Methods include horizon scanning, scenario building (exemplified by the work of the Intergovernmental Panel on Climate Change [186]), Delphi analysis, backcasting (working backwards from an envisioned future to understand the necessary intermediate actions), wind-tunneling (testing how existing strategies would fare in different futures), and methods for describing the future as if it exists today (e.g., imagining future newspaper headlines) [177]. Risk analysis, Bayesian statistics, and power-law mathematics can also assist prospective assessment of the effectiveness of alternative strategies prior to any implementation [134,187]. Other futures tools and frameworks include Three Horizons [160], Seeds of Good Anthropocenes [188], the Regenerative Lens [15], anticipatory action learning [189], the Value Proposition Toolkit [64], the ParEvo tool [177], and Futures Wheel [177].

Power principles

Four principles related particularly to power relations in transformation-focused evaluation and the responsibilities of evaluators to: promote justice, including through addressing power inequalities, upholding rights, decolonizing evaluation, and honoring diverse and marginalized perspectives; shift power towards evaluation users and blur the evaluator-evaluand dichotomy; support more autonomous evaluation practice; and cultivate mutualistic partnerships of knowledge and action.

6. Justice principle: Promote justice in power relations, rights, and use of knowledge.

Evaluators should promote justice as a key component of transformations, including through addressing power inequalities, upholding human and non-human rights, decolonizing evaluation, and honoring diverse and marginalized perspectives [136,89,100,106,108,190].

Evaluators can and should be ‘activists’ for democracy and justice [100,108], including: overcoming structural injustices in knowledge systems [191]; enabling self-determination of Indigenous peoples and upholding territorial rights [106]; supporting LGBTQ+ health and equity [111], intersectional justice [111] and human rights in general [175]; and promoting the rights of Nature and other sentient beings, avoiding anthropocentrism [89,192]. As well as applying this principle in stakeholder engagement, evaluators should see evaluation results as opportunities to lobby for justice [136], and local voluntary organizations for professional evaluators could support advocacy and social action [171]. This is a fundamental shift from the more neutral, independent stance traditionally adopted by many evaluators [4].

This principle requires that evaluators become aware of power relations, cultural differences, histories and underlying ontologies within the communities and societies of interest, including how these uphold the status quo and whether unjust power structures have been disrupted [136,190]. It also requires evaluator reflexivity (see Values Principle) to recognize the power implications of any evaluation approaches advocated and what forms of knowledge they treat as valid [111]. This is particularly pertinent to the Western evaluation field, its paradigms and its history. For instance, conventional Western evaluation paradigms are founded on historical violence against Indigenous communities, with evaluations performed in extractive ways without participants’ knowledge or consent, and often used to validate Western scientific methods whilst denigrating Indigenous ways of knowing [109]. Even supposedly neutral evaluation criteria may actually impose White values and knowledge [193], and have often been used as an excuse to ignore and perpetuate violence against Indigenous peoples [109]. Modern algorithms and technologies may be inherently racist in their formulation [194]. Western evaluation has also been seen to grow from White patriarchal logics [136].

Conversely, an important aspect of justice-oriented evaluations is to honor a diversity of perspectives and ways of knowing and evaluating [106,190], giving a voice to the most discriminated and marginalized groups of people (e.g., considering diversity of race, gender and ability) [113,190] and learning particularly from Indigenous, Global South and Eastern philosophies [106,194]. This is a particular shift from the domination of evaluation approaches from the West and Global North [194], and acknowledges, contrary to many Western beliefs, that there is no absolute, objective scientific truth that defines what appropriate evaluation is [111]. For instance, it may be important to: elevate unspoken felt-knowledge and embodied experience as a valid form of knowledge and truth [109]; include place-based ecological knowledge and ancestral wisdom communicated via dreams or intergenerational traditions [106]; ensure meaningful participation of youth [195], LGBTQ+ people [111] and those in poverty [191]; and include views of laypeople or those with on-the-ground experience of real-world problems alongside other ‘expert’ opinion [172,191]. This principle therefore requires evaluators to work in multi-actor, multi-disciplinary teams [113,187], remaining open to new ideas, willing to learn and continuously identifying opportunities for enhancing their own learning [111,101] (see Values and Learning Principles). There have additionally been calls for more South-South cooperation in evaluations to resist the dominance of Global North evaluation approaches [34].

Ethical dilemmas are likely to arise in justice-oriented evaluations, and evaluators should be transparent about these and how they were addressed [136], potentially using Dilemma Resolution techniques [78,196]. More appropriate evaluation questions than ‘Does the intervention work?’ would accordingly be ‘What works for whom, in what ways, under what conditions and with what results?’, because these questions recognize that one intervention will not work for everybody [56]. Indeed, dilemmas and contradictions can become valuable sources of transformative insight and learning if approached with the right attitude [165]. By remaining sensitive, reflexive and transparent in these ways helps evaluators to develop trusting relationships with stakeholders to unearth their perspectives and experiences [129] (see Values Principle).

This principle is important because failing to attend to local context and its power relations, and imposing externally developed evaluation approaches, can simply reproduce problematic power inequalities that hinder transformation [136]. Embracing this principle instead deepens appreciation of systemic complexity and interdependence [135,181], enables more meaningful judgements of the outcomes and impacts of transformation [67,102], encourages greater resilience through the ‘requisite variety’ in evaluation design and processes [134,106], and can generate more powerful learning between people with different perspectives [182]. As a particular example, including voices of young people is important because they tend to have more transformative values (e.g., fighting for social and environmental justice, peace, etc.), are good at building networks and alliances, have strong communication and technological savviness and connectivity, and offer fresh perspectives [195]. The Justice Principle is also an ethical imperative in itself [78].

Operationalizing this principle means applying evaluation approaches that are sensitive to cultural context, inequalities (including those related to race, gender, socio-economic status, sexual orientation, species, etc.) and conflict [34,89,111,122]. Transformative Evaluation has an explicit focus on promoting social justice, challenging power hierarchies in society and working in an inclusive and empowering way with minority and marginalized communities [54]. The Navigating the Intersection of Culture and Evaluation (NICE) Framework, Gates et al.’s ethics of evaluation for socio-ecological transformation, Prague Declaration, and Transformative Empowerment Evaluation, Transformative Participatory Evaluation, Culturally Responsive Evaluation, Equitable Evaluation, Equity-focused Developmental Evaluation, and Inclusive Evaluation, adopt similar stances (Table 2).

Approaches addressing more specific arenas of (in)justice include: Indigenous Evaluation and the Nation-to-Nation (N2N) Systems Evaluation Framework, which focus on elevating and empowering the knowledge and cultures of Indigenous peoples and First/Tribal Nations (Box 6); Feminist Evaluation, focused on advancing gender-based justice; LGBTQ+ Evaluation, which promotes LGBTQ+ inclusion and liberation; Aotearoa/New Zealand’s evaluation guidelines, which explicitly reference care toward the environment; and Made in Africa Evaluation, which aims to decolonize evaluation in Africa and support locally developed African evaluation approaches that promote African livelihoods and values (Table 2). The African Evaluation Principles similarly call on evaluations to empower Africans, address inequalities and power asymmetries, and embrace a diversity of perspectives particularly from the Global South and Indigenous cultures [88].

Box 6. Case study – Indigenous Feminist Evaluation in the ‘My Two Aunties’ program

Established in 2019, the ‘My Two Aunties’ (M2A) program is a home-visitation program for Indigenous communities in southern California run by Tribal Family Services (a department of the Indian Health Council) [109]. ‘Aunties’ traditionally provide mentorship for healthy families in these communities, and the program aims to decolonize and destigmatize social services by drawing on these Indigenous feminist and kinship traditions, thus combatting ‘settler colonial imposition of patriarchy and violence against Indigenous women and families’ [109]. The aunties also acted as evaluators of the program [109]. Accordingly, an Indigenous Feminist Evaluation approach was used [109]. This approach:

  • is rooted in a relational worldview, where ‘worlds are made up of extended networks of kin to whom we have reciprocal responsibilities’ (cf. Complex Systems Principle);
  • respects tribally-specific teachings and ‘land-based ontological pluralism’ rather than imposing single truths and approaches (cf. Context, Detective and Justice Principles);
  • centers Indigenous ways of knowing, including unspoken felt-knowledge and embodied experience, e.g., in the personal reflection surveys of parents conducted after each home visit (cf. Justice Principle);
  • respect the data sovereignty, privacy, and specific knowledge-exchange protocols of communities [109] (cf. Partnerships Principle);
  • is reflexive (e.g., naming the violence associated with Western evaluation) (cf. Justice and Values Principles);
  • is explicitly normative, aiming to have a transformative impact that fosters just futures for coming generations far into the future [109] (cf. Transformation Principle).

In this way the ‘My Two Aunties’ evaluation embodied many of the Complexity, Power and Purpose Principles presented in this paper, and particularly emphasizes the Justice Principle by promoting the decolonization of evaluation.

7. Power Shift Principle: Shift power towards evaluation users and blur the evaluator-evaluand dichotomy.

Evaluators should aim to be guides rather than directors or controllers of transformation-focused evaluations: evaluations should thus be genuinely participatory and empowering for evaluation users [101], allowing them to take greater ownership, leadership and responsibility of designing and delivering evaluations, and acting on their findings [136,121,130].

This principle sees participation as going beyond tokenistic consultation and instead involving a genuine transfer of power towards users of evaluations [101,130]. These stakeholders are not simply sources of data [136] and evaluators not mere data collectors [162]. Rather, users should be involved in deliberative discussions throughout the process of designing, conducting and drawing conclusions and recommendations from evaluations [122,197], and trusted as ultimate authorities on what is appropriate for their context, including the purpose of evaluation [122], evaluation questions [122], which data to collect [136], and what criteria and indicators to evaluate against [89]. Moreover, evaluation outputs (e.g., reports) should be disseminated to stakeholders with the agency to act on the findings, and not only to funders [35]. Indeed, evaluation data should be treated as autonomously owned by evaluated communities [111].

This principle signifies a fundamental shift in the role evaluators play [106], and blurs the traditional dichotomy between evaluator and evaluand [78]. Transformation-focused evaluators might be better described as participant observers [130], facilitators [136], change agents [4], ‘active design team member[s], fully participating in decisions’ [179], critical friends [179], or partners with other stakeholders [121] – rather than primary experts imposing their own view of reality and evaluation approaches on communities without consultation [101,130], and conducting top-down evaluations that externally judge evaluands without providing opportunities for development and reflection [198]. Rather, they are guided by critique and deliberation with evaluation users [106], and listen to voices that disagree with the perspectives of the evaluator or funder [179]. Evaluators may not even be the primary evaluators, but rather supporting communities to critically evaluate their own progress [121], building capacity for evaluative thinking and evaluation utilization in evaluated systems [171]. Evaluators may require training in developing these kinds of facilitation, communication and interpersonal skills (including conflict resolution), with which evaluators can help communities access their untapped collective wisdom and creativity, and nurture conditions of trust and inclusivity where insights can be safely and willingly shared [172,176]. Overall, however, evaluators may have to move fluidly between more ‘internal’ roles (invested in the system) and ‘external’ roles (independent from the system), and remain able to adopt the stance of a more external, independent, critical expert if this is considered most appropriate for the case in hand [100], such as where there is a risk of organizations cherry-picking results or greenwashing [78] (see Autonomous Evaluation and Values Principles).

Examples of this role reconfiguration include genuine youth-participatory evaluation, in which evaluation is not something ‘done to’ young people but ‘with or by’ young people, potentially with young people as co-evaluators or directors of evaluation and adult evaluators as coaches or facilitators [195]. In education, there have been calls for greater participation of students in self-assessment, self-reflection and peer assessment rather than didactic teacher-led approaches, disrupting the traditional power asymmetry [169].

This principle is important for a number of reasons. Overall, it embodies the kinds of ethically imperative power shifts sought in the transformation of wider social-ecological systems [190] (see Justice Principle). Shifting power towards the evaluand can deepen participants’ understanding of evaluation [136], and by building evaluations around dialogue in community settings, people can learn from each other, become more aware of social forces influencing their beliefs and behavior, and learn the importance of giving others feedback [182] (see Learning Principle). More technocratic evaluations – in which the methodology, questions and analysis are ultimately designed and imposed by the expert evaluator research team – may lead evaluators to ask the wrong questions and overlook important aspects, such as political dimensions [191]. Adopting this principle would also encourage commitment to act on evaluation results [136] and increase the likelihood that transformation-focused evaluation would become an integral, self-sustaining part of working cultures rather than a periodic externally applied exercise [101] (see Learning Principle).

Operationalizing this principle requires evaluators to build trust-based relationships with stakeholders in culturally appropriate ways that are context-aware [136], acknowledge and challenge power imbalances [190] (e.g., incorporating voices of the most marginalized [199] – see Justice Principle), and encourage deliberation about different values [89] (see Values Principle). A wide variety of evaluation approaches are built around greater participation and empowerment: as well as all those listed under the Justice Principle, they include Adversary, Collaborative, Deliberative, Democratic, Design-driven, Developmental, Participatory, Real-time and Reflexive Evaluation (Table 2). Empowerment Evaluation aims to enhance communities’ capacity to self-evaluate [101,200]. Utilization-focused Evaluation, meanwhile, encourages working with evaluation users in participatory ways that foster a sense of ownership by the users as they shape the evaluation approach [131]. Broadly speaking, participatory action research is often an appropriate method or framing for these kinds of participatory evaluations [197], and some analysis methods lend themselves well to dialogic and participatory contexts, such as Most Significant Change [162] (Box 5).

8. Autonomous evaluation principle: Support autonomous evaluation practice.

Evaluators should resist evaluation contracts and designs dictated principally by markets and individuals or organizations with vested interests and power in maintaining the unsustainable status quo [25,135]. Instead, evaluators should support stronger professionalization and autonomy of evaluation practice, which includes adopting clearer ethical and critical stances and permission to speak frankly in support of transformational change [25,146].

This principle is important because evaluations led by those in positions of power may be insufficiently critical (e.g., policy evaluations led by the associated government ministry) [19]. If evaluations are driven by globalized markets and vested interests in traditional kinds of economic growth, meanwhile, economic impacts (e.g., ‘value for money’ or ‘social return on investment’) will typically dominate the focus of evaluations, with less attention given to social and environmental considerations [25,135,139], and the evaluation may impose more homogenizing criteria that ignore individual uniqueness [166] (see Context Principle). For example, evaluation criteria for development are often based on industrial and technological paradigms focused on economic efficiency [181], and education evaluations underpinned by neoliberal thinking are product-focused, assessing student merit based on their ability to compete, whilst ignoring sociocultural diversity [166]. By supporting more autonomous and professionalized evaluation practice, however, evaluators could be more selective in choosing to conduct evaluations that align with ethical stances of supporting transformational change [25].

Operationalizing this principle will require collective action from evaluators to develop the ‘features of professionalism’, emphasized in the Prague Declaration [122], including ‘an ethical charter, expert knowledge, proven competencies and self-management’ [25]. As for particular evaluation approaches supporting this principle, Democratic Evaluation is notable for purposely avoiding excessive manipulation of evaluations by government agencies and the ‘information monopoly’ of program managers [98]. Adversary Evaluation has a goal of resisting neoliberal domination of evaluation and contributing to public good [87], whilst the African Evaluation Principles urge evaluators to guard against vested interests [88].

9. Partnerships principle: Cultivate mutualistic partnerships of knowledge and action.

Evaluators should cultivate mutualistic partnerships of knowledge generation and sharing when working together towards transformations, built upon strong ethical and professional codes of conduct [135,122]. They should also encourage and evaluate alignment and cooperation of different transformation efforts [78].

Partnerships of knowledge generation and sharing should promote mutual trust, shared values, transparency, honesty (including about mistakes and limitations), fairness, inclusivity, and shared responsibility for evaluation results [122]. The principle of ukama or ‘relatedness’ from African Ubuntu philosophy has been suggested as a foundation for nurturing solidarity and complementarity between Global North and South [165]. Partnerships should respect actors’ different responsibilities and role boundaries, and avoid letting egos and personal agendas dominate [163]. They should exist not only within evaluation teams, or with commissioners and users of evaluations, and other local stakeholders (including Indigenous peoples) – but also with other evaluators and national or international associations of evaluators, researchers, junior evaluators, students, interns, and other young people [135,122]. To develop strong, lasting relationships, it may be important to engage with stakeholders in person ‘in the field’ rather than only online engagement [122]. Another important aspect of this principle is that evaluators should encourage and assess alignment and cooperation of different transformation efforts [78]. This principle means taking care to avoid: developing only short-term, superficial relationships with partners [122]; failure to maximize accessibility of evaluation results [122]; and repetition, redundancy or incongruence of separate transformation initiatives [78]. Evaluation data are often considered proprietary, resulting in inadequate sharing across the field and therefore lack of coordination and integration [179].

Transparent and respectful knowledge exchange is a key aspect of this principle. Evaluators should devote intentional time and resources for sharing evaluation findings [179], and doing so in an honest and transparent way [146]. International fora and conferences provide suitable spaces for evaluators to share evaluation results and experiences with other evaluators, researchers and policy professionals [146], as well as writing articles [201]. Importantly, evaluators should make a particular effort to share their knowledge with actors working to transform systems [202], and with experts across different disciplines [78]. Evaluation reports should aim to be open-source, engaging, attractive and easily accessible, with clear, transparent and context-sensitive language that avoids unnecessary jargon [179,122]. Transparency means being honest about mistakes, failures and unintended consequences, and how theories of change evolved over time, for instance [179]. At the same time, evaluators should respect the unique protocols around knowledge-sharing and caretaking responsibilities around knowledge in particular (e.g., Indigenous) communities [109]. Overall, evaluators should reflect on whether they are being as effective as possible in helping to disseminate findings, influence others, and ultimately assist transformation [106,179,197].

This principle is important for numerous reasons. Inter- or transdisciplinary, multi-sector and multi-stakeholder approaches and partnerships are necessary for working with wicked problems [64,191,203]. Partnerships can provide a greater diversity of ideas and skills and encourage shared learning: for instance, evaluators could work productively with futurists and modelers to develop more powerful scenario planning (e.g., discussing what trends to pay attention to for the future) [56], and senior evaluators could act as mentors for junior evaluators, students and interns [122]. Indeed, the necessary shift to multiple and bricolaged methods (see Context and Detective Principles) means that having mixed teams of evaluators with different types of expertise – including newcomers, ‘outsiders’ and amateurs – may be critical (because one evaluator is unlikely to know every method) [57,122]. Greater knowledge-sharing by evaluators could furthermore enhance learning and innovation in the evaluation field as a whole [179]. Stronger coordination between different evaluations and ‘cross-case learning’ would enable them to complement each other in providing a more holistic understanding of a system [113,141], provide insights about how initiatives could be deepened, broadened and scaled up [204], and support more alignment and cooperation of initiatives to facilitate system change [78]. Evaluators should bear in mind that evaluation results may be used by new or previously unestablished actors in transitions, such as new NGOs or cooperatives, emphasizing the importance of making results transparent and accessible [113]. Finally, adopting these partnership values could encourage the implementation of evaluation recommendations [122].

While there are no specific evaluation methods to guide these forms of wider partnership-building per se, many of the evaluation approaches that promote decolonization, justice, and genuine participation (see Justice and Power Shift Principles), such as Transformative Evaluation, embed the building of trusting relationships with evaluation users (Table 2). Furthermore, the ‘Yin Yang Principle’ of Blue Marble Evaluation encourages engaging with different perspectives, harmonizing conceptual polarities and divisions, and transcending boundaries [90]. The African Evaluation Principles encourage a focus on developing relationships of reciprocity and mutual accountability [88], whilst the Prague Declaration also emphasizes the importance of mutualistic partnerships [122]. Utilization-focused Evaluation makes a point of encouraging wider dissemination of significant evaluation findings to expand influence [131].

Purpose principles

Three principles were considered to relate to the deeper purpose and values of evaluation in transformation contexts and were collectively called the Purpose Principles. They encourage evaluators to surface, question and uphold values, foster evaluation cultures centered around ongoing learning, and ultimately focus on evaluating and supporting systemic transformations towards regenerative futures.

10. Values principle: Surface, question and uphold values.

Evaluators should work with evaluands to surface and reflexively question the underlying values at play in transformation processes [89,101,205,206], and ensure that transformational values are upheld [136].

There are several dimensions to this principle. The first is that evaluators should surface, evaluate, and facilitate discussions about subjective aspects of evaluands, including their underlying values, worldviews and norms that shape behavior [89,101,205]. Evaluation does not take place in a value-free vacuum – there are always political, economic, cultural, and other value-laden forces at play [136]. For instance, different stakeholders will have different perspectives on where a system’s boundary lies [205], there will be plural values held for nature [89], desired system outcomes and evaluation results are likely to be contested [136,106], and political influence or vested interests may sway project outcomes [113]. As in the Context and Justice Principles, evaluators are therefore obliged to take into account the memories, heritage, and associated values of particular places, events, rituals, languages, arts and culture [132], and integrate the intellectual and emotive aspects of evaluation [168]. Indeed, evaluators should recognize the importance of people’s emotional and spiritual development during an evaluation as well as more objective dimensions, like gaining knowledge [182]. This requires a warm, nurturing social environment to encourage respect for other people’s beliefs and values [182].

The second dimension is the importance of reflexivity and questioning underlying values, assumptions, beliefs and worldviews. Evaluators should encourage critical reflection amongst stakeholders and help them to understand their own positionality [206], asking important and difficult reflexive questions even if the answers raise uncomfortable discussions [176], and potentially playing a ‘brokering’ role in mediating conflicts of values [89]. This may include questioning the acceptance of contextual aspects of evaluations (e.g., the policy strategies behind the policy impacts being evaluated, or the fundamental goals of a program or initiative more generally) as fixed and appropriate, whereas in reality they may have flawed assumptions that affect downstream actions [113,137]. Similarly, evaluators should beware blindly applying existing indicators to local contexts without reflecting on whether they truly are measures of success [136] (see Context Principle).

Equally, evaluators should recognize that they themselves are not value-free [25] and never perfectly neutral or objective [4], and reflect on how their own worldviews and values shape evaluations [106] (see Justice Principle). The worldview evaluators carry influences ontology, epistemology and axiology, and thus inevitably the way they evaluate [136]; evaluators see ‘through the eyes of [their] own history and environment’ such that their ‘view of reality is always partial and slanted’ [136]. Similarly, evaluators are never fully ‘external’ to and independent from evaluands – a second-order approach that recognizes evaluators’ embeddedness within the system being evaluated may be more appropriate [106,101]. A ‘good internal praxis’ [165] of reflexivity in evaluation teams would include articulating the assumptions guiding an evaluation and whose they are [132], transparency about power relations, learning from their own stories of success and failure, openness about their fears and vulnerabilities, and sharing this learning with others [165].

The third dimension is that transformation-focused evaluators should uphold and remain true to their own transformational values, whilst appreciating the many different perspectives that people have on what transformation means [78]. They should take an ethical stance of supporting positive transformation (see Transformation Principle), justice (see Justice Principle), and the public good [136], and integrate their ethical codes into evaluation activities, relations, organizations and training [171,122] (see Autonomous Evaluation Principle).

This principle involves a fundamental shift from evaluations built on a positivist worldview that assumes objectivity and neutrality of evaluators [136], leading to evaluations that typically focus on quantitative facts [101], omit important value-laden political and cultural dimensions [166], and lack reflexivity about their own values and approaches [106]. For example, much evaluation of education tends not to be reflexive about why something is learned [169]. While they may be well-suited to some contexts (e.g., evaluating medical interventions), more positivist, quantitative evaluations are often insufficient for the social contexts of transformation [78]. The principle is based on a more constructivist view of people as ‘socio-historical subject[s]’ with strong political-ethical dimensions [169].

There are various reasons why this principle is important to adopt. For one, it brings evaluations closer to the social reality of transformations [78]. Values are inextricably linked to more objective or quantitative aspects examined in evaluations [101]. Ignoring the underlying values and beliefs of participants – the ‘elephants in the room’ – could perpetuate a ‘culture of silence’ that simply reproduces a problematic status quo [101]. Shifts in social norms, values, beliefs, assumptions, and purposes are also powerful indicators of the depth of transformation [121]. Self-reflection by evaluators, meanwhile, requires humility, which disrupts the conventional power dynamic of the evaluator as directive expert [122] (see Power Shift Principle). Lastly, understanding what inhabitants value about their locale has been noted as encouraging regenerative mindsets and action, because these values can then be incorporated into manifesting the place’s potential [147].

Numerous evaluation approaches have an explicit focus on surfacing and reflexivity around people’s values, including: Adversary Evaluation; Deliberative Democratic Evaluation; Developmental Evaluation and especially Equity-focused Developmental Evaluation; Made in Africa Evaluation; the NICE Framework; and Principles-focused, Systemic, Transformative, Value(s)-based, and Visionary Evaluation (Table 2). In Reflexive Evaluation or Reflexive Monitoring in Action, evaluators act as ‘sparring partners’ [53] who continuously encourage innovation teams and networks to challenge underlying assumptions and systemic patterns in designing and managing transition initiatives [53,125]. Empowerment Evaluation and Feminist Evaluation encourage critical thinking particularly about power relations and prescribed roles [100,101]. Other tools and methods that help to elicit values include the Three Horizons framework, the Value Proposition Toolkit and Causal Layered Analysis (e.g., in encouraging participants to imagine radically different desired futures and the contrasting values underpinning them [64,160,170] – see also Foresight Principle), the Theory U framework [67], the Four Levels of Conversation framework [67], and the Most Significant Change method [207] (Box 5).

11. Learning principle: Develop evaluation cultures centered around ongoing learning.

Evaluators should develop evaluation cultures centered around ongoing learning during transformation processes [100,201], rather than driven by accountability requirements for showing positive results [139,162,191,114], and fear of failure [78].

Many evaluation contracts and designs, both today and historically, have been driven by requirements of accountability for the achievement of specific predetermined objectives, compliance with regulation, and the need to show positive results for funders and governments, rather than accountability for learning [139,162,191,114]. In international cooperation, for example, Global South countries are often required to present proof of a well-implemented project to donors, which satisfies donors’ accountability needs but not necessarily the learning needs of the country being evaluated [24]. This focus has given evaluation negative connotations of program problems, punishment and control [136], and created a fear of failure [78]. It also encourages simplistic, linear, ‘value-free’ designs and framings of interventions leading to particular outcomes, with limited timescales tied to funding cycles [139,208].

Transformation-focused evaluators should try to support a cultural shift in the meaning of accountability and purpose of evaluation. Accountability should instead be focused on things like alignment to principles and values, or ensuring rigorous and systematic learning processes and consequent adaptation rather than having particular impacts in complex systems [67,114]. Indeed, strategic learning may be the only outcome in system change that social innovators and evaluators can truly control [67]. A future-sensitive approach (see Foresight Principle) additionally suggests that initiatives should be accountable to future as well as present or past generations [78,177]. This then potentially shifts evaluation from being seen by evaluands as a punitive exercise and primarily for the benefit of funders and regulators, toward seeing the value of open, honest reporting and learning, and evaluation’s benefits for the people being evaluated, and beyond [100,162]. Organizations will need to be able to review negative feedback without defensiveness [201], be prepared to admit that original objectives were not met [179], and confront contradictions or tensions (e.g., between Global North and South perspectives and positionality), recognizing these outcomes as powerful opportunities for learning [165].

Overall, this shift means that adaptive learning for facilitating positive transformations becomes the central purpose of evaluation [139] (see also Developmental and Transformation Principles). Evaluators should advocate the value of evaluation to this end, especially persuading people in power [19], and helping others overcome associations of fear and blame around evaluations [201]. Avoiding the term ‘evaluation’ altogether may be useful in the initial stages of attracting participants in some evaluations [162] (instead using language like ‘review’ or ‘consultation’ [209]). Having stakeholder ‘champions’, who believe in the value of evaluation and are enthusiastic to take results to those with the power to implement change, is also useful [136]. Creating an evaluation culture centered around learning requires further steps to embed this mode of evaluation. For instance, evaluation should ideally become a core activity of all teams, rather than performed by just one person or unit [201]; moreover, longer time horizons are needed, since transformation should never be seen as ‘done’ and instead be understood as a continual learning process [121] (see Developmental Principle).

The real result of an evaluation, after all, is the learning from its process and findings and how this steers action [78,122]. Learning takes many forms, and takes place at many levels and at all stages of an initiative or transition (not just at the start or end) [210]. In transitions there is also a need to ‘unlearn’ old patterns and ways of doing things as well as learn new ones [210]. Some have distinguished between single-, double- and triple-loop learning [211]. ‘Single-loop learning’ (i.e., problem-detection-correction) is limited to asking whether intended outcomes have been achieved, and monitoring predetermined performance indicators [152] – in other words, ‘are we doing things right?’ [211,212]. Double-loop learning, meanwhile, involves examining the causal mechanisms underlying observed change, asking ‘are we doing the right things?’ [211,212]. Triple-loop learning goes even further, exploring and questioning underlying assumptions, values and beliefs, and asking ‘how do we decide what’s right?’ [104,211,212]. Reflexivity (see Values Principle) is therefore key for deeper learning to occur. To support learning towards system innovation, it is important not just to encourage reflexivity but also to monitor and record reflexivity changes and learning journeys (as in Reflective Monitoring in Action) – i.e., collective reflection on an initiative’s reflexivity by the actors involved will in turn help to increase the initiative’s reflexivity [213].

Many different frameworks, in addition to single-, double- and triple-loop learning [211,104], help to conceptualize and facilitate more transformational forms of learning. In Freirean pedagogy and the related Telessala Methodology, students ‘learn by doing’ and become empowered, reflective citizens, rather than unquestioningly absorbing information delivered didactically [101,182]. The concepts of experiential learning [171,206] and design-based learning [94] also emphasize a learning-by-doing approach. In generative learning, learning is based on integration of new and pre-existing knowledge in creative ways [162]. There are also ideas about learning being shaped by social and contextual interaction, such as social learning [191], the Critically Reflective Work Behavior framework (used particularly for social learning in businesses) [94], and the socio-cultural perspective on learning, whereby learning is understood to originate from social and material interaction and is shaped by cultural, historical and relational factors, such that learning involves the ‘whole person’ [43]. ‘Transformative learning’ shares features of triple-loop learning, experiential and social learning; it is action-oriented and includes critical reflection of underlying assumptions and beliefs and how these are shaped by context [171]. In futures research, ‘anticipatory action learning’ is used to reflexively interrogate possible and desired futures and how they could be brought about [189] (see Foresight Principle). While learning traditions including collaborative learning (from educational sciences), organizational learning (from management studies), and interactive learning in the learning economy (from institutional economics) have value for understanding learning in system change processes, they are not yet sufficient for working with complexity [210]. Hence, novel theories about complexity-informed learning have emerged from transformation-focused evaluation practices, such as system learning, whereby actors learn to redefine system barriers as opportunities and are thus inspired to design activities contributing to systemic change [213].

Many overarching evaluation approaches assimilate these ideas about learning, including Blue Marble, Design-driven, Developmental, Dynamic, Empowerment, Equitable, Participatory, Real-time, Reflexive, Transformative and Visionary Evaluation (Table 2). For instance: Transformative Evaluation draws on transformative learning theory [199]; Reflexive Monitoring in Action incorporates ideas about system learning [213]; and Empowerment Evaluation aligns with Freirean principles [101]. Several evaluation frameworks for evaluating complex systems change (e.g., those of Cabaj, Hargreaves, and Preskill et al.) also aim to create cultures of learning (Table 2). A number of other data collection and analysis methods are also geared towards supporting transformational learning, such as authentic dialogue [101], After Action Reviews [67], and appreciative inquiry [64]. Failure Reports and Intelligent Failure Learning Loops are structured processes to encourage admission and understanding of ‘failures’ in interventions and learn from them [67]. More generally, interviews and focus groups are important data collection methods for enabling reflection [123].

12. Transformation principle: Focus on evaluating and supporting systemic transformations towards regenerative futures.

Evaluators should facilitate systemic transformation towards regenerative futures, remembering the urgency for global transformation in a ‘time-is-running-out world’ [4,26], and therefore focus on evaluating transformations and integrate transformational principles across different facets of evaluations [63].

Whilst the nature of transformation is inevitably context-dependent, transformation should be understood as getting to deeper causes of problems, rather than incremental, superficial or reformist change that perpetuates problematic current systems [4,136]. Evaluators should aim to assist transformations to regenerative, resilient futures, which feature ‘adaptive sustainability’ [4], social justice [190] (see Justice Principle), and the flourishing of life [106]. Evaluation can support these kinds of transformations in multiple ways, including by understanding the implications and impacts of ongoing transformations, as well as providing developmental feedback for purposeful interventions in systems change [19]. This principle is vital, and ethically imperative, because the urgency of current crises means that fundamental change is needed across societies to safeguard the survival and thriving of life [26,113].

This principle shifts evaluation criteria. Many past evaluations have been geared towards increasing our understanding about current systems and measuring problems rather than evaluating transformative principles, visions and action [4,172,214]; while these have produced much useful information, they have tended to support only incremental improvement [113] which at best slows the ‘incremental destruction of life’ [172]. Such evaluations have tended to focus on criteria of economic growth and productivity [215], effectiveness in achieving micro-scale outcomes [24], and whether money, time and effort were spent efficiently [24], and give insufficient attention to environmental impacts [135]. Transformation-focused evaluation criteria, on the other hand, would aim to encourage transformative action [214] and regenerative thinking [216]. They would need to link interventions to system-level impacts [86] (such as how underlying structures, patterns and mindsets of systems are shifting [64,122]), developing and continually testing theories of system change [64]. Transformation-focused evaluation criteria would assess: visions of futures and transformation trajectories, and whether these are meaningful and genuinely oriented towards system change rather than more isolated or incremental change [63,113] (see Foresight Principle); experimentation [113]; the alignment and momentum amongst different actors, networks, institutions, ideas, initiatives and movements towards transformation [4] (see Partnerships Principle); and whether transformation rhetoric is backed up by appropriate action rather than greenwashing [4]. They would move away from assessing isolated pillars of sustainability (e.g., environmental, social, economic) towards assessing interactions between different dimensions, including co-evolutionary relationships between human systems and wider nature [217]. Transformation-focused evaluation would also assess the health of social-ecological systems over multiple generations into the future [109,215], and measure how conditions conducive to life are being created or maintained in a place, so people can see their role in increasing a place’s vitality or quality of life [172].

This principle also underscores the imperative to fundamentally redesign evaluation practice, as laid out in all the other principles, and integrate these transformational principles across different facets of evaluations [19,63]. This is a major paradigmatic shift, and not merely tweaking evaluation, like a greater reliance on technology as the evaluation field adapted to the COVID-19 pandemic [19]. The principle particularly highlights a shift from the traditional neutrality of evaluation, as it recognizes that evaluators are not independent of the system being evaluated (see Values Principle); rather, they have a personal interest in successful transformation (‘skin in the game’) [4]. There have been calls for evaluators to instead become more vocal communicators and advocates for transformation [146]. Evaluations are thus seen as part of transformation processes, not removed from them – i.e., seeing evaluation as system change, not just for or about system change [4,64]. Evaluators may have to be pickier about the contracts they choose to accept, refusing those that are not sufficiently geared towards transformation or restrict the evaluator’s ability to encourage more transformational approaches [78] (see Autonomous Evaluation Principle).

Numerous overarching evaluation approaches explicitly support transformational change as an ethical imperative. These include Blue Marble, Dynamic, Feminist, Footprint, Transformative, and Visionary Evaluation, Transformative Empowerment Evaluation, Transformative Participatory Evaluation, the African Evaluation Principles, Gates et al.‘s ethics of evaluation for socio-ecological transformation, and the Prague Declaration (Table 2). Many other approaches and frameworks, while not necessarily taking such a strong ethical stance, nonetheless focus on assessing transformational, systemic change, including Cabaj’s Inquiry Framework for Evaluating Systems Change, Hargreaves’ framework for evaluating system change, Latham’s Systems Change Evaluation framework, Mickwitz et al.’s theory-based framework for evaluating transitions, the Orders of Outcomes framework, Patton’s Evaluation Criteria for Evaluating Transformation, Patton & Felcis’ Principles for Evaluating Transformation, and the TIPC methodology (Table 2). The Systems Change Evaluation Canvas tool can aid reflection when designing a systems change evaluation [218].

Various theories and frameworks help to conceptualize transformations and other forms of change in complex social-ecological or socio-technical systems. These include transition theory and regime shifts [113,152,219], panarchy and adaptive cycles [152], Meadows’ leverage points and the related Intervention Level Framework [32,67,102], the Action Scales Model [102], the World Mandala [78,156,157] and Ambition Loops [78,220,221] (particularly for identifying synergistic actions), actor network theory [113], innovation systems theory [113], the multi-level perspective [113], the Stewards’ Pathway (a developmental theory of system change) [64], Kania et al.’s system change framework [64], and ‘directionality’ of system change [138,152]. The Signals of Transformational Change and related Four Dimensions of Transformational Change frameworks make an important shift from only tracking limited, specific quantitative metrics and indicators to tracking broader signals of systemic shifts at different stages of progress [62,222]. It is recommended to present evidence of contributions to system changes (e.g., changes in an initiative’s system maps over time and/or more narrative-based ‘contribution stories’) to different stakeholders to obtain their interpretations, because these will vary depending on whose perspective is taken into account [102]. This is supported by tools such as SenseMaker, based on the Cynefin framework from management science, which helps to obtain micro-narratives from the system context, drawn from people’s experiences, reflections, perceptions and attitudes, and make sense of emergent patterns from those narratives [102,223]. The Most Significant Change method can also help to identify anecdotal evidence of transformative impacts [206] (Box 5). Other frameworks such as Three Horizons, the Six Pillars framework, Causal Layered Analysis and the Theory U framework also help to convene conversations about transformational versus reformist change, and structure exercises of visioning radically different futures [67,78,170] (see Foresight Principle).

Various disciplines, frameworks and tools exist for focusing evaluation on regenerative systems and futures. Fields such as regenerative design, regenerative development and regenerative sustainability, as well as many Indigenous frameworks and philosophies, already embrace many of the principles laid out in this paper in support of the flourishing of life – such as considering change as more unpredictable and emergent, recognizing unique locales, embracing reflexivity and co-creation, and seeing the evaluand as part of a wider social-ecological system [15,172,224]. Regenerative design and development tools, such as the Regenerative Community Development Evaluation Tool [216], Regenerative Capacity Index [225], and LENSES framework [167], emphasize a shift from focusing on evaluating outcomes (e.g., of sustainable buildings) to evaluating processes of design and building community regenerative capacity [172,224], avoiding generic top-down forms of assessment [224]. Various other tools encourage ambition, imagination and regenerative thinking in exercises of envisioning desired and possible futures, such as the Regenerative Lens [15] or Seeds of Good Anthropocenes [188].

Discussion

This paper presents the first systematized review of literature on evaluation in transformation contexts, and provides the most comprehensive set of guiding principles and relevant methodologies for transformation-focused evaluation to date. The principles (Table 1; Fig 3) highlight the importance of transforming evaluation’s complexity awareness, its power relations, and its purpose and values, if evaluation is to assist wider transformations in societies. They call for transformation-focused evaluators to appreciate evaluands (e.g., a transition initiative, sector, or region) as unique, entangled, nested, dynamic and uncertain systems, and employ more diverse, developmental, contextually adapted and future-sensitive evaluation approaches. The principles also urge evaluators to promote justice, embrace diverse and marginalized perspectives, and reduce evaluator-evaluand polarization by shifting power towards evaluation users, whilst also fostering a more autonomous evaluation profession and mutualistic partnerships of knowledge and action. Most fundamentally, the principles advocate for a recentering of values, reflexivity and learning at the heart of evaluation practice, and ensure that the core purpose of evaluation is to support systemic, societal transformation towards regenerative futures.

Our review suggests that if these principles – which represent major shifts in evaluation conventions – are applied by evaluators (and supported by evaluation commissioners, sponsors and funders), they will significantly enhance the effectiveness of evaluation in assessing and assisting positive transformations. An important caveat is the tension between the Justice, Power Shift and Autonomous Evaluation Principles, which we discuss below. For an evaluation to be ‘transformation-focused’ as understood in this review, we suggest that it must apply Principles 1-6 and 9-12 (particularly because of the many interdependencies and complementarities between these principles, as illustrated in Fig 3), and also apply the Power Shift Principle where this does not contradict the Autonomous Evaluation Principle, and apply the Autonomous Evaluation Principle only where this does not contradict the Justice Principle (Fig 3). The principles are to be applied as an integrated set, not selected on a whim as from an à la carte menu. Together they provide a powerful approach to preparing for, and navigating, desired transformations as well as the transformations and social-ecological collapses that are already in train.

While the principles bear similarities to the Prague Declaration [122], they go further than the Declaration in several ways. Firstly, they clarify how transformation-focused evaluation is a fundamental contrast from how evaluation has often been done in the past, building on but going further than earlier paradigm shifts in evaluation. Secondly, they surface certain principles that, while possibly implicit in the Prague Declaration, are not obvious in that statement. These include the recognition that evaluands are unique, the need for more agile and developmental evaluation approaches, the importance of working with values, and the change in emphasis from results-based accountability to learning – all of which underscore the deeper paradigmatic shifts necessary for a more transformation-focused evaluation. Thirdly, our principles are more action-oriented in suggesting existing methodologies for putting the principles into practice. Our principles also complement and build on Patton and Felcis’ criteria and principles for evaluating transformations [4,63]. In particular, the latter have a stronger focus on what features of evaluands should be evaluated, whereas our principles focus on how evaluation practice itself should be conducted, and highlight the numerous ways in which evaluations can embody the transformations they aim to support by fundamentally shifting their modus operandi.

Viewing the principles as a whole, what stands out is the radical shift from what might be termed ‘first-order’ evaluation – where evaluators aim to be independent, value-neutral observers assessing systems from the outside – towards ‘second-order’ evaluation [226], where evaluators are embedded in system change processes, treat evaluation as an intervention that becomes part of system change, are reflexive about their own positionality, and actively support successful transformation because it is morally imperative and they have ‘skin in the game’ [4]. This shift is likely to be controversial for those who see evaluation as a neutral technical activity. Ultimately, however, a key skill for evaluators is to know when to ‘step into’ the system being evaluated, adopt a more cooperative power relation with other stakeholders and potentially build the capacity of initiatives to evaluate themselves, and when to ‘step out’ and adopt a more removed, independently critical approach. This will depend on context [227]. For instance, independent or external evaluation – where power cannot be exerted on the evaluator to give a particular judgment, and the evaluator has not been actively involved in the transformation effort up to that point – may be particularly important given the need to hold governments and businesses to account at a time when greenwashing and ‘transformation-washing’ is rife [10,228], and many investors and businesses are retreating from Environmental, Social and Governance (ESG) commitments due to their politicization by right-wing actors [229231]. Independent evaluation is also deemed important for intergovernmental financial institutions such as the International Monetary Fund, whose governance structure otherwise lacks democratic accountability [232]. In other cases, more second-order approaches may be appropriate. In the Netherlands, for instance, groups of civil servants at a few municipalities act as reflexive evaluators to work on the dual purpose of tackling societal challenges, such as the energy transition, as well as changing internal modes of operation, to turn into a learning organization [78]. External evaluations may also be inappropriate where they reinforce colonial issues, such as Global North evaluating the Global South, highlighting a tension between the Autonomous Evaluation and Justice Principles (Fig 3). We would emphasize that more independent, external evaluations can still be transformation-focused even if they do not apply the Power Shift Principle. For instance, they can still apply stronger ethical stances of assisting positive transformations. Such ‘bias’ does not preclude evaluators from offering useful critical feedback, as highlighted in approaches such as Reflexive Monitoring in Action [213]. The Power Shift Principle is an ideal that is pragmatically inappropriate in some contexts, but applicable and indeed imperative in others.

We call on evaluators, funders, policymakers, communities and other actors to attend to the presented principles in the commissioning and practice of evaluation if they aim to support transformations in social-ecological systems towards regenerative futures. This is because evaluation aligning to these principles not only provides insight into the impacts of transformations and developmental feedback for transformation initiatives, but also actively supports and embodies transformation in its ethics and approach. It would also encourage ongoing critical reflection in initiatives on whether ambition and transformational intent are being maintained, and not simply whether the status quo is being maintained and managed effectively. Sustaining such transformational intent is likely to be essential for achieving genuinely transformational change in societies, rather than more incremental change that fails to address global crises with the speed, breadth and depth that are urgently required [13]. There is likely to be a similar challenge in maintaining a transformational focus in evaluation practice and not falling back onto conventional approaches. Developing ‘communities of practice’, such as the Transformational Change Learning Partnership of the Climate Investment Funds [222], is likely to be important for embedding the kinds of transformation-focused principles presented in this paper.

Whilst some of the principles have been in practice for longer and are more mainstreamed, others may be more unfamiliar or contentious. Adaptive management applies developmental evaluation principles and has become increasingly common over the last two decades in conservation and natural resource management, as well as international development [233235]. For even longer, China’s approaches of ‘directed improvisation’ (whereby the central government facilitates local, decentralized experimentation) and co-evolution between governments and markets, have applied developmental principles and are key factors behind China’s extraordinary success in poverty alleviation [236238]. There has also been much progress in evaluations adopting complexity-aware and more genuinely participatory approaches, e.g., in the public health field [45,46].

However, the association between evaluation and objectivity, neutrality, quantification, standardization, controlled experimental designs, and counterfactual evidence, is widely institutionalized (e.g., in guidance from research funding bodies, and policymaking [174]). Such forms of evaluation are valuable in their appropriate context, but are rarely sufficient, and often unfeasible or inappropriate, where transformational change is sought in complex systems. There may also be significant pushback from governments and private sector organizations with vested interests in prolonging the status quo rather than fundamentally transforming their actions. These all present major hurdles to convincing actors of the importance, trustworthiness and rigor of transformation-focused evaluation. While transformation-focused evaluators can, and have a responsibility to, maintain rigor – including by remaining empirically grounded, justifying the choice of methods, and questioning underlying assumptions [85] – actors in wider society also have a responsibility to reset standards and expectations around what makes a rigorous evaluation. Particular responsibility lies on commissioners, sponsors and funders of evaluations to shift notions of accountability towards adaptation and learning rather than achievement of specific results, and to expand temporal horizons to enable more ongoing, developmental forms of evaluation.

Despite these barriers, there are also significant opportunities, such as in the growing importance of impact investing and evaluation in the private sector (e.g., social enterprises) and philanthropy space in support of transformative change [239241], and collaboration with the many science communities working at the nexus between environment and development, such as the Intergovernmental Panel on Climate Change and Intergovernmental Science-Policy Platform on Biodiversity and Ecosystem Services. The current crisis in development evaluation following major cuts to international aid budgets might spur evaluators to work more collaboratively, embrace responsible use of powerful new technologies such as AI, and both reframe and more strongly champion the vital role played by evaluation in assisting transformations. Moreover, the kaleidoscope of different approaches, tools, techniques and concepts now available offer many tangible routes into operationalizing transformation-focused evaluation principles. But transformation-focused evaluation is ultimately more than just applying particular evaluation methods, and requires a radical shift in our mindsets, including how we think about action and intervention more broadly. As our review has shown, transformation-focused evaluation means adopting a complexity-informed worldview, attending to issues of power, working in more coordinated, cooperative ways with diverse actors, and adopting an ethical stance of supporting social-ecological transformations towards regenerative futures by facilitating reflexivity and learning, seeing evaluation as part of the change process itself. It is also about reminding ourselves of how fundamental evaluation is to how we see the world and act within it, and its deep connection to the very things that make us human, including reasoning, critical judgment, understanding, intuition, values, empathy and morality [242]. Change the way we evaluate, and a whole new world of possibility emerges.

Supporting information

S1 Table. Information on database searches used to obtain literature on transformation-focused evaluation.

https://doi.org/10.1371/journal.pstr.0000164.s001

(DOCX)

S2 Table. Screening criteria for documents found in the literature review.

https://doi.org/10.1371/journal.pstr.0000164.s002

(DOCX)

S3 Table. Overarching evaluation approaches relevant to transformation-focused evaluation.

Numbers in the third column refer to the principles proposed in this paper. They are color-coded according to the three main groups of principles: green for Complexity Principles, purple for Power Principles, and blue for Purpose Principles.

https://doi.org/10.1371/journal.pstr.0000164.s003

(DOCX)

S5 Table. Preferred Reporting Items for Systematic reviews and Meta-Analyses extension for Scoping Reviews (PRISMA-ScR) Checklist.

https://doi.org/10.1371/journal.pstr.0000164.s005

(DOCX)

Acknowledgments

We are grateful to a number of additional people who read and provided comments on drafts of this paper, including Leo Beacroft, Amber Reed, Jenny Thompson and Susan Symonds from North Yorkshire Council, as well as several others from the University of York, including Lee Eyre and Suzanne Eugyen Om. If you found this paper useful (or not!) at your institution/organization, please let us know, as we are interested in impact beyond academic citations.

References

  1. 1. UNEP. The closing window: climate crisis calls for rapid transformation of societies (Emissions Gap Report 2022). Nairobi: United Nations Environment Programme; 2022. Available from: https://www.unep.org/resources/emissions-gap-report-2022
  2. 2. Chancel L, Piketty T, Saez E, Zucman G. World inequality report 2022. World Inequality Lab; 2021. Available from: https://wir2022.wid.world/
  3. 3. IPBES. Global assessment report on biodiversity and ecosystem services of the Intergovernmental Science-Policy Platform on Biodiversity and Ecosystem Services. Brondizio ES, Settele J, Díaz S, Ngo HT, editors. Bonn: IPBES secretariat; 2019. https://doi.org/10.5281/zenodo.6417333
  4. 4. Patton MQ. Evaluation criteria for evaluating transformation: implications for the coronavirus pandemic and the global climate emergency. Am J Eval. 2020;42(1):53–89.
  5. 5. Riedy C. Discourse coalitions for sustainability transformations: common ground and conflict beyond neoliberalism. Curr Opin Environm Sustain. 2020;45:100–12.
  6. 6. Boulangeat I, Allain S, Crouzat E, Girard S, Granjou C, Poirier C, et al. From human-nature dualism towards more integration in socio-ecosystems studies. In: Misiune I, Depellegrin D, Egarter Vigl L, editors. Human-nature interactions: exploring nature’s values across landscapes. Cham: Springer International Publishing; 2022. p. 37–49. https://doi.org/10.1007/978-3-031-01980-7_4
  7. 7. Hébert-Dufresne L, Allard A, Garland J, Hobson EA, Zaman L. The path of complexity. npj Complex. 2024;1(1):1–2.
  8. 8. Waddock S, Meszoely GM, Waddell S, Dentoni D. The complexity of wicked problems in large scale change. J Organ Chang Manag. 2015;28(6):993–1012.
  9. 9. Folke C, Polasky S, Rockström J, Galaz V, Westley F, Lamont M, et al. Our future in the Anthropocene biosphere. Ambio. 2021;50(4):834–69. pmid:33715097
  10. 10. Vogel C, O’Brien K. Getting to the heart of transformation. Sustain Sci. 2022;17(2):653–9. pmid:34413905
  11. 11. Díaz S, Settele J, Brondízio ES, Ngo HT, Agard J, Arneth A, et al. Pervasive human-driven decline of life on Earth points to the need for transformative change. Science. 2019;366(6471):eaax3100. pmid:31831642
  12. 12. Sachs JD, Schmidt-Traub G, Mazzucato M, Messner D, Nakicenovic N, Rockström J. Six transformations to achieve the sustainable development goals. Nat Sustain. 2019;2(9):805–14.
  13. 13. Fazey I, Leicester G. Archetypes of system transition and transformation: six lessons for stewarding change. Energy Res Soc Sci. 2022;91:102646.
  14. 14. Penuelas J, Nogué S. Catastrophic climate change and the collapse of human societies. Natl Sci Rev. 2023;10(6):nwad082. pmid:37181096
  15. 15. Buckton SJ, Fazey I, Sharpe B, Om ES, Doherty B, Ball P, et al. The regenerative lens: a conceptual framework for regenerative social-ecological systems. One Earth. 2023;6(7):824–42.
  16. 16. Geels FW. Regime resistance against low-carbon transitions: introducing politics and power into the multi-level perspective. Theory Cult Soc. 2014;31(5):21–40.
  17. 17. Paterson M. Climate change and international political economy: between collapse and transformation. Rev Int Political Econ. 2020;28(2):394–405.
  18. 18. Junge K, Cullen J, Iacopini G. Using contribution analysis to evaluate large-scale, transformation change processes. Evaluation. 2020;26:227–45.
  19. 19. Ofir Z, Rugg D. American journal of evaluation section on international developments in evaluation: transforming evaluation for times of global transformation. Am J Eval. 2021;42(1):47–52.
  20. 20. ANZEA, Superu. Evaluation standards for Aotearoa New Zealand. Superu; 2015. Available from: https://anzea.org.nz/assets/Key-ANZEA-Files/ANZEA-Superu-Evaluation-standards-final-020415.pdf.
  21. 21. Schwandt TA, Gates EF. Evaluating and valuing in social research. New York (NY): Guilford Press; 2021.
  22. 22. Ferguson M, Zayas V. Automatic evaluation. Curr Dir Psychol Sci. 2009;18(6):362–6.
  23. 23. Wanzer DL. What is evaluation?: Perspectives of how evaluation differs (or not) from research. Am J Eval. 2020;42(1):28–46.
  24. 24. Magro C, Van den Berg RD, Adrien M-H. Transforming evaluation for transformational changes. In: Van den Berg RD, Magro C, Adrien M-H, editors. Transformational evaluation for the global crises of our times. Exeter: IDEAS; 2021. p. 1–18. Available from: https://ideas-global.org/wp-content/uploads/2021/07/2021-IDEAS-book-Transformational-Evaluation.pdf
  25. 25. Picciotto R. Evaluation for transformation: What will it take? In: Van den Berg RD, Magro C, Adrien M-H, editors. Transformational evaluation for the global crises of our times. Exeter: IDEAS; 2021. p. 217–238. Available from: https://ideas-global.org/wp-content/uploads/2021/07/2021-IDEAS-book-Transformational-Evaluation.pdf
  26. 26. Hejnowicz AP, Ofir Z, Page GG. Evaluation for transformation: pathways to mobilize transformational change for sustainable development. Environ Sci Policy. under review.
  27. 27. Patton MQ. Transformation to global sustainability: implications for evaluation and evaluators. New Dir Eval. 2019;2019:103–17.
  28. 28. Feola G. Societal transformation in response to global environmental change: a review of emerging concepts. Ambio. 2015;44(5):376–90. pmid:25431335
  29. 29. Fazey I, Colvin J. Transformation: an introductory guide to fundamental change for researchers and change makers in a world of crises - a report for the transforming UK food systems SPF programme. York & Stroud: University of York & Emerald Network Ltd; 2023. Available from: https://ukfoodsystems.ukri.org/transformation-a-guide-to-fundamental-change-in-a-world-of-crisis/
  30. 30. O’Brien K, Sygna L. Responding to climate change: the three spheres of transformation. Transformation in a Changing Climate; 2013. p. 16–23. Available from: https://www.researchgate.net/publication/309384186_Responding_to_climate_change_The_three_spheres_of_transformation
  31. 31. Fazey I, Moug P, Allen S, Beckmann K, Blackwood D, Bonaventura M. Transformation in a changing climate: a research agenda. Clim Dev. 2018;10:197–217.
  32. 32. Davelaar D. Transformation for sustainability: a deep leverage points approach. Sustain Sci. 2021;16(3):727–47.
  33. 33. Cram F, Tibbetts K, LaFrance J. Editors’ notes: a stepping stone in indigenous evaluation. New Dir Eval. 2018;2018:7–16.
  34. 34. Ofir Z. Evaluation in transition: the promise and challenge of south-south cooperation. Can J Program Eval. 2021;36(2):120–40.
  35. 35. Chilisa B, Mertens D. Indigenous made in Africa evaluation frameworks: addressing epistemic violence and contributing to social transformation. Am J Eval. 2021;42(1):241–53.
  36. 36. van Mierlo B, Arkesteijn M, Leeuwis C. Enhancing the reflexivity of system innovation projects with system analyses. Am J Eval. 2010;31(2):143–61.
  37. 37. Kelly M, Maynard-Moody S. Policy analysis in the post-positivist era: engaging stakeholders in evaluating the economic development districts program. Public Admin Rev. 1993;53(2):135–42.
  38. 38. Parry O, Gnich W, Platt S. Principles in practice: reflections on a “postpositivist” approach to evaluation research. Health Educ Res. 2001;16(2):215–26. pmid:11345663
  39. 39. Mertens DM, Wilson AT. Program evaluation theory and practice: a comprehensive guide. New York (NY): The Guilford Press; 2012.
  40. 40. Guba E, Lincoln YS. Fourth generation evaluation. 1st ed. Thousand Oaks (CA): SAGE Publications; 1989.
  41. 41. Jack K. What is realist evaluation? Evid Based Nurs. 2022:ebnurs-2022-103608. pmid:35985802
  42. 42. Pawson R, Tilley N. Realistic evaluation. 1st ed. London: SAGE Publications Ltd; 1997.
  43. 43. Holmén J, Adawi T, Holmberg J. Student-led sustainability transformations: employing realist evaluation to open the black box of learning in a Challenge Lab curriculum. Int J Sustain High Educ. 2021;22(1):1–24.
  44. 44. Renger R, Wood S, Williamson S, Krapp S. Systemic evaluation, impact evaluation and logic models. Eval J Australas. 2011;11(2):24–30.
  45. 45. Skivington K, Matthews L, Simpson SA, Craig P, Baird J, Blazeby JM, et al. A new framework for developing and evaluating complex interventions: update of Medical Research Council guidance. BMJ. 2021;374:n2061. pmid:34593508
  46. 46. Harting J, Kruithof K, Ruijter L, Stronks K. Participatory research in health promotion: a critical review and illustration of rationales. Health Promot Int. 2022;37(Supplement_2):ii7–20. pmid:35748283
  47. 47. Jagosh J, Stott H, Halls S, Thomas R, Liddiard C, Cupples M, et al. Benefits of realist evaluation for rapidly changing health service delivery. BMJ Open. 2022;12(7):e060347. pmid:35896288
  48. 48. Ulrich W, Reynolds M. Critical systems heuristics. In: Reynolds M, Holwell S, editors. Systems approaches to managing change: a practical guide. London: Springer; 2010. p. 243–92. https://doi.org/10.1007/978-1-84882-809-4_6
  49. 49. Midgley G, Rajagopalan R. Critical systems thinking, systemic intervention, and beyond. In: Metcalf GS, Kijima K, Deguchi H, editors. Handbook of systems sciences. Singapore: Springer; 2021. p. 107–57. https://doi.org/10.1007/978-981-15-0720-5_7
  50. 50. Price L, Lotz-Sistka H, editors. Critical realism, environmental learning and social-ecological change. 1st ed. London: Routledge; 2015.
  51. 51. De Souza DE. A critical realist approach to systems thinking in evaluation. Evaluation. 2022;28(1):72–90.
  52. 52. Beck U, Bonss W, Lau C. The theory of reflexive modernization: problematic, hypotheses and research programme. Theory Cult Soc. 2003;20(2):1–33.
  53. 53. Arkesteijn M, van Mierlo B, Leeuwis C. The need for reflexive evaluation approaches in development cooperation. Evaluation. 2015;21(1):99–115.
  54. 54. Mertens DM. Transformative research and evaluation. 1st ed. New York: Guilford Press; 2008.
  55. 55. Schwandt TA. Post-normal evaluation? Evaluation. 2019;25(3):317–29.
  56. 56. Patton MQ. Blue marble evaluation perspective: how evaluations help solve global crises. In: Van den Berg RD, Magro C, Adrien M-H, editors. Transformational evaluation for the global crises of our times. Exeter: IDEAS; 2021. p. 19–36. Available from: https://ideas-global.org/wp-content/uploads/2021/07/2021-IDEAS-book-Transformational-Evaluation.pdf
  57. 57. Feinstein O. Dynamic evaluation for transformational change. In: van den Berg RD, Magro C, Mulder SS, editors. Evaluation for transformational change: opportunities and challenges for the sustainable development goals. Exeter: IDEAS; 2019. p. 17–31. Available from: https://ideas-global.org/wp-content/uploads/2019/11/2019-11-05-Final_IDEAS_EvaluationForTransformationalChange.pdf
  58. 58. van den Berg RD, Magro C, Mulder SS, editors. Evaluation for transformational change: opportunities and challenges for the sustainable development goals. Exeter: IDEAS; 2019. Available from: https://ideas-global.org/wp-content/uploads/2019/11/2019-11-05-Final_IDEAS_EvaluationForTransformationalChange.pdf
  59. 59. van den Berg RD, Magro C, Adrien M-H, editors. Transformational evaluation for the global crises of our times. Exeter: IDEAS; 2021. Available from: https://ideas-global.org/wp-content/uploads/2021/07/2021-IDEAS-book-Transformational-Evaluation.pdf
  60. 60. Uitto JI, Batra G, editors. Transformational change for people and the planet. Cham: Springer International Publishing; 2022. https://doi.org/10.1007/978-3-030-78853-7
  61. 61. Patton MQ. Incorporate environmental sustainability in evaluations. In: Change.org [Internet]; 2024 Apr 15 [cited 2024 May 9. ]. Available from: https://www.change.org/p/incorporate-environmental-sustainability-in-evaluations
  62. 62. Williams A, Dickman J, Smurthwaite R. Advancing evaluation and learning on transformational change: lessons from the climate investment funds’ transformational change learning partnership. Am J Eval. 2020;42(1):90–109.
  63. 63. Patton MQ, Felcis W. Transformational Eval week: principles for evaluating transformation by Michael Quinn Patton and Weronika Felcis. In: AEA365 [Internet]; 2023 [cited 2024 Mar 4. ]. Available from: https://aea365.org/blog/transformational-eval-week-principles-for-evaluating-transformation-by-michael-quinn-patton-and-weronika-felcis/
  64. 64. Gates EF, Fils-Aime F. System change evaluation: insights from The Rippel foundation and its ReThink health initiative. New Dir Eval. 2021;2021:125–38.
  65. 65. Rockefeller Philanthropy Advisors. Assessing systems change: a funders’ workshop report. New York: The Rockefeller Foundation; 2019. Available from: https://www.rockpa.org/wp-content/uploads/2019/10/Assessing-Systems-Change-A-Funders-Workshop-Report-Rockefeller-Philanthropy-Advisors-August-2019.pdf
  66. 66. Thompson Coon R, Vataja K, Parkkonen P. Sitra’s evaluation framework. Helsinki: Sitra; 2022. Available from: https://media.sitra.fi/app/uploads/2022/04/sitra-evaluation_framework_december_2021-006.pdf
  67. 67. Cabaj M. Evaluating systems change results: an inquiry framework. Tamarack Institute; 2019. Available from: https://www.tamarackcommunity.ca/hubfs/Resources/Publications/Paper%20Evaluating%20Systems%20Change%20Results%20Mark%20Cabaj.pdf?hsCtaTracking=2797ccdf-cfd3-4309-a6e0-c70b6a7ed5de%7Cfb84904f-568e-4e7f-b063-8040401998b4
  68. 68. Global Alliance for the Future of Food. Monitoring and evaluation. In: Global Alliance for the Future of Food [Internet]; 2020 Dec 22 [cited 2023 Jan 12. ]. Available from: https://futureoffood.org/our-approach/monitoring-and-evaluation/
  69. 69. Ofir Z, Chaplowe S. International Evaluation Academy (IEA) week: introducing the international evaluation academy. In: AEA365 [Internet]. 2022 Feb 20 [cited 2023 Aug 3. ]. Available from: https://aea365.org/blog/international-evaluation-academy-iea-week-introducing-the-international-evaluation-academy-by-zenda-ofir-and-scott-chaplow/
  70. 70. BetterEvaluation. In: BetterEvaluation [Internet]. [cited 2024 Apr 16. ]. Available from: https://www.betterevaluation.org/
  71. 71. Arksey H, O’Malley L. Scoping studies: towards a methodological framework. Int J Soc Res Methodol. 2005;8(1):19–32.
  72. 72. Tricco AC, Lillie E, Zarin W, O’Brien KK, Colquhoun H, Levac D, et al. PRISMA Extension for Scoping Reviews (PRISMA-ScR): checklist and explanation. Ann Intern Med. 2018;169(7):467–73. pmid:30178033
  73. 73. Martín-Martín A, Thelwall M, Orduna-Malea E, Delgado López-Cózar E. Google scholar, Microsoft academic, Scopus, dimensions, web of science, and OpenCitations’ COCI: a multidisciplinary comparison of coverage via citations. Scientometrics. 2021;126(1):871–906. pmid:32981987
  74. 74. Haddaway NR, Collins AM, Coughlin D, Kirk S. The role of google scholar in evidence reviews and its applicability to grey literature searching. PLoS One. 2015;10(9):e0138237. pmid:26379270
  75. 75. Hagstrom C, Kendall S, Cunningham H. Googling for grey: using Google and Duckduckgo to find grey literature. Abstracts of the 23rd Cochrane Colloquium (Cochrane database systematic reviews supplements); 2015.
  76. 76. Genç B, Bada E. English as a world language in academic writing. Read Matrix. 2010;10:142–51. Available from: https://www.semanticscholar.org/paper/4793156485f301c604dad5aeb2f53687d62e7a50
  77. 77. Stufflebeam DL, Coryn CLS. Evaluation theory, models, and applications. 2nd ed. San Francisco: Jossey-Bass; 2014. Available from: https://www.wiley.com/en-gb/Evaluation+Theory%2C+Models%2C+and+Applications%2C+2nd+Edition-p-9781118074053
  78. 78. Buckton SJ, Fazey I, Ball P, Ofir Z, Colvin J, Darby M, et al. Suggestion from the advisory group of co-authors; 2023.
  79. 79. Amarante V, Burger R, Chelwa G, Cockburn J, Kassouf A, McKay A, et al. Underrepresentation of developing country researchers in development research. Appl Econ Lett. 2021;29(17):1659–64.
  80. 80. Collyer F. Global patterns in the publishing of academic knowledge: global north, global South. Curr Sociol. 2018;66(1):56–73.
  81. 81. Cram F. Conclusion: lessons about indigenous evaluation. New Dir Eval. 2018;2018(159):121–33.
  82. 82. Jackson JL, Kuriyama A, Anton A, Choi A, Fournier J-P, Geier A-K, et al. The accuracy of google translate for abstracting data from non-english-language trials for systematic reviews. Ann Intern Med. 2019;171(9):677–9. pmid:31357212
  83. 83. Braun V, Clarke V. Using thematic analysis in psychology. Qual Res Psychol. 2006;3(2):77–101.
  84. 84. Patton MQ. Principles-focused evaluation: the GUIDE. 1st ed. New York (NY): Guilford Press; 2018.
  85. 85. Patton MQ. What is essential in developmental evaluation? On integrity, fidelity, adultery, abstinence, impotence, long-term commitment, integrity, and sensitivity in implementing evaluation models. Am J Eval. 2016;37(2):250–65.
  86. 86. Haddad CR, Bergek A. Towards an integrated framework for evaluating transformative innovation policy. Res Policy. 2023;52:104676.
  87. 87. Picciotto R. Is adversary evaluation worth a second look? Am J Eval. 2018;40(1):92–103.
  88. 88. African Evaluation Association. The African evaluation principles: 2021 version. Accra: African Evaluation Association; 2021. Available from: https://afrea.org/AEP/new/The-African-Evaluation-Principles.pdf
  89. 89. Gallagher L, Ofir Z. The rights of nature: an emerging transformation opportunity for evaluation. Can J Program Eval. 2021;36(2):141–61.
  90. 90. Patton MQ. Blue marble evaluation: premises and principles. 1st ed. New York: Guilford Press; 2019.
  91. 91. CECAN. Complexity evaluation framework: recognising complexity and key considerations for complexity-appropriate evaluation in the Department for Environment, Food and Rural Affairs, DEFRA. CECAN; 2019. Available from: https://www.betterevaluation.org/sites/default/files/14675_ComplexityEvaluationFramework.pdf
  92. 92. CECAN. The complexity evaluation toolkit (Version 1.0). CECAN; 2021. Available from: https://www.cecan.ac.uk/wp-content/uploads/2021/07/Toolkit-2021-web.pdf
  93. 93. Rodríguez-Campos L. Advances in collaborative evaluation. Eval Program Plann. 2012;35(4):523–8. pmid:22377419
  94. 94. Concilio G, Karimi M, Rössl L. Complex projects and transition-driven evaluation: the case of the easyRights European project. Sustainability. 2021;13(4):2343.
  95. 95. Kushnier L, Nadin S, Hill ME, Taylor M, Jun S, Mushquash CJ, et al. Culturally responsive evaluation: a scoping review of the evaluation literature. Eval Program Plann. 2023;100:102322. pmid:37315348
  96. 96. Anderson KC, Stern MJ, Powell RB, Dayer AA, Archibald TG. A culturally responsive evaluation framework and its application in environmental education. Eval Program Plann. 2022;92:102073. pmid:35364559
  97. 97. Bowman N, Dodge-Francis C. Culturally responsive indigenous evaluation and tribal governments: Understanding the relationship. New Dir Eval. 2018;2018:17–31.
  98. 98. Picciotto R. What is democratic evaluation? Eval Conn. 2013;special edition on EES “Public Hearing” at the European Parliament: Evaluation in Democracy:4–5. Available from: https://europeanevaluation.org/wp-content/uploads/2020/04/Evaluation-Connections-June-2013-Special-edition.pdf
  99. 99. Norman CD. Supporting systems transformation through design‐driven evaluation. New Dir Eval. 2021;2021(170):149–58.
  100. 100. Haylock L, Miller C. Merging developmental and feminist evaluation to monitor and evaluate transformative social change. Am J Eval. 2015;37(1):63–79.
  101. 101. Fetterman D. Transformative empowerment evaluation and freirean pedagogy: alignment with an emancipatory tradition. New Dir Eval. 2017;2017(155):111–26.
  102. 102. Luna Pinzon A, Stronks K, Dijkstra C, Renders C, Altenburg T, den Hertog K, et al. The ENCOMPASS framework: a practical guide for the evaluation of public health programmes in complex adaptive systems. Int J Behav Nutr Phys Act. 2022;19(1):33. pmid:35346233
  103. 103. Equitable Evaluation Initiative. Shifting the evaluation paradigm: the equitable evaluation framework. San Rafael (CA) and Washington (DC): Equitable Evaluation Initiative and Grantmakers for Effective Organizations; 2021. Available from: https://www.equitableeval.org/_files/ugd/21786c_7db318fe43c342c09003046139c48724.pdf
  104. 104. Reynolds M. Equity-focused developmental evaluation using critical systems thinking. Evaluation. 2014;20(1):75–95.
  105. 105. Davidson J, Macfarlan A, Rogers P, Rowe A, Stevens K. Sustainability-inclusive evaluation: why we need it and how to do it. In: A footprint evaluation guide. Footprint evaluation; 2023. Available from: https://www.betterevaluation.org/sites/default/files/2023-09/Sustainability-inclusive-evaluation-Why-we-need-it-and-how-to-do-it-Footprint-Evaluation-Initiative-v1-1.pdf
  106. 106. Gates E, Page G, Crespo J, Oporto M, Bohórquez J. Ethics of evaluation for socio-ecological transformation: Case-based critical systems analysis of motivation, power, expertise, and legitimacy. Evaluation. 2023;29:23–49.
  107. 107. Hargreaves MB. Evaluating system change: a planning guide. Princeton (NJ): Mathematica Policy Research, Inc.; 2010. Available from: https://www.mathematica.org/publications/evaluating-system-change-a-planning-guide
  108. 108. Mertens DM. Inclusive evaluation: implications of transformative theory for evaluation. Am J Eval. 1999;20(1):1–14.
  109. 109. Meissner S, Braithwaite J, Thorne K, Martinez A, Lycett E. Indigenous feminist evaluation methods: a case study in “My two aunties”. Can J Program Eval. 2023;38(1):186–207.
  110. 110. Latham N. A practical guide to evaluating systems change in a human services system context. Washington (DC) and San Francisco (CA): Center for Evaluation Innovation & Learning for Action; 2014. Available from: https://evaluationinnovation.org/wp-content/uploads/2014/07/Systems-Change-Evaluation-Toolkit_FINAL.pdf
  111. 111. Phillips G II, Felt D, Perez-Bill E, Ruprecht MM, Glenn EE, Lindeman P, et al. Transforming the paradigm for LGBTQ+ evaluation: advancing a praxis of LGBTQ+ inclusion and liberation in evaluation. Am J Eval. 2022;44(1):7–28.
  112. 112. Omosa O, Archibald T. Towards defining and advancing “Made in Africa Evaluation”. eVALUation Matters. 2019:32–42. Available from: https://idev.afdb.org/sites/default/files/Evaluations/2020-03/article%204%20Towards%20defining%20and%20advancing%20%E2%80%9CMade%20in%20Africa%20Evaluation.pdf
  113. 113. Mickwitz P, Neij L, Johansson M, Benner M, Sandin S. A theory-based approach to evaluations intended to inform transitions toward sustainability. Evaluation. 2021;27(3):281–306.
  114. 114. O’Leary S. Grassroots accountability promises in rights-based approaches to development: The role of transformative monitoring and evaluation in NGOs. Account Organ Soc. 2017;63(1):21–41.
  115. 115. Ostrom E. A polycentric approach for coping with climate change (Background paper to the 2010 world development report - policy research working paper 5095). Washington (DC): The World Bank; 2009. Available from: https://documents1.worldbank.org/curated/en/480171468315567893/pdf/WPS5095.pdf
  116. 116. Bowman N. Nation‐to‐nation in evaluation: utilizing an indigenous evaluation model to frame systems and government evaluations. New Dir Eval. 2020;2020:101–18.
  117. 117. Ofir Z. Made in Africa evaluation 4. The NICE framework. In: Evaluation for Transformation [Internet]. 2018 Aug 23 [cited 2023 Aug 8. ]. Available from: https://zendaofir.com/the-nice-framework-part-1/
  118. 118. Page GG. Blog # 7: the orders of outcomes. In: SustainaMetrix [Internet]. 2020 Nov 25 [cited 2024 Feb 23. ]. Available from: https://www.sustainametrix.com/20142018-blog/2014/12/22/blog-7-the-orders-of-outcomes-4gk9h
  119. 119. Olsen SB, Page GG, Ochoa E. The analysis of governance responses to ecosystem change: a handbook for assembling a baseline. LOICZ Reports & Studies No. 34. Geesthacht: GKSS Research Center; 2009. Available from: https://s3-eu-west-2.amazonaws.com/futureearthcoasts/wp-content/uploads/2018/05/30151144/LOICZ-RS34.pdf
  120. 120. Cousins JB, Earl LM. The case for participatory evaluation. Educ Eval Policy Anal. 1992;14:397–418.
  121. 121. Parsons B, Krenn H. PCI: A reflective evaluation framework for systems change. Found Rev. 2018;10:31–43.
  122. 122. Van den Berg RD, Svoboda D, Ocampo A, Uitto JI, Mulder SS, Agrawal R, et al. The prague declaration: meaning and testimonials. In: Van den Berg RD, Magro C, Adrien M-H, editors. Transformational evaluation for the global crises of our times. Exeter: IDEAS; 2021. p. 397–434. Available from: https://ideas-global.org/wp-content/uploads/2021/07/2021-IDEAS-book-Transformational-Evaluation.pdf
  123. 123. Preskill H, Gopal S, Cook J, Mack K. Evaluating complexity: propositions for improving practice. Boston (MA): FSG; 2014. Available from: https://search.issuelab.org/resource/evaluating-complexity-propositions-for-improving-practice.html
  124. 124. Rogers P. Real-time evaluation (Monitoring and evaluation for adaptive management working paper series number 4). BetterEvaluation; 2020. Available from: http://www.betterevaluation.org/en/monitoring-and-evaluation-adaptive-management-working-paper-series
  125. 125. Rijswijk K, Bewsell D, Small B, Blackett P. Reflexive monitoring in New Zealand: evaluation lessons in supporting transformative change. Eval J Australas. 2015;15(4):38–43.
  126. 126. Ofir Z, Schwandt T, Duggan C, McLean R. Research Quality Plus (RQ+): a holistic approach to evaluating research. Ottawa (ON): IDRC; 2016. Available from: http://hdl.handle.net/10625/56528
  127. 127. Reynolds M, Gates E, Hummelbrunner R, Marra M, Williams B. Towards systemic evaluation. Syst Res Behav Sci. 2016;33:662–73.
  128. 128. Velasco D, Witte J. Transforming Sweden’s food system. In: Transformative Innovation Policy Consortium [Internet]; 2022 Feb 10 [cited 2024 Mar 11. ]. Available from: https://tipconsortium.net/transforming-swedens-food-system/
  129. 129. Baptiste LJC. Governance pathways for the greater caribbean: transformative evaluation principles. In: Van den Berg RD, Magro C, Adrien M-H, editors. Transformational evaluation for the global crises of our times. Exeter: IDEAS; 2021. p. 133–53. Available from: https://ideas-global.org/wp-content/uploads/2021/07/2021-IDEAS-book-Transformational-Evaluation.pdf
  130. 130. Maclure R. Pragmatism or transformation? Participatory evaluation of a humanitarian education project in Sierra Leone. Can J Program Eval. 2006;21(1):107–29.
  131. 131. Patton MQ, Campbell-Patton CE. Utilization-focused evaluation. 5th ed. New York (NY): SAGE Publications, Inc; 2022.
  132. 132. Aronsson I-L, Hassnain H. Value-based evaluations for transformative change. In: van den Berg RD, Magro C, Mulder SS, editors. Evaluation for transformational change: opportunities and challenges for the sustainable development goals. Exeter: IDEAS; 2019. p. 89–104. Available from: https://ideas-global.org/wp-content/uploads/2019/11/2019-11-05-Final_IDEAS_EvaluationForTransformationalChange.pdf
  133. 133. Parsons B, Dhillon L, Keene M, editors. Visionary evaluation for a sustainable, equitable future. Charlotte (NC): Information Age Publishing; 2020.
  134. 134. Julnes G. Supporting transitions to sustainability: evaluation for managing processes in the public interest. New Dir Eval. 2019;2019(162):119–54.
  135. 135. Uitto JI. Evaluation at the nexus between nature and humanity for transformational change. In: Van den Berg RD, Magro C, Adrien M-H, editors. Transformational evaluation for the global crises of our times. Exeter: IDEAS; 2021. p. 87–104. Available from: https://ideas-global.org/wp-content/uploads/2021/07/2021-IDEAS-book-Transformational-Evaluation.pdf
  136. 136. Bilella PR, Mulder SS, Zaveri S. To be or not to be an evaluator for transformational change: perspectives from the global south. In: Van den Berg RD, Magro C, Adrien M-H, editors. Transformational evaluation for the global crises of our times. Exeter: IDEAS; 2021. p. 157–75. Available from: https://ideas-global.org/wp-content/uploads/2021/07/2021-IDEAS-book-Transformational-Evaluation.pdf
  137. 137. Kroll H. How to evaluate innovation strategies with a transformative ambition? A proposal for a structured, process-based approach. Sci Public Policy. 2019;46(6):635–47.
  138. 138. Kugelberg S, Bartolini F, Kanter DR, Milford AB, Pira K, Sanz-Cobena A, et al. Implications of a food system approach for policy agenda-setting design. Glob Food Sec. 2021;28:100451. pmid:33738183
  139. 139. Hejnowicz A, Chaplowe S. Catching the wave: harnessing data science to support evaluation’s capacity for making a transformational contribution to sustainable development. Can J Program Eval. 2021;36(2):162–86.
  140. 140. Morell JA. A complexity-based meta-theory of change for transformation towards green energy. In: Van den Berg RD, Magro C, Adrien M-H, editors. Transformational evaluation for the global crises of our times. Exeter: IDEAS; 2021. p. 345–369. Available from: https://ideas-global.org/wp-content/uploads/2021/07/2021-IDEAS-book-Transformational-Evaluation.pdf
  141. 141. Neij L, Sandin S, Benner M, Johansson M, Mickwitz P. Bolstering a transition for a more sustainable energy system: a transformative approach to evaluations of energy efficiency in buildings. Energy Res Soc Sci. 2021;72:101864.
  142. 142. Turnheim B, Berkhout F, Geels F, Hof A, McMeekin A, Nykvist B, et al. Evaluating sustainability transitions pathways: bridging analytical approaches to address governance challenges. Glob Environ Change. 2015;35:239–53.
  143. 143. Monno V, Conte E. Sustainability in the built environment: integrating scales of action and evaluation. EJSD. 2015;4(2):51.
  144. 144. Loring PA, Gerlach SC. Food, culture, and human health in Alaska: an integrative health approach to food security. Environ Sci Policy. 2009;12(4):466–78.
  145. 145. Cole RJ, Busby P, Guenther R, Briney L, Blaviesciunaite A, Alencar T. A regenerative design framework: setting new aspirations and initiating new discussions. Build Res Inf. 2012;40(1):95–111.
  146. 146. Hampton S, Fawcett T, Rosenow J, Michaelis C, Mayne R. Evaluation in an emergency: assessing transformative energy policy amidst the climate crisis. Joule. 2021;5(2):285–9.
  147. 147. Gibbons L, Pearthree G, Cloutier S, Ehlenz M. The development, application, and refinement of a regenerative development evaluation tool and indicators. Ecol Indic. 2020;108:105698.
  148. 148. Thorpe J, Guijt J, Sprenger T, Stibbe D. Multi stakeholder platforms as system change agents: a guide for assessing effectiveness. Wageningen: IDS Institute of Development Studies and Wageningen University & Research; 2021. Available from: https://library.wur.nl/WebQuery/wurpubs/fulltext/548294
  149. 149. Ajena F, Bossard N, Clément C, Hilbeck A, Oehen B, Thomas J, et al. Agroecology & Digitalisation: traps and opportunities to transform the food system. Brussels: IFOAM Organics Europe; 2020. Available from: https://www.organicseurope.bio/content/uploads/2022/06/IFOAMEU_Agroecology_Digitalization_2020.pdf?dd
  150. 150. Global Alliance for the Future of Food. Principles for food systems transformation: a framework for action. Global Alliance for the Future of Food; 2021. Available from: https://futureoffood.org/wp-content/uploads/2021/06/GA_PrinciplesDoc.pdf
  151. 151. Ramirez R, Mannervik U. Strategy for a networked world. London: Imperial College Press; 2017.
  152. 152. Jia X. Agro-food innovation and sustainability transition: a conceptual synthesis. Sustainability. 2021;13(12):6897.
  153. 153. HDR. Regenerative design framework: how we establish and track metrics. In: HDR [Internet]. 2023 Apr 23 [cited 2024 Jan 4. ]. Available from: https://www.hdrinc.com/insights/regenerative-design-framework-how-we-establish-and-track-metrics
  154. 154. Behrens TR, Kelly T. Paying the piper: Foundation evaluation capacity calls the tune. New Dir Eval. 2008;2008(119):37–50.
  155. 155. Valette E, Schreiber K, Conaré D, Bonomelli V, Blay-Palmer A, Bricas N, et al. An emerging user-led participatory methodology: mapping impact pathways of urban food system sustainability innovations. In: Sustainable food system assessment. 1st ed. London: Routledge; 2019. p. 19–41. https://doi.org/10.4324/9780429439896-2
  156. 156. Hodgson A. A transdisciplinary world model. Syst Res. 2012;29(5):517–26.
  157. 157. Hodgson T. Synergy with the World Mandala. In: H3Uni [Internet]; 2020 May 9 [cited 2024 Oct 14. ]. Available from: https://www.h3uni.org/facilitation-guide/synergy-with-the-world-mandala/
  158. 158. Treasury Board of Canada Secretariat. Theory-based approaches to evaluation: concepts and practices. Government of Canada; 2021. Available from: https://www.canada.ca/en/treasury-board-secretariat/services/audit-evaluation/evaluation-government-canada/theory-based-approaches-evaluation-concepts-practices.html
  159. 159. Victorian Food Security and Food Systems Working Group. Towards a healthy, regenerative and equitable food system in Victoria: a consensus statement; 2022. Available from: https://nutritionaustralia.org/app/uploads/2022/06/Food-Systems-Consensus-Statement_Web-20220324_.pdf
  160. 160. Sharpe B, Hodgson A, Leicester G, Lyon A, Fazey I. Three horizons: a pathways practice for transformation. Ecosyst Soc. 2016;21(1):47.
  161. 161. Davies R, Dart J. The “Most Significant Change” (MSC) technique: a guide to its use; 2005. Available from: https://www.mande.co.uk/wp-content/uploads/2005/MSCGuide.pdf
  162. 162. Cooper S. Transformative evaluation: organisational learning through participative practice. Learn Organ. 2014;21(2):146–57.
  163. 163. Richardson RA, Patton MQ. Leadership‐evaluation partnership: Infusing systems principles and complexity concepts for a transformational alliance. New Dir Eval. 2021;2021(170):139–47.
  164. 164. Nobles J, Wheeler J, Dunleavy-Harris K, Holmes R, Inman-Ward A, Potts A, et al. Ripple effects mapping: capturing the wider impacts of systems change efforts in public health. BMC Med Res Methodol. 2022;22(1):72. pmid:35300619
  165. 165. Abidi-Habib M, Burt J, Colvin J, Msukwa C, Mukute M. Contradictions and complementarities between south and north on transformation in the Anthropocene. In: Van den Berg RD, Magro C, Adrien M-H, editors. Transformational evaluation for the global crises of our times. Exeter: IDEAS; 2021. p. 105–32. Available from: https://ideas-global.org/wp-content/uploads/2021/07/2021-IDEAS-book-Transformational-Evaluation.pdf
  166. 166. Canen A, Tarré de Oliveira M. Institutional evaluation in the Mercosur: transformation or control? Int J Educ Dev. 2000;20(4):277–86.
  167. 167. Camrass K. Regenerative futures. FS. 2020;22(4):401–15.
  168. 168. Mertens DM. Inclusivity and transformation: evaluation in 2010. Am J Eval. 2001;22(3):367–74.
  169. 169. Ríos-Muñoz D, Herrera-Araya D. Contribución de la evaluación educativa para la formación democrática y transformadora de estudiantes. Rev Electr Educare. 2021;25(3):1–15.
  170. 170. Camrass K. Regenerative futures: eight principles for thinking and practice. J Futures Stud. 2023;28. Available from: https://jfsdigital.org/articles-and-essays/2023-2/vol-28-no-1-september-2023/regenerative-futures-eight-principles-for-thinking-and-practice/
  171. 171. Otieno D, Niewolny K, Archibald T, Schenk T, Nunoo N. Transformative learning to promote transformative evaluation of food system praxis. Front Sustain Food Syst. 2023;6.
  172. 172. Walsh EA, Moore SA. Regenerative design: redefining the limits of architectural judgment. In: Architecture beyond criticism. 1st ed. New York: Routledge; 2014. p. 286–98. https://doi.org/10.4324/9781315740652-42
  173. 173. Itad. Independent evaluation of the biodiverse landscapes fund. In: Itad [Internet]. 2022 Sep 12 [cited 2024 Mar 22. ]. Available from: https://www.itad.com/project/independent-evaluation-biodiverse-landscapes-fund-defra/
  174. 174. Magro C, Van den Berg RD. Of portals and paradigms: evaluation, systems thinking and the pandemic. In: Van den Berg RD, Magro C, Adrien M-H, editors. Transformational evaluation for the global crises of our times. Exeter: IDEAS; 2021. p. 371–94. Available from: https://ideas-global.org/wp-content/uploads/2021/07/2021-IDEAS-book-Transformational-Evaluation.pdf
  175. 175. Mertens DM. Social transformation and evaluation. Eval J Australas. 2010;10(2):3–10.
  176. 176. Noltemeyer AL, Harper EA. Chapter 15: systems change and program evaluation. In: Grapin SL, Kranzler JH, editors. School psychology: professional issues and practices. New York: Springer; 2023.
  177. 177. Mikkolainen P. Evaluation must become future-sensitive: easy-to-implement ideas on how to do it in practice. NIRAS International Consulting; 2022. Available from: https://www.niras.com/media/rghbpvri/evaluation_future_sensitive.pdf
  178. 178. Lawson HA, Anderson-Butcher D, Petersen N, Barkdull C. Design teams as learning systems for complex systems change. J Hum Behav Soc Environ. 2003;7(1–2):159–79.
  179. 179. Kelly T, Brown P, Yu HC, Colombo M. Evaluating for the bigger picture: breaking through the learning and evaluation barriers to advancing community systems-change field knowledge. Found Rev. 2019;11(2):10.
  180. 180. Patton MQ, McKegg K, Wehipeihana N, editors. Developmental evaluation exemplars: principles in practice. New York: The Guildford Press; 2016.
  181. 181. Du Plessis C, Brandon P. An ecological worldview as basis for a regenerative sustainability paradigm for the built environment. J Clean Prod. 2015;109:53–61.
  182. 182. Guimarães V. Transformative pedagogical evaluation: Freirean principles practiced in Brazilian public schools. New Dir Eval. 2017;2017(155):99–110.
  183. 183. Patton MQ. Developmental evaluation: applying complexity concepts to enhance innovation and use. 1st ed. New York: Guilford Press; 2010.
  184. 184. Gamble J, McKegg K, Cabaj M. A developmental evaluation companion. Redding (CA): The McConnell Foundation; 2021. Available from: https://www.ecala.org/wp-content/uploads/2021/06/A-Developmental-Evaluation-Companion-McConnell-Foundation-June-2021.pdf
  185. 185. Williams B. Qualitative data analysis. Available from: https://www.bobwilliams.co.nz/ewExternalFiles/analysis2.pdf
  186. 186. Pirani A, Fuglestvedt J, Byers E, O’Neill B, Riahi K, Lee J-Y. Scenarios in IPCC assessments: lessons from AR6 and opportunities for AR7. npj Clim Action. 2024;3:1–7.
  187. 187. Magro C, D. van den Berg R. Systems evaluations for transformational change. In: van den Berg RD, Magro C, Mulder SS, editors. Evaluation for transformational change: opportunities and challenges for the sustainable development goals. Exeter: IDEAS; 2019. p. 131–55. Available from: https://ideas-global.org/wp-content/uploads/2019/11/2019-11-05-Final_IDEAS_EvaluationForTransformationalChange.pdf
  188. 188. Seeds of Good Anthropocenes. Home. In: Seeds of Good Anthropocenes [Internet]. [cited 2023 Aug 9. ]. Available from: https://goodanthropocenes.net/
  189. 189. Inayatullah S. Anticipatory action learning: Theory and practice. Futures. 2006;38(6):656–66.
  190. 190. Mertens D. What does a transformative lens bring to credible evidence in mixed methods evaluations? New Dir Eval. 2013;2013:27–35.
  191. 191. Osinski A. evaluating transition pathways beyond basic needs: a transdisciplinary approach to assessing food assistance. Food ethics. 2020;5(1–2):19.
  192. 192. Vinnari M, Vinnari E. A Framework for sustainability transition: the case of plant-based diets. J Agric Environ Ethics. 2013;27(3):369–96.
  193. 193. Cochran-Smith M, Reagan E. Centering equity in teacher education evaluation: From principles to transformative enactment. J Teach Educ. 2022;73(3):449–62.
  194. 194. Sibanda A, Ofir Z. Evaluation in an uncertain world: a view from the global south. In: Van den Berg RD, Magro C, Adrien M-H, editors. Transformational evaluation for the global crises of our times. Exeter: IDEAS; 2021. p. 37–61. Available from: https://ideas-global.org/wp-content/uploads/2021/07/2021-IDEAS-book-Transformational-Evaluation.pdf
  195. 195. Montrosse-Moorhead B, Bitar K, Arévalo J, Rishko-Porcescu A. Revolution in the making: evaluation “done well” in the era of the SDGs with a youth participatory approach. In: van den Berg RD, Magro C, Mulder SS, editors. Evaluation for transformational change: opportunities and challenges for the sustainable development goals. Exeter: IDEAS; 2019. p. 33–50. Available from: https://ideas-global.org/wp-content/uploads/2019/11/2019-11-05-Final_IDEAS_EvaluationForTransformationalChange.pdf
  196. 196. Sharpe B. Dilemma navigation. In: H3Uni [Internet]; 2019 Nov 29 [cited 2024 Oct 15. ]. Available from: https://www.h3uni.org/facilitation-guide/dilemma-navigation-guide/
  197. 197. Edwards PB, Hindmarsh R, Mercer H, Bond M, Rowland A. A three-stage evaluation of a deliberative event on climate change and transforming energy. JDD. 2008;4(1).
  198. 198. Blythe C, Harré N. Encouraging transformation and action competence: a theory of change evaluation of a sustainability leadership program for high school students. J Environ Educ. 2019;51(1):83–96.
  199. 199. Cooper S, Morciano D, Scardigno F, Ord J. Transformative evaluation in youth work and its emancipatory role in southern Italy. Ital J Sociol Educ. 2019;11(3):133–52.
  200. 200. Fetterman D, Kaftarian SJ, Wandersman A, editors. Empowerment evaluation: knowledge and tools for self-assessment, evaluation capacity building, and accountability. 2nd ed. SAGE Publications, Inc; 2014.
  201. 201. Baptiste L, Iese V, Gordon V, Bailey A, Persaud N, Nicholson C, et al. The transformative agenda for evaluation in Small Island Developing States: The Caribbean and the Pacific. In: van den Berg RD, Magro C, Mulder SS, editors. Evaluation for transformational change: opportunities and challenges for the sustainable development goals. Exeter: IDEAS; 2019. p. 71–87. Available from: https://ideas-global.org/wp-content/uploads/2019/11/2019-11-05-Final_IDEAS_EvaluationForTransformationalChange.pdf
  202. 202. Waddock S. Catalyzing purposeful transformation: the emergence of transformation catalysts. Bus Soci Rev. 2022;127(1):167–70.
  203. 203. Herrero M, Thornton PK, Mason-D’Croz D, Palmer J, Bodirsky BL, Pradhan P, et al. Articulating the effect of food systems innovation on the Sustainable Development Goals. Lancet Planet Health. 2021;5(1):e50–62. pmid:33306994
  204. 204. Luederitz C, Schäpke N, Wiek A, Lang D, Bergmann M, Bos J, et al. Learning through evaluation – a tentative evaluative scheme for sustainability transition experiments. J Clean Prod. 2017;169:61–76.
  205. 205. Foster-Fishman PG, Nowell B, Yang H. Putting the system back into systems change: a framework for understanding and changing organizational and community systems. Am J Community Psychol. 2007;39(3–4):197–215. pmid:17510791
  206. 206. Horner CE, Morse C, Carpenter N, Nordstrom KL, Faulkner JW, Mares T, et al. Cultivating pedagogy for transformative learning: a decade of undergraduate agroecology education. Front Sustain Food Syst. 2021;5.
  207. 207. McDonald B, Stevens K, Nabben T. Most significant change. In: Better Evaluation [Internet]; 2021 [cited 2023 Sep 5. ]. Available from: https://www.betterevaluation.org/methods-approaches/approaches/most-significant-change
  208. 208. Picciotto R. From disenchantment to renewal. Evaluation. 2020;26(1):49–60.
  209. 209. Webb K, Pelletier D, Maretzki A, Wilkins J. Local food policy coalitions: evaluation issues as seen by academics, project organizers, and funders. Agric Human Values. 1998;15:65–75.
  210. 210. van Mierlo B, Beers PJ. Understanding and governing learning in sustainability transitions: a review. Environ Innov Soc Transit. 2020;34:255–69.
  211. 211. Tamarack Institute. TOOL: single, double and triple loop learning. Waterloo (ON): Tamarack Institute; Available from: https://www.tamarackcommunity.ca/hubfs/Events/Multi-Day%20Events/Community%20Change%20Institute%20-%20CCI/2017%20CCI%20Vancouver/Resources/Tool%20-%20Single%20Double%20Triple%20Loop%20Learning.pdf
  212. 212. Flood RL, Romm NRA. Plurality revisited: diversity management and triple loop learning. Syst Pract. 1996;9(6):587–603.
  213. 213. Beers P, van Mierlo B. Reflexivity and learning in system innovation processes. Sociol Ruralis. 2017;57(4):415–36.
  214. 214. Francis CA, Jensen ES, Lieblein G, Breland TA. Agroecologist education for sustainable development of farming and food systems. Agron J. 2017;109(1):23–32.
  215. 215. Dahlberg KA. A transition from agriculture to regenerative food systems. Futures. 1994;26:170–9.
  216. 216. Gibbons LV. Moving beyond sustainability: a regenerative community development framework for co-creating thriving living systems and its application. JSD. 2020;13(2):20.
  217. 217. Conte E, Monno V. Sustainability of the built environment: opening new frontiers for evaluation. Proceedings of the 1st Workshop on the State of the Art and Challenges of Research Efforts at POLIBA -Research Contributions. Gangemi; 2014. p. 79–83. Available from: https://iris.poliba.it/handle/11589/18325
  218. 218. Tamarack Institute. The systems change evaluation canvas. Waterloo (ON): Tamarack Institute; 2019. Available from: https://www.tamarackcommunity.ca/hubfs/Resources/Tools/TOOL%20%7C%20Systems%20Change%20Evaluation%20Canvas.pdf?hsCtaTracking=437cda01-d20d-461c-9967-1d81352a3b27%7C91c71aaa-a029-402c-a21f-e16cff42b2ac
  219. 219. Dijkshoorn-Dekker M, Linderhof V, Mattijssen TJM, Polman N. Food secure metropolitan areas: the transition support system approach. Sustainability. 2020;12(13):5376.
  220. 220. United Nations Global Compact, We Mean Business, World Resources Institute. The Ambition Loop: How business and government can advance policies that fast track zero-carbon economic growth. United Nations Global Compact, We Mean Business & World Resources Institute; 2018 Nov. Available from: https://ambitionloop.org/
  221. 221. Future Stewards, Leaders’ Quest, High-Level Climate Action Champions for UN COP. 10 tools for systems change to a zero carbon world. Future Stewards; 2021. Available from: https://futurestewards.com/wp-content/uploads/2021/12/10-tools-for-systems-change-to-a-zero-carbon-world.pdf
  222. 222. Savage M, McPherson S, Williams A, Dickman J, Smurthwaite R, Sharma N. Signals of transformational change: insights from the evaluation of transformational change in the climate investment funds. Climate Investment Funds and Itad; 2020. Available from: https://www.itad.com/wp-content/uploads/2020/02/tc_signals_brief-1.pdf
  223. 223. Van der Merwe SE, Biggs R, Preiser R, Cunningham C, Snowden DJ, O’Brien K, et al. Making sense of complexity: using SenseMaker as a research tool. Systems. 2019;7(2):25.
  224. 224. Cole RJ. Regenerative design and development: current theory and practice. Build Res Inf. 2012;40(1):1–6.
  225. 225. Hardman J. Leading for regeneration. 1st ed. London: Routledge; 2011. https://doi.org/10.4324/9780203155288
  226. 226. Fazey I, Schäpke N, Caniglia G, Patterson J, Hultman J, van Mierlo B, et al. Ten essentials for action-oriented and second order energy transitions, transformations and climate change research. Energy Res Soc Sci. 2018;40:54–70.
  227. 227. Conley-Tyler M. A fundamental choice: internal or external evaluation? Eval J Australas. 2005;4:3–11.
  228. 228. Montgomery AW, Lyon TP, Barg J. No end in sight? A greenwash review and research agenda. Organ Environ. 2023;10860266231168905.
  229. 229. Hilson C. Climate change and the politicization of ESG in the US. Front Polit Sci. 2024;6.
  230. 230. Sorkin AR, Giang V, Gandel S, Hirsch L, Livni E, Gross J, et al. The pushback on E.S.G. investing. The New York Times; 2022 May 11 [cited 2024 Apr 24. ]. Available from: https://www.nytimes.com/2022/05/11/business/dealbook/esg-investing-pushback.html
  231. 231. Winston A. Why business leaders must resist the anti-ESG movement. Harvard Business Review; 2023 Apr 5 [cited 2024 Apr 24. ]. Available from: https://hbr.org/2023/04/why-business-leaders-must-resist-the-anti-esg-movement
  232. 232. Picciotto R. The logic of evaluation independence and its relevance to international financial institutions. In: Lamdany R, Edison H, editors. Independent evaluation at the IMF: the first decade. Washington (DC): Independent Evaluation Office of the International Monetary Fund; 2012.
  233. 233. Allen CR, Garmestani AS. Adaptive management. Adaptive management of social-ecological systems. Dordrecht: Springer Netherlands; 2015. p. 1–10. Available from: https://link.springer.com/book/10.1007/978-94-017-9682-8
  234. 234. Cundill G, Cumming GS, Biggs D, Fabricius C. Soft systems thinking and social learning for adaptive management. Conserv Biol. 2012;26:13–20.
  235. 235. Williams BK, Brown ED. Adaptive management: from more talk to real action. Environ Manage. 2014;53:465–79.
  236. 236. Ofir Z. China’s transformation, Part 5. Directed improvisation. In: Evaluation for Transformation [Internet]. 2019 Apr 26 [cited 2024 May 20. ]. Available from: https://zendaofir.com/learning-from-chinas-transformation-part-5-directed-improvisation/
  237. 237. Ang YY. How China escaped the poverty trap. Ithaca (NY): Cornell University Press; 2016.
  238. 238. Ofir Z. Learning from China’s transformation, Part 6. Co-evolution - a “first principle”. In: Evaluation for Transformation [Internet]. 2019 Apr 30 [cited 2024 May 20. ]. Available from: https://zendaofir.com/learning-from-chinas-transformation-part-6-co-evolution-a-first-principle/
  239. 239. Neto M, Sobhani S. Engaging the private sector to achieve the 2030 Agenda. In: UNDP [Internet]; 2023 Oct 3 [cited 2024 Jun 20. ]. Available from: https://www.undp.org/policy-centre/istanbul/blog/engaging-private-sector-achieve-2030-agenda-0
  240. 240. Rockefeller Philanthropy Advisors. Philanthropy and the SDGs: getting started; 2019. Available from: https://www.rockpa.org/wp-content/uploads/2019/04/04-12-RPA-SDG-Guide-A-Final-WEB.pdf
  241. 241. Shome S, Hassan MK, Verma S, Panigrahi TR. Impact investment for sustainable development: a bibliometric analysis. Int Rev Econ Financ. 2023;84:770–800.
  242. 242. Ofir Z. Evaluation – a Beautiful practice. In: Evaluation for Transformation [Internet]; 2016 Mar 9 [cited 2024 Jun 19. ]. Available from: https://zendaofir.com/evaluation-a-beautiful-practice/