My Publications

Evaluating complex development programs: Integrating complexity thinking and systems analysis 

Michael Bamberger and Aaron Zazueta

This is Chapter 18 of a forthcoming Research Handbook on Program Evaluation edited by Katherine Newcomer and Steve Mumford (scheduled for publication in late 2024)

Abstract

This chapter provides an introduction to the evaluation of complex programs.  We discuss the importance of a complexity focus, and then review elements of complexity and systems thinking.  We introduce an Integrated Complexity-Responsive and Systems (ICS) evaluation approach, integrating compatible concepts from both systems thinking and the complexity literatures.  We use ICS as a generic term  that covers all of the of the evaluation tools and techniques that can be used, singly or in combination, to evaluate the outcomes and impacts of complex interventions operating in complex contexts.  The steps of the approach include: (a) understanding complexity in programs and in the systems in which they are embedded, (b) developing a theory-based evaluation framework, (c) selecting from among ICS approaches, (d) designing the evaluation for each approach, and (e) assessing the program’s contribution to broader development objectives. We also offer case studies illustrating how to implement the ICS approach.

Links to the two annexes included in the chapter

Annex 1  Evaluation tools and techniques for each stage of the complexity-responsive evaluation frameworkFor descriptions of each of the 5 steps see Section 2 
Stage of the evaluationTopics coveredReferences [the full references are given in the reference list at the end of the text]
Step 1 Understanding the complexity scenarioUnderstanding the system 
Conceptualizing systemsDavis and Guevara (2016)Raimondo et al (2016) p.34 Innovation network system mappingZazueta and Bahramalian et al 2021 
Understanding interactions among large numbers of stakeholders and organizationsUNIDO 2019Zazueta, Le et al 2021UNIDO 2020
Defining system boundariesWilliams and Hummelbrunner 2029 Chapter 19
Understanding dynamic processes of change
Describing and measuring interactions and change within and between project componentsBurke (2007)Williams and Hummelbrunner 2029 Chapter 1 
Describing the processes through which a project is implementedBeach and Pedersen (2013)
Describing and monitoring interactions among agents involved in a projectMorrell (YouTube reference)Brown, Page, Riolo, and Rand (2004) GEF/IEO 2012UNIDO 2019
Identifying short and long-term change indicators (DEMANTEL model)Si et al (2018)
How processes of change are influenced by initial conditions, traces left by actions of other actors, and resistance to changeMorell (2021)
Step 2: Theory-based evaluation frameworksComplex adaptive system (CAS) frameworkZazueta and Bahramalian (2021)GEF, IEO (2012) and (2016)Green Climate Fund (2022)
Theory of changeWeiss (1995)
Realist evaluationPawson and Tilley (1997)
Contribution analysisMayne (1999)
Critical systems practiceJackson (2020)
Blue marble evaluationPatton (2020)
Purposeful program theoryFunnell and Rogers (2011)
Step 3: Identifying components to be analyzed and selecting the unit of analysisDefining the unit of analysisGertler et al (2011) Chapter 2Bamberger, Raimondo and Vaessen (2016) Chapter 7
Step 4: Defining the methodology for the component evaluationsTracking and managing outcomes 
Tracking stakeholder opinions and perceptions on outcomesWilliams and Hummelbrunner (2009) Chapter 4
Understanding power relations and who can influence decisionsSi et al (2018)Zazueta, Bahramalian et al (2012)
Measuring changes over time and evaluating impacts 
Identifying the necessary and sufficient conditions of attributes required to achieve an outcome in groupsLam and Ostram (2010)GEF/IEO (2016)
Assessing the contribution of different organizations to the achievement of outcomes – when it is not possible to use quasi-experimental designs (contribution analysis)Mayne (2012)
Assessing project outcomes and impacts using conventional experimental, quasi-experimental designs and case-based comparisonsVaessen (2016)Vaessen, Raimondo and Bamberger (2016) Chapter 4
QCALam and Ostram (2010)
Contribution analysisMayne (2012)GEF/IEO (2016)
Randomized control trials and quasi-experimental designsGertler et al (2011)Vaessen, Raimondo and Bamberger (2016) Chapter 4
Interpretative, qualitative approaches (naturalistic enquiry, grounded theory). Structured approaches: configurational, within and between case comparisons, Narrative designs: participatory and empowerment evaluationsVaessen, Raimondo and Bamberger (2016) Chapter 4
Agency designs: learning by doing, policy dialogue, collaborative and action researchUNIDO (2019)
Step 5:  Reassembling the different components of a program to assess the big pictureSystems modellingDescriptive and inferential statisticsComparative case study approachesPortfolio analysisReview and synthesis approachesRating scalesBamberger, Raimondo and Vaessen (2016) Chapter 7
Articulating the program theoryCase study No. 2 Part 2 Case study No. 1 Part 1 Zazueta, B and Bahramalian 2021GEF, IEO 2012 and  2016Green Climate Fund 2022
Describing and measuring interactions and change within and between project componentsBurke (2007) in Williams and Imam Chapter 3Williams and Hummelbrunner 2029 Chapter 1
Describing the processes through which a project is implementedBeach and Pedersen 2013
Identifying short and long-term change indicatorsSi et al 2018GEF 2022GEF 2012GEF/IEO 2016Green Climate Fund 2022Si et al 2018UNIDO 2019UNIDO 2021Zazueta, Bahramalian 2021Green climate fund 2021 
Capturing stakeholders opinions and perceptions on outcomesWilliams and Hummelbrunner 2009 Chapter 4Wilson-Grau and Britt (2012)SMARTFish
 
Understanding power relations and who can influence decisions and outcomesSi et al 2018Zazueta, Bahramalian et al 2021 
Identifying the necessary and sufficient contribution of attributes require to achieve an outcome in groupsLam and Ostram (2010) GEF/IEO 2016
Assessing the contribution of different organizations to the achievement of outcomes – when it is not possible to use a quasi-experimental designMayne (2012)
Assessing project outcomes and impacts using conventional experimental and quasi-experimental designsVaessen (2016) 
Annex 2 Recommended readings
 
Bamberger, M; Vaessen, J & Raimondo, E (editors) (2016) Dealing with complexity in development evaluation: A practical approach. Sage Publications.This edited volume provides an overview of current thinking on complexity. It contains 5 parts: (1) dealing with complexity in development evaluation:  framework, (2) methodological approaches, (3) Emerging data and innovative approaches to deal with complexity in development evaluation, (4) the institutional challenges, (5) case studies.
Beach, D and Pedersen,R,B  (2013) Process-tracing methods: Foundations and guidelines. University of MichiganModeling and tracing the stages in the process of project implementation. A widely-used approach for defining and tracking the how projects are designed and implemented
Byrne, D & Callaghan, G (2015) Complexity theory and the social sciences: The state of the art. Routeledge.Discussion of the contribution of case based methods to the analysis of complexity with chapters illustrating a wide range of case-based approaches, including Qualitative Comparative Analysis (QCA)
Byrne, D & Ragin, C (eds) (2009). The Sage Handbook of case-based methods. SageOverview of the theory and practice of research into complexity
Forss, K, Marra, M,& Schwarz,R (eds) (2011) Evaluating the complex: attribution, contribution and beyond. Transaction.Readings on complexity-focused evaluations of development programs.
Funnell, S and Rogers, P (2011) Purposeful program theory: Effective use of theories of change and logic models. Jossey-BassThe book explains all of the stages in the design and use of program theory, including the theory of change, and with a specific focus on complexity.  They also introduce the widely-used framework of simple, complicated and complex evaluations.
Furubo, J, Rist, R, Speer, s (editors) (2013) Evaluation and turbulent times: Reflections on a discipline in disarray. Transaction.Readings on how evaluation is applied in situations of conflict and uncertainty. The focus is not directly on complexity, but many of the situations where the evaluations are applied are complex. Interesting discussion of the differences between political and economic turbulence and how this affects the evaluation approach.
Gallager, L et al 2020Combines transdisciplinary research design with  participatory causal loop diagramming processes, scenario modeling, and a new resilience analysis method to identify and test anticipated water-energy-food risks in Cambodia. Combines modeling techniques and participatory processes
Gertler, P et al (2011) Impact evaluation in practice. World BankReview the economic approach to quantitative impact  evaluations, including: causal inference and counterfactuals, randomized designs, regression discontinuity, difference-in-difference, and propensity score matching and other matching techniques
Global Evaluation Initiative. https://www.betterevaluation.orgA global resource site with material on evaluation courses and reference material
Jackson, M (2019) Critical systems thinking and the management of complexity.  WileyProvides a framework to assist decision makers and other stakeholders to understand “volatile, uncertain and complex and ambiguous world of general complexity”.  The focus is on managing, rather than evaluating, complexity, but the approach provides guidance on how to articulate complexity frameworks that is very helpful for step 1 on the approach discussed in this chapter.
Masset, E, Shreshta, S & Juden, M (2021) Evaluating complex interventions in international development. CEDIL Methods working paper No. 6. Available online: https://doi.org/10.51744/CMWP6.“ reviews promising methods for the evaluation of complex interventions that are new or have been used in a limited way. It offers a taxonomy of complex interventions in international development and draws on literature to discuss several methods that can be used to evaluate these interventions”
Mayne, J (2012) Contribution analysis: coming of age? Evaluation, 18(3), 270-280The approach provides practical guidance on how to analyze cause and effect in the many contexts where it is not possible to use experimental or quasi-experimental designs.  The approach is widely used in situations where there are multiple agencies involved in a program, and the challenge is to assess the contribution of a particular agency.
Morrell, J (2012) Evaluation in the face of uncertainty: Anticipating surprise and responding to the inevitable. GuildfordThis is a useful resource for addressing the often-overlooked issues of uncertainty, surprise and unintended outcomes.  The approach addresses many of the issues that complexity theorists often refer to “emergence”.  A strength of the book is that includes 18 case studies that help understand the concepts and the evaluation approaches.
Patton, M (2015) Qualitative research and evaluation methods. SageBroad review of participatory, observational and interview techniques and the rationale for their use.
Patton, M (2020) Blue Marble EvaluationBlue Marble evaluation addresses global, multidisciplinary issues that are beyond the scope of most evaluation approaches.  The focus is on the Anthropocene, and the recognition that these issues are complex.  While the issues are global, Patton emphasizes that they must build on the integration of local issues.  Transboundary issues must be addressed and silos broken down.
Pawson, R and Tilley, N (1997) Realistic evaluation Sage Publications; and Pawson, R (2013) The science of evaluation: A realist manifesto. Sage Publications.Outcomes are the result of the interactions between context and the mechanisms developed by the project. The focus is on outcome configurations, rather than single outcomes.  Programs have different outcomes in different contexts and for different people, and asks questions such as “Who benefits and who does not?”, “How?”, “where”, why?” and “when?”
Tashakkori[MB1] [MB2] , A; Johnson, B and Teddlie, C (2021) Foundations of Mixed Methods research: Integrating quantitative and qualitative approaches in the social and behavioral sciences. Sage.Comprehensive coverage of mixed methods designs which is an essential element of the complexity approach proposed in this chapter.
(Van der Merwe et al., 2019) “ The Cognitive Edge SenseMaker® presents a  method “for capturing and making sense of people’s attitudes, perceptions, and experiences. It is used for monitoring and evaluation” mapping ideas, mind-sets, and attitudes; and detecting trends and weak signals.  Various mapping techniques
Williams, B and Hummelbrunner, R 2011  Systems concepts in Action: A practitioners toolkit. Standford Business Books.A comprehensive, but not too technical overview of a wide range of systems analysis approaches that are applicable to the approach discussed in this chapter.
Williams, B and Imam, I (2007) Systems concepts in Evaluation: An expert anthology.  (2007) American evaluation Association.This anthology includes case studies illustrating all of the techniques described in the Williams and Hummelbrunner text.
Wilson Grau and Britt (2012) Outcome harvestingUsing outcome harvesting to capture the views of different stakeholders on the significant project outcomes
Zazueta, A and Bahramalian  et al 2021  Modeling of complex systems , Influential Relation Map and use of visualization and participation to assess contributions to long term goals.

 

Books and book-length publications

Deprez, S., Majale, M and Bamberger, M (2021) Impact Evaluation of UN-Habitat Housing Approach to Adequate, Affordable Housing and Poverty Reduction 2008 – 2019. Nairobi, UN-Habitat.

https://unhabitat.org/sites/default/files/2021/03/impact_evaluation_of_un-habitats_housing_approach_to_adequate_affordable_housing_and_poverty_reduction_2008-2019.pdf

York, P and Bamberger, M (2021) Measuring Results and Impact in the Age of Big Data: The nexus of evaluation, analytics and digital technology. New York: The Rockefeller Foundation

Bamberger, M and Mabry, L (2020 Third edition) Realworld Evaluation: Working Under Budget, Time, Data and Political Constraints. Thousand Oaks, CA: Sage Publications

Bamberger, M (2020) “Big data analytics and development evaluation: Optimism and caution” in Garcia, O and Kotturi, P (eds) Information and communication technologies for development evaluation. Chapter 3: 76-110. New York: Routledge

Bamberger, M (2017) Integrating gender into project-level evaluations. Economic Cooperation Group. Published by the African Development Bank. Abidjan. Ivory Coast.

Bamberger, M., Vaessen, J and Raimondo, E (2016) Dealing with Complexity in Development Evaluation: A practical Approach. Thousand Oaks, CA: Sage Publications

Bamberger,M., Segone, M and Tateossian, F. (2016) Evaluating the Sustainable Development Goals: With a “No-one Left Behind” lens through equity-focused and gender-responsive evaluations. New York: UN Women.

Bamberger, M (2016) Integrating Big Data into the Monitoring and Evaluation of Development Projects. New York: UN Global Pulse.

Bamberger, M (2014) Measuring resilience: Lessons from the design and implementation of monitoring and evaluation methodologies of the Horn of Africa Risk Transfer for Adaptation (HARITA/R4) Rural Resilience Initiative.  A case study.  Oxfam America and the Rockefeller Foundation.  June 2014 available at: http://www.oxfamamerica.org/static/media/files/Measuring_Resilience_-_R4_Case_Study.pdf and at:  Policy & Practice http://policy-practice.oxfamamerica.org/publications/measuring-resilience-r4-case-study/

Bamberger, M., Segone, M and Reddy, S (2014) National evaluation policies for sustainable and equitable development:  How to integrate gender equality and social equity in national evaluation policies and systems.  EvalPartners, UNWomen and IOCE.  http://www.mymande.org/sites/default/files/files/NationalEvaluationPolicies_web-single-color(1).pdf 2013

Raftree, R and Bamberger, M (2014) Emerging Opportunities: Monitoring and Evaluation in a Tech-enabled World. Discussion Paper. New York: The Rockefeller Foundation

Bamberger, M (2013) Engendering M&E  Special Series on “The Nuts & Bolts of Monitoring and Evaluation Systems” No. 27.  Poverty Reduction and Economic Management Department. Washington, DC: World BankWorld Bank 

Bamberger, M (2012) Introduction to mixed methods in impact evaluation.  Impact Evaluation Notes.  No. 3. Washington, DC: InterAction.

Bamberger, M., Rugh, J and Mabry, L (2012 Second edition) Realworld Evaluation: Working under budget, time, data and political constraints. Thousand Oaks, CA: Sage Publications

Bamberger, M and Segone, M (2011) How to Design and Manage Equity-focused evaluations. New York: UNICEF

Bamberger, M (2009) Institutionalizing Impact Evaluation with the Framework of a Monitoring and Evaluation System. Washington, DC: Independent Evaluation Group, The World Bank.

Bamberger, M (2009) “Enhancing the utilization of evaluation for evidence-based policymaking” in Marco Segone (editor) Bridging the Gap: the role of monitoring and evaluation in evidence-based policymaking. UNICEF. http://www.unicef.org/ceecis/evidence_based_policy_making.pdf

Bamberger, M., Rugh, J and Mabry, L (2012 First edition) Realworld Evaluation: Working under budget, time, data and political constraints. Thousand Oaks, CA: Sage Publications

Bamberger, M (2006)“Evaluation Capacity Building” in Marco Segone and Ada Ocampo (editors) Creating and Developing Evaluation Organizations: Lessons Learned from Africa, Americas, Asia, Australasia and EuropeInternational Organization for Cooperation in Evaluation.(Published by UNICEF). http://www.unicef.org/ceecis/07_IOCE_EvalOrg.pdf

Bamberger, M (2006) Conducting Quality Impact Evaluations Under Budget, Time and Data Constraints. Washington, DC: Independent Evaluation Group, The World Bank

Bamberger, M and Ooi, E (editors)(2005) Influential Evaluations: Detailed case Studies. Washington, DC: Operations Evaluation Department, World Bank

Bamberger, M with Ayata, A., Greenberg, M., Lingen, A., Sen, K.M., Pearson, R and Sarr, Fatou (2006) Evaluation of gender mainstreaming in UNDP. Evaluation Office. New York: UNDP

Bamberger, M and Ooi, E (2004) Influential Evaluations: Evaluations that improved performance and impacts of development programs. Washington, DC: Operations Evaluation Department, The World Bank. http://lnweb18.worldbank.org/oed/oeddoclib.nsf/24cc3bb1f94ae11c85256808006a0046/67433ec6c181c22385256e7f0073ba1c/$FILE/influential_evaluations_ecd.pdf

Bamberger, M with Kulessa,M., Alailima, P., Flaman,R and Forrester, A.  (2004) Evaluation of the Second Global Cooperation framework of UNDP. Evaluation Office. New York:  United Nations Development Programme.

 Bamberger, M and Fujita, N (2003)  Impact Evaluation of Development Assistance: A Practical Handbook for Designing Methodologically Sound Impact Evaluations under Budget, Time and Data Constraints.  Foundation for Advanced Studies in International Development. Tokyo. Japan 

Bamberger, M and Maramba, P (2001).  A Gender-Responsive Monitoring And Evaluation System For Rural Travel And Transport Programs In Africa.  Africa Rural Travel and Transport Program. World Bank.

Bamberger, M (editor) (2000) Integrating Quantitative and Qualitative Research in Development Projects. Directions in Dev elopement. Washington, DC: World Bank.

Valadez, J and Bamberger, M (1994) Monitoring and Evaluating Social Programs in Developing Countries. A Handbook for Policymakers, Managers, and Researchers  EDI Development Series. Washington D.C. The World Bank.  url: http://www-wds.worldbank.org/servlet/WDS_IBank_Servlet?pcont=details&eid=000009265_3970716143510

Book chapters

Bamberger, M (2020) “Big data analytics and development evaluation: Optimism and caution” in Garcia, O and Kotturi, P (eds) Information and communication technologies for development evaluation. Chapter 3: 76-110. New York: Routledge

Bamberger, M (2012) “Reconstructing baseline data for impact evaluation and results management”.  In Lopez-Acevedo, Krause and MacKay (editors)  Building better policies: The Nuts and Bolts of Monitoring and Evaluation Systems. Washington, DC: World Bank

Bamberger, M (2011) “Reconstructing baseline data for impact evaluation and results management”.  In Lopez-Acevedo, Krause and MacKay (editors)  Building better policies: The Nuts and Bolts of Monitoring and Evaluation Systems. World Bank

Bamberger, M., Rao, V and Woolcock, M (2010) Using mixed methods in monitoring and evaluation: Experiences from international development. in Tashakkori and Teddlie (editors) Sage Handbook of Mixed Methods in Social and Behavioral Research. Second Edition (Chapter 24 pp. 613-642)

Bamberger, M., Rao, V and Woolcock, M (2010)  t(February 1, 2010). BWPI Working Paper No. 107. Available at SSRN: http://ssrn.com/abstract=1545914

Bamberger, M and Rugh, J (2009) “RealWorld Evaluation: Conducting evaluations under budget, time, data and political constraints”  in Marco Segone (Editor) Country-led Monitoring and Evaluation Systems:  Better Evidence, Better Policies and Better Development Results. New York: UNICEF http://www.ceecis.org/remf/Country-ledMEsystems.pdf

Bamberger, M and Podems, D (2002) “Feminist Evaluation in the International Development Context” in Feminist Evaluation.  New Directions for Evaluation No. 96 Winter 2002

Bamberger, M, Blackden,M,  Fort, L and Manoukian, V (2002) “Gender” chapter of A Sourcebook for Poverty Reduction Strategies. World Bank. Washington D.C.

Articles

Bamberger, M (2018) “New challenges and opportunities for Multi and Mixed Methods Research (MMMR) in the Evaluation of International Development Programs”.  International journal of multiple research approaches, 2018 vol. 1, no. 1, 168–182 
https://doi.org/10.29034/ijmra.v10n1a11 

Raftree, L, Bamberger, M and Olazabal, V (2016)  “The role of new information and communication technologies in equity-focused evaluation: Opportunities and challenges”.  Evaluation 2016, Vol 22(2) 228-244.  

Bamberger, M (2015) “Innovations in the use of mixed methods in real-world evaluation”.  Journal of Development EffectivenessVol 7 Issue 3 Sept 2015 317-326

Bamberger, M (2009) “Strengthening the evaluation of development effectiveness through reconstructing baseline data  Journal of Development Effectiveness.  Volume 1 No. 1 March 2009.

Bamberger, M and White, H (2008). “Impact Evaluation in Official Development Agencies.”  IDS Bulletin Volume 39 No. 1 March 2008.

Bamberger, M and White, H  (2007)  “Using Strong Evaluation Designs in Developing Countries: Experience and Challenges”  Journal of MultiDisciplinary Evaluation.  October 2007 Vol 4 No. 8 pp. 58-73 [http://survey.ate.wmich.edu/jmde/index.php/jmde_1/article/view/31/78]

Bamberger, M (2006)  Book Review “Removing Unfreedoms:  Citizens as Agents of Change in Urban Development” (Edited by Jane Samuels. ITDG Publications). American Journal of Evaluation. Vol 27. No. 4 December 2006 498-502.

Bamberger, M, Rugh,J, Fort, L and  and Church, M (2004) “Shoestring Evaluation: Designing Impact Evaluations under Budget, Time and Data Constraints” American Journal of Evaluation 25(1), 5-38. 

Bamberger, M (2001). “Evaluation in developing countries: experience with agricultural research and development and the Annotated Bibliography of International Program Evaluation” Book Review Article.  American Journal of Evaluation. 22(1), 117-122.

Bamberger, M, Kaufmann, D and Velez, E (2000)  Interhousehold transfers: using research to inform policy.  PREMNotes. Development Economics Department. Washington, D.C.: The World Bank.

Bamberger, M (2000) “The Evaluation of International Development Programs: A View from the Front” American Journal of Evaluation. Vol. 21 No. 1 Winter 2000.