TIME keynote, November 2025
And:
Everything we—researchers and practitioners—have learned, points to the need to be modest about our knowledge, attuned to the role of context, and open to learning from different approaches
As development resources become scarce, the need to focus on learning becomes stronger than ever, especially for addressing the biggest questions
We went to study why people fought, what sustained the conflict, what explained the horrific abuses. Following work in Casamance and Mali and Peru, Uganda (Weinstein).
Focus on
Crisis when we find:
Both problems addressed by bringing in a focus on DDR.
The ex-combatants issued a nationwide OK for our work if we included questions on ex combatant welfare.
The (Irish) UN DDR coordinator wanted to learn about program effects and gave full helicopter access in return for including a DDR module.
DDR:
A “process of disarmament, demobilization, and reintegration has repeatedly proved to be vital to stability in a post-conflict situation” (United Nations 2000, 1; italics added).
But really little evidence.
At the time there was not (to my knowledge) a single study comparing places with and without DDR, or individuals exposed and not exposed
A colleague’s study highlights inferential challenge when you do not have a control group.
Blattman, short term patterns
Co-designed modules with UN, NGOs, and ex-combatants
Sought to interview participants and non participants nationwide
Reached over 1000 ex combatants from all groups throughout the country
Sierra Leone
Huge commitment of UN and NGO partners, invested in the question, and open to learning (Still, saw some manipulation risks)
Clash between qualitative insights and quantitative findings and challenges to integration of findings
Is DDR a good idea? Maybe! Clarification of twin goals of interventions: DDR interventions might make sense politically, even if they have no specific benefits for individual participants
Messaging: not ‘DDR does not work (at this level),’ but ‘no evidence that it does’
This is what’s called an “observational” study. The study was conducted after project implementation. Individuals self select into the intervention. Some do, some “self-reintegrate.”
We worried about three big risks to inference:
Selection biases: do people self-reintegrate because they are doing well? Or because they are doing very poorly?
Sampling biases: are successful reintegrators less sampleable?
Spillover biases: are non participants doing well because their friends did participate?
These risks are quite general across impact evaluations.
But they can all be minimized with randomized intervention.
Big push at this time to broaden partnerships between researchers and practitioners (EGAP network and others: DIME, 3IE, JPAL, IPA; DfID was a major actor).
Dream was (is!) to maintain twin goals:
scholars not plumbers: learn about what works to do good in the world and to get a sharper handle on generalizable social processes.Joined forces with the International Rescue Committee (IRC) as they started a learning agenda.
Two bigger RCTs with IRC in Liberia and Congo.
“Community-driven development operations produce two primary types of results: more and better distributed assets, and stronger, more responsive institutions.” (The World Bank)
But: almost no evidence to this effect at the time and almost none now.
The intervention:
Objectives: economic and governance:
Some clear priors:
“This program is exciting because it seeks to understand and rebuild the social fabric of communities. […] It’s a program that starts to rebuild trust, it’s a grassroots democratization program.”
Strong identification:
Strong measurement strategy:
Pre-registration: lots of sign offs, buy-in, enthusiasm
On measure after measure after measure, the distribution of outcomes in treatment and control were identical.
A collection of other RCTs similarly found null or mixed results, Liberia, Sierra Leone, Afghanistan…
A lot of work by IRC and DfID to make sense of this.
Major implication seems to be picked up: CDR (CDD) might make sense for implementing projects but not for improving governance. Allocate resources accordingly.
Perhaps bigger check on the idea that externals can (or should) be trying to alter local governance structures.
Since then:
IRC has largely shifted out of governance interventions
Irish development NGOs and academic researchers playing a big role.
What kinds of findings are emerging?
An average increase of less than 8 percent of the current average profit, and less than 5 percent of the standard deviation, is not likely to be a transformative change for a household. (Meager 2019)
(Dunning et al. 2019) No evidence that getting information to voters about political malfeasance affects vote choice!
Recent Liberia study (Blattman) looks specifically at Cognitive Behavioral Therapy with at risk youths post conflict.
CBT shows some surprisingly encouraging results after 10 years
CBT Liberia project; Blattman et al
Community monitoring reduces extraction rates (and has other benefits) but heterogeneous and only observed across studies.
Meta-analysis of 115 studies of 72 UCT programs in middle and low income countries: strong and positive average treatment effects on 10 of 13 outcomes: monthly household total and food consumption, monthly income, labor supply, school enrollment, food security, psychological well-being, total assets, financial assets, and children height-for-age. Crosta et al
Multifacted components to help move people out of poverty traps Banerjee et al. (2015) (assets, skills, training)
A lot of work, including by TIME, about optimizing the design – e.g. understanding gender aspects.
Many nulls (both in individual studies and meta-analyses)
Much heterogeneity
Heterogeneity across fields
It seems lots and lots of interventions don’t work
We systematically over-estimate program effectiveness
But for all that, clearly cumulation of knowledge
They are more likely to:
An irony for impact evaluation is that, for the most part, there is not good evidence on the impact of impact.
They have all the features that make impact hard to study
In preparation for the talk I spoke with a few people working on policy and research at
Where did they see research having an impact?
Exposed policy makers to single experimental study, single observational study (Ext), meta-analyses of experimental studies
Learning about learning: Practitioner put weight on RCTs, on meta-analyses, and on negative evidence
We saw null results from the Sierra Leone observational DDR project
20 years later, what do we know?
We urged the UN and others to implement some DDR RCTs
We tried with UN in Haiti but initial fieldwork suggested the program was unlikely to be effective
There is still not a single completed RCT of a UN style DDR program
Major design questions remain unanswered:
However the Blattman study is now having influence (though it’s still just one study).
We are currently working on a German funded project in Nigeria directly influenced by the Blattman study, following literature reviews.
In general, it is easier to use evidence to stop programmes. … It has far larger reputational risk – there is evidence and documents that could be requested … that state advice the minister should not do something that is a risk.
It is a much smaller risk if a minister did not do what someone advised based on evidence. The argument is that there are many good things they could do so they can pick and not pick.
CDR: Reduced in some portfolios, restructured in others
Microcredit
Bednets example
Citizens were historically charged for bednets in malarial zones based on quasi-behavioral arguments.
Cohen and Dupas varied whether bednets are subsidized or not and found no deterioration in quality of usage (but gains in uptake)
In 2009, the British government cited the study in calling for the abolition of user fees for health products and services in poor countries. (IPA)
Targetting
World Food Program recently worrying about effectiveness of aid targetting algorithms:
Many of these:
WFP is now regularly doing “lean” evaluations: rapid low cost randomized pilot interventions to decide whether or how to proceed.
WFP Jordan study uses a school feeding menu change pilot to learn about optimal design, boosting school attendance
When it should have:
World Food Program decided to largely shift to cash, before the evidence came in. But were able to refer to the evidence to support the policy.
DfID also shifted to cash before the evidence, but relied on evidence later to defend it
Germany much more hesitant
When it maybe shouldn’t have:
Community monitoring of health workers
Build trust at all levels
Invest in partnerships: Joint communities of practice, be present
Work across methodological divides: seek to integrate quantitative and qualitative knowledge rather than treating them as alternative perspectives
Engage ethically: Minimize interference; step back when unexpected risks arise; adhere to principle of justice
Align “central” (HQ, global) learning agendas and “local” agendas: neither pure top down or pure bottom up approaches are working, for different reasons
Communicate: Use MOUs and share PAPs: clarity about the purpose and what implications will be drawn from different findings
A best practice model from WFP:
Twin focus:
Biggest lesson is maybe that many development interventions are likely ineffective and those that are likely are not effective everywhere
Expect to have your priors challenged.
Knowing that things so often do not work out as we expect has welfare implications and political implications
Two responses:
Approach A is to focus on “best buys” in development
A lot to be said for this, especially when resources are scarce and you want to be sure of impact.
But a focus on doing “what works” means reduced investment in things that
This:
Approach B: Focus on learning. In particular: use impact evaluations more as a tool for learning than for accountability
Accountability goal
Can threaten trust
Discourage risk taking
Impact evaluation a noisy tool for accountability anyhow
implementers are not (always) responsible for impact; impact depends on the design (commissioners also responsible)
Learning agenda:
Knowledge is a public good and the impacts (in principle) can be large: a contribution in its own right
Even if the evidence on individual projects is be noisy it cumulates
Implications for what interventions or aspects of interventions to be examined:
Learning agenda is a shared agenda, shared with development partners