Effects of economic and social incentives on bureaucratic quality

Experimental Evidence from Sierra Leone

Maarten Voors, Macartan Humphreys, Salif Jaiteh, Niccolo Meriggi, Carlo Prato, Peter van der Windt

Outline

  • Motivation
  • Model
  • Design
  • Results
  • Structural model

Motivation

Motivation

A challenge:

  • Quality of service provision is low in many developing countries

  • Recruiting and incentivizing service providers is difficult

A case: post Ebola epidemic in West Africa

  • Caught many by surprise, no early warning
  • Need for stronger decentralized bureaucracy
  • 7/10 emerging infectious diseases spill over from animals

Motivation

  • Local delivery models since 1990s: locally recruited actors, to deliver interventions or to do government tasks

  • Popular, …. but is little known when and how they work

  • Here: focus on agricultural extension workers

  • Largely unpaid, using entrepreneurial worker model

  • Agripreneur programs (World Bank, etc), but part of government system

Motivation

  • Most people live in rural areas and crucially depend on agriculture (crops and livestock)
  • Local animal welfare and disease risk go hand in hand
    • Sick animals are often eaten and cause health risk
    • Diseases are highly contagious
  • Need for surveillance: case-reporting, contact tracing, vaccine tracking, data management, response

Motivation

Promising work on improving service provision:

  1. Recruitment

  2. Community monitoring

  3. Financial incentives

We show:

  • evidence 2 and 3 matter
  • little movement on 1.
  • we use a structural model to explore why, and when these might be more effective.

Past Work on Recruitment

Theoretically:

Empirically:

Past Work on Recruitment

Past Work on Financial Incentives

By and large a success:

Past Work on Community Monitoring

Monitoring: involving communities

  • Original enthusiasm: “Power to the People” (Björkman and Svensson 2009)

    • How to improve service delivery and health outcomes in rural Uganda
  • Intensive treatment

    • Assess quality of services provided in clinic/community-specific report cards

    • Disseminate report cards to community and to clinic staff

    • Develop joint action plans

    • Interface meetings between clinic staff and community

Monitoring: involving communities

  • A small sample (n = 50)

  • And amazing results:

    • Infant weights significantly higher

    • Under-5 mortality 33% lower

    • Immunization rates higher

    • Staff absenteeism and waiting times at clinics lower

Raffler et al (2024) replication …

  • Same country, larger sample, teasing apart treatment components

Where does this leave us?

  • Candidate explanation: differences as baseline

Policy question

  1. Are performance-based payment schemes more or less effective than community monitoring mechanisms?
  2. Are there interaction effects between recruitment strategies and incentive mechanisms?
  • E.g. community monitoring may be less effective when service providers are selected by the state
  • They may be more effective if selected by communities, trusted, knowledgeable, present

Study design

We study these questions in the context of a randomized agricultural intervention in Sierra Leone.

Core design elements (more on context to come)

treatment label
T1 Agent is selected by community (T1 = 1) or by state (T1 = 0)
T2 Community can impose social sanctions on agent (T2 = 1) or not (T2 = 0)
T3 Agents receive pay for performance (T3 = 1) or not (T3 = 0)

Model

Model {.smaller} (base model)

After Francois (2000). A worker has type \(\theta = (\eta,\mu)\) where:

  • \(\eta \in \{ \underline{\eta} , 1 \}\) denotes embeddedness
  • \(\mu \in \{ 0 ,\overline{\mu} \}\) denotes intrinsic motivation

Drawn according to:

\({\mu}{\backslash}\eta\) \(\underline{\eta}\) 1
0 \(\frac{1+ \rho}{4}\) \(\frac{1- \rho}{4}\)
\(\overline{\mu}\) \(\frac{1- \rho}{4}\) \(\frac{1+ \rho}{4}\)

So \(\rho\) governs the correlation between embeddedness and motivation: embedded people are possibly more effective, but are they more likely to be highly motivated?

Selection (\(T_s\)) and type

  • Under state selection (\(T_s=0\)) the worker is a random draw from the candidate pool.
  • Under villager selection (\(T_s=1\)), the worker is drawn from the pool with high embeddedness.

So state is more likely to select workers with high intrinsic motivation if and only if \(\rho < 0\).

Performance

  • Workers choose effort \(e\)

  • Performance depends on effort and embeddedness: \(\pi = e \eta\)

  • Stochastic performance signal \(\sigma\), which is high (\(\sigma=1\)) with probability \(\pi\) or low (\(\sigma=0\)) with probability \(1-\pi\)

  • Cost of effort: \(c(e) = \frac{e^2}{2\kappa}\) (alternative use of time, with wage elasticity of effort \(\kappa\))

Villager (Monitoring, \(T_m\))

  • A representative villager selects a social sanctioning plan, which depends on the performance signal \(\sigma\):

\[s=(s(0),s(1)) \in [-T_m, T_m]^2.\]

  • Interpretable as commitments (additional) to reward (\(s(\pi)<0\)) or punish (\(s(\pi)>0\)) performance.

  • in line with Community Action Plan in experimental intervention (in CM)

Villager (Monitoring, \(T_m\))

  • Punishment may be psychologically costly, according to \(\gamma\).

Payoff to villager:

\[u(\pi, s) = \pi - \gamma s.\]

  • term \(\zeta\) captures village norms, impacts the probability that punishment is not costly for the villager

In crowding out version we have \(u(\pi, s) = \pi - s(\gamma - \xi T_p).\) (and replace \(s\) with \(-r\) to indicate rewards rather than sanctioning interpretation)

CAHW utility

Under pay-for-performance (\(T_p=1\)) and high performance signal (\(\sigma=1\)), the worker receives a payment normalized to 1

In addition:

  • \(\phi\): sensitivity to social sanctions
  • \(\mu\): intrinsic benefits from performing well
    • recall \(\pi = e \eta\)

Putting it all together:

\[w(e;s,\sigma;\theta) = \sigma T_p + \pi\mu - \phi s(\sigma) T_m - \frac{e^2}{2 \kappa},\]

Timeline

The sequence of events is as follows:

  1. Worker is selected by village (\(T_s=1\)) or state (\(T_s=0\)).
  2. Villagers and worker learn whether the latter is subject to community monitoring (\(T_m=1\)) and/or P4P (\(T_p=1\)).
  3. If \(T_m=1\), villager with guilt parameter \(\gamma\) commits to a conditional plan of social sanctioning \(\{s(\sigma)\}_{\sigma \in \{0,1\}}\).
  4. The worker chooses effort \(e\).
  5. The worker’s performance signal \(\sigma\) is realized.
  6. Payoffs are realized and the game ends.

Solution concept

An equilibrium is a pair of strategies \(\{ s,e(s;\theta) \}\) such that

  1. effort is individually rational for each type given the sanctioning strategy
  2. the sanctioning strategy maximizes the villager’s expected payoff given anticipated effect on the workers’ effort \(\hat{e}(s;\theta)\),
  3. expectations are correct (\(\hat{e}(s;\theta)=e(s;\theta)\) for each \(\theta\)) :

\[e(s;\theta) \in \arg \max_{e \in [0,1]} \mathbb{E}_{\pi } \Big\{ w(e;s,\sigma;\theta) \Big\}\] for \(T_m\)\[s \in \arg \max_{\{ (s(0),s(1)) \in [-T_m,T_m]^2 \}} \mathbb{E}_{\theta,\pi} \Big\{ u(\pi,s) \mid \hat{e}(s;\theta) \Big\}\]

Results

  • Sanctioning (\(T_m\)):
    • following a high signal: always praise the worker \(s(1)=-1\)
    • following a low signal: choose maximal shaming \(s(0)=1\) if guilt is low (\(\gamma=0\)) and minimal shaming \(s(0)=-1\) .
  • Effort:

\[e(s^*;\gamma,\theta) = \begin{cases} \kappa \eta (\mu + T_p + 2 \phi \zeta T_m) & \mbox{if} \quad \gamma =0 \\ \kappa \eta (\mu + T_p) \quad & \mbox{otherwise}. \end{cases}\]

Results: expected performance

Full closed form solution for all treatment conditions:

\((T_m,T_p) | T_s\)

State selection

(\(T_s=0\))

Community Selection

(\(T_s=1\))

Control \({\pi^*(T_s,0,0)}/{\kappa}\) \(\frac{\overline{\mu}}{2} \frac{1 + \rho + \underline{\eta}^2(1-\rho)}{2}\) \(\frac{\overline{\mu}}{2} (1 + \rho)\)
P4P \({\pi^*(T_s,0,1)}/{\kappa}\) \(\frac{\overline{\mu}}{2} \frac{1 + \rho + \underline{\eta}^2(1-\rho)}{2}+ \frac{1+ \underline{\eta}^2}{2}\) \(\frac{\overline{\mu}}{2} (1 + \rho)\) \(+1\)
CM \({\pi^*(T_s,1,0)}/{\kappa}\) \(\frac{\overline{\mu}}{2} \frac{1 + \rho + \underline{\eta}^2(1-\rho)}{2} +(1+ \underline{\eta}^2) \phi \zeta\) \(\frac{\overline{\mu}}{2} (1 + \rho) +2 \phi \zeta\)
P4P + CM \({\pi^*(T_s,1,1)}/{\kappa}\) \(\frac{\overline{\mu}}{2} \frac{1 + \rho + \underline{\eta}^2(1-\rho)}{2} +(1+ \underline{\eta}^2) \phi \zeta + \frac{1+ \underline{\eta}^2}{2}\) \(\frac{\overline{\mu}}{2} (1 + \rho) +2 \phi \zeta +1\)

Hypotheses derived from the model

H1: Village selection increases performance.

  • Explanation: Requires gains from embeddedness not crowded-out by any reductions in intrinsic motivation

H2: Community monitoring increases performance.

  • Explanation: Community monitoring (\(T_m\)=1) incentivizes worker performance through the associated probability of being publicly shamed for poor performance.

H3: Pay-for-performance increases performance.

  • Explanation: Pay for performance (\(T_p\)=1) increases performance due to the probability of receiving a monetary reward.

Hypotheses

H4: Positive interaction village selection and pay-for-performance.

  • Explanation: More productive workers are more responsive to extrinsic incentives (\(T_p\)=1).

H5: Positive interaction village selection and community monitoring.

  • Explanation: For more productive workers it is easier for villagers to commit to sanctioning

Community monitoring and pay-for-performance could be positive (if independent) or negative (if crowding out)

Design

Context: Sierra Leone

  • Poor service provision: mortality is 10% of live birthsqualk

  • Threat of zoonosis is high and expected to increase

  • Animals are essential for livelihoods (90% keep animals), comprising >50% of income

  • Low information and poor/risky animal husbandry practices

    • ~50% have frequent animal diseases, with high mortality
  • Animal diseases are highly contagious

  • Large impacts on households and largely preventable

Context: study site

  • In the wake of the Ebola epidemic, together with the GoSL, we implement a “One Health” intervention, addressing animal and human health

  • 363 villages across 7 chiefdoms in Kono district

Cross randomized treatments

One Health Program

  • Implemented from July 2017 to July 2019

  • Recruitment of Community Animal Health Worker: CAHW

    • 21-day training, implemented by veterinarians

    • Tool kit and drugs for treatment of basic animal diseases

    • Creation One Health platform

  • CAHW:

    • provides basic animal health services
    • demonstrates and provides information on best practices
    • reports suspected disease symptoms to MAF supervisors

One Health Program: 21-day training

Animal disease reporting form

Results

One Health Program results

  • Strong enthusiasm for the program: reporting throughout the period and beyond

  • No impact on human health, but impacts on intermediary outcomes

One Health Program: Results

  • From planner’s perspective it matters if we can improve CAHW performance:

    1. Recruitment

    2. Community monitoring

    3. Pay-for-performance

Recruitment

  • Community selection: MAF facilitators hold village meetings to nominate and select the CAHW candidate

  • “Business as usual”: CAHWs are recruited by state authorities in collaboration with Paramount Chiefs

  • Eligibility requirements, ability tests, language, residence

  • Training before assignment to incentives

Recruitment

  • Key design element: double listing

    • Obtain nominees from both Paramount Chief and community
    • Allows us to have the counter-factual

Incentives

Community Monitoring

  • CAHW and community develop joint action plans

  • Interface meetings between ministry staff and community:

    • Two times: after 6 months and 1.5 years
    • CAHW performance revealed (relative and overall)
    • Allows community members to reward or impose social sanctions

Incentives

Pay-for-Performance

  • CAHWs are paid based on their rank (across three groups)
  • Two times: after 6 months and 1.5 years

Note: rank ensure revenue neutrality but introduces small risk of non-independence / collusion

Treatment Implementation

  • Community monitoring and pay-for-performance implemented as designed

  • The recruitment treatment not so much

    • Only in 100 villages do we have true double-listing (in many other cases either same person listed twice or no one listed by one party)

    • Implication: smaller sample and non-experimental interactions

Cases: Assignments

Small amount of attrition

Does community selection affect CAHW type?

\(T_m\) Manipulation Check

\(T_m\) Manipulation Check

\(T_m\) Manipulation Check

Primary Outcome: Performance (Effort)

Measuring number and detail of disease reporting, worker availability: SD = 0.19

Main design based results (effort)

Changes in distributions

Main design based results

  • Both incentives shift reporting behavior

  • Correlation between performance and observed effort

    • CAHW more days observed working (survey)
  • No change in animal health, stocks, deaths

    • (ie no change in reporting denominator)
  • Increase in reporting quality

Outcome 2: Quality of Reporting

  • Surveillance performance driven by events, effort and ability

  • Records contain both Type 1 and Type 2 errors

  • Both matter for containing pandemics

    • triggering responses to false events
    • failing to report events

Quality of Reporting

We measure the quality of CAHW reporting

  • Practical: Central location to assess CAHW

    • Animal exam: CAHWs observe and diagnose sick animals, and record
    • Form review: Assess 5 (random) recently submitted forms
  • Village visit: MAF representatives visit the CAHW’s village

    • Sick animal observation: Verify if sick animals had been reported

    • Form follow-up: Verify reported sick animals

Components

(Less happy with raw data distributions)

Results: Quality of Reporting

There is strong relationship between surveillance quality and performance

Results: Quality of Reporting

Summary

  • Evidence for both community monitoring and pay for performance
  • No evidence for selection
  • Noisy evidence for interactions: signs in wrong direction

Some of these results challenge the model more than others:

  • No effect for \(T_s\) can be consistent with the model
  • But we really do expect from H4: that more productive workers are more responsive to extrinsic incentives (\(T_mT_p>0\))
  • No average effect for \(T_s\) a bit hard to reconcile with suggestive evidence of \(T_s\)T_m$ interaction. Are community selections worse when there is monitoring? (cannot reject null)

Structural model

Structural model

  • Our model provides predictions for effort as a function of parameters and treatments
  • We integrate these into a model with chiefdom level fixed effects and random errors
  • We then estimate model parameters using stan

\[e_{ij} = \alpha_{j} + \kappa \times p(T_s, T_m, T_p | \bar{\mu}, \underline{\eta}, \rho, \nu, \phi, \zeta) + \epsilon_i\]

Parameter estimates

Wide distributions, but show movement. \(\mu\) moves, small range on \(\eta\), \(\rho\) is flat

Parameter estimates

Posterior distributions on model parameters

Experiments

Two distinct types of experiments possible:

  1. In what kind of world (for what parameter values) are different types of interventions most likely to be optimal?

  2. Given some type of policy is likely to be in place, what kind of parameters would one want to alter in order to improve outcomes (so: thinking of parameters as possible sites of intervention)

Both can be seen from next figure.

Experiments

Comparing packages

Experiments

  • Intrinsic motivation matters across the board
  • Social sensitivity (\(\phi\)) matters only if \(T_m\) in place; appears to matter more than detachment or crowding out on the villager side
  • Type correlation matters across the board, though we do not see it here; harder to intervene on

Experiments: Selection \(T_s\)

Where and when might we expect stronger selection effects?

Experiments: Comm. Monitoring \(T_m\)

The Björkman and Svensson (2009) effect.

Experiments: Pay for Performance \(T_p\)

The payment effect.

Overall comparisons

Comparing packages

Discussion & Conclusion

Program

  • Decentralised delivery are an important development tool
  • Successful intervention: increase in good animal and human health practices, and animal stock
  • Important as outbreaks like Ebola, Covid may be more likely to happen in the future

Discussion & Conclusion

  • Evidence for two distinct institutional strategies

  • Increase institutional capital that can become even more critical during crisis

  • Tradeoff between P4P and CM?

    • No social cost of P4P, private gains and more efficient delivery
    • In CM: time costs seem small, but some social and status gains for CAHWs
    • So CM may build institutional capital, more sustainable?

Discussion & Conclusion

  • No evidence here that selection changes outcomes
    • But plausibly case specific issues
    • Hard to improve selection margin (selection vs eligibility)
    • Selection can undermine community sanctions due to norms and guilt

Discussion & Conclusion

Paper

  • Highlights power of combining model and data for understanding lessons
  • But also risks (are we too model dependent e.g. there are no negative interactions when data suggest there might be)
  • Perhaps set rival models in competition?

End

Extra slides

One Health Program

Created new positions in villages

One Health Program

Sustained engagement

Results: Animal Husbandry Practices

- Data on disease control, response, knowledge, preventative practices

Observed effort

Results: Behavioral test

- Covid knowledge

Results: Social gains

- Social relations, trust

Results: Reporting vs Cases

- Animal Health

Parameters

Symbol Name Interpretation
\(\mu \in \{0, \overline{\mu}\}\) (Intrinsic) motivation Strength of motivation
\(\eta \in \{\underline{\eta},1\}\) Embeddedness Strength of a worker’s ties to the community
\(\theta = (\mu,\eta)\) Worker’s type
\(\rho \in [-1,1]\) Type correlation \(Corr(\mu,\eta)\)
\(\phi\) Social sensitivity The worker’s susceptibility to social sanctioning
\(\gamma \in \{0,\overline{\gamma}\}\) Guilt The villager’s psychological cost of social sanctioning
\(\zeta \in [0,1]\) Strength of village-level norms Probability of low-guilt villager
\(\kappa\) Productivity Effort elasticity to wage
\(e \in [0,1]\) Effort How hard the worker works
\(\pi \in [0,1]\) Performance Outcome of worker’s activity (\(\pi= e \eta\))
\(\sigma \in \{0,1\}\) Performance signal Public signal of performance (\(\Pr(\sigma=1)=\pi\))
\(s(\sigma) \in [-1,1]\) Social sanctioning plan Villager commits to it before observing \(\sigma\)
Alatas, Vivi, Abhijit Banerjee, Rema Hanna, Benjamin A. Olken, and Julia Tobias. 2012. “Targeting the Poor: Evidence from a Field Experiment in Indonesia.” American Economic Review 102 (4): 1206–40. https://doi.org/10.1257/aer.102.4.1206.
Ashraf, Nava, Oriana Bandiera, and Scott Lee. 2020. “Losing Prosociality in the Quest for Talent? Sorting, Selection, and Productivity in the Delivery of Public Services.” American Economic Review 110: 1355–94.
Balán, Pablo, Augustin Bergeron, Gabriel Tourek, and Jonathan L. Weigel. 2022. “Local Elites as State Capacity: How City Chiefs Use Local Information to Increase Tax Compliance in the Democratic Republic of the Congo.” American Economic Review 112 (3): 762–97. https://doi.org/10.1257/aer.20201159.
Banerjee, Abhijit, and Esther Duflo. 2006. “Addressing Absence.” Journal of Economic Perspectives 20 (1): 117–32.
Basurto, Maria Pia, Pascaline Dupas, and Jonathan Robinson. 2020. “Decentralization and Efficiency of Subsidy Targeting: Evidence from Chiefs in Rural Malawi.” Journal of Public Economics 185: 104047. https://doi.org/https://doi.org/10.1016/j.jpubeco.2019.07.006.
Björkman, Martina, and Jakob Svensson. 2009. “Power to the People: Evidence from a Randomized Field Experiment on Community-Based Monitoring in Uganda.” The Quarterly Journal of Economics 124 (2): 735–69.
Callen, Michael, Saad Gulzar, Ali Hasanain, Muhammad Yasir Khan, and Arman Rezaee. 2015. “Personalities and Public Sector Performance: Evidence from a Health Experiment in Pakistan.” Working Paper, Working paper series, no. 21180 (May). https://doi.org/10.3386/w21180.
Christensen, Darin, Oeindrila Dube, Johannes Haushofer, Bilal Siddiqi, and Maarten Voors. 2021. “Healthcare Delivery During Crises: Experimental Evidence from Sierra Leone’s Ebola Outbreak.” Quarterly Journal of Economics 136 (2): 1145–98.
Dal Bó, Ernesto, Frederico Finan, Nicholas Li, and Laura Schechter. 2021. “Information Technology and Government Decentralization: Experimental Evidence from Paraguay.” Econometrica 89 (2): 677–701.
Dal Bó, Ernesto, Frederico Finan, and Martı́n A Rossi. 2013. “Strengthening State Capabilities: The Role of Financial Incentives in the Call to Public Service.” The Quarterly Journal of Economics 128 (3): 1169–1218.
Deserranno, Erika. 2017. “Financial Incentives as Signals: Experimental Evidence from the Recruitment of Health Workers.”
Duflo, Esther, Michael Greenstone, Rohini Pande, and Nicholas Ryan. 2018. “The Value of Regulatory Discretion: Estimates from Environmental Inspections in India.” Econometrica 86: 2123–60.
Francois, Patrick. 2000. “‘Public Service Motivation’as an Argument for Government Provision.” Journal of Public Economics 78 (3): 275–99.
Hussam, Reshmaan, Natalia Rigol, and Benjamin N. Roth. 2022. “Targeting High Ability Entrepreneurs Using Community Information: Mechanism Design in the Field.” American Economic Review 112 (3): 861–98. https://doi.org/10.1257/aer.20200751.
Leaver, Clare, Owen Ozier, Pieter Serneels, and Andrew Zeitlin. 2021. “Recruitment, Effort, and Retention Effects of Performance Contracts for Civil Servants: Experimental Evidence from Rwandan Primary Schools.” American Economic Review 111 (7): 2213–46.
Mohanan, Manoj, Katherine Donato, Grant Miller, Yulya Truskinovsky, and Marcos Vera-Hernández. 2021. “Different Strokes for Different Folks: Experimental Evidence on the Effectiveness of Input and Output Incentive Contracts for Health Care Providers with Different Levels of Skills.” American Economic Journal: Applied Economics.
Molina, Ezequiel, Laura Carella, Ana Pacheco, Guillermo Cruces, and Leonardo Gasparini. 2016. “Community Monitoring Interventions to Curb Corruption and Increase Access and Quality of Service Delivery in Low- and Middle-Income Countries: A Systematic Review.” Campbell Systematic Reviews 12 (1): 1–204. https://doi.org/https://doi.org/10.4073/csr.2016.8.
Muralidharan, Karthik, and Venkatesh Sundararaman. 2011. “Teacher Performance Pay: Experimental Evidence from India.” Journal of Political Economy 119 (1): 39–77.
Pradhan, Menno, Daniel Suryadarma, Amanda Beatty, Meiyan Wong, Arya Gaduh, Armida Alisjahbana, and Tessa Prama Artha. 2014. “Improving Educational Quality Through Enhancing Community Participation: Results from a Randomized Field Experiment in Indonesia.” American Economic Journal: Applied Economics 6 (2): 105–26.
Raffler, Pia, Daniel Posner, and Donald Parkerson. 2024. “Can Citizen Pressure Be Induced to Improve Public Service Provision?” Journal of Politics, Forthcoming.