The GiveWell Blog

Aid’s track record

While Elie’s been investigating the Carter Center, I’ve been scanning literature (mostly academic) on general questions about aid: what has worked in the past? What’s promising for the future? etc.

Rather than trying to come to an independent conclusion on each debate, I’ve been trying to establish which beliefs are supported by evidence that is largely undisputed among scholars (and where there is no agreement, get a sense of what each side believes and what its most frequently pointed to evidence is).

At this point I’ve looked in a lot of places (though I’m still far from done) and I’m going to start sharing where we stand (and what we’re still missing) on various questions. First I’ll discuss what I’ve read about the track record of aid to developing-world areas.

The vast majority of the large-scale “success stories” I’ve seen come from health initiatives (particularly “vertical” health initiatives, i.e., large-scale campaigns against particular diseases).

The main non-health “success stories” I’m aware of:

I don’t believe that any of the sources cited above are fully comprehensive (or necessarily fully reliable) as lists of large-scale success stories. We’re still looking for more.

Aid has also had major failures – but insufficient monitoring and evaluation means that few are thoroughly documented or discussed.

  • Dissatisfaction with the accomplishments of aid to date is fairly widespread. The latest report on the Millennium Development Goals shows mostly inadequate/less-than-hoped-for progress (and no progress in many cases). William Easterly argues that

    systematic testing would not just count the alleged ‘success stories’ of aid, but also the larger number that got the same amount of aid as the ‘success stories’ and failed: Guinea-Bissau, Somalia, The Gambia, Mali, Rwanda, Nicaragua, Burundi, Guyana, Zambia, the Central African Republic, Senegal, Suriname, Chad, Niger, Togo, Haiti, and so on. Further testing shows that these outcomes were not an artifact of selection bias or reverse causality.

  • Yet while project-level evaluations provide scattered analysis of projects gone wrong (one example of a failed World Bank project here), we have not found well-documented “failure stories” along the lines of the “success stories” above – examining major humanitarian (as opposed to political) initiatives that had little or negative impact.
  • Part of the reason may be the general lack of evaluation and documentation in international aid. The sentiments of this Center for Global Development paper, arguing that “very few programs benefit from studies that could determine whether or not they actually made a difference. This absence of evidence is an urgent problem,” are common; the Paris Declaration on Aid Effectiveness, an agreement between major aid agencies, includes an expressed commitment to more monitoring and evaluation.

I have the sense that failure stories in aid are common, but I continue to look for more concrete examples of what has gone wrong and how.

The relationship between aid and growth at the macro/country level has been extensively studied, but is not well established one way or the other.

Some believe in a moderate positive relationship, often with the caveat that aid works better where existing institutions are stronger (more below) or that aid has diminishing returns. Others believe that there is no relationship or that there is insufficient evidence. Two particularly accessible summaries: A Primer on Foreign Aid (from the optimistic side) and Macro Aid Effectiveness Research: a Guide for the Perplexed (from the skeptical side).

There are no established broad patterns in the sorts of environments where aid has worked, though many believe that aid works better where existing institutions are stronger.

A Primer on Foreign Aid states,

the view that aid works better (or in a stronger version, aid works only) in countries with good policies and institutions has become the conventional wisdom among donors, partly based on [empirical] research and partly due to development practitioners that believe this to be the case based on their own experience.

However, as this paper acknowledges, the empirical research has been questioned repeatedly (as in this paper, which claims that the statistical work underlying most such claims is excessively fragile). In addition, Millions Saved (the success stories compilation referred to above) gives examples that “Success is possible even in the world’s most underdeveloped and remote regions, in the face of grinding poverty and weak health systems” (quote available at this page).

One caveat to keep in mind about all of this analysis is that most academic literature focuses on official aid flows – aid from developed-world governments (or multilateral institutions such as the World Bank), which usually goes through developing-world governments. This sort of aid is different in many ways from private donations going through nonprofits.

The Carter Center – part II

In Part I, I laid out my case for The Carter Center, and why we think they’re worth investigating deeply. I also said that we’re not yet ready to recommend them because we have some unanswered questions. Here they are:

Relevance of your donation

When I give, I want to know that the organization needs my support and is going to accomplish more good (than they otherwise would have) because of my support. Because of The Carter Center’s strong track record, I wonder whether they can already effectively raise all the money they want. That is, I wonder whether money isn’t the bottleneck to serving more people. That question is supported by looking at Carter’s current financial situation. Revenues have risen roughly in-line with expenses over the past few years (see chart 1 below), maintaining their (relatively large) assets:expense ratio of 3:1 (see chart 2, below). All data comes from Carter Center annual reports, available on their website here.

Chart 1
 

Chart 2
 

Less-proven and less-monitored programs.

Many of The Carter Center’s programs have extremely strong independent evidence supporting them (e.g., their guinea worm, river blindness, lymphatic filariasis, and schistosomiasis programs). These programs also have the type of strong, consistent monitoring I mentioned in the previous post. However, the evidence case for some of their other programs is weaker (in particular their trachoma, agriculture, and public health training initiative), and the monitoring provided for those programs is both a) less consistent and b) less compelling than the monitoring provided for the others.

Running some less-proven projects doesn’t necessarily mean we won’t recommend them, especially if those programs are relatively small, but:

Lack of financial transparency

The Carter Center is one of the (if not the) most transparent organizations I’ve ever come across when it comes to monitoring their activities. Carter not only lays out what they do in extreme detail — for the intrepid, the most recent river blindness program report (PDF) is representative of Carter Center monitoring — they discuss program trade-offs they make, lessons they’ve learned, and potential future obstacles. But, they don’t publish any cost data on their website, other than the very broad and general 990 data and some sporadic program-specific information. To properly evaluate The Carter Center, we need to know the portion of the budget that’s allocated to any less-proven programs. In addition, we want tell you what your donation accomplishes to The Carter Center. Does $1,000 prevent 1 case of blindness? 2? 10? We need better information on costs to make that call.

Those are the questions we’re working on answering now. Whatever we find, you’ll get our answers soon.

Malaria “success story” questioned

Aid Watch on questionable claims of success against malaria:

Real victories against malaria would be great, but false victories can mislead and distract critical malaria efforts. Alas, Mr. and Mrs. Gates are repeating numbers that have already been discredited. This story of irresponsible claims goes back to a big New York Times headline on February 1, 2008: “Nets and New Drug Make Inroads Against Malaria,” which quoted Dr. Arata Kochi, chief of malaria for the WHO, as reporting 50-60 percent reductions in deaths of children in Zambia, Ethiopia, and Rwanda, and so celebrated the victories of the anti-malaria campaign. Alas, Dr. Kochi had rushed to the press a dubious report. The report was never finalized by WHO, it promptly disappeared, and its specific claims were contradicted by WHO’s own September 2008 World Malaria Report, by which time Dr. Kochi was no longer WHO chief of malaria.

Video with the Gateses available here.

Trachoma: An example of the need for long-term monitoring

When is a measured program-impact not a real impact? When it doesn’t last.

A study published today in PloS NTDs evaluated the impact of four doses of azithromycin (one every 6 months), and monitored trachoma prevalence throughout the drug administration period and for 2 years after the last dose.

In the first 24 months (from the start of the program through 6 months after the last treatment) prevalence decreased from 63.5% to 2.6%. In the 18 months after the last treatment, prevalence increased to 25.2%.

There was wide village-to-village variation in the prevalence before, during, and after treatment. While all villages ended up with lower infection rates of trachoma than they started with, in some (in particular village numbers 10 and 12 in this table), the infection rate 18 months post-treatment was close to pre-treatment levels.

Another reason to think that more monitoring is probably worth the cost.

The Carter Center

I’ve spent a good part of the past month reviewing The Carter Center in depth. We found out about The Carter Center though referrals from an advisor and donor, and through the Gates Award for Global Health.

We chose to investigate The Carter Center further for four reasons:

  1. Strong track record of success. The Carter Center has been credited with leading the global effort to eradicate guinea worm. In 1986, when Carter got involved, there were 3.5m annual cases of guinea worm. In 2008, there were fewer than 5,000. Guinea worm has been eliminated from 12 of the 18 countries that were once endemic. This chart shows the decline in cases from 1989-2007 (PDF). (More information about guinea worm and its symptoms is available on The Carter Center website.)
  2. Strong program selection. Other Carter Center programs have strong evidence of effectiveness supporting them. The interventions they implement to control or eliminate neglected tropical disease including river blindness, trachoma, lymphatic filariasis, and schistosomiasis are recommended by the experts at the Disease Control Priorities Project.
  3. Monitoring and evaluation. For its health programs, The Carter Center often monitors not only the number of drugs distributed in each region but disease prevalence directly. This monitoring isn’t evident for every single program or region, but it is far more consistent than any other complex organization we’ve seen before.
  4. Transparency. All the information we’ve used to research The Carter Center is available publicly on their website. For example, the detailed program reports I mentioned above are available here. For an example of the type of monitoring conducted for its guinea worm program, see the latest monthly report linked here.

I still have some large, unanswered questions about The Carter Center, and we aren’t yet ready to recommend them.

Financial Times on microfinance (and the need for better info about it)

PDF here (via Innovators for Poverty Action, whose research is featured). After discussing the Karlan/Zinman study showing benefits for loans (which we summarize here), it continues:

Karlan is the first to warn against extrapolating too much from a single experiment. “This is the last thing in the world that I would use to develop policy,” he warns. “You’ve got to replicate.”

The trouble is that the replication just isn’t happening. For all the optimism about microfinance – and the [Karlan/Zinman] experiment only encourages that optimism – it is striking how much we do not know about when it works, and why.

This matters because non-commercial microfinance projects often depend on donor subsidies. And while microfinance has a good reputation among development professionals, that doesn’t mean guaranteed access to those subsidies. Not everyone is convinced that a donor grant is best used to subsidise a loan rather than, say, pay directly for a primary school. More credible evaluation would help preserve the programmes that deserve to be preserved.

Already, solidly held beliefs about microfinance have been shaken. The “group liability” system, in which a group of borrowers guarantee one another’s loans, is still supposed by many to be the secret behind Grameen Bank’s low default rates. But a randomised trial in the Philippines conducted by Karlan and aWorld Bank economist, Xavier Gine, found that group liability was discouraging new customers without improving repayment rates. Grameen itself quietly dropped group liability some time ago.

Another sacred cow of microfinance is that women make best use of the money – the Grameen Bank says that 97 per cent of its borrowers are women. But another randomised trial, conducted in Sri Lanka by a team of researchers including David McKenzie of the World Bank, found that male borrowers seemed to make a far higher return on their capital. As with the ZaFinCo study, it’s just one experiment in one country. Yet it raises a worrying question: for how long will donors fund microfinance projects with so little compelling evidence about exactly what kinds of project really work?

It also discusses why the Karlan/Zinman’s results might not be generalizable:

[The experiment’s] clients were ideal customers for a commercial lender: they were city-dwellers and therefore cheap to reach; they were poor enough to want loans but rich enough that the loans were profitably chunky. A peasant farmer in Ethiopia or Sudan ticks none of those boxes: living in the middle of nowhere, he is expensive to reach and he is so poor that he can only afford tiny loans. Trapped in a barren economic ecosystem, he has no job that a ZaFinCo-style loan can help preserve, and no business prospects either. A mere loan will not catapult him into the ranks of the entrepreneurial class. Then there are the destitute, the disabled, the elderly and the orphans. Such people cannot repay loans at a rate that would cover costs. Heavy subsidies or outright grants would be needed. “All people are entrepreneurs,” says Muhammad Yunus. If only he were right.

And indeed, a more recent randomized trial of microfinance has preliminarily found far less encouraging results (discussed in this interview on Philanthropy Action, co-maintained by GiveWell Board member Tim Ogden).