What’s the growth elasticity of development across Africa today?

“Tomorrow’s Africa is going to be an economic force,” says a report from Goldman Sachs. KPMG trumpets the Africa story as “the rise of the phoenix.” Many factors have made this possible. After decades of stagnation, in recent years most African countries began to reform their economies. Wars, coups, political instability and disease have declined since the late 1990s. And rising commodity prices have lured investment in African resources. Mobile technology is leapfrogging ahead (Africa has become one of the fastest-growing markets for Canadian firm Research in Motion’s BlackBerry) and a new consumer class has been born. Multinational retailers are leaping in, and even Wal-Mart recently acquired a chain with nearly 300 stores in 14 African countries. The prosperity of China has been a particular spark, with about 2,000 Chinese companies investing $32-billion in Africa by the end of 2010.

But what is the truth behind the hype?

On a continent with a long history of foreign domination and colonial exploitation, this wave of external investment has the potential to repeat some of the errors of the past. There is still a power imbalance between huge multinational investors and weak governments, with officials tempted by quick payoffs and sometimes willing to sell out the people who live on the land. Mining and oil companies can generate big sums of money for governments while employing less than 1 per cent of the African work force.


From the first installment of a six part series by Globe and Mail‘s Geoffrey York on Africa’s growth boom (link). It promises to be great.

“Growth elasticity of poverty” refers to percentage reduction in poverty per percent growth in average income (link). The concept can be extended to consider how access to infrastructure and services, access to legal protection, and other general features of development accompany a rise income. Analyses of growth such as that which Sierra Leone is experiencing (profiled in York’s article) ought to keep stock of such relationships.

Over the past year I’ve traveled quite rapidly through middle income and lower income countries in Asia, Latin America, and Africa, as well as middle and lower income areas in each of these countries. Naturally this has led to mental comparisons. One feature that sharply distinguishes the two contexts is the extent to which a local entrepreneur has some non-laughable chance of realizing and growing a market for an idea, even to the point of industrial level production. In reading York’s piece it is hard to see how the types of large scale extractive investments that Sierra Leone is hosting will promote local or national level entrepreneurship. So does development need to come from somewhere else, perhaps leveraging income from these investments, but not fundamentally drawing its momentum from it? If so, then in addition to tracking growth elasticity of development indicators, one would want to consider the proportion of income gains attributable to direct sale of extractive commodities versus through other transactions.

Share

`rdd’: A new R package for nonparametric estimation with regression discontinuity designs

Drew Dimmery, a current NYU PhD student, has just finished coding up a new package for R that applies state-of-the-art non-parametric methods to estimate treatment effects for regression discontinuity designs. The package is now in beta, and we are looking for feedback from users. Drew has a nice description at his blog here: link. Drew also posted an introduction to the package to the POLMETH listserv: link.

Some great feedback has already come in, focusing in particular on possible problems with the Imbens-Kalynaraman bandwidth selection algorithm. If you have suggestions on improvements on that front, we are especially interested.

We are hoping to add more functionality to future versions (e.g., nonparametric covariate adjustment).

We would love to receive more feedback, whether in comments here, on Drew’s blog, or via email.

Share

Get funded to replicate & extend analyses of development interventions!

The International Initiative for Impact Evaluation (3ie) has an open call for proposals for researchers to replicate and extend the analysis of data from major development interventions and publications:

The primary objective of the replication programme is to improve the quality of evidence available for development policymaking and programme design. More specifically, the replication programme seeks to:
  • Verify the findings, and examine the robustness of findings, in selected influential, innovative, or controversial impact evaluations of development interventions;
  • Improve the incentives of researchers to conduct careful and responsible data analysis.
Completed replications will be published through the 3ie website and working paper series. In addition, a special issue of the Journal of Development Effectiveness will be published containing a selection of studies resulting from this first request for applications.



Details for this CFP are here: link. More details on 3ie’s replication program are here: link. Questions about the CFP need to be submitted by July 13, and applications need to be in by July 31. Note that this CFP is only about replicating data analyses, not replicating studies in their entirety. 3ie has prioritized a set of “candidate studies” for this round of replication funding: link.

I use this kind of data-analysis replication regularly in my teaching, and I applaud this effort. It would be particularly useful for researchers and graduate students considering either (i) designing new research that builds on the findings any of the candidate studies or (ii) new analytical methods that explore, e.g, effect heterogeneity, distributional effects, or violations of standard assumptions (e.g., violations of no interference assumptions). In either case, replication and exploration of the robustness of past findings would be a crucial first step. This is a chance to build that first step up and turn it into an output of its own.

Share

Ellis on “Africa in the World”

[T]he true postcolonial age came to an end in the last quarter of the twentieth century. Since then, Africa has entered a phase in its history that as yet has no name. (We can’t call it ‘post-post-colonial’.)…Among the dynamic new factors shaping Africa’s environment in recent years are a rapid rise in foreign investment in Africa, particularly from Asia, although Western countries are still the leading investors in Africa, and large-scale immigration by Chinese entrepreneurs. It is already apparent that Asian businesspeople and diplomats do not come to Africa with the same expectations as their European and American counterparts, nor with the same ideological baggage, and that they make different demands…As African societies respond to new demands and as people develop new strategies, new forms of insertion in the world are emerging.


So writes Stephen Ellis in Season of Rain: Africa in the World (Hurst & Co., 2011), a panoramic account of how local interests across Africa interact with international business, development, and major power strategy. A dense read (counting 170 pages but feeling like 300), Ellis makes the case that African societies have had to contend with strong international currents despite state weakness. Thus, the most important types of exchange have been between international currents and familial and patronage networks operating in the “shadows” of the state. There are echoes of Reno and Bayart, but updated. The book is quite sweeping and sometimes fails to back assertions with clear references to data. But it would make an excellent read for undergraduates or others seeking an introduction to modern predicaments facing policy makers, business people, and ordinary folks across Africa as they contend with international actors pursuing their commercial, strategic, as well as development aid and humanitarian interests.

Share

Power for cluster randomized studies

Over at the Development Impact Blog, Berk warns researchers to “beware of studies with a small number of clusters” (link), and raises the worry that we really don’t have good tools for assessing power for cluster randomized studies when the number of clusters is rather small. Berk’s general message is absolutely right. I agree that the available tools are imperfect. Nonetheless, there are better and worse ways to go about thinking through power for a cluster randomized study. Below are some thoughts on this based on my understanding (please correct or comment below!):

First, let’s lay down some fundamental concepts. Getting correct confidence intervals is hard even in the simplest scenarios—e.g., without even considering clustering. Let’s look under the hood for a second to see what is going on in a simple case. The “correct” confidence interval for an experimental treatment effect depends on a few things: the estimand, the population that is the target for your inferences, and the type of variability you are trying to measure. For example, maybe the estimand is the difference in means over treatment and control, the target of your inferences is the specific set of subjects over which treatment was randomized, and the type of variability is that which is induced by the random assignment. That is, you are just trying to make inferences about treatment effects for the subjects that were observed as part of the treatment and control groups. This is perhaps the most basic inferential scenario for an experiment, covering experiments with convenience samples or cases where there was no possibility of random sampling from a larger population. As it happens, the variance for the difference in means in this case is easy to express. Unfortunately, the expression is, generally, unidentified, as it requires knowledge of potential outcome covariances. So right off the bat, we’re in the land of approximation and we haven’t even yet gotten to the question of a reference distribution to use to do hypothesis tests or construct a confidence interval. The good thing is that we can get a conservative approximation of the variance using the usual (heteroskedastic) sampling distribution for difference in means.

Moving ahead, we know the asymptotic distribution for the difference in means is normal. (Well, this is actually something that was only proved recently by David Freedman in a 2008 paper, but I digress. See here: link). But we’re not in asymptopia. So, why not try to fatten the tails of our reference distribution, say, by applying the result that the finite sample distribution for the difference in means with normal data is t? Our underlying data aren’t normal, but a little tail fattening can’t hurt, right? (Though it may not “help” enough…)

Or wait, maybe some resampling method is available—e.g., a permutation test? Well, this works for testing the “sharp null hypothesis,” but standard permutation test procedures are not valid, generally, for generating confidence intervals otherwise. In fact they are more anti-conservative than the approximate method using the t distribution. So we stick with our kludgy variance approximation and use it against a t distribution that hopefully does enough to correct for the fact that we aren’t in asymptopia. Et voila. It’s ugly, but probably good enough for government work. (Some colleagues and I are actually looking into new permutation methods to see if we can do better that what current methods allow. I’ll update once we figure out how well it works.)

Suppose now that instead of a convenience sample, your sample was drawn as a random sample from the target population. Then, a happy coincidence is that the variance approximation described above is unbiased for the true randomization-plus-sampling variance. We still need a reference distribution; for lack of a better alternative, we could use the same t distribution as a tail-fattened finite sample approximation to the asymptotic distribution. An alternative would be to use a bootstrap. But the validity of the bootstrap procedure depends on how well the sample represents the population. While the bootstrap is unbiased over repeated samples, for any given sample, the expected divergence from representing the target population depends on sample size, alas. For that reason, bootstrap confidence intervals can be larger or smaller than their analytically derived counterparts; the two will tend to agree as the sample size grows larger.

So even in these simple cases, without clustering even entering the picture, we’ve got approximations going on all over the place. These approximations are what we feed into our power analysis calculators, hoping for the best.

Clustering doesn’t fundamentally change the situation. We just need to take a few extra details into account. First, the relevant type of variability now has to take into account randomization plus whether we have sampling of clusters, sampling within clusters, or both. In a manner that is similar to what we saw above, conventional (heteroskedastic, cluster-robust) variance expressions coincide with the type of variability associated with randomization plus sampling of and within clusters, making it a conservative approximation to the variance when one or the other kind of sampling is not relevant.

Second, we need to appreciate that the effective sample size from a cluster randomized study is somewhere between the number of units of observation and the number of clusters. The location of the effective sample size between these bookends is a function of intra-class correlation of the outcomes. Working with the exact expression is cumbersome, and so most power calculation packages use a kludge that assumes constant intraclass correlation over treatment and control outcomes (a conservative way to do this is to assume it is equal to the larger of the two). Under random sampling of and within clusters, this gives rise to the classic design effect expression, 1+(m-1)rho, where m is the average cluster size and rho is the intra-class correlation. (This is what Stata’s sampclus applies.) However, while sampclus accounts for the higher variance with the design effect, it does not change the reference distribution. For few clusters, this is probably anti-conservative: the fattening of the tails of our reference distribution ought to take into account these issues of effective sample size. Stata’s kludge is to fatten tails using t(C-1), where C is the number of clusters. (Cameron, Gelbach, and Miller (2008; see here or my Quant II files linked at top right) consider t(C-k) with k being the number of regressors; if you look at their Table 4, this kludge actually performs as well as any of the bootstrap procedures, even with only 10 clusters.) So ideally, you’d want to adjust sampclus results on that basis too (a point that Berk makes in his blog post). It should be an easy thing to program for anyone enterprising and with some time on his or her hands, or you can fudge it by tinkering with the alpha and power levels. I think Optimal Design accounts for degrees of freedom in a better way while using the same design effect adjustment, but as far as I know, the results are only defined for balanced designs. With balanced designs, many of these problems actually disappear completely (see my 2012 paper with Peter Aronow in SPL: PDF), but results for balanced designs are often anti-conservative for non-balanced designs.

The same logic as above would also apply when considering the bootstrap with clustered data: its accuracy depends on how well the sample of clusters represents the target population of clusters. So, as with the above, bootstrap confidence intervals may be larger or smaller than the analytically derived ones. Berk mentioned the wild bootstrap procedure discussed in Cameron, Gelbach and Miller (2008), implemented in a Stata .ado file from Judson Caskey’s website (link). If you play with the example data from the helpfile of that .ado, you will find that wild bootstrap confidence intervals are indeed narrower in some cases and for some variables than what you get from “reg …, cluster(…).”

One could test all of these ideas via simulation, and I think this is an underused complement to analytical power analyses implemented by, e.g., sampsi and sampclus. That is, if you have data from a study or census that resembles what you expect to encounter in your planned field experiment, you should simulate the sampling and treatment assignment and assess the true variability of effect estimates over repeated samples and assignments, adjusting the sample sizes. You can also assess the validity of analytical or bootstrap alternatives against the simulated benchmark.

(For those interested in more details on variance estimation and inference, see the syllabus, slides, and readings for my Quant II and Quant Field Methods courses linked at top right.)

Share