Carolina Torreblanca, William Dinneen, Guy Grossman, and Yiqing Xu have an impressive new study on trends in quantitative research design in political science—specifically, on whether the field has truly undergone a “credibility revolution.”
Read the working draft here.
One finding jumps off the page: the sheer rise of survey experiments. By their count, survey experiments now make up nearly half of all
design-based explanatory quantitative research—the category they treat as the hallmark of credibility-revolution work.
That conclusion deserves more scrutiny. I do not think survey experiments, in general, are emblematic of the credibility revolution. Some are already treating them as such, and that risks muddling two distinct intellectual streams. If we want to track the credibility revolution, we need at least to separate survey experiments that estimate effects of real-world interventions from those that are essentially measurement tools or light-touch priming exercises, although even then I am skeptical about the connection to the credibility revolution.
To be clear: I use survey experiments, RCTs, observational-causal methods, and descriptive analysis in my own work. This is not about elevating or denigrating any one method. It is about clarifying what belongs to which intellectual lineage. For a broader statement of this view, see my “problem solving” essay:
read the preprint here.
Intellectual origins matter
Small priming experiments have a long history in psychology and behavioral science. They are the direct ancestors of today’s survey experiments. Political science has always been an eclectic discipline, drawing from psychology, sociology, economics, and more. Even without the identification-strategy turn, political science would still have survey experiments—perhaps fewer, but they would be here. Their growth is better understood as a side effect of the credibility revolution, not a direct indication of its spread.
This is why Angrist and Pischke never talk about survey experiments. Their work is about how to study big, messy policy processes with disciplined empirical strategies. As Adi Dasgupta put it on BlueSky (@adasgupta.bsky.social): the “credibility revolution was about empirical strategies to overcome real-world endogeneity.” Survey experiments are simply not what the credibility revolution was proposing.
The credibility revolution pushes against “only experiments are causal”
Many who do survey experiments convey an old-school psychology view: only tightly controlled experiments identify causal effects; observational work does not, and field experiments are too messy. The credibility revolution directly challenges that mindset. It showed that careful observational designs and real-world RCTs can credibly identify, albeit with well understood limitations, the effects of major interventions and policies—problems once assumed to be too uncontrolled to study rigorously.
Some psychologists today are trying to bring this insight into their own field: as I understand it, they are trying to usher in a credibility revolution in psychology by moving colleagues beyond the “experiments or nothing” framework and teaching the logic of identification strategies.
Two categories of survey experiments
Most survey experiments fall into one of two buckets:
-
Measurement-oriented experiments
These include conjoint designs, list experiments, endorsement experiments, randomized-response techniques, and many priming interventions. They are invaluable tools. But their goal is not to estimate the effects of real-world interventions or policies. -
Survey-platform RCTs that mimic real interventions
These are experiments targeting causal effects relevant to policy design. They belong on a different track and have a stronger relationship to the credibility revolution, but even then do not necessarily reflect the willingness to try to study change in the real world.
These various distinctions matter. Lumping everything together inflates and obscures the advance of the credibility revolution, which is about empirical strategies that credibly estimate causal effects in complex real-world settings.
