Natural Experiments in Labor Economics and Beyond

The 2021 Nobel Prize in Economic Sciences was recently awarded to David Card of the University of California, Berkeley, ‘for his empirical contributions to labour economics’, and to Joshua Angrist of MIT and Guido Imbens of Stanford University ‘for their methodological contributions to the analysis of causal relationships’. As has become an annual tradition at the IGM, we invited our panels to express their views on the work of the new laureates. We asked the experts whether they agreed or disagreed with the following statements, and, if so, how strongly and with what degree of confidence:

a) The introduction of natural experiments to economic analysis of the labor market and related areas has led to a more precise understanding of cause and effect.

b) The ‘credibility revolution’ in empirical economics has improved our understanding of a number of public policy issues, including education, immigration and the minimum wage.

c) In pursuit of credible research designs, researchers often seek good answers instead of good questions.

Background on the first two questions has been provided in much of the coverage of the new Nobels, including a VoxEU column by Jörn-Steffen Pischke, who has co-authored books and papers with Joshua Angrist. The third was stimulated by their 2010 Journal of Economic Perspectives paper on the credibility revolution, which closes by discussing ‘the claim that the experimentalist paradigm leads researchers to look for good experiments, regardless of whether the questions they address are important.’

Of our 43 US experts, 41 participated in this survey; of our 48 European experts, 39 participated – for a total of 80 expert reactions.

Understanding cause and effect

On the first statement about natural experiments leading to a more precise understanding of cause and effect, a big majority of respondents on both panels agree. Weighted by each expert’s confidence in their response, 72% of the US panel strongly agree, 24% agree, 0% are uncertain, and 3% disagree (the totals don’t always sum to 100 because of rounding). Among the European panel (again weighted by each expert’s confidence in their response), 66% strongly agree, 31% agree, 3% are uncertain and 0% disagree.

Overall, across both panels, 69% strongly agree, 27% agree, 2% are uncertain, and 2% disagree.

Among the short comments that the experts are able to include in their responses, Larry Samuelson at Yale says: ‘Natural experiments are a welcome complement to other tools.’ Jan Pieter Krahnen at Goethe University Frankfurt comments: ‘Using natural experiments is an intelligent way to build causal inference on the diversity of institutions, shocks and behavior.’ And Franklin Allen at Imperial remarks: ‘As long as the natural experiment is well designed, these studies have the potential to increase our knowledge.’

Others mention applications beyond labor economics. Christian Leuz at Chicago replies:
‘Agree but not only in labor and micro, but also many other applied fields like finance and accounting.’ Pete Klenow at Stanford provides a reference to a survey of natural experiments in macroeconomics by European panel member Nicola Fuchs-Schundeln and a colleague. And Christopher Udry at Northwestern links to two experimental studies in development economics.

A few panelists express reservations. Anil Kashyap at Chicago notes: ‘Can’t answer all questions with these approaches, but we have definitely learned a lot.’ Patrick Honohan at Trinity College Dublin observes: ‘Although, as with randomized control trials, the evidence often comes from particular contexts that may not be generalizable.’ And Costas Meghir at Yale, who votes uncertain, adds: ‘Use of natural experiments without models relies on behavioral assumptions that are often unstated. May learn little about mechanisms.’

The ‘credibility revolution’ in empirical economics

On the second statement about the impact of the credibility revolution on policy debates in education, immigration and the minimum wage, again there is a big majority in agreement on both panels. Weighted by each expert’s confidence in their response, 60% of the US panel strongly agree, 32% agree, 7% are uncertain, and 0% disagree. Among the European panel (again weighted by each expert’s confidence in their response), 48% strongly agree, 52% agree, and 0% are uncertain or disagree.

Overall, across both panels, 54% strongly agree, 42% agree, 4% are uncertain, and 0% disagree.

Among the comments, Nicholas Bloom at Stanford says: ‘The research on the minimum wage has been pathbreaking – I absolutely changed my mind based on the evidence, and it has driven policy.’ Ricardo Reis at LSE adds: ‘Definitely improved. Of course, one would hope that progress had been even larger and more decisive. But no doubts on the sign of progress.’ And Costas Meghir notes: ‘There has been increased emphasis on justifying sources of exogenous variation in both structural and quasi-experimental studies.’

Several panelists refer to where natural experiments sit relative to other methods in research and policy evaluation. Larry Samuelson comments: ‘I view it as credibility evolution, with techniques that have built upon and evolved alongside other methods.’ Antoinette Schoar at MIT suggests: ‘Policy evaluation is complex and we need to draw on as many methods as possible for progress. But the credibility revolution is a key tool.’ Christian Leuz adds: ‘For policy, causal inferences are critical so need credible designs.’

Robert Shimer at Chicago, who votes uncertain, comments: ‘The devil is in the details with identification; and researchers often explore limited outcomes, e.g. only the short run.’ Also voting uncertain is Angus Deaton at Princeton, who says: Some plusses lots of minuses.’

Good answers and good questions

The third question – inviting views on whether in pursuit of credible research designs, researchers often seek good answers instead of good questions – generated considerably more differences of opinion. Weighted by each expert’s confidence in their response, 4% of the US panel strongly agree, 63% agree, 23% are uncertain, 7% disagree and 4% strongly disagree. Among the European panel (again weighted by each expert’s confidence in their response), 17% strongly agree, 36% agree, 37% are uncertain, and 10% disagree.

Overall, across both panels, 10% strongly agree, 49% agree, 30% are uncertain, 9% disagree, and 2% strongly disagree.

Among the comments of those who agree, Franklin Allen notes: ‘Unfortunately, there is some truth in this statement. It’s difficult to identify good natural experiments for many questions.’ Larry Samuelson adds: ‘This is the case with much of economics, and the social sciences more generally.’ And Anil Kashyap responds: ‘Running joke in the faculty lounge after some shock: “we will see a diff-in-diff paper on the job market using that variation next year…”’.

Others comment on incentives within economic research. Aaron Edlin at Berkeley states: ‘The profession continues to place a premium on clever identification strategies.’ Kenneth Judd at Stanford argues: ‘Too many economists focus on tractability, and even demand that one should know what the results will be before doing the analysis.’ And Luigi Guiso at the Einaudi Institute for Economics and Finance observes: ‘There is indeed a risk that the method drives the question and the risk is already real.’

Several panelists demur at ‘often’ in the statement. David Autor at MIT says: ‘This sometimes happens, for sure. But “often” is too high a bar for me. So, I disagree’; and Barry Eichengreen at Berkeley states: ‘If the question had been worded “sometimes” rather than “often” I would have agreed.’ David Cutler at Harvard concurs: ‘There are certainly some studies like this. “Often” is the key word here’; as does Daron Acemoglu at MIT: ‘I would say “sometimes” rather than “often”. Credibility revolution is fantastic for economics but we can/should not sacrifice big questions.’

Other panelists go further. Oriana Bandiera at LSE argues: ‘I agree that some researchers do this but you can’t blame the method, it’s like saying knives are bad because people use them to hurt others.’ Similarly, Antoinette Schoar comments: ‘I believe that methods can indeed shape the questions researchers ask. But this is true for all tools, including structural estimation, etc.’

Still others are critical of the way the question is framed. Richard Thaler at Chicago says: ‘Don’t love the wording here. Researchers try to answer some question, often it is not the most interesting one.’ Pol Antras at Harvard comments: ‘I don’t know what’s meant by “good”. I see a trade-off between credibility and generality (or external validity), but questions aren’t “bad”.’ And Abhjit Banerjee at MIT protests: ‘I really don’t know what that means. Should we pursue good questions which can’t be answered? Wittgenstein had it right.’

In similar vein, Richard Schmalensee at MIT argues: ‘Working on good questions that can’t be given good answers is a waste of effort; researchers aren’t wrong to consider quality of answers.’ And Michael Greenstone at Chicago objects: ‘This is a straw man debating style criticism. Is the alternative bad answers to good questions? Count me out for that!’

A final set of comments portray a trade-off. Ricardo Reis remarks: ‘Within set of good answers, they pick the better questions. Researchers trade off the two, unclear that they do so sub-optimally.’ Nicholas Bloom adds: ‘There is a trade-off between the quality of the question and the quality of the answer, but as long as we are on the frontier, all is good.’ And Olivier Blanchard at the Peterson Institute concludes: ‘The initial phase was indeed good answers. But we have moved over time to good questions.’

All comments made by the experts are in the full survey results.

Romesh Vaitilingam
@econromesh
October 2021

Question A:

The introduction of natural experiments to economic analysis of the labor market and related areas has led to a more precise understanding of cause and effect.

Responses weighted by each expert's confidence

Question B:

The ‘credibility revolution’ in empirical economics has improved our understanding of a number of public policy issues, including education, immigration and the minimum wage.

Responses weighted by each expert's confidence

Question C:

In pursuit of credible research designs, researchers often seek good answers instead of good questions.

Responses weighted by each expert's confidence

Question A Participant Responses

Participant University Vote Confidence Bio/Vote History
Allen
Franklin Allen
Imperial College London
Agree
6
Bio/Vote History
As long as the natural experiment is well designed, these studies have the potential to increase our knowledge.
Antras
Pol Antras
Harvard
Agree
1
Bio/Vote History
Bandiera
Oriana Bandiera
London School of Economics
Agree
9
Bio/Vote History
Blanchard
Olivier Blanchard
Peterson Institute
Agree
8
Bio/Vote History
Bloom
Nicholas Bloom
Stanford
Strongly Agree
8
Bio/Vote History
Blundell
Richard William Blundell
University College London
Agree
7
Bio/Vote History
Carletti
Elena Carletti
Bocconi
No Opinion
Bio/Vote History
Danthine
Jean-Pierre Danthine
Paris School of Economics
Strongly Agree
5
Bio/Vote History
De Grauwe
Paul De Grauwe
LSE Did Not Answer Bio/Vote History
Eeckhout
Jan Eeckhout
UPF Barcelona
Strongly Agree
7
Bio/Vote History
Fehr
Ernst Fehr
Universität Zurich
Strongly Agree
8
Bio/Vote History
Freixas
Xavier Freixas
Barcelona GSE Did Not Answer Bio/Vote History
Fuchs-Schündeln
Nicola Fuchs-Schündeln
Goethe-Universität Frankfurt
Strongly Agree
10
Bio/Vote History
Galí
Jordi Galí
Barcelona GSE Did Not Answer Bio/Vote History
Giavazzi
Francesco Giavazzi
Bocconi Did Not Answer Bio/Vote History
Griffith
Rachel Griffith
University of Manchester
Strongly Agree
10
Bio/Vote History
Guerrieri
Veronica Guerrieri
Chicago Booth
Strongly Agree
6
Bio/Vote History
Guiso
Luigi Guiso
Einaudi Institute for Economics and Finance
Strongly Agree
6
Bio/Vote History
Guriev
Sergei Guriev
Sciences Po
Strongly Agree
9
Bio/Vote History
Honohan
Patrick Honohan
Trinity College Dublin
Agree
3
Bio/Vote History
Although, as with randomized control trials, the evidence often comes from particular contexts that may not be generalizable
Javorcik
Beata Javorcik
University of Oxford
Strongly Agree
9
Bio/Vote History
Krahnen
Jan Pieter Krahnen
Goethe University Frankfurt
Strongly Agree
7
Bio/Vote History
Using natural experiments is an intelligent way to build causal inference on the diversity of institutions, shocks and behavior.
Kőszegi
Botond Kőszegi
Central European University
Strongly Agree
10
Bio/Vote History
La Ferrara
Eliana La Ferrara
Harvard Kennedy
Strongly Agree
10
Bio/Vote History
Leuz
Christian Leuz
Chicago Booth
Agree
8
Bio/Vote History
Agree but not only in labor and micro, but also many other applied fields like finance and accounting.
Mayer
Thierry Mayer
Sciences-Po Did Not Answer Bio/Vote History
Meghir
Costas Meghir
Yale
Uncertain
9
Bio/Vote History
Use of natural experiments without models relies on behavioral assumptions that are often unstated. Me learn little about mechanisms
Pagano
Marco Pagano
Università di Napoli Federico II
Agree
8
Bio/Vote History
Pastor
Lubos Pastor
Chicago Booth
Strongly Agree
5
Bio/Vote History
Persson
Torsten Persson
Stockholm University Did Not Answer Bio/Vote History
Pissarides
Christopher Pissarides
London School of Economics and Political Science Did Not Answer Bio/Vote History
Portes
Richard Portes
London Business School
Agree
4
Bio/Vote History
Prendergast
Canice Prendergast
Chicago Booth
Strongly Agree
8
Bio/Vote History
Propper
Carol Propper
Imperial College London
Strongly Agree
7
Bio/Vote History
Rasul
Imran Rasul
University College London
Strongly Agree
9
Bio/Vote History
Reichlin
Lucrezia Reichlin
London Business School Did Not Answer Bio/Vote History
Reis
Ricardo Reis
London School of Economics
Strongly Agree
9
Bio/Vote History
As explained by Nobel prize committee in its, as usual, very useful report.
-see background information here
Repullo
Rafael Repullo
CEMFI
Strongly Agree
6
Bio/Vote History
Rey
Hélène Rey
London Business School
Agree
8
Bio/Vote History
Schoar
Antoinette Schoar
MIT
Strongly Agree
9
Bio/Vote History
Storesletten
Kjetil Storesletten
University of Minnesota
Strongly Agree
7
Bio/Vote History
Sturm
Daniel Sturm
London School of Economics
Strongly Agree
10
Bio/Vote History
Van Reenen
John Van Reenen
LSE
Strongly Agree
9
Bio/Vote History
Vickers
John Vickers
Oxford
Agree
4
Bio/Vote History
Voth
Hans-Joachim Voth
University of Zurich
Agree
8
Bio/Vote History
Whelan
Karl Whelan
University College Dublin Did Not Answer Bio/Vote History
Wyplosz
Charles Wyplosz
The Graduate Institute Geneva
Agree
3
Bio/Vote History
Zilibotti
Fabrizio Zilibotti
Yale University
Agree
8
Bio/Vote History

Question B Participant Responses

Participant University Vote Confidence Bio/Vote History
Allen
Franklin Allen
Imperial College London
Agree
6
Bio/Vote History
Identifying causation is a difficult problem but some progress has been made in a number of areas in recent years.
Antras
Pol Antras
Harvard
Agree
8
Bio/Vote History
Bandiera
Oriana Bandiera
London School of Economics
Agree
9
Bio/Vote History
Blanchard
Olivier Blanchard
Peterson Institute
Agree
7
Bio/Vote History
Bloom
Nicholas Bloom
Stanford
Strongly Agree
10
Bio/Vote History
The research on the minimum wage has been pathbreaking - I absolutely changed my mind based on the evidence, and it has driven policy.
Blundell
Richard William Blundell
University College London
Strongly Agree
8
Bio/Vote History
Carletti
Elena Carletti
Bocconi
Agree
3
Bio/Vote History
Danthine
Jean-Pierre Danthine
Paris School of Economics
Strongly Agree
5
Bio/Vote History
De Grauwe
Paul De Grauwe
LSE Did Not Answer Bio/Vote History
Eeckhout
Jan Eeckhout
UPF Barcelona
Strongly Agree
7
Bio/Vote History
Fehr
Ernst Fehr
Universität Zurich
Strongly Agree
8
Bio/Vote History
Freixas
Xavier Freixas
Barcelona GSE Did Not Answer Bio/Vote History
Fuchs-Schündeln
Nicola Fuchs-Schündeln
Goethe-Universität Frankfurt
Agree
10
Bio/Vote History
Galí
Jordi Galí
Barcelona GSE Did Not Answer Bio/Vote History
Giavazzi
Francesco Giavazzi
Bocconi Did Not Answer Bio/Vote History
Griffith
Rachel Griffith
University of Manchester
Strongly Agree
10
Bio/Vote History
Guerrieri
Veronica Guerrieri
Chicago Booth
Agree
6
Bio/Vote History
Guiso
Luigi Guiso
Einaudi Institute for Economics and Finance
Strongly Agree
7
Bio/Vote History
Guriev
Sergei Guriev
Sciences Po
Agree
8
Bio/Vote History
Honohan
Patrick Honohan
Trinity College Dublin
Agree
3
Bio/Vote History
Javorcik
Beata Javorcik
University of Oxford
Strongly Agree
9
Bio/Vote History
Krahnen
Jan Pieter Krahnen
Goethe University Frankfurt
Agree
5
Bio/Vote History
Kőszegi
Botond Kőszegi
Central European University
Strongly Agree
10
Bio/Vote History
La Ferrara
Eliana La Ferrara
Harvard Kennedy
Strongly Agree
10
Bio/Vote History
Leuz
Christian Leuz
Chicago Booth
Agree
6
Bio/Vote History
For policy, causal inferences are critical so need credible designs. In many areas, still far from real evidence-based policy making.
-see background information here
Mayer
Thierry Mayer
Sciences-Po Did Not Answer Bio/Vote History
Meghir
Costas Meghir
Yale
Agree
8
Bio/Vote History
There has been increased emphasis on justifying sources of exogenous variation in both structural and quasi experimental studies.
Pagano
Marco Pagano
Università di Napoli Federico II
Agree
7
Bio/Vote History
Pastor
Lubos Pastor
Chicago Booth
Strongly Agree
1
Bio/Vote History
Persson
Torsten Persson
Stockholm University Did Not Answer Bio/Vote History
Pissarides
Christopher Pissarides
London School of Economics and Political Science Did Not Answer Bio/Vote History
Portes
Richard Portes
London Business School
Agree
5
Bio/Vote History
Prendergast
Canice Prendergast
Chicago Booth
Agree
8
Bio/Vote History
Propper
Carol Propper
Imperial College London
Strongly Agree
7
Bio/Vote History
Rasul
Imran Rasul
University College London
Strongly Agree
9
Bio/Vote History
Reichlin
Lucrezia Reichlin
London Business School Did Not Answer Bio/Vote History
Reis
Ricardo Reis
London School of Economics
Strongly Agree
8
Bio/Vote History
Definitely improved. Of course, one would hope that progress had been even larger and more decisive. But no doubts on the sign of progress.
Repullo
Rafael Repullo
CEMFI
Agree
6
Bio/Vote History
Rey
Hélène Rey
London Business School
Agree
8
Bio/Vote History
Schoar
Antoinette Schoar
MIT
Agree
9
Bio/Vote History
policy evaluation is complex and we need to draw on as many methods as possible for progress. But the credibility revolution is a key tool
Storesletten
Kjetil Storesletten
University of Minnesota
Strongly Agree
7
Bio/Vote History
Sturm
Daniel Sturm
London School of Economics
Strongly Agree
10
Bio/Vote History
Van Reenen
John Van Reenen
LSE
Strongly Agree
8
Bio/Vote History
Vickers
John Vickers
Oxford
Agree
4
Bio/Vote History
Voth
Hans-Joachim Voth
University of Zurich
Agree
8
Bio/Vote History
Whelan
Karl Whelan
University College Dublin Did Not Answer Bio/Vote History
Wyplosz
Charles Wyplosz
The Graduate Institute Geneva
Agree
3
Bio/Vote History
Zilibotti
Fabrizio Zilibotti
Yale University
Agree
8
Bio/Vote History

Question C Participant Responses

Participant University Vote Confidence Bio/Vote History
Allen
Franklin Allen
Imperial College London
Agree
6
Bio/Vote History
Unfortunately, there is some truth in this statement. It's difficult to identify good natural experiments for many questions.
Antras
Pol Antras
Harvard
Uncertain
5
Bio/Vote History
I don't know what's meant by "good". I see a trade-off between credibility and generality (or external validity), but questions aren't "bad"
Bandiera
Oriana Bandiera
London School of Economics
Uncertain
9
Bio/Vote History
I agree that some researchers do this but you can't blame the method, it's like saying knives are bad because people use them to hurt others
Blanchard
Olivier Blanchard
Peterson Institute
Uncertain
7
Bio/Vote History
The initial phase was indeed good answers. But we have moved over time to good questions.
Bloom
Nicholas Bloom
Stanford
Agree
10
Bio/Vote History
There is a trade-off between the quality of the question and the quality of the answer, but as long as we are on the frontier all is good.
Blundell
Richard William Blundell
University College London
Uncertain
5
Bio/Vote History
Carletti
Elena Carletti
Bocconi
Agree
6
Bio/Vote History
Danthine
Jean-Pierre Danthine
Paris School of Economics
Uncertain
5
Bio/Vote History
De Grauwe
Paul De Grauwe
LSE Did Not Answer Bio/Vote History
Eeckhout
Jan Eeckhout
UPF Barcelona
Agree
6
Bio/Vote History
Fehr
Ernst Fehr
Universität Zurich
Agree
6
Bio/Vote History
Freixas
Xavier Freixas
Barcelona GSE Did Not Answer Bio/Vote History
Fuchs-Schündeln
Nicola Fuchs-Schündeln
Goethe-Universität Frankfurt
Agree
6
Bio/Vote History
Galí
Jordi Galí
Barcelona GSE Did Not Answer Bio/Vote History
Giavazzi
Francesco Giavazzi
Bocconi Did Not Answer Bio/Vote History
Griffith
Rachel Griffith
University of Manchester
Disagree
10
Bio/Vote History
Guerrieri
Veronica Guerrieri
Chicago Booth
Uncertain
6
Bio/Vote History
Guiso
Luigi Guiso
Einaudi Institute for Economics and Finance
Strongly Agree
7
Bio/Vote History
There is indeed a risk that the method drives the question and the risk is already real
Guriev
Sergei Guriev
Sciences Po
Agree
6
Bio/Vote History
Honohan
Patrick Honohan
Trinity College Dublin
Uncertain
3
Bio/Vote History
If poorly addressed, even good questions can elicit bad answers.
Javorcik
Beata Javorcik
University of Oxford
Strongly Agree
9
Bio/Vote History
Krahnen
Jan Pieter Krahnen
Goethe University Frankfurt
Disagree
5
Bio/Vote History
Kőszegi
Botond Kőszegi
Central European University
Agree
6
Bio/Vote History
La Ferrara
Eliana La Ferrara
Harvard Kennedy
Uncertain
10
Bio/Vote History
Leuz
Christian Leuz
Chicago Booth
Uncertain
6
Bio/Vote History
While this tendency can be observed, I am uncertain about distortions from it. On net, emphasis on credible designs has moved us forward.
Mayer
Thierry Mayer
Sciences-Po Did Not Answer Bio/Vote History
Meghir
Costas Meghir
Yale
Uncertain
8
Bio/Vote History
Pagano
Marco Pagano
Università di Napoli Federico II
Strongly Agree
8
Bio/Vote History
Pastor
Lubos Pastor
Chicago Booth
Strongly Agree
5
Bio/Vote History
Persson
Torsten Persson
Stockholm University Did Not Answer Bio/Vote History
Pissarides
Christopher Pissarides
London School of Economics and Political Science Did Not Answer Bio/Vote History
Portes
Richard Portes
London Business School
Agree
5
Bio/Vote History
Prendergast
Canice Prendergast
Chicago Booth
Agree
8
Bio/Vote History
Propper
Carol Propper
Imperial College London
Strongly Agree
7
Bio/Vote History
Rasul
Imran Rasul
University College London
Disagree
5
Bio/Vote History
Reichlin
Lucrezia Reichlin
London Business School Did Not Answer Bio/Vote History
Reis
Ricardo Reis
London School of Economics
Uncertain
7
Bio/Vote History
Within set of good answers, they pick the better questions. Researchers trade off the two, unclear that they do so sub-optimally.
Repullo
Rafael Repullo
CEMFI
No Opinion
Bio/Vote History
Rey
Hélène Rey
London Business School
Agree
7
Bio/Vote History
Schoar
Antoinette Schoar
MIT
Agree
8
Bio/Vote History
I believe that methods can indeed shape the questions researchers ask. But this is true for all tools, including structural estimation etc.
Storesletten
Kjetil Storesletten
University of Minnesota
Uncertain
3
Bio/Vote History
Sturm
Daniel Sturm
London School of Economics
Uncertain
8
Bio/Vote History
Van Reenen
John Van Reenen
LSE
Disagree
6
Bio/Vote History
Vickers
John Vickers
Oxford
Uncertain
4
Bio/Vote History
Voth
Hans-Joachim Voth
University of Zurich
Uncertain
5
Bio/Vote History
Whelan
Karl Whelan
University College Dublin Did Not Answer Bio/Vote History
Wyplosz
Charles Wyplosz
The Graduate Institute Geneva
Strongly Agree
5
Bio/Vote History
Zilibotti
Fabrizio Zilibotti
Yale University
Agree
10
Bio/Vote History