US

Natural Experiments in Labor Economics and Beyond

The 2021 Nobel Prize in Economic Sciences was recently awarded to David Card of the University of California, Berkeley, ‘for his empirical contributions to labour economics’, and to Joshua Angrist of MIT and Guido Imbens of Stanford University ‘for their methodological contributions to the analysis of causal relationships’. As has become an annual tradition at the IGM, we invited our panels to express their views on the work of the new laureates. We asked the experts whether they agreed or disagreed with the following statements, and, if so, how strongly and with what degree of confidence:

a) The introduction of natural experiments to economic analysis of the labor market and related areas has led to a more precise understanding of cause and effect.

b) The ‘credibility revolution’ in empirical economics has improved our understanding of a number of public policy issues, including education, immigration and the minimum wage.

c) In pursuit of credible research designs, researchers often seek good answers instead of good questions.

Background on the first two questions has been provided in much of the coverage of the new Nobels, including a VoxEU column by Jörn-Steffen Pischke, who has co-authored books and papers with Joshua Angrist. The third was stimulated by their 2010 Journal of Economic Perspectives paper on the credibility revolution, which closes by discussing ‘the claim that the experimentalist paradigm leads researchers to look for good experiments, regardless of whether the questions they address are important.’

Of our 43 US experts, 41 participated in this survey; of our 48 European experts, 39 participated – for a total of 80 expert reactions.

Understanding cause and effect

On the first statement about natural experiments leading to a more precise understanding of cause and effect, a big majority of respondents on both panels agree. Weighted by each expert’s confidence in their response, 72% of the US panel strongly agree, 24% agree, 0% are uncertain, and 3% disagree (the totals don’t always sum to 100 because of rounding). Among the European panel (again weighted by each expert’s confidence in their response), 66% strongly agree, 31% agree, 3% are uncertain and 0% disagree.

Overall, across both panels, 69% strongly agree, 27% agree, 2% are uncertain, and 2% disagree.

Among the short comments that the experts are able to include in their responses, Larry Samuelson at Yale says: ‘Natural experiments are a welcome complement to other tools.’ Jan Pieter Krahnen at Goethe University Frankfurt comments: ‘Using natural experiments is an intelligent way to build causal inference on the diversity of institutions, shocks and behavior.’ And Franklin Allen at Imperial remarks: ‘As long as the natural experiment is well designed, these studies have the potential to increase our knowledge.’

Others mention applications beyond labor economics. Christian Leuz at Chicago replies:
‘Agree but not only in labor and micro, but also many other applied fields like finance and accounting.’ Pete Klenow at Stanford provides a reference to a survey of natural experiments in macroeconomics by European panel member Nicola Fuchs-Schundeln and a colleague. And Christopher Udry at Northwestern links to two experimental studies in development economics.

A few panelists express reservations. Anil Kashyap at Chicago notes: ‘Can’t answer all questions with these approaches, but we have definitely learned a lot.’ Patrick Honohan at Trinity College Dublin observes: ‘Although, as with randomized control trials, the evidence often comes from particular contexts that may not be generalizable.’ And Costas Meghir at Yale, who votes uncertain, adds: ‘Use of natural experiments without models relies on behavioral assumptions that are often unstated. May learn little about mechanisms.’

The ‘credibility revolution’ in empirical economics

On the second statement about the impact of the credibility revolution on policy debates in education, immigration and the minimum wage, again there is a big majority in agreement on both panels. Weighted by each expert’s confidence in their response, 60% of the US panel strongly agree, 32% agree, 7% are uncertain, and 0% disagree. Among the European panel (again weighted by each expert’s confidence in their response), 48% strongly agree, 52% agree, and 0% are uncertain or disagree.

Overall, across both panels, 54% strongly agree, 42% agree, 4% are uncertain, and 0% disagree.

Among the comments, Nicholas Bloom at Stanford says: ‘The research on the minimum wage has been pathbreaking – I absolutely changed my mind based on the evidence, and it has driven policy.’ Ricardo Reis at LSE adds: ‘Definitely improved. Of course, one would hope that progress had been even larger and more decisive. But no doubts on the sign of progress.’ And Costas Meghir notes: ‘There has been increased emphasis on justifying sources of exogenous variation in both structural and quasi-experimental studies.’

Several panelists refer to where natural experiments sit relative to other methods in research and policy evaluation. Larry Samuelson comments: ‘I view it as credibility evolution, with techniques that have built upon and evolved alongside other methods.’ Antoinette Schoar at MIT suggests: ‘Policy evaluation is complex and we need to draw on as many methods as possible for progress. But the credibility revolution is a key tool.’ Christian Leuz adds: ‘For policy, causal inferences are critical so need credible designs.’

Robert Shimer at Chicago, who votes uncertain, comments: ‘The devil is in the details with identification; and researchers often explore limited outcomes, e.g. only the short run.’ Also voting uncertain is Angus Deaton at Princeton, who says: Some plusses lots of minuses.’

Good answers and good questions

The third question – inviting views on whether in pursuit of credible research designs, researchers often seek good answers instead of good questions – generated considerably more differences of opinion. Weighted by each expert’s confidence in their response, 4% of the US panel strongly agree, 63% agree, 23% are uncertain, 7% disagree and 4% strongly disagree. Among the European panel (again weighted by each expert’s confidence in their response), 17% strongly agree, 36% agree, 37% are uncertain, and 10% disagree.

Overall, across both panels, 10% strongly agree, 49% agree, 30% are uncertain, 9% disagree, and 2% strongly disagree.

Among the comments of those who agree, Franklin Allen notes: ‘Unfortunately, there is some truth in this statement. It’s difficult to identify good natural experiments for many questions.’ Larry Samuelson adds: ‘This is the case with much of economics, and the social sciences more generally.’ And Anil Kashyap responds: ‘Running joke in the faculty lounge after some shock: “we will see a diff-in-diff paper on the job market using that variation next year…”’.

Others comment on incentives within economic research. Aaron Edlin at Berkeley states: ‘The profession continues to place a premium on clever identification strategies.’ Kenneth Judd at Stanford argues: ‘Too many economists focus on tractability, and even demand that one should know what the results will be before doing the analysis.’ And Luigi Guiso at the Einaudi Institute for Economics and Finance observes: ‘There is indeed a risk that the method drives the question and the risk is already real.’

Several panelists demur at ‘often’ in the statement. David Autor at MIT says: ‘This sometimes happens, for sure. But “often” is too high a bar for me. So, I disagree’; and Barry Eichengreen at Berkeley states: ‘If the question had been worded “sometimes” rather than “often” I would have agreed.’ David Cutler at Harvard concurs: ‘There are certainly some studies like this. “Often” is the key word here’; as does Daron Acemoglu at MIT: ‘I would say “sometimes” rather than “often”. Credibility revolution is fantastic for economics but we can/should not sacrifice big questions.’

Other panelists go further. Oriana Bandiera at LSE argues: ‘I agree that some researchers do this but you can’t blame the method, it’s like saying knives are bad because people use them to hurt others.’ Similarly, Antoinette Schoar comments: ‘I believe that methods can indeed shape the questions researchers ask. But this is true for all tools, including structural estimation, etc.’

Still others are critical of the way the question is framed. Richard Thaler at Chicago says: ‘Don’t love the wording here. Researchers try to answer some question, often it is not the most interesting one.’ Pol Antras at Harvard comments: ‘I don’t know what’s meant by “good”. I see a trade-off between credibility and generality (or external validity), but questions aren’t “bad”.’ And Abhjit Banerjee at MIT protests: ‘I really don’t know what that means. Should we pursue good questions which can’t be answered? Wittgenstein had it right.’

In similar vein, Richard Schmalensee at MIT argues: ‘Working on good questions that can’t be given good answers is a waste of effort; researchers aren’t wrong to consider quality of answers.’ And Michael Greenstone at Chicago objects: ‘This is a straw man debating style criticism. Is the alternative bad answers to good questions? Count me out for that!’

A final set of comments portray a trade-off. Ricardo Reis remarks: ‘Within set of good answers, they pick the better questions. Researchers trade off the two, unclear that they do so sub-optimally.’ Nicholas Bloom adds: ‘There is a trade-off between the quality of the question and the quality of the answer, but as long as we are on the frontier, all is good.’ And Olivier Blanchard at the Peterson Institute concludes: ‘The initial phase was indeed good answers. But we have moved over time to good questions.’

All comments made by the experts are in the full survey results.

Romesh Vaitilingam
@econromesh
October 2021

Question A:

The introduction of natural experiments to economic analysis of the labor market and related areas has led to a more precise understanding of cause and effect.

Responses weighted by each expert's confidence

Question B:

The ‘credibility revolution’ in empirical economics has improved our understanding of a number of public policy issues, including education, immigration and the minimum wage.

Responses weighted by each expert's confidence

Question C:

In pursuit of credible research designs, researchers often seek good answers instead of good questions.

Responses weighted by each expert's confidence

Question A Participant Responses

Participant University Vote Confidence Bio/Vote History
Acemoglu
Daron Acemoglu
MIT
Agree
6
Bio/Vote History
Altonji
Joseph Altonji
Yale
Strongly Agree
10
Bio/Vote History
Auerbach
Alan Auerbach
Berkeley
Strongly Agree
7
Bio/Vote History
Autor
David Autor
MIT
Strongly Agree
10
Bio/Vote History
Baicker
Katherine Baicker
University of Chicago
Strongly Agree
3
Bio/Vote History
Banerjee
Abhijit Banerjee
MIT
Strongly Agree
8
Bio/Vote History
Bertrand
Marianne Bertrand
Chicago
Strongly Agree
8
Bio/Vote History
Brunnermeier
Markus Brunnermeier
Princeton
Strongly Agree
8
Bio/Vote History
Chetty
Raj Chetty
Harvard
Strongly Agree
10
Bio/Vote History
Chevalier
Judith Chevalier
Yale Did Not Answer Bio/Vote History
Cutler
David Cutler
Harvard
Strongly Agree
10
Bio/Vote History
Deaton
Angus Deaton
Princeton
Disagree
10
Bio/Vote History
Duffie
Darrell Duffie
Stanford
Strongly Agree
3
Bio/Vote History
Edlin
Aaron Edlin
Berkeley
Agree
7
Bio/Vote History
Eichengreen
Barry Eichengreen
Berkeley
Strongly Agree
5
Bio/Vote History
Einav
Liran Einav
Stanford
Strongly Agree
7
Bio/Vote History
Fair
Ray Fair
Yale
Agree
5
Bio/Vote History
Finkelstein
Amy Finkelstein
MIT
Strongly Agree
10
Bio/Vote History
Goldberg
Pinelopi Goldberg
Yale
Agree
8
Bio/Vote History
Goolsbee
Austan Goolsbee
Chicago
Strongly Agree
10
Bio/Vote History
Greenstone
Michael Greenstone
University of Chicago
Strongly Agree
9
Bio/Vote History
Hall
Robert Hall
Stanford
Agree
4
Bio/Vote History
Hart
Oliver Hart
Harvard
Strongly Agree
7
Bio/Vote History
Holmström
Bengt Holmström
MIT
Strongly Agree
5
Bio/Vote History
Hoxby
Caroline Hoxby
Stanford
Strongly Agree
10
Bio/Vote History
Hoynes
Hilary Hoynes
Berkeley
Strongly Agree
9
Bio/Vote History
Judd
Kenneth Judd
Stanford
Agree
4
Bio/Vote History
Kaplan
Steven Kaplan
Chicago Booth
Agree
5
Bio/Vote History
Kashyap
Anil Kashyap
Chicago Booth
Agree
5
Bio/Vote History
Can't answer all questions with these approaches, but we have definitely learned a lot
Klenow
Pete Klenow
Stanford
Strongly Agree
5
Bio/Vote History
Levin
Jonathan Levin
Stanford
Strongly Agree
8
Bio/Vote History
Maskin
Eric Maskin
Harvard
Agree
8
Bio/Vote History
Nordhaus
William Nordhaus
Yale
Strongly Agree
7
Bio/Vote History
Obstfeld
Maurice Obstfeld
Berkeley
Strongly Agree
5
Bio/Vote History
Saez
Emmanuel Saez
Berkeley
Strongly Agree
8
Bio/Vote History
Samuelson
Larry Samuelson
Yale
Agree
8
Bio/Vote History
Natural experiments are a welcome complement to other tools.
Scheinkman
José Scheinkman
Columbia University Did Not Answer Bio/Vote History
Schmalensee
Richard Schmalensee
MIT
Agree
5
Bio/Vote History
Shapiro
Carl Shapiro
Berkeley
Strongly Agree
10
Bio/Vote History
Shimer
Robert Shimer
University of Chicago
Agree
6
Bio/Vote History
Stock
James Stock
Harvard
Strongly Agree
4
Bio/Vote History
Thaler
Richard Thaler
Chicago Booth
Strongly Agree
5
Bio/Vote History
Udry
Christopher Udry
Northwestern
Strongly Agree
10
Bio/Vote History

Question B Participant Responses

Participant University Vote Confidence Bio/Vote History
Acemoglu
Daron Acemoglu
MIT
Agree
7
Bio/Vote History
Altonji
Joseph Altonji
Yale
Strongly Agree
10
Bio/Vote History
Auerbach
Alan Auerbach
Berkeley
Strongly Agree
7
Bio/Vote History
Autor
David Autor
MIT
Strongly Agree
10
Bio/Vote History
Baicker
Katherine Baicker
University of Chicago
Agree
3
Bio/Vote History
Banerjee
Abhijit Banerjee
MIT
Strongly Agree
8
Bio/Vote History
Bertrand
Marianne Bertrand
Chicago
Strongly Agree
8
Bio/Vote History
Brunnermeier
Markus Brunnermeier
Princeton
Agree
8
Bio/Vote History
Chetty
Raj Chetty
Harvard
Strongly Agree
10
Bio/Vote History
Chevalier
Judith Chevalier
Yale Did Not Answer Bio/Vote History
Cutler
David Cutler
Harvard
Strongly Agree
10
Bio/Vote History
Deaton
Angus Deaton
Princeton
Uncertain
10
Bio/Vote History
Some plusses lots of minuses
Duffie
Darrell Duffie
Stanford
Agree
2
Bio/Vote History
Edlin
Aaron Edlin
Berkeley
Agree
7
Bio/Vote History
Eichengreen
Barry Eichengreen
Berkeley
Strongly Agree
5
Bio/Vote History
Einav
Liran Einav
Stanford
Strongly Agree
7
Bio/Vote History
Fair
Ray Fair
Yale
Agree
5
Bio/Vote History
Finkelstein
Amy Finkelstein
MIT
Strongly Agree
8
Bio/Vote History
Goldberg
Pinelopi Goldberg
Yale
Agree
8
Bio/Vote History
Goolsbee
Austan Goolsbee
Chicago
Strongly Agree
9
Bio/Vote History
Greenstone
Michael Greenstone
University of Chicago
Strongly Agree
10
Bio/Vote History
Hall
Robert Hall
Stanford
Uncertain
1
Bio/Vote History
Hart
Oliver Hart
Harvard
Agree
6
Bio/Vote History
Holmström
Bengt Holmström
MIT
Agree
5
Bio/Vote History
Hoxby
Caroline Hoxby
Stanford
Strongly Agree
10
Bio/Vote History
Hoynes
Hilary Hoynes
Berkeley
Strongly Agree
9
Bio/Vote History
Judd
Kenneth Judd
Stanford
Uncertain
5
Bio/Vote History
Kaplan
Steven Kaplan
Chicago Booth
Uncertain
3
Bio/Vote History
Kashyap
Anil Kashyap
Chicago Booth
Agree
5
Bio/Vote History
Has not settled all issues, but has helped us make progress on some fronts
Klenow
Pete Klenow
Stanford
Strongly Agree
5
Bio/Vote History
Levin
Jonathan Levin
Stanford
Strongly Agree
8
Bio/Vote History
Maskin
Eric Maskin
Harvard
Agree
7
Bio/Vote History
Nordhaus
William Nordhaus
Yale
Agree
7
Bio/Vote History
Obstfeld
Maurice Obstfeld
Berkeley
Strongly Agree
5
Bio/Vote History
Saez
Emmanuel Saez
Berkeley
Strongly Agree
8
Bio/Vote History
Samuelson
Larry Samuelson
Yale
Agree
8
Bio/Vote History
I view it as credibility evolution, with techniques that have built upon and evolved alongside other methods.
Scheinkman
José Scheinkman
Columbia University Did Not Answer Bio/Vote History
Schmalensee
Richard Schmalensee
MIT
Agree
5
Bio/Vote History
Shapiro
Carl Shapiro
Berkeley
Strongly Agree
10
Bio/Vote History
Shimer
Robert Shimer
University of Chicago
Uncertain
1
Bio/Vote History
The devil is in the details with identification; and researchers often explore limited outcomes, e.g. only the short-run
Stock
James Stock
Harvard
Agree
8
Bio/Vote History
Thaler
Richard Thaler
Chicago Booth
Strongly Agree
5
Bio/Vote History
Is this a question a reliability check?
Udry
Christopher Udry
Northwestern
Strongly Agree
8
Bio/Vote History

Question C Participant Responses

Participant University Vote Confidence Bio/Vote History
Acemoglu
Daron Acemoglu
MIT
Agree
5
Bio/Vote History
I would say "sometimes" rather than "often". Credibility revolution is fantastic for economics but we can/should not sacrifice big questions
Altonji
Joseph Altonji
Yale
Agree
9
Bio/Vote History
Auerbach
Alan Auerbach
Berkeley
Agree
5
Bio/Vote History
Autor
David Autor
MIT
Disagree
8
Bio/Vote History
This sometimes happens, for sure. But "often" is too high a bar for me. So, I disagree.
Baicker
Katherine Baicker
University of Chicago
Agree
3
Bio/Vote History
Banerjee
Abhijit Banerjee
MIT
No Opinion
Bio/Vote History
I really don't know what that means. Should we pursue good questions which can't be answered? Wittgenstein had it right
Bertrand
Marianne Bertrand
Chicago
Agree
8
Bio/Vote History
Brunnermeier
Markus Brunnermeier
Princeton
Uncertain
8
Bio/Vote History
Chetty
Raj Chetty
Harvard
Agree
10
Bio/Vote History
Chevalier
Judith Chevalier
Yale Did Not Answer Bio/Vote History
Cutler
David Cutler
Harvard
Uncertain
5
Bio/Vote History
There are certainly some studies like this. "Often" is the key word here.
Deaton
Angus Deaton
Princeton
Agree
10
Bio/Vote History
Not sure what a "good" answer is, unless it is one derived by an approved method.
Duffie
Darrell Duffie
Stanford
Agree
3
Bio/Vote History
Edlin
Aaron Edlin
Berkeley
Strongly Agree
8
Bio/Vote History
The profession continues to place a premium on clever identification strategies.
Eichengreen
Barry Eichengreen
Berkeley
Uncertain
5
Bio/Vote History
If the question had been worded "sometimes" rather than "often" I would have agreed.
Einav
Liran Einav
Stanford
Agree
7
Bio/Vote History
Fair
Ray Fair
Yale
Disagree
5
Bio/Vote History
Finkelstein
Amy Finkelstein
MIT
Uncertain
5
Bio/Vote History
Goldberg
Pinelopi Goldberg
Yale
Agree
8
Bio/Vote History
Goolsbee
Austan Goolsbee
Chicago
Agree
8
Bio/Vote History
Greenstone
Michael Greenstone
University of Chicago
Strongly Disagree
8
Bio/Vote History
this a straw man debating style criticism. is the alternative bad answers to good questions? count me out for that!
Hall
Robert Hall
Stanford
Uncertain
1
Bio/Vote History
Hart
Oliver Hart
Harvard
Uncertain
5
Bio/Vote History
Holmström
Bengt Holmström
MIT
Agree
3
Bio/Vote History
Hoxby
Caroline Hoxby
Stanford
Uncertain
7
Bio/Vote History
Hoynes
Hilary Hoynes
Berkeley
Agree
8
Bio/Vote History
Judd
Kenneth Judd
Stanford
Agree
8
Bio/Vote History
Too many economists focus on tractability, and even demand that one should know what the results will be before doing the analysis.
Kaplan
Steven Kaplan
Chicago Booth
Uncertain
3
Bio/Vote History
Kashyap
Anil Kashyap
Chicago Booth
Agree
5
Bio/Vote History
running joke in the faculty lounge after some shock, "we will see a diff in diff paper on the job market using that variation next year..."
Klenow
Pete Klenow
Stanford
Agree
10
Bio/Vote History
Levin
Jonathan Levin
Stanford
Disagree
3
Bio/Vote History
Maskin
Eric Maskin
Harvard
No Opinion
Bio/Vote History
Nordhaus
William Nordhaus
Yale
No Opinion
Bio/Vote History
Obstfeld
Maurice Obstfeld
Berkeley
Uncertain
5
Bio/Vote History
Saez
Emmanuel Saez
Berkeley
Uncertain
7
Bio/Vote History
Samuelson
Larry Samuelson
Yale
Agree
8
Bio/Vote History
This is the case with much of economics, and the social sciences more generally.
Scheinkman
José Scheinkman
Columbia University Did Not Answer Bio/Vote History
Schmalensee
Richard Schmalensee
MIT
Agree
5
Bio/Vote History
Working on good questions that can't be given good answers is a waste of effort; researchers aren't wrong to consider quality of answers.
Shapiro
Carl Shapiro
Berkeley
Uncertain
1
Bio/Vote History
Shimer
Robert Shimer
University of Chicago
Agree
7
Bio/Vote History
Stock
James Stock
Harvard
Agree
5
Bio/Vote History
Thaler
Richard Thaler
Chicago Booth
Uncertain
1
Bio/Vote History
Don't love the wording here. Researchers try to answer some question, often it is not the most interesting one.
Udry
Christopher Udry
Northwestern
Agree
8
Bio/Vote History
"Often" may overstate it, but yes it happens.