Experiments, Quasi-experiments, and Definitely not Experiments

What’s so great about randomization and control anyway?

An important question to ask for any study design is: what causal effects are, or could be identified here? Often the answer is ‘none’, and that is an important thing to know. Not only is purely descriptive research important in itself, but all causal inferences from data are still descriptive studies when the causal claims are removed.

More typically we must ask: what assumptions - in the broadest sense - would it be necessary to make in order to identify an effect of interest in this study?

This week we revisit the ‘experimental ideal’ for causal inference with a critical eye. In most policy contexts we cannot, or should not, randomize or control ‘treatment’ assignment, so one question is how close we can get to the ideal.

Conversely, when we do manage to get an experiment running, particularly a field experiment, then lots of things can happen that make our happy randomization and control go wrong, e.g. non-compliance, attrition (‘drop out’) and missing data of a less drastic variety. At this point we have an partly ‘observational’ study again, so we may as well get used to it.

Either way it’ll be useful to work in a framework that doesn’t strongly distinguish experimental and non-experimental work. The general issue will be: how close to the experimental ideal can we get?

Finally we will consider methods of adaptive experimentation where the sample size and design is dynamically allocated. This methodology was developed in computer science for ‘active’, and more generally ‘reinforcement learning’ applications, but is now used for large scale experimentation, e.g.  in the tech sector. We will look at an example using Thompson sampling.

Readings (for weeks 1 and 2)

Lecture