Ablation studies, knockout studies, lesion studies
In order to understand it we must be able to break it
October 26, 2016 — October 28, 2024
Can we work out how a complex or adaptive system works by destroying bits of it?
This comes up often in the study of complicated things, because there are too many variables, and too many ethical constraints, to do all the experiments we might imagine.
In various fields, it has been asked how far we can push the simple experiment of working out what breaks, which is called an ablation study in neural nets, a lesion study in neuroscience, and a knockout study in genetics.
Questions about the limits of this methodology arise in Can a biologist fix a radio (Lazebnik 2002)? Could a neuroscientist understand a microprocessor (Jonas and Kording 2017)? Should an alien doctor stop human bleeding by removing blood?
The answer, I claim, is it depends. These are classic identifiability from observational data problems.
What do we miss in such studies? Here is one answer: Interpretability Creationism:
[…] Stochastic Gradient Descent is not literally biological evolution, but post-hoc analysis in machine learning has a lot in common with scientific approaches in biology, and likewise often requires an understanding of the origin of model behaviour. Therefore, the following holds whether looking at parasitic brooding behaviour or at the inner representations of a neural network: if we do not consider how a system develops, it is difficult to distinguish a pleasing story from a useful analysis. In this piece, I will discuss the tendency towards “interpretability creationism” — interpretability methods that only look at the final state of the model and ignore its evolution over the course of training — and propose a focus on the training process to supplement interpretability research.
That essay clearly connects to model explanation, and I think, observational studies generally.