Excitation Pullbacks - looking behind the curtain of gradient noise
We present a potent explanation method for (pretrained) Deep Neural Networks, the Excitation Pullback, which is a simple modification of the vanilla gradient.
For details, check out our paper and its corresponding code repository.
Note that there is still a lot of room for improvement as we use just a single architectural hyperparameter (temp) for every hidden neuron (ideally, it should be adjusted for every neuron independently).