Visualize the distribution of Activations

I want to get the distribution of the activations in a MLP that uses ReLU. Any convenient ways to do this?

It can be the activations per layer, or activations of the network as a whole. I know these activations fire per input, so I am not really sure how to approach this.

Take a look at pytorch hooks I think it is what you are looking for.