deep-rules icon indicating copy to clipboard operation
deep-rules copied to clipboard

Treat it as a statistical model

Open blengerich opened this issue 5 years ago • 1 comments

Have you checked the list of proposed rules to see if the rule has already been proposed?

  • [x] Yes

This rule is related to @ttriche's comments in #5. Because DL models can be difficult to interpret intuitively, there is a temptation to anthropomorphize DL models. We should resist this temptation. We would not expect a linear model to learn beyond pattern recognition, so we should also not expect an overparameterized non-linear model to do so. Because of this, we need to pay just as much attention to statistical caveats with DL models as we do with traditional models. For instance:

  • Don't use a deep learning model to extrapolate beyond the domain of the training set. While architectures with saturating nonlinearities will produce outputs in a reasonable range outside the training domain, that does not mean the predictions are reliable.
  • Don't use predictive accuracy or interpretability as evidence of casual reasoning (#9). We wouldn't interpret R^2 or MSE as evidence of causal reasoning in statistical models; don't do it for DL models either.
  • Interrogate the model for input sensitivity. Does your model respond to confounding effects, translations on the data domain, etc.? Do you want it to? Just as we would inspect the coefficients of a linear model, we should inspect the sensitivity of DL models to understand which signals they identify.

blengerich avatar Nov 12 '18 14:11 blengerich

"There is a temptation to anthropomorphize DL models. We should resist this temptation."

Well said, particularly with respect to #26. Both for better and for worse, DL models are not humans. Recognizing their capabilities and limitations is key for using them properly.

Benjamin-Lee avatar Nov 13 '18 01:11 Benjamin-Lee