#### Model Interpretability

Model interpretability refers to the ability for a human to understand and articulate the relationship between a model’s predictors and its outcome. For linear models, including linear and logistic regression, these relationships are seen directly in the model coefficients. For black-box models like neural nets,…