1 R-squared

\[R^2 = \frac{\sum(y_i - \bar{y})^2 - \sum(y_i - \hat{y_i})^2}{\sum(y_i - \bar{y})^2}\] Perfect model: \(\hat{y_i} = y_i\) so \(R^2 = 1\)

2 Gini

Ratio of how close model is to perfect model and how far it is from a random model. Gini = 1 → perfect model, Gini = 0 → random model

This article derives Gini using three methods: from the CAP curve, from the Lorenz curve, and from the ROC curve.

From CAP curve:

From Lorenz curve:

From ROC curve:

3 Lift Chart

3.1 Single Lift

  • Order data by prediction/weight, group into bins (e.g. deciles) with equal weight (e.g. earned car years)

  • Within each bin, get average predicted and actual value

3.2 Double Lift

For comparing 2 predictions to actual

  • Order data by prediction_1/prediction_2, group into bins

  • Within each bin, get averaged prediction_1, prediction_2, and actual value

4 Binomial Deviance

Deviance = \(2 \sum o_i log \left( \frac{o_i}{e_i} \right)\) where \(o_i\) denotes observed, \(e_i\) denotes expected, and the sum is over both successes and failures for each i

5 Metrics from Confusion Matrix

6 Confidence Intervals for Machine Learning

See here for more details

Nonparametric Confidence Interval

7 Nested Cross Validation

Motivation/Theory, Method/Figure

K-fold cross-validation is used both in the selection of model hyperparameters to configure each model and in the selection of configured model and can lead to overfitting. Each time a model with different model hyperparameters is evaluated on a dataset, it provides information about the dataset.

Method

Cost of validation