R-squared is a statistical measure of how close the data are to the fitted regression line. It is also known as the coefficient of determination, or the coefficient of multiple determination for multiple regression.

This metric is specifically designed for regression-based algorithms where the output is a real value.

Computing Coefficient of determination

Let x, y, y^ be the input, output and predicted output vectors in linear regression.

Lets assume there are n points So for a point with input xi, output yi and predicted output yi^

We can define error ei be defined as ei = yi – yi^

Sum of Squares Total

Let say the mean of all the outputs in the training data be y_mean.

Hence

y_mean = (y1+y2+….yn)/n

or

Then we can define sum of squares total as

Sum of Squares, SStotal = ((y1 – y_mean)2 + (y2 – y_mean)2… (yn – y_mean)2)

Lets assume a very naive model for which for every new query point, we just return the average y_mean for any new query point.

In this case Sum of Squares would be 0.

Sum of Squares Residual

Similarly we can define SSresidue as the sum of sqaures of difference between expected output and predicted output.

SSresidue = ((y1 – y1^)2 + (y2 – y2^)2… (yn – yn^)2)

where ei = yi − fi 

For the naive model case,since every predicted value is mean

SSresidue=SStotal

Coefficient of determination

Hence we can say coefficient of determination is

R2 can be seen to be related to the fraction of variance unexplained (FVU), since the second term compares the unexplained variance (variance of the model’s errors) with the total variance (of the data)

Now there are 4 cases related to this

Case 1: SSresidue=0 In this case R2=1,

which is the best case we could have.

Case 2: SSresidue=SStotal, In this case R2=0, which is the same as the naive mean model case.

Case 3:SSresidue < SStotal In this case R2 lies between 0 and 1, which would be for most cases.

Case 4:SSresidue > SStotal In this case R2 is less than 0.

This would be the worst case. This means our model is behaving worse than the naive mean model.

>>> from sklearn.metrics import r2_score
>>> y_true = [3, -0.5, 2, 7]
>>> y_pred = [2.5, 0.0, 2, 8]
>>> r2_score(y_true, y_pred)  
0.948...
>>> y_true = [[0.5, 1], [-1, 1], [7, -6]]
>>> y_pred = [[0, 2], [-1, 2], [8, -5]]
>>> r2_score(y_true, y_pred,
...          multioutput='variance_weighted') 
0.938...
>>> y_true = [1, 2, 3]
>>> y_pred = [1, 2, 3]
>>> r2_score(y_true, y_pred)
1.0
>>> y_true = [1, 2, 3]
>>> y_pred = [2, 2, 2]
>>> r2_score(y_true, y_pred)
0.0
>>> y_true = [1, 2, 3]
>>> y_pred = [3, 2, 1]
>>> r2_score(y_true, y_pred)
-3.0

0 Comments

What do you think?