202405182101
Status: #idea
Tags: Hypothesis Testing in Multiple Linear Regression

Extra Sum of Squares Method
This is a similar idea to what we do when we are checking significance of model except that instead of checking the full set of regressors at once with Analysis of Variance (ANOVA), we select a subset of coefficients that we suspect might be useless and test them.
Intuition
When I have the intuition that a subset of regressors are useless, I am essentially intuiting that the smaller model called Restricted Model does not have significantly less predictive power than the model with all the regressors called Full Model.
In other terms, we say that we can with impunity remove those parameters and keep the power the same. This is not usually expected since in general as the number of regressors increase, the predictive power of our model increases. This can be stated as "as the number of regressors increase, the
So our null hypothesis will essentially say that the restricted model is enough, in other terms that
From there we proceed.
Method
The idea is to split our
In so doing, we effectively have two models to compare. And from these two models we can compute a
So we are comparing these models through their respective
By taking the difference between the
This difference will have
Note that the full model has
We now want to see how significant of a difference the added parameters make.
As we often do in statistic, whenever we want to compare two things, and that those two things happen to be independently distributed according to a
If this F statistic is bigger than the appropriate
After all, if the difference between the two models is not squashed to oblivion by dividing by MSE, it means that the full model has much more (statistically significantly more) explanatory power than the restricted model.
Video Screenshots

