One could get a better R squared for having more variables, but that potentially induces overfitting for the sample of data you are looking at. One can look at BIC where the model selection criterion is penalized by having too many variables.
3 Answers
One could get a better R squared for having more variables, but that potentially induces overfitting for the sample of data you are looking at. One can look at BIC where the model selection criterion is penalized by having too many variables.
Additional variables may improve R^2 but may lead to multicollinearity
To comment on this, Sign In or Sign Up.
Would you like us to review something? Please describe the problem with this {0} and we will look into it.
Your feedback has been sent to the team and we'll look into it.
having more covariates will in general give a better fit, however, this does not necessarily mean a better model (e.g., in terms of generalization). Thus, model comparison is carried out at the end in terms of (1) how the model explains the data (e.g., likelihood, R2, etc), and (2) how simple the model is (i.e., Occam's razor)