Lasso Regression

Describing Lasso Regression

  • Lasso regression is a type of regression that performs L1 regularization on the OLS coefficients
  • Specifically, the penalty term associated with L1 regularization is the absolute value of the magnitude of OLS coefficients
  • Lasso regression mitigates the problem of multicollinearity in regression, and thus mitigates the problem of overfitting
  • Lasso regression achieves this by shrinking the OLS cofficients to exactly zero

Mathematics behind Lasso Regression

  • When using OLS, the coefficients can often create a ridge in the coefficient space, meaning many different coefficients on the space can do as well (or nearly as well)
  • By adding the penalty (or tuning factor), the coefficient space can be lifted up to provide better coefficient estimates compared to the OLS coefficient space
  • This adjusted coefficient space doesn't guarantee better coefficient estimates, but it can be helpful to explore when looking at additional regression techniques

Effects of Lasso Regression

  • Correlation in parameter estimates is reduced, which will mitigate the problem of multicollinearity
  • Parameter estimates won't be very large in magnitude if the RSS for small parameters aren't much worse, which also mitigates the problem of overfitting
  • Some parameter estimates will quickly drop to zero, which provides the function of variable selection

References

Previous
Next

Ridge Regression

Spline Regression