14.03.2017 Views

policymakers demonstrate

EwB_Final_WithCover

EwB_Final_WithCover

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

Big Data in Economics: Evolution or Revolution? 617<br />

Box 14.2: The Ridge Regression Estimator<br />

The ridge regression estimator is given by<br />

[ ]<br />

ˆβ ridge = argmin β ‖Y − Xβ‖<br />

2<br />

2<br />

+ λ 2 ‖β‖ 2 2<br />

= (X ′ X + λ 2 I) −1 X ′ (14.5)<br />

Y<br />

where I is the identity matrix and λ 2 > 0 is the so-called ‘regularization<br />

parameter’, which, as seen from (14.5), reduces the impact of the smallest<br />

eigenvalues of X ′ X, at the origin of the instability of the OLS estimator.<br />

An alternative to quadratic penalties that allows for variable selection by<br />

enforcing sparsity, that is, the presence of zeroes in the vector β of the regression<br />

coefficients, has been popularized in the statistics and machine-learning<br />

literature under the name of ‘Lasso regression’ by Tibshirani (1996). It consists<br />

in replacing the L 2 -norm penalty used in ridge regression by a penalty<br />

proportional to the L 1 -norm of β (see Box 14.3).<br />

In the case of orthonormal regressors, it is easily seen that the Lasso penalty<br />

provides a nonlinear shrinkage of the components of ˆβ ols , which are shrunk<br />

differently according to their magnitude, as well as sparsity, since the jth coefficient<br />

[ ˆβ lasso ] j = 0if|[X ′ Y] j |

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!