Subsampling estimates of the Lasso distribution.
Subsampling estimates of the Lasso distribution.
Subsampling estimates of the Lasso distribution.
Create successful ePaper yourself
Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.
22 Application to <strong>the</strong> <strong>Lasso</strong> estimator<br />
for some c > 0.Then<br />
almost surely.<br />
1<br />
n<br />
n∑<br />
ξ i → 0<br />
i=1<br />
Theorem 3.1.0.13. (Convergence in probability) If C is nonsingular and λ n /n →<br />
λ 0 ≥ 0, <strong>the</strong>n<br />
where<br />
ˆβ n → P arg min(Z)<br />
p∑<br />
Z(φ) = (φ − β) ′ C(φ − β) ′ + λ 0 |φ j |.<br />
j=1<br />
Thus if λ n = o(n), arg min(Z) = β and so ˆβ n is consistent.<br />
Pro<strong>of</strong>. Maps Z n defined in 3.0.0.1 have convex sample paths and are minimized at ˆβ n .<br />
Sample paths <strong>of</strong> Z are strictly convex since C is nonsingular, hence have unique minimizers.<br />
Following Corollary 2.1.0.3, it is sufficient to show that Z n (φ) → P Z(φ) + σ 2 for every<br />
point φ ∈ R p .<br />
We have<br />
Set ξ i = ε i (β − φ) ′ x i with<br />
Z n (φ) = 1 n∑<br />
(Y i − x ′<br />
n<br />
iφ) 2 + λ n<br />
p∑<br />
|φ j |<br />
n<br />
i=1<br />
j=1<br />
= 1 n∑<br />
(x ′<br />
n<br />
iβ + ε i − x ′ iφ) 2 + λ n<br />
p∑<br />
|φ j |<br />
n<br />
i=1<br />
j=1<br />
( )<br />
1<br />
n∑<br />
= (β − φ) ′ x i x ′ i (β − φ) + 1 n∑<br />
ε i<br />
n<br />
n<br />
i=1<br />
i=1<br />
+ 2 n∑<br />
ε i (β − φ) ′ x i + λ n<br />
p∑<br />
|φ j |<br />
n<br />
n<br />
i=1<br />
j=1<br />
i=1<br />
.<br />
E(ξ 2 i ) = σ 2 |〈β − φ, x i 〉| 2<br />
≤ σ 2 ‖(β − φ)‖ 2 ‖x i ‖ 2 .<br />
Under assumption 3.0.0.4 i −1 ‖x i ‖ 2 is asymptotically bounded above by i −δ for some δ > 0,<br />
thus we obtain<br />
∞∑ 1<br />
∞<br />
i 2 E(ξ2 i ) ≤ σ 2 ‖(β − φ)‖ 2 ∑<br />
( ) 1 2<br />
i ‖x i‖ < ∞.<br />
Now it follows from <strong>the</strong>orem 3.1.0.12 that<br />
2<br />
n<br />
i=1<br />
n∑<br />
ε i (β − φ) ′ x i → 0<br />
i=1