28
Introduction Review Spline LASSO Spline LASSO with Thresholding Summary Spline LASSO with thresholding in high-dim regression Bing-Yi Jing June 10, 2013 Bing-Yi Jing Spline LASSO with thresholding in high-dim regression

Bing-Yi Jing

  • Upload
    votruc

  • View
    277

  • Download
    6

Embed Size (px)

Citation preview

Page 1: Bing-Yi Jing

IntroductionReview

Spline LASSOSpline LASSO with Thresholding

Summary

Spline LASSO with thresholding in high-dimregression

Bing-Yi Jing

June 10, 2013

Bing-Yi Jing Spline LASSO with thresholding in high-dim regression

Page 2: Bing-Yi Jing

IntroductionReview

Spline LASSOSpline LASSO with Thresholding

Summary

Outline

1 Introduction

2 Review

3 Spline LASSO

4 Spline LASSO with Thresholding

5 Summary

Bing-Yi Jing Spline LASSO with thresholding in high-dim regression

Page 3: Bing-Yi Jing

IntroductionReview

Spline LASSOSpline LASSO with Thresholding

Summary

High Dimensional Problems with Ordered Features

Small n and large p problem: p � nSparsity:

Number of influential features k is small.

Feathers are ordered, and correlated.Examples:

protein mass spectroscopy datagene expression data

Bing-Yi Jing Spline LASSO with thresholding in high-dim regression

Page 4: Bing-Yi Jing

IntroductionReview

Spline LASSOSpline LASSO with Thresholding

Summary

Objectives in model selection:

Model sparsityFeathers (variables) selectionPrediction power

Bing-Yi Jing Spline LASSO with thresholding in high-dim regression

Page 5: Bing-Yi Jing

IntroductionReview

Spline LASSOSpline LASSO with Thresholding

Summary

Linear regression model

Given (Yi , xi1, ..., xip), i = 1, ..., n, assume

Yi = β0 + β1xi1 + ... + βpxip + εi , i = 1, ..., n.

In matrix form,Y = Xβ + ε

where

Y =

Y1Y2...

Yn

, X =

1 x11 · · · x1p1 x21 · · · x2p...

.... . .

...1 xn1 · · · anp

, β =

β0β1...

βp

, ε =

ε1ε2...εn

,

Bing-Yi Jing Spline LASSO with thresholding in high-dim regression

Page 6: Bing-Yi Jing

IntroductionReview

Spline LASSOSpline LASSO with Thresholding

Summary

One Example: protein mass spectroscopy data, Adamet al (2003)

For each blood serum sample i ,xij : intensity for many time-of-flight value tjtime-of-flight: related to mass over charge ratio m/zp = 48538 m/z-sitesn1 = 157 healthy patients, n2 = 167 with cancer.

Objective:find m/z-sites discriminating between 2 groups.

Bing-Yi Jing Spline LASSO with thresholding in high-dim regression

Page 7: Bing-Yi Jing

IntroductionReview

Spline LASSOSpline LASSO with Thresholding

Summary

Least square estimate (LSE)

LSE:β̂ lse = {

∑i

(yi −∑

j

xijβj)2} = (X T X )−1XY

ill-posed if p > n

Bing-Yi Jing Spline LASSO with thresholding in high-dim regression

Page 8: Bing-Yi Jing

IntroductionReview

Spline LASSOSpline LASSO with Thresholding

Summary

Ridge regression

Ridge Regression: (Hoerl and Kennard, Technometrics, 1970)

β̂ridge = argmin{∑

i

(yi −∑

j

xijβj)2}+ λ

p∑j=1

β2j

= (X T X + λI)−1XY

When X T X = I, then

β̂ridge =β̂ lse

1 + λ

L1 penalty does shrinkage of LSE,but no variable selection.

Bing-Yi Jing Spline LASSO with thresholding in high-dim regression

Page 9: Bing-Yi Jing

IntroductionReview

Spline LASSOSpline LASSO with Thresholding

Summary

LASSO (least absolute shrinkage and selectionoperator)

Lasso (Tibshirani (1996), JRSSB):

β̂ lasso = argmin{∑

i

(yi −∑

j

xijβj)2}+ λ

p∑j=1

|βj |

When X T X = I, then

β̂ lasso = sgn(β̂ lse)(|β̂ lse| − λ

)+

L1 penalty does shrinkage and variable selection simultaneously.It works for p > n as well.Computation: LARS (Efron et al, 2004), very efficient

Bing-Yi Jing Spline LASSO with thresholding in high-dim regression

Page 10: Bing-Yi Jing

IntroductionReview

Spline LASSOSpline LASSO with Thresholding

Summary

Variants of LASSO

Method Reference DetailElastic Net Zou and Hastie(2005) λ

∑β2

jFused Lasso Tibshirani et al .(2005) λ

∑|βj+1 − βj |

Adaptive Lasso Zou(2006) λ∑

wj |βj |Grouped Lasso Yuan and Lin(2007)

∑g ||βg ||2

Dantzig selector Candes and Tao(2007) min{||X T (y − Xβ)||∞}||β||1 < t

Bing-Yi Jing Spline LASSO with thresholding in high-dim regression

Page 11: Bing-Yi Jing

IntroductionReview

Spline LASSOSpline LASSO with Thresholding

Summary

Fused LASSO

Fused lasso (Tibshirani et al .(2005)):

β̂ = argmin{N∑

i=1

(yi − xTi β)2 + λ1

p∑j=1

|βj |+ λ2

p∑j=2

|βj − βj−1|}

Equivalently,

β̂ = argmin{∑

i

(yi −∑

j

xijβj)2}

subject to∑p

j=1 |βj | ≤ s1 and∑p

j=2 |βj − βj−1| ≤ s2.

1st penalty =⇒ sparse βj , while 2nd penalty =⇒ flatness of βj .(i.e., maintain grouping effects and sparsity of the coefficients).Fussed lasso picks grouped ordered features.

Bing-Yi Jing Spline LASSO with thresholding in high-dim regression

Page 12: Bing-Yi Jing

IntroductionReview

Spline LASSOSpline LASSO with Thresholding

Summary

Figure: N = 50, p = 80, X ∈ N(0, Σ) where Σii = 1 and Σij = 0.9|i−j|.

Drawbacks:Hard to keep shape of βj ’s within the same group.Computation:

very intensive for large p,not fully automatic

Bing-Yi Jing Spline LASSO with thresholding in high-dim regression

Page 13: Bing-Yi Jing

IntroductionReview

Spline LASSOSpline LASSO with Thresholding

Summary

Smooth LASSO (Hebiri, M. and Van de Geer,S.(2011))

β̂ = argmin{N∑

i=1

(yi − xTi β)2 + λ1

p∑j=1

|βj |+ λ2

p∑j=2

(βj − βj−1)2}

Advantages:Capture the smooth change in features in a groupComputation efficient

Disadvantages:Cannot capture changes in curvature.Less prediction power for large p,

Bing-Yi Jing Spline LASSO with thresholding in high-dim regression

Page 14: Bing-Yi Jing

IntroductionReview

Spline LASSOSpline LASSO with Thresholding

Summary

Spline LASSO

β̂ = argmin{N∑

i=1

(yi − xTi β)2 + λ1

p∑j=1

|βj |+ λ2

p−1∑j=2

(βj−1 − 2βj + βj+1)2}

The first penalty encourages sparse solution.The second penalty mimics cubic spline, i.e. penalizing largesecond-order derivatives of coefficients.

Bing-Yi Jing Spline LASSO with thresholding in high-dim regression

Page 15: Bing-Yi Jing

IntroductionReview

Spline LASSOSpline LASSO with Thresholding

Summary

Computation

Spline LASSO can be solved by LARS.

Proposition

Given dataset (Y , X ) and (λ1, λ2), define an artificial data set(Y ∗, X ∗) by

X ∗(N+p−2)×p =

(X ,

√λ2L

)T, Y ∗

(N+p−2) = (Y , 0)T.

where L is a (p − 2)× p matrix with Li,i = Li,i+2 = 1, Li,i+1 = −2 andLi,j = 0 otherwise. Then the spline lasso optimization can be writtenas

(Y ∗ − X ∗β)T (Y ∗ − X ∗β) + λ1

p∑j=1

|βj |,

which is an equivalent lasso problem and can be solved efficiently.

Bing-Yi Jing Spline LASSO with thresholding in high-dim regression

Page 16: Bing-Yi Jing

IntroductionReview

Spline LASSOSpline LASSO with Thresholding

Summary

Simulations

Model:Y = Xβ + Z ,

where (X1, . . . , Xp)T ∈ N(0,Σ), and Z ∈ N(0, σ) where

σ ∈ [0.1, 1].p > n.β’s are generated from a continuous function with k non-zeroterms, (k > n or k < n).Comparison in variable selection and prediction:

fused-lasso,smooth-lasso,spline,spline-lasso

Bing-Yi Jing Spline LASSO with thresholding in high-dim regression

Page 17: Bing-Yi Jing

IntroductionReview

Spline LASSOSpline LASSO with Thresholding

Summary

Case 1: N = 60, p = 200, k = 50, σ = 110 and

neighboring correlation is high

Figure: X ∼ N(0, Σ) where Σii = 1 and Σij = 0.9|i−j|.

Bing-Yi Jing Spline LASSO with thresholding in high-dim regression

Page 18: Bing-Yi Jing

IntroductionReview

Spline LASSOSpline LASSO with Thresholding

Summary

Case 2:N = 60, p = 300, k = 70, σ = 15 and

neighboring correlation is median

Figure: X ∼ N(0, Σ) where Σii = 1 and Σij = 0.5|i−j|.

Bing-Yi Jing Spline LASSO with thresholding in high-dim regression

Page 19: Bing-Yi Jing

IntroductionReview

Spline LASSOSpline LASSO with Thresholding

Summary

Case 3:N = 60, p = 500, k = 70, σ = 1 and all featuresare i.i.d.

Figure: X ∼ N(0, I).

Bing-Yi Jing Spline LASSO with thresholding in high-dim regression

Page 20: Bing-Yi Jing

IntroductionReview

Spline LASSOSpline LASSO with Thresholding

Summary

Prediction Accuracy

We generated 1000 testing data to compare the predictionaccuracy of the above estimates and a summary is given asfollow:

MSE Fused Lasso Smooth Lasso Spline OLS Spline LassoCase 1 73.9222 52.7745 64.8594 50.2912Case 2 201.9074 63.1969 59.2204 48.2225Case 3 258.1292 106.0806 133.276 63.4307

Bing-Yi Jing Spline LASSO with thresholding in high-dim regression

Page 21: Bing-Yi Jing

IntroductionReview

Spline LASSOSpline LASSO with Thresholding

Summary

Estimation Accuracy

We also summarize the L2 norm of the difference between theseestimates and the true β:

||β̂ − β||2 Fused Lasso Smooth Lasso Spline OLS Spline LassoCase 1 2.1621 1.3238 1.7614 1.2262Case 2 15.9208 1.7069 1.5989 1.1407Case 3 17.5091 3.1576 4.0761 1.747

Bing-Yi Jing Spline LASSO with thresholding in high-dim regression

Page 22: Bing-Yi Jing

IntroductionReview

Spline LASSOSpline LASSO with Thresholding

Summary

Why Threshold?

Spline LASSO is good at prediction, but not quite sparse.

Specificity Fused Lasso Smooth Lasso Spline OLS Spline LassoCase 1 0.9603 0.7881 0 0.7483Case 2 0.8371 0.4208 0 0.1674Case 3 0.9121 0.3682 0 0.1378

WHY? Lasso penalty (λ1) is often quite small.Solution: Applying thresholding after estimation.Thresholding can also be applied to smooth lasso and splineOLS, etc.

Bing-Yi Jing Spline LASSO with thresholding in high-dim regression

Page 23: Bing-Yi Jing

IntroductionReview

Spline LASSOSpline LASSO with Thresholding

Summary

Level of thresholding?

thresholding level ±σ̂√

2 log(N), where σ̂ is standard error ofsmall coefficients.In wavelet, Donoho and Johnstone (1998)Theoretical study is under investigation.

Bing-Yi Jing Spline LASSO with thresholding in high-dim regression

Page 24: Bing-Yi Jing

IntroductionReview

Spline LASSOSpline LASSO with Thresholding

Summary

Applying the Threshold

The specificity improved a lot while sensitivity was barely affected.

Specificity Smooth Lasso Smooth Lasso Thr Spline OLS Spline OLS Thr Spline Lasso Spline Lasso ThrCase 1 0.7881 0.9603 0 1 0.7483 0.9669Case 2 0.4208 0.9683 0 0.9864 0.1674 0.9729Case 3 0.3682 0.9549 0 0.9786 0.1378 0.9501

Sensitivity Smooth Lasso Smooth Lasso Thr Spline OLS Spline OLS Thr Spline Lasso Spline Lasso ThrCase 1 0.9592 0.9388 1 0.9184 0.9796 0.9592Case 2 0.9873 0.9747 1 0.9873 1 0.9873Case 3 0.9873 0.962 1 0.9241 1 0.9747

Bing-Yi Jing Spline LASSO with thresholding in high-dim regression

Page 25: Bing-Yi Jing

IntroductionReview

Spline LASSOSpline LASSO with Thresholding

Summary

Prediction and estimation also improved:

MSE Smooth Lasso Smooth Lasso Thr Spline OLS Spline OLS Thr Spline Lasso Spline Lasso ThrCase 1 52.7745 52.7739 64.8594 42.3359 50.2912 49.698Case 2 63.1969 59.1765 59.2204 47.7498 48.2225 45.1844Case 3 106.0806 104.24 133.276 96.0668 63.4307 61.1429||β̂ − β||2 Smooth Lasso Smooth Lasso Thr Spline OLS Spline OLS Thr Spline Lasso Spline Lasso ThrCase 1 1.3238 1.3219 1.7614 0.8796 1.2264 1.1983Case 2 1.7069 1.546 1.5989 1.12 1.1407 1.0048Case 3 3.1576 3.07 4.0761 2.7968 1.747 1.6269

Bing-Yi Jing Spline LASSO with thresholding in high-dim regression

Page 26: Bing-Yi Jing

IntroductionReview

Spline LASSOSpline LASSO with Thresholding

Summary

Summary

Spline lasso captures smooth-changing feathersIt works well in high dimensional settings with ordered features.Prediction is better than existing methods.Thresholding improves variable selection, prediction andestimation.

Bing-Yi Jing Spline LASSO with thresholding in high-dim regression

Page 27: Bing-Yi Jing

Appendix Reference

Reference

Candes, E. and Tao, T. (2007) The dantzig selector statisticalestimation when p is much larger than n. Ann. Statist., 35,2313-2351.David L. Donoho and Iain M. Johnstone. (1998) MinimaxEstimation via Wavelet Shrinkage. Ann. Statist., 26, 879-921.Hebiri, M. and Van de Geer, S. (2011) The Smooth-Lasso andOther `1 + `2−penalized methods. Electron. J. Statist. 5,1184-1226.Tibshirani, R. (1996) Regression Shrinkage and Selection via theLasso. J. R. Statist. Soc. B, 58, 267-288.Tibshirani, R., Saunders, M., Rosset, S., Zhu, J. and Knight, K.(2005) Sparsity and smoothness via the fused lasso. J. R.Statist. Soc. B, 67, 91-108.

Bing-Yi Jing Spline LASSO with thresholding in high-dim regression

Page 28: Bing-Yi Jing

Appendix Reference

Reference

Yuan, M. and Lin, Y. (2007) Model selection and estimation inregression with grouped variables. J. R. Statist. Soc. B, 68,49-67.Zou, H. (2006) The adaptive lasso and its oracle properties. J.Am. Statist. Ass., 101, 1418-1429.Zou, H. and Hastie, T. (2005) Regularization and variableselection via the elastic net. J. R. Statist. Soc. B, 67, 301-320.

Bing-Yi Jing Spline LASSO with thresholding in high-dim regression