47
Spline-based separable expansions for approximation, regression and classification Nithin Govindarajan Nico Vervliet, Lieven De Lathauwer IPAM Workshop I: Tensor Methods and their Applications in the Physical and Data Sciences, UCLA, United States, April 1, 2021

Spline-based separable expansions for approximation

  • Upload
    others

  • View
    5

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Spline-based separable expansions for approximation

Spline-based separable expansions for

approximation, regression and classification

Nithin Govindarajan

Nico Vervliet, Lieven De Lathauwer

IPAM Workshop I: Tensor Methods and their Applications in the Physical and DataSciences, UCLA, United States, April 1, 2021

Page 2: Spline-based separable expansions for approximation

What are we trying to accomplish?

Introduce a new technique for modeling functions in several variables:

Regression tasks

Classification tasks

Our recent submission to Frontiers:Regression and classification with spline-based separable expansions.

N. Govindarajan, N. Vervliet, L. De Lathauwer.

2

Page 3: Spline-based separable expansions for approximation

The main challenge of approximating functions in high dimensions

Curse-of-dimensionality in approximation theory:In general, to approximate a n-times differentiable function in D variableswithin ε-tolerance (measured in the uniform norm), one typically requires M &(1ε )D/n parameters

Optimal nonlinear approximation. DeVore et al., Manuscripta mathematica, 1989.

caveat:Many high-dimensional functions in applications are inherently of “low com-plexity”

3

Page 4: Spline-based separable expansions for approximation

Focus of this talk: exploiting low-rank structures through sums of separable functions

f (x) =R∑

r=1

(D∏

d=1

φr ,d(xd)

)

f =

φ1,3

φ1,1

φ1,2+ · · ·+

φR,3

φR,1

φR,2

Sums of separable functions = continuous analogs of polyadic decompositions

4

Page 5: Spline-based separable expansions for approximation

Revisiting this problem: are there any benefits of using splines over polynomials?

Past work (e.g., Mohlenkamp & Beylkin) mostly considered polynomials toapproximate the component functions φr ,d(·),

why not use piece-wise polynomials a.k.a. splines?

5

Page 6: Spline-based separable expansions for approximation

What to expect next?

Spline basics and splines in higher dimensions: exploiting low-rank structures

Performing regression and classification

A Gauss–Newton algorithm exploiting sparsity

Numerical examples (regression)

Numerical examples (classification)

Key take-aways and future work

6

Page 7: Spline-based separable expansions for approximation

The knot set and B-spline basis terms

Let T = {ti}N+Mi=0 denote the set of knots:

a = t0 = . . . = tN−1 ≤ tN ≤ tN+1 ≤ . . . ≤ tM+1 = . . . = tM+N = b.

The B-spline basis terms {Bm,N}Mm=0 are defined through the recursion formula

Bm,N(x) :=x−tm

tm+N−tmBm,N−1(x) +

tm+N+1 − x

tm+N+1 − tm+1Bm+1,N−1(x),

where Bm,0(x) :=

{1 x ∈ [tm, tm+1)

0 otherwise.

7

Page 8: Spline-based separable expansions for approximation

The B-spline basis elements Bm,N(·) are compactly supported!

0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1

x

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

y

N=2, M=14

Bm,N(x) = 0, x ∈ (−∞, tm) ∪ [tm+N+1,∞).

8

Page 9: Spline-based separable expansions for approximation

The B-spline basis elements Bm,N(·) are compactly supported!

0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1

x

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

y

N=3, M=14

Bm,N(x) = 0, x ∈ (−∞, tm) ∪ [tm+N+1,∞).

8

Page 10: Spline-based separable expansions for approximation

The B-spline basis elements Bm,N(·) are compactly supported!

0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1

x

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

y

N=4, M=14

Bm,N(x) = 0, x ∈ (−∞, tm) ∪ [tm+N+1,∞).

8

Page 11: Spline-based separable expansions for approximation

The B-spline basis elements Bm,N(·) are compactly supported!

0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1

x

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

y

N=5, M=14

Bm,N(x) = 0, x ∈ (−∞, tm) ∪ [tm+N+1,∞).

8

Page 12: Spline-based separable expansions for approximation

The B-spline function

Any continuous function can be approximated arbitrarily well by

S(x) =[B0,N(x) · · · BM,N(x)

] c0...cM

= BT ,N(x)c .

through either increasing the knot density and order of the spline.

9

Page 13: Spline-based separable expansions for approximation

Taking direct tensor products of splines leads to exponential blow-up of coefficients...

f (x ; C) =

M1∑m1=0

· · ·MD∑

mD=0

cm1···mD

D∏d=1

B(d)

md ,N(d)(xd) = C ·1 Bd(x1) · · · ·D BD(xD)

C

∏Dd=1(Md + 1) parameters

10

Page 14: Spline-based separable expansions for approximation

Exploit low-rank structure: C(Γ1, . . . , Γd) = JΓ1, . . . , ΓDK, to alleviate this blow-up!

f (x ; Γ1, . . . , ΓD) = C(Γ1, . . . , ΓD) ·1 B1(x1) · · · ·D BD(xD) =R∑

r=1

D∏d=1

Bd(xd)γr ,d

C =

γ1,3

γ1,1

γ1,2

+ · · ·+

γR,3

γR,1

γR,2

∏Dd=1(Md + 1) parameters R(

∑Dd=1Md + 1) parameters

11

Page 15: Spline-based separable expansions for approximation

Spline basics and splines in higher dimensions: exploiting low-rank structures

Performing regression and classification

A Gauss–Newton algorithm exploiting sparsity

Numerical examples (regression)

Numerical examples (classification)

Key take-aways and future work

12

Page 16: Spline-based separable expansions for approximation

Regression is performed with the quadratic objective function

Given samples {(xi , yi )}Ii=1 ⊂ [0, 1]D × R from a underlying target functionf ∈ C ([0, 1]D), we minimize:

Q(Γ1, . . . , ΓD) :=1

2

I∑i=1

(f (xi ; Γ1, . . . , ΓD)− yi

)2.

13

Page 17: Spline-based separable expansions for approximation

A level-set approach to modeling a binary classification function

Binary classification function g : [0, 1]D → {0, 1} can be modeled by the function

g(x) =

{0 f (x) ≤ 0

1 f (x) > 0

(copyright wikimedia)

14

Page 18: Spline-based separable expansions for approximation

Replace step function with the logistic function σα : t 7→ 1/(exp(−αt) + 1)

Replace g withgα(x) := (σα ◦ f )(x) = σα(f (x)),

where α > 0 controls sharpness of transition.

15

Page 19: Spline-based separable expansions for approximation

Logistic objective function

gα is replaced by the approximant

gα(x ; Γ1, . . . , ΓD) := σα ◦ f (x ; Γ1, . . . , ΓD),

Given a collection of labeled data {(xi , yi )}Ii=1 ⊂ [0, 1]D × {0, 1}, the performance ofgα is optimized by maximizing the quantity

0 ≤∏yi=0

(1− gα(xi ; Γ1, . . . , ΓD))∏yi=1

gα(xi ; Γ1, . . . , ΓD) ≤ 1,

16

Page 20: Spline-based separable expansions for approximation

Logistic objective function

gα is replaced by the approximant

gα(x ; Γ1, . . . , ΓD) := σα ◦ f (x ; Γ1, . . . , ΓD),

Equivalent to minimizing the objective function

Lα(Γ1, . . . , ΓD) := −I∑

i=1

yi log gα(xi ; Γ1, . . . , ΓD) + (1− yi ) log (1− gα(xi ; Γ1, . . . , ΓD)) .

16

Page 21: Spline-based separable expansions for approximation

Spline basics and splines in higher dimensions: exploiting low-rank structures

Performing regression and classification

A Gauss–Newton algorithm exploiting sparsity

Numerical examples (regression)

Numerical examples (classification)

Key take-aways and future work

17

Page 22: Spline-based separable expansions for approximation

Minimization of objective functions is effectively done with Gauss-Newton doglegalgorithm

Exploit multi-linear structure of the objective functions, see:

Optimization-based algorithms for tensor decompositions: Canonical polyadicdecomposition, decomposition in rank-(Lr , Lr , 1) terms, and a new generalization.Sorber et al., SIAM J. Optim., 2013.Numerical optimization-based algorithms for data fusion. Vervliet et al., DataHandling in Science and Technology, 2019.

Main computational burden:

evaluating gradients and Grammian-vector products.

18

Page 23: Spline-based separable expansions for approximation

Benefit of compactly supported B-splines:significant speed-ups in Grammian and gradient by exploiting sparsity!

Gradient:

gr ,d = Ad

((D∗

k=1,k 6=dATkγr ,k

)∗η).

Grammian (of the Jacobian) vector product

wr ,d = Ad

( D∗k=1,k 6=d

ATkγr ,k

)∗ ξ∗

D∑d=1

R∑r=1

(D∗

k=1,k 6=dATkγ r ,k

)∗AT

dzr ,d

.

19

Page 24: Spline-based separable expansions for approximation

Benefit of compactly supported B-splines:significant speed-ups in Grammian and gradient by exploiting sparsity!

If the order of the B-spline is kept low:

O (DIMR) → O (DIR) flops19

Page 25: Spline-based separable expansions for approximation

Benefit of compactly supported B-splines:significant speed-ups in Grammian and gradient by exploiting sparsity!

20 40 60 80 100 120 140 160 180 200

M

0.005

0.01

0.015

0.02

0.025

0.03

0.035

0.04

Com

p. tim

e (

s)

Chebyshev basis

spline basis

average required computation time to pass through one cycle of the GN algorithm.N = 4, R = 3, I = 1000.

19

Page 26: Spline-based separable expansions for approximation

Spline basics and splines in higher dimensions: exploiting low-rank structures

Performing regression and classification

A Gauss–Newton algorithm exploiting sparsity

Numerical examples (regression)

Numerical examples (classification)

Key take-aways and future work

20

Page 27: Spline-based separable expansions for approximation

A R = 3 separable function

Consider the following example

f (x) = |x1||x2|︸ ︷︷ ︸non-smooth term

+ sin(2πx1) cos(2πx2) + x21x2, x ∈ [−1, 1]× [−1, 1].

21

Page 28: Spline-based separable expansions for approximation

As expected... an R = 3 is sufficient for a good approximation

R = 1

-1 -0.5 0 0.5 1

x1

-1

-0.8

-0.6

-0.4

-0.2

0

0.2

0.4

0.6

0.8

1

x2

0

0.005

0.01

0.015

0.02

0.025

0.03

0.035

0.04

0.045

0.05

(Knots are uniformly distributed on the approximation domain)22

Page 29: Spline-based separable expansions for approximation

As expected... an R = 3 is sufficient for a good approximation

R = 2

-1 -0.5 0 0.5 1

x1

-1

-0.8

-0.6

-0.4

-0.2

0

0.2

0.4

0.6

0.8

1

x2

0

0.005

0.01

0.015

0.02

0.025

0.03

0.035

0.04

0.045

0.05

(Knots are uniformly distributed on the approximation domain)22

Page 30: Spline-based separable expansions for approximation

As expected... an R = 3 is sufficient for a good approximation

R = 3

-1 -0.5 0 0.5 1

x1

-1

-0.8

-0.6

-0.4

-0.2

0

0.2

0.4

0.6

0.8

1

x2

0

0.005

0.01

0.015

0.02

0.025

0.03

0.035

0.04

0.045

0.05

(Knots are uniformly distributed on the approximation domain)22

Page 31: Spline-based separable expansions for approximation

As expected... an R = 3 is sufficient for a good approximation

R = 4

-1 -0.5 0 0.5 1

x1

-1

-0.8

-0.6

-0.4

-0.2

0

0.2

0.4

0.6

0.8

1

x2

0

0.005

0.01

0.015

0.02

0.025

0.03

0.035

0.04

0.045

0.05

(Knots are uniformly distributed on the approximation domain)22

Page 32: Spline-based separable expansions for approximation

Unlike for splines, Runge’s phenomenon can adversely affect quality of approximation

5 10 15 20 25 30 35

no. of basis terms M

10-3

10-2

10-1

100

RM

SE

B-spline, 400 samples

Chebyshev, 400 samples

B-spline, 800 samples

Chebyshev, 800 samples

No of separable terms R = 3.

(Knots are uniformly distributed on the approximation domain)23

Page 33: Spline-based separable expansions for approximation

Taming Runge’s phenomenon with splines: keep order low and increase knots

-1 -0.8 -0.6 -0.4 -0.2 0 0.2 0.4 0.6 0.8 1

x

0

0.2

0.4

0.6

0.8

1

1.2

y

A least-squares fit of |x| with 135 uniformly distributed samples

true

Chebyshev basis of degree 50

B-spline M= 50, N= 4 with uniform knot distribution

Runge’s phenomenon can adversely contribute to the overfitting problem

24

Page 34: Spline-based separable expansions for approximation

Low-rank structures in real life datasets - an example

NASA dataset from the UCI machine learning repository:

independent variables:

frequency,angle of attack,chord length,free-stream velocity,suction-side displacement thickness.

dependent variable: self-noise generated by airfoil.

randomly split data into a training (1202 samples) and a test (301 samples) sets.

25

Page 35: Spline-based separable expansions for approximation

An R = 5 separable function is sufficient to model the NASA dataset

1 2 3 4 5 6 7

R

0.05

0.06

0.07

0.08

0.09

0.1

0.11

0.12

0.13

RM

SE

(tr

ain

ing)

1 2 3 4 5 6 7

R

0.06

0.07

0.08

0.09

0.1

0.11

0.12

0.13

RM

SE

(te

st)

M=8

M=10

M=12

M=14

26

Page 36: Spline-based separable expansions for approximation

Spline basics and splines in higher dimensions: exploiting low-rank structures

Performing regression and classification

A Gauss–Newton algorithm exploiting sparsity

Numerical examples (regression)

Numerical examples (classification)

Key take-aways and future work

27

Page 37: Spline-based separable expansions for approximation

The separable rank can be increased to account for complexity of the classification sets

Consider the labeled dataset:

28

Page 38: Spline-based separable expansions for approximation

The separable rank can be increased to account for complexity of the classification sets

-1 -0.5 0 0.5 1

-1

-0.8

-0.6

-0.4

-0.2

0

0.2

0.4

0.6

0.8

1R = 1

28

Page 39: Spline-based separable expansions for approximation

The separable rank can be increased to account for complexity of the classification sets

-1 -0.5 0 0.5 1

-1

-0.8

-0.6

-0.4

-0.2

0

0.2

0.4

0.6

0.8

1R = 2

28

Page 40: Spline-based separable expansions for approximation

The separable rank can be increased to account for complexity of the classification sets

-1 -0.5 0 0.5 1

-1

-0.8

-0.6

-0.4

-0.2

0

0.2

0.4

0.6

0.8

1R = 3

28

Page 41: Spline-based separable expansions for approximation

The separable rank can be increased to account for complexity of the classification sets

-1 -0.5 0 0.5 1

-1

-0.8

-0.6

-0.4

-0.2

0

0.2

0.4

0.6

0.8

1R = 4

28

Page 42: Spline-based separable expansions for approximation

The separable rank can be increased to account for complexity of the classification sets

-1 -0.5 0 0.5 1

-1

-0.8

-0.6

-0.4

-0.2

0

0.2

0.4

0.6

0.8

1R = 5

28

Page 43: Spline-based separable expansions for approximation

Our method compared with well-established techniques for classification

102

103

104

no. of training samples

0

1

2

3

4

5

6

7

8

9

10

FiC

tra

inin

g (

%)

102

103

104

no. of training samples

0

5

10

15

FiC

te

st

(%)

CPD spline with R=7, M=16 SVM with RBF kernel SVM with order 9 polynomial kernel Patternnet with 50 nodes

29

Page 44: Spline-based separable expansions for approximation

CPU time for training grows more moderately with dataset size

102

103

104

no. of training samples

10-2

10-1

100

101

102

Exe

cu

tio

n t

ime

(s)

CPD spline with R=7, M=16

SVM with RBF kernel

SVM with order 9 polynomial kernel

Patternet with 50 nodes

30

Page 45: Spline-based separable expansions for approximation

Spline basics and splines in higher dimensions: exploiting low-rank structures

Performing regression and classification

A Gauss–Newton algorithm exploiting sparsity

Numerical examples (regression)

Numerical examples (classification)

Key take-aways and future work

31

Page 46: Spline-based separable expansions for approximation

Key take-aways and future work

Important take-aways:

With B-splines, sparsity can be exploited to further accelerate GN algorithm

Runge phenomenon effects are easily suppressed by keeping order of the spline low

Low-rank structures do appear in practice!

A new promising technique for (binary) classification

Future work:

Extend to other decompositions, e.g., Hierarchical Tucker, Tensor Train,

multi-class classification,

knot optimization

32

Page 47: Spline-based separable expansions for approximation

Spline-based separable expansions for

approximation, regression and classification

Nithin Govindarajan

Nico Vervliet, Lieven De Lathauwer

IPAM Workshop I: Tensor Methods and their Applications in the Physical and DataSciences, UCLA, United States, April 1, 2021