Fast Sparse Spectral Methods for Higher Dimensional PDEs Jie Shen - - PowerPoint PPT Presentation

fast sparse spectral methods for higher dimensional pdes
SMART_READER_LITE
LIVE PREVIEW

Fast Sparse Spectral Methods for Higher Dimensional PDEs Jie Shen - - PowerPoint PPT Presentation

Fast Sparse Spectral Methods for Higher Dimensional PDEs Jie Shen Purdue University Collaborators: Li-Lian Wang, Haijun Yu and Alexander Alekseenko Research supported by AFOSR and NSF ICERM workshop, June 3-7, 2013 Outlines Approximation by


slide-1
SLIDE 1

Fast Sparse Spectral Methods for Higher Dimensional PDEs

Jie Shen Purdue University

Collaborators: Li-Lian Wang, Haijun Yu and Alexander Alekseenko

Research supported by AFOSR and NSF

ICERM workshop, June 3-7, 2013

slide-2
SLIDE 2

Outlines

Approximation by hyperbolic cross Sparse grids for bounded domains Sparse grids for unbounded domains Application to electronic Schr¨

  • dinger equation

Adaptive sparse tensor products with multi-wavelets for the Boltzmann collision operator Concluding remarks

slide-3
SLIDE 3

Motivations

  • Electronic Schr¨
  • dinger equation: find the eigenvalues and

eigenfunctions of the Hamilton operator H = −1 2

N

  • i=1

∆i −

N

  • i=1

K

  • ν=1

Zν |xi − aν| +

N

  • i=1

N

  • j>i

1 |xi − xj|, where x1, · · · , xN ∈ R3 are coordinates of N electrons.

  • Boltzmann Equation:

∂f ∂t + p m · ∇xf = G(f , f ) where x, p are position and momentum of particle in R3, m is the particle mass and f is the density distribution function.

slide-4
SLIDE 4
  • Fokker-Planck Navier-Stokes equations for FENE dumbbell

model:                ∂tu + (u · ∇x)u + ∇xp = ν Re∆xu + 1 − ν Re De∇x · τp, ∇x · u = 0, ∂tf + ∇x · (uf ) + ∇q · (∇xu · qf ) = 1 De∇q · (F(q)f ) + 1 De∆qf , τp(x, t) =

  • W

(F(q) ⊗ q) f (x, q, t)dq. x ∈ Ω ⊂ R3, Ω is the fluid space; q ∈ W ⊂ R3, W is the configuration space; F(q) is the configuration force.

slide-5
SLIDE 5

A model problem

Consider the model elliptic equation: αu − ∆u = f in Ω = (−1, 1)d; u|∂Ω = 0. Given an approximation space Vn and an interpolation operator In, the (weighted) spectral-Galerkin method is to find un ∈ Vn such that (∇un, ∇(vnω)) = (Inf , vn)ω, ∀vn ∈ Vn.

slide-6
SLIDE 6

A model problem

Consider the model elliptic equation: αu − ∆u = f in Ω = (−1, 1)d; u|∂Ω = 0. Given an approximation space Vn and an interpolation operator In, the (weighted) spectral-Galerkin method is to find un ∈ Vn such that (∇un, ∇(vnω)) = (Inf , vn)ω, ∀vn ∈ Vn. Error estimate: ∇(u − un)L2

ω inf

vn∈Vn ∇(u − vn)L2

ω + f − Inf L2 ω.

approximation error interpolation error

slide-7
SLIDE 7

Usual spectral method:

Let Vn = Pn+1 ∩ H1

0(Ω) where Pn+1 is the space of polynomials of

degree ≤ n + 1 in EACH variable. N = dim(Vn) = nd; infvn∈Vn ∇(u − vn)L2

ω n1−suHs ω N(1−s)/duHs ω.

The # of degree of freedom increases exponentially fast with d — the convergence rate deteriorates rapidly as d increases. This is the so called Curse of dimensionality!

slide-8
SLIDE 8

Hyperbolic cross (Korobov, Babenko ’57)

The hyperbolic cross is defined as K(n) =

  • k ∈ Zd

+ : d

  • i=1

max(ki, 1) ≤ n

  • .

The cardinality of K(n) is of order n(log n)d−1.

slide-9
SLIDE 9

Hyperbolic cross (Korobov, Babenko ’57)

The hyperbolic cross is defined as K(n) =

  • k ∈ Zd

+ : d

  • i=1

max(ki, 1) ≤ n

  • .

The cardinality of K(n) is of order n(log n)d−1. Set Vn = span{

d

  • i=1

φki(xi) :

d

  • i=1

max(ki, 1) ≤ n}, in what functional spaces does Vn provide good approximations?

slide-10
SLIDE 10

Table : Comparison of grids and hyperbolic cross n d=2 d=3 d=4 Grids |K(n)|

|K(n)| nlog(n)

Grids |K(n)|

|K(n)| nlog(n)2

Grids |K(n)| 9 100 42 2.1 1000 141 3.2 10000 424 13 196 64 1.9 2744 228 2.7 38416 720 17 324 87 1.8 5832 321 2.4 104976 1041 21 484 113 1.8 10648 435 2.2 234256 1457 25 676 138 1.7 17576 546 2.1 456976 1877 29 900 162 1.7 27000 646 2.0 810000 2229 33 1156 190 1.6 39304 778 1.9 1336336 2745 37 1444 217 1.6 54872 904 1.9 2085136 3239 41 1764 243 1.6 74088 1021 1.8 3111696 3683 45 2116 273 1.6 97336 1165 1.8 4477456 4243

slide-11
SLIDE 11

Korobov spaces for periodic functions

Given f ∈ L2

p(0, 2π)d and its Fourier series

f (x) =

  • k∈Zd

ˆ f (k)eik·x with ˆ f (k) = 1 |Ω|

f (x)e−ik·xdx. Let us define r(k) := d

j=1 max(|kj|, 1),

k = (k1, . . . , kd), then, the Korobov space of order α ∈ R is K α

p :=

  • f ∈ L2

p(Ω) :

  • k∈Zd

|ˆ f (k)|2r(k)2α < ∞

  • =
  • ∂q1+...+qd

∂xq1

1 . . . ∂xqd d

f ∈ L2

p(Ω) : 0 ≤ qj ≤ α, 1 ≤ j ≤ d

  • .
slide-12
SLIDE 12

Setting Xn :=

  • f : f (x) =

r(k)≤n ˆ

f (k)eik·x . Then, one can prove inf

vn∈Xn v − vnK r

p nr−svK s p.

(e.g. Griebel & Hamaekers 2007)

slide-13
SLIDE 13

Setting Xn :=

  • f : f (x) =

r(k)≤n ˆ

f (k)eik·x . Then, one can prove inf

vn∈Xn v − vnK r

p nr−svK s p.

(e.g. Griebel & Hamaekers 2007) Remarks: The estimate still depends on d since N = O(n(log n)d−1). It is also possible to remove the dependence on d by using the

  • ptimized (energy based) hyperbolic cross space (Bungartz &

Griebel ’04): Kσ(n) =

  • k ∈ Zd

+ : d

  • i=1

max(ki, 1)| |k|−σ

∞ ≤ n1−σ

  • .

Then, for σ ∈ (0, 1), Card(Kσ(n)) = O(n). In this case, the estimate holds only for σ ≤ r

s .

slide-14
SLIDE 14

Setting Xn :=

  • f : f (x) =

r(k)≤n ˆ

f (k)eik·x . Then, one can prove inf

vn∈Xn v − vnK r

p nr−svK s p.

(e.g. Griebel & Hamaekers 2007) Remarks: The estimate still depends on d since N = O(n(log n)d−1). It is also possible to remove the dependence on d by using the

  • ptimized (energy based) hyperbolic cross space (Bungartz &

Griebel ’04): Kσ(n) =

  • k ∈ Zd

+ : d

  • i=1

max(ki, 1)| |k|−σ

∞ ≤ n1−σ

  • .

Then, for σ ∈ (0, 1), Card(Kσ(n)) = O(n). In this case, the estimate holds only for σ ≤ r

s .

These estimates were extended to the non-periodic case and unbounded domains in S. & Wang (SINUM ’10).

slide-15
SLIDE 15

Korobov type spaces for non-periodic functions

Let ωα,β = (1 − x)α(1 + x)β and J(α,β)

n

be the Jacobi weight and Jacobi polynomial with index α, β > −1. We define the multi-dimensional Jacobi weights and Jacobi polynomials by ω ¯

α,¯ β(¯

x) = Πd

i=1ωαi,βi(xi), Φ¯ α,¯ β ¯ k

(¯ x) = Πd

i=1Jαi,βi ki

(xi). We then define (for any integer t ≥ 0) K t

¯ α,¯ β := {u : u2 K t

¯ α, ¯ β :=

r|∞≤t

∂¯

ru2 ω ¯

α+¯ r, ¯ β+¯ r < ∞}.

slide-16
SLIDE 16

Projection errors in the multi-D case

Denote P ¯

α,¯ β n

= span{Φ¯

α,¯ β ¯ k

(¯ x) : Πd

i=1 max(ki, 1) ≤ n}, and define

Π¯

α,¯ β n

: L2

ω ¯

α, ¯ β → P ¯

α,¯ β n

be the orthogonal projector defined by (u − Π¯

α,¯ β n

u, vK)ω ¯

α, ¯ β = 0, ∀vn ∈ P ¯

α,¯ β n

.

  • Theorem. (S. & Wang ’10) For u ∈ K s

¯ α,¯ β and 0 ≤ t < s, we have

u − Π¯

α,¯ β n

uK t

¯ α, ¯ β nt−s|u|K s ¯ α, ¯ β,

where |u|2

K s

¯ α, ¯ β :=

r|∞=s

∂¯

ru2 ω ¯

α+¯ r, ¯ β+¯ r .

slide-17
SLIDE 17

Projection errors in the multi-D case

Denote P ¯

α,¯ β n

= span{Φ¯

α,¯ β ¯ k

(¯ x) : Πd

i=1 max(ki, 1) ≤ n}, and define

Π¯

α,¯ β n

: L2

ω ¯

α, ¯ β → P ¯

α,¯ β n

be the orthogonal projector defined by (u − Π¯

α,¯ β n

u, vK)ω ¯

α, ¯ β = 0, ∀vn ∈ P ¯

α,¯ β n

.

  • Theorem. (S. & Wang ’10) For u ∈ K s

¯ α,¯ β and 0 ≤ t < s, we have

u − Π¯

α,¯ β n

uK t

¯ α, ¯ β nt−s|u|K s ¯ α, ¯ β,

where |u|2

K s

¯ α, ¯ β :=

r|∞=s

∂¯

ru2 ω ¯

α+¯ r, ¯ β+¯ r .

Approximation results in usual H1-norm and for elliptic equations are also established in S. & Wang ’10.

slide-18
SLIDE 18

200 400 600 800 1000 200 400 600 800 1000

  • 40
  • 35
  • 30
  • 25
  • 20
  • 15
  • 10
  • 5

Figure : Spectrum of a highly anisotropic exact solution

slide-19
SLIDE 19

Figure : Convergence history: comparison with the full grid

slide-20
SLIDE 20

10 20 30 40 50 60 10 20 30 40 50 60

  • 40
  • 35
  • 30
  • 25
  • 20
  • 15
  • 10
  • 5

Figure : Spectrum of an isotropic exact solution

slide-21
SLIDE 21

1e-16 1e-14 1e-12 1e-10 1e-08 1e-06 0.0001 0.01 1 10 100 1000 10000 100000 1e+06

L2 error # of grid points L2 error of solution

d=2 full d=3 full d=4 full d=5 full d=1 d=2 sparse d=3 sparse d=4 sparse d=5 sparse

Figure : Convergence history: comparison with the full grid

slide-22
SLIDE 22

Sparse grids

Sparse grid construction (Smolyak ’63): Start from a sequence of (preferably nested) univariate quadrature/interpolation formulas: Qif :=

ni

  • j=1

f (xi

j )ωi j,

∆i = Qi − Qi−1 with Q0 = 0. Define a multidimensional quadrature rule by: Q(d)

l

f :=

  • d≤|i|≤l

(∆i1 ⊗ · · · ⊗ ∆id)f .

slide-23
SLIDE 23

−1 −0.8 −0.6 −0.4 −0.2 0.2 0.4 0.6 0.8 1 −1 −0.8 −0.6 −0.4 −0.2 0.2 0.4 0.6 0.8 1

Figure : Q(2)

5

based on Chebyshev-Gauss-Lobatto quadrature.

slide-24
SLIDE 24

Sparse “grids” in frequency space

Similarly, we can define sparse “grids” in frequency space. For example, by splitting the 1-D basis {P0(x), P1(x), P2(x), ....} into ∆1 = {P0, P1, P2}, ∆2 = {P3, P4}, ∆3 = {P5, P6, P7, P8}, · · · we can define sparse “grids” in frequency space by H(d)

l

:=

  • d≤|i|≤l

(∆i1 ⊗ · · · ⊗ ∆id).

slide-25
SLIDE 25

2 4 6 8 10 12 14 16 2 4 6 8 10 12 14 16

Figure : H(2)

5

in frequency space.

slide-26
SLIDE 26

Optimized hyperbolic cross/sparse grid

  • 1
  • 0.5

0.5 1

  • 1
  • 0.5

0.5 1 5 10 15 20 25 30 35 5 10 15 20 25 30 35

slide-27
SLIDE 27

Hierarchical basis and interpolation on sparse grid

The key ingredient for constructing a fast transform between the space grids in physical and frequency spaces is the Hierarchical basis.

slide-28
SLIDE 28

Hierarchical basis and interpolation on sparse grid

The key ingredient for constructing a fast transform between the space grids in physical and frequency spaces is the Hierarchical basis. Given a sparse grid generated by nested 1-D quadratures with the index set Ii = {0, 1, · · · , ni − 1.} A set of function {φk} form a hierarchical bases if φk(xi

j ) = 0, ∀j

if k ≥ ni. Then, we can rearrange {xi

j } as {xj} and define an interpolation

  • perator by

f (xj) =

  • k∈I i

bkφk(xj), for any j ∈ I i, i = 1, 2, . . . where {bk} do not depend on the level index i.

slide-29
SLIDE 29

Figure : Left: The 1-d linear hierarchical finite element bases; Right: The 2-d construction of sparse grid (from Bungartz and Griebel 2004)

slide-30
SLIDE 30
  • 2
  • 1.5
  • 1
  • 0.5

0.5 1 1.5 2

  • 1
  • 0.924
  • 0.707
  • 0.383

0.383 0.707 0.924 1

  • 2
  • 1.5
  • 1
  • 0.5

0.5 1 1.5 2

  • 1
  • 0.924
  • 0.707
  • 0.383

0.383 0.707 0.924 1

  • 2
  • 1.5
  • 1
  • 0.5

0.5 1 1.5 2

  • 1
  • 0.924
  • 0.707
  • 0.383

0.383 0.707 0.924 1

  • 2
  • 1.5
  • 1
  • 0.5

0.5 1

  • 1
  • 0.924
  • 0.707
  • 0.383

0.383 0.707 0.924 1

Figure : Hierarchical basis functions with Chebyshev-Gauss-Lobatto quadrature: φk = Tk, for k ∈ I 1; φk = Tk − T2i−k, for k ∈ ˜ I i = I i\I i−1, i > 1.

slide-31
SLIDE 31

Multi-dimensional case:

Define the multi-D index set corresponding to the sparse grid I (d)

l

:=

  • d≤|i|≤l

(˜ Ii1 ⊗ · · · ⊗ ˜ Iid), and the multi-dimensional hierarchical bases by φk(x) = Πd

i=1φki(xi).

Then, we can define the interpolation operator U(d)

l

  • n the sparse

grid by (U(d)

l

f )(xj) =

  • k∈I (d)

l

bkφk(xj), ∀j ∈ I (d)

l

. The transform between the values at the sparse grid and the coefficients of the hierarchical basis can be performed with quasi-optimal computational complexity!

slide-32
SLIDE 32

Implementation of hyperbolic-cross/sparse-grid solver

αu − ∆u = f , x ∈ Ω = (−1, 1)d, u|∂Ω = 0.

  • Sparse grid in frequency space:

Iq

d =

  • (k1, k2, . . . , kd) : 0 ks < 2is − 1, is 0, |i|1 = q
  • ,

V q

d =

  • ˜

φk, k ∈ Iq

d

  • ;
  • Sparse grid in physical space:

J q

d =

  • (k1, k2, . . . , kd) : 0 ks < 2is + 1, is 0, |i|1 = q
  • ,

X q

d =

  • xj, j ∈ J q

d

  • ;
  • Uq

d the interpolation on X q d .

  • Find uq

d ∈ V q d such that

α(uq

d, v)ω − (∆uq d, v)ω = (Uq d f , v)ω,

∀v ∈ V q

d .

slide-33
SLIDE 33

Choice of basis functions for V q

d

Remark: While the hierarchical basis plays a critical role in fast transform, it lacks good orthogonality property.

slide-34
SLIDE 34

Choice of basis functions for V q

d

Remark: While the hierarchical basis plays a critical role in fast transform, it lacks good orthogonality property.

  • Chebyshev-Galerkin method:

φk(x) = Tk(x) − Tk+2(x), ω(x) = (1 − x2)−1/2 1-d mass matrix: banded; 1-d stiff matrix: triangular; the multi-D stiffness matrix is not sparse.

slide-35
SLIDE 35

Choice of basis functions for V q

d

Remark: While the hierarchical basis plays a critical role in fast transform, it lacks good orthogonality property.

  • Chebyshev-Galerkin method:

φk(x) = Tk(x) − Tk+2(x), ω(x) = (1 − x2)−1/2 1-d mass matrix: banded; 1-d stiff matrix: triangular; the multi-D stiffness matrix is not sparse.

  • Legendre-Galerkin method:

φk(x) = Lk(x) − Lk+2(x), ω(x) = 1 1-d mass matrix: banded; 1-d stiff matrix: diagonal; Legendre-Gauss type points not nested; lack of fast transform.

slide-36
SLIDE 36

Choice of basis functions for V q

d

Remark: While the hierarchical basis plays a critical role in fast transform, it lacks good orthogonality property.

  • Chebyshev-Galerkin method:

φk(x) = Tk(x) − Tk+2(x), ω(x) = (1 − x2)−1/2 1-d mass matrix: banded; 1-d stiff matrix: triangular; the multi-D stiffness matrix is not sparse.

  • Legendre-Galerkin method:

φk(x) = Lk(x) − Lk+2(x), ω(x) = 1 1-d mass matrix: banded; 1-d stiff matrix: diagonal; Legendre-Gauss type points not nested; lack of fast transform.

  • Chebyshev-Legendre-Galerkin method (Don & Gottlieb ’94, S.

’96)

slide-37
SLIDE 37

Linear algebraic system

Denote φk(x) = φk1(x1) · · · φkd(xd), then V q

d =

  • φk, k ∈ Iq

d

  • .

The Galerkin method leads to (αM + S)u = f, where M = (mk1,j1 · · · mkd,jd)k,j∈Iq

d ,

S = (sk1,j1mk2,j2 · · · mkd,jd)k,j∈Iq

d + · · · + (mk1,j1 · · · mk2,j2skd,jd)k,j∈Iq d ,

mk,j = (φk, φj)ω, sk,j = (−∆φk, φj)ω f = (fk)k∈Iq

d ,

fk = (Uq

d f , φk).

slide-38
SLIDE 38

(a) (b) (c) (d)

Figure : Sparsity of Legendre-Galerkin method: (a) stiff X 8

2 ; (b) system

X 8

2 ; (c) stiff X 9 3 ; (d) system X 9 3 .

slide-39
SLIDE 39

d q n nnz0 nnz1 cond1 pcond1 2 4 17 45 49 5.008 2.035 5 49 157 193 13.248 2.819 6 129 461 625 55.992 5.941 7 321 1229 1777 156.295 8.670 8 769 3085 4657 746.584 20.689 9 1793 7437 11569 2125.186 30.454 4 5 9 29 29 1.938 1.836 6 49 281 281 5.510 3.795 7 209 1809 1809 33.153 7.297 8 769 9009 9025 212.514 14.519 9 2561 37601 37873 995.447 30.715

Table : nnz0: number of non-zeros of stiff matrix; nnz1: number of non-zeros of system matrix (α = 1); cond1: the condition number of system matrix (α = 1); pcond1: the condition number of system matrix (α = 1) with diagonal pre-conditioner.

slide-40
SLIDE 40

PCG with fast matrix-vector product algorithm

For the Chebyshev sparse Galerkin method: The evaluation of f can be performed using the fast transform algorithm. The stiff matrix S is not sparse but the matrix-vector product Mu, Su can be performed with quasi-optimal computational complexity. Solve the linear system by a preconditioned CG type method — How to construct an optimal pre-conditioner is still an

  • pen problem.
slide-41
SLIDE 41

PCG with fast matrix-vector product algorithm

For the Chebyshev sparse Galerkin method: The evaluation of f can be performed using the fast transform algorithm. The stiff matrix S is not sparse but the matrix-vector product Mu, Su can be performed with quasi-optimal computational complexity. Solve the linear system by a preconditioned CG type method — How to construct an optimal pre-conditioner is still an

  • pen problem.

Not optimal computational complexity, but memory requirement is small since it does not need explicit formation of stiffness/mass matrices.

slide-42
SLIDE 42

Direct sparse solvers

In order to use an efficient sparse solver, we need to form explicitly the system matrix — only possible with Chebyshev-Legendre sparse grid method The evaluation of f is done by the fast Chebyshev-Legendre transform (by FMM, Alpert & Rokhlin ’91). The sparse mass and stiff matrices can be assembled with a fast algorithm. Solve the linear system by using an efficient sparse solver

UMFPack (T. Davis) — competitive with PCG in some cases AMG (Y. Notay) — competitive with PCG in some cases SuperMF (J. Xia et al.) — can lead to quasi-optimal computational complexity.

slide-43
SLIDE 43

Direct sparse solvers

In order to use an efficient sparse solver, we need to form explicitly the system matrix — only possible with Chebyshev-Legendre sparse grid method The evaluation of f is done by the fast Chebyshev-Legendre transform (by FMM, Alpert & Rokhlin ’91). The sparse mass and stiff matrices can be assembled with a fast algorithm. Solve the linear system by using an efficient sparse solver

UMFPack (T. Davis) — competitive with PCG in some cases AMG (Y. Notay) — competitive with PCG in some cases SuperMF (J. Xia et al.) — can lead to quasi-optimal computational complexity.

Quasi-optimal computational complexity, but memory requirement is large.

slide-44
SLIDE 44

Sparse grid by Hermite Gauss quadrature

“The natural choice”: Construct Sparse grids from the Hermite-Gauss points and Laguerre-Gauss points. The problem: The Hermite-Gauss and Laguerre-Gauss points are not nested — leads to too many points in the sparse grid so it is not computationally effective.

−5 −4 −3 −2 −1 1 2 3 4 5 −5 −4 −3 −2 −1 1 2 3 4 5 2 4 6 8 10 12 14 16 2 4 6 8 10 12 14 16

Figure : Left: Hermite sparse grid X 5

2 ; Right: the corresponding index

set I5 in the frequency space.

slide-45
SLIDE 45

Mapped Chebyshev method

  • Given a mapping x = x(ξ) : (−1, 1) → R and its inverse g

ξ = ξ(x) : R → (−1, 1). The mapped Chebyshev functions ˆ Tk(x) = Tk(ξ(x))µ(ξ(x)), µ(ξ) =

  • ω(ξ)/x′(ξ) with ω(ξ) = 1/
  • 1 − ξ2

which satisfies ( ˆ Tk, ˆ Tj) =

  • R

ˆ Tk(x) ˆ Tj(x)dx = 1

−1

Tk(ξ)Tj(ξ)ω(ξ)dξ = δkj.

  • The convergence rate will depend on the rate of decay at infinity
  • A class of rational mapping: (−1, 1) → R:

x′(ξ) = L (1 − ξ2)1+r/2 , r 0,

slide-46
SLIDE 46

Two most useful cases: r = 0, 1: x(ξ) = L

2 log 1+ξ 1−ξ,

r = 0

1−ξ2 ,

r = 1 ; ξ(x) =

  • tanh( 1

Lx)

r = 0

x √ x2+L2 ,

r = 1 .

−2.5 −2 −1.5 −1 −0.5 0.5 1 1.5 2 2.5 −2.5 −2 −1.5 −1 −0.5 0.5 1 1.5 2 2.5 −5 −4 −3 −2 −1 1 2 3 4 5 −5 −4 −3 −2 −1 1 2 3 4 5

Figure : Mapped Chebyshev sparse grids. Left: X 5

2 with r = 0; Right:

X 5

2 with r = 1.

slide-47
SLIDE 47

Let us define our hyperbolic approximation space by Xd

N := span

  • Tk : k ∈ ΥH

N

  • ,

with ΥH

N =

  • k ∈ Nd

0 : 1 ≤ |k|mix := d

  • j=1

max{1, kj} ≤ N

  • .

We define the orthogonal projection πd

N : L2(Rd) → Xd N, and the

mapped derivative Dxu := a(x)dˆ u dx with a(x) = dx dξ , ˆ u(x) = u(x) µ(ξ(x)), Dk

xu = Dk1 x1 · · · Dkd xd u,

̟(1+r)/2+k =

d

  • j=1

(1 − ξ2

j )(1+r)/2+kj.

slide-48
SLIDE 48

Approximation results

  • Theorem. For r ≥ 0 and any u ∈ Km(Rd), we have

πd

Nu − uL2(Rd) ≤ CN−m|u|Km(Rd),

m ≥ 0, and ∇(πd

Nu − u)L2(Rd) ≤ CN1−m|u|Km(Rd),

m ≥ 1. where Km(Rd) is the Korobov-type space Km(Rd) :=

  • u : Dk

xu ∈ L2 ̟(1+r)/2+k(Rd), 0 ≤ |k|∞ ≤ m

  • ,

∀m ∈ N0, and |u|Km(Rd) =

  • |k|∞=m
  • Dk

xu

  • L2

̟(1+r)/2+k(Rd)

1

2 .

slide-49
SLIDE 49

Mass matrix M = I; Stiff matrix S: banded — leads to system matrix with much less non-zero entries. S = S ⊗ M ⊗ · · · ⊗ M + M ⊗ S ⊗ M · · · ⊗ M + M ⊗ · · · ⊗ M ⊗ S Fast transform available on mapped Chebyshev sparse grids.

Figure : Sparsity of the mapped Chebyshev method: Left: d = 2, q = 8; Right: d = 4, q = 8

slide-50
SLIDE 50

The linear system for Poisson type equations can be solved efficiently by using the sparse fast transform and PCG or a sparse solver such as CHOLMOD.

2 4 6 8 10 12 14 16 2 4 6 8 10 12 14 16

10 10

1

10

2

10

3

10

4

10

5

10

6

10

−6

10

−5

10

−4

10

−3

10

−2

10

−1

10 10

1

number of grid points L2 error d=1 d=2 full d=3 full d=4 full d=2 sparse d=3 sparse d=4 sparse

Figure : Comparison between the MCFG method and MCSG method (with r = 1) for solving Poisson-type equation with an anisotropic solution with algebraic delay.

slide-51
SLIDE 51

10 10

1

10

2

10

3

10

4

10

5

10

6

10

−15

10

−10

10

−5

10 number of grid points L2 error d=1 d=2 full d=3 full d=2 sparse d=3 sparse

Figure : Left: the optimal frequency index set for tensor product of 1-D function with geometric convergence; Right: Comparison between the MCFG method and MCSG method (with r = 1).

slide-52
SLIDE 52

10 10

1

10

2

10

3

10

4

10

5

10

6

10

−15

10

−10

10

−5

10 number of grid points L2 error d=1 d=2 full d=3 full d=4 full d=2 sparse d=3 sparse d=4 sparse

Figure : Left: the optimal frequency index set for tensor product of 1-D function with subgeometric convergence; Right: Comparison between the MCFG method and MCSG method (with r = 1).

slide-53
SLIDE 53

Problems with variable coefficients

  • Use a constant coefficient problem as preconditioner
  • While the system matrix is full, but the matrix-vector product

can be performed in quasi-optimal complexity

  • Ex. 1: r = 0
  • Ex. 1: r = 1
  • Ex. 2: r = 0

d q iter# CPU iter# CPU iter# CPU 1 7 13 3.000(-3) 13 4.000(-3) 6 1.000(-3) 8 13 7.000(-3) 13 6.000(-3) 5 2.000(-3) ) 9 15 1.000(-2) 13 1.000(-2) 5 4.000(-3) 2 8 12 4.400(-2) 13 5.200(-2) 6 2.600(-2) 9 13 1.040(-1) 14 1.150(-1) 5 4.200(-2) 10 13 2.270(-1) 14 2.410(-1) 5 8.700(-2) 3 9 14 6.560(-1) 13 7.780(-1) 6 2.920(-1) 10 14 1.510(-0) 13 1.911(-0) 6 6.830(-1) 11 14 4.071(-0) 13 4.749(-0) 6 1.857(-0)

Table : The numbers of iterations and CPU time of using BICGSTAB solving equations with non-constant coefficients.

slide-54
SLIDE 54

Application to electronic Schr¨

  • dinger equation

Much effort have been devoted to this problem using density function theory, but not much work on solving the equation directly; It has been shown (Yserentant ’04) that its solution lives in the Korobov space; The problem is high-dimensional and set in unbounded domains; previous work (Gribel & Hamaekers ’07) was based

  • n domain truncation and Fourier method — the effect of

domain truncation is not easy to quantify; Our fast sparse spectral method in unbounded domain is well suited for this problem.

slide-55
SLIDE 55

Consider the 1-D electronic Schr¨

  • dinger equation:

Hψ = λψ with H = T + V := −1 2

N

  • i=1

∆i +

  • N
  • i=1

Z|xi| −

N

  • i=1

N

  • j>i

|xi − xj|

  • .

We look for the ground state energy which is the smallest eigenvalue of H.

slide-56
SLIDE 56

Consider the 1-D electronic Schr¨

  • dinger equation:

Hψ = λψ with H = T + V := −1 2

N

  • i=1

∆i +

  • N
  • i=1

Z|xi| −

N

  • i=1

N

  • j>i

|xi − xj|

  • .

We look for the ground state energy which is the smallest eigenvalue of H. Sparse spectral-Galerkin approximation: Find ψ ∈ V q

N, λ ∈ R

such that (Tψ, φ)+ < V ψ, φ >disc= λ(ψ, φ), ∀φ ∈ V q

N,

where V q

N consists of mapped Chebyshev polynomials.

slide-57
SLIDE 57

Using the mapped Chebyshev polynomials, the above reduces to: (S + V)E = λME where the mass matrix M is diagonal, the stiffness matrix S is sparse, the matrix V is full, but the matrix-vector product V¯ x can be performed in quasi-optimal complexity thanks to the fast Chebyshev transform.

slide-58
SLIDE 58

How to find the smallest eigenvalue?

The direct application of Arnoldi method for computing the smallest eigenvalue converges very slowly; To accelerate the Arnoldi method, we employed a shifted-inverse technique which requires solving (S + V − δI)¯ x = ¯ f ; We used S − δI to precondition the above system. The above approach appears to be very efficient and robust. For example, six-dimensional problem with q = 12 can be solved on a laptop using single core within 3 hours.

slide-59
SLIDE 59

q dof L = 0.5 L = 0.6 L = 0.75 L = 1 6 257 2.0139266 1.9089328 1.8596860 1.8463230 7 577 1.9411353 1.8801002 1.8579695 1.8511776 8 1281 1.9053250 1.8713405 1.8624749 1.8607346 9 2817 1.8838480 1.8657901 1.8622006 1.8610317 10 6145 1.8734093 1.8646394 1.8634796 1.8632008 11 13313 1.8678381 1.8638844 1.8634728 1.8632195 12 28673 1.8655343 1.8639148 1.8638087 1.8637460 13 61441 1.8644735 1.8638533 1.8638094 1.8637472 14 106497 1.8640681 1.8639032 1.8638934 1.8638778 15 188417 1.8639412 1.8639003 1.8638934 1.8638778 16 335873 1.8639259 1.8639160 1.8639144 1.8639104

Table : Two electrons: the smallest eigenvalue should lie in [1.8639144,1.8639259] with a reletive error of 3.1 × 10−6.

slide-60
SLIDE 60

q dof Memory L = 0.5 L = 0.6 L = 0.75 L = 1 6 729 14.19322 10.62023 7.14268 4.13509 7 3645 1M 11.53182 8.98147 6.56414 4.57501 8 14337 7M 9.84225 8.20117 6.84163 6.28908 9 49761 30M 9.13927 8.08522 7.40965 7.55470 10 159489 158M 8.81094 8.15501 7.88803 8.19936 11 483201 879M 8.65394 8.26734 8.21052 8.36132 12 1403137 5163M 8.57456 8.35870 8.37154 8.44094 13 3940609 16689M 8.52448 8.42065 8.44336 8.46657

Table : Six electrons: the smallest eigenvalue should lie in [8.46657, 8.52448] with a reletive error of 3.4 × 10−3.

slide-61
SLIDE 61

q dof memory L = 0.5 L = 0.6 L = 0.75 L = 1 8 6561 2M 19.011 14.176 9.525 5.514 9 41553 30M 16.487 12.652 9.076 6.059 10 193185 312M 14.387 11.764 9.591 8.622 11 768609 1621M 13.428 11.723 10.682 10.210 12 2772225 16587M 13.054 11.995 11.601 11.339

Table : Eight electrons: the smallest eigenvalue should lie in [11.339, 13.054] with a relative error of 7.6 × 10−2.

slide-62
SLIDE 62

N q DoF δ n1 n2 N q DoF δ n1 n2 1 4 4 17 0.79 7 10 7 2769 3.84 9 41 5 33 0.80 7 12 8 7681 4.27 9 44 6 65 0.80 7 13 9 20481 4.58 7 51 7 129 0.80 7 14 10 52933 4.64 7 56 8 257 0.80 7 15 11 133889 4.70 5 69 9 513 0.80 7 16 12 331777 4.70 5 74 10 1025 0.80 7 17 13 808961 4.70 5 87 2 6 5 113 1.72 9 31 7 3645 4.13 9 37 6 257 1.79 9 31 8 14337 4.57 21 49 7 577 1.84 7 33 9 49761 6.28 23 80 8 1281 1.84 7 36 10 159489 7.55 13 110 9 2817 1.84 7 36 11 483201 8.19 9 128 10 6145 1.84 9 37 12 1403137 8.36 7 141 11 13313 1.84 9 37 13 3940609 8.42 7 148 Table : Numbers of iterations used in ARPack and BICGSTAB.

slide-63
SLIDE 63

Adaptive sparse tensor products using multi-wavelets

The one domain approach is not suitable for problems with local features, such as those in many challenging applications. For this type of problems, an efficient adaptive procedure, e.g., wavelets, is needed.

slide-64
SLIDE 64

Adaptive sparse tensor products using multi-wavelets

The one domain approach is not suitable for problems with local features, such as those in many challenging applications. For this type of problems, an efficient adaptive procedure, e.g., wavelets, is needed. Standard wavelets are generated by a single function: the

  • rder of accuracy is low and fixed.
slide-65
SLIDE 65

Adaptive sparse tensor products using multi-wavelets

The one domain approach is not suitable for problems with local features, such as those in many challenging applications. For this type of problems, an efficient adaptive procedure, e.g., wavelets, is needed. Standard wavelets are generated by a single function: the

  • rder of accuracy is low and fixed.

We propose to use the Legendre multi-wavelets (Alpert ’93), which allow spectral accuracy, and we propose to use a sparse tensor product with these multi-wavelets so that it is feasible for higher-dimensional problems.

slide-66
SLIDE 66

Legendre multi-wavelets in L2(−1, 1)

Let φj(x) be the Legendre polynomial of degree j and V k

0 = span{φj(x)} : 0 ≤ j ≤ k − 1}. For n ≥ 1, we define V k n as

the space of piece-wise polynomials of degree less than k on a regular mesh with h = 21−n. Then, a set of basis function for V k

n

is given by φn

jl(x) = 2

n 2 φj(2nx − l),

j = 0, . . . , k − 1, l = 0, . . . , 2n − 1.

slide-67
SLIDE 67

Legendre multi-wavelets in L2(−1, 1)

Let φj(x) be the Legendre polynomial of degree j and V k

0 = span{φj(x)} : 0 ≤ j ≤ k − 1}. For n ≥ 1, we define V k n as

the space of piece-wise polynomials of degree less than k on a regular mesh with h = 21−n. Then, a set of basis function for V k

n

is given by φn

jl(x) = 2

n 2 φj(2nx − l),

j = 0, . . . , k − 1, l = 0, . . . , 2n − 1. Let W k

0 = V k 0 and define the multi-wavelet subspace W k n ,

n = 1, 2, . . . as the L2-orthogonal compliment of V k

n−1 in V k n , i.e.,

V k

n−1 ⊕ W k n = V n k .

Then, we have V k

n = W k 0 ⊕ W k 1 ⊕ . . . ⊕ W k n .

slide-68
SLIDE 68

let ψ0, . . . , ψk−1 be the orthogonal basis for W k

1 . Then,

W k

n (n ≥ 1) is spanned by

ψn

jl(x) = 2

n 2 ψj(2nx − l),

j = 0, . . . , k − 1, l = 0, . . . , 2n − 1.

slide-69
SLIDE 69

let ψ0, . . . , ψk−1 be the orthogonal basis for W k

1 . Then,

W k

n (n ≥ 1) is spanned by

ψn

jl(x) = 2

n 2 ψj(2nx − l),

j = 0, . . . , k − 1, l = 0, . . . , 2n − 1. We have 1

−1

ψn

il(x)ψn′ i′l′(x) dx = δii′δll′δnn′,

and they form a complete orthonormal basis in L2(−1, 1).

slide-70
SLIDE 70

Sparse tensor products in (−1, 1)d

Let Ω = (−1, 1)d. The regular tensor product approximation space Vk

L = V k L ⊕ · · · ⊕ V k L ,

which can be written as Vk

L =

  • 0≤li≤L

W k

l1 ⊕ W k l2 · · · ⊕ W k ld .

However, dim(Vk

L) = O(2Ld) which is not feasible for d large.

slide-71
SLIDE 71

Sparse tensor products in (−1, 1)d

Let Ω = (−1, 1)d. The regular tensor product approximation space Vk

L = V k L ⊕ · · · ⊕ V k L ,

which can be written as Vk

L =

  • 0≤li≤L

W k

l1 ⊕ W k l2 · · · ⊕ W k ld .

However, dim(Vk

L) = O(2Ld) which is not feasible for d large.

The sparse tensor product approximation space ˆ Vk

L =

  • 0≤l1+···+ld≤L

W k

l1 ⊕ W k l2 · · · ⊕ W k ld ,

with dim(ˆ Vk

L) = O(Ld−12L).

slide-72
SLIDE 72

Galerkin approximation of the collision operator with sparse multi-wavelets

Let L = ( 1, · · · , ld) with 0 ≤ l1 + · · · + ld ≤ L. Then φk

L ∈ ˆ

Vk

L,

and the inner product of the collision operator with ˆ Vk

L is

I(φk

L) =

  • R3 φk

L(v)dv

  • R3
  • S2[f (x, v′)f (x, w′) − f (x, v)f (x, w)]|v − w|dSdw

=

  • R3
  • R3 f (x, v)f (x, w)dvdw

|v − w| 2

  • S2[φk

L(v′) + φk L(w′) − φk L(v) − φk L(w)]dS

=

  • R3
  • R3 f (x, v)f (x, w)A(v, w; φk

L)dvdw,

where A(v, w; φk

L) = |v − w|

2

  • S2[φk

L(v′) + φk L(w′) − φk L(v) − φk L(w)]dS.

slide-73
SLIDE 73

Proposed algorithm

In practice, we can replace the sparse Legendre multi-wavelets by corresponding Lagrangian basis functions on a sparse grid. Given a sparse grid for v and w, we pre-compute A(v, w; φk

L)

for all φk

L ∈ ˆ

Vk

L.

The total # of entries in this trilinear tensor is of order (Ld−12L)3. However, due to the local feature of φk

L, most of

A(v, w; φk

L) are zeros.

We are in the process of working out the detail, and hopefully this will be a feasible alternative approach for the collision

  • perator.
slide-74
SLIDE 74

Concluding remarks

Fast sparse spectral algorithms for elliptic equations in high-dimensional bounded and unbounded domains

fast transforms between mapped Chebyshev sparse grids and corresponding “hyperbolic cross” space; fast solvers for the resulting linear system.

slide-75
SLIDE 75

Concluding remarks

Fast sparse spectral algorithms for elliptic equations in high-dimensional bounded and unbounded domains

fast transforms between mapped Chebyshev sparse grids and corresponding “hyperbolic cross” space; fast solvers for the resulting linear system.

Preliminary results to electronic Schr¨

  • dinger equation indicate

that this approach is promising for solving moderately high-dimensional PDEs.

slide-76
SLIDE 76

Concluding remarks

Fast sparse spectral algorithms for elliptic equations in high-dimensional bounded and unbounded domains

fast transforms between mapped Chebyshev sparse grids and corresponding “hyperbolic cross” space; fast solvers for the resulting linear system.

Preliminary results to electronic Schr¨

  • dinger equation indicate

that this approach is promising for solving moderately high-dimensional PDEs. A Galerkin sparse tensor product approach with multi-wavelets is proposed for dealing with the collision operator.

Thank You!