materials process design and control laboratory 1 high-dimensional model representation technique...

49
C C O O R R N N E E L L L L U N I V E R S I T Y Materials Process Design and Control Laboratory Materials Process Design and Control Laboratory 1 High-dimensional model representation High-dimensional model representation technique for the solution of stochastic PDEs technique for the solution of stochastic PDEs Nicholas Zabaras and Xiang Ma Materials Process Design and Control Laboratory Sibley School of Mechanical and Aerospace Engineering 101 Frank H. T. Rhodes Hall Cornell University Ithaca, NY 14853-3801 Email: [email protected] URL: http://mpdc.mae.cornell.edu/ CSE09, SIAM Conference on Computational Science and Engineering, Miami, FL, March 2-6, 2009

Upload: kerry-spencer

Post on 02-Jan-2016

219 views

Category:

Documents


1 download

TRANSCRIPT

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 11

High-dimensional model representation High-dimensional model representation technique for the solution of stochastic PDEs technique for the solution of stochastic PDEs

Nicholas Zabaras and Xiang MaMaterials Process Design and Control Laboratory

Sibley School of Mechanical and Aerospace Engineering101 Frank H. T. Rhodes Hall

Cornell University Ithaca, NY 14853-3801

Email: [email protected]: http://mpdc.mae.cornell.edu/

CSE09, SIAM Conference on Computational Science and Engineering, Miami, FL, March 2-6, 2009

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 22

Outline of the presentation

Problem definition

Basic concepts of high-dimensional model

representation (HDMR)

Adaptive sparse grid collocation (ASGC) method for

interpolating component functions

Examples

Conclusions

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 33

Motivation

Why uncertainty modeling ?

Assess product and process reliability

Estimate confidence level in model predictions

Identify relative sources of randomness

Provide robust design solutions

All physical systems have inherent associated randomness

SOURCES OF UNCERTAINTIES

• Multiscale nature inherently statistical

• Uncertainties in process conditions

• Material heterogeneity

• Model formulation – approximations,

assumptions

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 44

Motivation Previous developed conventional and adaptive collocation methods are not suitable for high- dimensional problems due to their weakly dependence on the dimensionality (logarithmic) in the error estimate. Although ASGC can alleviate the problem to some extent, it depends on the regularity of the problem and it is only effective when some random dimensions are more important than others.

High-dimensional model representation (HDMR) is an efficient tool to capture the model output as finite number of hierarchical correlated function-expansions in terms of the inputs starting from lower-order to higher-order. There does not exist a closed form for implementation to higher-order expansion. We try to develop a general formulation of HDMR for efficient computer implementation and to achieve higher accuracy. To represent the component functions of HDMR, either low-dimensional data table or tensor-product type Lagrange polynomial interpolation are currently used, which significantly affects its accuracy of interpolation.

In the current work, we combine the strength of the HDMR and ASGC methods to develop a stochastic dimension reduction method.

CSGC

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 55

Problem definition

Define a complete probability space . We are interested to find a stochastic function such that for P-almost everywhere (a.e.) , the following equation holds:

, ,F P

:u D

, ; , ,

, ,

L u f D

B u g D

x x x

x x x

where are the coordinates in , L is (linear/nonlinear) differential operator, and B is a boundary operator.

1 2, , , dx x xx d

In the most general case, the operators L and B as well as the driving terms f and g, can be assumed random.

In general, we require an infinite number of random variables to completely characterize a stochastic process. This poses a numerical challenge in modeling uncertainty in physical quantities that have spatio-temporal variations, hence

necessitating the need for a reduced-order representation.

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 66

Representing randomness

S

Sample space of elementary events

Real line

Random

variable

MAP

Collection of all possible outcomes

Each outcome is mapped to a

corresponding real value

Interpreting random variables as functions

A general stochastic process is a random field with variations along space and time – A

function with domain (Ω, Τ, S)

1. Interpreting random variables

2. Distribution of the random variable

e.g. inlet velocity, inlet temperature

1 0.1o

3. Correlated data

ex. presence of impurities, porosity

Usually represented with a correlation function

We specifically concentrate on this

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 77

Representing Randomness

1. Representation of random process

- Karhunen-Loève, Polynomial Chaos expansions

2. Infinite dimensions to finite dimensions

- depends on the covariance

Karhunen-Loèvè expansion

Based on the spectral decomposition of the covariance kernel of the stochastic process

Random process Mean

Set of random variables to

be found

Eigenpairs of covariance

kernel

• Need to know covariance

• Converges uniformly to any second order process

Set the number of stochastic dimensions, N

Dependence of variables

Pose the (N+d) dimensional problem

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 88

Karhunen-Loeve expansion

1

( , , ) ( , ) ( , ) ( )i ii

X x t X x t X x t Y

Stochastic process

Mean function

ON random variablesDeterministic functions

Deterministic functions ~ eigen-values, eigenvectors of the covariance function

Orthonormal random variables ~ type of stochastic process

In practice, we truncate (KL) to first N terms

1( , , ) fn( , , , , )NX x t x t Y Y

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 99

The finite-dimensional noise assumption

By using the Doob-Dynkin lemma, the solution of the problem can be described by the same set of random variables, i.e.

1, , , , Nu u Y Y x x

So the original problem can be restated as: Find the stochastic function such that

:u D

, ; , ,

, ; , ,

L u f D

B u g D

x Y x Y x

x Y x Y x

In this work, we assume that are independent random variables with probability density function . Let be the image of . Then

1

, N

i ii

Y

Y Y

1

N

i iY

i i iY

is the joint probability density of with support 1, , NY YY

1

NN

ii

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 1010

High dimensional model representation (HDMR)1

Let be a real-value multivariate stochastic function: , which depends on a N-dimensional random vector . A HDMR of can be described by

f Y N 1 2, , 0,1

N

NY Y Y Y

f Y 1 1

1

01

, ,s s

s

N N

i i i is i i

f f f Y Y

Y

where the interior sum is over all sets of integers , that satisfy

s1, , si i

11 si i N .This relation means that

1 2 1 2 1 2 3 1 2 3

1 2 1 2 3

1 1

1

01

12 1

, , ,

, , , ,s s

s

N N N

i i i i i i i i i i i ii i i i i i

N

i i i i N Ni i

f f f Y f Y Y f Y Y Y

f Y Y f Y Y

Y

can be viewed as a finite hierarchical correlated function expansion in terms of the input random variables with increasing dimensions.

1N S N

The basic conjecture underlying HDMR is that the component functions arising in typical “real problems” will not likely exhibit high-order

cooperativity among the input variables such that the significant terms in the HDMR expansion are expected to satisfy the relation: for

1. O. F. Alis, H. Rabitz, General foundations of high dimensional representations, Journal of Mathematical Chemistry, 25 (1999) 127-142.

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 1111

High dimensional model representation (HDMR)

1 2 1 2 1 2 3 1 2 3

1 2 1 2 3

1 1

1

01

12 1

, , ,

, , , ,s s

s

N N N

i i i i i i i i i i i ii i i i i i

N

i i i i N Ni i

f f f Y f Y Y f Y Y Y

f Y Y f Y Y

Y

In this expansion: denotes the zeroth-order effect which is a constant.

The component function gives the effect of the variable acting independently of the other input variables.

The component function describes the interactive effects of the variables and . Higher-order terms reflect the cooperative effects of increasing numbers of variables acting together to impact upon .

The last term gives any residual dependence of all the variables locked together in a cooperative way to influence the output .

Depending on the method to represent component functions, there are two types of HDMR: ANOVA-HDMR and Cut-HDMR.

i if Y iY

1 2 1 2

,i i i if Y Y

1iY

2iY

12 1, ,N Nf Y Y

0f

f Y

f Y

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 1212

ANOVA - HDMR

ANOVA-HDMR is the same as the analysis of variance (ANOVA) decomposition used in statistics. Its component functions are given as

1

21 2 1 2 1 1 2 2

1 2

0

0

0 ,

,

N

N

N

i i jj i

i i i i j i i i ij i i

f f d

f Y f dY f

f Y Y f dY f Y f Y f

Y Y

Y

Y

This expansion is useful for measuring the contributions of the variance of individual component functions to the overall variance of the output.

A significant drawback of this method is the need to compute the above integrals to extract the component functions, which in general are high- dimensional integrals and thus difficult to carry out.

To circumvent this difficulty, a computationally more efficient cut-HDMR expansion which can be combined with ASGC is introduced in the next few slides.

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 1313

Cut - HDMR

In this work, Cut – HDMR is used for the multivariate interpolation problem. For Cut-HDMR, we fix a reference point . For this work, we fix it at the mean of the random input. The component functions

of Cut-HDMR are given as follows:

1 2, , , NY Y YY

1 2

1 2 1 2 1 2 1 1 2 2

1 2 3

1 2 3 1 2 3 1 2 3 1 2 1 2 1 3 1 3

2 3 2 3 1 1 2 2 3 3

1

1 1 1 1 1

0

0

0

0

, ,

, , , , , ,

,

, ,s

s s s s

ii i i

i ii i i i i i i i i i

i i ii i i i i i i i i i i i i i i i i

i i i i i i i i i i

i ii i i i i i i i i

f f

f Y f Y f

f Y Y f Y Y f Y f Y f

f Y Y Y f Y Y Y f Y Y f Y Y

f Y Y f Y f Y f Y f

f Y Y f Y Y f Y

Y

1 1

1 1 1 2

0 1

,s

s s

N N

i i ii i i i i i

Y f Y f

where the notation stands for the function with all the remaining variables of the input random vector set to , e.g.

1

1,s

s

i ii if Y Y f Y

Y i

if Y stands for 1 1 1, , , , ,i i i Nf Y Y Y Y Y which is a univariate function.

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 1414

Cut - HDMR

The cut-HDMR expansion up to the pth order with minimizes the following functional:

0 p N

1 1

1

2

01

,N p p

p

N

i i i i i i i ii i i i

J f f f Y f Y Y d d Y Y dY

Y Y Y

with the constraints 1 1, , 0, 1, 2, ,

p pi is s

i i i iY Y

f Y Y s p

The null property introduced above serves to assure that the functions are orthogonal with respect to the inner product induced by the measure

1 1 1 1

, , , , 0p p s s

N

i i i i j j j jf Y Y f Y Y d

Y

for at least one index differing in and , and may be the same as . This orthogonality condition implies that the cut-HDMR expansion to pth order is exact along lines, planes and hyperplanes of dimension up to p which pass through the reference point .

1, , pi i 1, , sj j sp

Y

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 1515

Cut – HDMR VS Taylor expansion

Suppose defined in a unit hypercube can be expanded as a convergent Taylor series at reference point , i.e.,

f Y

Y

2

1 , 1

1

2!

N N

i i i i j ji i ji i j

f ff f Y Y Y Y Y Y

Y Y Y

Y Y

Y Y

It is shown that the first order component function is the sum of all the Taylor series terms which contain and only contain variable . Similarly, the second order component function is the sum of all the Taylor series terms which contain and only contain variables and .

i if Y

iY

,ij i jf Y Y

iY

jY

Thus, the infinite number of terms in the Taylor series are partitioned into finite groups and each group corresponds to one Cut-HDMR component functions.

Therefore, any truncated Cut-HDMR expansion gives a better approximation of than any truncated Taylor series because the later only contains a finite number of terms of Taylor series.

f Y

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 1616

Truncation error of Cut - HDMR

Assume that all mixed derivatives that include not more than one differentiation with respect to each variable are piecewise continuous. These are the derivatives

1

1for 1 ; 1,2, , .s

s

si i

fi i N s N

Y Y

Denote 1 1sup /

s s

si i i i

YA f Y Y

and let1

11s

s

N N

i is p i i

A A

Then the truncation error1 up to pth order is pf Y f Y A

1. I. M. Sobol, Theorems and examples on high dimensional model representation, Reliability Engineering and System safety 79 (2003) 187-193.

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 1717

A general formulation of Cut-HDMR

In its current form, each component function is defined through lower order component functions which are unknown a priori. Therefore, it is not easy for computer programming. We develop here a general form of Cut-HDMR.

For lower-order component functions, it is easy to rewrite them as

1 2 1 2

1 2 1 2 1 2 1 2

1 2 3 1 31 2

1 2 3 1 2 3 1 2 3 1 2 1 3

2 3 31 2

2 3 1 2 3

0

0

0

0

, ,

, , , , , ,

,

ii i i

i i i ii i i i i i i i

i i i i ii ii i i i i i i i i i i i i

i i ii ii i i i i

f f

f Y f Y f

f Y Y f Y Y f Y f Y f

f Y Y Y f Y Y Y f Y Y f Y Y

f Y Y f Y f Y f Y f

Y

Each component function is expressed in terms of the function output at a specific input. However, the procedure to derive these formulas becomes tedious to carry out along these lines.

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 1818

A general formulation of Cut-HDMR

Instead, since HDMR has an analogous structure to the many body expansion used in molecular physics to represent a potential energy surface, we generalize its idea to the current work and get the following formulation of a order expansion of Cut-HDMR:

1 1

1

01

, ,s s

s

P N

i i i is i i

f f f Y Y

Y

1

1 1 11

01

, 1 1 , ,j jL

s s j jLL

S SS S L i i

i i i i i iL j j

f Y Y f f Y Y

where each component function can be obtained via the Mobius inversion approach from number theory as follows:

thP

where follows the notation defined before with all the remaining variables of the input random vector set to .

1

1, ,j jL

j jL

i i

i if Y Y

Y

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 1919

A general formulation of Cut-HDMR

Now we have a general formulation of truncated Cut-HDMR to pth order:

1 1

1

01

, ,s s

s

P N

i i i is i i

f f f Y Y

Y

1

1 1 1

1

01

, 1 1 , ,j jL

s s j jL

L

S SS S L i i

i i i i i iL j j

f Y Y f f Y Y

Previously, the components functions (i.e. , ) of Cut-HDMR were typically provided numerically, at discrete values (often tensor

product) of the input variables to construct the numerical data tables. Then the value of the function for arbitrary input was determined by performing only low-dimensional interpolation over , , …

i if Y 1 2 1 2

,i i i if Y Y

i if Y 1 2 1 2

,i i i if Y Y

This method is computational expensive and not accurate since construction of each component function involves lower order inaccurate component functions.

Now, through the general formulation, the problem is transformed to several small sub problems to interpolate the function 1

1, ,j jL

j jL

i i

i if Y Y

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 2020

Cut – HDMR and ASGC

1

1 1 11

01

, 1 1 , ,j jL

s s j jLL

S SS S L i i

i i i i i iL j j

f Y Y f f Y Y

Interpolated with ASGC

Now, instead of interpolating one N-dimensional function using ASGC, the problem is transformed to several at most P-dimensional

interpolation problems, where, in general, for real problems, forP N 1N

By using ASGC for the approximation, there is no need to search the numerical data tables. Interpolation is done quickly through simple weighted sum of the interpolation function and corresponding

hierarchical surplus. In addition, the mean and variance can be deduced easily.

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 21212121

Stochastic Collocation based framework

Function value at any point is simply

Stochastic function in 2 dimensions

Need to represent this function

Sample the function at a finite set of points

Use polynomials (Lagrange polynomials) to get a approximate representation

: Nf ( )

1 i MM iΘ Y

( ) ( )

1

( ) ( ) ( )M

i ii

i

f f a

Y Y YI

( )f ISpatial domain is approximated using a FE, FD, or FV discretization.

Stochastic domain is approximated using multidimensional interpolating functions

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 2222

LET OUR BASIC 1D INTERPOLATION SCHEME BE SUMMARIZED AS

IN MULTIPLE DIMENSIONS, THIS CAN BE WRITTEN AS

Denote the one dimensional interpolation formula as

In higher dimension, a simple case is the tensor product formula

Conventional sparse grid collocation (CSGC)

Using the 1D formula, the sparse interpolant , where is the depth of sparse grid interpolation and is the number of stochastic dimensions, is given by the Smolyak algorithm as

,q NA

00,q q N

q

1i i iU U 0 0U

Here we define the hierarchical surplus as:

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 23232323

Choice of collocation points and nodal basis functions In the context of incorporating adaptivity, we have used the Newton-Cotes grid using equidistant support nodes and use the linear hat function as the univariate nodal basis.

ijY

12i ijY

12i ijY

1

Furthermore, by using the linear hat function as the univariate nodal basis function, one ensures a local support in contrast to the global support of Lagrange polynomials. This ensures that discontinuities in the stochastic space can be resolved. The piecewise linear basis functions can be defined as

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 24242424

Adaptive sparse grid collocation (ASGC)1

Let us first revisit the 1D hierarchical interpolation For smooth functions, the hierarchical surpluses tend to zero as the interpolation level increases.

On the other hand, for non-smooth function, steep gradients/finite discontinuities are indicated by the magnitude of the hierarchical surplus.

The bigger the magnitude is, the stronger the underlying discontinuity is.

Therefore, the hierarchical surplus is a natural candidate for error control and implementation of adaptivity. If the hierarchical surplus is

larger than a pre-defined value (threshold), we simply add the 2N neighbor points of the current point.

1 .X. Ma, N. Zabaras, An hierarchical adaptive sparse grid collocation method for the solution of stochastic differential equations, JCP, in press, 2009.

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 2525

Integrating HDMR and ASGC

1

1 1 1

1

01

, 1 1 , ,j jL

s s j jL

L

S SS S L i i

i i i i i iL j j

f Y Y f f Y Y

1 1

1

01

, ,s s

s

P N

i i i is i i

f f f Y Y

Y

Unlike using numerical tables, there is no need to search in the table. Interpolation is done quickly through simple weighted sum of the basis functions and the corresponding hierarchical surpluses.

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 2626

Integrating HDMR and ASGC

The mean of the random solution can be evaluated as follows:

Denoting we rewrite the mean as

2u To obtain the variance of the solution, we need to first obtain an approximate expression for

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 2727

Implementation

The first procedure is to find the different ordered index set from the set, i.e. . 1, 2 , , 1,2, , Si i i N

The number of components is given by

0

!

( )! !

P

s

N

N s s

Although this number still grows quickly with increasing number of dimensions, we can solve each of the sub-problems more efficiently than solving the N-dimensional problem directly.

We use a recursive loop to generate all the index sets. After solving each sub problem, the surpluses for each interpolant are stored.

Each sub-problem is uniquely determined by its corresponding index set. Therefore, it is important to transform the index set to a unique key to store surplus and perform interpolation.

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 2828

Implementation

We use the following convention to generate the unique key as a string

“d(dimension of subproblem)_(index set)” Only using index set is not enough. For example, for a 25 dimensional

problem, the index set 12 can be either considered as a one-dimensional problem in the 12th dimension, or as a two-dimensional

problem in the 1st and 2nd dimension. Therefore, we put a dimension in front of the index set, e.g. “1_12” and “2_12”.

To compute the value using the HDMR, it is simple to perform interpolation for different sub-problems according to the index set.

Therefore, we use <map> from the C++ standard template library to store the different interpolants. We store each interpolant according to its unique key value defined above. Then we can perform the

interpolation by calling

map[key]->interpolate()

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 2929

Numerical example: Mathematical functions

It is noted that according to the right hand side of HDMR, if the function has an additive structure, the HDMR is exact, i.e. when can be

additively decomposed into functions of single variables, then the first-order expansion is enough to exactly represent the original function.

f Y

i if Y

Let us consider the following simple function:

2

1

N

ii

f Y

Y

It is expected that, no matter what the dimensionality is, it is enough to use just 1st order expansion of HDMR.

A 3rd order expansion is used although only 1st order is enough. After constructing the HDMR, we generate 100 random points in the domain and compute the normalized L2 interpolation error with the exact value. The result is shown in the following table.

Error HDMR exact

exact

y y

y

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 3030

Numerical example: Mathematical functions

Next, we compare it using 1st HDMR with ASGC with same epsilon and CSGC.

N = 10, Level 7 CSGCN = 10, Level 7 CSGC #points#points LL2 2 ErrorError

11stst order HDMR order HDMR 12911291 2.9794046540e-052.9794046540e-05

22ndnd order HDMR order HDMR 3301633016 2.9794046540e-052.9794046540e-05

33rdrd order HDMR order HDMR 340336340336 2.9794046540e-052.9794046540e-05

Therefore, when the target function has an additive structure, the converged HDMR

is better than both ASGC and CSGC.

The #points for CSGC at q = 6 is 171425. However, to go to the next level, the #points is 652065.

As expected, 1st order HDMR is

enough to capture the function.

101

102

103

104

105

106

10-5

10-4

10-3

10-2

10-1

100

# Points

L2 e

rro

r

CSGC: q = 6

ASGC: q = 7, = 10-4

1st HDMR + CSGC : q = 7

1st HDMR + ASGC : q = 7, = 10 -4

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 3131

Numerical example: Mathematical functions

N = 10N = 10 #points(HDMR)#points(HDMR) MaxError MaxError (HDMR)(HDMR)

#points( ASGC)#points( ASGC) MaxError MaxError ( ASGC)( ASGC)

εε = 1e-1 = 1e-1 5151 1.3886863491e-1.3886863491e-0101

221221 1.3886863491e-1.3886863491e-0101

εε = 1e-2 = 1e-2 171171 8.3629048275e-8.3629048275e-0303

14211421 8.3629048275e-8.3629048275e-0303

εε = 1e-3 = 1e-3 331331 2.1349220166e-2.1349220166e-0303

30213021 2.1349220166e-2.1349220166e-0303

εε = 1e-4 = 1e-4 12911291 1.3026665672e-1.3026665672e-0404

1262112621 1.3026656424e-1.3026656424e-0404

εε = 1e-5 = 1e-5 51315131 8.3142450165e-8.3142450165e-0606

5102151021 8.3141714420e-8.3141714420e-0606

If however, the multiplicative nature of the function is dominant then all the right hand side components of HDMR are needed to obtain the best result. However, if HDMR requires all 2N components to obtain a desired accuracy, the method

becomes very expensive. This is demonstrated through the following example.

Therefore, as expected the error threshold εε also controls the error of a also controls the error of a converged HDMR. This can be partially explained that each component has converged HDMR. This can be partially explained that each component has error error εε, and HDMR is a linear combination of components, therefore the error is , and HDMR is a linear combination of components, therefore the error is M M εε, where M is a constant., where M is a constant.

Next, we investigate the convergence with respect to the error threshold while fixing the expansion order to 1.

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 3232

Numerical example: Mathematical functions Next we consider the function “product peak” from GENZ test package:

1222

1

N

i i ii

f c Y w

Y

1 2 3 4 510

-4

10-3

10-2

10-1

100

HDMR order

L2 e

rro

r

HDMR + CSGC: q = 6

Interpolation ErrorIntegration Error

The normalized interpolation error is defined the same as before. The integration error is defined as the relative error to the exact value.

So we need at least 5th order HDMR to get a converged

expansion, due to the structure of this function.

It is also interesting to note that for integration, the HDMR

converges faster than forinterpolation.

N = 10

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 3333

Numerical example: Mathematical functions

Thus the HDMR method is indeed not optimal in this case which requires about two billion points. Therefore, HDMR is not equally useful for interpolation of all types of mathematical functions.

We will investigate further these aspects of HDMR for the solution of stochastic PDEs.

100

102

104

106

108

10-4

10-3

10-2

10-1

100

# Points

Inte

rpo

latio

n L 2

Err

or

CSGC : q = 6

ASGC : = 10 -5

5th HDMR + CSGC : q = 6

5th HDMR + ASGC : q = 6, = 10 -5

100

102

104

106

108

10-4

10-3

10-2

10-1

100

# Points

Inte

grat

ion

L 2 Err

or

CSGC : q = 6

ASGC : = 10 -5

5th

HDMR + CSGC : q = 6

5th

HDMR + ASGC : q = 6, = 10 -5

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 3434

Numerical example: Mathematical functions

1 2 3 4 510

-3

10-2

10-1

100

101

Order

L2 in

terp

olat

ion

erro

r

Reference point 1: (0.5,,0.5)Reference point 2: (0.55,,0.55)Reference point 3: (0.1,,0.1)

1 2 3 4 510

-4

10-3

10-2

10-1

# Points

L2 in

tegr

atio

n er

ror

Reference point 1: (0.5,,0.5)Reference point 2: (0.55,,0.55)Reference point 3: (0.1,,0.1)

We consider three different reference points in the above plots. It is shown that the results near the center of the domain is more accurate. Therefore, we always take our reference point as the mean vector.

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 3535

Numerical example: Stochastic elliptic problem Here, we adopt the model problem

, , , in

, 0 on

N Na u f D

u D

with the physical domain . To avoid introducing errors of any significance from the domain discretization, we take a deterministic smooth load with homogeneous boundary conditions.

2, 0,1D x x y

, , cos sinNf x y x y

The deterministic problem is solved using the finite element method with 900 bilinear quadrilateral elements. Furthermore, in order to eliminate the errors associated with a numerical K-L expansion solver and to keep the random diffusivity strictly positive, we construct the random diffusion

coefficient with 1D spatial dependence as ,Na x

1/ 2

12

log , 0.5 12

N

N n n nn

La x Y x Y

where are independent uniformly distributed random variables in the interval

, 1,nY n N 3, 3

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 3636

Numerical example: Stochastic elliptic problem

In the earlier expansion,

2

1/ 2 2: exp , if 1

8n

nL

L n

and

2sin , if even

:

2cos , if odd

p

n

p

nx

nL

xn

xn

L

The parameter can be taken as and the parameter L is

pL max 1,2p cL L/cL L Lp

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 3737

Numerical example: Stochastic elliptic problem This expansion is similar to a K-L expansion of a 1D random field with stationary covariance

21 21 2 2

log , 0.5 , expNc

x xa x x x

L

Small values of the correlation correspond to a slow decay, i.e. each stochastic dimension weighs almost equally. On the other hand, large

values of result in fast decay rates, i.e., the first several stochastic dimensions corresponding to large eigenvalues weigh relatively more. By using this expansion, it is assumed that we are given an analytic stochastic input. Thus, there is no truncation error. This is different from the discretization of a random filed using the K-L expansion, where for different correlation lengths we keep different terms accordingly.

In this example, we fix N and change to adjust the importance of each stochastic dimension. In this way, we investigate the effects of .

cL

cL

cL

cL

The normalized error is the same as before, where the exact solution is taken as the statistics from 106 Monte Carlo samples

2L

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 3838

Numerical example: Stochastic elliptic problem N=7

1 2 3 40

0.002

0.004

0.006

0.008

0.01

0.012

Order

L2 E

rror

s

Mean : HDMR + ASGC = 10-6

Std : HDMR + ASGC = 10-6

Lc = 1/2

1 2 3 40

0.005

0.01

0.015

0.02

0.025

0.03

0.035

Order

L2 E

rror

s

Mean : HDMR + ASGC = 10-7

Std : HDMR + ASGC = 10-7

Lc = 1/4

1 2 3 40

0.005

0.01

0.015

0.02

Order

L2 E

rror

s

Mean : HDMR + ASGC = 10-8

Std : HDMR + ASGC = 10-8

Lc = 1/16

1 2 3 40

0.002

0.004

0.006

0.008

0.01

Order

L2 E

rror

s

Mean : HDMR + ASGC = 10-9

Std : HDMR + ASGC = 10-9

Lc = 1/64

Convergence with orders of

expansion

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 3939

Numerical example: Stochastic elliptic problem N=7

-0.01 0 0.01 0.02 0.03 0.040

10

20

30

40

50

60

70

80

90

u

PDF

MC

HDMR, 1st

order

HDMR, 2nd

order

HDMR, 3rd

order

Lc = 1/2

-0.01 0 0.01 0.02 0.03 0.040

20

40

60

80

100

u

PDF

MC

HDMR, 1st order

HDMR, 2nd order

HDMR, 3rd order

Lc = 1/4

0 0.005 0.01 0.015 0.020

20

40

60

80

100

120

140

160

u

PDF

MC

HDMR, 1st

order

HDMR, 2nd

order

HDMR, 3rd

order

Lc = 1/16

0.005 0.01 0.0150

50

100

150

200

250

300

350

u

PDF

MC

HDMR, 1st

order

HDMR, 2nd

order

HDMR, 3rd

order

Lc = 1/64

PDF at

(0.5,0.5)

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 4040

Numerical example: Stochastic elliptic problem N=7

101

102

103

104

105

106

10-4

10-3

10-2

10-1

100

# Points

L2 E

rror

of

Std

CSGC : q = 7

ASGC : = 10-6

3rd HDMR + CSGC : q = 7

3rd HDMR + ASGC : = 10-6

MC = 1,000,000

Lc = 1/2

101

102

103

104

105

10-4

10-3

10-2

10-1

100

# Points

L2 e

rror

of

Std

CSGC : q = 7

ASGC : = 10-8

2nd HDMR + CSGC : q = 7

2nd HDMR + ASGC : = 10-8

MC = 1,000,000

Lc = 1/16

101

102

103

104

105

10-4

10-3

10-2

10-1

100

# Points

L2 e

rror

of

Std

CSGC : q = 7

ASGC : = 10-9

2nd

HDMR + CSGC : q = 7

2nd

HDMR + ASGC : = 10-9

MC = 1,000,000

Lc = 1/64

101

102

103

104

105

106

10-3

10-2

10-1

100

# Points

L2 e

rror

of

Std

CSGC : q = 7

ASGC : = 10-7

3rd

HDMR + CSGC : q = 7

3rd

HDMR + ASGC : = 10-7

MC = 1,000,000

Lc = 1/4

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 4141

Numerical example: Stochastic elliptic problem N=7

The convergence of mean is faster than that of std, since mean is a first order statistics and first order expansion is enough to give good result.

For large correlation length, in order to achieve the same accuracy as other methods, a higher order expansion is needed since the problem is

highly anisotropic and higher order effects for std cannot be neglected. However, as seen from the plots, as the correlation length decreases, the problem becomes smoother in the random space, even first order

expansion can give enough accuracy for std, e.g. when Lc = 1/64, 1st order can give a error of O(10-3).

For large correlation length, ASGC is still the most effective method. When correlation length is small, the effect of ASGC becomes the same as CSGC. In this case, lower order expansion of HDMR is enough and thus becomes favorable than others.

From the PDF figures, we can have the same conclusions. For small correlation length, the PDF of first order expansion is nearly the same as

that of MC result.

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 4242

Numerical example: Stochastic elliptic problem N=21

101

102

103

104

105

106

10-3

10-2

10-1

100

# Points

L2 e

rror

of

std

CSGC : q = 4

2nd

HDMR + CSGC : q = 4MC = 1,000,000

N = 21, Lc = 1/64

101

102

103

104

105

106

10-4

10-3

10-2

10-1

# Points

L2 e

rror

of

mea

n

CSGC : q = 4

2nd

HDMR + CSGC : q = 4MC = 1,000,000

N = 21, Lc = 1/64

0 0.2 0.4 0.6 0.8 10

0.002

0.004

0.006

0.008

0.01

x

Mea

n

MC = 1,000,000

1st

HDMR + CSGC : q = 4

2nd

HDMR + CSGC : q = 4

N = 21, Lc = 1/64

0 0.2 0.4 0.6 0.8 10

0.2

0.4

0.6

0.8

1

1.2

1.4x 10

-3

x

Std

MC = 1,000,000

1st

HDMR + CSGC : q = 4

2nd

HDMR + CSGC : q = 4

N = 21, Lc = 1/64

Std along y = 0.5

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 4343

Numerical example: Stochastic elliptic problem N=21

For such a small correlation length, the effect of ASGC is the same as CSGC. So we only use HDMR + CSGC.

In such a high dimensional problem, CSGC or ASGC is not optimal in the sense that a large number of collocation points is needed to get good

accuracy. The level 4 dimension 21 CSGC has 145377 points, where the convergence rate is even worst than MC for std.

On the other hand, 2nd order HDMR + level 4 CSGC can give us a much higher accuracy for both mean and std, and the convergence rate is much better than MC method.

To improve the accuracy for CSGC method, we need to go to the next interpolation level, where the number of points is 1285761 which is definitely not good compared with MC method . Therefore, it is clear that sparse grid collocation method still depends weakly on the dimensionality

of the problem.

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 4444

Numerical example: Flow through random media

Basic equation for pressure and velocity in a domain

where denotes the source/sink term.

To impose the non-negativity of the permeability, we will treat the permeability as a log- normal random field obtained from the K-L expansion

injection well

production well

1 1 f x

exp( )k Y

where is a zero mean Gaussian random field with covariance function Y

2

2, expr

Cov x yL

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 4545

Numerical example: Flow through random media

We first fix and vary the correlation length, where the KLE is truncated such that the eigenvalues correspond to

99.7% of the energy.

2 1

0 0.2 0.4 0.6 0.8 10.002

0.004

0.006

0.008

0.01

0.012

0.014

0.016

0.018

x

Stan

dard

dev

iati

on

MC (1,000,000)

HDMR, 1st

order

HDMR, 2nd

order

HDMR, 3rd

order

L = 1.0, N = 6

0 0.2 0.4 0.6 0.8 10.006

0.008

0.01

0.012

0.014

0.016

0.018

0.02

0.022

x

Stan

dard

dev

iati

on

MC (1,000,000)

HDMR, 1st order

HDMR, 2nd

order

L = 0.5, N = 10

0 0.2 0.4 0.6 0.8 10.016

0.018

0.02

0.022

0.024

0.026

0.028

0.03

x

Stan

dard

dev

iati

on

MC (1,000,000)

HDMR, 1st

order

L = 0.1, N = 25

Std along y = 0.5

It is noted that in all three cases, even first-order expansioncan give accuracy as good as O(10-3). Higher-order terms do not provide significant improvements.

The reason is that when the truncated KLE can accurate represent the random field, the random variables are

uncorrelated. Therefore, their cooperative effects on the output are very weak while the individual effects have strong impact on the output.

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 4646

Numerical example: Flow through random media

101

102

103

104

105

106

10-3

10-2

10-1

100

101

# Points

L2 e

rror

of

std

CSGC : q = 8

ASGC : = 10-4

1st

HDMR + CSGC : q = 10

1st

HDMR + ASGC : = 10-4

MC (1,000,000)

L = 1.0

101

102

103

104

105

106

10-3

10-2

10-1

100

101

102

# Points

L2 e

rror

of

std

CSGC : q = 4

ASGC : = 10-4

1st

HDMR + CSGC : q = 10

1st

HDMR + ASGC : = 10-5

MC (1,000,000)

L = 0.1

101

102

103

104

105

106

10-3

10-2

10-1

100

101

102

#Points

L2 e

rror

of

std

CSGC : q = 6

ASGC : = 10-5

1st

HDMR + CSGC : q = 10

1st

HDMR + ASGC : = 10-5

MC (1,000,000)

L = 0.5

Since only first-order expansion is used, the convergence of HDMR + ASGC is very obvious. The # points is nearly two-orders of magnitude lower compared with the MC method.

The employed covariance kernel has a fast decay of eigen-values. Therefore, even correlation L= 0.1 will result in

different importance in each dimension, which can be effectively detected by the ASGC method.

The MC result is obtained with 50, 100, 100 terms in KLE expansion, respectively, which also shows that the truncation error of KLE is negligible.

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 4747

Numerical example: Flow through random media

101

102

103

104

105

106

10-3

10-2

10-1

100

101

# Points

L2 e

rror

of

std

CSGC: q = 6

ASGC : = 10-5

1st HDMR + CSGC : q = 10

1st HDMR + ASGC : = 10-5

Lc = 0.5, 2 = 0.25

Next we fix the correlation length and vary the variance of the covariance to account for large variability in the input

uncertainty.

101

102

103

104

105

106

10-3

10-2

10-1

100

101

102

# Points

L2 e

rror

of

std

CSGC : q = 6

ASGC : = 10-4

1st

HDMR + CSGC : q = 10

1st

HDMR + ASGC : = 10-4

MC : 1,000,000

L = 0.5, 2 = 4.0

101

102

103

104

105

106

10-3

10-2

10-1

100

101

102

# Points

L 2 err

or o

f st

d

CSGC : q = 6

ASGC : = 10-5

1st

HDMR + CSGC : q = 10

1st

HDMR + ASGC : = 10-5

MC : 1,000,000

L = 0.5, 2 = 2.0

The coefficient of variation is 53%, 253%, 732%, respectively. Although 10 KLE terms are used for calculation, 100 terms are

kept for MC simulation .

Again, first-order expansion successfully captures the input uncertainty

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 4848

Numerical example: Flow through random media

0 0.2 0.4 0.6 0.8 1

0.02

0.022

0.024

0.026

0.028

0.03

x

Stan

dard

Dev

iati

on

1st HDMR + ASGC : = 10-5

MC (1,000,000)

L = 1/64, N = 150

102

103

104

105

10-3

10-2

10-1

100

101

# Points

L2 e

rror

of

std

1st

order HDMR + ASGC : = 10-5

-1

-1/2

L = 1/64, N = 150

1/ 64, 150L N

Std along y = 0.5 Error convergence

Relative error is 36.635 10

CCOORRNNEELLLL U N I V E R S I T Y

CCOORRNNEELLLL U N I V E R S I T Y

Materials Process Design and Control LaboratoryMaterials Process Design and Control Laboratory 4949

Conclusions

A general framework for stochastic high-dimensional model representation technique is developed. It applies the deterministic HDMR in the stochastic space

together with ASGC to represent the stochastic solutions.

Various statistics can be easily extracted from the final reduced representation. Numerical examples show the efficiency and accuracy for both stochastic

interpolation and integration.

Numerical examples show that for problems with not-equally weighted random dimensions, a higher-order expansion is needed to get accurate results so that ASGC remains favorable in this case. On the other hand, for small correlation length when all dimensions weigh equally, the lower-order HDMR combined with ASGC or CSGC has the best convergence rate over MC and direct CSGC method.

For high-dimensional problems with equally weighted random dimensions, the sparse grid collocation method depends strongly on the dimensionality and HDMR is the best choice.

It is interesting that when the random input was truncated from KLE, only 1st order expansion was enough to capture the input uncertainty