kriging

18
Kriging 1

Upload: zan

Post on 22-Feb-2016

53 views

Category:

Documents


2 download

DESCRIPTION

Kriging. Cost of surrogates. In linear regression, the process of fitting involves solving a set of linear equations once. - PowerPoint PPT Presentation

TRANSCRIPT

Page 1: Kriging

Kriging

1

Page 2: Kriging

Cost of surrogates• In linear regression, the process of fitting involves solving a set

of linear equations once.

• With radial basis neural networks we have to optimize the selection of neurons, which will again entail multiple solutions of the linear system. We may find the best spread by minimizing cross-validation errors.

• Kriging, our next surrogate is even more expensive, we have a spread constant in every direction and we have to perform optimization to calculate the best set of constants (hyperparameters).– With many hundreds of data points this can become significant

computational burden.

Page 3: Kriging

Introduction to Kriging• Method invented in the 1950s by South African

geologist Danie G. Krige (1919-2013) for predicting distribution of minerals.– Formalized by French engineer, Georges Matheron in

1960.– Statisticians refer to a more general Gaussian Process

regression.• Became very popular for fitting surrogates to

expensive computer simulations in the 21st century.• It is one of the best surrogates available.• It probably became popular late mostly because of

the high computer cost of fitting it to data.

Page 4: Kriging

Kriging philosophy• We assume that the data is sampled from an unknown

function that obeys simple correlation rules.• The value of the function at a point is correlated to the

values at neighboring points based on their separation in different directions.

• The correlation is strong to nearby points and weak with far away points, but strength does not change based on location.

• Normally Kriging is used with the assumption that there is no noise so that it interpolates exactly the function values.

Page 5: Kriging

Reminder: Covariance and Correlation• Covariance of two random variables X and Y

• The covariance of a random variable with itself is the square of the standard deviation.

• Covariance matrix for a vector contains the covariances of the components.

• Correlation

• The correlation matrix has 1 on the diagonal.

cov( , ) [( )( )] [ ]X Y X YX Y E X Y E XY

cov( , )( , ) 1 ( , ) 1X Y

X Ycor X Y cor X Y

( ) cov( , )cov( , ) ( )Var X X YX Y Var Y

2( ) [ ( )]Var X X

Page 6: Kriging

Correlation between function values at nearby points for sine(x)

• Generate 10 random numbers, translate them by a bit (0.1), and by more (1.0)x=10*rand(1,10)8.147 9.058 1.267 9.134 6.324 0.975 2.785 5.469 9.575 9.649

xnear=x+0.1; xfar=x+1; • Calculate the sine function at the three sets.ynear=sin(xnear)0.9237 0.2637 0.9799 0.1899 0.1399 0.8798 0.2538 -0.6551 -0.2477 -0.3185

y=sin(x)0.9573 0.3587 0.9551 0.2869 0.0404 0.8279 0.3491 -0.7273 -0.1497 -0.2222

yfar=sin(xfar)0.2740 -0.5917 0.7654 -0.6511 0.8626 0.9193 -0.5999 0.1846 -0.9129 -0.9405

• Compare corelations: r=corrcoef(y,ynear) 0.9894; rfar=corrcoef(y,yfar) 0.4229

• Decay to about 0.4 over one sixth of the wavelength.– Wavelength on sine function is 2~6

Page 7: Kriging

Gaussian correlation function• Correlation between point x and point s

• We would like the correlation to decay to about 0.4 at one sixth of the wavelength .

• Approximately • For the function we would like to estimate

21

exp( ), ( ), ( )x s θvN

i i ii

C Z Z x s

Page 8: Kriging

Universal Kriging

• Trend function is most often a low order polynomial

• We will cover Ordinary Kriging, where linear trend is just a constant to be estimated by data.

• There is also Simple Kriging , where constant is assumed to be known (often as 0).

• Assumption: Systematic departures Z(x) are correlated.

• Kriging prediction comes with a normal distribution of the uncertainty in the prediction.

x

y

Kriging

Sampling data points

Systematic Departure

Linear Trend Model

1

ˆ( ) ( ) ( )x x xvN

i ii

y Z

Trend model

Systematic departure

Page 9: Kriging

Notation• The function values are given at points , with the point having

components (number of dimensions).• The function value at the ith point is =y(), and the vector of

function values is denoted y.• Given decay rates , we form the covariance matrix of the data

• The correlation matrix R above is formed from the covariance matrix, assuming a constant standard deviation which measures the uncertainty in function values.

• For dense data, will be small, for sparse data will be large.• How do you decide whether the data is sparse or dense?

2 ( ) ( ) 2 2

1

( , ) exp ( ) , 1,...,n

i ji j k k k ij y

k

Cov y y x x R i j n

Page 10: Kriging

Prediction and shape functions• Ordinary Kriging prediction formula

• The equation is linear in r, which means that the exponentials may be viewed as basis functions.

• The equation is linear in the data y, in common with linear regression, but b is not calculated by minimizing RMS.

• Note that far away from data (not good for substantial extrapolation).

1 ( ) 2

1

ˆ ˆ ˆ( ) exp ( )n

T T ii k k k

k

y R r x x

x r y b r

Page 11: Kriging

Fitting the data• Fitting refers to finding the parameters • We fit by maximizing the likelihood that the data comes from a

Gaussian process defined by .

• Once they are found, the estimate of the mean and standard deviation is obtained as

• Maximum likelihood is a tough optimization problem.• Some kriging codes minimize the cross validation error.• Stationary covariance assumption: smoothness of response

is quite uniform in all regions of input space.

112

1

ˆ ˆˆ ˆ

TT

T

RRR n

y 1 y 11 y1 1

Page 12: Kriging

Prediction variance

Square root of variance is called standard error.

The uncertainty at any x is normally distributed.

represents the mean of Kriging prediction.

1

2 11

(1 )ˆ( ) 1T

TT

RV y RR

1 rx r r1 r

Page 13: Kriging

Kriging fitting issues• The maximum likelihood or cross-validation optimization

problem solved to obtain the kriging fit is often ill-conditioned leading to poor fit, or poor estimate of the prediction variance.

• Poor estimate of the prediction variance can be checked by comparing it to the cross validation error.

• Poor fits are often characterized by the kriging surrogate having large curvature near data points (see example on next slide).

• It is recommended to visualize by plotting the kriging fit and its standard error.

Page 14: Kriging

Example of poor fits.

True function

Kriging ModelTrend model : Covariance function :

-8 -6 -4 -2 0 2 4 6 8

-20

0

20

40

60

80

100

x

y

True functionData points

Page 15: Kriging

selection

Good fit and poor standard error

-8 -6 -4 -2 0 2 4 6 8

-20

0

20

40

60

80

100

x

y

True functionTest pointsKrigin interpolation2SE bounds

SE: standard error

-8 -6 -4 -2 0 2 4 6 8

-20

0

20

40

60

80

100

x

y

True functionTest pointsKriging interpolation2SE bounds

Page 16: Kriging

Kriging with nuggets• Nuggets – refers to the

inclusion of noise at data points.

• The more general Gaussian Process surrogates or Kriging with nuggets can handle data with noise (e.g. experimental results with noise).

0 0.2 0.4 0.6 0.8 1-10

-5

0

5

10

15

20

x

y

True function value y(x)Surrogate PredictionData

Page 17: Kriging

Introduction to optimization with surrogates

• Based on cycles. Each consists of sampling design points by simulations, fitting surrogates to simulations and then optimizing an objective.

• Construct surrogate, optimize original objective, refine region and surrogate.– Typically small number of cycles with large number of simulations in

each cycle.

• Adaptive sampling – Construct surrogate, add points by taking into account not only

surrogate prediction but also uncertainty in prediction.– Most popular, Jones’s EGO (Efficient Global Optimization).– Easiest with one added sample at a time.

Page 18: Kriging

Problems Kriging• Fit the quadratic function of Slide 13 with

kriging using different options, like different covariance (different s) and trend functions and compare the accuracy of the fit.

• For this problem compare the standard error with the actual error.