annealing techniques applied to reservoir modeling and the

325
ANNEALING TECHNIQUES APPLIED TO RESERVOIR MODELING AND THE INTEGRATION OF GEOLOGICAL AND ENGINEERING (WELL TEST) DATA a dissertation submitted to the department of applied earth sciences and the committee on graduate studies of stanford university in partial fulfillment of the requirements for the degree of doctor of philosophy By Clayton Vernon Deutsch May 1992

Upload: others

Post on 09-Feb-2022

4 views

Category:

Documents


0 download

TRANSCRIPT

ANNEALING TECHNIQUES APPLIED TO RESERVOIR

MODELING AND THE INTEGRATION OF

GEOLOGICAL AND ENGINEERING (WELL TEST) DATA

a dissertation

submitted to the department of applied earth sciences

and the committee on graduate studies

of stanford university

in partial fulfillment of the requirements

for the degree of

doctor of philosophy

By

Clayton Vernon Deutsch

May 1992

c© Copyright 1992 by Clayton Vernon Deutsch

All Rights Reserved

ii

I certify that I have read this thesis and that in my opin-

ion it is fully adequate, in scope and in quality, as a

dissertation for the degree of Doctor of Philosophy.

Andre Journel(Principal Adviser)

I certify that I have read this thesis and that in my opin-

ion it is fully adequate, in scope and in quality, as a

dissertation for the degree of Doctor of Philosophy.

Thomas Hewett(Petroleum Engineering)

I certify that I have read this thesis and that in my opin-

ion it is fully adequate, in scope and in quality, as a

dissertation for the degree of Doctor of Philosophy.

Paul Switzer(Statistics/Applied Earth Sciences)

Approved for the University Committee on Graduate

Studies:

Dean of Graduate Studies

iii

Abstract

Stochastic reservoir models must honor as much input data as possible to be reliable

numerical models of the reservoir under study. Traditional simulation algorithms

are unable to honor either complex geological/morphological patterns or engineering

data from well tests. The technique developed in this dissertation may be used to

incorporate such information into stochastic reservoir models.

This dissertation develops the application of the optimization methods known as

simulated annealing, to stochastic simulation. The essential feature of the method is

the formulation of stochastic imaging as an optimization problem with some speci-

fied objective function. The additional information to be matched by the stochastic

images is built into the objective function. Complex geological patterns and effec-

tive properties inferred from well tests may be incorporated into stochastic reservoir

models with relatively modest computational effort.

Complex geological patterns or spatial features require multivariate spatial statis-

tics (n > 2) in addition to conventional bivariate (n = 2) statistics. By considering

selected multivariate spatial statistics it is possible to impose such geological patterns

on stochastic images.

The effective permeability inferred from a well test constrains the possible spatial

distribution of elementary grid block permeability values near the well bore. Once

again, it is possible to impose this well test information through an understanding

and heuristic quantification of the averaging process near the well bore.

iv

Acknowledgements

It is a great pleasure to thank my advisor, Professor Andre Journel, for all the guid-

ance and assistance given during my studies at Stanford. Andre’s contagious enthu-

siasm, accessibility for discussion, and pedagogic skill helped my research greatly.

I have also appreciated the constructive advise and support of Professors Thomas

Hewett, Paul Switzer, and Roland Horne.

I would like to thank Andre, the Department of Applied Earth Sciences, the

Stanford Center for Reservoir Forecasting, and the Alberta Research Council for

providing the financial assistance needed to complete this dissertation.

Further, I would like to acknowledge the encouragement of my parents, Leonard

and Anne Deutsch, which ultimately led to the completion of this dissertation.

Finally, I would like to thank my wife Pauline and my children Jared, Rebecca, and

Matthew for their inspiration and patience throughout my studies. This dissertation

is dedicated to Pauline for her love and constant support.

v

Contents

Abstract iv

Acknowledgements v

1 Introduction 1

2 Stochastic Reservoir Modeling: Concepts and Algorithms 10

2.1 Reservoir Performance Forecasting . . . . . . . . . . . . . . . . . . . 11

2.2 Goodness Criteria for Stochastic Simulation Techniques . . . . . . . . 15

2.3 Spatial Statistics and Stochastic Simulation . . . . . . . . . . . . . . 18

2.3.1 The Random Function Concept . . . . . . . . . . . . . . . . . 22

2.3.2 Inference and Stationarity . . . . . . . . . . . . . . . . . . . . 24

2.3.3 Kriging . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 27

2.3.4 Models of Uncertainty . . . . . . . . . . . . . . . . . . . . . . 29

2.3.5 The Sequential Approach to Stochastic Simulation . . . . . . . 30

2.3.6 Sequential Gaussian Simulation (SGS) . . . . . . . . . . . . . 33

2.3.7 Sequential Indicator Simulation (SIS) . . . . . . . . . . . . . . 35

2.4 Geological Structures and Well Test Data . . . . . . . . . . . . . . . . 39

2.4.1 Geological Structures . . . . . . . . . . . . . . . . . . . . . . . 40

2.4.2 Well Test Data . . . . . . . . . . . . . . . . . . . . . . . . . . 46

2.5 Annealing Techniques for Stochastic Simulation . . . . . . . . . . . . 48

2.5.1 Simulated Annealing . . . . . . . . . . . . . . . . . . . . . . . 52

2.5.2 The Maximum A Posteriori (MAP) Variant . . . . . . . . . . 56

vi

3 Application of Annealing Techniques 60

3.1 Framework for a General Purpose Annealing Simulation Program . . 61

3.2 The Objective Function . . . . . . . . . . . . . . . . . . . . . . . . . 68

3.3 Multiple-Point Statistics . . . . . . . . . . . . . . . . . . . . . . . . . 73

3.4 Well Test-Derived Effective Permeability . . . . . . . . . . . . . . . . 85

3.4.1 Empirical Relationship for the Well Test Effective Permeability 91

3.4.2 A Wrong Averaging Volume . . . . . . . . . . . . . . . . . . . 98

3.4.3 Advanced Well Test Interpretation Techniques: . . . . . . . . 102

4 A Comparative Study of Various Simulation Techniques 103

4.1 Setting of the Problem . . . . . . . . . . . . . . . . . . . . . . . . . . 104

4.2 Stochastic Simulation and Output Uncertainty . . . . . . . . . . . . . 112

4.3 Uncertainty due to Ergodic Fluctuations . . . . . . . . . . . . . . . . 137

4.4 Conditioning to Local Data . . . . . . . . . . . . . . . . . . . . . . . 144

4.5 Summary of the Results . . . . . . . . . . . . . . . . . . . . . . . . . 145

5 Advanced Applications of Annealing Techniques 153

5.1 General Geostatistical Problems . . . . . . . . . . . . . . . . . . . . . 154

5.1.1 Conditioning to Connectivity Functions . . . . . . . . . . . . . 154

5.1.2 Multivariate Spatial Transformation . . . . . . . . . . . . . . . 163

5.1.3 Accounting for a Secondary Variable . . . . . . . . . . . . . . 169

5.2 Geological Structures . . . . . . . . . . . . . . . . . . . . . . . . . . . 179

5.2.1 An Example Application: eolian Sandstone . . . . . . . . . . . 180

5.2.2 Impact on Output Response Uncertainty . . . . . . . . . . . . 187

5.3 Well Tests . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 204

6 Concluding Remarks 225

A Acquisition of Geological Images 231

B Fluctuations in the cdf and Variogram due to Ergodicity 243

vii

C A Detailed Look at Spatial Entropy 253

C.1 Entropy of Continuous Distributions . . . . . . . . . . . . . . . . . . 254

C.2 Entropy of Discrete Distributions . . . . . . . . . . . . . . . . . . . . 256

C.3 Spatial Entropy . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 258

C.4 Some Examples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 259

C.4.1 A Discrete Variable Example . . . . . . . . . . . . . . . . . . 259

C.4.2 A Continuous Variable Example . . . . . . . . . . . . . . . . . 264

C.5 Final Thoughts . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 265

D Kriging in a Finite Domain 268

E Documentation 281

E.1 Program sasimi: . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 282

E.2 Program sasimr: . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 289

Bibliography 294

viii

List of Tables

3.1 The number of classes in an N point histogram for different numbers

of univariate classes. . . . . . . . . . . . . . . . . . . . . . . . . . . . 77

3.2 The one dimensional indices of a four-point histogram with K=2. . . 78

4.1 The lag vectors used with annealing to generate unconditional simula-

tions with the Berea distribution and normal scores semivariogram. . 121

4.2 Summary of all output response values for unconditional Berea real-

izations. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 136

4.3 Summary of output response values for SGS and SIS before and after

removing ergodic fluctuations in the cdf. . . . . . . . . . . . . . . . . 143

4.4 Summary of all output response values. . . . . . . . . . . . . . . . . . 146

5.1 Summary of output response values for SAS realizations before and

after post processing to the connectivity function. . . . . . . . . . . . 160

5.2 Reference indicator variogram parameters. . . . . . . . . . . . . . . . 194

5.3 Summary of output response values for cross stratified sands and silty-

sands example.. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 200

5.4 The reservoir properties used for the numerical well test integration

example. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 208

5.5 Summary of output response values for the realizations before and after

conditioning to the well test-derived effective permeability. . . . . . . 223

B.1 The grid sizes considered to evaluate the fluctuations in the histogram

and variogram due to ergodicity . . . . . . . . . . . . . . . . . . . . . 247

ix

C.1 The entropy in a unit vertical lag for the bombing model, the multi-

Gaussian realizations, and the annealing results. . . . . . . . . . . . . 262

C.2 The entropy in a unit vertical lag for the mosaic model and multiGaus-

sian realizations. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 265

x

List of Figures

1.1 Reference reservoir model for introductory example. . . . . . . . . . . 4

1.2 Reference distribution and variogram for introductory example. . . . 4

1.3 The fractional flow curves of the Gaussian simulations. . . . . . . . . 6

1.4 The fractional flow curves of the simulations once post-conditioned to

honor the well test response. . . . . . . . . . . . . . . . . . . . . . . . 7

1.5 Conventional and post-conditioned distributions of output uncertainty. 7

2.1 A schematic illustration of the data available for a typical reservoir

modeling exercise. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13

2.2 A schematic illustration of how stochastic simulation works. . . . . . 14

2.3 An illustration of accuracy and precision - the two criteria for a good

statistical prediction technique. . . . . . . . . . . . . . . . . . . . . . 17

2.4 An example, using eolian sandstone data, of the inadequacy of two-

point information. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 42

2.5 An example, using cross stratified sands and silty sands, of the inade-

quacy of two-point information. . . . . . . . . . . . . . . . . . . . . . 43

2.6 An example of a linear structure, well characterized by bivariate statis-

tics, and a curvilinear structure, poorly characterized by bivariate

statistics. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 45

2.7 A curvilinear training image, the areal extent of exhaustive two-point

statistics, and a simulated realization honoring the two-point statistics. 46

2.8 The distribution of permeability derived from 100 stochastic reservoir

models. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 48

2.9 A flow chart illustrating the simulated annealing algorithm. . . . . . . 55

xi

2.10 The decision rule used by MAP, threshold accepting, and simulated

annealing. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 58

3.1 A schematic illustration of the grid network system used by sasim. . 63

3.2 A general flow chart of the stochastic relaxation technique to impose

multipoint statistics (including well test derived data). . . . . . . . . 65

3.3 An example of updating a two-point covariance. . . . . . . . . . . . . 69

3.4 Examples of 1, 2, 3, 4, and 9-point configurations. . . . . . . . . . . . 73

3.5 An illustration of the indexing convention for a four-point configuration. 79

3.6 An example of the three-point configurations that need updating after

a perturbation. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 82

3.7 Square four point configuration. . . . . . . . . . . . . . . . . . . . . . 83

3.8 Three illustrations of four-point histograms. . . . . . . . . . . . . . . 84

3.9 An example, using berea sandstone data, of using partial quadrivariate

information. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 86

3.10 An example, using a difficult synthetic example, of using partial quadri-

variate information. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 87

3.11 Examples of six point configurations. . . . . . . . . . . . . . . . . . . 88

3.12 A schematic illustration of the volume measured by a given well test

interpretation. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 94

3.13 The normalized weighting function for three different instants in time. 95

3.14 A schematic illustration of the radial weight function for the power

average. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 95

3.15 A schematic illustration of a typical situation: a vertical well in the

center of a block imbedded within a sequence of such blocks. . . . . . 99

3.16 An illustration of the weight function for different dimensionless grid

block sizes. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 101

3.17 An experimentally derived weight function. . . . . . . . . . . . . . . . 102

4.1 Reference Berea image. . . . . . . . . . . . . . . . . . . . . . . . . . . 105

4.2 Histogram, summary statistics, and normal probability paper plot of

the 1600 reference Berea permeability values. . . . . . . . . . . . . . . 106

xii

4.3 Experimental normal scores semivariogram and model fit. . . . . . . . 108

4.4 Theoretical biGaussian indicator variograms and the experimental in-

dicator variograms computed from the reference Berea image. . . . . 110

4.5 Schematic illustration of the flow scenario. . . . . . . . . . . . . . . . 111

4.6 Fractional flow of oil for the reference Berea permeability distribution. 112

4.7 Four SGS realizations using the Berea permeability histogram and nor-

mal scores semivariogram. . . . . . . . . . . . . . . . . . . . . . . . . 114

4.8 A quantile-quantile plot of all 100 SGS realizations. . . . . . . . . . . 115

4.9 The normal scores semivariogram for all 100 SGS realizations in the

two principal directions. . . . . . . . . . . . . . . . . . . . . . . . . . 116

4.10 Four SIS realizations using the Berea permeability histogram and nor-

mal scores semivariogram. . . . . . . . . . . . . . . . . . . . . . . . . 118

4.11 A quantile-quantile plot of all 100 SIS realizations. . . . . . . . . . . . 119

4.12 The normal scores semivariogram for all 100 SIS realizations in the two

principal directions. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 120

4.13 The lag vectors used with annealing to generate unconditional simula-

tions of Berea. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 122

4.14 Objective function versus the number of swaps (CPU time) for the slow

annealing schedule. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 123

4.15 Objective function versus the number of swaps (CPU time) for the fast

annealing schedule. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 124

4.16 Four annealing realizations using the Berea permeability histogram and

normal scores semivariogram. . . . . . . . . . . . . . . . . . . . . . . 125

4.17 A quantile-quantile plot of all 100 annealing realizations. . . . . . . . 126

4.18 The normal scores semivariogram for all 100 annealing realizations in

the two principal directions. . . . . . . . . . . . . . . . . . . . . . . . 127

4.19 The lag vectors used with annealing to generate simulations with arti-

fact checkerboard appearance. . . . . . . . . . . . . . . . . . . . . . . 128

4.20 Two SAS realizations with four lags. . . . . . . . . . . . . . . . . . . 129

4.21 Two SAS realizations with the temperature set to zero. . . . . . . . . 130

xiii

4.22 The simulated output distributions generated by the unconditional se-

quential Gaussian realizations. . . . . . . . . . . . . . . . . . . . . . . 131

4.23 The simulated output distributions generated by the unconditional se-

quential indicator realizations. . . . . . . . . . . . . . . . . . . . . . . 132

4.24 The simulated output distributions generated by the unconditional an-

nealing realizations. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 133

4.25 The SGS realizations that yield the shortest and the longest time to

reach a water cut of 5%. . . . . . . . . . . . . . . . . . . . . . . . . . 134

4.26 The SIS realizations that yield the shortest and the longest time to

reach a water cut of 5%. . . . . . . . . . . . . . . . . . . . . . . . . . 135

4.27 The SAS realizations that yield the shortest and the longest time to

reach a water cut of 5%. . . . . . . . . . . . . . . . . . . . . . . . . . 135

4.28 Two SGS realizations before and after being transformed to the refer-

ence Berea permeability distribution. . . . . . . . . . . . . . . . . . . 140

4.29 The simulated output distributions generated by the transformed un-

conditional Gaussian realizations. . . . . . . . . . . . . . . . . . . . . 141

4.30 The simulated output distributions generated by the transformed un-

conditional indicator realizations. . . . . . . . . . . . . . . . . . . . . 142

4.31 Four sequential Gaussian realizations using the Berea permeability his-

togram, the normal scores semivariogram model, and two conditioning

data. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 147

4.32 Four sequential indicator realizations using the Berea permeability his-

togram, the normal scores semivariogram model, and two conditioning

data. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 148

4.33 Four annealing realizations using the Berea permeability histogram,

the normal scores semivariogram model, and two conditioning data. . 149

4.34 The simulated output distributions generated by the conditional Gaus-

sian realizations. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 150

4.35 The simulated output distributions generated by the conditional indi-

cator realizations. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 151

xiv

4.36 The simulated output distributions generated by the conditional an-

nealing realizations. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 152

5.1 The reference Berea image, the first SGS realization, the first SIS re-

alization, the first SAS realization and the corresponding connectivity

functions for 10 lags. . . . . . . . . . . . . . . . . . . . . . . . . . . . 156

5.2 A realization generated with the full SIS algorithm and the correspond-

ing connectivity functions for 10 lags. . . . . . . . . . . . . . . . . . . 157

5.3 The reference Berea image, the first SGS realization (before and af-

ter post-processing), the first SIS realization (before and after post-

processing), the first SAS realization (before and after post-processing),

and the corresponding connectivity functions for 10 lags. . . . . . . . 159

5.4 A post-processed SGS realization with a connectivity function close to

the reference Berea image. . . . . . . . . . . . . . . . . . . . . . . . . 160

5.5 The connectivity functions of all SAS realizations before and after post

processing. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 161

5.6 The simulated output distributions generated by the annealing realiza-

tions before and after conditioning by the connectivity function. . . . 162

5.7 An example of bivariate transformation with annealing. . . . . . . . . 167

5.8 An example of the bivariate transformation procedure applied to a

simulated realization. . . . . . . . . . . . . . . . . . . . . . . . . . . . 168

5.9 Example calibration scatterplot and a conditional pdf of z given a class

of y-values. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 170

5.10 Location map with gray-level coded porosity values. . . . . . . . . . . 172

5.11 Histograms of 74 2-D vertically-averaged well porosity values and 16900

seismic energy values. . . . . . . . . . . . . . . . . . . . . . . . . . . . 173

5.12 Normal scores semivariogram based on 74 porosity values. . . . . . . 173

5.13 Gray scale map of seismic data. . . . . . . . . . . . . . . . . . . . . . 174

5.14 Calibration scatterplot of the porosity values with the seismic data. . 174

5.15 Two sequential Gaussian realizations of the porosity. . . . . . . . . . 175

xv

5.16 The lag vectors used with annealing to condition the realizations of

porosity. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 176

5.17 The component objective functions versus the number of swaps. . . . 177

5.18 Two realizations generated with annealing to match specified lags of

the variogram and the correlation with the seismic data. . . . . . . . 178

5.19 The scatterplot of all simulated porosity values and the seismic data. 179

5.20 The original scanned image of the eolian sandstone with the upscaled

reference. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 181

5.21 Normal scores variogram of the eolian sandstone . . . . . . . . . . . . 182

5.22 Two Gaussian-based simulated realizations of the eolian sandstone. . 184

5.23 Two indicator-based simulated realizations of the eolian sandstone . . 185

5.24 Two post processed indicator-based simulated realizations of the eolian

sandstone . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 186

5.25 A scanned photograph of core from a distributary-mouth bar sequence. 187

5.26 The distribution of the reference permeability values . . . . . . . . . 188

5.27 Normal scores semivariogram and fitted model. . . . . . . . . . . . . 189

5.28 Two realizations from a Gaussian model. . . . . . . . . . . . . . . . . 190

5.29 Two realizations from the median IK/modiac model. . . . . . . . . . 193

5.30 Reference indicator variograms and their fitted models . . . . . . . . 195

5.31 Two realizations from the multiple indicator model. . . . . . . . . . . 197

5.32 Lags for two-point histogram control. . . . . . . . . . . . . . . . . . . 198

5.33 Two realizations from the annealing model. . . . . . . . . . . . . . . . 199

5.34 The distribution of effective permeabilities obtained from the Gaussian,

median IK, indicator and annealing RF models. . . . . . . . . . . . . 201

5.35 The distribution of late breakthrough times obtained from the Gaus-

sian, median IK, indicator and annealing RF models. . . . . . . . . . 202

5.36 Schematic illustration of five spot pattern used for well test example. 207

5.37 Histogram, probability plot, normal scores semivariogram, and refer-

ence gray scale image of permeability for the well test example. . . . 209

5.38 The Miller-Dyes-Hutchinson plot resulting from a well test with the

reference distribution of permeability. . . . . . . . . . . . . . . . . . . 210

xvi

5.39 Miller-Dyes-Hutchinson plots resulting from a well test in uniform per-

meability fields of 25md, 50md, and 75md. . . . . . . . . . . . . . . . 212

5.40 The first four initial realizations generated with annealing. . . . . . . 214

5.41 A q-q plot comparing the reference distribution (in the normal space)

to the distribution of all 100 realizations. . . . . . . . . . . . . . . . . 215

5.42 A plot comparing the reference variogram (in the normal space) to

variograms computed from all 100 realizations. . . . . . . . . . . . . . 215

5.43 The histogram of 100 effective permeability values from initial realiza-

tions. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 216

5.44 The mean normalized absolute deviation and mean normalized error . 217

5.45 A scatterplot of the power average approximation and the true well

test-derived effective permeabilities for all 100 initial realizations. . . 218

5.46 The annular region measured by the well test. . . . . . . . . . . . . . 219

5.47 The objective function versus time for direct annealing simulation to

match variogram and well test. . . . . . . . . . . . . . . . . . . . . . 220

5.48 Four realizations of direct annealing simulation to match variogram

and well test. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 221

5.49 The histogram of 100 effective permeability values after post processing.222

5.50 The simulated output distributions generated by the annealing real-

izations before and after conditioning to the well test-derived effective

permeability . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 224

A.1 Unprocessed output from the scanning program. . . . . . . . . . . . . 235

A.2 Scanned image filtered with a 3 by 3 pixel sum. . . . . . . . . . . . . 236

A.3 Scanned image filtered to remove most artifacts. . . . . . . . . . . . . 236

A.4 Example of an eolian sandstone . . . . . . . . . . . . . . . . . . . . . 238

A.5 Example of an wedge and tabular cross strata in an eolian sandstone 238

A.6 Example of ripple cross laminations in an eolian sandstone . . . . . . 239

A.7 Example of migrating ripples . . . . . . . . . . . . . . . . . . . . . . . 239

A.8 Example of convoluted and deformed laminations from a fluvial envi-

ronment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 240

xvii

A.9 Example of “Starved Current Ripples” from a deltaic environment . . 241

A.10 Example of large scale cross laminations from a deltaic environment . 242

A.11 Example of anastomosing mud layers around sand ripples from an es-

tuarine environment . . . . . . . . . . . . . . . . . . . . . . . . . . . 242

B.1 P-P plots for 100 realizations of three different grid sizes. . . . . . . . 245

B.2 Variograms for 100 realizations of three different grid sizes. . . . . . . 246

B.3 mAD of the Gaussian realization cdfs from the model cdfs. . . . . . . 248

B.4 mAD of the Gaussian realization variograms from the model variograms.248

B.5 P-P plots for 100 realizations of three different grid sizes. . . . . . . . 250

B.6 Variograms for 100 realizations of three different grid sizes. . . . . . . 251

B.7 mAD of the SIS realization cdfs from the model cdfs. . . . . . . . . . 252

B.8 mAD of the SIS realization variograms from the model variograms. . 252

C.1 Two realizations of a 2-D bombing model. . . . . . . . . . . . . . . . 260

C.2 The analytical variogram model and the variogram corresponding to

the two bombing model realizations. . . . . . . . . . . . . . . . . . . 261

C.3 Two realizations of a multiGaussian random function model with the

variogram of a 2-D bombing model process. . . . . . . . . . . . . . . 261

C.4 The analytical variogram model and the variogram corresponding to

the two multiGaussian realizations. . . . . . . . . . . . . . . . . . . . 262

C.5 Two realizations of a simulated annealing random function model with

the variogram of a 2-D bombing model process. . . . . . . . . . . . . 263

C.6 The analytical variogram model and the variogram corresponding to

the two annealing realizations. . . . . . . . . . . . . . . . . . . . . . . 263

C.7 Two realizations of the mosaic model generated by median indicator

simulation. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 265

C.8 The analytical variogram model and the variogram corresponding to

the two mosaic realizations. . . . . . . . . . . . . . . . . . . . . . . . 266

C.9 Two realizations of a multilognormal random function model. . . . . 266

C.10 The analytical variogram model and the variogram corresponding to

the two multilognormal realizations. . . . . . . . . . . . . . . . . . . . 267

xviii

D.1 Two commonly encountered situations when finite strings of contigu-

ously aligned data are used in kriging. . . . . . . . . . . . . . . . . . 269

D.2 An illustration of the effect of kriging with a contiguous string of values.271

D.3 The change in the kriging weights applied to a string of contiguous

values as the point being estimated gets closer to the string. . . . . . 272

D.4 The change in the kriging weights when simple kriging is used and the

range is increased from one to four units. . . . . . . . . . . . . . . . . 273

D.5 An illustration of the effect of ordinary block kriging to estimate the

local mean followed by a point simple kriging using that local mean. . 278

D.6 The change in the kriging weights applied to a string of contiguous

values as the point being estimated gets closer to the string after the

clustering due to kriging is corrected. . . . . . . . . . . . . . . . . . . 280

E.1 Example parameter file for sasimi . . . . . . . . . . . . . . . . . . . 283

E.2 Example parameter file for sasimr . . . . . . . . . . . . . . . . . . . 291

xix

Chapter 1

Introduction

Flow simulation contributes to reservoir management by providing the means to pre-

dict the reservoir response before expensive implementation of an actual recovery

process. A realistic reservoir description is necessary for flow simulators to predict

accurately future performance.

Accurate reservoir description and modeling is difficult given the considerable un-

certainty in the spatial distribution of reservoir rock properties. The petrophysical

properties needed for flow simulation and in-situ resource estimation are typically

sampled at very few locations within the reservoir. This sparse knowledge leads us to

consider a stochastic reservoir modeling approach as opposed to a single determinis-

tic model. The idea is to construct numerical models of the reservoir lithofacies and

rock properties that honor all available data (core measurements, well logs, seismic

and geological interpretations, analog outcrops, well test interpretations, . . . ). By

considering multiple realizations, each of which are consistent with the available in-

formation, the uncertainty in the spatial distribution of reservoir properties and the

reservoir response to various actions / production schemes can be quantified. Decision

analysis techniques may then be used to make risk-qualified decisions.

This is not a new concept. Stochastic models of physical systems are used ex-

tensively in many scientific disciplines. The idea of stochastic reservoir modeling has

already been considered extensively in the petroleum industry [59, 86]. The abil-

ity to evaluate alternate recovery processes and the uncertainty associated with the

1

CHAPTER 1. INTRODUCTION 2

reservoir response allows better, risk-conscious, reservoir management.

However, there is no single stochastic modeling algorithm that can create real-

izations of petrophysical properties that simultaneously honor all of the available

information. Some techniques are well suited to the modeling of discrete or cate-

gorical variables like lithofacies; others are suited to continuous variables like poros-

ity, saturation, and permeability. Unfortunately, information like complex geological

structures and effective properties derived from well tests cannot be easily integrated

into traditional stochastic models.

Certain geological patterns, created by complex interacting physical processes, are

not well characterized by the two-point variogram/covariance functions used by these

traditional stochastic models. There exists a need to describe better these complex

geological processes and to impose that description on stochastic reservoir models.

Promising descriptive techniques, involving the inference of multiple-point spatial

statistics from control patterns or training images, are proposed in this dissertation.

The methodology to impose these multiple-point statistics on stochastic reservoir

models is developed.

Another source of information, that has largely been ignored in traditional stochas-

tic reservoir modeling, is pressure transient well tests. Pressure transient well tests

measure the effective permeability of some region around the wellbore. This informa-

tion does not resolve local details of the spatial distribution of permeability; however,

it does constrain the average permeability around the wellbore. The methodology

developed to incorporate geological structures will be extended to constrain prior

stochastic images to effective permeabilities inferred from well test results.

The common denominator in the incorporation of these two disparate sources

of information is the use of stochastic relaxation or annealing techniques where the

stochastic imaging problem is formulated as an optimization problem. An overall

objective function is constructed as the sum of component objective functions. Each

component ensures that a particular source of data is honored. The optimization

problem will then be solved by annealing techniques.

The essential feature of annealing methods is to iteratively perturb (relax) an

easily constructed initial realization. The initial realization could be constructed by

CHAPTER 1. INTRODUCTION 3

randomly assigning all nodal values from a representative histogram. The perturba-

tions are then accepted or rejected with some decision rule. The decision rule is based

on how the objective function has changed, i.e., how the perturbation has brought

the candidate image closer to having the desired properties. One possible decision

rule is based on an analogy with the physical process of annealing, hence the name

simulated annealing [1, 93].

Annealing techniques tend to be computationally expensive. To ease the com-

putational burden, and to allow a more conservative assessment of uncertainty, the

techniques are applied to prior realizations of more conventional geostatistical stochas-

tic simulation techniques. The conventional techniques used are sequential Gaussian

simulation and sequential indicator simulation.

Successful incorporation of all sources of information will have a practical influence

on reservoir modeling; the predictive ability of the models will be better and a fair

assessment of uncertainty will be possible. The descriptive techniques and method-

ology developed in this dissertation may be applied in many other fields including

hydrogeology, environmental engineering, mineral resource assessment, agriculture,

forestry,. . . .

An Introductory Example

Figure 1.1 shows a schematic illustration of one quarter of a five spot injection /

production pattern. This introductory example is concerned with predicting the wa-

terflooding performance of this pattern. In particular, consider the generation of

stochastic images of absolute permeability and their impact on the predicted water-

flooding performance. The following data are available to map permeability:

• The absolute horizontal permeability at each well location: in this case, the

permeability at the injector and producer are known to be 1.52 md and 18.22

md respectively.

• The histogram of permeability values: see Figure 1.2.

CHAPTER 1. INTRODUCTION 4

injector

producer

Figure 1.1: Reference distribution of permeability and the location of the one quarter fivespot injection/production pattern.

Fre

quen

cy

Variable

.0 10.0 20.0 30.0 40.0

.000

.050

.100

.150

.200

Reference Distribution Number of Data 1000

mean 10.2121std. dev. 15.7115

coef. of var 1.5385

maximum 223.2879upper quartile 13.0391

median 5.2324lower quartile 1.1526

minimum .0000

Var

iogr

am

Distance

Reference Variogram

0. 20. 40. 60. 80. 100.

.00

.20

.40

.60

.80

1.00

1.20

Figure 1.2: The reference histogram of permeability and the reference normal score vari-ogram (EW - solid line, and NS - dashed line).

CHAPTER 1. INTRODUCTION 5

• A semivariogram model γ(h) modeling the spatial continuity of the normal score

transformed permeability values: see Figure 1.2. Note that the wells are 71 grid

units apart (the square described by the injection/production pattern is 50 grid

units by 50 grid units).

• A well test-derived effective permeability at each well location: the effective

permeability derived at the injector and producer are 0.85 md and 6.45 md

respectively.

The permeability at the well locations and the well test-derived effective permeabili-

ties were taken from a specific simulated realization considered as the reference field.

The absolute permeability at the well locations (1.52 and 18.22md) was read directly

from that reference realization. A drawdown well test was then forward simulated us-

ing Eclipse [48] at both the injecting and producing well. The two well test responses

were interpreted with standard interpretation techniques to arrive at the effective

permeabilities of 0.85 md and 6.45 md respectively for the injector and producer.

Conventional stochastic simulation techniques used to build models of permeabil-

ity can account for the first three types of data, i.e., local conditioning data, a univari-

ate distribution model, and a variogram/covariance model. The annealing technique

developed in later chapters allows the fourth data, well test-derived effective proper-

ties, to be accomodated. The impact of this well test information is demonstrated in

this introductory example by first predicting flow performance with realizations gen-

erated by a conventional method, then with the same realizations post-conditioned to

account for the well test data. Repeating the flow simulation on the post-conditioned

realizations allows a reduction in the uncertainty of the performance predictions.

The performance of the recovery process is judged by the fractional flow of oil

versus time1. Fifteen realizations of the permeability field were generated with a

conventional sequential Gaussian simulation technique (details of this technique are

given in Chapter 2). The Eclipse [48] flow simulator was then used to compute

the fractional flow of oil for every realization, see Figure 1.3. Note the bias and

considerable uncertainty in predicting the breakthrough time (the time beyond which

1The units of time are important only in a relative sense.

CHAPTER 1. INTRODUCTION 6

Fractional Flow of Oil versus Time

Time

Oil

Fra

ctio

n

0. 50. 100. 150. 200. 250.

.00

.20

.40

.60

.80

1.00

Figure 1.3: The fractional flow of oil versus time for all 15 Gaussian simulations. Thethicker dotted gray line is the response obtained from the reference distribution.

a significant fraction of water is being produced).

The same initial 15 realizations of permeability were post-processed by an anneal-

ing technique (documented in Chapters 3 and 5) to honor the two well test-derived

effective permeabilities. Repeating the flow simulation with exactly the same param-

eters (except for the spatial distribution of absolute permeability) yields the results

shown on Figure 1.4. Note the narrower spread in the fractional flow curves as com-

pared to Figure 1.3.

The nominal time to achieve breakthrough is recorded as the time at which the

fractional flow of water reaches 5% (the fractional flow of oil reaches 95%). A

histogram of this response variable for the conventional realizations and the post-

processed realizations are shown on Figure 1.5. The reference value of 13.40 time

units shown below the abscissa axis on these histograms is the result of performing

the same flow simulation on the reference image from which the conditioning data

were taken. Both distributions of uncertainty contain this reference value. The post-

processed realizations appear better centered around the reference value with less

spread. This improvement in the prediction is the direct result of integrating the

additional well test data.

CHAPTER 1. INTRODUCTION 7

Fractional Flow of Oil versus Time

Time

Oil

Fra

ctio

n

0. 50. 100. 150. 200. 250.

.00

.20

.40

.60

.80

1.00

Figure 1.4: The fractional flow of oil versus time for all 15 simulations once post-conditionedto honor the well test responses. The thicker dotted gray line is the response obtained fromthe reference distribution.

Fre

quen

cy

Time to 5% Water Cut

.0 10.0 20.0 30.0 40.0 50.0 60.0 70.0

.000

.050

.100

.150

.200Conventional Realizations

mean 30.61std. dev. 16.89

Time to 5% Water Cut

.0 10.0 20.0 30.0 40.0 50.0 60.0 70.0

.00

.10

.20

.30

.40Post-Conditioned Realizations

mean 11.76std. dev. 6.77

Figure 1.5: The distribution of the time to achieve a 5% water cut generated by the con-ventional technique (not taking into account the well test results) and the same distributiononce the well tests are taken into account. The black dot shown below the abscissa axisindicates the value obtained from the reference distribution.

CHAPTER 1. INTRODUCTION 8

Dissertation Outline

Chapter 2 presents the basic concepts underlying this research. The general problem

of reservoir modeling and the place of stochastic simulation are discussed. The criteria

to judge between alternate simulation techniques are developed: the best technique

accounts for the most input information and yet provides the largest space of output

uncertainty for a given transfer function (e.g., flow simulator) processing the input

numerical models. A presentation of the random function model-based approaches

to stochastic simulation helps to establish the notation and the place of annealing-

based algorithms. The motivation for annealing techniques arises from the inability

of conventional techniques to account for either curvilinear geological structures or

effective/average properties derived from pressure transient well testing. Chapter 2

concludes with a general presentation of the annealing approach and a more detailed

look at simulated annealing algorithms.

Chapter 3 documents a general purpose annealing-based simulation program with

the goal of integrating more geological and engineering data than conventional tech-

niques. Implementation details such as choosing the initial image, the perturbation

mechanism, the decision rule and details of the objective function are discussed. The

details of how multiple-point spatial statistics, involving more than two-points at a

time, enter into the objective function are given. An important reservoir property

inferred from pressure transient well testing is the effective absolute permeability of a

radial volume around the wellbore. The quantification of the volume and type of av-

eraging measured by pressure transient well tests is also developed. Once quantified,

that average can enter the objective function of an annealing simulation program.

Chapter 4 develops an extended example comparing stochastic simulation based

on annealing to the more conventional sequential Gaussian and sequential indicator

simulation algorithms. These various techniques are compared in terms of the input

and output space of uncertainty that they generate. The reduction in uncertainty

due to local conditioning data is investigated.

Chapter 5 presents a number of advanced applications of annealing-based sim-

ulation techniques. The first advanced application is to longstanding geostatistical

CHAPTER 1. INTRODUCTION 9

problems such as multivariate spatial transformation, conditioning to multiple-point

connectivity functions, and accounting for a secondary variable. The second area of

application is the use of multiple-point statistics to characterize better and simulate

geological structures. Once again, the improvement in the characterization is shown

in the output space of uncertainty. The final area of application documented in this

dissertation is in conditioning to well test-derived effective absolute permeabilities. A

number of cases are given that show the benefit of accounting for this information.

Finally, Chapter 6 contains conclusions and recommendations. Many practical

concerns with the annealing methodology and avenues of research were identified

during the preparation of this dissertation. Research avenues, which are beyond the

scope of this dissertation, are noted in this last chapter.

The appendices contain some in-depth information that does not properly belong

in the main text of the dissertation. Appendix A contains a detailed look at how

geological images may be scanned and coded into a useful format for training im-

ages. Appendix B presents a detailed look at ergodicity and the fluctuations from

model statistics that can be expected with conventional random function-based ap-

proaches. Appendix C presents a detailed look at spatial entropy and compares the

spatial entropy of realizations generated by a variety of simulation techniques includ-

ing annealing. Appendix D presents a a detailed look at kriging in a finite domain

and some of the problems that are encountered. Appendix E contains the documen-

tation and pseudo code for two sasim Simulated Annealing SIMulation) programs

developed for this dissertation. The sasim programs allow the direct simulation, or

the post-conditioning of previously simulated stochastic images, to honor multiple

point statistics and well test-derived properties in addition to conventional two-point

variogram/covariance functions.

Chapter 2

Stochastic Reservoir Modeling:

Concepts and Algorithms

This chapter discusses the probabilistic or stochastic approach to issues in reservoir

management. The concepts described in this chapter provide the motivation and the

background for methodology developed in later chapters.

Section 2.1 discusses the general problem of reservoir management and the place

of stochastic simulation techniques.

Section 2.2 considers how to compare different stochastic simulation techniques.

The important conclusion is that good techniques are those that directly account for

the most prior relevant information/data while exploring the largest space of output

uncertainty.

Section 2.3 presents the statistical concepts and notations for the conventional

random function approach to stochastic simulation. The notations developed in this

section are used consistently throughout the remaining text . The essential feature of

the random function approach is the priority placed on the determination of posterior

probability distribution models. The sequential Gaussian simulation (SGS) technique

and the sequential indicator simulation (SIS) technique are described in detail because

of their current popularity and common usage in later chapters.

Section 2.4 presents specific types of geological structures and engineering (well

test) data that are not accounted for by conventional simulation techniques. Based

10

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS11

on the goodness criteria established in section 2.2 there is a place for techniques that

could account for these data.

Section 2.5 discusses the “annealing” approach to the generation of stochastic

realizations. This alternative, which poses the generation of a stochastic model as an

optimization problem, has the potential to account for the geological structures and

well test data documented in section 2.4.

2.1 Reservoir Performance Forecasting

The primary objective of reservoir performance forecasting is to “predict future per-

formance of a reservoir and to find ways and means of optimizing ultimate recovery”

[11]. The idea is to model the recovery for a number of alternate production schemes.

Then, after selecting a particular recovery scheme, the predicted future performance

can be used for production planning and economic forecasting.

Exact reservoir performance forecasting would require exhaustive knowledge of

the spatial distribution of reservoir rock and fluid properties. Although a given reser-

voir, at any specific instant in time, has a single true distribution of petrophysical

properties, this distribution is unknown to those predicting future performance. The

true distribution was created by the complex interaction of many different chemical

and physical processes over geological time and would be accessible only through

exhaustive sampling. Therefore, in all practical situations the true distribution will

remain unknown.

Without complete knowledge of the reservoir properties, the exact behavior or

response of a reservoir to some future action or recovery scheme is unknown. Al-

though the real response is unknown, a numerical model can be constructed that

approximates the behavior of the real reservoir. In the past, many different kinds

of models were considered including analog1 and physical models2. Since the early

1980’s, computer models have replaced all other types for predictive purposes.

1Common analog models were based on the use of electrical potential and current as analogvariables for pressure and flow rate. Models of resistor networks or resistivity paper have beenconstructed and used to approximate the flow of fluid in porous reservoir rock.

2Scaled models using actual or synthetic reservoir rock and fluids.

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS12

Stochastic simulation and Monte Carlo methods are names used interchangeably

for methods that use a model rather than a real system with some random or un-

known component being present. There are two distinct problem areas in reservoir

performance forecasting where models are typically used:

1. Fluid flow in porous media is governed by fundamental laws based on the conser-

vation of mass, momentum, and energy. These laws and empirical relationships

such as Darcy’s law form the basis for the mathematical equations used to model

fluid flow. The formulation and numerical solution of fluid flow equations fall

under the heading of flow simulation or reservoir simulation. Flow simulation is

not “stochastic” since conventional numerical solution methods provide unique

solutions for a given set of input parameters. The underlying mathematical

equations do not acknowledge any random or unknown component. A com-

puter program for flow simulation is often referred to as a transfer function.

A transfer function is defined as a numerical model of some real operation or

system.

2. Prior to flow simulation, petrophysical properties such as the porosity, perme-

ability, and fluid saturations are needed for every grid block or element in the

flow simulation model. Given incomplete sampling there is typically a great

deal of uncertainty in the assignment of grid block properties. Building numeri-

cal models or alternative images of petrophysical properties, accounting for the

unknown aspects of the spatial distribution, is generally referred to as stochastic

reservoir modeling or stochastic imaging.

This dissertation is concerned with the second problem area, that is, how to build

better stochastic reservoir models.

A schematic illustration of a typical problem setting is shown on Figure 2.1. A

reservoir, shown in 2-D for convenience, is to be modeled using a limited number

of good quality well data, a greater number of indirect seismic data, knowledge of

the geological setting, and interpretations from a limited number of well tests. The

reservoir management problem is to assess the performance of a number of alter-

nate production scenarios. A flow simulation program provides the needed response

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS13

areal extent of reservoirwell location

Available Data- Seismic Interpretation- Well Logs, Core Measurements- Geological Interpretation- Well Test Interpretation

Figure 2.1: A schematic illustration of the data available for a typical reservoir modelingexercise. A plan view of the reservoir is shown on the left.

variables given a numerical model of the rock and fluid properties.

Given sparse sampling and uncertainty in the available data there is no unique

model of the distributions of petrophysical properties. The idea behind stochastic

reservoir modeling is to generate a number of alternate numerical models (called

realizations) that are all consistent with the known data. Running a flow simulation

program with a number of alternate numerical models allows the uncertainty in the

prediction, due to uncertainty in the input rock/fluid properties, to be appreciated.

Figure 2.2 illustrates this concept versus the reality of a single true distribution

of rock properties. The first step in a reservoir modeling exercise is to establish the

spatial distribution of rock and fluid properties (upper portion of Figure 2.2). In

reality, there is only one true distribution of these properties, yet, there are many

stochastic models of the spatial distribution, each of which is consistent with the

available data. Only three stochastic images are shown in this schematic figure; in

practice, many more (sometimes several hundreds) may be considered. Note that

each stochastic image honors the available well data.

The next step is to consider the proposed recovery scheme (central portion of

Figure 2.2). The actual recovery scheme, symbolized by the drilling rig, could be

implemented only once in the actual reservoir. A flow simulation program, symbol-

ized by the computer, provides a numerical model of the recovery scheme for each

realization.

Finally, as illustrated at the bottom of Figure 2.2, there is only one true value

for each response variable (e.g., hydrocarbon recovery, breakthrough time, flow rate,

bottom hole pressure, . . . ). Each realization potentially yields a different response

providing a probability distribution for each response variable. In practice, the true

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS14

Reality Model

Distribution of Rock/Fluid Properties

Distribution of Rock/Fluid Properties

Recovery Process

Recovery Process

Field Response Field Response

single true distribution

multiple stochastic models

actual process implemented

numerical model of process

single true response distribution of possible responses

Fre

quen

cy

Fre

quen

cy

Figure 2.2: A schematic illustration of stochastic reservoir modeling. The first step consistsof establishing the spatial distribution of rock and fluid properties. In reality, there is onlyone true distribution of these properties, yet, there can be many stochastic models of thatdistribution. The next step is the implementation of the recovery scheme. In reality, therecovery scheme may be implemented only once in the field. A flow simulator provides anumerical model of the recovery scheme using each alternate input model. Finally, there isonly one true reservoir response, but there is a distribution of possible responses given thealternate stochastic models which can be generated.

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS15

response remains unknown until it is too late to alter the recovery scheme. The

simulated distributions of response variables can be used to assess the risk involved

with any particular recovery scheme. Decision analysis techniques exist to allow

optimum risk-qualified decisions [68, 78, 124].

As mentioned earlier, this dissertation is concerned with creating better input

numerical models of reservoir properties. Before presenting the details of any one

technique it is useful to specify the properties of a good simulation technique.

2.2 Goodness Criteria for Stochastic Simulation

Techniques

There are many techniques for stochastic simulation. Given a choice between two

techniques which one should be retained? The first practical criterion is that all po-

tentially good methods must be feasible, i.e., they must generate plausible realizations

in a reasonable amount of time (both human and CPU time). When two candidate

techniques pass this first criterion the next criteria relates to the output distribu-

tion generated by the different techniques (see the bottom of Figure 2.2). As in any

statistical prediction a good technique generates an output distribution that is both

accurate and precise. An output distribution is accurate if some fixed probability

interval, say the 95% probability interval, contains the true response. An output

distribution is precise if it is as narrow as possible.

Figure 2.3 shows output distributions that are accurate but not precise, precise but

not accurate, neither accurate nor precise, and both accurate and precise. Clearly, a

good technique will generate distributions of response variables that are both accurate

and precise. Accuracy can be achieved at the expense of precision by an output

distribution with a large spread - if the 95% probability interval is very large then

it is likely to include the true value. Similarly, precision can be achieved at the

expense of accuracy - a very narrow distribution may be obtained by generating the

same response value all the time; however, any one value is unlikely to be the true

value. Accuracy is the most important - an output distribution that is precise but

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS16

not accurate is not suited to risk analysis since it gives a false sense of confidence

for a wrong prediction. An output distribution that is accurate but not precise

acknowledges uncertainty and includes the true value in the distribution. Precision

becomes a priority only after accuracy is obtained.

In practice, it is straightforward to assess the relative precision of different tech-

niques simply be measuring the spread of the output distributions. However, it is

not possible to assess accuracy since the true value is never known. A minimum

condition for accuracy is that a technique must account for all of the important input

data. The next condition to ensure accuracy is to maximize the spread of the output

distribution, i.e., forsake precision altogether, subject to the condition that all input

data are accounted for. This corresponds to the maximum entropy3 criterion adopted

by researchers in information theory. The idea is to aim for accuracy by incorporating

as much prior information as possible:

. . . the only way to set up a probability distribution that honestly repre-

sents a state of incomplete knowledge is to maximize the entropy, subject

to all the information we have. Any other distribution would necessarily

either assume information that we do not have, or contradict information

that we do have. E.T. Jaynes [71]

An important aspect of the maximum entropy approach is to make the output distri-

bution subject to all the information we have. This implies that the input stochastic

models of rock/fluid properties must be subject to all of the available information. For

example, the spread or entropy of the output distribution should not be artificially

expanded by geologically implausible realizations or those inconsistent with observed

data. An appreciation for the plausibility of a model, based on experience and an

understanding of the geological processes that created the reservoir, is yet another

piece of information that constrains the output distribution.

Another important point is that the entropy to be maximized is that of the re-

sponse (output) distributions and not that of the input realizations. The output

response variables are related to the input spatial distributions through a specific

3Entropy is another measure of the spread of a distribution, see equation (2.7).

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS17

Accurate but not Precise

true value response value

prob

abili

ty d

ensi

ty

Precise but not Accurate

true value response value

prob

abili

ty d

ensi

ty

Neither Accurate nor Precise

true value response value

prob

abili

ty d

ensi

ty

Both Accurate and Precise

true value response value

prob

abili

ty d

ensi

ty

Figure 2.3: An illustration of accuracy and precision. The vertical line in the center of eachgraph represents the true value and the shaded area represents the distribution of outcomesgenerated by a Monte Carlo technique.

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS18

transfer function (flow simulator); however, that transfer function is usually very

complex and non-linear. Even though spatial entropy of the input realizations can

be defined and predicted (see [87] and Appendix B), it is, in general, not related to

the entropy of the response or output distribution.

The spread of a response distribution is sometimes referred to as a space of uncer-

tainty. To recapitulate, the goodness criteria that can be used to compare alternate

stochastic modeling techniques are as follows:

1. A good technique must generate plausible realizations in a reasonable amount

of time. The time refers to the human and the CPU time required for the initial

set up and the repeated application of the technique.

2. A good technique is one that allows the maximum prior information to be ac-

commodated. This is the only direct way to ensure that the output distribution

is as accurate as possible.

3. Finally, a good technique is one that explores the largest space of uncertainty,

i.e., one that generates a maximum entropy distribution of response variables.

These criteria will be recalled throughout the dissertation when alternate techniques

must be assessed.

2.3 Spatial Statistics and Stochastic Simulation

The geostatistical approach to stochastic simulation is often taken as synonymous with

the application of random function models based on two-point or bivariate (covariance

or variogram) statistics. It is important to understand that the application of random

function models to geological phenomena is a matter of pure convenience. These

random function models do not represent any of the physical, chemical, or mechanistic

processes that created the true spatial distribution.

The following discussion of random variables and random functions is largely taken

from Deutsch and Journel [40]. More details and theoretical demonstrations can be

found in the following references [42, 50, 51, 90, 102, 112, 139].

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS19

The basic approach taken by predictive statistics is to model the uncertainty

about an unsampled value z as a random variable (RV) Z, the probability distribu-

tion of which characterizes the uncertainty about z. A random variable is a variable

which can take a certain number of outcome values according to some probability

(frequency) distribution. The random variable is traditionally denoted with a capital

letter, say Z, while its outcome values are denoted with the corresponding lower case,

say z. The RV model Z, and more specifically its probability distribution, is usually

location-dependent; hence the notation Z(u), with u being the coordinate location

vector. The RV Z(u) is also information-dependent in the sense that its probability

distribution changes as more data about the unsampled value z(u) becomes available.

Both continuously varying quantities such as petrophysical properties (porosity, per-

meability, saturation) and categorical variables such as rock or facies types can be

effectively modeled by RV’s.

The cumulative distribution function (cdf) of a continuous RV Z(u) is denoted:

F (u; z) = Prob {Z(u) ≤ z} (2.1)

When the cdf is made specific to a particular information set, e.g., (n) consisting of

n neighboring data values Z(uα) = z(uα), α = 1, . . . , n, the notation “conditional to

(n)” is used, defining the conditional cumulative distribution function (ccdf):

F (u; z|(n)) = Prob {Z(u) ≤ z|(n)} (2.2)

In the case of a categorical RV Z(u) that can take any one of K outcome values

k = 1, . . . , K, a similar notation is used:

F (u; k|(n)) = Prob {Z(u) ∈ category k|(n)} (2.3)

Note that since categorical variables do not necessarily have any predefined ordering

the probability distribution given above in (2.3) is a probability density function

(pdf) and not a cumulative distribution function (cdf). In many cases a naturally

continuous variable, such as permeability, will be classified into K classes by K − 1

cutoff values zk, k = 1, . . . , K − 1. That is, a categorical RV Y (u) is defined that can

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS20

take one of K outcome values k = 1, . . . , K, depending on the class k that z(u) falls

within:

z(u) ∈ (−∞, z1] ⇒ y(u) = 1,

z(u) ∈ (z1, z2] ⇒ y(u) = 2,...

z(u) ∈ (zK−1, +∞] ⇒ y(u) = K

The cumulative distribution F (u; z), as shown in expression (2.1), characterizes

the uncertainty about the unsampled value z(u) prior to using the information set (n);

the conditional cumulative distribution function, as shown in expression (2.2), char-

acterizes the posterior uncertainty once the information set (n) has been accounted

for. The goal of any predictive algorithm is to update prior models of uncertainty

such as (2.1) into posterior models such as (2.2). Note that the ccdf F (u; z|(n)) is a

function of the location u and the available conditioning data (the sample size n, the

geometric configuration (the data locations uα, α = 1, . . . , n), and the sample values

z(uα)’s).

From the ccdf (2.2) one can derive various optimal estimates for the unsampled

value z(u). The ccdf mean or expected value

m(u) = E{Z(u)} =∫ 1

0zdF (u; z|(n)) (2.4)

is a common central measure for a ccdf of a continuous variable. In the case of a

categorical variable, the mode or class(es) k′ with the largest probability,

F (u; k′|(n)) ≥ F (u; k|(n)), ∀ k = 1, . . . , K, (2.5)

is an important value.

One can also calculate measures of spread or dispersion such as the variance for a

continuous variable or Shannon’s entropy [119] for a categorical variable. The variance

is defined as:

σ2(u) = E{[Z(u) − m(u)]2} =∫ 1

0[z − m(u)]2 dF (u; z|(n)) (2.6)

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS21

Shannon’s entropy (or simply the entropy) is defined as:

s(u) = −k=K∑k=1

ln [F (u; k|(n))]F (u; k|(n)) (2.7)

In general, the variance is not applicable to categorical variables since the arbitrary

numerical coding of the categories affects the variance measure. Similarly, unless

the ccdf has an analytical expression, the entropy is not applicable to a continuous

variable since the way in which Z(u) is separated into classes will affect the entropy

measure.

In the case of a continuous variable one can also derive various probability intervals

such as the 95% interval [q(0.025);q(0.975)] such that

Prob {Z(u) ∈ [q(0.025); q(0.975)]|(n)} = 0.95,

with q(0.025) and q(0.975) being the 0.025 and 0.975 quantiles of the ccdf, e.g.,

q(0.025) is such that F (u; q(0.025)|(n)) = 0.025

Moreover, one can draw any number of simulated outcome values z(l)(u), l = 1, . . . , L,

from the ccdf. A simulated outcome z(l)(u) is drawn by generating a uniform ran-

dom number p(l) between 0 and 1 and determining the p(l)-quantile z(l)(u) such that

F (u; z(l)(u)|(n)) = p(l). Calculation of posterior ccdf’s and Monte Carlo drawings of

outcome values is at the heart of the random function approach to stochastic simu-

lation.

In stochastic reservoir modeling most of the information related to an unsampled

value z(u) comes from sample values at neighboring locations u′, whether defined on

the same attribute z or on some related attribute y. Thus, it is important to model the

degree of correlation or dependence between any number of RV’s Z(u), Z(uα), α =

1, . . . , n and more generally Z(u), Z(uα), α = 1, . . . , n, Y (u′β), β = 1, . . . , n′. The

concept of a random function (RF) allows such modeling and updating of prior cdf’s

into posterior ccdf’s.

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS22

2.3.1 The Random Function Concept

A random function (RF) is a set of RV’s defined over some field of interest, e.g.,

{Z(u),u ∈ study area A} also denoted simply as Z(u). Usually the RF definition is

restricted to RV’s related to the same attribute , say z, hence another RF would be de-

fined to model the spatial variability of a second attribute, say {Y (u),u ∈ study area}.Just as a RV Z(u) is characterized by its cdf (2.1), a RF Z(u) is characterized by

the set of all its N -variate cdf’s4 for any number N and any choice of the N locations

ui, i = 1, . . . , N within the study area A:

F (u1, . . . ,uN ; z1, . . . , zN) = Prob {Z(u1) ≤ z1, . . . , Z(uN) ≤ zN} (2.8)

Just as the univariate cdf of the RV Z(u) is used to characterize uncertainty about

the value z(u), the multivariate cdf (2.8) is used to characterize joint uncertainty

about the N values z(u1), . . . , z(uN ).

Bivariate (Two-Point) Distributions

The bivariate (N = 2) cdf of any two RV’s Z(u1), Z(u2), or more generally Z(u1),

Y (u2), is particularly important since conventional geostatistical procedures are re-

stricted to univariate (F (u; z)) and bivariate distributions:

F (u1,u2; z1, z2) = Prob {Z(u1) ≤ z1, Z(u2) ≤ z2} (2.9)

One important summary of the bivariate cdf F (u1,u2; z1, z2) is the covariance function

defined (if it exists) as:

C(u1,u2) = E {Z(u1)Z(u2)} − E {Z(u1)}E {Z(u2)} (2.10)

However, when a more complete summary is needed, the bivariate cdf F (u1,u2; z1, z2)

is more completely described by considering binary indicator transforms of Z(u) de-

fined as

I(u; z) =

1, if Z(u) ≤ z

0, otherwise(2.11)

4The scalar N used here is not to be confused with the information set which is enclosed byparentheses “(n)” or explicitly referred to as the set of conditioning information.

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS23

Then, the previous bivariate cdf (2.9) at various thresholds z1 and z2 appears as the

non-centered covariance of the indicator variables:

F (u1,u2; z1, z2) = E {I(u1; z1)I(u2; z2)} (2.12)

Relation (2.12) is the key to the indicator geostatistics formalism [77]: it shows that

inference of bivariate cdf’s can be done through sample indicator covariances.

In the case of a categorical variable one could also maintain all of the information

provided by the probability density function (pdf) representation (equivalent to the

“cdf” representation given in (2.9)). For example,

f(u1,u2; k1, k2) = Prob {Z(u1) ∈ category k1, Z(u2) ∈ category k2, } (2.13)

k1, k2 = 1, . . . , K

is the bivariate or two-point distribution of Z(u1) and Z(u2). This two-point dis-

tribution, when established from experimental proportions, is also referred to as a

two-point histogram [49].

Recall that the categorical variable Z(u), which takes K outcome values k =

1, . . . , K (see 2.3), may arise from a naturally occurring categorical variable or from

a continuous variable separated into K classes.

Multivariate (Multiple-Point) Distributions

Most practical applications of the theory of random functions do not consider multiple-

point cdfs beyond the two point cdf (2.9). The principal reason is that inference of

experimental multiple-point cdfs is usually not practical. Thus, random function

models have not been developed that explicitly account for multiple-point statistics.

However, this dissertation will present one method, based on annealing, to account

for multiple point statistics. For this reason, the following brief exposition on multiple

point cdfs and their summary statistics is proposed.

Recall that a particular N -variate or N -point cdf is written (see also 2.8):

F (u1, . . . ,uN ; z1, . . . , zN) = Prob {Z(u1) ≤ z1, . . . , Z(uN) ≤ zN}

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS24

In some applications this multiple point cdf may be summarized by considering

binary transforms of Z(u) as defined in (2.11). The multiple point cdf described

above appears as the product (non-centered covariance) of N indicator variables:

F (u1, . . . ,uN ; z1, . . . , zN) = E {I(u1; z1) · I(u2; z2) · . . . · I(uN ; zN )} (2.14)

Specific multiple-point non-centered indicator covariances have been introduced

in the geostatistics literature by Journel and Alabert [85]. In the case of Journel and

Alabert [85] the N -points were spatially separated by a fixed lag separation vector h

(i.e., u2 = u1 +h,u3 = u2 +h, . . .). A critical cutoff zc is considered and the N -point

connectivity function φ(u; N, zc) is defined as the expected value of the product of N

indicator variables:

φ(u; N, zc) = E

N∏j=1

I(u + (j − 1)h; zc)

(2.15)

this could be interpreted as the probability of having N points, starting from u1,

aligned in the direction of h being jointly below cutoff zc.

In the case of a categorical variable one could also maintain all of the information

provided by the probability density function (pdf) representation (an extension of the

two-point case presented above (2.13)). For example,

f(u1, . . . ,uN , ; k1, . . . , kN) = Prob{ Z(u1) ∈ category 1, . . . , (2.16)

Z(uN) ∈ category N }k1, . . . , kn = 1, . . . , K

is the multivariate or multiple-point distribution of Z(u1), . . . , Z(uN). This type

of multiple-point distribution, when established from experimental proportions, is

referred to as a multiple-point histogram or an N-point histogram.

2.3.2 Inference and Stationarity

The purpose of conceptualizing a RF for {Z(u),u ∈ study area A} is not to study

the case where the variable Z is completely known. If all the z(u)′s are known for

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS25

all u ∈ study area A there would be no problem nor any need for the concept of

a random function. The ultimate goal of a RF model is inference, that is, to make

some predictive statement about locations u where the outcome z(u) is unknown.

Inference of any statistic requires some repetitive sampling. For example, repeti-

tive sampling of the variable z(u) is needed to evaluate the cdf

F (u; z) = Prob {Z(u) ≤ z}

from experimental proportions. However, in many applications at most one sample

is available at any single location u in which case z(u) is known (ignoring sampling

errors), and the need to consider the RV model Z(u) vanishes. The paradigm under-

lying statistical inference processes is to trade the unavailable replication at location

u for another replication available somewhere else in space and/or time. For example,

the cdf F (u; z) may be inferred from the sampling distribution of z-samples collected

at other locations, uα �= u, within the same field, or at the same location u but at

different times if a time series is available.

This trade of replication corresponds to the hypothesis or decision of stationarity.

Stationarity is a property of the RF model, not of the underlying physical spatial

distribution. Thus, it cannot be checked from data. The decision to pool data into

statistics across rock types is not refutable a priori from data; however, it can be

shown inappropriate a posteriori if differentiation per rock type is critical to the

undergoing study. For a more extensive discussion see [69, 82].

The RF {Z(u),u ∈ A} is said to be stationary within the field A if its multivariate

cdf (2.8) is invariant under any translation of the N coordinate vectors uk, that is:

F (u1, . . . ,uN ; z1, . . . , zN ) = F (u1 + l, . . . ,un + l; z1, . . . , zn), (2.17)

∀ translation vector l.

Invariance of the multivariate cdf entails invariance of any lower order cdf, including

the univariate and bivariate cdfs, and invariance of all their moments, including all

covariances of type (2.12) or (2.10). The decision of stationarity allows inference. For

example, the unique stationary cdf

F (z) = F (u; z), ∀ u ∈ A

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS26

can be inferred from the cumulative sample histogram of the z-data values available

at various locations within A. The stationary mean (2.4) and variance (2.6) can then

be calculated from the stationary cdf F (z):

E{Z(u)} =∫ 1

0zdF (z) = m, ∀ u

E{[Z(u) − m]2} =∫ 1

0[z − m]2 dF (z) = σ2, ∀ u

The decision of stationarity also allows inference of the stationary covariance

C(h) = E {Z(u + h)Z(u)} − [E {Z(u)}]2 (2.18)

∀u,u + h ∈ A

from the sample covariance of all pairs of z-data values approximately separated by

vector h. At h = 0 the stationary covariance C(0) equals the stationary variance σ2.

In certain situations the standardized stationary correlogram is preferred:

ρ(h) =C(h)

C(0)

In other cases another second order (two-point) moment called the variogram is con-

sidered:

2γ(h) = E{[Z(u + h) − Z(u)]2

}(2.19)

∀u,u + h ∈ A

Under the decision of stationarity the covariance, correlogram, and variogram are

equivalent tools for characterizing two-point correlation:

C(h) = C(0) · ρ(h) = C(0) − γ(h) (2.20)

The decision of stationarity is critical for the appropriateness and reliability of

geostatistical simulation methods. Pooling data across geological facies may mask

important geological differences; on the other hand, splitting data into too many sub-

categories may lead to unreliable statistics based on too few data per category. The

rule in statistical inference is to pool the largest amount of relevant information to

formulate predictive statements.

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS27

Stationarity is a property of the RF model; thus, the decision of stationarity may

change if the scale of the study changes or if more data become available. If the goal

of the study is global then local details may be unimportant; conversely, the more

data available the more statistically significant differentiations become possible.

2.3.3 Kriging

Kriging was initially introduced to provide estimates for unsampled values z(u).

Within the context of stochastic simulation, however, it is used to build probabilistic

models of uncertainty about these unknown values (see (2.2) and Lesson 4 in [83]).

Essentially, the kriging algorithm provides a minimum error variance estimate of any

unsampled value based on a stationary two-point covariance (2.18). Contouring a

grid of kriging estimates is the traditional mapping application of kriging. However,

since kriging estimates are weighted moving averages of the data, their distribution

smooths out details and extreme values of the original data set. Thus, it is not an

appropriate algorithm to generate input numerical models for flow simulation since

the extreme permeability values, i.e., flow paths and barriers are critical to the fluid

flow response [63].

The minimum error variance property of kriging allows it to approximate, and

in some cases identify, the conditional expectation of the variable being estimated.

Thus, kriging can be used to determine a series of posterior conditional probability

distributions from which stochastic images of the attribute spatial distribution can

be drawn. This aspect of kriging which is key to the random function approach

to stochastic simulation, will be discussed in the next section. The following is a

summary of kriging. More details and theoretical proofs can be found in [22, 29, 54,

65, 69, 83, 90, 98].

All versions of kriging are elaborations on the same basic linear regression algo-

rithm and corresponding estimator:

[Z∗SK(u) − m(u)] =

n∑α=1

λα(u) [Z(uα) − m(uα)] (2.21)

where Z(u) is the RV model at location u, the uα’s are the n data locations, m(u)

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS28

= E{Z(u)} is the expected value of RV Z(u), and Z∗SK(u) is the linear regression

estimator, also called the “simple kriging” (SK) estimator.

The SK weights λα(u) are given by the system of normal equations written in

their general non-stationary form as follows:

n∑β=1

λβ(u)C(uβ,uα) = C(u,uα), α = 1, . . . , n (2.22)

The SK algorithm requires prior knowledge of the (n+1) means m(u), m(uα), α =

1, . . . , n , and the (n + 1) by (n + 1) covariance matrix [C(uα,uβ), α, β = 0, 1, . . . , n]

with u0 = u. In most practical situations, inference of these means and covariance

values requires a prior decision of stationarity of the random function Z(u) (see the

discussion in the previous section). If the RF Z(u) is stationary with constant mean

m, and covariance function C(h) = C(u,u + h), ∀u, the SK estimator reduces to its

stationary version:

Z∗SK(u) =

n∑α=1

λα(u)Z(uα) +

[1 −

n∑α=1

λα(u)

]m (2.23)

where the weights λα(u) are established to minimize the error variance or estimation

variance:

σ2E(u) = E

{[Z(u) − Z∗

SK(u)]2}

(2.24)

Solution of the traditional stationary SK system:

n∑β=1

λβ(u)C(uβ − uα) = C(u− uα), α = 1, . . . , n (2.25)

provides the weights λα(u) such that the estimator (2.23) minimizes the estimation

variance (2.24). The minimized estimation variance or kriging variance is given by:

σ2SK(u) = C(0) −

n∑α=1

λα(u)C(u − uα) (2.26)

where, C(0) is the stationary variance σ2.

According to strict stationary theory, SK is the algorithm that should be applied;

however, stationary SK does not adapt to local trends in the data since it relies on the

mean value m assumed known and constant throughout the area. Consequently, when

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS29

enough data are available to identify trends, the more robust ordinary kriging (OK)

algorithm is used. Ordinary kriging (OK) filters the mean from the SK estimator

(2.23) by requiring that the kriging weights sum to one. This results in the following

ordinary kriging (OK) estimator:

Z∗OK(u) =

n∑α=1

να(u)Z(uα) (2.27)

and the stationary OK system:

∑n

β=1 νβ(u)C(uβ − uα) + µ(u) = C(u − uα), α = 1, . . . , n∑nβ=1 νβ(u) = 1

(2.28)

where the να(u)’s are the OK weights and µ(u) is the Lagrange parameter associated

to the constraint∑n

β=1 νβ(u) = 1. Comparing expression (2.25) and (2.28), note that

the SK weights are different from the OK weights.

2.3.4 Models of Uncertainty

At the beginning of the previous section it was noted that the primary usefulness of

kriging, in the context of stochastic simulation, is in the determination of posterior

distribution models. There are two characteristic properties of kriging that are the

basis for, respectively, the multiGaussian (MG) approach and the indicator kriging

(IK) approach to the determination of posterior ccdf’s:

(i) - The multiGaussian Approach: If the RF model Z(u) is multivariate nor-

mal or Gaussian5, then the simple kriging estimate (2.23) and variance (2.26)

identify the mean and variance of the posterior ccdf. In addition, since that

ccdf is Gaussian, it is fully determined by these two parameters, see [9] and

[134]. This rather simple result is the basis for multiGaussian (MG) kriging

and simulation. The MG approach is said to be parametric in the sense that it

determines the ccdf’s through their parameters (mean and variance).

5Since one cannot expect all sample histograms to be normal, a normal score-transform is per-formed on the original z-data. A multiGaussian model Y (u) is then adopted for the normal scoredata. Kriging and simulation are performed on the y-data with the results appropriately back-transformed into z-values.

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS30

(ii) - The Indicator Kriging Approach: If the value to be estimated is the ex-

pected value (mean) of a distribution, then least squares (LS) regression, i.e.,

kriging, is a priori the preferred algorithm. The reason is that the LS estimator

of the variable Z(u) is also its conditional expectation E{Z(u)|(n)}, i.e., the

mean of the ccdf (2.2) (see [90], p. 566). Now, instead of the variable Z(u), con-

sider its binary indicator transform I(u; z) as defined in relation (2.11). Kriging

of the indicator RV I(u; z) provides an estimate which is also the LS estimate

of the conditional expectation of I(u; z). Moreover, the conditional expectation

of I(u; z) is itself equal to the ccdf of Z(u); indeed:

E {I(u; z)|(n)} = 1 · Prob {I(u; z) = 1|(n)}+ 0 · Prob {I(u; z) = 0|(n)}

= 1 · Prob {Z(u) ≤ z)|(n)} ≡ F (u; z|(n)), as defined in (2.2)

Thus, the kriging algorithm applied to indicator data provides LS estimates of

the ccdf (2.2). Note that indicator kriging (IK) is not aimed at estimating the

unsampled value z(u) or its indicator transform i(u; z) but at providing a ccdf

model of uncertainty about z(u). The IK algorithm is said to be non-parametric

in the sense that it does not approach the ccdf through its parameters, rather,

the ccdf values for various threshold values z are estimated directly.

Given these two algorithms for updating prior models of uncertainty (2.1) to

posterior models of uncertainty (2.2) it is possible to discuss an important family of

stochastic simulation techniques.

2.3.5 The Sequential Approach to Stochastic Simulation

Consider a random function (RF) defined over some field of interest, e.g.,

{Z(u),u ∈ study area A}

Stochastic simulation is the process of building alternate, equally probable models

of the spatial distribution of z(u); each realization is denoted with the superscript l:

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS31

{z(l)(u),u ∈ A}. The simulation is said to be “conditional” if the resulting realizations

honor the hard data values at their locations:

z(l)(uα) = z(uα), ∀ l, α = 1, . . . , n (2.29)

The simulated realizations are built such that they honor aspects of the multivariate

probability distribution of the RF chosen to model Z(u). In all cases, except for the

multiGaussian RF model, those aspects of the multivariate distribution are limited

to moments of its bivariate distribution.

Consider the joint distribution of N random variables Zi with N very large. The

N RV’s Zi may represent a single attribute, say permeability, at the N nodes of a

dense grid discretizing the field A, or they can represent N attributes measured at

the same location, or they could represent a combination of K different attributes

defined at the N ′ nodes of a grid with N = KN ′.

Next, consider the conditioning of these N RV’s by a set of n data of any type

symbolized by the notation |(n). The corresponding N -variate ccdf is denoted:

F(N)(z1, . . . , zN |(n)) = Prob{Zi ≤ zi, i = 1, . . . , N |(n)} (2.30)

Expression (2.30) is completely general with no intrinsic limitations; some or all of

the variables Zi could be categorical.

Successive application of Bayes’ relation shows that drawing an N -variate sample

from the ccdf (2.30) can be done in N successive steps, each involving a univariate

ccdf with increasing levels of conditioning:

• draw a value z(l)1 from the univariate ccdf of Z1 given the original data (n). The

value z(l)1 is now considered as a conditioning datum for all subsequent drawings;

thus, the information set (n) is updated to (n + 1) = (n) ∪ {Z1 = z(l)1 }.

• draw a value z(l)2 from the univariate ccdf of Z2 given the updated data set

(n + 1), then update the information set to (n + 2) = (n + 1) ∪ {Z2 = z(l)2 }.

• sequentially consider all N RV’s Zi.

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS32

The set {z(l)i , i = 1, . . . , N} represents a simulated joint realization of the N dependent

RV’s Zi. If another realization is needed, {z(l′)i , i = 1, . . . , N}, the entire sequential

drawing process is repeated.

This sequential simulation procedure requires the determination of N univariate

ccdf’s, more precisely:

Prob{Z1 ≤ z1|(n)} (2.31)

Prob{Z2 ≤ z2|(n + 1)}Prob{Z3 ≤ z3|(n + 2)}

. . .

P rob{ZN ≤ zN |(n + N − 1)}

The sequential simulation principle is the same regardless of the algorithm or model

used to establish the sequence (2.31) of univariate ccdf’s. In the multiGaussian ap-

proach all ccdf’s (2.31) are assumed Gaussian and their means and variances are

given by a series of N simple kriging systems (see the previous section 2.3.4). In the

indicator kriging approach, the ccdf’s are obtained directly by indicator kriging (see

the previous section 2.3.4).

Implementation Considerations:

• Strict application of the sequential simulation principle calls for the determina-

tion of more and more complex ccdfs, in the sense that the size of the condi-

tioning data set increases from (n) to (n + N − 1). In practice, the argument is

that the closer6 data screens the influence of more remote data; therefore, only

the closest data are retained to condition any of the N ccdfs (2.31).

• The neighborhood limitation of the conditioning data entails that statistical

properties of the (N +n) set of RV’s will be reproduced only up to the maximum

distance found in the neighborhood7. For example, the search must be at least

6“Closer” is not necessarily taken in terms of Euclidean distance, particularly if the original dataset (n) and the N RV’s include different attribute values. The data “closest” to each particular RVZi being simulated are those that have the most influence on its ccdf.

7Larger scale conditioning could be imparted from local conditioning data [99]

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS33

as large as the distance to which the covariance is to be reproduced; this requires

extensive conditioning as the sequence progresses from 1 to N . One solution

is provided by the multiple grid concept which is to simulate the N nodal

values in two or more steps [55]: First, a coarse grid, say every tenth node is

simulated using a large data neighborhood. The large neighborhood allows the

reproduction of large scale covariance structures. Second, the remaining nodes

are simulated with a smaller neighborhood.

• Theory does not specify the sequence in which the N nodes should be simulated;

practice has shown that it is better to use a random sequence [68].

2.3.6 Sequential Gaussian Simulation (SGS)

The most straightforward algorithm for generating realizations of a multivariate Gaus-

sian field is provided by the sequential principle described above. Each variable is

simulated sequentially according to its normal ccdf fully characterized through a SK

system of type (2.25). The conditioning data consist of all original data and all previ-

ously simulated values found within a neighborhood of the location being simulated.

The conditional simulation of a continuous variable z(u) modeled by a Gaussian-

related stationary RF Z(u) proceeds as follows:

1. Determine the univariate cdf FZ(z) representative of the entire study area and

not only of the available z-data. Declustering may be needed if the z-data are

preferentially located [37, 40, 77].

2. Perform the normal score transform of the z-data, with the FZ(z) cdf, into

y-data with a standard normal cdf [40, 90].

3. Although not a part of the algorithm, it is good practice to check for bivariate

normality of the normal score y-data by comparing the experimental indicator

variograms to the ones expected from multiGaussian theory (see [40, 138]). If

the data do not show a bivariate Gaussian behavior, then alternate models

such as a mixture of Gaussian populations [141] or an indicator-based approach

should be considered.

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS34

4. If a multivariate Gaussian RF model is adopted for the y-variable then proceed

with sequential simulation, i.e.,

• define a random path that visits each node of the grid (not necessarily

regular) once. At each node u, retain a specified number of neighboring

conditioning data including both original y data and previously simulated

grid node y-values.

• use SK to determine the parameters (mean and variance) of the ccdf of

the RF Y (u) at location u.

• draw a simulated value y(l)(u) from that ccdf

• add the simulated value y(l)(u) to the data set

• proceed to the next node, and loop until all nodes are simulated.

5. Backtransform the simulated normal values {y(l)(u),u ∈ A} into simulated

values for the original variable {z(l)(u) = ϕ−1(y(l)(u)),u ∈ A}. Within-class

interpolations and tail extrapolations are usually called for, see [40] for details.

If multiple realizations are desired {z(l)(u),u ∈ A}, l = 1, . . . , L, the previous

algorithm is repeated L times with a different random path for each realization.

The prior decision of stationarity requires that simple kriging (SK) with zero mean

(for Y (u)) be used in step 4 of the SGS algorithm. However, if there are enough data

to indicate that a non-stationary RF model would be more appropriate, one may

• either split the area into distinct sub-zones and consider for each sub-zone a dif-

ferent RF model, which implies inference of a different normal score covariance

for each sub-zone,

• or consider a stationary normal score covariance, inferred from the entire pool of

data, and a non-stationary mean for Y (u). The non-stationary mean, E{Y (u)},at each location u, is implicitly re-estimated from the neighborhood data through

ordinary kriging (OK), see section 2.3.3. Locally rescaling the model mean usu-

ally results in a poorer reproduction of the stationary Y -covariance model.

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS35

A number of implementations of the sequential Gaussian simulation (SGS) algo-

rithm exist. The program gsim3d.c written by Isaaks [68] and the sgsim program of

GSLIB [40] are two published versions.

2.3.7 Sequential Indicator Simulation (SIS)

Rather than adopt a multiGaussian RF model one could consider a non-parametric

indicator kriging approach as described in section 2.3.4. For details and the theoretical

development of the sequential indicator simulation methodology see [8, 55, 83, 89,

115]. In this approach, kriging is applied to binary indicator transforms of the data

to directly estimate conditional probabilities:

• if the indicator variable being kriged arises from a categorical variable, i.e., i(u)

set to 1 if the location u belongs to category k, to zero otherwise, then:

Prob{I(u) = 1|(n)} = E{I(u)|(n)} (2.32)

• if the variable z(u) to be simulated is continuous, its ccdf can also be written

as an indicator conditional expectation:

Prob{Z(u) ≤ z|(n)} = E{I(u; z)|(n)} (2.33)

with I(u; z) = 1 if Z(u) ≤ z, =0 otherwise.

In both cases, the problem of evaluating the conditional probability is mapped

onto that of evaluating the conditional expectation of a specific indicator RV. The

evaluation of a conditional expectation calls for well-established regression theory,

i.e., kriging (see section 2.3.3).

The sequential simulation algorithm proceeds somewhat differently for categorical

and continuous variables:

(i) - Categorical Variables: At each node u along the random path, indicator krig-

ing followed by order relation corrections8 provides K estimated probabilities

8Order relation corrections amount to ensuring that the estimated distribution follows the axiomsof a probability distribution: a cdf is never less than 0, greater than 1, and must be non-decreasing.The probabilities of a pdf must all be greater (or equal to) zero and sum to one.

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS36

p∗k(u|(·)). The conditioning information (·) consists of both the original ik-data

and the previously simulated ik-values for category k.

Next define any ordering of the K categories, say 1,2,. . . ,K. This ordering

defines a cdf-type scaling of the probability interval [0, 1] with K intervals, say:

[0, p∗1(·)], (p∗1(·), p∗2(·) + p∗1(·)], . . . , (1 −K−1∑k=1

p∗k(·), 1]

Draw a random number p uniformly distributed in [0, 1]. The interval in which

p falls determines the simulated category at location u.

Update all K indicator data sets with this new simulated information, and

proceed to the next location u′ along the random path.

The arbitrary ordering of the K probabilities p∗k(·) does not affect which category

is drawn nor the spatial distribution of categories [8].

(ii) - Continuous Variables: The continuous variable z(u) discretized into K mu-

tually exclusive classes k : (zk−1, zk], k = 1, . . . , K can be interpreted and simu-

lated as the spatial distribution of K class indicators.

One advantage of considering the continuous variable z(u) as a paving (mosaic)

of K classes is the flexibility to model the spatial distribution of each class by a

different indicator covariance. For example, in the absence of any facies infor-

mation, the class of highest permeability corresponding to a complex network of

fractures, may be modeled by a zonal anisotropic indicator covariance with the

maximum direction of continuity in the fracture direction; while the geometry

of the classes of low-to-median permeability values could be modeled by more

isotropic indicator covariances. The indicator formalism allows for modeling

mixtures of populations loosely9 defined as classes of values of a continuous

attribute z(u) [30, 80].

9Note that major heterogeneities characterized by actual categorical variables, such as lithofaciestypes, should be dealt with (simulated) first, e.g., through categorical indicator simulation or consid-ering a Boolean process (the simulation of objects with a predefined shape). However, there are casesand/or scales where the only property recorded is a continuous variable such as a porosity value oracoustic log; yet experience tells us that this continuous variable is measured across heterogeneouspopulations. In this case the indicator formalism allows a “loose” separation of populations throughdiscretization of the range of the continuous attribute measured [80].

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS37

At each node u to be simulated along the random path indicator kriging (SK

or OK) provides a ccdf through K probability estimates:

F ∗(u; zk|(n)) = Prob∗{Z(u) ≤ z|(n)}, k = 1, . . . , K

Within-class interpolation, see [40], provides the continuum for all threshold

values z ∈ [zmin, zmax].

Monte-Carlo simulation of a realization z(l)(u) is obtained by drawing a uniform

random number p(l) ∈ [0, 1] and retrieving the ccdf p(l)-quantile:

z(l)(u) = F ∗−1(u; p(l)|(n)) (2.34)

such that : F ∗(u; z(l)(u)|(n)) = p(l)

The indicator data set (for all cutoffs zk) is updated with the simulated value

z(l)(u), and indicator kriging is performed at the next location u′ along the

random path.

Once all locations u have been simulated, a stochastic image {z(l)(u), u ∈ A} is

obtained. The entire sequential simulation process with a new random path can be

repeated to obtain another independent realization {z(l′)(u),u ∈ A}, l′ �= l.

The indicator approach is well suited to categorical variables which are naturally

expressed as binary indicators. The indicator approach is not as natural for continuous

variables because of the need for within class interpolations. However, there are a

number of advantages of the indicator formalism:

• The major advantage of the indicator kriging approach to generating posterior

conditional distributions (ccdf’s) is its ability to account for soft data. As long as

the soft or fuzzy data can be coded into prior local probability values, indicator

kriging can be used to integrate that information into a posterior probability

value [6, 45, 81, 92].

The prior information can take one of the following forms:

– local hard indicator data i(uα; z) originating from local hard data z(uα):

i(uα; z) = 1 if z(uα) ≤ z, = 0 if not (2.35)

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS38

or ik(uα) = 1 if uα ∈ category k, = 0 if not

– local hard indicator data j(uα; z) originating from ancillary information

that provides hard inequality constraints on the local value z(uα). If

z(uα) ∈ (aα, bα], then:

j(uα; z) =

0 if z ≤ aα

undefined (missing) if z ∈ (aα, bα]

1 if z > bα

(2.36)

– local soft indicator data y(uα; z) originating from ancillary information

providing prior (pre-posterior) probabilities about the value z(uα):

y(uα; z) = Prob{Z(uα) ≤ z| local information} (2.37)

∈ [0, 1]

�= F (z) : global prior as defined below

However, usually E{Y (u; z)} = F (z).

– global prior information common to all locations u within the stationary

area A:

F (z) = Prob{Z(u) ≤ z}, ∀ u ∈ A (2.38)

At any location u ∈ A, prior information about the value z(u) is characterized

by any one of the four previous prior probability distributions. The IK process

consists of a Bayesian updating of the local prior cdf into a posterior cdf using

information supplied by neighboring local prior cdf’s [17, 92, 140].

• A second major advantage of the indicator formalism over most other ap-

proaches to estimation and simulation is the possibility of accounting for soft

structural information. That is, the bivariate distribution can be more fully

specified by data rather than summarized by a single covariance. Inferring

different indicator covariance models at each threshold value allows a better

representation of the spatial correlation; however, the ability to incorporate

soft information may be retained while only retaining one indicator covariance

(see the median IK approach documented below).

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS39

The Median IK Approach

In indicator kriging the K cutoff values zk are usually chosen so that the corresponding

indicator covariances CI(h; zk) are different from one another. However, there are

cases when the sample indicator covariances/variograms appear proportional to each

other, i.e., the sample indicator correlograms are all similar. The corresponding

continuous RF model Z(u) is the so-called “mosaic” model [79] such that:

ρZ(h) = ρI(h; zk) = ρI(h; zk; zk′), ∀zk, zk′ (2.39)

where ρZ(h) and ρI(h; zk) are the correlograms and cross-correlograms of the contin-

uous RF Z(u) and its indicator transforms.

Then, the single correlogram function is better estimated either directly from the

sample Z-correlogram or from the sample indicator correlogram at the median cutoff

zk = M , such that F (M) = 0.5. Indeed, at the median cutoff, the indicator data are

evenly distributed as 0 and 1 values with, by definition, no extreme values.

Indicator kriging under the model (2.39) is called “median indicator kriging” [77].

It is a particularly simple and fast procedure since it calls for a single easy-to-infer

median indicator variogram which is used for all K cutoffs. Moreover, if the indicator

data configuration is the same for all cutoffs10, one single IK system needs to be solved

with the resulting weights being used for all cutoffs.

The median IK algorithm applied in conjunction with sequential indicator simu-

lation is used extensively in later chapters.

2.4 Geological Structures and Well Test Data

The sequential Gaussian simulation (SGS) and sequential indicator simulation (SIS)

algorithms are state-of-the-art simulation methods [55, 56, 68, 86, 85] that have con-

tributed significantly to reservoir modeling and risk-qualified decision making. Nev-

ertheless, these methods and all their enhancements [92, 128, 140] are limited to cases

10Unless inequality constraint-type data z(uα) ∈ (aα, bα] are considered, the indicator data con-figuration is the same for all cutoffs zk’s as long as the same data locations uα’s are retained for allcutoffs.

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS40

when the spatial continuity is characterized by stationary two-point statistics and to

data that is defined on the same support11.

This section documents two important cases when conventional geostatistical tech-

niques fail to account for important prior information:

1. Most common geological structures have curvilinear directions of continuity

and multiple-point connectivity that are difficult to account for with the con-

ventional RF “covariance-based” techniques. In many cases this information is

crucial to yield geologically realistic stochastic models of reservoir properties.

2. Pressure transient well testing provides critical information about the reservoir

properties at a significantly larger scale than can be measured by core or well

logging tools. Specifically, the average absolute permeability inferred from a

well test constrains the spatial distribution of permeability near the well bore,

i.e., the stochastic reservoir models.

The next two sections (2.4.1 and 2.4.2) illustrate these types of data and show the

inadequacy of conventional stochastic simulation techniques.

2.4.1 Geological Structures

Conventional stochastic simulation methods rely on a prior decision of stationarity

and two-point covariance functions C(h) to describe the spatial correlation within

each stationary zone A. The decision of stationarity implies that the major direc-

tion of continuity is the same throughout A, recall the definition of the stationarity

covariance (2.18):

C(h) = E {Z(u + h)Z(u)} − [E {Z(u)}]2 (2.40)

∀u,u + h ∈ A

Thus, the specific locations of two points separated by h is not information retained

in the stationary covariance (2.40). Any curvilinear patterns of continuity, or the

11Data of different supports may only be considered if the averaging is linear which is not the casefor permeability.

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS41

presence of subregions with different directions of continuity, within the stationary

zone A are represented with an average covariance showing only the average direction

of continuity12.

For example, a cross section through an eolian sandstone is shown on the top

of Figure 2.4 (Appendix A documents the procedure used to acquire the geological

images used as examples here and in later chapters). The simulated realization,

generated with the sequential Gaussian simulation (see section 2.3.6), reproduces the

two-point indicator covariances; however, the curvilinear features are not reproduced

due to the decision of stationarity and the restriction to two-point statistics.

A second example of cross stratified sands and silty sands from a distributary bar

deltaic environment is shown on Figure 2.5. The simulated realization was created by

a full indicator simulation (see section 2.3.7) with seven cutoffs (at all the seven gray

scale thresholds). Once again, the multiple-point connectedness and the curvilinear

features are not characterized by stationary two-point statistics.

As a last example, consider the two control patterns shown at the top of Figure 2.6;

both are the result of Boolean simulations whereby predefined shapes are located at

random. The left control pattern shows linear structures and the right image shows

curvilinear structures. All stationary two-point histograms were retained for 25 lags

in 2 directions and used to condition two simulations of each control pattern (the

methodology to accomplish this will be discussed in section 2.5 and in later chapters).

The two inadequacies of stationary two-point statistics are illustrated by this simple

example:

1. The reproduction of the linear control pattern is quite good. The anisotropy is

well characterized; however, the fact that all of the anisotropic dark gray rect-

angles have exactly an 8:1 anisotropy (i.e., the rectangles were all constructed

1 unit thick and 8 units long) is not captured by two-point statistics. Multiple

point connectivity, even if linear, is not captured by two-point statistics.

2. The reproduction of the curvilinear control pattern is poor. Stationary two-

point statistics are inadequate for capturing any continuity that is not linear;

12In certain cases it is possible to consider curvilinear coordinate systems to avoid the problemsintroduced by a single stationary zone [28, 41, 121].

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS42

Eolian Sandstone Image

Gaussian Simulation

Figure 2.4: The upper image shows a cross section through an eolian sandstone. The lowerimage was created by Gaussian simulation (see section 2.3.6) with the exhaustive normalscores variogram of the upper image.

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS43

Cross Stratified Sands and Silty Sands

Indicator Simulation

Figure 2.5: The upper image shows an image of cross stratified sands and silty sands.The lower image was created with two-point information in the form of complete indicatorvariogram models (see the sequential indicator simulation described in section 2.3.7.

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS44

at least trivariate statistics are required for curvilinear features.

In certain cases, non-linear features are essential for a realistic representation

of the spatial distribution. There are two evident ways that information on these

features could be accounted for: 1) the decision of stationarity could be relaxed

by separating the stationary area A into subzones each with different directions of

continuity [41, 121], and 2) multiple-point statistics could be considered to more fully

describe the multivariate distribution of Z(u) within A.

Relaxing the decision of stationarity to make up for the inadequacy of two-point

statistics has been somewhat successful when locally stationary subregions can be

established. Many geological features are non-linear even at a small scale which

implies that the subregions will be too small to allow inference of the needed statistics

or anisotropy directions. In this case, multiple-point statistics are called for to more

completely define the multivariate distribution.

A Comment on Exhaustive Two Point Statistics

It can be shown that exhaustive bivariate (two point) statistics contain enough in-

formation to reproduce a finite image exactly [49]. This is interesting but inconse-

quential; in fact, without assuming stationarity univariate statistics also allow exact

reproduction of an image13.

The reproduction of an image with bivariate statistics is possible because the

outermost points are established first, then the second outermost, . . . , until the central

point is finally established. In stochastic simulation, the point is not to reproduce

exactly the training image but to extract its essence with limited statistics.

Bivariate statistics available for every possible lag vector h over a specific areal

extent B, smaller than the training image A, do not necessarily capture the features

smaller than B. For example, if there were curvilinear features smaller than B or

there was a multiple-point connectivity, still less than B, then exhaustive two-point

statistics for all vectors h ∈ B are not sufficient. For example, consider the training

image on the left of Figure 2.7 where the circular shapes are less than four pixels in

13The histogram at each location is exactly the value specified by the training image. The valuesare not pooled into a histogram representative of a larger volume.

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS45

Linear Control Pattern Curvilinear Control Pattern

Simulation One Simulation One

Simulation Two Simulation Two

Figure 2.6: An example of a linear structure, well characterized by bivariate statistics,and a curvilinear structure, poorly characterized by bivariate statistics. Two simulationsof each are shown below the control patterns. All four simulations reproduce the two pointstatistics for 20 lags in two directions.

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS46

Curvilinear Training Image

Subarea B forControl Statistics

Simulated Realization

Figure 2.7: A curvilinear training image, the areal extent of exhaustive two-point statistics,and a simulated realization honoring the two-point statistics.

diameter. Reproducing all two point statistics for a subarea B four by four pixels

does not reproduce the circular shapes.

There is a need for simulation techniques that could directly account for multiple-

point statistics and, indirectly, the curvilinear features described above. There is also

a need for techniques to create realizations conditional to information available at a

significantly larger scale, e.g., that measured by pressure transient well testing.

2.4.2 Well Test Data

A pressure transient well test is conducted by altering the production conditions

(flow rate) and monitoring the reservoir pressure response [66]. Different reservoir

properties may be interpreted from the pressure response:

• The effective permeability or conductivity (permeability-thickness product) near

the well is an important parameter needed to predict future performance and

design well spacing.

• The reservoir pressure which is needed to predict the in-situ resource and an

appropriate recovery scheme for future exploitation.

• In many cases the reservoir limits may be predicted. These limits are essential

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS47

to establish the in-situ resource.

• In some cases it is possible to identify the presence of faults, fluid fronts, strati-

graphic interfaces, and other barriers.

The latter three pieces of information typically enter the reservoir model determin-

istically. That is, the reservoir pressure, limits, and major faults are all directly ac-

counted for in the stochastic models or when setting up the flow simulator. The first

piece of information, which informs the spatial distribution of absolute permeability,

could be used directly in a homogeneous or layer-cake reservoir model14. However,

it can not be accounted for by any conventional geostatistical/statistical technique

that constructs heterogeneous models (see section 2.2) of the reservoir properties (in

particular permeability).

The permeability inferred from a well test carries important information because

all other measurements are on a scale many orders of magnitude smaller than the

grid blocks typically used in flow simulators. It could be argued that well test, or

engineering data in general, are the most important because they are intimately

related to the recovery process and the response variables that will form the basis for

reservoir management decisions.

For example, an actual well test provides a single true effective permeability near

the well at which the test was conducted. If the same well test is numerically simulated

on stochastic models, that are conditional only to core measurements and the global

histogram of permeability values, the test results may not match the single known

permeability value. This is illustrated by the histogram of well test-derived effective

permeability values which is shown on Figure 2.8. These values were all derived

by numerically simulating a well test on 100 multiGaussian simulated realizations

that honored the permeability at the well location, the univariate distribution, and

the stationary covariance C(h) of the permeability values (the details of this study

are given in chapter 3). Ideally, if all the information were accounted for, the 100

simulated values should be all equal to the single value inferred from the well test

performed in the field (1.73 md).

14The permeanbility of the homogeneous model or the appropriate layers could be set equal tothe well test-derived value.

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS48

Fre

quen

cy

Well Test-Derived Permeability, md

.00 1.00 2.00 3.00 4.00 5.00 6.00 7.00

.000

.040

.080

.120

mean 2.56std. dev. 1.12

coef. of var .44

maximum 6.44upper quartile 3.43

median 2.25lower quartile 1.79

minimum .52

Figure 2.8: The distribution of permeability derived from well tests performed on 100stochastic reservoir models.

2.5 Annealing Techniques for Stochastic Simula-

tion

In the “annealing” approach to stochastic simulation there is no explicit random

function model, rather, the creation of a simulated realization is formulated as an

optimization problem to be solved with a stochastic relaxation or “annealing” tech-

nique. The first requirement of this class of methods is an objective (or energy)

function which is some measure of difference between the desired spatial characteris-

tics and those of a candidate realization. The essential feature of stochastic relaxation

methods is to perturb iteratively (relax) the candidate realization and then accept or

reject the perturbation with some decision rule. The decision rule is based on how

much the perturbation has brought the candidate image closer to having the desired

properties. One possible decision rule is based on an analogy with the physical process

of annealing, hence the name simulated annealing15. Annealing is the process where

15Technically the name “simulated annealing” applies only to those stochastic relaxation methodsbased strictly on simulated annealing [1, 93]; however, through common usage the name “anneal-ing” is used to describe the entire family of methods that are based on the principle of stochastic

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS49

a metallic alloy is heated without leaving the solid phase so that molecules may move

positions relative to one another and reorder themselves into a low energy crystal (or

grain) structure. The probability that any two molecules will swap locations is known

to follow the Boltzmann distribution. Simulated annealing is the application of the

annealing mechanism of perturbation and the Boltzmann probability distribution for

accepting perturbations.

At first glance this approach appears terribly inefficient. For example, if every

nodal value z(l)(ui), i = 1, . . . , N in the candidate image is initially assigned by ran-

domly drawing from the stationary cdf F (z) then millions of perturbations may be

required to arrive at an image that has the desired spatial structure. In practice,

these methods are more efficient than they might seem since remarkably few arith-

metic operations are required to update the objective function after a perturbation;

virtually all conventional global spatial statistics (e.g., a covariance) may be updated

locally rather than globally recalculated after a local perturbation. The real interest

in annealing techniques stems from a hope that they will be able to integrate better

complex geological structures and engineering data. .

The conventional stochastic simulation techniques documented in section 2.3 are

limited to two-point statistics. Annealing is not limited to two-point statistics. The

objective function is defined as some measure of difference between a set of reference

properties and the corresponding properties of a candidate realization. The refer-

ence properties could consist of any quantified geological, statistical, or engineering

property.

In many applications the reference properties/statistics are derived from a full

valued array of values referred to as a training image or control pattern. The concept

is that these training images will come from outcrop mapping, zones with a significant

amount of conditioning data, or perhaps from the output of programs that simulate

geological processes from first principles [131]. An advantage of using a training image

is that the reference statistics are, by definition, consistent in that there exists at least

one realization that reproduces them all. If a suitable training image is unavailable

the reference statistics may be inferred from sparsely sampled data or analog data.

relaxation.

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS50

Annealing does not require a training image; the method requires reference properties

or statistics regardless of their origin.

To illustrate the application of the annealing methodology, consider a categorical

variable Z(u) that can take any one of K outcome values k = 1, . . . , K. The stochastic

simulation problem considered here is to generate a spatial distribution for Z(u) at N

grid node locations: z(ui), i = 1, . . . , N . To be useful each simulated realization must

share certain spatial statistics with those inferred from actual data. As an example

statistic, consider the stationary16 two-point histogram, see also (2.13):

f(h; k1, k2) = Prob {Z(u) ∈ category k1, Z(u + h) ∈ category k2, } (2.41)

k1, k2 = 1, . . . , K

Considering the two-point histogram for a pre-defined number of lags h1, . . . ,hL

an objective function could be constructed as [49]:

O =L∑

l=1

K∑k1=1

K∑k2=1

[f(hl; k1, k2)reference − f(hl; k1, k2)realization

]2(2.42)

In Chapter 3 it will be shown that this objective function can be made to include

more complex components such as multiple-point statistics and engineering data.

The general annealing methodology is as follows:

1. Establish the reference components in the objective function, e.g.,

f(hl; k1, k2)reference, l = 1, . . . , L, k1 = 1, . . . , K, k2 = 1, . . . , K

2. Generate an easily constructed initial realization17 z(ui), i = 1, . . . , N .

3. Compute the realization components in the objective function, e.g.,

f(hl; k1, k2)realization, l = 1, . . . , L, k1 = 1, . . . , K, k2 = 1, . . . , K

16Adopting a stationary model amounts to assuming that the two-point histogram depends onlyon the separation vector h between the two points u1 and u2.

17One straightforward way of constructing an initial realization would be to draw the values atrandom from the stationary univariate F (z) distribution.

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS51

4. Compute the objective function O based on the reference and the realization

statistics, see equation (2.42).

5. Perturb the realization z(ui), i = 1, . . . , N by some simple mechanism to gener-

ate a new realization z′(ui), i = 1, . . . , N . One common perturbation mechanism

is to swap the value at two locations ui, uj , i �= j, 0 < i, j ≤ N . Another alter-

native is to simply change the value at one node location, i.e., reset z(ui) to a

different category knew different from the original k.

6. Update all components in the objective function and recompute the total ob-

jective function O′.

7. The perturbation is either accepted or rejected based on a specified decision

rule. One approach would be to accept all helpful perturbations O′ ≤ O and to

reject all disruptive perturbations O′ > O. This would correspond to a steepest

descent approach. Accepting the perturbation causes the image (z(ui), i =

1, . . . , N) and the objective function O to be updated.

8. If the objective function O is close to zero then the realization is finished since

it now honors the reference characteristics; otherwise, return to step 5 and

continue the perturbation process.

Given this introduction it is interesting to consider annealing in light of criteria

for good simulation techniques, (see section 2.2):

1. Regarding its feasibility, there is some evidence that annealing techniques gen-

erate realizations in a practical amount of CPU time. However, there are no

full field scale case studies to confirm this.

2. Regarding its ability to incorporate more prior data, there is hope that more

prior information can be accounted for with annealing techniques. This would

be a significant contribution.

3. Regarding its output space of uncertainty, there is no evidence to suggest that

realizations generated by annealing techniques will fully explore the uncertainty

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS52

prevailing beyond the input data; since annealing exactly honors control statis-

tics one would a priori expect a smaller space of uncertainty than that generated

by techniques that allow ergodic fluctuations. This possible disadvantage is mit-

igated by the fact that more input data can be accounted for.

There is a need to understand better annealing techniques and to provide a better

assessment of when it qualifies as a good simulation technique. The remaining chapters

in this dissertation will provide some insight as to when annealing is appropriate.

There are many different algorithms that fall within this general annealing ap-

proach. These algorithms differ in the way the initial image is chosen, the compo-

nents that enter the objective function, the perturbation mechanism, or the type of

decision rule that is adopted. Two of the more common choices, known as simulated

annealing, and the maximum a posteriori (MAP) approach, have become reasonably

well known and are documented below. A general program to consider many other

permutations will be developed in Chapter 3.

2.5.1 Simulated Annealing

The central idea behind simulated annealing is an analogy with thermodynamics,

specifically with the way liquids freeze and crystallize, or metals cool and anneal.

At high temperatures the molecules can move freely. As the temperature is slowly

lowered the molecules line up in crystals which represent the minimum energy state

for the system. The Boltzmann probability distribution, P{E} ∼ e−EkbT , expresses the

idea that a system in thermal equilibrium at a temperature T has the energies of its

component molecules probabilistically distributed among all different energy states

E. The Boltzmann constant kb is a natural constant which relates temperature to

energy. Even at a low temperature there is a probability that the energy is quite

high, in other words, a system will sometimes give up a low energy state in favor of

a higher energy state which may lead to a global minimum energy state [110].

Metropolis and his coworkers [105] extended these principles to simulate numer-

ically how molecules behave. A system will change from a configuration of energy

E1 to a configuration of energy E2 with probability p = e−(E2−E1)

kbT . The system will

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS53

always change if E2 is less than E1 (i.e., a favorable step will always be taken), and

sometimes an unfavorable step is taken: this has come to be known as the Metropolis

algorithm. More generally, any optimization procedure that draws upon the thermo-

dynamic analogy of annealing is known as simulated annealing.

In the early 1980’s Kirkpatrick et al. [93] and independently Cerny [133] ex-

tended these concepts to combinatorial optimization, i.e., they formulated an anal-

ogy between the objective function and the free energy of a thermodynamical system

[1, 110]. A control parameter, analogous to temperature, is used to control the iter-

ative optimization algorithm until a state with a low objective function (energy) is

reached.

One of the first direct applications to spatial phenomena was published by Ge-

man and Geman [52] who applied the method to the restoration of degraded images.

About the same time Rothman [117] at Stanford applied the method to nonlinear

inversion and residual statics estimation in geophysics. Independent research by C.L.

Farmer [49] led to the publication of a simulated annealing algorithm for the genera-

tion of rock type models. This triggered considerable interest in the method among

geostatisticians [38, 44, 125].

The essential contribution of simulated annealing is a prescription for when to

accept or reject a given perturbation. The acceptance probability distribution is

given by:

P{accept} =

1, if Onew ≤ Oold

eOold−Onew

t , otherwise(2.43)

All favorable perturbations (Onew ≤ Oold) are accepted and all unfavorable pertur-

bations are accepted with an exponential probability distribution. The parameter t

of the exponential distribution is analogous to the “temperature” in annealing. The

higher the temperature the more likely an unfavorable perturbation will be accepted.

The temperature t must not be lowered too fast or else the image may get trapped

in a sub-optimal situation and never converge. However, if lowered too slowly then

convergence may be unnecessarily slow. The specification of how to lower the tem-

perature t is known as the “annealing schedule”. There are mathematically based

annealing schedules that guarantee convergence [1, 52]; however, they are much too

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS54

slow for a practical application. The following empirical annealing schedule is one

practical alternative [49, 110].

The idea is to start with an initially high temperature t0 and lower it by some

multiplicative factor λ whenever enough perturbations have been accepted (Kaccept)

or too many have been tried (Kmax). The algorithm is stopped when efforts to lower

the objective function become sufficiently discouraging. The following parameters

describe this annealing schedule (see also the chart of Figure 2.9):

t0: the initial temperature.

λ: the reduction factor 0 < λ < 1.

Kmax: the maximum number of attempted perturbations at any one temperature (on

the order of 100 times the number of nodes). The temperature is multiplied by

λ whenever Kmax is reached.

Kaccept: the acceptance target. After Kaccept perturbations are accepted the temperature

is multiplied by λ (on the order of 10 times the number of nodes).

S: the stopping number. If Kmax is reached S times then the algorithm is stopped

(usually set at 2 or 3).

∆O : a low objective function indicating convergence.

The objective function could be established to reproduce traditional geostatistical

constraints. That is, a spatial distribution that honors a given histogram, variogram

model, and the data values at their locations [83, 76].

The work of C.L. Farmer [49] consisted of a direct application of simulated an-

nealing to the simulation of integer coded rock types. A two-point histogram (2.13)

and/or a correlation function for some specified number of lags and directions enter

the objective function to control the simulation. A flowchart of the algorithm is shown

on Figure 2.9.

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS55

Generate an image with the correctproportion of each rock type

Compute the Objective Function: O(old) Iend = 0, c = c0

Naccept = 0, Ntry = 0

Ntry = Ntry + 1Choose i and j at random(not conditioning nodes)

Swap i and j to form new imageCompute new Objective Function: O(new)

O(new) < O(old) - c ln(u) u = random[0,1]

?

O(new) < convergence limit ?

Ntry = K(max)?

Naccept = K(accept)?

No

No

No

No

Yes

Yes

Yes

Yes

Iend = Iend + 1

Naccept = Naccept + 1Update Image

O(old) = O(new)

write resultsand END

Iend > S?

c = Lambda c

No

Yes

Figure 2.9: A flow chart illustrating the simulated annealing algorithm.

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS56

Features of Farmer’s Algorithm (Simulated Annealing):

• The proportion of each rock type is honored at the beginning by randomly

assigning rock types at all nodes according to their respective proportions.

• Conditioning data values are honored by fixing them in the initial image and

never perturbing them in the subsequent optimization.

• Virtually all of the computational effort is taken to update the objective func-

tion; the time needed to choose two locations at random is negligible. Therefore,

it is important to code this part of the algorithm with a minimum number of

arithmetic operations. Many spatial statistics are easily updated (this will be

demonstrated in Chapter 3).

• Different realizations, i.e., conditional simulations, are achieved by starting with

different random images.

2.5.2 The Maximum A Posteriori (MAP) Variant

Bayesian classification schemes provide a variant to the simulated annealing algorithm

[10, 18, 44]. The method described below for generating rock type images is due to the

work by Geman and Geman [52], Besag [18], and others. The method was recently

applied in geostatistics by Doyen [44].

The same basic stochastic relaxation or annealing algorithm is followed. The

characteristic feature of the MAP algorithm is a different prescription for when to

accept or reject a given perturbation. The acceptance probability distribution is given

by:

P{accept} =

1, if Onew ≤ Oold

0, otherwise(2.44)

The relaxation algorithm is continued until there are no further changes in the objec-

tive function. The resulting image is called a Maximum A Posteriori (MAP) model.

One implication of the MAP decision rule is that the image can not jump out of local

minima, i.e., the methodology corresponds to a steepest descent-type approach.

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS57

An interesting aspect of the MAP algorithm applied by Doyen [44] was the in-

corporation of seismic information. A two part objective function was considered;

the first part was the deviation of the simulated rock types from those indicated by

a seismic impedance profile and the second part was the deviation of the two-point

histogram from that inferred from a control pattern. This two-part objective function

yields images that are simultaneously constrained by seismic data and the statistical

properties of the reservoir.

Threshold Accepting (TA)

A third published stochastic relaxation technique is a method called threshold accept-

ing (TA) [46]. Threshold accepting differs from annealing and the MAP approaches

only in the decision rule to accept or not an unfavorable perturbation. An unfavor-

able perturbation will be accepted if the change in the objective function is less than

a specified threshold.

P{accept} =

1, if Onew − Oold ≤ threshold

0, otherwise(2.45)

As the simulation proceeds the threshold is lowered in much the same way as the

temperature parameter t is lowered in simulated annealing.

Figure 2.10 graphically illustrates the decision rule for the MAP, threshold accept-

ing (TA), and simulated annealing (SA) approaches. A perturbation that decreases

the objective function would show to the left of the vertical axes and is accepted in

all three cases. A perturbation that increases the objective function would show to

the right of the axes:

• An unfavorable perturbation is never accepted with the MAP algorithm.

• Threshold accepting will accept a perturbation as long as the increase in the

objective function is below a specified threshold. This threshold decreases as

the simulation exercise proceeds.

• Simulated annealing will accept a perturbation with an exponential distribution.

The probability of accepting an unfavorable swap decreases as the simulation

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS58

Maximum A Posteriori (MAP)

0O_new - O_old

11

Threshold Accepting (TA)

0O_new - O_old

Decreasing Acceptance Threshold

11

Simulated Annealing (SA)

0O_new - O_old

Decreasing Temperature11

Figure 2.10: The acceptance probability distributions for the MAP, TA, and SA ap-proaches: a favorable transition (to the left of the vertical axis) is always accepted. Anunfavorable transition is never accepted with the MAP criterion. The TA criterion speci-fies that all unfavorable transitions are accepted below a certain threshold (the thresholddecreases to zero as the optimization proceeds). An exponential or Boltzmann distributionwith a decreasing parameter is considered for simulated annealing (SA).

CHAPTER 2. STOCHASTIC RESERVOIR MODELING: CONCEPTS AND ALGORITHMS59

exercise proceeds, i.e., as the temperature is decreased.

Many different variants of the same basic annealing algorithm are possible by con-

sidering different objective functions, ways to create the initial image, perturbation

mechanisms, and decision rules. A general purpose annealing simulation program

which allows simulated annealing, MAP, threshold accepting, and many other per-

mutations is documented in the next chapter. The motivation for such a program is

to explore more fully the potential of the method and evaluate the circumstances in

which it will establish itself as a good technique (see section 2.2).

Chapter 3

Application of Annealing

Techniques

This chapter discusses the elements of an annealing-based stochastic simulation pro-

gram sasim that accounts for conventional covariance/variogram models, geological

structures (through multiple-point statistics), and well test-derived effective perme-

abilities (through weighted power averages).

Section 3.1 presents the framework of the annealing simulation program sasim.

True simulated annealing, the MAP alternative, and other variants of the general

annealing algorithm are integrated into this single program. Choosing an initial

realization, the perturbation mechanism, and the decision rule to accept or reject a

perturbation are discussed.

Section 3.2 documents the criteria for a quantity to enter the objective function.

The most important consideration is that after each local perturbation the numerical

quantity must be updated locally rather than globally recalculated. Special con-

siderations for handling multiple components in a global objective function are also

documented in this section.

Section 3.3 considers multiple-point (including two-point) statistics and how they

enter into annealing-based simulation. The initial calculation and local updating of

multiple-point indicator covariances and multiple-point histograms is documented.

Section 3.4 presents the quantification of well test-derived effective permeability

60

CHAPTER 3. APPLICATION OF ANNEALING TECHNIQUES 61

in a form that may be locally updated. A non-linear power average is used as a

substitute for the full well test response. This power average, weighted by radial

distance from the well bore, is easy to update locally and captures most of the full

well test response. The theoretical basis for the radially decreasing weight function

and the procedure to calibrate the averaging power are documented

Although the principle of annealing applied to stochastic simulation is straight-

forward, the detailed implementation considerations documented in this chapter are

very important for the successful application of the technique. Recall that “success-

ful” is measured by the three criteria described in section 2.2: the technique must

generate plausible realizations in a practical amount of time, it must account for a

maximum amount of prior information, and it should explore the largest space of

output uncertainty for any given transfer function.

3.1 Framework for a General Purpose Annealing

Simulation Program

The motivation behind an annealing-based simulation program is to consider the

different possible implementations of annealing and evaluate the place of such tech-

niques. The central thesis of this dissertation is that annealing may be used to con-

dition stochastic models to geological structures and well test information that are

beyond the ability of conventional techniques. An efficient general purpose annealing

program is required to address both the place of annealing and to demonstrate how

additional information may be accounted for in stochastic models.

Recall that the concept behind annealing techniques is to formulate the simula-

tion exercise as an optimization problem. The first step is to construct an objective

function which measures the difference between reference spatial features and those

of a candidate realization. The optimization problem then consists of systematically

modifying an easily constructed initial realization so that all parts of the objective

function are lowered to zero. As discussed in section 2.5, this difficult optimization

problem may be addressed with annealing-based techniques. The problems involved

CHAPTER 3. APPLICATION OF ANNEALING TECHNIQUES 62

with the calculation, updating, and interaction between various components of the

objective function will be addressed in the next section (3.2). The subtleties of these

topics are easier to appreciate once details of the solution methods have been pre-

sented. The remainder of this section presents the framework of a general solution

method that allows any hybrid combination of the common annealing approaches

documented in section 2.5.

Annealing based techniques could apply to the simulation of continuous or cate-

gorical attributes Z at any predefined number of locations, ui, i = 1, . . . , N , within

some area of interest A. There is no intrinsic limitation that the N points be on a

regular grid; however, many of the spatial features which enter the objective function,

in particular most multiple-point statistics, are simpler to compute when the z-values

are on a regular grid. For this reason, regular grid networks are considered through-

out this dissertation. Note that the grid network need not be full valued, i.e., the

reservoir or area of interest A, may be defined (or clipped) by any irregular surface.

The coordinate system is established by specifying the coordinates at the first

block (xmn, ymn, zmn), the number of grid nodes (nx, ny, nz), and the spacing of

the grid nodes (xsiz, ysiz, zsiz). In many cases the nodal values refer to the entire

volume surrounding the node, i.e., to a block centered at the node location. Figure 3.1

illustrates these parameters on two 2-D sectional views. The following conventions

are used to define the grid network:

• The X axis is associated to the east direction. Grid node indices ix increase

from 1 to nx in the positive x direction, i.e., to the east.

• The Y axis is associated to the north direction. Grid node indices iy increase

from 1 to ny in the positive y direction, i.e., to the north.

• The Z axis is associated to the elevation. Grid node indices iz increase from 1

to nz in the positive z direction, i.e., upward.

These three axes can be associated to any coordinate system that is appropriate for

the problem at hand. For example, if the phenomenon being studied is within global

stratigraphic boundaries, then some type of stratigraphic coordinates relative to a

marker horizon would be relevant [75, 89].

CHAPTER 3. APPLICATION OF ANNEALING TECHNIQUES 63

PLAN VIEW CROSS SECTION

Easting (X axis)

Easting (X axis)N

ort

hin

g (

Y a

xis)

Ele

vati

on

(Z

axi

s)

xmnxsiz

ymn

ysiz

zmn

zsiz

ix = 1 2 ... nx

iy =

12

...ny

iz =

12

...nz

Figure 3.1: Plan and vertical cross section views to illustrate the grid definition used inGSLIB.

Depending on the application, a continuous variable Z ∈ (zmin, zmax] or a cate-

gorical variable Z = k, k = 1, . . . , K may be appropriate. As mentioned in section

2.3 a naturally continuous variable could be classified into a categorical variable.

The program sasim works with both continuous and categorical variables but certain

statistics are suited to one and not the other, e.g., a direct z-covariance or variogram

is appropriate for a continuous variable but not for categorical variables, whereas

an n-point histogram is appropriate for categorical variables but not for continuous

variables.

The general annealing program sasim is written in ANSI standard Fortran 77.

Fortran was retained as the programming language because of its familiarity, common

usage, and to maintain consistency with the programming standards of GSLIB [40].

Adhering to a standard and common programming language makes the code portable

to a wide variety of computers.

A schematic flowchart for sasim is shown on Figure 3.2. The principle behind

annealing methods is to perturb an image (realization) and accept or reject the per-

turbation based on some decision rule. sasim can be applied to generate realizations

from initially uncorrelated images (easy to create), or it can be used to post-process

CHAPTER 3. APPLICATION OF ANNEALING TECHNIQUES 64

prior realizations obtained from a conventional technique or even realizations gener-

ated by a previous sasim run. These two applications are possible within the same

program with different approaches to the perturbation mechanism and decision rule.

The following discussion of the steps in sasim, as illustrated on Figure 3.2, will clarify

many implementation details (the source code and some example runs are contained

in Appendix D):

Read Parameters: All input parameters are read from a prepared parameter file

rather than from a graphical user interface (GUI). This maintains machine inde-

pendence, simplifies multiple executions of the program, and facilitates record-

keeping.

Establish Control Statistics: The control or reference statistics involve two-point

covariances, multiple-point indicator covariances, multiple-point histograms . . . .

In principle, there is no limit to the number of statistics that can enter the

objective function (this is discussed further in section 3.2). The control values

for each statistic are established in one of three ways (not necessarily the same

for each statistic):

1. The control values may be explicitly entered into the program, e.g., as

experimental covariance values that have been smoothed or as numerical

values specified by an analytical model.

2. A second alternative is to use a control pattern or training image to com-

pute directly the needed control statistics. Note that a different control

image may be used for each statistic. sasim will compute the required

statistics from each input training image.

3. Certain control statistics may also be computed from the starting realiza-

tion of the simulation exercise. This is particularly useful when annealing

is applied to integrate another source of data by post-processing prior sim-

ulated realizations.

Establish Initial Image: An initial image is needed as a starting realization for

each simulation exercise. The initial image may be established in one of three

CHAPTER 3. APPLICATION OF ANNEALING TECHNIQUES 65

yes

no

no

yes

no

yes

Figure 3.2: A general flow chart of the stochastic relaxation technique to impose high orderstatistics (including well test derived data).

CHAPTER 3. APPLICATION OF ANNEALING TECHNIQUES 66

ways:

1. A different realization can be generated for each new simulation with

each nodal value drawn at random from the stationary cdf F (z). This

is straightforward and very fast.

2. The same input image, read directly from an input file, could be used as

the starting point for all realizations. Depending on the application this

may artificially narrow the output space of uncertainty.

3. Different initial images, read from an input file, could be used (e.g., various

sequential indicator simulation realizations could be taken and modified

with annealing to integrate additional prior information). These images

could be obtained from any source including output from a previous sasim

run.

Compute Initial Image Statistics: As discussed in the next section it is impor-

tant to update rather than recompute image statistics; however, an initial global

calculation is required before updating can take place. Note that the initial im-

age statistics may be both “control” values and “realization” values.

Compute Initial Objective Function: At this point all “control” and “realiza-

tion” values for each component of the global objective function are known.

Each component of the global objective function is calculated and weighted so

that all components may be lowered to zero during the optimization/simulation.

Details of the components entering the objective function and their weighting

scheme are given in section 3.2.

Objective Function Low Enough?: When the objective function has been low-

ered enough, or if attempts to improve the image are sufficiently discouraging,

the realization will be written to the output file. At that point the program will

begin the next simulation if more simulations are called for, or it will stop.

Perturb Image, Update Statistics and Objective: One of the following three

perturbation mechanisms is performed a specified number of times:

CHAPTER 3. APPLICATION OF ANNEALING TECHNIQUES 67

1. The values at two nodal locations ui and uj can be swapped.

2. The value at a node location ui is replaced by a random drawing from the

global stationary cdf F (z).

3. The value at a node location ui is replaced by a random drawing from a

local conditional ccdf F (ui; z) (this approach is documented by Srivastava

in [125]). The ccdf F (ui; z) is arrived at by a local weighted average or

indicator kriging. The local kriging weights are required input to sasim.

The decision rule is applied after a given number of perturbations (set by the

user). After each perturbation all components in the objective function, that

can be updated locally, are updated. All components that require a global

recalculation are updated just before the decision rule is applied.

Decision to Accept Perturbation: The acceptance criteria of simulated anneal-

ing, threshold accepting, or MAP simulation may be applied. That is, all per-

turbations (or series of perturbations) that decrease the objective function are

accepted and those that increase the objective function are accepted with the

following probability:

• Simulated Annealing (SA): P{accept} = eOold−Onew

t with the temperature

parameter t controlled by either a default annealing schedule or one preset

by the user.

• Maximum A Posteriori (MAP): P{accept} = 0 whenever the objective

function has increased.

• Threshold Accepting (TA): P{accept} = 1 if Oold − Onew < threshold ; 0,

otherwise. As the simulation proceeds the threshold is decreased with a

user defined schedule.

Update Image: The image, the image statistics, and the image objective function

are updated if the perturbation is accepted.

Update Decision Rule: In the case of simulated annealing and threshold accept-

ing the decision rule is updated after every perturbation (whether accepted or

CHAPTER 3. APPLICATION OF ANNEALING TECHNIQUES 68

not). In many cases the decision rule remains unchanged. In certain cases, this

updating will set the current objective function to zero to force the algorithm

to stop perturbing the image and move to the next simulation.

Local conditioning data are honored by fixing them in all prior realizations and

leaving then unperturbed:

z(l)(uα) = z(uα), ∀l, α = 1, . . . , n

Although conditioning to local data is very straightforward, experience has shown

that honoring local conditioning statistics may require special consideration in the

objective function. The issue of conditioning is discussed in more detail in Chapter

4.

3.2 The Objective Function

Annealing techniques rely on many, often millions, of perturbations to achieve a final

acceptable realization. The implications are that each component of the objective

function must be reasonably simple and there should not be too many components nor

conflicting components. In general, it will be essential to replace a global recalculation

of each component with a local updating; otherwise, the required computer time may

become excessive.

It is often possible to replace global recalculations with local updating since most

spatial statistics interact with only a limited number of other elements in the system.

For example, consider the variogram for a specific lag h:

γ(h) =1

2N(h)

N(h)∑i=1

[z(u) − z(u + h)]2 (3.1)

The number of pairs N(h) contributing to the variogram depends on the size of the

grid and the separation vector h. For example, vector h as a unit distance xsiz in

the x-direction of a square nx = 200 by ny = 200 grid. A global calculation of

γ(h) would require an evaluation of N(h) = (nx − 1) · ny = 39800 pairs of values.

However, when any one value is perturbed only two pairs in γ(h) must be updated.

CHAPTER 3. APPLICATION OF ANNEALING TECHNIQUES 69

Separation Vector: h

h

Small Grid

Unchanged grid node location

Perturbed grid node location

+ h- h

Figure 3.3: An example of the two pairs that require updating after perturbing a grid nodelocation. Only the two lags (+h and -h) require updating regardless of how large the gridnetwork.

This is illustrated on Figure 3.3 with a small 5 by 5 grid. An important feature of

the variogram (or the covariance/correlogram) is that regardless of how large the grid

network (or the number of pairs N(h)) there are only two lags that must be updated.

When the value at one node location z(u) is perturbed to to z′(u) the variogram is

updated to γ(h)new from γ(h)old by:

γ(h)new = γ(h)old − [z(u) − z(u + h)]2

− [z(u − h) − z(u)]2

+ [z′(u) − z(u + h)]2

+ [z(u − h) − z′(u)]2

(3.2)

A recalculation of γ(h) would require 39800 pairs to be considered whereas updating

only requires 2 pairs.

This example illustrates that an intelligent coding of the components in the objec-

tive function is essential. Regardless of how efficiently the objective function is coded,

the large number of perturbations required by annealing techniques cause them to be

CPU intensive. Therefore, from the standpoint of CPU time it is desirable to restrict

the number and complexity of components entering the objective function. However,

CHAPTER 3. APPLICATION OF ANNEALING TECHNIQUES 70

this may be contrary to the goal of integrating a maximum amount of prior informa-

tion. Given a slow technique and a fast technique that are equally good (see section

2.2), the fast technique is clearly preferred1. A very slow technique may be preferred

if it accounts for more prior information or it allows a more complete sampling of the

space of uncertainty. One hour of workstation CPU time to generate a realization,

versus one second, would be acceptable only if the realization generated in an hour is

significantly better. Given the exponential increase in computer speed over the past

few years and projected future increases there is no reason to limit artificially the

development of annealing programs to too simple objective functions.

In general, the overall objective function O is made up of the weighted sum of C

components:

O =C∑

c=1

wcOc (3.3)

where wc and Oc are the weights and component objective functions respectively. The

component objective functions measure how certain features of the simulated image

differ from the desired control or reference properties. For example, one component

could be a measure of difference between a variogram function modeled from actual

data and the experimental variogram function of the realization, a second component

could measure reproduction of a particular multiple-point statistic intended to reflect

the short scale structure (see section 3.3), and a third component could measure the

fidelity to a well test-derived effective property (see section 3.4).

Each component objective function Oc could be expressed in widely different

units of measurement. For example, a component measuring variogram departure

(Oc1 =∑[

γreference − γrealization

]2) may be in units of variance squared, while a

component measuring the fidelity to a well test-derived effective permeability (Oc2 =

[Kreference − Krealization]2) may be in units of millidarcies squared. Equally weighting

components with widely different units will cause the component with the largest

units to dominate the global objective function.

The purpose behind the weights wc, c = 1, . . . , C, is to have each component c play

1The speed of a technique is the time required to set it up plus the actual CPU time needed togenerate the realizations.

CHAPTER 3. APPLICATION OF ANNEALING TECHNIQUES 71

an equally important role in the global objective function2. Note that the magnitude

of the objective function only indirectly affects the decision to accept or reject a

perturbation: all decision rules are based on the change to the objective function,

i.e.,

∆O = Onew − Oold

∆O =C∑

c=1

wc∆Oc =C∑

c=1

wc [Ocnew − Ocold]

The weights wc, c = 1, . . . , C are established so that, in average, each component con-

tributes equally to the change in the objective function ∆O. That is, each weight wc

is inversely proportional to the average change of that component objective function:

wc =1¯|∆Oc|

, c = 1, . . . , C (3.4)

In practice, the average change of each component |∆Oc| can not be computed

analytically; however, it can be numerically approximated by evaluating the average

change of M (say 1000) independent perturbations:

|∆Oc| =1

M

M∑m=1

|O(m)c − Oc|, c = 1, . . . , C (3.5)

where |∆Oc| is the average change for component c, O(m)c is the perturbed objective

value, and Oc is the initial objective value. Each of the M perturbations m = 1, . . . , M

arises from the perturbation mechanism that will be employed for the annealing

simulation.

The overall objective function used in sasim is finally written as,

O =1

O(0)·

C∑c=1

wc · Oc (3.6)

where the weights wc given in equation (3.4) are computed at the beginning of the

simulation (annealing) procedure. The objective function O is restandardized by

its initial value, O(0), so that it always starts at 1.0 and a standardized annealing

schedule3 may be used.

A number of considerations regarding the objective function should be noted:

2The program sasim also allows the user to change arbitrarily the relative importance of eachcomponent.

3The initial temperature t0 is also set to 1.0 for simulated annealing.

CHAPTER 3. APPLICATION OF ANNEALING TECHNIQUES 72

• The gradient or rate of decrease of each component is not accounted for in the

weights (3.4). One component may quickly go to zero while another could take

much longer. Consequently, the weights may need to be re-evaluated periodi-

cally during the simulation.

• In addition to weighting each component on the basis of its average change

(|∆Oc|) one could also consider the relative importance of each component. For

example, it may be more desirable to have a good variogram match than repro-

ducing partial quadrivariate information obtained from a questionable training

image. Unequal weighting is particularly useful in cases where the objective

function cannot be lowered close to zero, e.g., due to conflicting components.

sasim allows the user to define arbitrarily the relative weights of each compo-

nent.

• Another question is whether the components should receive the same weight for

the entire duration of the optimization procedure. For example, at the begin-

ning it may be advantageous to establish the coarse features by preferentially

weighting two-point statistics, then, at the end multiple-point statistics and well

test results could be given more weight to establish the small scale details. This

has not been considered in sasim and could be a subject of future research.

Annealing techniques make no explicit reference to a random function (RF) model;

however, the very act of generating multiple spatial realizations defines an implicit

multivariate distribution or RF model. This implicit RF model depends on the com-

ponents entering the objective function, and on other implementation decisions such

as the initial image, the perturbation mechanism, and the decision rule. An effort

will be made in the case studies of later chapters to understand better the implicit

RF model underlying annealing techniques.

The following two sections describe how multiple-point statistics and well test-

derived effective permeability may be expressed as component objective functions.

CHAPTER 3. APPLICATION OF ANNEALING TECHNIQUES 73

One-Point Configuration

Two-Point Configuration

Three-Point Configuration

Four-Point Configuration

Nine-Point Configuration

Figure 3.4: Examples of 1, 2, 3, 4, and 9-point configurations. Each gray square representsa grid node location. The black dot in the center of a grid node location represents the nullvector.

3.3 Multiple-Point Statistics

There are few published references [85] and little practical experience with multiple-

point spatial statistics beyond the bivariate or two-point level; intuition about which

multiple-point statistics are appropriate to characterize any particular geological fea-

ture will develop with time. In section 2.4, multiple-point statistics were considered

necessary to model many commonly encountered geological features where the pre-

dominant characteristics are curvilinear and connected beyond the two-point level.

Developing a program, such as sasim, with the capability to integrate multiple-point

statistics will allow the evaluation of the importance of that information. Recall that

the importance of an input information is measured in terms of the output space of

uncertainty and not necessarily in the appearance of the stochastic models.

Throughout this dissertation a multiple-point or N -point configuration is defined

by N separation lag vectors h1, . . . ,hN , with by convention h1 = 0. Figure 3.4 shows

examples of 1, 2, 3, 4, and 9 point configurations.

The first multiple-point statistic coded into sasim is the semivariogram (2.19),

which carries the same two-point information as the the covariance or correlogram.

CHAPTER 3. APPLICATION OF ANNEALING TECHNIQUES 74

The semivariogram for lag hl is defined as:

γ(hl) =1

2 · N(hl)

N(hl)∑[Z(u) − Z(u + hl)]

2 (3.7)

Where any number l = 1, . . . , L of separation vectors hl may be specified, N(hl) is the

number of pairs for lag hl, and γ(hl) is the semivariogram for lag hl. A component c

in the objective function O is then:

Oc =L∑

l=1

|γcontrol(hl) − γrealization(hl)| (3.8)

with γcontrol(hl) being the control or reference semivariogram and γrealization(hl) the

experimental semivariogram of the candidate realization.

The second multiple-point statistics coded into sasim are multiple-point non-

centered indicator covariances. The indicator transform of a categorical variable is

defined as, see also (2.11):

I(u; k) =

1, if Z(u) ∈ category k

0, otherwise(3.9)

The indicator transform of a continuous variable may be constructed in one of two

ways:

I(u; z) =

1, if Z(u) ≤ z

0, otherwise(3.10)

J(u; z) =

1, if Z(u) > z

0, otherwise(3.11)

with I(u; z) = 1 − J(u; z).

The multiple-point non-centered indicator covariance defined by N lag separation

vectors h1, . . . ,hN , with by convention h1 = 0, is defined as:

ϕI(h1, . . . ,hN ; z1, . . . , zN) = E

N∏j=1

I(u + hj ; zj)

(3.12)

The non-centered indicator covariance for J is similarly defined:

ϕJ(h1, . . . ,hN ; z1, . . . , zN) = E

N∏j=1

J(u + hj ; zj)

(3.13)

CHAPTER 3. APPLICATION OF ANNEALING TECHNIQUES 75

In general, ϕI and ϕJ contain different information: ϕI is the probability of the N

values being jointly below the given threshold values whereas ϕJ is the probability of

the N values being jointly above the given threshold values. Recall from section 2.3

equation (2.15), that the connectivity function of Journel and Alabert [85] corresponds

to an I indicator defined at the same cutoff zc with the N -points spatially separated

by multiples of the same lag separation vector h.

Given L such multiple-point covariances a component c of the objective function

O is then:

Oc =L∑

l=1

| ϕcontrolI (h

(l)1 , . . . ,h

(l)N ; z

(l)1 , . . . , z

(l)N )

− ϕrealizationI (h

(l)1 , . . . ,h

(l)N ; z

(l)1 , . . . , z

(l)N ) | (3.14)

Where each of the l = 1, . . . , L contributions is a specific N (l)-point indicator covari-

ance. Note that ϕJ could be used in place of ϕI .

The univariate cdf F (z), at a specific threshold z, may be expressed as a non-

centered indicator covariance ϕI(h1; z) with N = 1 and h1 the null vector. Thus, the

univariate cdf may explicitly enter the objective function in the form of non-centered

indicator covariances.

Traditional bivariate indicator covariances, CI(h; z1, z2), may also enter the ob-

jective function as specific N = 2 non-centered multiple-point indicator covariances:

ϕI(h1,h; z1, z2) = CI(h; z1, z2) + F (z1) · F (z2)

where ϕI(h1,h; z1, z2) is the non-centered indicator covariance, h1 is the null vector,

h is the pair separation vector, and z1 and z2 are indicator thresholds.

Note that an N -point statistic includes lower order statistics M < N whose config-

urations are part of the N -point geometric configuration. For example, when consider-

ing non-centered two-point indicator covariances it is unneccessary to allow explicitly

for the univariate cdf at the same thresholds.

Specific non-centered indicator covariances ϕI(h1, . . . ,hN ; k1, . . . , kN) for categor-

ical variables may also enter the objective function. When considering a categorical

variable that takes outcomes k = 1, . . . , K, the collection of all direct and cross non-

centered indicator covariances is referred to as a multiple-point histogram denoted

CHAPTER 3. APPLICATION OF ANNEALING TECHNIQUES 76

with notation f (see also (2.17) in section 2.3.1):

ϕI(h1, . . . ,hN ; k1, . . . , kN) = f(h1, . . . ,hN ; k1, . . . , kN) (3.15)

∀ k1, . . . , kN = 1, . . . , K

These cross non-centered indicator covariances can be interpreted as the set of all

transition probabilities for the K categories for a pattern described by the N vectors

h1, . . . ,hN .

Any number L of multiple-point histograms, specified as configurations of N (l)

points defined by N (l) separation vectors h1, . . . ,hN(l), may be considered in the

objective function as:

Oc =L∑

l=1

K∑k1=1

K∑k2=1

· · ·K∑

kN(l)=1

| f control(h(l)1 , . . . ,h

(l)N ; k1, . . . , kN(l))

− f realization(h(l)1 , . . . ,hN(l); k1, . . . , kN(l)) | (3.16)

There is no essential difference between N -point covariances, which enter the ob-

jective as (3.14), and N -point histograms which enter as (3.16). The difference in

the sasim program is that the thresholds for each multiple-point covariance must

be explicitly specified whereas selecting a multiple-point histogram causes sasim to

consider automatically all possible classes or combinations of threshold values.

Given an N -point histogram of a categorical variable that can take one of K

outcomes, there are KN classes. Table 3.1 shows the number of classes for different

K and N values. Although there is a practical limit to the number N of points

that can be considered, sasim has been coded with no such restriction (only the

dimensioning will have to be changed when RAM memory in the gigabyte range

becomes available).

Commonly, the KN classes are assigned a one dimensional index for computer

storage and to allow N -point histograms to be displayed in the traditional way, i.e., as

bar charts of frequency versus class number. Given N categorical variables z(hi), i =

1, . . . , N that can take values k = 1, . . . , K a unique one dimensional index may be

computed as:

index = 1 +i=N∑i=1

[z(hi) − 1] · Ki−1 (3.17)

CHAPTER 3. APPLICATION OF ANNEALING TECHNIQUES 77

N=2 N=3 N=4 N=9K=2 4 8 16 512K=3 9 27 81 19,683K=4 16 64 256 262,144K=5 25 125 625 1,953,125K=6 36 216 1,296 10,077,696K=7 49 343 2,401 40,353,607K=8 64 512 4,096 134,217,728K=9 81 729 6,561 387,420,489

K=10 100 1,000 10,000 1,000,000,000

Table 3.1: The number KN of classes for various multiple-point statistics (N = 2, 3, 4, 9)and numbers of categories (K = 2, 3, 4, 5, 6, 7, 8, 9, 10).

Each index value corresponds to a unique pattern, i.e., a particular combination of

the N points. For example, given N = 4 and K = 2 there are KN = 24 = 16 possible

patterns with indices given by:

index = z(h1) + [z(h2) − 1] · 2 + [z(h3) − 1] · 4 + [z(h4) − 1] · 8 (3.18)

If the four points constitute a square configuration, with k = 1 corresponding to

white, and k = 2 to black, the 16 patterns are enumerated in Table 3.2 and shown

graphically on Figure 3.5.

This indexing convention will be used throughout the code and text of this dis-

sertation.

Updating Multiple-Point Statistics

As discussed in section 3.2, an important practical criterion for any statistic is that

is should allow the objective function to be updated locally rather than having to be

globally recalculated after each local perturbation. Equation (3.2) in section 3.2 shows

how the semivariogram could be locally updated. It turns out that local updating is

also possible for multiple-point statistics.

Let Z be a categorical variable that can take one of K integer categories k =

CHAPTER 3. APPLICATION OF ANNEALING TECHNIQUES 78

index z(h1) z(h2) z(h3) z(h4)1: 1 1 1 12: 2 1 1 13: 1 2 1 14: 2 2 1 15: 1 1 2 16: 2 1 2 17: 1 2 2 18: 2 2 2 19: 1 1 1 2

10: 2 1 1 211: 1 2 1 212: 2 2 1 213: 1 1 2 214: 2 1 2 215: 1 2 2 216: 2 2 2 2

Table 3.2: The one dimensional indices of a four-point histogram with K=2.

CHAPTER 3. APPLICATION OF ANNEALING TECHNIQUES 79

1 2

3 4Locations: h_1, h_2, h_3, h_4

index = 1 index = 2 index = 3 index = 4

index = 5 index = 6 index = 7 index = 8

index = 9 index = 10 index = 11 index = 12

index = 13 index = 14 index = 15 index = 16

Figure 3.5: An illustration of the indexing convention for a four-point configuration: N =4,K = 2. The configuration is illustrated at the top of the figure followed by all 16 possiblepatterns (white and black correspond to k = 1 and 2 respectively).

CHAPTER 3. APPLICATION OF ANNEALING TECHNIQUES 80

1, . . . , K. Consider any realization with values z(ui), i = 1, . . . , N and any arbi-

trary N -point histogram defined by N lag separation vectors h1, . . . ,hN with, by

convention, h1=0. This specific multiple-point histogram is globally calculated from

the starting image before any perturbation, i.e., all of the following proportions are

enumerated:

f(h1, . . . ,hN ; k1, . . . , kn) =

proportion of patterns z(h1) = k1, . . . , z(hN ) = kN (3.19)

∀k1, . . . , kN = 1, . . . , K

Now, consider that a particular nodal value z(u) = k is perturbed into z(u) = k′.

The multiple-point histogram has changed in that all configurations involving location

u are now different. The updating concept consists of subtracting what was there

before and adding what comes in after the perturbation.

For a two point statistic, defined by two vectors h1 = 0 and h2, there are two

configurations where the pattern has changed (as shown on Figure 3.3). The first is

generated by offsetting the configuration by minus the first lag vector h1:

point one : (u + h1) − h1

point two : (u + h2) − h1

The second configuration that has changed is that generated by offsetting the config-

uration by minus the second lag vector h2:

point one : (u + h1) − h2

point two : (u + h2) − h2

For both of these configurations the original pattern with z(u) = k is subtracted from

the global statistic and the new pattern with z(u) = k′ is added in.

As another example, consider the specific three-point statistic shown on Figure 3.6.

Perturbing the central grid node location, shown by the black dot, changes three

patterns/configurations. The first configuration is that generated by offsetting the

CHAPTER 3. APPLICATION OF ANNEALING TECHNIQUES 81

configuration of N = 3 points by lag −h1 (a null vector by convention):

point one : (u + h1) − h1

point two : (u + h2) − h1

point three : (u + h3) − h1

The second configuration is that generated by offsetting the N = 3 point configuration

by lag −h2:

point one : (u + h1) − h2

point two : (u + h2) − h2

point three : (u + h3) − h2

The third and final configuration that needs updating, regardless of the size of the

grid network, is that generated by offsetting the N = 3 point configuration by lag

−h3:

point one : (u + h1) − h3

point two : (u + h2) − h3

point three : (u + h3) − h3

In general, for any N -point statistic there are N configurations where the pattern

changes if a nodal value z(u) = k is perturbed to z(u) = k′. Updating the N -

point statistic amounts to subtracting the previous contribution and adding the new

contribution for all N configurations:

pattern of Npoints offset by − h1

pattern of Npoints offset by − h2

...

pattern of Npoints offset by − hN

Within the sasim program the subtraction and addition are done simultaneously

for each of the N configurations. The updating principle is the same regardless of the

specific multiple-point statistic and how many points N are involved.

CHAPTER 3. APPLICATION OF ANNEALING TECHNIQUES 82

A Specific Three-Point Statistic

1 2

3

First Pattern to Update

Unchanged grid node location

Perturbed grid node location

Second Pattern to Update

Unchanged grid node location

Perturbed grid node location

Third Pattern to Update

Unchanged grid node location

Perturbed grid node location

Figure 3.6: An example of the three-point configurations that need updating after a per-turbation.

CHAPTER 3. APPLICATION OF ANNEALING TECHNIQUES 83

1

2 3

4

Figure 3.7: The four point configuration, and three direction vectors, corresponding to theexamples shown on Figures 3.8, 3.9, and 3.10.

Examples with Four-Point Histograms

For an additional acquaintance with multiple-point statistics consider the four-point

histogram corresponding to four points in a unit square configuration as in Figure 3.7.

Three examples of four point histograms with K = 2 (white and gray) are shown on

Figure 3.8. All three images are 200 by 200 pixels and have exactly 50% white and

50% gray pixels. The four point histogram is shown to the right of each image. There

are Kn = 24 = 16 classes indexed from 1 through 16. Note that the horizontal axes

on each histogram plot is the one dimensional index (3.17) of each class. The upper

image shows 20 alternating white and gray stripes each 5 pixels thick. There are only

four non-zero classes in the corresponding four-point histogram (all white, all gray,

2 white over 2 gray, and 2 gray over 2 white). The central image shows 5 by 5 gray

squares located at random. All classes in the four-point histogram are represented;

the histogram shows only six significant classes, all white, all gray, 2 white over 2 gray,

2 gray over 2 white, 2 gray beside 2 white, and 2 white beside 2 gray). The random

image, shown at the bottom, leads to an equal probability to be in each class4.

Figure 3.9 attempts to quantify visually how much information is carried in such

four-point histograms. The training image is shown at the top of the figure. The

two middle realizations are conditional to the two-point histogram in one lag in all

directions (horizontal, vertical, and the two diagonal directions). The two lower real-

izations are conditional to one particular four-point statistic (a 2 by 2 pixel configu-

ration as in Figure 3.7). The two-point statistics cover the same areal extent as the

four-point information but clearly do not carry the same information. The difference

4This would not be true if there was an unequal fraction of white and gray pixels

CHAPTER 3. APPLICATION OF ANNEALING TECHNIQUES 84

Stripes

Fre

quen

cy

class

0 4 8 12 16

.000

.100

.200

.300

.400

Stripes: Four Point Histogram

Squares

Fre

quen

cy

class

0 4 8 12 16

.000

.100

.200

.300

.400

Squares: Four Point Histogram

Random

Fre

quen

cy

class

0 4 8 12 16

.0000

.0100

.0200

.0300

.0400

.0500

.0600

Random: Four Point Histogram

Figure 3.8: Three illustrations of four-point histograms. The four-point histogram is forfour-points in a unit square configuration. The images are shown on the left and the fourpoint histograms on the right.

CHAPTER 3. APPLICATION OF ANNEALING TECHNIQUES 85

quantified in terms of output uncertainty is given in Chapter 5.

A second example shown on Figure 3.10 compares images generated with two-

point and four-point statistics. Two-point statistics for a unit lag in each direction

were retained to condition the two middle realizations and the four-point information

corresponding to a unit square configuration was retained for the two lower real-

izations. The areal coverage of both sources of information is the same, yet, the

four-point statistic seems to capture more of the salient features of the upper training

image such as the diagonal banding.

In practice, the multiple point configurations to be used must be chosen as char-

acteristics of the geometric features deemed important. For example, the four point

configuration shown on Figure 3.7 or any arbitrary configuration such as the six-point

or hexivariate configurations shown on Figure 3.11 may be retained. The specification

of multiple-point statistics requires the number of points N and the N −1 separation

lag vectors (offsets from the first vector h1 set to zero).

A combination of different “multiple” information is recommended. For example,

consider retaining two-point statistics specified for larger separation distances with

some multiple-point statistics used for short scale features. This point is further

developed with some example applications in Chapter 5.

3.4 Well Test-Derived Effective Permeability

Well test data is already quantified in the sense that the pressure response to some

flow impulse is measured with extremely precise gauges [66, 111]. The challenge is to

interpret this detailed pressure response and relate it to the petrophysical properties

that are important for flow modeling. Interpretation of the pressure transient is

necessary. Early-time effects such as wellbore storage and near wellbore effects such

as the skin effect are not relevant to the reservoir modeling problem being addressed

in this dissertation. Similarly, the modeling of reservoir boundaries and other late-

time effects could be accounted for deterministically and are not considered in this

research. For example, fault boundaries can be explicitly entered into the numerical

model of the reservoir. Moreover, the effects of other wells can also be explicitly

CHAPTER 3. APPLICATION OF ANNEALING TECHNIQUES 86

Berea Sandstone Training Image

Two-point Statistics Two-point Statistics

Four-point Statistics Four-point Statistics

Figure 3.9: The Berea control pattern, two realizations using two-point information, andtwo realizations using four-point information. Both the two-point and four-point informa-tion cover the same extent (a unit pixel in each direction).

CHAPTER 3. APPLICATION OF ANNEALING TECHNIQUES 87

Training Image

Two-point Statistics Two-point Statistics

Four-point Statistics Four-point Statistics

Figure 3.10: A difficult control pattern, two examples using partial bivariate information,and two examples using quadrivariate information. The bivariate data was retained for 25lags in four directions. The quadrivariate data was a 2 by 2 pixel configuration.

CHAPTER 3. APPLICATION OF ANNEALING TECHNIQUES 88

Six Points: discontinuous pattern

1 2

3 4 5 6

Six Points: curvilinear pattern

1

2 3

4 5 6

Six Points: rectangular pattern

1 2 3

4 5 6

Six Points: linear pattern

1 2 3 4 5 6

Figure 3.11: Examples of 6-point configurations. In practice, the configuration is cus-tomized to reflect the spatial features deemed important.

accounted for by digitizing the drainage area of each well.

One way of conditioning stochastic reservoir models to a well test response would

be to discard all models that do not yield a forward simulated well test response close

enough to the actual measured pressure response. Such selection procedures may be

practical when building models with a single well test; however, it is not practical in

the presence of multiple well test interpretations, some with more advanced multirate

tests which inform a number of different permeability averages near the well. In

general, a prohibitively large number of realizations would be required to find a few

that simultaneously match all well test data.

The idea of this research is to take realizations that already match other available

data (e.g., core, well logs, seismic) and modify them, with annealing, to match addi-

tionally the well test responses. Given the computer resources available at this time

it is not possible to perform a full 3-D flow simulation after every perturbation called

for in the annealing algorithm; such 3-D flow simulation would amount to a global

updating. Somehow the well test data must be translated into an easily updateable

property while retaining the flexibility to differentiate a wide variety of heterogeneous

systems encountered in practice.

CHAPTER 3. APPLICATION OF ANNEALING TECHNIQUES 89

The absolute effective permeability near the well bore inferred from the well test

response is a critical parameter provided by well test interpretation. The essential

assumptions and mathematical basis for well test interpretation must be discussed

before presenting a quantification of the permeability averaging process associated to

a well test.

Pressure transient well tests are performed by generating some flow rate impulse

in the reservoir (e.g., start production, change the flow rate, stop production, . . . )

and measuring the pressure response. Well test interpretation or pressure transient

analysis consists of interpreting the pressure response by using some appropriate

mathematical model to relate the pressure response (output) to flow rate history (in-

put) [66]. Provided that the mathematical model is appropriate the model parameters

can be associated to certain reservoir parameters.

Pressure transient analysis is based on what seem a-priori like limiting assumptions

[66, 103], these being,

• There is radial flow into a well opened over the entire vertical thickness of the

formation.

• The formation is homogeneous and isotropic with constant porosity and con-

stant permeability (both are independent of pressure).

• The fluid has a small and constant compressibility and a constant viscosity.

• The pressure gradients are small and the gravity forces are negligible.

Some of these assumptions may be relaxed by considering variable transformations,

adding terms to the flow equations to account for non-Darcy effects, or by considering

more sophisticated flow equations.

The equation governing fluid flow given the preceding assumptions is the following:

∂2p

∂r2+

1

r

∂p

∂r=

φµc

k

∂p

∂t(3.20)

where p is pressure, r is the radial distance from the wellbore, φ is the porosity, µ

is the viscosity, c is the fluid compressibility, k is the absolute permeability, and t is

time.

CHAPTER 3. APPLICATION OF ANNEALING TECHNIQUES 90

In the case of an infinite reservoir and a line source well the pressure solution to

this equation may be written.

p(r, t) = pi − qµ

2πkh

{−1

2Ei

(−φµcr2

4kt

)}(3.21)

where Ei is the exponential integral function,

−Ei(−x) =∫ ∞

x

e−u

udu

The exponential integral function is approximated very well by a log function for

all but early times, therefore, we may write (see equation 2.32, page 11, Matthews

and Russell [103]):

p(r, t) = pi − qµ

4πkh

[ln

kt

φµcr2+ 0.80907

](3.22)

In the previous development the standard SPE notations have been used as much as

possible [123]. Recall that p(r, t) is the pressure (psi) at radial distance r (feet) and

time t (hours), pi is the initial reservoir pressure (psi) (the subscript i is not to be

confused with a grid node index), q is the flow rate (STB/d), µ is the fluid viscosity

(cp), k is the absolute permeability (md), h is the formation thickness (feet), φ is

the porosity (pore volume/bulk volume), and c is the compressibility (/psi). In many

situations dimensionless variables are used to simplify the notation and to provide

solutions that are independent of any particular unit system [66].

The dimensionless pressure pD is defined in oilfied units as:

pD =kh

141.2qBµ(pi − pwf) (3.23)

where all variables are as above and pwf is the well flowing pressure (the pressure

directly measured in a well test). In a consistent unit set, pD is defined as:

pD =2πkh

qBµ(pi − pwf) (3.24)

The dimensionless time tD is defined in oilfield units as:

tD =0.000264kt

φµctr2w

(3.25)

CHAPTER 3. APPLICATION OF ANNEALING TECHNIQUES 91

where all variables are as above and rw is the wellbore radius in feet. In a consistent

unit set, tD is defined as:

tD =kt

φµctr2w

(3.26)

Finally, the dimensionless radius rD is defined as:

rD =r

rw(3.27)

Given a well test performed in the field, and estimates for µ, h, φ, and c, it is

possible to estimate an effective permeability ke that relates to some volume near

the well. The idea proposed by Alabert [7] and developed in this dissertation is to

impose this effective permeability on stochastic models through the objective function

of annealing-based techniques. An important issue that must be addressed is the

volume and type of averaging represented by the well test-derived permeability.

3.4.1 Empirical Relationship for the Well Test Effective Per-

meability

The nature of the radial averaging process depends on the exact spatial configuration

of the reservoir attributes; therefore, in all rigor, a full 3-D flow simulation is required

to quantify this averaging process. In practice, the constraint of annealing calls for a

very fast yet reasonably accurate evaluation of each parameter entering the objective

function. It would not be practical to perform a full 3-D flow simulation to obtain the

effective permeability after each (set of) perturbation(s). An empirical scaling law

must be devised that relates the well test-derived effective permeability to elementary

block absolute permeability values.

Alabert [7] proposes a power average of the block permeabilities within a specified

averaging volume V to model the full non-linear averaging of block permeabilities as

measured by a well test. The assumption is that the elementary block permeability

values average linearly after a non-linear power transformation, i.e.,

k(ω) =

1

N

∑ui∈V

k(ui)ω

(3.28)

CHAPTER 3. APPLICATION OF ANNEALING TECHNIQUES 92

Where k(ω) is the ω-power average permeability of the N permeability values k(ui), i =

1, . . . , N , at locations ui within the volume of interest V . The power ω varies between

the bounding values of -1 and 1 corresponding to the harmonic and arithmetic aver-

ages respectively. Note that the geometric average is obtained for ω → 0.0. Previous

work with power averaging [27, 36, 35, 88, 94, 95] has shown that the averaging power

parameter ω is remarkably robust.

Once the averaging volume V and the averaging type, as characterized by the

averaging power ω, are known a component objective function is considered as:

Oc = [kwelltest − krealization]2 (3.29)

where kwelltest is the well test-derived effective permeability and krealization is the ef-

fective permeability of the candidate realization, see equation (3.28). When the value

at one node location k(u) is perturbed to k′(u) the effective permeability is updated

to knewrealization from kold

realization by:

knewrealization =

[(kold

realization

)ω − 1

Nk(u)ω +

1

Nk′(u)ω

] 1ω

(3.30)

This fast numerical updating replaces a full flow simulation to arrive at a new

effective permeability for each candidate realization. To implement this numerical

approximation both the averaging volume V and averaging power ω must be defined.

The Averaging Volume

To define the appropriate averaging volume it is necessary to consider the portion of

the pressure response used to derive the well test effective permeability kwelltest. In

practice, kwelltest is obtained by interpreting the pressure response during the time at

which the response resembles infinite-acting radial flow. Early-time effects such as

wellbore storage and late-time boundary effects are not considered in the interpre-

tation. It is possible to define an inner radius rmin and an outer radius rmax that

correspond to the limits of infinite acting radial flow since the pressure response, at

CHAPTER 3. APPLICATION OF ANNEALING TECHNIQUES 93

any time t, may be related to block permeabilities within a time-dependent radius-

of-drainage r(t). Consider a typical pressure response shown on the Miller-Dyes-

Hutchinson (MDH) plot5 at the bottom of Figure 3.12. The pressure response be-

tween 1 hour and 10 hours is used to derive an estimate of the effective permeability

kwelltest. The inner and outer limits of the shaded region (on the schematic illustra-

tion of the reservoir) correspond to the radius-of-drainage at 1 hour and 10 hours

respectively.

Before proceeding with details of how to calculate rmin and rmax it is useful to

explain why the volume of investigation is an annular volume. Intuitively, it is rea-

sonable to expect that for a given well test duration the volume of investigation is

limited by an outer radius; however, the reason why the near wellbore permeability

values are excluded is less intuitive. First, it should be noted that the well test effec-

tive permeability is based on the pressure derivative during the infinite acting radial

flow portion of the response. The slope on the straight line fit through the response

on Figure 3.12 is linearly related to the permeability, i.e.,

ke = 162.6qBµ

|m|h (3.31)

where m is the slope of the semilog straight line. At any instant in time the pres-

sure derivative is informed by a particular annular volume centered at the wellbore.

Figure 3.13 shows the normalized weighting function, derived6 by Oliver [107], for

tD = 102, 104, and 106. By not considering the early time pressure response the block

permeability values near the wellbore do not contribute to the well test derived effec-

tive permeability. Figure 3.14 shows the contribution of the block permeability values

plotted as a function of the radial distance from the wellbore.

The time interval that the pressure response resembles infinite-acting radial flow is

easily determined by standard interpretation techniques 7. Evaluating the radius-of-

drainage r(t) at the time limits is not as straightforward; depending on the arbitrary

5A plot of pressure versus the logarithm of time.6The derivation calls for the permeability field to be expressed as a constant value plus a small

variation. The consequences of this assumption have not been evaluated since this weight functionis only used for explanation purposes.

7Note that the outer limit may be at the end of the test; boundary effects may not be seen on“short” well tests.

CHAPTER 3. APPLICATION OF ANNEALING TECHNIQUES 94

Miller-Dyes-Hutchinson Plot

Time, hrs

Pre

ssur

e, p

si

Early-time effects Late-time effectsInfinite Acting

Radial Flow

0.01 0.1 1 10 100

270.0

274.0

278.0

282.0

286.0

290.0

Plan View of the Reservoir

wellVo

lum

e Inform ed by Infinite Acting R

adial

Flo

w

Figure 3.12: A schematic illustration of the volume measured by a given well test inter-pretation. The inner and outer limits of the shaded region, on the schematic illustrationof the reservoir, correspond to the radius-of-drainage at 1 hour and 10 hours (the limits ofinfinite acting radial flow).

CHAPTER 3. APPLICATION OF ANNEALING TECHNIQUES 95

Normalized Weighting Function for Three tD values

Dimensionless Distance

Nor

mal

ized

Wei

ghtin

g F

unct

ion

td = 100 td = 10000 td = 1000000

1 10 100 1000 10000

0.00

0.10

0.20

0.30

0.40

Figure 3.13: The normalized weighting function versus the dimensionless radius rd forthree different instants in time.

Weight

Inner limits

Outer limits

Radial Distance Radial Distance

Figure 3.14: A schematic illustration of the radial weight function for the power averageapproximation to the well test effective permeability.

CHAPTER 3. APPLICATION OF ANNEALING TECHNIQUES 96

definition chosen for r(t), the radius can change by as much as a factor of 4. As

documented by Alabert [7] it will be necessary to calibrate the radius-of-drainage

r(t) by repeated flow simulations.

The radius-of-drainage defined below follows the classical formulation by van

Poollen [132] and Johnson [72]. The work of Alabert [7] and Oliver [107, 108] present

additional interpretations and numerical examples. The block permeabilities con-

tributing to the pressure response measured up to time t are approximately enclosed

by a circular volume centered at the well defined by a time-dependent radius r(t)

written as [7, 72, 132]:

r(t) = A

√ket

φµct

(3.32)

where A is a constant, ke is the reservoir permeability around the well, φ is the

porosity, µ is the fluid viscosity, and ct is the total compressibility. This relation

(3.32) implicity assumes that the heterogeneous permeability field is described by a

single permeability, i.e., ke is independent of time. Note that (3.32) can be written

in dimensionless units as:

rD(t) = A√

tD (3.33)

with rD and tD defined in equations (3.27) and (3.26).

Depending on the definition chosen for the radius-of-drainage the value of A varies

from 0.023 to 0.07 (for oil field units) or between 1.4 and 4.3 in consistent units.

Alabert [7], in the context of evaluating the averaging volume of a well test, and for

specified levels of discretization and test durations, found a robust optimal Aopt value

of 0.010 in oil field units (0.6 in consistent units). Thus, the annular volume V (A)

may be defined by rmin(A) and rmax(A) where A is taken as Aopt = 0.010 or calibrated

by repeated flow simulations.

The Averaging Power

The averaging power ω describes the type of averaging within the volume V (A). In

many cases, this averaging power is close to the geometric average (ω = 0). For

example, Butler [23] and Alabert [7] show cases where the effective permeabilities

obtained from classical well test analyses converge towards the geometric average. For

CHAPTER 3. APPLICATION OF ANNEALING TECHNIQUES 97

practical test durations and for complex heterogeneous permeability distributions, the

type of averaging can differ significantly from the geometric average.

The constant A = 0.010 and the averaging power ω = 0.0 may be used as first

approximations in equation (3.28) to calculate the effective permeability krealization.

However, in practice, both the constant A and the averaging power ω should be

calibrated ahead of time with the following calibration procedure:

1. Generate ns (20-100) multiple realizations of the permeability field with relevant

statistical properties.

2. Forward simulate a well test, with conditions as close as possible to those used in

the field to arrive at kwelltest, on each realization to obtain ns pressure response

curves.

3. Deduce an effective permeability ki, i = 1, . . . , ns from each pressure curve using

established well test interpretation techniques [66].

4. Compute average permeabilities k(A, ω)i, i = 1, . . . , ns for A values between

the practical bounding limits of 0.005 and 0.035 and for ω values between the

bounding limits of -1 and 1. For example, taking A-increments of 0.001 and

ω-increments of 0.05 would lead to

ns ·(

0.035 − 0.005

0.001+ 1

)·(

1.0 − (−1.0)

0.05+ 1

)= ns · 1271

calculations of k(ω, A) values.

5. Choose the pair (Aopt, ωopt) that yields the closest agreement between the refer-

ence ki, i = 1, . . . , ns values and the approximate k(Aopt, ωopt)i, i = 1, . . . , ns val-

ues. Alabert suggests taking the mean normalized absolute deviation (mNAD)

and the mean normalized error (mNE) as criteria for selecting (Aopt, ωopt). That

is, choose the pair (Aopt, ωopt) that jointly minimize:

mNAD(A, ω) =i=ns∑i=1

|k(A, ω)i − ki|ki

(3.34)

mNE(A, ω) = |∑i=ns

i=1 k(A, ω)i −∑i=nsi=1 ki∑i=ns

i=1 ki

| (3.35)

CHAPTER 3. APPLICATION OF ANNEALING TECHNIQUES 98

The quality of the power average approximation can be judged by plotting a scat-

terplot of the approximate values k(Aopt, ωopt)i, i = 1, . . . , ns versus the reference well

test effective permeabilities ki, i = 1, . . . , ns. An example illustrating the integration

of well test data is given in Section 5.3. For that example, it will be shown that the

power average approximation and the well test-derived effective permeability have a

0.9 correlation coefficient.

3.4.2 A Wrong Averaging Volume

In the case of radial flow, one would naturally consider the permeability values close

to the well as more consequential than permeability values farther away from the

well. Initially, this research considered that the averaging volume was specified by

a radially decreasing weight function. This concept is not valid since it does not

account for the implicit fitting of a slope to the infinite-acting radial flow period of

the response. The early time behavior and the effect of permeability values close

to the wellbore are filtered from the interpretation8. This incorrect weight function

will be documented below for completeness. This subsection will conclude with a

numerical example which verifies the final approach documented earlier.

The key assumption behind a radially decreasing weight function is that the weight

attributable to an elementary volume is proportional to the pressure drop across the

volume:

f(r, t) ∝ ∆p(r, t) (3.36)

This pressure drop is calculable from the pressure-diffusion equation. If the blocks be-

come infinitely small then the weight function would be proportional to the derivative

of p(r, t) with respect to r, i.e.,

f(r, t) ∝ d

drp(r, t)

where,

d

drp(r, t) =

d

dr

(pi − qµ

4πkh

[ln

kt

φµcr2+ 0.80907

])

8The near wellbore permeability effects are grouped with the skin effect.

CHAPTER 3. APPLICATION OF ANNEALING TECHNIQUES 99

Vertical Well of radius r_w

blocks of equal length l

0 1 2 3 i n

length = l

Figure 3.15: A schematic illustration of the typical situation: a vertical well in the centerof a block imbedded within a sequence of such blocks. Note that there are a symmetricnumber of grid blocks to the left of the vertical well.

=−qµ

4πkh

d

dr

(ln

kt

φµcr2

)

=qµ

2πkh

1

r(3.37)

Now, the weight function for any radial distance greater than the wellbore radius

r ≥ rw and within the drainage area A may be written:

f(r) =qµ

4πkh1r∫

Aqµ

4πkh1r

=1r∫

A1r

(3.38)

Note that the flow rate q, viscosity µ, permeability k, and formation thickness h, do

not contribute to the weight function f(r) (3.38) once it is standardized to sum to

1.0. Further, note that the weight (3.38) and the the derivative (3.37) do not depend

on the duration of the well test; the pressure diffusion equation calls for the pressure

impulse to be instantaneously felt an infinite distance from the source.

In practice, the reservoir is not modeled by infinitesimal blocks. The most common

model considered in flow simulators is that illustrated on Figure 3.15, i.e., a vertical

well of radius rw is in the center of a block within a sequence of such blocks. The

weight attributable to the first block “0” is proportional to the pressure drop from

the block boundary to the well bore, i.e.,

f0(r = 0, t) ∝ p(l

2, t) − p(rw, t)

∝pi − qµ

4πkh

ln kt

φµc l2

2 − lnkt

φµcr2w

CHAPTER 3. APPLICATION OF ANNEALING TECHNIQUES 100

= − qµ

4πkhln

[4(

rw

l

)2]

(3.39)

Note that f0(r, t) in equation 3.39 does not depend on time.

Similarly, the weight for any arbitrary block i, i = 1, 2, . . . , n (see Figure 3.15) is

written,

fi(r, t) ∝pi − qµ

4πkh

ln kt

φµc[(i − 1

2)l]2 − ln

kt

φµc[(i + 1

2)l]2

= − qµ

4πkhln

( i − 1

2

i + 12

)2 (3.40)

Note that fi(r, t), i = 1, 2, . . . , n is independent of both the time t and the block

length l. The decrease in the function is approximately inversely proportional to the

block number, i.e., to the ln term in equation (3.40),

fi(r, t) ∝ f(r) = ln

(

i − 12

i + 12

)2 (3.41)

The actual weights f(r) depend on the the weight assigned to the first block (3.39)

because, they are all standardized to sum one. The weight assigned to the first block

is directly proportional to the ln term in equation (3.39),

f0(r, t) ∝ f(0) = ln

[4(

rw

l

)2]

(3.42)

Note that lrw

can be interpreted as the relative or dimensionless grid block size.

Finally, given N blocks at radial distances ri, i = 0, . . . , N , the weight function

can be written as:

f(ri) =f(ri)∑N

i=0 f(ri)(3.43)

where the 0’th block corresponds to the block with the well (see equation (3.42)

for f(0)), and i = 1, . . . , N correspond to blocks within the radius of drainage (see

equation (3.41) for f(ri)).

As the dimensionless grid block size increases the weight for the second, and

subsequent blocks, drops off faster. This is illustrated on Figure 3.16 where the

CHAPTER 3. APPLICATION OF ANNEALING TECHNIQUES 101

Weight versus Grid Block Number

Grid Block (number)

Wei

ght

l = 25

l = 400

0 2 4 6 8 10

.00

.20

.40

.60

.80

1.00

Figure 3.16: An illustration of the weight function for different dimensionless grid blocksizes (25, 50, 100, 200, and 400). The abscissa axis is the grid block index number (notdistance). The weights are normalized such that the weight applied to the first (well) blockis 1.0.

weight for dimensionless block sizes of 25, 50, 100, 200, and 400 is plotted versus the

grid block index. To better illustrate the effect of the block size the weight functions

have been standardized such that the weight applied to the first (well) grid block is

one. Also note that the weight function is not continuous; it is a step function with

a constant value for the entire block.

As mentioned earlier, the concept of a radially decreasing weight function was

found to be incorrect. The first indication that the radially decreasing weight function

was inappropriate was the poor correlation between actual well test derived effective

permeabilities kwelltest and weighted power average approximations k(ω).

The weight function finally adopted was also verified by a numerical experiment.

Starting from a uniform permeability field, the grid block permeabilities were per-

turbed, one at a time, at successively greater distances from the wellbore. After each

perturbation the effective permeability of the field can be computed. The weight

function should then be proportional to the deviation from the intial uniform perme-

ability. Figure 3.17 illustrates the result of this numerical experiment. Note that this

experimental weight applies approximately over an annular region.

CHAPTER 3. APPLICATION OF ANNEALING TECHNIQUES 102

Experimental Weight Function

Grid Block (number)

Wei

ght

.00 1.00 2.00 3.00 4.00 5.00 6.00 7.00

.000

.050

.100

.150

.200

Figure 3.17: An experimentally derived weight function generated by perturbing elemen-tary permeability values successively farther away from the well location.

3.4.3 Advanced Well Test Interpretation Techniques:

There are a number of newer approaches to performing and interpreting well tests

that bring information that could be integrated through annealing techniques. The

first method based on the instantaneous slope of the semilog pressure response, has

the potential of providing a great deal of information about the distribution of perme-

ability around the wellbore. The second method, based on cyclic flow rate variations,

also has significant potential but requires a specific periodic flow rate impulse.

Interpretation of the instantaneous slope of the pressure response curve, due to

Oliver [107, 108], provides an estimate of the effective permeability for a concentric

regions centered around the wellbore (see Figure 3.13). The significant limitations of

the technique is that it relies on very “clean” and accurate pressure measurements

and also on small permeability deviations from a constant overall mean.

Interpretation of the pressure response from a well test performed with cyclic flow

variations, as proposed by Rosa [114], is more robust in the sense that the results are

based on a more significant impulse to the reservoir.

Chapter 4

A Comparative Study of Various

Simulation Techniques

This chapter develops an extended example comparing stochastic simulation based

on annealing to the more conventional sequential Gaussian and sequential indicator

simulation techniques. A number of realizations are displayed on gray scale maps for

a visual comparison; however, the real criteria for comparison are those developed in

section 2.2, i.e., a good technique must generate realizations in a plausible amount of

time, it must account for the maximum amount of relevant prior information, and it

must yield the largest space of output uncertainty.

Section 4.1 presents the setting of the comparative study. The spatial distribution

of 40 by 40 air permeameter measurements, taken from a slab of Berea sandstone [53]

will provide the true reference for the study. The transfer function is a flow simulation

program and the response variables are measures of flow performance. The problem

will be to generate output distributions of those response variables, from limited input

data, that meet the criteria of a good simulation technique.

Section 4.2 documents the results of sequential Gaussian simulation (SGS), se-

quential indicator simulation (SIS), and annealing-based simulation (SAS) using only

a distribution model F (z) and a normal scores covariance model CY (h) for input data.

Only those implementations which generate realizations in a reasonable amount of

CPU time are considered. Moreover, since exactly the same amount of input data

103

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES104

is considered by each technique, the best technique will be that which generates the

largest space of output uncertainty while being accurate.

Section 4.3 considers how much of the uncertainty defined by the conventional

random function techniques (SGS and SIS) is due to fluctuations from the model

distribution F (z). These ergodic fluctuations become important when the size of the

field being simulated is small with respect to the range of two-point correlation.

Section 4.4 documents how each of the three techniques account for local condi-

tioning data and evaluates the impact of that conditioning on the output response

distributions.

This comparative study is important and relevant since it illustrates the applica-

tion of annealing with a concrete example. However, it will not be possible to draw

general conclusions; annealing methods applied in other situations will lead to dif-

ferent results and conclusions. This example was chosen because the data are real,

public domain, widely known, and few enough to allow repeated applications of a

flow simulator.

4.1 Setting of the Problem

The goal of this chapter is to evaluate how different simulation techniques succeed

to explore the output space of uncertainty for a given transfer function. A reference

image/model is necessary to establish the accuracy of the different simulation tech-

niques. Permeability data from a slab of Berea sandstone is taken as the reference

spatial distribution. In addition to displaying visual differences between realizations,

a flow-related transfer function is considered to quantify the output uncertainty gen-

erated by the different simulation techniques. The transfer function will be taken as a

two-phase numerical flow simulation program. The output response variables are flow

characteristics such as breakthrough time and recovery of an immiscible displacement

process.

The Berea sandstone permeability data were first published in 1985 by Giordano

and others [53]. Since its publication, the Berea data have served as reference values

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES105

Berea

30.0

40.0

50.0

60.0

70.0

80.0

90.0

Figure 4.1: Gray scale map of reference Berea permeability values.

for many geostatistical studies [6, 39, 85, 128]. The square grid of 40 by 40 perme-

ability values originate from air permeameter measurements taken from a 2 by 2 foot

vertical slab of Berea sandstone. The spatial distribution of the 1600 values is shown

on the gray scale map of Figure 4.1. Note that the scale of the permeability values is

in millidarcies. This scale will be used consistently for all gray scale representations

of Berea-related images. Further, note the strong diagonal banding, the connectivity

of the low permeability values in the lower left corner, and the concentration of the

high permeability values in the upper right corner.

A histogram of the permeability values, summary statistics, and a normal prob-

ability paper plot are shown on Figure 4.2. This distribution F (z), defined by 1600

values, is used as a reference for the different simulation techniques. All simulation

algorithms considered in this chapter generate realizations with a standard normal

univariate distribution. The simulated standard normal y-values are then back trans-

formed to the previous z-permeability distribution F (z), using a look-up table ap-

proach or a graphical transformation procedure [40, 90, 134]. The look-up table of

1600 paired z, y values is established by transforming the original z-values to standard

normal y-values with the following procedure,

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES106

Cum

ulat

ive

Pro

babi

lity

Permeability, md

Berea

.01

.1

.2

125

10

20304050607080

90959899

99.899.9

99.99

0. 10. 20. 30. 40. 50. 60. 70. 80. 90. 100. 110. 120.

Fre

quen

cy

Permeability, md

0. 40. 80. 120.

.0000

.0100

.0200

.0300

.0400

.0500

.0600

Berea Permeability Number of Data 1600

mean 55.53std. dev. 15.78

coef. of var 0.28

maximum 111.50upper quartile 65.00

median 55.00lower quartile 45.00

minimum 19.50

Figure 4.2: Histogram, summary statistics, and normal probability paper plot of the 1600reference Berea permeability values.

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES107

1. The z values are sorted in ascending order zi, i = 1, . . . , 1600, with

zi ≥ zj ∀ i > j.

2. An equal probability is assigned to each datum: pi = 1/1600. The cumulative

probability of each value zi is given by cpi =∑j=i

j=1 pj.

3. The normal score transform yi associated with each zi is

yi = G−1(

cpi + cpi−1

2

)(4.1)

with G(y) being the standard normal cdf, ycp = G−1(cp) being the corresponding

standard normal cp-quantile, and cp0 = 0.0.

Linear interpolation between the 1600 values, in the lower tail to a minimum of 15md,

and in the upper tail to a maximum of 120md allows for the transformation of any

z-value and the back transformation of any y-value.

The minimum spatial information needed by conventional stochastic simulation

techniques such as sequential Gaussian simulation (SGS) and sequential indicator

simulation (SIS) algorithm (see sections 2.3.6 and 2.3.7) is a stationary covariance or

variogram model. The experimental variogram of the 1600 normal score-transformed

y-values was computed in the two primary directions. The major direction, at an

angle of 123o (measured clockwise from the north or vertical axis), is that of the

diagonal banding with greatest continuity. The minor direction, at an angle of 33o,

is that perpendicular to the banding. The normal scores y-semivariogram is used as

a measure of spatial correlation for Gaussian, indicator1, and annealing2 techniques.

The experimental semivariogram points, shown as black dots, and the fitted model in

the two principal directions, shown as the solid line, are given on Figure 4.3. Note that

all distances are relative to the discretization units of the original Berea image with

the image being 40 distance units by 40 distance units. The following analytical model

1The indicator thresholds apply to the standard normal distribution and the resulting simulatedy-values are back transformed.

2The annealing procedure is applied in the normal space, i.e., with a standard normal distribution,and the resulting simulated y-values are appropriately back transformed.

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES108

Sem

ivar

iogr

am

Distance

Normal Scores Semivariogram

.0 5.0 10.0 15.0 20.0 25.0

.00

.20

.40

.60

.80

1.00

1.20

Figure 4.3: The experimental normal scores semivariogram in the two principal directions(123o and 33o) are shown as the black dots. The model fit in the same two directions isshown as the solid line.

was considered, with (h1, h2) being the coordinates in the two principal directions:

γY (h) = 0.10 + 0.20 · Sph

h21

3.62+

h22

1.82

+ 0.40 · Exp

h21

6.52+

h22

6.52

+ 0.30 · Sph

h21

144.02+

h22

12.02

(4.2)

Sph(h) is the unit range spherical variogram model defined as:

Sph(h) =

[1.5h − 0.5h3] , if h ≤ 1

1, if h ≥ 1

and Exp(h) is the exponential variogram function defined as:

Exp(h) = [1 − exp(−h)]

There are four components in the semivariogram model (4.2):

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES109

1. an isotropic nugget effect that explains 10% of the total variability,

2. a short scale anisotropic spherical structure (the longer range of 3.6 is in the

major 123o direction and the shorter range of 1.8 is in the minor 33o direction)

that explains an additional 20% of the variability,

3. an isotropic long range exponential structure (the range parameter is 6.5, thus,

the effective range is 19.5) that explains 40% of the variability, and

4. finally, an anisotropic spherical structure to account for the zonal3 anisotropy

(the range parameter is 12.0 in the 33o direction and 144.0 in the 123o direction)

that explains the remaining 30% of the variability.

The stationary normal scores covariance CY (h) is related to the normal scores

variogram:

CY (h) = 1.0 − γY (h) (4.3)

and is used interchangeably in the following text.

The normal scores variogram or covariance is a specific two-point or bivariate

measure of the spatial variability of Berea. There is potentially much more bivariate

information, in the form of two-point indicator covariances. Beyond the bivariate

level there is multiple-point information in the form of multiple-point indicator co-

variances (see section 2.3.1) that could be considered to describe more completely the

spatial characteristics of the Berea image (Figure 4.1). The only case where the sole

normal scores covariance is sufficient to describe completely the spatial distribution

corresponds to a multivariate Gaussian model.

One test for bivariate Gaussianity consists of comparing the theoretically pre-

dicted indicator variograms, based on a bivariate Gaussian distribution model, to the

corresponding experimental indicator variograms (see [40, 91, 138]). A close corre-

spondence between the theoretical and experimental variograms would imply that the

data are bivariate Gaussian, hence the complete multivariate Gaussian assumption

3A zonal anisotropy relates to a variogram which reaches different sill values in different directions[90]. The range parameter of 144.0 is not reached within the boundary of the figure (40x40) andamounts to fitting a zonal anisotropy.

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES110

Var

iogr

am

Distance

Indicator Variograms (5th percentile)

.0 5.0 10.0 15.0 20.0 25.0

.0000

.0100

.0200

.0300

.0400

.0500

.0600

Var

iogr

am

Distance

Indicator Variograms (95th percentile)

.0 5.0 10.0 15.0 20.0 25.0

.0000

.0100

.0200

.0300

.0400

.0500

.0600

Figure 4.4: The theoretical biGaussian indicator variograms and the experimental indicatorvariograms computed from the reference Berea image: the left figure shows the results forthe 0.05 quantile and the right figure shows the results for the 0.95 quantile. The solidcurves are the indicator variograms predicted by the biGaussian distribution in the 33o and123o directions using the normal score variogram model (4.2). The dashed lines are theexperimental indicator variograms in the same directions.

is not invalidated. The result of this test for the 0.05 quantile and the 0.95 quantile

are shown on Figure 4.4. There is a poor agreement between the theoretical results,

shown by the solid lines, and the experimental results shown by the dashed lines. In-

deed, the reference Berea data set is known not to be multivariate Gaussian, Journel

and Alabert in [85].

The main reason to restrict the simulation algorithms to the cdf F (z) of Figure 4.2

and the variogram model (4.2) is that Gaussian methods can only be conditioned to

these data. Thus, these two statistics represent the lowest common set of global statis-

tics that can be used by all algorithms considered here, i.e., the Gaussian, indicator,

and annealing-based simulation algorithms. Indicator methods could be extended

to integrate more completely the bivariate distribution [85] and, conceptually, the

annealing-based methods could be extended to integrate any arbitrary multiple-point

statistic (see Section 3.3).

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES111

Flow D

irect

ion

Injector

Producer

Figure 4.5: Schematic illustration of the flow scenario.

Flow Scenario

The transfer function considered is a two-phase numerical flow simulation program

(Eclipse [48]). The 2-D grid of 40 by 40 point permeability values is associated to a

much larger horizontal grid of block horizontal permeabilities. This grid represents

one quarter of a five spot waterflooding production/injection scheme. The injector is

located in the lower left corner and the producer is located in the upper right corner,

see Figure 4.5. Thus, flow is perpendicular to the diagonal streak of low permeability

values aligned in the 123o direction in the lower left corner. All other flow properties,

except the block absolute permeabilities, have been held constant. The initially oil

saturated grid blocks are subjected to waterflooding and the fractional flow of oil

versus time4 is recorded. The three response variables retained are the time to achieve

breakthrough (5% water cut 5), the time to achieve a 95% water cut, and the time

required to recover 50% of the oil.

4The time unit is simply referred to as generic “time units” since the time is used only in arelative sense. For a constant rate injection scheme, as adopted in this problem, the actual timewould be directly related to the pore volumes injected.

5The water cut is the fractional flow rate of water. In cases, where only oil and water are flowingthe fractional flow of water is 1.0 minus the fractional flow of oil.

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES112

Fractional Flow of Oil versus Time

Time

Oil

Fra

ctio

n

0 10.0 20.0 30.0 40.0

.00

.20

.40

.60

.80

1.00

Figure 4.6: Fractional flow of oil for the reference Berea permeability distribution.

Figure 4.6 shows the reference waterflooding results, more precisely, the fractional

flow of oil versus time. The three reference response values are:

Time to reach 5% water cut = 4.178 time units

Time to reach 95% water cut = 29.956 time units

Time to recover 50% of the oil = 17.091 time units

The problem is to assess the uncertainty about these response variables using only

the univariate distribution of permeability values F (z) as shown on Figure 4.2, and

the normal scores semivariogram γY (h) shown on Figure 4.3 and modeled in equation

(4.2).

4.2 Stochastic Simulation and Output Uncertainty

The sequential Gaussian simulation (SGS) technique (see section 2.3.6), the sequen-

tial indicator simulation (SIS) technique (see section 2.3.7), and the annealing (SAS)

method (see section 2.5 and Chapter 3) have each been used to generate 100 uncon-

ditional (no local conditioning data) realizations of a square 40 by 40 grid the same

size as the reference Berea image of Figure 4.1. The acronyms SGS, SIS, and SAS

are used in the following text and figures to shorten the presentation.

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES113

The model distribution F (z) of 1600 values was used along with the normal scores

semivariogram model (4.2) as prior input information. Other implementation details

are given below:

Sequential Gaussian Simulation:

The sgsim program of GSLIB [40] was used to generate realizations with sequential

Gaussian simulation (SGS). A random path was followed in the sequential algorithm.

At each node the 24 closest previously simulated nodal values were retained for the

required simple kriging.

Each realization was generated in a reasonable 7.741 seconds of CPU time on a

DEC 5000 workstation (less than 15 minutes for the one hundred realizations). The

first four realizations, as shown on Figure 4.7, reproduce the important features of

the Berea sandstone. For example, the histogram of permeability, as visually appre-

ciated by the areal extent of different gray levels, is generally reproduced as is the

overall pattern of variability and anisotropy. These simulated realizations are consid-

ered plausible realizations of the Berea sandstone. The underlying model distribution

(cdf) was that of the 1600 reference values of Figure 4.2. However, statistical fluctu-

ations about this cdf model cause any one realization to have a different cdf. This

is illustrated on Figure 4.8 which shows a quantile-quantile plot with all 100 realiza-

tions. In average, the distribution of the SGS realizations is seen to reproduce the

reference distribution; however, there can be significant departures from the refer-

ence. These ergodic fluctuations contribute to the space of uncertainty defined by

SGS realizations. Section 4.3 considers how much of the uncertainty is due to these

ergodic fluctuations and how much is due to the higher order multivariate spatial

distribution.

The normal scores semivariogram of each realization also deviates from the model

semivariogram. This is illustrated on Figure 4.9 which shows the semivariogram,

for all 100 realizations, along the two principal directions. These fluctuations are

substantial and can not be removed by any conventional transformation procedure.

A more detailed look at ergodicity and the statistical fluctuations that can be expected

is given in Appendix B.

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES114

SGS Realization - One SGS Realization - Two

SGS Realization - Three SGS Realization - Four

Figure 4.7: The first four SGS realizations using the Berea permeability histogram andnormal scores semivariogram.

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES115

Q-Q Plot: SGS Realizations versus Reference

Reference

Sim

ulat

ed

0. 20. 40. 60. 80. 100. 120.

0.

20.

40.

60.

80.

100.

120.

Figure 4.8: A quantile-quantile plot of all 100 SGS realizations. Two identical distributionswould plot as a straight line at 45o-degrees (the gray line). In average, the distribution ofthe SGS realizations is close to the reference distribution.

Sequential Indicator Simulation:

The sisim program of GSLIB [40] was used to generate realizations with sequential

indicator simulation (SIS). A median IK approach (see section 2.3.7) was considered

with the stationary normal scores covariance model (4.2).

When the indicator thresholds or cutoffs are defined on p-quantiles qp the indicator

data i(u; qp) are the same regardless of the univariate distribution. For example,

choosing a first decile cutoff on the z-permeability data, zp=0.10 such that FZ(zp=0.10) =

0.10, yields the same indicator data as a first decile cutoff applied to the y-normal

scores data, yp=0.10 such that FY (yp=0.10) = 0.10, i.e.,

i(u; zp) = i(u; yp) when FZ(zp) = FY (yp)

Thus, the same results are obtained by performing the indicator simulation in the

normal space (with the nine deciles of the standard normal distribution) or directly

in the z-permeability space (with the nine deciles of the original Berea permeability

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES116

Normal Scores Semivariogram: 123 degrees

Distance

Sem

ivar

iogr

am

0 5.0 10.0 15.0 20.0 25.0

.00

.40

.80

1.20

1.60

Normal Scores Semivariogram: 33 degrees

Distance

Sem

ivar

iogr

am

0 5.0 10.0 15.0 20.0 25.0

.00

.40

.80

1.20

1.60

Figure 4.9: The normal scores semivariogram for all 100 SGS realizations in the twoprincipal directions. The model semivariogram is shown by the gray curve in the centerof the variogram cloud. The horizontal dashed line at a value of 1.0 represents the modelvariance. In average, the semivariogram of the SGS realizations is close to the modelsemivariogram.

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES117

distribution). To be consistent with SGS and annealing, and to make it straightfor-

ward to check reproduction of the stationary normal scores semivariogram, simulation

was performed in the normal space (with the nine standard normal deciles: -1.282,

-0.842, -0.524, -0.253, 0.000, 0.253, 0.524, 0.842, 1.282)

The normal scores semivariogram model (4.2) is adopted for all standardized in-

dicator semivariograms, and simple kriging was performed at all thresholds using the

24 closest previously simulated nodal values. At any particular location the data

configuration and the standardized semivariogram do not change from one cutoff to

the next. Therefore, at any one location the kriging weights do not change from one

cutoff to the next and only one kriging system need be solved with the same weights

applied to the different indicator data at all thresholds.

It took 12.12 seconds of CPU time on a DEC 5000 workstation to generate 1

realization (slightly over 20 minutes for the one hundred realizations). This is con-

sidered within practical limits. The first four realizations as shown on Figure 4.10,

are considered plausible realizations of the Berea permeability distribution since they

reasonably reproduce the histogram and general variability of the reference image.

Figure 4.11 shows a quantile-quantile plot of 100 realization cdf’s compared to the

reference distribution. In average, the distribution of a SIS realization is close to the

reference distribution. Once again the statistical (or ergodic) fluctuations are seen to

be significant (compare with the fluctuations obtained with the Gaussian method as

shown on Figure 4.8).

The semivariogram of each realization also fluctuates around the model semivari-

ogram. This is illustrated on Figure 4.12 which shows the semivariogram, for all 100

realizations, in the two principal directions. These fluctuations are significantly larger

than those observed with the SGS realizations, see Figure 4.9.

Annealing Simulation:

The sasim program (see Chapter 3, Appendix D, and GSLIB [40]) was used to gener-

ate realizations with the annealing (SAS) technique. The realizations were generated

in the normal space with the same distribution and semivarigoram as used by SGS

and SIS. The initial realization for each simulation was created by assigning the 1600

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES118

SIS Realization - One SIS Realization - Two

SIS Realization - Three SIS Realization - Four

Figure 4.10: The first four SIS realizations (generated with median indicator simulation)using the Berea permeability histogram and normal scores semivariogram.

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES119

Q-Q Plot: SIS Realizations versus Reference

Reference

Sim

ulat

ed

0. 20. 40. 60. 80. 100. 120.

0.

20.

40.

60.

80.

100.

120.

Figure 4.11: A quantile-quantile plot of all 100 SIS realizations. In average, the distributionof the SIS realizations is close to the reference distribution.

nodal values standard normal random deviates. Annealing was carried out with the

simulated annealing decision rule and the following objective function:

O =nh∑l=1

[γmodel(hl) − γrealization(hl)]2 (4.4)

The nh = 16 separation vectors h are defined by the offsets6 given on Table 4.1 and

graphically illustrated on Figure 4.13.

γmodel(hl) is the control or model variogram specified in (4.2), and γrealization(hl)

is the experimental semivariogram of the candidate realization.

It took 54.348 seconds of CPU time on a DEC 5000 workstation to perform 500,000

swaps and generate 1 realization when using a slow annealing schedule, see section

2.5, with parameters t0=1.0, λ = 0.5, Kmax = 160000, Kaccept = 16000, S = 5,

and ∆O = 0.000001, and running sasim until completion. Without changing the

annealing schedule, this time can be reduced by 50% by noticing that the objective

6It is relevant to state explicitly the offsets in the X and Y direction since this is the format usedby the sasim program.

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES120

Normal Scores Semivariogram: 123 degrees

Distance

Sem

ivar

iogr

am

0 5.0 10.0 15.0 20.0 25.0

.00

.40

.80

1.20

1.60

Normal Scores Semivariogram: 33 degrees

Distance

Sem

ivar

iogr

am

0 5.0 10.0 15.0 20.0 25.0

.00

.40

.80

1.20

1.60

Figure 4.12: The normal scores semivariogram for all 100 SIS realizations in the twoprincipal directions. The model semivariogram is shown by the gray curve in the center ofthe variogram cloud.

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES121

Lag Number X (horizontal offset) Y (vertical offset)1 1 02 0 13 1 14 1 -15 2 06 0 27 2 28 2 -29 2 110 1 211 1 -212 2 -113 3 -214 6 -415 2 316 4 6

Table 4.1: The lag vectors used with annealing to generate unconditional simulations withthe Berea distribution and normal scores semivariogram. These vectors are graphicallyillustrated on Figure 4.13.

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES122

1

2 3

4

5

6 7

8

9

10

11

12

13

14

15

16

Figure 4.13: The lag vectors used to generate unconditional realizations of Berea. The lagnumbers correspond to those given on Table 4.1.

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES123

Objective Function versus Time

Number of swaps (1000’s)

Obj

ectiv

e F

unct

ion

0 100 200 300 400 500

.00

.20

.40

.60

.80

1.00

Figure 4.14: Objective function versus the number of swaps (CPU time) for the slowannealing schedule.

function O had reached a near zero value half way through the simulation (after

250,000 swaps), see Figure 4.14. Each swap takes essentially the same number of

arithmetic operations7; therefore, the CPU time is directly related to the number

of swaps. The sudden vertical drops in the objective function happen when the

temperature is decreased. When generating the first realization for a new case it is

good practice to start with a slow annealing schedule and then move to faster ones

once a reference realization has been generated. In some cases a too fast annealing

schedule will cause artifacts such as banding and positioning of the high (or low)

values near the edges of the simulated field.

The default annealing schedule with parameters t0=1.0, λ = 0.1, Kmax = 160000,

Kaccept = 16000, S = 5, and ∆O = 0.000001, arrives at realizations even faster. As

shown on Figure 4.15 the objective function reaches a near zero value after 80000

swaps or 8.696 seconds of CPU time on a DEC 5000 workstation. This compares

favorably with 7.741 seconds for a SGS realization and 12.120 seconds for a SIS real-

ization. The first four realizations generated with the default annealing schedule are

shown on Figure 4.16. These realizations are seen to reproduce the general variability

and anisotropy of the Berea image and are considered plausible realizations.

7Fewer operations are needed near the edges of the grid since there are fewer pairs to update.

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES124

Objective Function versus Time

Number of swaps (1000’s)

Obj

ectiv

e F

unct

ion

0 100 200 300 400 500

.00

.20

.40

.60

.80

1.00

Figure 4.15: Objective function versus the number of swaps (CPU time) for the fastannealing schedule.

Figure 4.17 shows a quantile-quantile plot of all 100 realization cdf’s compared

to the reference distribution. Each of the annealing realization cdf’s is seen to be

very close to the reference distribution since the initial cdf is not changed by the

annealing/perturbation process. The 1600 nodal values are each initially assigned

by randomly drawing from a standard normal distribution. The fluctuations on Fig-

ure 4.17 are the result of this drawing. Compare with the fluctuations obtained with

SGS and SIS shown on Figures 4.8 and 4.11.

The semivariogram of each realization deviates from the model semivariogram to

a much lesser degree than either SGS or SIS. This is illustrated on Figure 4.18 (to

be compared with Figures 4.9 and 4.12) which shows the semivariogram, for all 100

realizations, in the two principal directions. The fluctuations beyond 7-8 distance

units arise because the corresponding model semivariogram values do not enter the

objective function.

In most practical situations the model statistics (univariate distribution and semi-

variogram) have been inferred from a limited number of sample data; thus, one expects

a certain amount of sampling uncertainty. In such cases, the tight control provided

by the SAS realizations may not be desireable.

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES125

SAS Realization - One SAS Realization - Two

SAS Realization - Three SAS Realization - Four

Figure 4.16: The first four annealing realizations using the Berea permeability histogramand normal scores semivariogram.

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES126

Q-Q Plot: SAS Realizations versus Reference

Reference

Sim

ulat

ed

0. 20. 40. 60. 80. 100. 120.

0.

20.

40.

60.

80.

100.

120.

Figure 4.17: A quantile-quantile plot of all 100 annealing realizations. The distribution ofeach annealing realization is close to the reference distribution.

Practical Considerations in Choosing Annealing Simulation Parameters:

The annealing-based simulated realizations shown on Figure 4.16 appear reasonable

when compared to the SGS and SIS realizations shown on Figures 4.7 and 4.10.

Depending on the objective function and the annealing schedule it is possible to get

unreasonable realizations. The following empirical rules can be followed to avoid such

unreasonable results:

1. In general, the separation vectors chosen to enter the objective function should

cover a compact volumetric extent as illustrated on Figure 4.13 (in this case,

the two lags at longer distances help the results and do not lead to artifacts).

For example, choosing four lags that are not contiguous, i.e., (2,3), (4,6), (3,-

2), (6,-4), (shown on Figure 4.19) will yield realizations like those shown on

Figure 4.20. The objective function can be lowered to zero with realizations

showing a checkerboard appearance. A checkerboard is perfectly continuous for

every second lag.

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES127

Normal Scores Semivariogram: 123 degrees

Distance

Sem

ivar

iogr

am

0 5.0 10.0 15.0 20.0 25.0

.00

.40

.80

1.20

1.60

Normal Scores Semivariogram: 33 degrees

Distance

Sem

ivar

iogr

am

0 5.0 10.0 15.0 20.0 25.0

.00

.40

.80

1.20

1.60

Figure 4.18: The normal scores semivariogram for all 100 annealing realizations in the twoprincipal directions. The model semivariogram is shown by the gray curve in the center ofthe variogram cloud.

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES128

1

2

3

4

Figure 4.19: The four lag vectors used to generate unconditional realizations with artifactcheckerboard appearance. The lag numbers correspond to offsets (2,3), (4,6), (3,-2), (6,-4).This figure is to be compared to Figure 4.13 which does not lead to any artifacts.

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES129

SAS with four lags SAS with four lags

Figure 4.20: Two SAS realizations using the offset pairs of (2,3), (4,6), (3,-2), (6,-4) shownin Figure 4.19. The separation of the pairs leads to an unreasonable checkerboard appear-ance.

2. To reduce the required CPU time it is tempting to reduce quickly the temper-

ature in the annealing schedule. In some cases the temperature can be lowered

quickly with no consequent artifacts. In this case, if the temperature is set to

zero each realization is generated in 10000 swaps or 1.587 seconds on a DEC

5000. Two realizations are shown on Figure 4.21 with no apparent artifacts;

however, in general a too fast annealing schedule will cause a lack of convergence

and poor quality realizations. The default annealing schedule documented in

section 2.5 and coded in sasim is typically slow enough to lead to reasonable

realizations.

Flow Simulation Results:

Flow simulation, with a constant pressure drop between the injector and producer,

was performed on each of the 3 · 100 = 300 realizations. On average, it took 281

seconds on a DEC 5000 for each flow simulation. Recall that it took 7.74, 12.12, and

8.70 DEC 5000 seconds to generate a realization with sequential Gaussian, sequential

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES130

SAS with Temperature=zero SAS with Temperature=zero

Figure 4.21: Two SAS realizations with the temperature set to zero. These realizationsappear quite reasonable.

indicator, and annealing techniques.

The space of output uncertainty defined by each method is characterized by the

distribution (histogram) of each of the three output response variables. The his-

tograms for time8 to achieve breakthrough, the time to achieve a water cut of 95%,

and the time required to recover 50% of the oil, are shown on Figures 4.22, 4.23, and

4.24 for sequential Gaussian, sequential indicator, and annealing simulation, respec-

tively.

All simulated output distributions contain the true reference value as obtained

from the reference Berea image: the reference black dots on Figures 4.22, 4.23, and

4.24 are within the 95% probability intervals. In all cases the center (mean or median)

of the output distributions is higher than the reference value. The breakthrough time

and oil recovery is faster with the reference image than the average predicted by any

of the three simulation algorithms.

This systematic difference is explained by particular features of the reference Berea

image of Figure 4.1. The low permeability near the injector and the high permeability

8For this example, the units of time are not relevant since the same response variable is beingconsidered in all cases. The time is real time and not pore volumes injected.

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES131

Fre

quen

cy

.0 2.0 4.0 6.0 8.0 10.0

.000

.040

.080

.120

.160

Response One: unconditional SGS

True = 4.178

mean 4.8717std. dev. .8532

coef. of var .1751

maximum 7.2928upper quartile 5.4651

median 4.6694lower quartile 4.2416

minimum 3.3951

Fre

quen

cy

20.0 30.0 40.0 50.0 60.0

.000

.040

.080

.120

Response Two: unconditional SGS

True = 29.956

mean 33.0602std. dev. 4.4479

coef. of var .1345

maximum 44.7830upper quartile 36.0873

median 31.9523lower quartile 30.0914

minimum 25.2867

Fre

quen

cy

10.0 15.0 20.0 25.0 30.0

.000

.040

.080

.120Response Three: unconditional SGS

True = 17.091

mean 18.7598std. dev. 2.3542

coef. of var .1255

maximum 24.8604upper quartile 20.3861

median 18.3413lower quartile 17.0085

minimum 14.4716

Figure 4.22: The simulated output distributions generated by the unconditional sequentialGaussian realizations. The upper histogram is for the time to achieve 5% water cut, themiddle histogram is for the time to achieve a 95% water cut, and the lower histogram isfor the time to recover 50% of the oil. The black dot in the box plot below each histogramis the true value obtained from the reference image, the three vertical lines are the 0.025quantile, the median (0.5 quantile), and the 0.975 quantile of the output distribution.

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES132

Fre

quen

cy

.0 2.0 4.0 6.0 8.0 10.0

.000

.040

.080

.120

.160

Response One: unconditional SIS

True = 4.178

mean 4.9474std. dev. .8028

coef. of var .1623

maximum 7.2935upper quartile 5.4046

median 4.7717lower quartile 4.3191

minimum 3.5474

Fre

quen

cy

20.0 30.0 40.0 50.0 60.0

.000

.040

.080

.120

Response Two: unconditional SIS

True = 29.956

mean 33.6503std. dev. 4.1778

coef. of var .1242

maximum 46.2240upper quartile 36.2371

median 32.9109lower quartile 30.2623

minimum 26.0475

Fre

quen

cy

10.0 15.0 20.0 25.0 30.0

.000

.040

.080

.120

Response Three: unconditional SIS

True = 17.091

mean 19.1450std. dev. 2.2606

coef. of var .1181

maximum 25.8093upper quartile 20.2333

median 18.8895lower quartile 17.4985

minimum 14.7144

Figure 4.23: The simulated output distributions generated by the unconditional sequentialindicator realizations. The upper histogram is for the time to achieve 5% water cut, themiddle histogram is for the time to achieve a 95% water cut, and the lower histogram isfor the time to recover 50% of the oil. The black dot in the box plot below each histogramis the true value obtained from the reference image, the three vertical lines are the 0.025quantile, the median (0.5 quantile), and the 0.975 quantile of the output distribution.

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES133

Fre

quen

cy

.0 2.0 4.0 6.0 8.0 10.0

.000

.040

.080

.120

Response One: unconditional SAS

True = 4.178

mean 5.1005std. dev. 1.0397

coef. of var .2038

maximum 7.6889upper quartile 5.7904

median 5.1028lower quartile 4.1890

minimum 3.2988

Fre

quen

cy

20.0 30.0 40.0 50.0 60.0

.0000

.0100

.0200

.0300

.0400

.0500

.0600

.0700

.0800Response Two: unconditional SAS

True = 29.956

mean 34.5353std. dev. 5.7947

coef. of var .1678

maximum 48.7835upper quartile 38.7852

median 34.2986lower quartile 29.7469

minimum 24.6263

Fre

quen

cy

10.0 15.0 20.0 25.0 30.0

.000

.020

.040

.060

.080

Response Three: unconditional SAS

True = 17.091

mean 19.6551std. dev. 3.2340

coef. of var .1645

maximum 27.9956upper quartile 21.9916

median 19.3641lower quartile 17.0208

minimum 14.3032

Figure 4.24: The simulated output distributions generated by the unconditional annealingrealizations. The upper histogram is for the time to achieve 5% water cut, the middlehistogram is for the time to achieve a 95% water cut, and the lower histogram is for thetime to recover 50% of the oil. The black dot in the box plot below each histogram is thetrue value obtained from the reference image, the three vertical lines are the 0.025 quantile,the median (0.5 quantile), and the 0.975 quantile of the output distribution.

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES134

SGS realization with short time (3.40)

Injector

ProducerSGS realization with long time (7.29)

Injector

Producer

Figure 4.25: The SGS realizations that yield the shortest and the longest time to reach awater cut of 5%.

near the producer are factors. Another important factor is the continuity perpendic-

ular to the flow direction. Intuitively, one would think that the continuity between

the wells (parallel to the flow direction) is the most important factor for early break-

through. However, the wells are separated by 5 times the variogram range in that

direction and there is virtually no possibility for a direct flow path between the wells.

Continuity perpendicular to the pressure gradient as present in Berea allows a more

direct and less tortuous path from the injector to producer. This is confirmed by

visually examining the SGS, SIS, and SAS realizations that generate the shortest

“time to reach a 5% water cut” shown on the left of Figures 4.25, 4.26, and 4.27.

For all three simulation algorithms the short time cases are those where the overall

permeability is high and connected in the direction perpendicular to flow.

Figures 4.25, 4.26, and 4.27 also show the the realizations that give the longest

“time to reach a 5% water cut”. The long time realizations are characterized by overall

low permeability and a lack of continuity either between the wells or perpendicular

to the flow direction.

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES135

SIS realization with short time (3.55)

Injector

ProducerSIS realization with long time (7.29)

Injector

Producer

Figure 4.26: The SIS realizations that yield the shortest and the longest time to reach awater cut of 5%.

SAS realization with short time (3.29)

Injector

ProducerSAS realization with long time (7.69)

Injector

Producer

Figure 4.27: The SAS realizations that yield the shortest and the longest time to reach awater cut of 5%.

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES136

Response Variable Technique 95% Probability Intervalq0.025 q0.975 width

Time to reach 5% SGS 3.60 6.88 3.28SIS 3.67 6.98 3.31SAS 3.49 7.25 3.76

Time to reach 95% SGS 26.15 42.77 16.62SIS 26.78 43.71 16.93SAS 24.78 46.75 21.97

Time to recover 50% SGS 15.43 24.04 8.61SIS 15.79 25.39 9.60SAS 14.57 26.52 11.95

Table 4.2: Summary of output response for SGS, SIS, and SAS realizations.

The space of output uncertainty generated by SGS, SIS, and SAS may be sum-

marized by the 95% probability intervals of the three response variables that they

generate. The 95% probability interval is bound by the 2.5% quantile (q0.025) and

the 97.5% quantile (q0.975). The width of the 95% probability interval is defined as

q0.975 − q0.025. Table 4.2 gives the 95% probability interval for the output response

distributions generated by the SGS, SIS, and SAS simulated realizations.

In all cases the width of the 95% probability interval is largest for the annealing

(SAS) realizations. This was not expected since the Gaussian and indicator input im-

ages showed significantly more ergodic fluctuations (see Figures 4.8 and 4.11 versus

4.17 and also Figures 4.9 and 4.12 versus 4.18). Moreover, the annealing realizations

are indirectly the result of an optimization procedure whereas the Gaussian and indi-

cator realizations are generated by more traditional RF-based simulation procedures.

These facts would have a-priori led one to believe that annealing would generate a

smaller space of output uncertainty.

The SGS and SIS realizations clearly reflect more univariate and bivariate vari-

ability; however, the flow simulator response is a complex multivariate characteristic.

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES137

There is no theoretical or empirical relation between input uncertainty at the uni-

variate and bivariate level and output uncertainty. By construction SGS realizations

have a multivariate Gaussian distribution. It is also known that these Gaussian im-

ages have maximum entropy (see Appendix C) which may lead to realizations with

too similar flow characteristics (i.e., a smaller space of output uncertainty).

The implicit multivariate distribution of annealing realizations is the result of

many implementation decisions (see Chapter 3) and the components entering the

objective function. The result of Table 4.2 allows the conjecture that departure from

a multivariate Gaussian distribution and its maximum entropy feature in general

leads to a larger sampled space of output uncertainty.

4.3 Uncertainty due to Ergodic Fluctuations

All three simulation algorithms generate realizations {z(l)(u),u ∈ A} whose cdf

F (l)(z) and variogram γ(l)(h) deviate from the corresponding model parameters F (z)

and γ(h). These deviations, referred to as ergodic fluctuations, lead to a larger space

of input uncertainty, i.e., realizations that honor exactly the model cdf and variogram

can be considered as a subset of all the realizations. The ergodic fluctuations of the

conventional RF-based SGS and SIS algorithms were significantly more pronounced

than those of the annealing-based algorithm. The larger the space of possible out-

comes the more conservative assessment of output uncertainty; thus, at the univariate

and bivariate level the SGS and SIS realizations present a larger space of uncertainty.

There are two relevant questions related to ergodic fluctuations: 1) how much fluc-

tuation is acceptable?, and 2) how much is the space of output uncertainty expanded

by these fluctuations? Some fluctuations are acceptable and desirable since the model

statistics are typically affected by sampling fluctuations. Note that a higher order

model of uncertainty for the model parameters (F (z) and γY (h)) would be needed to

rigorously evaluate how much fluctuation is acceptable.

In many applications, the model statistics are inferred from sparse samples and

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES138

cannot be considered as perfectly known. Therefore, deviations from the model statis-

tics are desirable because they address a real aspect of the inherent uncertainty. Al-

though ergodic fluctuations have the same impact on the output response variables

as uncertain or randomized input model parameters; the source of the variability is

quite different. Ergodic fluctuations arise only because the physical size of the domain

being simulated is not infinite (see Appendix B). All simulation algorithms, regardless

of the ergodic fluctuations, implicitly consider that the model statistics are perfectly

known.

A worthwhile goal would be to introduce explicitly the uncertainty in the model

statistics, e.g., in some parameters of the cdf F (z) and variogram γ(h). A model

of uncertainty in these parameters could be constructed and Monte Carlo drawings

from these models would yield the parameters to be used for each realization. Thus,

fluctuations from the model parameters would be explicitly controlled. This approach

has not been considered in this dissertation and may be an interesting topic of future

research.

The other question regarding ergodic fluctuations which has been investigated

in this section is “how much is the output space of uncertainty expanded by these

fluctuations?”. To address this question one needs a subset of realizations that exactly

honor the conditioning statistics.

One idea would be to generate many realizations and keep only those that honor

the model distribution F (z) and model variogram γY (h) with some strict tolerance.

This idea was not pursued since a prohibitively large number of realizations would

have to be created to find enough that simultaneously honor F (z) and γY (h). Rather

than generate new realizations, a univariate graphical transformation procedure (see

[90], p. 478 and [40], p. 191) was used to transform the 100 SGS and the 100

SIS realizations to exactly match the model distribution F (z). Thus, the ergodic

fluctuations of F (z) can be isolated while leaving the fluctuations in the variogram9.

The transform is achieved by replacing each z-value of a particular image by a

9There is no such straightforward bivariate transform to remove fluctuations from a referencecovariance or variogram.

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES139

rank-preserving graphical transform, i.e,

z = F−1(F (l)(z′)) (4.5)

where F−1(·) is the inverse of the reference (model) distribution, F (l)(·) is the distri-

bution for simulation (l), z′ is a simulated value from F (l)(z), and z is the transformed

value.

The results of this histogram identification are illustrated on Figure 4.28 where

the third and fourth SGS realizations (the bottom of Figure 4.7) have been trans-

formed to reproduce exactly the distribution of Figure 4.2. The transform does not

change the spatial arrangement of high and low permeabilities; only the magnitude is

changed so that the reference histogram is reproduced. For example, the permeability

distribution shown on the upper left of Figure 4.28 had a too large proportion of low

permeability (light colored pixels). These low permeabilities have been rescaled to

obtain the distribution shown on the upper right of Figure 4.28 which has the correct

proportion of high and low permeability. Similarly, the distribution shown on the

lower left of Figure 4.28 had a too large proportion of high permeability (dark colored

pixels). The rescaled distribution shown on the lower right has the correct proportion

of each permeability type.

The univariate distribution of all 100 SGS and 100 SIS realizations F (l)(z) have

been transformed to identify the reference distribution F (z). The numerical flow

simulation was repeated to obtain the new response values. The histograms of output

response values are shown on Figures 4.29 and 4.30. The space of output uncertainty

generated before and after the univariate transformation is summarized by the 95%

probability intervals of the three response variables shown on Table 4.3.

In general, the space of uncertainty, as measured by the width of the 95% proba-

bility interval, decreases when the ergodic fluctuations in the histogram are removed.

The width of the 95% probability interval has been reduced by an average of 9% for

the SGS realizations and by an average of 25% reduction for the SIS realizations. The

larger reduction in the case of the SIS realizations is expected since the original 100

realizations showed significantly more ergodic fluctuations in the histogram than the

SGS realizations (compare Figures 4.8 and 4.11).

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES140

SGS Realization - Three SGS Realization - Transformed Three

SGS Realization - Four SGS Realization - Transformed Four

Figure 4.28: Two SGS realizations (the third and fourth realizations on Figure 4.7) beforeand after being transformed to identify the reference Berea permeability histogram.

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES141

Fre

quen

cy

.0 2.0 4.0 6.0 8.0 10.0

.000

.040

.080

.120

.160Response One: transformed SGS

True = 4.178

mean 4.9101std. dev. .6673

coef. of var .1359

maximum 6.6414upper quartile 5.3533

median 4.9087lower quartile 4.3895

minimum 3.6375

Fre

quen

cy

20.0 30.0 40.0 50.0 60.0

.000

.040

.080

.120

Response Two: transformed SGS

True = 29.956

mean 33.3955std. dev. 3.4843

coef. of var .1043

maximum 42.9719upper quartile 35.1463

median 33.2680lower quartile 30.7249

minimum 27.2026

Fre

quen

cy

10.0 15.0 20.0 25.0 30.0

.000

.040

.080

.120

Response Three: transformed SGS

True = 17.091

mean 19.0120std. dev. 1.9876

coef. of var .1045

maximum 25.2365upper quartile 19.9726

median 18.7707lower quartile 17.6491

minimum 15.5901

Figure 4.29: The simulated output distributions generated by the transformed uncondi-tional Gaussian realizations. The upper histogram is for the time to achieve 5% water cut,the middle histogram is for the time to achieve a 95% water cut, and the lower histogram isfor the time to recover 50% of the oil. The black dot in the box plot below each histogramis the true value obtained from the reference image, the three vertical lines are the 0.025quantile, the median (0.5 quantile), and the 0.975 quantile of the output distribution.

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES142

Fre

quen

cy

.0 2.0 4.0 6.0 8.0 10.0

.000

.040

.080

.120

.160

Response One: transformed SIS

True = 4.178

mean 4.7795std. dev. .6098

coef. of var .1276

maximum 6.5578upper quartile 5.2285

median 4.7133lower quartile 4.3086

minimum 3.5482

Fre

quen

cy

20.0 30.0 40.0 50.0 60.0

.000

.040

.080

.120

.160

Response Two: transformed SIS

True = 29.956

mean 32.5745std. dev. 3.0376

coef. of var .0933

maximum 42.3085upper quartile 34.7588

median 32.0443lower quartile 30.4497

minimum 26.4870

Fre

quen

cy

10.0 15.0 20.0 25.0 30.0

.000

.040

.080

.120

Response Three: transformed SIS

True = 17.091

mean 18.5452std. dev. 1.7229

coef. of var .0929

maximum 23.4879upper quartile 19.7219

median 18.2691lower quartile 17.1133

minimum 15.0275

Figure 4.30: The simulated output distributions generated by the transformed uncondi-tional indicator realizations. The upper histogram is for the time to achieve 5% water cut,the middle histogram is for the time to achieve a 95% water cut, and the lower histogram isfor the time to recover 50% of the oil. The black dot in the box plot below each histogramis the true value obtained from the reference image, the three vertical lines are the 0.025quantile, the median (0.5 quantile), and the 0.975 quantile of the output distribution.

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES143

Response Variable Technique 95% Probability Interval % reductionq0.025 q0.975 width

Time to reach 5% SGS (before) 3.60 6.88 3.28SGS (after) 3.80 6.98 3.18 -3%SIS (before) 3.67 6.98 3.31SIS (after) 3.69 6.48 2.79 -16%

Time to reach 95% SGS (before) 26.15 42.77 16.62SGS (after) 27.85 42.37 14.52 -13%SIS (before) 26.78 43.71 16.93SIS (after) 27.06 38.88 11.82 -30%

Time to recover 50% SGS (before) 15.43 24.04 8.61SGS (after) 16.23 23.83 7.60 -11%SIS (before) 15.79 25.39 9.60SIS (after) 15.54 22.48 6.94 -28%

Table 4.3: Summary of SGS and SIS output response results before and after removingergodic fluctuations in the cdf.

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES144

The ergodic fluctuations of the SAS (annealing) realizations were minimal (see

Figure 4.17) and removing them would have yielded the same results as those shown

on Table 4.2. Although the output uncertainty is significantly higher for the SAS

realizations, this uncertainty is due to the multivariate (1600-variate) distribution

well beyond the variability controlled by the one and two-point distribution.

In practice, the output uncertainty is constrained by more than just a cdf and

variogram model. The permeability (and other rock properties) are known at the

well locations. The next section investigates how conditioning to local data affects

the output uncertainty.

4.4 Conditioning to Local Data

The realizations generated up until now have not considered any local conditioning

data z(uα), α = 1, . . . , n, so that the visual appearance and output uncertainty are

the result of only the implicit RF models. In practice, there are always some local

conditioning data; in particular, the petrophysical properties are typically known at

the well locations through direct measurements or interpretation of various well log

responses.

Sequential Gaussian, sequential indicator, and annealing simulation as documented

in section 2.2 have been repeated with two local conditioning data as taken from the

reference map of Figure 4.1: the permeability at the injector (45.0 md) and the per-

meability at the producer (88.0 md). In all three cases the conditioning is achieved

by fixing these two nodal values at the start of the simulation procedure. In the case

of sequential Gaussian and sequential indicator simulation these two data are used

to condition all remaining 1598 nodes. In the case of annealing these two data are

left unperturbed during the entire annealing procedure. Four conditional realizations

for SGS, SIS, and SAS are shown on Figures 4.31, 4.32, and 4.33 respectively. In all

cases the permeability in the upper right corner is higher (due to the producing well

permeability of 88 md) and the permeability in the lower left corner is lower (due to

the injecting well permeability of 45 md).

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES145

The flow simulation was repeated with the 3 · 100 = 300 realizations. The his-

tograms of flow response variables are shown on Figures 4.34, 4.35, and 4.36. The

results for all runs, including these conditional realizations, are shown on Table 4.4.

It appears that adding these two conditioning data has not decreased the space of

uncertainty. Two data are not sufficient to constrain the spatial distribution to an

amount noticeable by the transfer function used here (flow simulator).

4.5 Summary of the Results

The primary result of this chapter is that the annealing-based simulation results

appear to fulfill the requirements of a good simulation technique. That is,

• Plausible realizations were generated in a reasonable amount of computer time.

It took 8.696 seconds of CPU time on a DEC 5000 workstation for an SAS

realization. This compares with 7.741 seconds for a SGS realization and 12.120

seconds for a SIS realization.

• All of the input data were honored. The univariate distribution and the vari-

ogram were honored more exactly than either the SGS or the SIS realizations

(see Figure 4.17 versus 4.8 and 4.11 and also Figure 4.18 versus 4.9 and 4.12).

This is not necessarily an advantage since the ergodic fluctuations of SGS and

SIS could be considered as accounting for sampling uncertainty in the model

statistics.

• Finally, the annealing-based realizations seem to generate a space of output

uncertainty comparable to both SGS and SIS. In fact, in all cases the SAS

realizations generated a slightly larger space of uncertainty (see Table 4.4).

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES146

Case Time to reach 5% Time to reach 95% Time for 50% recoveryq0.025 q0.975 width q0.025 q0.975 width q0.025 q0.975 width

Unconditional SimulationsSGS 3.60 6.88 3.28 26.15 42.77 16.62 15.43 24.04 8.61SIS 3.67 6.98 3.31 26.78 43.71 16.93 15.79 25.39 9.60SAS 3.49 7.25 3.76 24.78 46.75 21.97 14.57 26.52 11.95

Transformed Unconditional SimulationsSGS 3.80 6.98 3.18 27.85 42.37 14.52 16.23 23.83 7.60SIS 3.69 6.48 2.79 27.06 38.88 11.82 15.54 22.48 6.94

Conditional SimulationsSGS 3.58 6.84 3.26 26.06 42.94 16.88 14.80 23.96 9.16SIS 3.66 7.27 3.61 26.33 45.32 18.99 15.69 26.04 10.35SAS 3.49 7.47 3.98 25.59 48.59 23.00 14.66 28.30 13.64

Table 4.4: Summary of all output response values. The unconditional realizations consideronly a distribution model F (z) and a variogram model γY (h). The transformed resultscorrespond to the same unconditional SGS and SIS results with the ergodic fluctuations inthe cdf F (l)(z) removed. The conditional results consider two local conditioning data; oneat each well. The output uncertainty generated by each simulation method is the width ofthe 95% probability interval (q0.975 − q0.025).

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES147

SGS Conditional Realization - One SGS Conditional Realization - Two

SGS Conditional Realization - Three SGS Conditional Realization - Four

Figure 4.31: The first four sequential Gaussian realizations using the Berea permeabilityhistogram, and normal scores semivariogram model, and two conditioning data.

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES148

SIS Conditional Realization - One SIS Conditional Realization - Two

SIS Conditional Realization - Three SIS Conditional Realization - Four

Figure 4.32: The first four sequential indicator realizations using the Berea permeabilityhistogram, and normal scores semivariogram model, and two conditioning data.

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES149

SAS Conditional Realization - One SAS Conditional Realization - Three

SAS Conditional Realization - Three SAS Conditional Realization - Four

Figure 4.33: The first four annealing realizations using the Berea permeability histogram,and normal scores semivariogram model, and two conditioning data.

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES150

Fre

quen

cy

.0 2.0 4.0 6.0 8.0 10.0

.000

.040

.080

.120

Response One: conditional SGS

True = 4.178

mean 4.9582std. dev. .8605

coef. of var .1736

maximum 8.7430upper quartile 5.3872

median 4.8200lower quartile 4.3492

minimum 3.4646

Fre

quen

cy

20.0 30.0 40.0 50.0 60.0

.000

.020

.040

.060

.080

.100

Response Two: conditional SGS

True = 29.956

mean 33.2936std. dev. 4.4735

coef. of var .1344

maximum 48.7133upper quartile 35.4617

median 32.9098lower quartile 30.1443

minimum 25.7926

Fre

quen

cy

10.0 15.0 20.0 25.0 30.0

.000

.040

.080

.120

Response Three: conditional SGS

True = 17.091

mean 18.7995std. dev. 2.4061

coef. of var .1280

maximum 27.1050upper quartile 20.0251

median 18.7472lower quartile 16.9775

minimum 14.6602

Figure 4.34: The simulated output distributions generated by the conditional Gaussianrealizations. The upper histogram is for the time to achieve 5% water cut, the middlehistogram is for the time to achieve a 95% water cut, and the lower histogram is for thetime to recover 50% of the oil. The black dot in the box plot below each histogram is thetrue value obtained from the reference image, the three vertical lines are the 0.025 quantile,the median (0.5 quantile), and the 0.975 quantile of the output distribution.

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES151

Fre

quen

cy

.0 2.0 4.0 6.0 8.0 10.0

.000

.040

.080

.120

.160Response One: conditional SIS

True = 4.178

mean 5.0428std. dev. .8358

coef. of var .1657

maximum 8.3816upper quartile 5.4509

median 4.9631lower quartile 4.4377

minimum 3.4332

Fre

quen

cy

20.0 30.0 40.0 50.0 60.0

.000

.040

.080

.120

Response Two: conditional SIS

True = 29.956

mean 34.1732std. dev. 4.5198

coef. of var .1323

maximum 51.9611upper quartile 36.5571

median 33.4862lower quartile 31.1628

minimum 26.0228

Fre

quen

cy

10.0 15.0 20.0 25.0 30.0

.000

.040

.080

.120

Response Three: conditional SIS

True = 17.091

mean 19.4290std. dev. 2.5186

coef. of var .1296

maximum 28.7117upper quartile 20.5566

median 18.8896lower quartile 17.7288

minimum 14.6489

Figure 4.35: The simulated output distributions generated by the conditional indicatorrealizations. The upper histogram is for the time to achieve 5% water cut, the middlehistogram is for the time to achieve a 95% water cut, and the lower histogram is for thetime to recover 50% of the oil. The black dot in the box plot below each histogram is thetrue value obtained from the reference image, the three vertical lines are the 0.025 quantile,the median (0.5 quantile), and the 0.975 quantile of the output distribution.

CHAPTER 4. A COMPARATIVE STUDY OF VARIOUS SIMULATION TECHNIQUES152

Fre

quen

cy

.0 2.0 4.0 6.0 8.0 10.0

.000

.020

.040

.060

.080

.100

Response One: conditional SAS

True = 4.178

mean 5.1294std. dev. 1.1318

coef. of var .2207

maximum 8.0487upper quartile 5.8276

median 5.0934lower quartile 4.1983

minimum 3.3466

Fre

quen

cy

20.0 30.0 40.0 50.0 60.0

.000

.020

.040

.060

.080

Response Two: conditional SAS

True = 29.956

mean 34.8327std. dev. 6.4776

coef. of var .1860

maximum 52.2646upper quartile 38.8941

median 34.3124lower quartile 29.5458

minimum 24.7293

Fre

quen

cy

10.0 15.0 20.0 25.0 30.0

.0000

.0100

.0200

.0300

.0400

.0500

.0600

.0700

.0800Response Three: conditional SAS

True = 17.091

mean 19.9429std. dev. 3.6634

coef. of var .1837

maximum 29.4332upper quartile 22.2814

median 19.5170lower quartile 16.9574

minimum 14.5196

Figure 4.36: The simulated output distributions generated by the conditional annealingrealizations. The upper histogram is for the time to achieve 5% water cut, the middlehistogram is for the time to achieve a 95% water cut, and the lower histogram is for thetime to recover 50% of the oil. The black dot in the box plot below each histogram is thetrue value obtained from the reference image, the three vertical lines are the 0.025 quantile,the median (0.5 quantile), and the 0.975 quantile of the output distribution.

Chapter 5

Advanced Applications of

Annealing Techniques

This chapter builds on the concepts developed in earlier chapters and presents a

number of advanced applications of annealing-based simulation algorithms.

Section 5.1 develops the application of annealing to three general geostatistical

problems. The first problem is that of conditioning stochastic realizations to multiple-

point connectivity functions. The annealing-based algorithm to achieve this condi-

tioning is demonstrated by post-processing selected unconditional Berea realizations

of Chapter 4. The improvement brought by these multiple-point statistics is shown

visually and quantified by the output distribution of three flow-related response vari-

ables. A second application addresses the problem of transforming a data set so that

it matches either a full bivariate distribution model, or some summary bivariate mo-

ments such as the variogram. The third general application relates to the integration

of a secondary variable. The algorithm is illustrated by mapping porosity in a West

Texas reservoir using both well and seismic data.

Section 5.2 presents the application of annealing to the characterization and sim-

ulation of complex geological structures. A number of examples are displayed on

gray and color scale maps for a visual demonstration of the improvement brought by

multiple-point statistics. A case study is then developed with reference data taken

from a cross-stratified sands and silty sands distributary-mouth bar environment.

153

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES154

The improvement brought by annealing and multiple-point statistics is quantified in

the output space of uncertainty.

Section 5.3 develops a case study showing the integration of well test-derived ef-

fective absolute permeabilities. Once again, the contribution of this data is quantified

by showing the impact on the distributions of output response variables.

5.1 General Geostatistical Problems

There are a number of geostatistical problems that could be approached with the

annealing methodology. The first problem is conditioning stochastic images to spatial

connectivity functions.

5.1.1 Conditioning to Connectivity Functions

Many flow-related problems are highly sensitive to the spatial connectivity of extreme

permeability values. This has been documented by many researchers in the field

of stochastic reservoir modeling, e.g., [63]. Specifically, Journel and Alabert [85]

discuss this issue and present the indicator methodology for conditioning to two-point

connectivity of extreme values. They also propose a measure of N -points connectivity

(see also equation (2.15) in section 2.3.1):

φ(h; N ; zc) = E

N∏j=1

I(u + (j − 1)h; zc)

(5.1)

where the indicator transform I(u; zc) is defined as 1 if the value at location u is

below the critical threshold zc and 0 otherwise. The N -point connectivity function

φ(h; N ; zc) could be interpreted as the probability of having N points aligned in the

direction of h jointly below threshold zc. The direction h is chosen as an important

direction for the transfer function being considered. For the Berea image considered in

Chapter 4 (Figure 4.2), the continuity in the direction of diagonal banding (123o) and

the lack of continuity in the perpendicular direction (33o) are considered important

for the five spot injection/production scheme.

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES155

Journel and Alabert [85] used this measure of N -point connectivity to compare

different simulation techniques. The sequential indicator algorithm was preferred over

the sequential Gaussian algorithm because it led to an N -point connectivity closer to

the reference image. The approach taken here is to show that annealing could use the

measure φ(h; N ; zc) directly as conditioning data rather than as a diagnostic statistic.

The methodology to incorporate this statistic in the objective function of annealing

was fully developed in section 3.3. The remainder of this section will develop the

application to the Berea image.

In Chapter 4, one hundred unconditional realizations of the Berea image were

generated by sequential Gaussian simulation (SGS), sequential indicator simulation

(SIS) and simulated annealing (SAS). Figure 5.1 shows the reference Berea image and

the corresponding connectivity function φ(h; N ; zc) for h approximately aligned in the

123o direction (horizontal offset = 3 and vertical offset = -2), N = 1, . . . , 10, and zc

at the first decile threshold of 35.5 md (same as that used by Journel and Alabert

[85]). The first SGS, SIS, and SAS realizations are also shown with the corresponding

connectivity function. The connectivity function for the SGS realization drops the

quickest which corresponds to the maximum entropy property of the multiGaussian

random function model (see Appendix C). Although the SGS realization is the farthest

from the Berea reference, none of the simulation algorithms come close to reproducing

the strong connectivity (diagonal banding) seen on the Berea reference image.

Recall that the SIS realizations generated in Chapter 4 were generated by median

indicator simulation (Section 2.3.7) with a single covariance. A better alternative

would have been to use experimental indicator variograms at a series of thresholds

including the first decile 35.5 md. This was not done in Chapter 4 because all tech-

niques, including annealing, were limited to the univariate distribution and a single

covariance function. The single covariance function retained was good, on average,

for the entire range of permeability values; however, it did not characterize the par-

ticularly strong continutity of the low permeability values. A full sequential multiple

indicator simulation was performed to generate the realization shown on Figure 5.2.

Note that the connectivity function is now better reproduced because, by construc-

tion, full indicator simulation identifies the two-point connectivity φ(h; 2; zc) at all

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES156

Berea Control PatternBerea Connectivity Function

Lag Number

Pro

babi

lity

of th

e C

onne

ctio

n

0 2 4 6 8 100.00

0.02

0.04

0.06

0.08

0.10

SGS RealizationBerea and SGS Realization

Lag Number

Pro

babi

lity

of th

e C

onne

ctio

n

0 2 4 6 8 100.00

0.02

0.04

0.06

0.08

0.10

SIS RealizationBerea and SIS Realization

Lag Number

Pro

babi

lity

of th

e C

onne

ctio

n

0 2 4 6 8 100.00

0.02

0.04

0.06

0.08

0.10

SAS RealizationBerea and SAS Realization

Lag Number

Pro

babi

lity

of th

e C

onne

ctio

n

0 2 4 6 8 100.00

0.02

0.04

0.06

0.08

0.10

Figure 5.1: The Berea control pattern is shown on the upper left. The connectivity functionfor 10 lags in the 1230 direction is shown to the right (this reference connectivity functionis repeated as the solid line on all following connectivity plots). The first SGS realization,the first SIS realization, and the first SAS realization are shown with the correspondingconnectivity functions (shown as dashed lines on the plot to the right).

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES157

Full SIS Realization

Connectivity: Berea and Full SIS Realization

Lag Number

Pro

babi

lity

of th

e C

onne

ctio

n

.0 2.0 4.0 6.0 8.0 10.0 .000

.020

.040

.060

.080

.100

Figure 5.2: A realization generated with full Sequential Indicator Simulation and the cor-responding connectivity function (dashed line). The reference Berea connectivity functionis shown as the solid line.

thresholds.

The SGS algorithm does not have the flexibility to consider multiple-point con-

nectivity measures, φ(h; N ; zc), as conditioning data; and the SIS algorithm is unable

to consider multiple-point connectivity measures, φ(h; N ; zc), for N > 2. However,

annealing could be formulated to consider these connectivity functions by adding a

component to the objective function:

Oc =10∑i=1

[φmodel(h; i; zc) − φrealization(h; i; zc)]2 (5.2)

where the 10 connectivity functions correspond to i = 1, . . . , 10. Note that i = 2

corresponds to a two-point connectivity measure. Thus, Oc calls for a combination

of 1,2,. . . ,10-variate information.

To illustrate the post-processing potential of annealing, the realizations generated

in Chapter 4 and shown on Figure 5.1, were taken as initial images for further an-

nealing processing. The temperature in the annealing schedule was set to zero since a

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES158

high starting temperature would destroy the initial structure by accepting all pertur-

bations early in the annealing procedure. The Berea reference, the initial realizations

of Figure 5.1, the post-processed realizations and all corresponding connectivity func-

tions are shown on Figure 5.3. The post-processing yields both a visual improvement

and a noticeable improvement in the connectivity function.

Although the connectivity function is closer to the reference after post processing it

remains significantly lower for most N . The objective function in the annealing was to

match both the variogram and the connectivity function. Moreover, to minimize the

perturbation from the original images, the temperature in the annealing schedule was

set to zero; consequently, the image is frozen before reaching a global optimum (i.e.,

a perfect match to both the two-point statistics and the connectivity function). Were

the additional requirement to match the two-point statistics relaxed, the connectivity

function would be honored more closely. For example, Figure 5.4 shows the result of

post processing the first SGS realization by retaining only two lags of the variogram (a

unit distance in the vertical direction and a unit distance in the horizontal direction)

and all previous 10 connectivity values. The post-processed connectivity function,

shown by the short dashed line, is now well reproduced. Note the artifact checkerboard

aspect of the post-processed image arising from the fact that the unit lag h in the

connectivity function is three horizontal grid units and two vertical grid units rather

than being defined as a continuous line.

Figure 5.5 shows the connectivity function of all 100 SAS realizations generated in

Chapter 4 before and after post-processing. The better reproduction of the connectiv-

ity function and the visual improvement in the images are interesting features. A more

consequential measure is the impact on the output space of uncertainty. One hundred

additional flow simulations, using exactly the same flow scenario as documented in

Chapter 4, were performed using the post-processed images. Figure 5.6 shows the

output space of uncertainty before and after conditioning to the connectivity func-

tion. Table 5.1 shows the 95% probability interval before and after conditioning to

the connectivity function. Note the significant 30% reduction in the width of the 95%

probability intervals.

This reduction in the output uncertainty is the direct result of adding relevant

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES159

Berea Control PatternBerea Connectivity Function

Lag Number

Pro

babi

lity

of th

e C

onne

ctio

n

0 2 4 6 8 100.00

0.02

0.04

0.06

0.08

0.10

SGS Realization SGS Post ProcessedConnectivity: Berea and SGS Realization

Lag Number

Pro

babi

lity

of th

e C

onne

ctio

n

0 2 4 6 8 100.00

0.02

0.04

0.06

0.08

0.10

SIS Realization SIS Post ProcessedConnectivity: Berea and SIS Realization

Lag Number

Pro

babi

lity

of th

e C

onne

ctio

n

0 2 4 6 8 100.00

0.02

0.04

0.06

0.08

0.10

SAS Realization SAS Post ProcessedConnectivity: Berea and SAS Realization

Lag Number

Pro

babi

lity

of th

e C

onne

ctio

n

0 2 4 6 8 100.00

0.02

0.04

0.06

0.08

0.10

Figure 5.3: The Berea control pattern is shown on the upper left. The connectivity func-tion for 10 lags in the 1230 direction is shown to the right (reproduced as the solid line onall following connectivity plots). The first SGS realization, the first SIS realization, andthe first SAS realization before and after post processing are shown with the correspond-ing connectivity functions. The long dashed line is the connectivity function before postprocessing and the short dashed line is after post processing.

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES160

SGS Post Processed

Connectivity: Berea and SGS Realization

Lag Number

Pro

babi

lity

of th

e C

onne

ctio

n

0 2 4 6 8 100.00

0.02

0.04

0.06

0.08

0.10

Figure 5.4: The first SGS realization post-processed to match the Berea connectivity func-tion. The solid line corresponds to the reference Berea image, the long dashed line tothe initial SGS realization (shown on Figure 5.3), and the short dashed line to the imageshown to the left. Note the artifacts caused by relaxing the constraint to match two-pointstatistics.

Response Variable SAS 95% Probability Interval %Realizations q0.025 q0.975 width reduction

Time to reach 5% before 3.49 7.25 3.76after 3.64 6.41 2.77 -26%

Time to reach 95% before 24.78 46.75 21.97after 27.52 41.11 13.59 -38%

Time to recover 50% before 14.57 26.52 11.95after 15.35 23.88 8.53 -29%

Table 5.1: Summary of SAS output response results before and after conditioning to includethe connectivity function.

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES161

Connectivity: SAS Realizations

Lag Number

Pro

babi

lity

of th

e C

onne

ctio

n

0 2 4 6 8 10

0.00

0.02

0.04

0.06

0.08

0.10

Connectivity: SAS Realizations After Post Processing

Lag Number

Pro

babi

lity

of th

e C

onne

ctio

n

0 2 4 6 8 10

0.00

0.02

0.04

0.06

0.08

0.10

Figure 5.5: The connectivity function for all SAS realizations before and after post pro-cessing.

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES162

Fre

quen

cy

.0 2.0 4.0 6.0 8.0 10.0

.000

.040

.080

.120

Response One: unconditional SASmean 5.1005

std. dev. 1.0397coef. of var .2038

maximum 7.6889upper quartile 5.7904

median 5.1028lower quartile 4.1890

minimum 3.2988F

requ

ency

20.0 30.0 40.0 50.0 60.0

.0000

.0100

.0200

.0300

.0400

.0500

.0600

.0700

.0800Response Two: unconditional SAS

mean 34.5353std. dev. 5.7947

coef. of var .1678

maximum 48.7835upper quartile 38.7852

median 34.2986lower quartile 29.7469

minimum 24.6263

Fre

quen

cy

10.0 15.0 20.0 25.0 30.0

.000

.020

.040

.060

.080

Response Three: unconditional SASmean 19.6551

std. dev. 3.2340coef. of var .1645

maximum 27.9956upper quartile 21.9916

median 19.3641lower quartile 17.0208

minimum 14.3032

Fre

quen

cy

.0 2.0 4.0 6.0 8.0 10.0

.000

.040

.080

.120

Response One: post-processed SASmean 4.9223

std. dev. .7714coef. of var .1567

maximum 6.6964upper quartile 5.5296

median 4.8612lower quartile 4.2540

minimum 3.5347

Fre

quen

cy

20.0 30.0 40.0 50.0 60.0

.000

.040

.080

.120Response Two: post-processed SAS

mean 33.3572std. dev. 4.0088

coef. of var .1202

maximum 43.1688upper quartile 36.7992

median 32.6902lower quartile 30.2014

minimum 26.8260

Fre

quen

cy

10.0 15.0 20.0 25.0 30.0

.000

.020

.040

.060

.080

Response Three: post-processed SASmean 18.9260

std. dev. 2.2749coef. of var .1202

maximum 24.0021upper quartile 20.6730

median 18.5639lower quartile 17.1932

minimum 14.8614

Figure 5.6: The simulated output distributions generated by the annealing realizationsbefore and after conditioning by the connectivity function. The upper histograms are forthe time to achieve 5% water cut, the middle histograms are for the time to achieve a 95%water cut, and the lower histograms are for the time to recover 50% of the oil. The blackdot in the box plot below each histogram is the true value obtained from the referenceimage, the three vertical lines are the 0.025 quantile, the median (0.5 quantile), and the0.975 quantile of the output distribution.

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES163

global conditioning statistics; it is not the result of a different random function model.

Once again, accounting for uncertainty in the model statistics used in the objective

function has been left for future research.

There are times when a single realization is all that is required. That single real-

ization may be used for predictive purposes with no attached measure of uncertainty,

in which case, uncertainty in the model statistics is not relevant. To ensure that this

lone realization yields results close to the center of the output distributions it would

be desirable to reproduce exactly the model statistics. The following section describes

how annealing could be used for this purpose.

5.1.2 Multivariate Spatial Transformation

Realizations generated by conventional simulation techniques typically show signif-

icant ergodic fluctuations, see Figure 4.9, Figure 4.18, and Appendix B. In certain

situations these ergodic fluctuations are desirable since they may be used to represent

(model) uncertainty in the input statistics due to sparse sampling. When the goal is

a close match to the RF model statistics it is desirable to remove these fluctuations.

Removing the departures from the model univariate distribution allows the output

uncertainty attributable to the multivariate distribution (beyond the univariate level)

to be assessed. A bivariate transformation procedure would be needed to assess the

output uncertainty attributable beyond the bivariate level. An interesting avenue of

research would be to assess independently each of the factors contributing to the out-

put uncertainty. These contributing factors would depend on the underlying random

function model; thus, availability of a multivariate transformation procedure would

provide a tool to understand better the impact of different RF models.

Another motivation for a multivariate transformation procedure arises from the

increasingly common practice of retaining a single simulated realization as an im-

proved “estimated” map. Indeed, stochastic simulation algorithms have proven to

be much more versatile than traditional interpolation algorithms in reproducing the

full spectrum of data spatial variability and in accounting for data of different types

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES164

and sources, whether hard or soft [137]. A single realization does not allow an assess-

ment of uncertainty; however, it is quicker to generate and process through a complex

transfer function1. Given that only one realization is going to be used, it is desirable

that this realization identifies all the input spatial statistics. In other words, the

sole realization to be retained should be such that its response values are as near as

possible to the center of the output response distributions.

A unique analytical univariate transformation is possible with the rank-preserving

graphical transform procedure discussed in section 4.3 (see also [90], p. 478), i.e,

z = F−1(F (l)(z′)) (5.3)

where F−1(·) is the inverse of the model distribution, F (l)(·) is the distribution for the

simulation, z′ is a simulated value from F (l)(z), and z is the transformed value. This

transformation procedure was used in section 4.3 to evaluate the reduction in the space

of uncertainty due to removing ergodic fluctuations in the univariate distribution.

Consider the transformation of a bivariate distribution

F (l)(u,u + h; z, z′), ∀ h, z, z′ of interest

to some reference bivariate distribution

F (u,u + h; z, z′), ∀ h, z, z′ of interest

The general algorithm for such a bivariate transformation involving more than two

points at a time would require much more than the bivariate cdf F (u,u + h; z, z′).

Indeed, consider the problem of drawing a joint realization of N RV’s Z(u + hi), i =

1, . . . , N conditioned by the set of all bivariate cdf’s F (u + hi,u + hj ; z, z′), i, j =

1, . . . , N . The general algorithm would proceed as follows:

• draw a uniform [0, 1] random number y1. The first z-simulated value is as in

(5.3):

z(u + h1) = F−1(y1) = z1

1For example, in the petroleum industry it is common practice to history match the final numericalmodel of the reservoir before using it for predictive purposes. History matching consists of fine tuningthe rock/fluid properties so that a forward simulation of past production yields the actual measuredresponse. It may not be practical to perform history matching on more than one stochastic model.

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES165

• draw a second uniform random number y2 independent of y1. The second z-

simulated value is:

z(u + h2) = F−1u+h2|u+h1

(y2|z1) = z2,

where Fu+h2|u+h1(z|z1) is the ccdf of Z(u+h2) given Z(u+h1) = z1. Everything

is fine as long as there are only two RV’s Z(u + h1) and Z(u + h2). A problem

arises already for the third variable, because Z(u+h3) must be related not only

to Z(u + h2) but also to Z(u + h1).

• draw a third independent random number y3. The third z-simulated value is:

z(u + h3) = F−1u+h3|u+h1,u+h2

(y3|z1, z2) = z3,

where F−1u+h3|u+h1,u+h2

(y3|z1, z2) is the ccdf of Z(u + h3) given Z(u + h1) =

z1 and Z(u + h2) = z2. This last ccdf actually calls for the trivariate cdf

F (u + h1,u + h2,u + h3; z, z′, z′′). That trivariate cdf ensures the consistency

between the corresponding three bivariate cdf’s.

Annealing provides an approximate solution that does not call for the multivariate

cdf by constructing an objective function that measures the difference between the

realization’s bivariate distribution and the model distribution:

O =N∑

i=1

[F (u,u + hi; z, z

′) − F (u,u + hi; z, z′)]2

(5.4)

where N is the number of point (lags) to be considered, F (u,u + hi; z, z′) is the

experimental bivariate distribution of the image being transformed, and

F (u,u + hi; z, z′) is the model bivariate distribution assumed internally consistent.

In addition to these N bivariate possibilities, the objective function may include

summaries of the full bivariate distribution, e.g., a covariance or variogram function,

∑h

[γ(h) − γ(h)]2 (5.5)

where γ(h) is the realization variogram, and γ(h) is the reference variogram for

selected separation vectors h.

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES166

The annealing procedure to lower such objective functions, (5.4) plus (5.5), was

described in Chapter 3 and is possible with the sasim program documented in Ap-

pendix E.

As an example, consider the simulation of an 1/0 indicator variable representing

the presence or absence of shale. An anisotropic indicator semivariogram model is

available that fully characterizes the bivariate spatial distribution of that binary shale

indicator. An initial image2 is shown on the upper left of Figure 5.7. The model semi-

variogram in the two directions (aligned with the sides of the square image) is shown

by the solid lines on the semivariogram plot to the right of the initial image. The

bullets represent the experimental semivariogram values computed from that initial

image. The sasim program was used to transform this initial image into identifying

the model semivariogram. The initial image was transformed into the lower left cor-

ner image of Figure 5.7. The resulting experimental indicator semivariogram, shown

as the bullets on the plot at the right, is extremely close to the model. It took 11.4

CPU seconds on a DEC 5000 to perform the initial indicator simulation (one cutoff)

and 7.2 CPU seconds to perform the annealing post processing.

As another example, consider a non-conditional simulated realization of the Berea

permeability data. A specific sequential Gaussian simulation (SGS) realization in

the normal space, before back transformation to permeability, is shown at the upper

right of Figure 5.8. The semivariogram for four lags in the two principal directions of

continuity are shown as the bullets on the plot at the lower left. The corresponding

model is shown as the solid lines. Annealing has been applied to transform this

image so that the semivariogram model is more closely reproduced. The image and

corresponding semivariogram after post processing are shown to the right. The local

structure of the image was changed significantly since there were no local conditioning

data. It took 7.74 CPU seconds on a DEC 5000 for the original SGS simulation and

11.72 seconds for the annealing post-processing. Recall that it only took 8.69 seconds

to generate a realization with annealing starting from a random image. Depending on

the spatial features of the initial image it may be faster to generate a realization with

2That initial image could be derived from any technique. In the case of Figures 5.7 and 5.8,realizations that show large departures (due to a lack of ergodicity) from the input variogram modelwere selected for post-processing.

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES167

Realization to be Transformed

Var

iogr

am

Distance

Indicator Semivariogram

.0 5.0 10.0 15.0 20.0

.000

.100

.200

.300

After Transformation

Var

iogr

am

Distance

Indicator Semivariogram

.0 5.0 10.0 15.0 20.0

.000

.100

.200

.300

Figure 5.7: An initial realization of a 1/0 indicator variable is shown in the upper left.Annealing is used to transform this image to identify the semivariogram model (shown asthe solid line on both semivariogram plots).

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES168

SGS Realization (Normal Space) Post Processed (Normal Space)V

ario

gram

Distance

Semivariogram Before Transformation

.0 4.0 8.0 12.0

.00

.20

.40

.60

.80

1.00

1.20

Var

iogr

am

Distance

Semivariogram After Transformation

.0 4.0 8.0 12.0

.00

.20

.40

.60

.80

1.00

1.20

Figure 5.8: A SGS realization (in the normal space) is shown in the upper left corner.Below this realization is the corresponding variogram reproduction for four lags in the twoprincipal directions (123o and 33o). Applying annealing with the objective to honor thevariogram model (shown as the solid line) leads to the realization shown in the upper right.The variogram of this post processed image honors the variogram very closely.

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES169

full simulated annealing starting from a random image rather than post-processing.

However, the main advantage of annealing over more conventional RF-based

stochastic simulation techniques is that the objective function can include more than

two-point statistics. The following section illustrates how a correlation coefficient with

a secondary variable could be used as a component in a global objective function.

5.1.3 Accounting for a Secondary Variable

In stochastic reservoir modeling there are usually very few direct (hard) measurements

of the attributes being mapped. Often, direct measurements are available only at well

locations and, typically, there are few wells within any one zone of the reservoir. Geo-

physical techniques provide additional measurements of sonic properties, available at

many locations, that may be correlated with the rock properties being mapped. The

quality of geophysical measurements has increased, through the use of high resolution

3-D seismic and inter-well measurements, and it now brings significant information

about the spatial distribution of rock properties such as porosity [60]. Any good sim-

ulation methodology must integrate all relevant information. Integrating geophysical

information is possible in an interpolation mode with cokriging approaches [43, 140]

or some type of trend model [100]; it is also possible to perform that integration with

annealing in a stochastic simulation mode.

The annealing algorithm described below to integrate seismic data may be gener-

alized to any situation where a secondary variable is available. For example, mapping

natural methane seepage per unit land area given sparse direct measurements and

near exhaustive satellite measurements [106].

An essential piece of information needed to account for a y-secondary variable is

a measure of correlation between the Z-primary and Y -secondary variables. When

the Z-Y correlation is very good then stochastic images of the primary z-variable will

look like the map of the secondary variable rescaled to the z-units; however, when

there is little correlation the map of the z-variable need not bear any resemblance

to the map of the y-variable. One aspect of the Z-Y correlation is reflected by the

scatterplot of collocated z-y measurements. Figure 5.9 shows an example calibration

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES170

Z -

val

ues

Y - values

Calibration Scattergram

Y(l-1) Y(l)

Prior z-pdf for class [ Y(l-1),Y(l)]

Figure 5.9: Example calibration scatterplot and the prior z-probability pdf where thesecondary variable is y in the range (yl−1, yl].

scatterplot and the corresponding conditional distribution of z-values given a specific

range of y values. A number of such conditional distributions may be extracted from

the calibration scatterplot to allow for a reasonably complete characterization of the

Z-Y relationship [140].

In many cases, the z-y correlation may be further summarized by the linear cor-

relation coefficient:

ρ =Cov(Y, Z)

σZ · σY(5.6)

The correlation coefficient between measured y-secondary values and simulated z

values at the same location could be considered as a component in the objective

function of an annealing simulation, i.e.,

Oc = [ρcalibration − ρrealization]2 (5.7)

Recall that the key criterion for a quantity to enter the objective function is that it

must be locally updatable. The correlation coefficient meets this criterion. Consider

the following computational formula,

ρ =E{Z · Y } − E{Z} · E{Y }√

(E{Z · Z} − E{Z} · E{Z}) · (E{Y · Y } − E{Y } · E{Y })(5.8)

where the expected values are replaced by discrete summations over the N locations

in the area of interest, e.g., E{Z · Y } is evaluated by ( 1N

∑Ni=1 zi · yi).

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES171

The correlation coefficient of the N simulated nodes, ρrealization, is locally updat-

able by updating each of the 5 summations, e.g., given a change from zoldi to znew

i the1N

∑Ni=1 zi · yi summation is updated as follows:

[1

N

N∑i=1

zi · yi

]new

=

[1

N

N∑i=1

zi · yi

]old

− 1

N

[(zold

i − znewi

)· yi

].

The other four summations are similarly updated. Moreover, if the annealing pertur-

bation mechanism is swapping then only 1N

∑Ni=1 zi ·yi need be updated. The measured

y-values never change and, in the case of swapping, the univariate distribution of the

z-values does not change.

An Example

A data set with 74 wells and a 3-D seismic survey were provided to the Stanford Center

for Reservoir Forecasting (SCRF) by Amoco Production Company in December 1990.

This data set, taken from a producing oil field in West Texas, was subjected to an

exhaustive geostatistical study [25]. One goal of the study was to generate realizations

of the 2-D vertically-averaged well porosity using both the well data and the 3-D

seismic energy data. This stochastic simulation exercise is repeated here with the

annealing-based approach.

The 74 well locations with gray-level coded porosity values are shown on the plan

map of Figure 5.10. The area enclosed by the solid border is 14000 feet by 15000

feet3. The inner area enclosed by the dashed line is a square 10400 feet by 10400 feet

area covered by 16900 CDP locations at which the sonic properties of the reservoir

have been measured. All future images of this reservoir are shown over the 10400 feet

by 10400 inner feet area covered by the seismic at an 80 by 80 foot resolution (130

by 130 blocks).

A histogram of the 74 porosity values is shown on the left of Figure 5.11. The

mean is 7.9 %, the coefficient of variation is a low 0.23, and there is no significant

skewness. The omnidirectional normal scores semivariogram of the porosity values

3The coordinates have been transformed to ensure confidentiality while preserving the spatialfeatures.

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES172

12000-3000-2000

12000

Location Map of Well Data

4.0

4.8

5.6

6.4

7.2

8.0

8.8

9.6

10.4

11.2

12.0

Figure 5.10: A location map of the 74 well data with the porosity coded by the gray-levelof the location circles. The inner dashed line defines the area covered by the 3-D seismicsurvey.

computed from the 74 well data is shown on Figure 5.12. The horizontal dashed line

is at the unit variance of the standard normal distribution. The solid line on this

semivariogram plot is a model with a nugget effect of 20% and a single structure

spherical model with a range of 7500 feet. This semivariogram model is retained as

the sole measure of spatial continuity of porosity.

Figure 5.13 shows the gray-level coded seismic energy map. A histogram of the

16900 seismic data values is shown on the right of Figure 5.11. As shown on Fig-

ure 5.10, 19 of the 74 wells fall outside the area covered by the seismic; this leaves 55

pairs of z-porosity values and y-seismic values to establish the calibration scatterplot

shown on Figure 5.14. The linear correlation coefficent (ρ) is 0.535.

Stochastic models of the porosity should honor the sample histogram of porosity

(right of Figure 5.11), the spatial variability of porosity (variogram of Figure 5.12),

and the correlation with the seismic data. Conventional sequential Gaussian simu-

lation (SGS) would yield realizations like the two shown on Figure 5.15. Note that

the gray scale is identical to that used for the location map shown on Figure 5.10.

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES173

Fre

quen

cy

Variable

0.0 4.0 8.0 12.0 16.0

0.000

0.040

0.080

0.120

Well Porosity Data Number of Data 74

mean 7.89std. dev. 1.78

coef. of var 0.23

maximum 11.38upper quartile 9.60

median 7.57lower quartile 6.49

minimum 3.87

Fre

quen

cy

Variable

0. 5000. 10000. 15000. 20000. 25000.

0.0000

0.0100

0.0200

0.0300

0.0400

0.0500

Seismic Energy Number of Data 16900

mean 9309.8std. dev. 5776.2

coef. of var 0.62

maximum 27700.0upper quartile 13200.0

median 8910.0lower quartile 4850.0

minimum 0.0

Figure 5.11: Histograms of 74 2-D vertically-averaged well porosity values and 16900seismic energy values.

Var

iogr

am

Distance

Porosity Normal Score Semivariogram

0. 2000. 4000. 6000. 8000. 10000.

0.00

0.40

0.80

1.20

Figure 5.12: The normal scores semivariogram based on 74 porosity values is shown by thebullets. The fitted model is shown by the solid line.

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES174

Seismic Energy

0

2800

5600

8400

11200

14000

16800

19600

22400

25200

28000

Figure 5.13: A gray scale map of the 16900 seismic energy values.

Por

osity

Energy

Porosity versus Seismic Energy

0. 5000. 10000. 15000. 20000. 25000.

0.0

4.0

8.0

12.0

Number of data 55Number trimmed 19

X Variable: mean 8980.2std. dev. 5925.3

Y Variable: mean 8.145std. dev. 1.856

correlation 0.535rank correlation 0.544

Figure 5.14: A calibration scatterplot of the 55 porosity values collocated with seismicCDP data.

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES175

First SGS Realization Second SGS Realization

Figure 5.15: Two sequential Gaussian realizations of the porosity.

These realizations are not built to honor the correlation with the calibration seis-

mic data. The correlation coefficient between the simulated porosity values and the

seismic values is 0.42 and 0.46 respectively for the first and second realization4.

Annealing could be used either to post-process the SGS realizations or to gener-

ate realizations with the sample histogram (Figure 5.11) and variogram model (Fig-

ure 5.12). The latter option was considered here: the sasim program was set up to

simulate directly the porosity with the objective function:

O = λ1

N∑i=1

[γmodel(hi) − γrealization(hi)]2 + λ2 [ρcalibration − ρrealization]2 (5.9)

The N = 18 separation vectors hi are defined by the most compact arrangement of 18

lags, see Figure 5.16. The ρcalibration = 0.535 is taken from the calibration scatterplot

shown on Figure 5.14.

Two realizations were generated with the two part objective function (5.9). Fig-

ure 5.17 shows how each component in the objective function changes as the sim-

ulation proceeds (the solid line is the deviation from the model variogram and the

4The correlation coefficient is high thanks to the many conditioning data used. With no condi-tioning data the correlation coefficient would be zero.

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES176

Figure 5.16: The lag vectors used to condition the realizations of porosity.

dashed line is the deviation from the input correlation coefficient). At 600,000 swaps

the component objective function due to the correlation coefficient drops significantly

below the component due to the variogram5. Note that the sudden drops in the objec-

tive function are due to successive temperature decreases. Both component objective

functions go to zero at the end of the annealing procedure. It took 18.7 minutes on

a DEC 5000 to perform the 1,500,000 swaps (the far right of the horizontal scale on

Figure 5.17) needed to complete the simulation. Once again, the temperature reduc-

tion parameter λ in the annealing schedule could be lowered to decrease the CPU

time (the CPU time can be reduced to 2.4 minutes without affecting the quality of

the realizations). For comparison, it takes 22.9 minutes of DEC 5000 CPU time to

generate a realization with the Markov/Bayes [140] algorithm (the mbsim program of

GSLIB [40])6.

Figure 5.18 shows two simulated realizations generated with annealing and the

objective function (5.9). Almost perfect reproduction of the model variogram is shown

below each realization. The univariate distribution of porosity is honored exactly

5This is a time at which the importance of each component objective function could be renor-malized (see the discussion in section 3.2)

6This time corresponds to a full Markov-Bayes simulation with 7 cutoffs. A Markov-Bayes sim-ulation under the median IK approximation, i.e., conditioned to a single variogram model such asconsidered in Figure 5.12, would take roughly 1/7 of the CPU time, i.e., about 3 minutes. Also, theMarkov-Bayes algorithm reproduces the corresponding prior pdf, as shown in Figures 5.9, at eachlocation.

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES177

Objective Function versus Time

Number of swaps (1000’s)

Obj

ectiv

e F

unct

ion

0 400 800 1200

0.0

0.2

0.4

0.6

0.8

1.0

Figure 5.17: The two component objective functions are plotted versus the number ofswaps. The solid line is due to the deviation from the model variogram and the dashed lineis due to the deviation of the correlation coefficient from the observed value of 0.535.

at the beginning by assigning the original nodal values by random drawing from

that reference distribution. The swapping perturbation mechanism considered in this

case does not alter the initial histogram. In both cases, the correlation coefficient

(0.535) between the simulated porosity values and seismic data has been exactly

reproduced. Figure 5.19 shows a scatterplot of the simulated porsity values from the

first realization and the seismic data, to be compared with the sample scatterplot of

Figure 5.14.

These annealing realizations appear as plausible images of the 2-D vertically av-

eraged porosity; the input control statistics are almost perfectly reproduced. These

realizations indicate a continuous zone of high porosity in the center of the map with

low porosity zones in the lower left and lower right corners. The SGS realizations

(Figure 5.15), which did not consider the seismic data, do not reveal these features

as clearly.

This example illustrates how annealing can be used to integrate geophysical in-

formation. The next section presents more examples of how annealing, with multiple

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES178

First Annealing Realization Second Annealing Realization

Sem

ivar

iogr

am

Distance

Semivariogram: Realization One

0. 1000. 2000. 3000. 4000. 5000. 6000.

.00

1.00

2.00

3.00

Sem

ivar

iogr

am

Distance

Semivariogram: Realization Two

0. 1000. 2000. 3000. 4000. 5000. 6000.

.00

1.00

2.00

3.00

Figure 5.18: Two realizations generated with annealing to match specified lags of thevariogram (see Figure 5.16) and the correlation with the seismic data. The omnidirectionalsemivariogram of each image is shown below the corresponding realization.

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES179

Por

osity

Seismic

Porosity versus Seismic Energy

0. 5000. 10000. 15000. 20000. 25000.

0.0

4.0

8.0

12.0

Number of data 16900

X Variable: mean 9309.763std. dev. 5776.214

Y Variable: mean 7.898std. dev. 1.829

correlation 0.535rank correlation 0.540

Figure 5.19: A scatterplot of the simulated porosity values (first realization) and the seismicdata. Note that the correlation coefficient is exactly the reference value of 0.535.

component objective functions, can be used to integrate complex geological struc-

tures.

5.2 Geological Structures

The objective function can be formulated to account for multiple-point statistics

(section 3.3) so that complex geological structures could be more realistically charac-

terized, and reproduced. Consequently, the resulting distributions of output response

variables should be more accurate.

An example is presented in the next section 5.2.1 where a reference image (ob-

tained by scanning a photograph) is analyzed through traditional two-point statistics

and then with multiple-point statistics. Simulated realizations are shown to demon-

strate visually the improvement brought by considering annealing to reproduce these

multiple-point statistics.

The visual appearance of simulated realizations is important; the human eye can

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES180

be very sensitive to the features shared by a simulated realization and a reference im-

age. However, goodness criteria must be based on the output distributions generated

after processing these realizations through a transfer function. A case study based on

reference data taken from a cross-stratified sands and silty sands distributary-mouth

bar environment is developed in section 5.2.2. The improvement brought by annealing

and multiple-point statistics is quantified in the output space of uncertainty.

5.2.1 An Example Application: eolian Sandstone

The eolian sandstone environment illustrated by the image shown at the top of Fig-

ure 5.20 poses severe problems for traditional stochastic imaging methods. This image

represents one side of a thin rectangular parallelepiped (1 cm by 10 cm by 20 cm) cut

from an eolian sandstone. The scanned image shown at the top of Figure 5.20 has

328 by 171 square pixels each with a continuous gray level value between 0 (white)

and 1 (black). The white areas represent clean well sorted sandstone with high per-

meability and the dark areas represent finer grained sandstone with low permeability.

For computational ease this original image was upscaled to the image shown at the

bottom of Figure 5.20. The upscaled image has 164 by 85 square integer-coded pixels

with an integer gray level code ranging from 1 (white) to 8 (black).

The normal scores variogram of the eolian sandstone is shown on Figure 5.21.

Note the significant anisotropic periodicity and the 6:1 anisotropy. All distances are

relative to the discretization units of the upscaled image with the image being 164

distance units by 85 distance units. The following model was considered, with (h1, h2)

being the coordinates in the two directions corresponding to the sides of the image:

γY (h) = 0.45 · Sph

h21

2.02+

h22

5.02

+ 0.22 · Sph

h21

0.52+

h22

5.02

+ 0.45 · Exp

h21

0.000012+

h22

40.02

+ 0.27 · DH17.0,0.298

(√h2

1

)

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES181

Original Eolian Sandstone

Upscaled Eolian Sandstone

Figure 5.20: The original scanned image of the eolian sandstone with the upscaled reference.

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES182

Var

iogr

am

Distance

Normal Scores Semivariogram

.0 10.0 20.0 30.0 40.0 50.0

.00

.20

.40

.60

.80

1.00

1.20

Figure 5.21: Normal scores variogram of the eolian sandstone in the two principal direc-tions.

(5.10)

Sph(h) is the unit range spherical variogram model defined as:

Sph(h) =

[1.5h − 0.5h3] , if h ≤ 1

1, if h ≥ 1

Exp(h) is the exponential variogram function with effective range 3 defined as:

Exp(h) = [1 − exp(−h)]

and DHd,a(h) is a dampened hole effect cosine model defined as:

DHd,a(h) = 1.0 − exp

(−h · ad

)· cos (h · a)

There are four components in the semivariogram model (5.10):

1. a short scale anisotropic spherical structure that explains 40% of the total vari-

ability (the longer range is in the horizontal h2 direction and the shorter range

is in the vertical h1 direction).

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES183

2. a second short scale anisotropic spherical structure (with a more pronounced

anisotropy along the same horizontal and vertical directions) that explains an

additional 20% of the variability,

3. a third anisotropic long range exponential structure (the range parameter is

40.0, thus, the effective range is 120.0 in the horizontal direction and 0.0 in the

vertical direction ) that explains 40% of the variability, and

4. finally, a dampened hole effect model in the vertical direction to account for the

periodicity.

Two realizations based on a Gaussian simulation algorithm (program sgsim in

GSLIB [40]) were generated and are shown on Figure 5.22. As expected the real-

izations fail to show the curvilinear well connected structures of the reference image.

The indicator variograms for the quintiles of the univariate distribution of the

eolian sandstone have been modeled with the same nested structures as the normal

scores variogram. Two realizations based on a indicator simulation algorithm (pro-

gram sisim in GSLIB [40]) were generated and are shown on Figure 5.23. Once again,

the realizations do not show the curvilinear structures or the connected features of

the reference image.

Even limited to two-point statistics the image is better characterized with full

two-point histograms which contain all of the direct and cross indicator covariances.

This possibility was considered by using the annealing program sasim (described in

Chapter 3 and documented in Appendix E). The two-point histogram for 10 lags

in four directions (offsets [1,0], [0,1], [2,1], [2,-1]) were imposed. The post-processed

realizations are shown on Figure 5.24.

Two-points statistics do a poor job of characterizing the curvilinear highly con-

nected features of this image. Boolean or marked-point processes would also be inap-

propriate due to the complex interaction between the curvilinear shapes. Simulation

procedures using multiple-point statistics would be needed to capture the curvilinear

features of this image.

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES184

Gaussian Simulation Two

Gaussian Simulation One

Figure 5.22: Two Gaussian-based simulated realizations of the eolian sandstone.

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES185

Indicator Simulation Two

Indicator Simulation One

Figure 5.23: Two indicator-based simulated realizations of the eolian sandstone.

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES186

Indicator Simulation Two: processed

Indicator Simulation One: processed

Figure 5.24: Two post processed indicator-based simulated realizations of the eolian sand-stone.

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES187

Figure 5.25: A scanned photograph of core from a distributary-mouth bar sequence [118].The darker gray levels correspond to a greater proportion of clayey/silty material. Theactual image is approximately 6 inches by 3 inches.

5.2.2 Impact on Output Response Uncertainty

In the following case-study an exhaustively sampled reference image will be repro-

duced through stochastic simulations based on RF models differing by the statistics

which constrain them. It will be shown that different models though sharing the

same covariance function C(h) and the same univariate distribution lead to real-

izations with different spatial characteristics [87] and, more importantly, result in a

different assessment of the response uncertainty.

Figure 5.25 is an image of cross-stratified sands and silty sands from a distributary-

mouth bar taken from page 151 of Sandstone Depositional Environments [118]. Fig-

ure 5.25 is approximately at 1:1 scale (6 inch by 3 inch). The grayness of the image

has been arbitrarily scaled through a monotonic function into permeability values to

yield the discrete histogram of Figure 5.26.

The reference image of Figure 5.25 is discretized into 225 x 125 square pixels. For

the rest of this study, the data and image of Figure 5.25 will be considered as the

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES188

Fre

quen

cy

Permeability, md

0. 20. 40. 60. 80. 100.

0.000

0.050

0.100

0.150

0.200

0.250Reference Permeability Distribution

Number of Data 28125mean 10.0

std. dev. 13.96coef. of var 1.39

maximum 323.8upper quartile 13.4

median 5.23lower quartile 1.05

minimum 0.0

Figure 5.26: Distribution of the reference permeability values (md).

“true” reference to be reproduced.

The reference image will be simulated using various RF models conditioned by

various global statistics such as the covariance. However, no local conditioning will

be considered, that is, no samples will be retained from the reference image.

Gaussian Model

The reference permeability values are first transformed into a standard normal distri-

bution through a normal scores transform. Then a multivariate Gaussian RF model

is adopted for these normal scores transforms. This Gaussian RF model is fully char-

acterized by the variogram of the normal scores transforms shown in Figure 5.27. The

experimental (exhaustive) variogram of the reference normal scores has been modeled

by the nested sum of three structures, a nugget effect of 0.20, a short scale exponen-

tial structure contributing 0.55 of the total unit variance and a larger scale spherical

structure contributing 0.25; both structures are anisotropic:

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES189

Var

iogr

am

Distance

Normal Scores Variogram

.0 10.0 20.0 30.0 40.0 50.0

.00

.20

.40

.60

.80

1.00

Figure 5.27: Semivariograms of the normal score transforms of the reference data of Fig-ure 5.25 along the vertical (upper set of curves) and horizontal directions (lower set). Thesolid lines are the model fits.

γ(hx, hy) = 0.20 + 0.55Exp

√√√√(hx

10

)2

+

(hy

1.6

)2 (5.11)

+ 0.25Sph

√√√√( hx

180

)2

+

(hy

15

)2

with Exp(h) and Sph(h) as defined in (5.10). hx and hy are the horizontal and vertical

coordinates.

The sequential Gaussian simulation algorithm and GSLIB program sgsim [40] was

used to generate 100 simulated realizations of the permeability normal scores field,

conditioned to the previous variogram model. These normal scores realizations are

then back-transformed using the reference permeability distribution of Figure 5.26.

Consequently, all back-transformed realizations exactly reproduce the original his-

togram of Figure 5.26. The first two generated permeability realizations are shown

on Figure 5.28.

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES190

Figure 5.28: Two realizations from the Gaussian model.

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES191

Figure 5.28 is to be compared to Figure 5.25. Note the characteristic “salt-and-

pepper” appearance of the Gaussian images, a consequence of the maximum entropy

character of the underlying RF model (see Appendix C).

The slight 10o dip of the structures seen on Figure 5.25 is not reproduced on

Figure 5.28 for the simple reason that only the vertical to horizontal anisotropy was

modeled.

Median IK Model

The median IK model represents an alternative to the Gaussian RF model in that it

is also fully characterized by a single covariance function, yet its multivariate distri-

bution is non-Gaussian.

The median IK or mosaic model is a bivariate mixture of two binormal distribu-

tions with common univariate Gaussian distribution. The first binormal distribution

has a correlation coefficient of 1 (i.e., all values are equal) and the second a correla-

tion coefficient of 0 (i.e., all values are independent). A mixture of these two bivariate

Gaussian distributions with proportion p(h) yields the following bivariate cumulative

distribution function [101].

Prob{Y (u) ≤ y, Y (u + h) ≤ y′} (5.12)

= p(h)G(min(y, y′)) + [1 − p(h)]G(y)G(y′)

where Y (u) represents the stationary RF model with standard normal cumulative

distribution function (cdf): G(y) = Prob{Y (u) ≤ y}, ∀u.

The previous bivariate distribution (5.12) corresponds to a RF model Y (u) such

that Y (u) and Y (u + h) are equal with probability p(h) and independent with the

complement probability [1 − p(h)]. It can be shown that the correlogram and all

indicator correlograms of the RF Y (u) are equal to p(h):

CY (h)

CY (0)= p(h) (5.13)

with: CY (h) = Cov{Y (u), Y (u + h)}, and:

CI(h; y, y′)CI(0; y, y′)

= p(h) ∀y, y′ (5.14)

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES192

with: CI(h; y, y′) = Cov{I(u; y), I(u + h; y′)}and: I(u; y) = 1 if Y (u) ≤ y; = 0 if not.

One hundred realizations of the median IK model were generated using the pro-

gram sisim of GSLIB [40]; the two first generated realizations are given on Fig-

ure 5.29. Once again, all realizations share, by construction, the original histogram

of Figure 5.26.

Figure 5.29 is to be compared to Figure 5.25 for the reference and to Figure 5.28

for the Gaussian model. Remember that all images share the same histogram, i.e.,

the same proportion of high/median/low permeability values, and the same variogram

model (5.11) for normal score transforms. Hence any difference is due to the implicit

RF model beyond histogram and covariance.

The visual difference between the Gaussian and median IK realizations is quite

striking, with the median IK realizations presenting clearer spatial structures.

Multiple Indicator Model

One should question retaining the covariance as the sole spatial characteristic of an

image (or equivalently the covariance of its normal scores transform). By extracting

more structural information from the reference or sampled image, one should be able

to reproduce more of the spatial features of the reference image. Experience has

shown that whenever sampling allows inference of the attribute covariance C(h) it

also allows inference of its indicator covariances, at least for threshold values z, z′ that

are not too extreme. Thus, consider simulation of the reference image of Figure 5.25

by retaining some of its indicator covariances/variograms.

Seven threshold values zk = 0.1, 2.5, 5.0, 7.5, 10.0, 15.0, and 30.0 md, k = 1, . . . , 7

were retained which correspond to the 0.20, 0.35, 0.48, 0.60, 0.67, 0.80, and 0.92

quantiles of the exhaustive reference distribution of Figure 5.26. The indicator RF is

then defined for each threshold zk as:

I(u; zk) = 1, if permeability Z(u) ≤ zk; = 0 if not.

The corresponding exhaustive variograms in the vertical and horizontal directions,

together with their fitted models are given in Figure 5.30. All models are standardized

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES193

Figure 5.29: Two realizations from the median IK model.

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES194

Threshold Exponential Structure Spherical Structurek quantile C0() C1() a1() b1() C2() a2() b2()1 0.20 0.17 0.50 18.0 7.8 0.33 100.0 10.02 0.35 0.11 0.54 47.7 14.4 0.35 150.0 15.03 0.48 0.13 0.58 90.0 18.0 0.29 170.0 17.04 0.60 0.13 0.61 90.0 14.5 0.26 160.0 16.05 0.67 0.12 0.68 108.0 17.4 0.20 91.0 12.86 0.80 0.12 0.68 108.0 15.2 0.20 85.0 12.07 0.92 0.22 0.69 144.0 21.5 0.09 66.0 11.2

Table 5.2: Reference indicator variograms parameters. The ranges a1(), b1(), a2(), b2() areexpressed in pixel units. The relative sill values C0(), C1(), C2() are dimensionless.

to a unit sill with the following combination of nugget effect, short range exponential

structure, and longer range spherical structure:

γI(h; zk) = Co(zk) (5.15)

+ C1(zk)Exp

√√√√( hx

a1(zk)

)2

+

(hy

b1(zk)

)2

+ C2(zk)Sph

√√√√( hx

a2(zk)

)2

+

(hy

b2(zk)

)2

with Exp() and Sph() defined as in (5.11).

Table 5.2 gives the corresponding model parameters. Some remarks:

• The relative nugget effect is larger for the two extreme threshold values (first

and last); however, this “destructuring” of extreme values is not symmetric,

as would be implied by a Gaussian model [83]: the nugget effect for the 0.20

quantile is different from that for the 0.80 quantile (0.17 �= 0.12).

• The relative contribution, ranges and anisotropy ratios, of the medium range

exponential structure increases consistently from the first to last threshold.

• The relative contribution of the larger range spherical structure decreases con-

sistently from the first through the last threshold.

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES195

Var

iogr

am

Distance

First Cutoff

.0 10.0 20.0 30.0 40.0 50.0

.0000

.0100

.0200

.0300

.0400

.0500

.0600

.0700

Var

iogr

am

Distance

Second Cutoff

.0 10.0 20.0 30.0 40.0 50.0

.000

.040

.080

.120

Var

iogr

am

Distance

Third Cutoff

.0 10.0 20.0 30.0 40.0 50.0

.000

.040

.080

.120

.160

Var

iogr

am

Distance

Fourth Cutoff

.0 10.0 20.0 30.0 40.0 50.0

.000

.050

.100

.150

.200

Var

iogr

am

Distance

Fifth Cutoff

.0 10.0 20.0 30.0 40.0 50.0

.000

.050

.100

.150

.200

.250

Var

iogr

am

Distance

Sixth Cutoff

.0 10.0 20.0 30.0 40.0 50.0

.000

.050

.100

.150

.200

.250

Var

iogr

am

Distance

Seventh Cutoff

.0 10.0 20.0 30.0 40.0 50.0

.000

.040

.080

.120

.160

Figure 5.30: Reference indicator variograms and their fitted models.

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES196

• Therefore, high and low permeability values are less correlated than the medium

values, with low values being better correlated than high values. These remarks

are somewhat corroborated by visual inspection of Figure 5.25.

One hundred realizations of the multiple indicator model were generated using

the program sisim of GSLIB [40]; the first two realizations are given on Figure 5.31.

Figure 5.31 is to be compared to Figures 5.25, 5.28, and 5.29 respectively. Re-

member that all images share the exact same histogram of permeability values, that

of Figure 5.26.

The realizations created with multiple indicator simulation are seen to reproduce

more of the reference spatial features. Another alternative is to use annealing with

multiple-point spatial statistics to obtain a better reproduction.

Annealing

By extracting more structural information from the reference image, one should be

able to reproduce more of the spatial features of the reference image. However,

extracting too much information would lead to too close reproduction of the reference

image; this is not the goal. The goal is to extract the salient or transportable features

that, in practice, could be used for mapping a sparsely sampled reservoir.

In this case the two-point histogram for 33 lags (see Figure 5.32) and four different

five-point indicator covariances aligned in the horizontal direction were retained.

One hundred realizations of the annealing model were generated using the program

sasimi documented in Chapter 3 and Appendix E; the first two realizations are given

on Figure 5.33.

Figure 5.33 is to be compared to Figures 5.25, 5.28, 5.29, and 5.31 respectively.

Note that the annealing realizations share more spatial features with the reference

image due to the additional global statistics that were retained for the simulation.

Flow Modeling

The output uncertainty given by each of the four RF models is quantified by pro-

cessing all 401 images (the reference image and 100 realizations from each of the four

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES197

Figure 5.31: Two realizations from the multiple indicator model.

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES198

Figure 5.32: Lags for two-point histogram control.

RF models) through a flow-related transfer function. The 2-D images were taken to

represent vertical cross sections between an injecting and a producing well.

The constant porosity grid blocks are initially saturated with oil and then sub-

jected to a water flooding with no-flow boundary conditions on the upper and lower

boundaries of the cross section. The Eclipse flow simulator [48] has been used for the

flow simulations. Both wells operate at constant bottom hole pressure, straight line

relative permeabilities were used, and both the oil and water have a unit mobility

ratio.

The following two flow response variables were isolated to characterize the flow

characteristics of each image:

• The effective permeability of the image. This parameter is computed once the

oil is completely swept from the image and the water flow rate has stabilized.

This response variable provides an overall steady state flow characteristic of the

image.

• The time at which the water cut reaches 90%. This response is a measure of

sweepage and late flow characteristics of the image.

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES199

Figure 5.33: Two realizations from the annealing model.

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES200

Case Time to reach 90% Effective Permeabilityq0.025 q0.5 q0.975 q0.025 q0.5 q0.975

Reference 8.33 3.70Gaussian 7.84 11.36 16.45 1.39 2.04 3.11Median IK 7.24 11.10 16.46 1.26 2.21 4.31Full Indicator 5.83 13.03 33.65 0.83 2.13 5.79Annealing 4.69 10.37 18.60 0.91 2.36 7.24

Table 5.3: Summary of output response values. The reference values of 8.33 md and 3.70time units are followed by the median q0.05 and the 95% probability interval (q0.025 andq0.975).

The flow simulation exercise described above was also carried out using the refer-

ence spatial distribution (Figure 5.25). The reference effective permeability was 8.33

md and the time to achieve a 90% water cut was 3.70 time units.

Figure 5.34 shows the histograms of effective permeability obtained from the four

RF models being considered. Figure 5.35 shows the histograms of the time to reach

90% water cut obtained from the four RF models. The effective permeability and the

time to achieve 90% water cut have a correlation of about -0.75 in all cases.

Table 5.3 gives the median (q0.5) and the 95% probability bounds (q0.025 and q0.975)

of the flow response variables for the four sets of simulations.

Observations

• The actual reference time value (3.70 time units) is not within the range of

the 100 Gaussian model-derived response values and barely within the ranges

of the median IK and indicator model-derived distributions. The reference

effective permeability (8.33 md) is only barely in the range of the Gaussian

model derived-distribution and fits only slightly better in the ranges of the

median IK and indicator model-derived distributions.

• Although all 400 realizations of the permeability field reproduce exactly the ref-

erence univariate distribution (Figure 5.25) they all yield distributions whose

centers (mean or median) deviate considerably from the reference value. The

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES201

Fre

quen

cy

.0 10.0 20.0 30.0 40.0

.000

.100

.200

.300

.400

Gaussian Model: Effective Permeabilitymean 11.77

std. dev. 2.14coef. of var 0.182

maximum 18.16upper quartile 13.15

median 11.36lower quartile 10.21

minimum 7.23

Fre

quen

cy

.0 10.0 20.0 30.0 40.0

.000

.100

.200

.300

.400

Median IK Model: Effective Permeabilitymean 11.42

std. dev. 2.36coef. of var 0.21

maximum 18.84upper quartile 12.69

median 11.10lower quartile 9.67

minimum 6.33

Fre

quen

cy

.0 10.0 20.0 30.0 40.0

.000

.050

.100

.150

.200Indicator Model: Effective Permeability

mean 14.04std. dev. 6.12

coef. of var 0.44

maximum 35.03upper quartile 17.30

median 13.03lower quartile 9.05

minimum 5.03

Fre

quen

cy

.0 10.0 20.0 30.0 40.0

.000

.050

.100

.150

.200

Annealing Model: Effective Permeabilitymean 11.14

std. dev. 3.86coef. of var 0.35

maximum 19.78upper quartile 14.86

median 10.37lower quartile 8.10

minimum 3.73

Figure 5.34: Distribution of effective permeabilities obtained from the Gaussian, medianIK, indicator, and annealing RF models. The dot location gives the reference image value(8.33 md).

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES202

Fre

quen

cy

.0 2.0 4.0 6.0 8.0

.000

.100

.200

.300

.400 Gaussian Model: Late Breakthroughmean 2.11

std. dev. 0.45coef. of var 0.22

maximum 3.48upper quartile 2.38

median 2.04lower quartile 1.77

minimum 1.18

Fre

quen

cy

.0 2.0 4.0 6.0 8.0

.000

.100

.200

.300

Median IK Model: Late Breakthroughmean 2.34

std. dev. 0.78coef. of var 0.33

maximum 6.81upper quartile 2.82

median 2.21lower quartile 1.83

minimum 1.14

Fre

quen

cy

.0 2.0 4.0 6.0 8.0

.000

.050

.100

.150

.200

Indicator Model: Late Breakthroughmean 2.41

std. dev. 1.30coef. of var 0.54

maximum 7.15upper quartile 3.12

median 2.14lower quartile 1.43

minimum 0.71

Fre

quen

cy

.0 2.0 4.0 6.0 8.0

.000

.050

.100

.150

.200

Annealing Model: Late Breakthroughmean 2.80

std. dev. 1.66coef. of var 0.59

maximum 8.60upper quartile 3.64

median 2.36lower quartile 1.45

minimum 0.78

Figure 5.35: Distribution of late breakthrough times obtained from the Gaussian, medianIK, indicator, and annealing RF models. The dot location gives the reference image value(3.70 time units).

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES203

Gaussian model is the most inaccurate in the sense that it generates distribu-

tions of response values that do not even contain the true values. The annealing

model is centered closest to the reference values. This is expected since it re-

produces the most reference statistics.

• Note that the distributions of response values need not be Gaussian, nor even

symmetric: see the indicator model-derived histograms of Figures 5.34 and 5.35.

• The multiple indicator RF model is imprecise, in the sense that it yields the

largest variance for the posterior response distribution; but that imprecision

allows accuracy, in the sense that these distributions do include the actual

response values.

• A possible reason for the actual effective permeability (8.33md) to be so low

may be the cross-bedding of the low (white) permeability streaks visible on

Figure 5.25 and their concentration next to the left vertical well. Such infor-

mation is not accounted for in any of the three RF models. Were the 10o dip

of the direction of maximum continuity modeled and the continuity of the low

indicator RV’s increased (e.g., by setting the relative nugget effect C0(1) to zero

in Table 2), the multiple indicator RF model would likely have yielded better

results, i.e., response distributions better centered on the actual values.

• Purposely not to confuse the discussion on entropy, the stochastic simulations

were made non-conditional to local data, as obtained from the producer and

injector. In practice, no matter the RF model chosen, the realizations will be

made to honor local data whether originating from well logs or seismic data

(impedance or velocity); consequently, the realizations would be less different

one from another and also more accurate, i.e., closer to the actual image; con-

sequently, the posterior response distributions would be both more precise and

more accurate.

There is much more actual information to be collected from data than a mere

covariance of normal score transforms. In which case, the output distributions can

be estimated more accurately and precisely.

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES204

.

5.3 Well Tests

An important reservoir parameter measured by a pressure transient well test is the

effective absolute permeability near the well bore. This effective permeability does

not directly inform the block permeabilities near the well bore; however, it does

constrain a complex non-linear average of the small scale block permeabilities. Many

flow processes are directly influenced by this effective permeability. For example,

the average flow rate for a fixed pressure gradient or the reservoir pressure behavior

under constant flow rate conditions are given directly by the effective permeability

near the well. Historically, reservoir performance forecasting [111] used homogeneous

reservoir models with all blocks assigned the well test-derived permeability. Two

shortcomings of such homogeneous and deterministic models are that they do not

allow an assessment of uncertainty and they do not allow for the important influence

of small scale permeability heterogeneities. Stochastic modeling techniques would

allow the small scale permeability heterogeneities to be accounted for. However,

traditional stochastic models cannot account for the effective absolute permeabilities

measured by well tests.

Section 3.4 documents the framework of an annealing approach to include the well

test-derived effective permeability as a component objective function:

Oc =[kwelltest − krealization

]2(5.16)

where kwelltest is the well test derived result and krealization is an approximation of

the well test calculated from the candidate realization. Ideally, krealization would be

derived from a forward simulated well test; however, a full flow simulation amounts to

a global updating and would be too expensive given the computer resources available

at the time of writing this dissertation. A power average of the block permeability

values within an annular volume of investigation V is considered as a numerically

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES205

simpler substitute for the well test forward simulation (see equation (3.28)):

k(ω) =

1

N

∑ui∈V

k(ui)ω

(5.17)

Where k(ω) is the ω-power average permeability of the N permeability values k(ui), i =

1, . . . , N , at locations ui within the volume of interest V . Before application of this

numerical equation, the averaging power ω and the volume V investigated by the well

test must be determined. The volume V is defined by an inner radius rmin and outer

radius rmax that depend on the constant A in the following equations, see also (3.32)

and Figure 3.12:

rmin = A

√√√√kwelltesttmin

φµct

= A√

tdmin(5.18)

rmax = A

√√√√kwelltesttmax

φµct= A

√tdmax (5.19)

where tmin and tdminare the real and dimensionless times at the start of the infinite-

acting radial flow period, tmax and tdmax are the real and dimensionless times at the

end of the infinite-acting radial flow period, φ is the porosity, µ is the viscosity, and ct

is the total compressibility, and kwelltest is the well test-derived effective permeability.

The constant A = 0.010 and the averaging power ω = 0.0 may be used to establish

the volume and type of averaging.

The steps needed to integrate well test-derived effective properties into stochastic

models are:

Calibrate the Volume and Type of Averaging: it is good practice to calibrate

both the constant A and the averaging power ω for each reservoir. The following

procedure (see also [7] and Section 3.4) allows establishing an optimal Aopt and

averaging power ωopt:

1. Generate ns (20-100) multiple realizations of the permeability field with

relevant statistical properties.

2. Forward simulate a well test on each realization to obtain ns pressure

response curves. The conditions should be as close as possible to those

used in the field to arrive at kwelltest.

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES206

3. Deduce an effective permeability ki, i = 1, . . . , ns from each pressure curve

using established well test interpretation techniques [66].

4. Compute average permeabilities k(A, ω)i, i = 1, . . . , ns for values A be-

tween the practical bounding limits of 0.001 and 0.020 and for values ω

between practical bounding limits of -0.5 and 0.5.

5. Choose the pair (Aopt, ωopt) that yields the closest agreement between the

reference ki, i = 1, . . . , ns values and the approximate k(Aopt, ωopt)i, i =

1, . . . , ns values.

Once the appropriate Aopt and ωopt values have been established the goodness

of the power average approximation can be checked by plotting a scatterplot of

the k(Aopt, ωopt)i values versus the well test-derived ki values.

Construct Realizations honoring the numerical approximation to the well test.

Annealing can be used directly or as a post-processing step to generate realiza-

tions with an objective function containing the component:

Oc =[kwelltest − krealization

]2

Validate Reproduction of Well Test: by forward simulating the well test on some

of the well test-conditioned realizations it is possible to evaluate how well the

power average approximation characterizes the full well test result. The cal-

ibrated power average (5.17) should lead to a close agreement between the

known well test result and the interpreted effective permeability from a forward

simulated well test.

The resulting post-processed models may then be used for reservoir performance

forecasting. The output distributions of uncertainty will be better in the sense that

the input stochastic models now honor more relevant data (see Section 2.2).

An Example

All of the steps to incorporate well test data into stochastic models will be illustrated

in this example. The idea is to consider a realistic reservoir performance forecasting

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES207

Injector

Injector

Injector

Injector

Producer

-2.0 6.7 md

-1.6 9.3 md

-1.2 13.0 md

-0.8 18.2 md

-0.4 25.4 md

0.0 35.4 md

0.4 49.3 md

0.8 68.8 md

1.2 96.0 md

1.6 133.9 md

2.0 187.0 md

Figure 5.36: The five spot injection/production pattern that will be used to illustrate theintegration of well test data into stochastic models.

(RPF) problem with and without well test data. The expected result is that by

integrating well test data the output response distributions will be more accurate

and precise.

The RPF problem considered here is to predict the performance (breakthrough

time, oil recovery, . . . ) of a five spot injection/production pattern. A schematic

illustration of the particular pattern is shown on Figure 5.36. The permeability

values at the producer (71.8 md) and the injector wells (clockwise from the upper

left: 231.9 md, 19.9 md, 12.6 md, and 7.6 md) are shown by a gray level scaling. All

stochastic models of the permeability for this spatial domain are conditioned to these

five local data. The gray scale legend provides both a scaling in the normal space and

in millidarcies. All future gray scale images will use this scaling.

These conditioning values were taken from an initial unconditionally simulated

realization, see Figure 5.37.

The goal of this example is to evaluate uncertainty before and after integrating

the well test data. For this reason, all flow parameters including the univariate dis-

tribution and the variogram of block permeabilities have been fixed. The reference

univariate distribution of block permeabilities is lognormal with a mean of 50.0 md

and a variance of 2500 md2 (coefficient of variation = 1.0). A histogram and log-

normal probability plot of the reference permeability values are shown at the top

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES208

Reservoir Parameter valueNumber of grid blocks (x,y,z) 51 x 51 x 1Size of grid blocks (x,y,z) 70 x 70 x 10 feetFormation thickness 10.0 feetWellbore radius 0.33 feetPorosity 0.3 p.u.Viscosity 0.3 cpCompressibility 0.00005 1/psiFormation volume factor 1.4Production Rate 100 STB/dReservoir Boundaries no flow

Table 5.4: The parameters used for the well test integration example.

of Figure 5.37. The isotropic normal scores semivariogram, shown at the lower left

of Figure 5.37, is the sole statistical measure (aside from the reference well test re-

sponse) retained to characterize the spatial distribution of the permeability over the

area of the five spot pattern. All distance units are relative to the 51 by 51 block

discretization used to describe the five spot pattern7. The isotropic practical range

of the reference exponential semivariogram model is one half the spatial extent of the

five spot pattern. The reference realization, generated by annealing (program sasim)

using the reference univariate distribution and reference normal scores semivariogram,

is shown at the lower right of Figure 5.37.

The Eclipse black oil flow simulator [48] was used to simulate a drawdown well

test at the producing well using the reference (true) distribution of permeability. The

reservoir parameters are not important in an absolute sense; however, they are given

on Table 5.4 for completeness.

The well test pressure response for the reference image is shown on Figure 5.38.

Interactive non-linear regression software [66] could be used to interpret this well test;

however, the more traditional Miller-Dyes-Hutchinson (MDH) plot8 suffices for the

exercise considered in this dissertation. The effective permeability is inferred from

7The 2-D square pixels represent 70 feet by 70 feet blocks.8A semi-log plot of pressure at the well versus time.

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES209

Fre

quen

cy

Variable

0. 100. 200. 300. 400. 500.

.000

.040

.080

.120

.160

Histogram Number of Data 2601

mean 50.43std. dev. 50.63

coef. of var 1.00

maximum 864.8upper quartile 62.9

median 36.5lower quartile 20.3

minimum 2.6

Cum

ulat

ive

Pro

babi

lity

Variable

Lognormal Probability Plot

.01

.1

.2

125

10

20304050607080

90959899

99.899.9

99.99

1.0 10. 100. 1000.

Sem

ivar

iogr

am

Distance

Reference Normal Scores Semivariogram Model

.0 10.0 20.0 30.0 40.0 50.0

.00

.20

.40

.60

.80

1.00

1.20

Gray Scale Map

Figure 5.37: The reference univariate distribution of block permeability values is illustratedat the top of this figure by a histogram and lognormal probability plot. The reference normalscores semivariogram model is shown at the lower left and the reference image (which yieldsthe reference well test result) is shown at the lower right.

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES210

Reference Miller-Dyes-Hutchinson Plot

Time, hrs

Pre

ssur

e, p

si

0.1 1 10 100 1000

5800.

5820.

5840.

5860.

5880.

5900.

Figure 5.38: The Miller-Dyes-Hutchinson plot resulting from a well test with the referencedistribution of permeability.

the straight line infinite acting portion of the semi-log curve, i.e.,

k = 162.6q B µ

|m| h(5.20)

where k is the permeability in milliDarcies, q is the flowrate in STB/day, B is the

formation volume factor, µ is the viscosity in centipoise, m is the slope of the semilog

straight line (pressure in psi and time in hours), and h is the thickness of the for-

mation. Note some early time effects due to numerical precision and the late time

pressure decrease due to depletion within the area of the five spot (all sides of the five

spot pattern represent no flow boundaries). The estimated slope |m| of the semi-log

straight line leads to an effective permeability of 41.1 md.

The time interval for infinite acting radial flow was taken between dimensionless

times of 5.46 x 108 and 113.8 x 108 (oil field units). These times agree with the ex-

pected result for a no-flow boundary [103]. Note that the actual time limits (shown on

Figure 5.38 by the gray vertical lines) depend on the interpreted effective permeability

ke. For this reason the effective permeability is evaluated iteratively as follows:

1. Choose an initial guess ktry and compute the time limits:

tmin =tdmin

φµctr2w

ktry, tmax =

tdmaxφµctr2w

ktry

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES211

using the dimensionless time limits of tdmin= 5.46 x 108 and tdmax = 113.8 x

108.

2. Estimate a new effective permeability ktry by computing the straight line slope

between tmin and tmax.

3. If the effective permeability has changed significantly go back to step 1 and

compute new limits,. . .

No more than 10 iterations were necessary to obtain a stable effective permeability.

This interpretation procedure has been verified by simulated well tests under the

same conditions with uniform permeability fields (k= 25md, 50md, and 75md). The

interpreted results (shown on Figure 5.39) are 24.8 md, 49.8 md, and 75.5 md.

The five local conditioning values (see the well locations on Figure 5.36), the

univariate distribution of permeability (Figure 5.37), the normal scores semivariogram

(Figure 5.37), and the well test-derived effective permeability (41.1 md) are the data

available to generate stochastic models of the reservoir permeability. The importance

of the well test-derived effective permeability will be judged by integrating it into

stochastic models. The first step is to calibrate the volume and type of averaging

represented by the well test-derived effective permeability.

Volume and Type of Averaging

The four step procedure to establish the appropriate volume and type of averaging was

recalled at the beginning of this section (5.3). The first step is to generate a significant

number of realizations ns (100 in this case) that are conditional to the local data,

the univariate distribution, and the reference variogram model. In practice, these

realizations should be conditional to seismic data and all other sources of data except

the well test.

The 100 initial realizations were generated with annealing to match the variogram

model, i.e., the following objective function was used with the sasim program:

O =∑h

[γ(h) − γ(h)]2 (5.21)

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES212

MDH PLOT - Uniform K=25

Time, hours

Pre

ssur

e

0.1 1 10 100 1000 5760.

5800.

5840.

5880.

MDH PLOT - Uniform K=50

Time, hours

Pre

ssur

e

0.1 1 10 100 1000 5870.0

5880.0

5890.0

5900.0

5910.0

5920.0

5930.0

5940.0

5950.0

MDH PLOT - Uniform K=75

Time, hours

Pre

ssur

e

0.1 1 10 100 1000 5900.0

5910.0

5920.0

5930.0

5940.0

5950.0

5960.0

Figure 5.39: Miller-Dyes-Hutchinson plots resulting from a well test in uniform permeabil-ity fields of 25md, 50md, and 75md.

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES213

where γ(h) is the realization variogram, and γ(h) is the reference variogram for

selected separation vectors h. The nh = 18 separation vectors h are defined by the

most compact arrangement of 18 lags, see Figure 5.16 used for the previous Amoco

example. The local five well data are honored by never perturbing them and the

histogram is honored by drawing the initial realization from the reference univariate

distribution.

The first four realizations are shown on Figure 5.40. They appear as plausible

realizations of the distribution of permeability. The effect of the local conditioning

data is seen by the high permeability near the injector in the upper left corner and low

permeability at the other three injector locations. Figure 5.41 shows a q-q plot com-

paring the reference distribution (in the normal space) to the distributions of all 100

realizations. The univariate distributions are all near the 45o line indicating a close

agreement to the reference distribution. Figure 5.42 gives the reference variogram

model, shown as the thick gray line, and the isotropic variograms computed from

all 100 realizations. The variograms of the simulated realizations appear to have a

slightly shorter range and reach a higher sill than the reference model; otherwise there

is a reasonable agreement between the realizations and the model. Note that it took

1.91 minutes of CPU time on a DEC 5000 workstation to generate each realization.

The next step is to use a flow simulator to simulate numerically a well test on

each of these 100 realizations. This was done with the Eclipse [48] flow simulator

under the exact same conditions as used for the reference distribution of permeability

shown in Figure 5.37. The well test-derived effective permeability was established for

each of the 100 initial realizations, see Figure 5.43. The effective permeabilities for

the first four realizations shown on Figure 5.40 are 41.8 md, 43.8 md, 38.9 md, and

66.8 md respectively.

The constant A is required to define rmin (5.18) and rmax (5.19) and the averag-

ing power ω is needed to define the type of averaging. The average permeabilities

k(A, ω)i, i = 1, . . . , ns for values A between the practical bounding limits of 0.001 and

0.020 and for values ω between practical bounding limits of -0.5 and 0.5 have been

computed using all 100 initial realizations. The criteria for an optimal pair (Aopt, ωopt)

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES214

First Realization Second Realization

Third Realization Fourth Realization

Figure 5.40: The first four initial realizations generated with annealing. Note the effect ofthe five local conditioning data - one at each corner and the center.

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES215

Q-Q Plot: Simulated versus Reference

Reference

Sim

ulat

ed

-4 -3 -2 -1 0 1 2 3

-4

-3

-2

-1

0

1

2

3

Figure 5.41: A q-q plot comparing the reference distribution (in the normal space) to thedistribution of all 100 realizations.

Normal Scores Semivariogram

Distance

Sem

ivar

iogr

am

0.0 5.0 10.0 15.0 20.0 25.0

0.0

0.2

0.4

0.6

0.8

1.0

1.2

Figure 5.42: A plot comparing the reference normal scores variogram (shown as the thickergray line) to variograms computed from all 100 realizations.

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES216

Fre

quen

cy

20. 40. 60. 80. 100. 120.

.000

.040

.080

.120

Initial Realizations: Well Test-Derived Effective Permeabilitymean 58.30

std. dev. 15.45coef. of var 0.26

maximum 102.1upper quartile 70.8

median 54.6lower quartile 46.0

minimum 33.1

Figure 5.43: The histogram of 100 effective permeability values from initial realizations.

is to minimize the mean normalized absolute deviation and the mean normalized error

defined as:

NAD(A, ω) =i=ns∑i=1

|k(A, ω)i − ki|ki

(5.22)

NE(A, ω) = |∑i=ns

i=1 k(A, ω)i −∑i=nsi=1 ki∑i=ns

i=1 ki

| (5.23)

NAD measures correlation and NE measures the bias between the power average

approximation and the true well test values. Figure 5.44 shows gray level maps of

NAD and NE for values A ranging from 0.001 to 0.020 and values ω ranging from

-0.5 to 0.5. Note that the optimal pair Aopt = 0.003 and ωopt = 0.10 minimizes both

error terms.

Figure 5.45 shows a scatterplot of the power average numerical approximation and

the true well test-derived effective permeability. Note the unbiasedness (the average

effective permeabilities are 58.3 in both cases) and the excellent correlation value of

0.98. The physical volume informed by the well test can be determined from the

constant Aopt and knowledge of the time limits (tdminand tdmax):

rmin = A ·√

tdmin= 0.003 ·

√5.46x108 = 70.1feet

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES217

a = 0.020

a = 0.010

a = 0.001w = -0.5 w = 0 w = +0.5

NAD

.0

.04

.08

.12

.16

.20

.24

.28

.32

.36

.40

a = 0.020

a = 0.010

a = 0.001w = -0.5 w = 0 w = +0.5

NE

.0

.04

.08

.12

.16

.20

.24

.28

.32

.36

.40

a = 0.020

a = 0.010

a = 0.001w = -0.5 w = 0 w = +0.5

NAD + NE

.0

.08

.16

.24

.32

.40

.48

.56

.64

.72

.80

Figure 5.44: Gray level maps of the mean normalized absolute deviation NAD, the meannormalized error NE, and the sum of NAD and NE, for A values ranging from 0.001to 0.020 and ω values ranging from -0.5 to 0.5. Note the optimal pair Aopt = 0.003 andωopt = 0.10 shown by the small circle.

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES218

Pow

er A

vera

ge A

ppro

xim

atio

n

Well Test-Derived Effective Permeability

Approximate Permeability versus Well Test Derived Permeability

20. 40. 60. 80. 100. 120.

20.

40.

60.

80.

100.

120.

X Variable: mean 58.3variance 238.8

Y Variable: mean 58.3variance 290.3

correlation (r) 0.976rank correlation (r) 0.975

Figure 5.45: A scatterplot of the power average approximation and the true well test-derived effective permeabilities for all 100 initial realizations.

rmax = A ·√

tdmin= 0.003 ·

√113.8x108 = 320.0feet

The annular region defined by these limits are shown on Figure 5.46.

Conditioning to Well Test Permeability

Once the volume and type of averaging have been defined, the annealing program

developed in Chapter 3 sasim can be used to impose the power average (equation

5.17 with Aopt = 0.003 and ωopt = 0.1). The objective function is to match jointly

the spatial structure, as quantified by the variogram, and the well test derived per-

meability:

O = λ1

nh∑l=1

[γmodel(hl) − γrealization(hl)]2 + λ2

[kwelltest − krealization

]2(5.24)

The same 18 lags nh as considered for the initial 100 realizations were used for the

final simulations.

Figure 5.47 shows how each component in the objective function changes as the

simulation proceeds (the solid line is the deviation from the model variogram and

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES219

Injector

Injector

Injector

Injector

Producer

Figure 5.46: An illustration of the annular region measured by the well test.

the dashed line is the deviation due to the well test permeability). Both component

objective functions go to zero at about 150,000 swaps. It took 1.7 minutes on a DEC

5000 to perform the 250,000 swaps (the far right of the horizontal scale on Figure 5.47)

needed to complete the simulation. It took 3.41 minutes of CPU time on a DEC 5000

workstation to generate each realization, versus 1.91 minutes of CPU time to generate

realizations that are not conditional to the well test. Once again, the temperature

reduction parameter λ in the annealing schedule could be lowered to decrease the

CPU time.

The first four realizations are shown on Figure 5.48. The conditioning data and

general characteristics (refer back to Figure 5.40) are still honored; now the calibrated

power average has also been imposed on the realizations.

Although the annealing approach imposes almost exactly the calibrated power

average, it does not ensure reproduction of either the actual effective permeability

of the actual field or the full pressure response. Another 100 well test flow simula-

tions are needed to establish the information brought by honoring the power average.

Figure 5.49 shows the distribution of the 100 well test derived effective permeability

values after post processing. Note that the distribution of effective permeabilities, af-

ter conditioning to the well test, is much more closely centered on the value inferred

from the reference image (41.1 md). Further, note the significant reduction in the

variance as compared to Figure 5.43.

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES220

Objective Function versus Time

Number of swaps (1000’s)

Obj

ectiv

e F

unct

ion

0. 50. 100. 150. 200. 250.

.00

.20

.40

.60

.80

1.00

1.20

Figure 5.47: The objective function versus time for direct annealing simulation to matchvariogram (solid line) and the well test (dashed line).

The true reference value (41.1 md in this case) would be measured in practice and

the corresponding improvement in the distribution of effective permeabilities would

be accessible.

Impact on Output Uncertainty

The well test-derived effective permeability has been honored reasonably well by the

post-processed realizations. The important question addressed in this section is how

does integrating the well test-derived effective permeability impact the final reservoir

performance forecasting and the associated measure of uncertainty?

The impact of integrating the well test data is evaluated by simulating the perfor-

mance of the five spot pattern using the reference image (the truth), with the initial

realizations (without accounting for the well test data), and last with the realizations

conditional to the well test-derived effective permeability. The initially oil saturated

reservoir is produced by pumping from the central well and injecting water at the four

corner injector wells (see Figure 5.36). The following three output response variables

are recorded from the flow simulation output:

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES221

First Well Test Conditioned Realization Second Well Test Conditioned Realization

Third Well Test Conditioned Realization Fourth Well Test Conditioned Realization

Figure 5.48: Four realizations of direct annealing simulation to match variogram and welltest.

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES222

Fre

quen

cy

20. 40. 60. 80. 100. 120.

.000

.100

.200

.300

.400

Conditioned Realizations: Well Test-Derived Effective Permeabilitymean 92.0

std. dev. 3.26coef. of var 0.035

maximum 101.0upper quartile 94.0

median 92.2lower quartile 89.4

minimum 85.1

Figure 5.49: The histogram of 100 effective permeability values after post processing.

1. the time of first water arrival (breakthrough time),

2. the time to reach a 50% water cut,

3. and the final oil in place after a fixed period of production.

The three reference response values are:

Breakthrough time = 0.809 time units

Time to reach 50% water cut = 2.732 time units

Final oil recovery = 0.376

The problem is to assess the uncertainty related to prediction of these response

variables using only the five local conditioning data, the univariate distribution of

permeability values, the variogram, then in addition the well test result. Two hun-

dred additional flow simulations were performed on the initial and post-processed

realizations. Figure 5.50 shows the output space of uncertainty before and after con-

ditioning to the well test result. Table 5.5 shows the 95% probability interval before

and after conditioning to the well test data. The significant reduction in the width

of the 95% probability interval is the direct result of adding the well test-derived

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES223

Response Variable 95% Probability Interval %q0.025 q0.5 q0.975 reduction

Breakthrough reference 1.46time before 0.45 1.11 2.15

after 0.63 1.28 2.16 -10%

Time to reach reference 5.5350% water cut before 2.22 4.12 7.36

after 3.14 4.66 6.83 -28%

Final oil reference 0.47recovery before 0.35 0.41 0.53

after 0.37 0.42 0.47 -44%

Table 5.5: Summary of output response results before and after conditioning to includethe well test-derived effective permeability.

effective permeability. Further, note that the central value (as characterized by the

median q0.5) of the post-processed realizations is closer to the true reference. Thus,

the output distributions of uncertainty are both more accurate and more precise after

accounting for more input data.

One should not expect perfect accuracy with only five well data even if the well

test data is used. In practice, the true reference value (shown as the black dot on

Figure 5.50) is not available and the width of a fixed probability interval, say the

95% probability interval, is used to assess the possible true response value. One

expects this assessment of uncertainty to be reduced as more relevant data such as

well test-derived properties are considered in the stochastic modeling.

CHAPTER 5. ADVANCED APPLICATIONS OF ANNEALING TECHNIQUES224

Fre

quen

cy

.00 .50 1.00 1.50 2.00 2.50

.000

.040

.080

.120

.160

Time of Water Breakthroughmean 1.17

std. dev. 0.42coef. of var 0.36

maximum 2.29upper quartile 1.39

median 1.11lower quartile 0.83

minimum 0.43

Fre

quen

cy

1.00 2.00 3.00 4.00 5.00 6.00 7.00 8.00 9.00

.000

.020

.040

.060

.080

Time to Reach 50 percent Water Cutmean 4.22

std. dev. 1.35coef. of var 0.32

maximum 7.97upper quartile 5.07

median 4.12lower quartile 3.14

minimum 1.76

Fre

quen

cy

.300 .350 .400 .450 .500 .550

.000

.020

.040

.060

.080

.100

Final Oil in Placemean 0.42

std. dev. 0.047coef. of var 0.11

maximum 0.55upper quartile 0.45

median 0.41lower quartile 0.38

minimum 0.34

Fre

quen

cy

.00 .50 1.00 1.50 2.00 2.50

.000

.050

.100

.150

.200

.250

Time of Water Breakthroughmean 1.2552

std. dev. .3562coef. of var .2838

maximum 2.1574upper quartile 1.3795

median 1.2804lower quartile 1.0201

minimum .6323

Fre

quen

cy

1.00 2.00 3.00 4.00 5.00 6.00 7.00 8.00 9.00

.000

.040

.080

.120

.160

Time to Reach 50 percent Water Cutmean 4.6369

std. dev. .8658coef. of var .1867

maximum 6.9745upper quartile 5.2162

median 4.6608lower quartile 4.0584

minimum 3.0233

Fre

quen

cy

.300 .350 .400 .450 .500 .550

.000

.050

.100

.150

.200

Final Oil in Placemean .4171

std. dev. .0237coef. of var .0568

maximum .4923upper quartile .4291

median .4156lower quartile .4017

minimum .3671

Figure 5.50: The simulated output distributions generated by the annealing realizationsbefore (left side) and after (right side) conditioning to the well test result. The upperhistograms are for the breakthrough time, the middle histograms are for the time to achievea 50% water cut, and the lower histograms are for final oil recovery. The black dot in thebox plot below each histogram is the true value obtained from the reference image, thethree vertical lines are the 0.025 quantile, the median (0.5 quantile), and the 0.975 quantileof the output distribution.

Chapter 6

Concluding Remarks

Reservoir performance forecasting calls for running flow simulators with numerical

models of the reservoir petrophysical properties (saturation, porosity, permeability,

. . . ). To yield unbiased predictions these numerical models must reflect both the

major flow units and the short scale heterogeneities [84]. The requirement to reflect

the short scale heterogeneities implies that smoothed models generated by kriging,

splines, or any conventional interpolation algorithm are inappropriate for reservoir

modeling. A stochastic simulation algorithm must be used to generate acceptable

input numerical models. The numerical models generated by stochastic simulation

algorithms are non-unique since the data spacing is typically large with respect to

the scale of these heterogeneities; there are many models which will honor both the

local conditioning data and the spatial variability.

Processing multiple stochastic models of the reservoir allows constructing an out-

put probability distribution of each critical response variable. For example, in an

enhanced oil recovery project a critical response variable could be the time at which

the flooding agent breaks through at the production wells. A probability distribution

of this time would be constructed by simulating the recovery project on a number

of stochastic reservoir models. These output distributions are used to judge the en-

gineering and economic viability of the project. A number of alternate production

scenarios could be considered to choose one that optimizes the ultimate recovery, with

minimal predicted uncertainty. Finally, once a production scenario has been chosen,

225

CHAPTER 6. CONCLUDING REMARKS 226

the stochastic models may be used for production planning and economic forecasting.

There are many techniques available for stochastic simulation. A recent article [59]

summarized seven different modeling techniques each of which could be implemented

in a number of different ways. The stochastic relaxation or annealing techniques

documented in this dissertation were not on the list. At the time of writing this

dissertation stochastic simulation is “the most vigorous sector of research and ap-

plications in geostatistics” [84] and new techniques are continually being developed.

Given the number of alternate techniques, criteria had to be developed to compare

simulation techniques. The following three criteria were proposed in Section 2.2:

1. A good technique must generate plausible realizations in a reasonable amount

of time. Time refers to the human involvement and CPU time required for the

initial set up and the repeated application of the technique.

2. A good technique is one that allows the maximum prior relevant information

to be accounted for. This is the only direct way to ensure that the output

distribution is as accurate as possible.

3. Finally, a good technique is one that explores the largest space of output un-

certainty, i.e., one that generates a maximum entropy distribution of response

variables.

These three criteria have been recalled and used in every chapter of this dissertation.

The motivation for considering annealing techniques arises from their ability to

integrate more prior information (see point 2 noted above) such as complex geolog-

ical structures and well test-derived effective properties. Chapter 3 of this disser-

tation developed the elements of an annealing-based stochastic simulation program

sasim that accounts for conventional covariance/variogram models, geological struc-

tures (through multiple-point statistics), and well test-derived effective permeabilities

(through appropriate power averages).

One concern with annealing-based techniques is that by exactly honoring control

statistics they may artificially limit the space of output uncertainty (see point 3 noted

above). Chapter 4 developed an extended example comparing stochastic simulation

CHAPTER 6. CONCLUDING REMARKS 227

based on annealing to the more conventional sequential Gaussian and sequential indi-

cator simulation techniques. It was shown that annealing actually generates a slightly

larger space of uncertainty. The reason for this was conjectured to be the very par-

ticular multivariate distributions used by the Gaussian and indicator techniques.

Following up on Chapter 3 and point 2 noted above, Chapter 5 presented some

advanced applications of annealing techniques where more prior information such

as multiple-point connectivity functions, seismic data, and well test-derived effective

properties were integrated into stochastic models. The importance of integrating

these additional data was displayed visually and quantified by output distributions

of flow-related variables. Integrating more relevant prior information leads to more

accurate and precise distributions of output uncertainty.

The research presented in this dissertation has uncovered many different ways

to improve the results obtained: 1) special coding and determining fast annealing

schedules would allow generating realizations much faster than the code given with

this dissertation, 2) many other types of engineering data, such as tracer tests, could

be integrated into stochastic models to yield even more accurate predictions, and 3)

accounting for uncertainty in the control statistics would lead to a larger space of

output uncertainty.

The original contribution of this dissertation has been the development of the

annealing algorithm for stochastic imaging in a reservoir modeling context. Anneal-

ing allows data, that conventional algorithms are incapable of accouting for, to be

integrated into stochastic models. The improvement in the accuracy and precision of

output distributions of uncertainty has been demonstrated with example applications

considering multiple-point spatial connectivity functions and well test data.

The immediate practical application of this technique is to stochastic modeling of

petroleum reservoirs and groundwater aquifers. The ability to integrate the results

of pressure transient well tests to integrate seismic data and to reproduce complex

geological patterns (geometries) will lead to better reservoir performance forecasting.

CHAPTER 6. CONCLUDING REMARKS 228

Ideas for Future Research

Most of the research needs discussed below may be related back to one of the three

criteria for a good stochastic simulation technique.

Implementation

The implementation of annealing for production use would warrant some effort to

fine tune the algorithm and reduce the CPU time requirements. For example, the

CPU time required to simulate the 1600 node Berea image was reduced by a factor

of 10 simply by adjusting the annealing schedule. Adjusting the annealing schedule

is problem specific since some problems require a slow schedule to arrive at plausi-

ble realizations. For example, a too fast cooling schedule can lead to artifact edge

effects. There is some evidence that analagous thermodynamic properties such as

the critical temperature [15], specific heat, or entropy [1] could be used to determine

automatically when the temperature can be decreased faster.

A number of odd artifacts, such as the checkerboard appearance of Figure 5.3,

were generated with certain objective functions. There is a need for flexible interac-

tive software that would allow real-time visualization of the stochastic model. For

example, the color frame buffer on workstation computers could be used as memory

for the simulated realization during the annealing simulation. In this way, the trans-

formation of a random image to a plausible stochastic model could be dynamically

visualized. It would be possible to appreciate which spatial features appear first and

how many components can practically enter the objective function. This visualization

would also help to understand better the annealing approach and would allow unre-

alistic realizations to be quickly recognized and discarded. One could interactively

modify the objective function with an immediate graphical display of the results.

The sasim code developed during the preparation of this dissertation should be

considered as the starting point for operating programs and further research.

CHAPTER 6. CONCLUDING REMARKS 229

Control Statistics

The annealing algorithm aims at reproducing specific control statistics. In many

cases it is not clear which statistics should be used and how to infer them. For

example, control patterns obtained by scanning reference geological images may be of

little practical value depending on the relationship between color or grayness and the

actual rock property being mapped. There are two other alternatives: 1) the result

of sedimentary clastic simulation [131] could be used to provide appropriate reference

images at the reservoir scale required, and 2) outcrops or detailed geological maps

constructed from zones of dense information could yield reference control patterns.

In all cases there is the risk that the control statistics may not be representative

of the actual reservoir; consequently, the distribution of output uncertainty may be

inaccurate. More research into the use of these real control patterns is warranted.

Any full assessment of uncertainty output should consider the uncertainty in the

input control statistics. Future research should address the problem of accounting for

uncertain control statistics in a number of ways. One possibility would be to consider

the control statistics as the result of some random function model. Then, prior to the

creation of a stochastic realization, the control statistics could be drawn from such

a model. The realizations would match the input statistics exactly and the space

of uncertainty would be expanded since these input statistics vary from one set of

realizations to another.

Engineering Data: Well Tests

One worthwile research avenue is the possibility of using engineering data based on a

convective process rather than a diffusive process. The pressure transient well tests

considered in this dissertation are based on a diffusive process; therefore, they are ap-

propriate for measuring average properties and are insensitive to local heterogeneities.

Measurements that depend on a convective process, such as tracer tests, would be

more suited to identifying local heterogeneities.

The well test information considered in this dissertation could also be extended

CHAPTER 6. CONCLUDING REMARKS 230

to multiple permeability averages within concentric annular volumes around the well-

bore. Alternate reservoir models such as fractured or faulted reservoirs could be

considered. For example, the positioning of fault boundaries could be formulated as

an optimization problem where the position and extent of the fault(s) is perturbed

to obtain a close match to an observed well test response.

Another interesting avenue of research opened up by this thesis work would be to

help in designing a test that would bring the most information about the reservoir. For

any given piece of data, as obtained from a particular well test, it should be possible

to judge its impact on output uncertainty before obtaining the information. This

would be done through stochastic simulation exercises similar to those documented

in Section 5.1 and Section 5.3. This would allow balancing the cost of obtaining

the data with its predicted impact on the output uncertainty. The data that most

significantly constrains the output realizations at the least cost is preferable.

Appendix A

Acquisition of Geological Images

This appendix documents the procedure used to acquire images in a format useful

for annealing and simulation applications. Real geological images, with all of their

complexity, motivate the development of stochastic simulation techniques that can

account for their characteristic yet complex features. The geological images here

considered correspond to clastic sedimentary environments. Clastic sediments are

the most important petroleum reservoir rock and present challenging characteristic

features for the methodology developed in this dissertation.

Depositional sedimentary environments can be classified from non-marine to deep

marine, i.e., according to topographic elevation starting from the mountains and

ending with deep sea fans.

Glacial Sediments are deposited directly from glacial ice. They are generally poorly

sorted and stratified and are rarely petroleum reservoir rocks (except in Oman).

Eolian Deposits originate from arid and semi-arid deserts. Most common are dune,

interdune, sand sheet, and extradune deposits. These are very important reser-

voir units and present considerable heterogeneity in the form of lateral discon-

tinuities, interspersed impermeable zones, cross-bedded units, and anisotropic

properties across individual laminae.

Alluvial Fan Deposits accumulate at the base of a mountain or other upland area.

These are not generally reservoir rocks for petroleum; however, some production

231

APPENDIX A. ACQUISITION OF GEOLOGICAL IMAGES 232

has been realized from ancient fan deltas in Texas, Oklahoma, New Mexico, and

Colorado.

Lacustrine Environments originate from lakes or lake systems. The major conti-

nental source of petroleum in North America, China, and North Africa is from

large lacustrine depositional systems.

Fluvial Facies Models are due to the activities of rivers, such as meandering,

straight, braided, and anastomosing environments. The coarse grained nature

of fluvial sediments form potentially good reservoir rocks. There are significant

corresponding petroleum deposits in Alberta, Montana, Texas, and Prudhoe

Bay which is largely an ancient braided river deposit.

Deltaic Environments result from interacting dynamic processes (wave energy,

tidal regime, currents, climate, . . . ) which modify and disperse riverborne flu-

vial clastic deposits. These are important reservoir source rocks.

Estuarine Deposits form in semi-enclosed marginal marine water (an estuary) where

the salinity is diluted from fluvial discharge. Estuarine deposits have excellent

oil and gas potential.

Tidal Flats occur on open coasts of low relief. The North Sea and Georgia coast of

the USA are important examples of tidal flats.

Barrier-Island and Strand Plain Deposits are sandy islands or peninsulas elon-

gated parallel to the shore. Strand-plains are wider in the land-sea direction

and generally lack well developed lagoons and inlets. These environments are

supplied and molded almost entirely by marine processes and are important oil

and gas reservoirs.

Continental Shelf is that part of the sea floor between the shoreline and the upper

edge of the continental slope. The potential for economic oil and gas accumu-

lations in sandstone facies of ancient shelf deposits is high.

APPENDIX A. ACQUISITION OF GEOLOGICAL IMAGES 233

Continental Slope is that part of the continental margin that has gradients of

greater than 1 in 40 (2.5%).

Submarine Fans are deep sea deposits formed by a whole family of mass flow pro-

cesses. Important petroleum sources are located in ancient submarine fan de-

posits (including nearly all the oil recovered in Southern California to date and

several areas within the North Sea Basin).

Each of these depositional environments is well documented in the literature. The

preceeding is to recall the origin of reservoirs and establish the important environ-

ments.

Another approach to classifying geological structures would be on the basis of the

process that creates a particular structure, e.g., structural, depositional, chemical.

This classification scheme was not considered because the above classification provides

better geometric information.

It is essential that the techniques developed in this dissertation apply to real

life problems. Consider the complexities shown on the geological images collected

later. These real images present characteristics that repeat themselves in space with

some variance; the random component makes the structures difficult to quantify with

marked point processes and the deterministic aspect of the structures poses problems

for the conventional stochastic simulation techniques.

Acquisition of Characteristic Features in an Electronic Format

Before discussing the detailed steps required to encode digitally a photograph, it is

worthwile to point out a prior fundamental assumption. The assumption is that the

rock properties are somehow characterized by a monotonic or known transform of the

grayness or color of the image. In many cases this assumption is not too limiting,

e.g., the shaleyness in a shaley-sandstone relates not only to the permeability of the

rock but also to its color. In other cases the color intensity or grayness may have no

relation to the petrophysical properties. Ideally, rock properties should be measured

directly at a very detailed scale as was done for the Berea sandstone example [53].

There are very few examples of this type of direct measurements; acquisition of such

APPENDIX A. ACQUISITION OF GEOLOGICAL IMAGES 234

exhaustive spatial data would have been prohibitively expensive for the purposes of

this dissertation.

When petrophysical properties are measured directly, a photograph or a geolo-

gist’s drawing can be digitally encoded by the procedure documented hereafter. One

significant limitation in the acquisition of geological features is that one must not to

violate any copyright protection. Permission to copy or scan an image is typically

obtainable from the publisher or original author/photographer. Some books com-

piled by educational facilities or geological associations grant limited permission to

make copies for non-commercial educational purposes, e.g., Sandstone Depositional

Environments [118]. Images from such sources have been used for this dissertation.

The procedure to encode digitally an image is:

1. The first step is to scan the image. A “Complete PC” [26] full page scanner was

purchased. The scanner will scan 8.5” by 11” pages at up to 256 gray scales and

at 300 dots per inch (dpi). Note that the gray scales are made up of particular

patterns of black pixels; the device will only scan black and white pixels at 300

dpi.

2. The scanner generates output files in a variety of commercial formats. The

difficulty is that all file formats are compressed and non-readable by standard

C or Fortran programs. Moreover, the compression algorithm is proprietary so

it is impossible to read the file directly.

This major hurdle (converting the format of the image) was crossed by using

the Microsoft Paintbrush program. The scanning software will write out a .pcx

format which the Paintbrush program can read. The Paintbrush program is

then used to write out a bit map file. This file is still binary but at least it

is not compressed. A C program was written to convert the binary file to files

compatible with the GSLIB programs [40].

3. It is now possible to read the file, create standard data files, and plot PostScript

gray scale or color maps. However, we are still not at a point where the image is

useful. The pixels are all black or white and the pattern indicates the grayness

APPENDIX A. ACQUISITION OF GEOLOGICAL IMAGES 235

Scanned Image

Figure A.1: Unprocessed output from the scanning program. Note the gray levels inducedby the pattern of black and white pixels.

level. Figure A.1 illustrates this on a small image (the original is less than one

inch square).

4. The next step is to measure the grayness by some moving window statistic

and replace the binary variable by a continuous variable. One example of this

is shown on Figure A.2 where the sum of blacks in a 3 by 3 pixel (black=1,

white=0) has replaced the original variable. Windows of various sizes, vary-

ing amounts of overlap, and varying weight functions (as described below) are

considered to remove artifacts. For example, Figure A.3 shows the image with

most artifacts removed.

The discontinuous 1/0 pixel measurements is converted to a continuous gray

scale measurement by applying a general inverse distance weighting scheme to

pixels falling within a neighborhood:

wi =

1(di+c)p∑n

i=11

(di+c)p

(A.1)

where, there are n pixels within the neighborhood, di is the anisotropic distance

APPENDIX A. ACQUISITION OF GEOLOGICAL IMAGES 236

Filtered Image

Figure A.2: Filtered with a 3 by 3 pixel sum of black pixels. Note that the pattern is lessgrainy relative to Figure A.1 but it still needs some artifacts removed.

Filtered Image

Figure A.3: Filtered so that most artifacts are removed.

APPENDIX A. ACQUISITION OF GEOLOGICAL IMAGES 237

from pixel i to the central pixel, c is a constant that smooths the weight function

(as c increases the weight function gets smoother), and p is another constant

that determines how fast the weight decreases with distance (as p increases the

weight decreases faster).

The concept of filtering or smoothing these types of images is discussed in

Switzer, 1983 [130]. A more extensive discussion on the importance of bound-

aries may be found in the same reference.

5. Sometimes the levels of grayness obtained from the moving windows are re-

combined to a subset of gray levels. There may be relatively few identifiable

classes. Considering the smallest number will simplify the analysis presented in

later chapters.

An Abbreviated Catalog of Geological Features

The following digitally encoded geological structures show characteristic features that

are not easily captured by two-points statistics. These images will motivate the de-

velopment of more advanced techniques and provide control patterns for statistical

analysis. These images do not represent an exhaustive illustration of clastic variabil-

ity. There are many other features that may be more important that those shown.

Note that some of the following images still show artifacts due to the scanning and

subsequent manipulation.

Figures A.4, A.5, A.6, and A.7 can generally be classified within the eolian en-

vironment. Figure A.4 is an example of characteristic cross bedding in an eolian

sandstone. Figure A.5 is an example of large-scale wedge and tabular cross strata

in an eolian environment (Zion National Park, Utah, U.S.A.). The photograph was

taken from A Study of Global Sand Seas [104]. Figure A.6 is an example of character-

istic ripple cross lamination structures. Figure A.7 is an example of migrating ripples

from the U.S. Wind Tunnel Laboratory.

Figures A.8 can generally be classified within the fluvial environment. Figure A.8

shows convoluted, deformed laminations in the upper part of a point bar of the Brazos

River, Texas.

APPENDIX A. ACQUISITION OF GEOLOGICAL IMAGES 238

Eolian Sandstone

Figure A.4: A real example of an eolian sandstone. The sandstone was cut into a thinrectangular parralelipiped (about 4 inches by 6 inches), photocopied, and then scanned at300 dpi to provide this image.

Wedge and Tabular Cross Strata

Figure A.5: An example of wedge and tabular cross strata (Zion National Park, Utah,U.S.A.) in an eolian sandstone (the image represents an exposed face of about 75 by 125feet). Photograph taken from A Study of Global Sand Seas [104].

APPENDIX A. ACQUISITION OF GEOLOGICAL IMAGES 239

Ripple Cross Lamination

Figure A.6: A core scale example of ripple cross laminations in an eolian sandstone.

Migrating Ripples

Figure A.7: An example of migrating ripples in a man-made eolian sandstone (Wind TunnelLaboratory).

APPENDIX A. ACQUISITION OF GEOLOGICAL IMAGES 240

Deformed Laminations

Figure A.8: A real example of convoluted and deformed laminations from a fluvial environ-ment. The original core photograph was copied from page 131 of Sandstone DepositionalEnvironments [118].

Figures A.9 and A.10 can generally be classified within a deltaic environment.

Figure A.9 shows lenticular sand laminations representing “starved current ripples”

common in the transition zone between the distal bar and distributary-mouth bar.

Figure A.10 shows large scale cross laminations common near the top of distributary-

mouth bar deposits.

Figure A.11 can generally be classified within an estuarine environment. Fig-

ure A.11 shows anastomosing layers of mud around sand ripples.

The inadequacy of two-points statistics is documented in chapter 2; however,

readers familiar with stochastic simulation techniques may have recognized that many

of the features documented in this section could not be reproduced with simulation

methods based on two-points or bivariate statistics.

APPENDIX A. ACQUISITION OF GEOLOGICAL IMAGES 241

Starved Current Ripples

Figure A.9: A real example of “Starved Current Ripples” from a deltaic environment. Theoriginal photograph was copied from page 162 of Sandstone Depositional Environments[118].

APPENDIX A. ACQUISITION OF GEOLOGICAL IMAGES 242

Large Scale Cross Laminations

Figure A.10: A real example of large-scale cross laminations from a deltaic environment.The original photograph was copied from page 162 of Sandstone Depositional Environments[118].

Anastomosing Mud Layers Around Sand

Figure A.11: A real example of anastomosing mud layers around sand ripples from anestuarine environment. The original photograph was copied from page 182 of SandstoneDepositional Environments [118].

Appendix B

Fluctuations in the cdf and

Variogram due to Ergodicity

Ergodicity is not a property of reality; it is a property of random function models.

The specific theoretical definition of ergodicity is not relevent for the applications

addressed in this appendix, it is enough to quote the ergodic characteristic theorem,

. . . Informally, a random process is said to be ergodic if the statistics of

a single realization of the process, observed over a finite spatial domain,

converge to their expected values as the size of the domain of observation

increases. G.R. Luster page 205 [99]

. . . The RF Z(u) is said to be “ergodic” in the parameter µ if the corre-

sponding realization statistics µ(l), ∀ realizations l, tends toward µ, as

the size of the field A increases. C.V. Deutsch and A.G. Journel page 121

[40]

The statistic or parameter µ is usually taken as the realization mean m =∫A z(u)du.

Ergodicity in the mean can be ensured if the covariance C(h) exists and tends to zero,

as h → ∞. Thus, if the covariance tends to zero (the variogram has a sill) and the

simulated field A becomes large with respect to the range of the covariance or vari-

ogram then the spatial mean mA of any one realization will converge to that of the

RF model.

243

APPENDIX B. FLUCTUATIONS IN THE CDF AND VARIOGRAM DUE TO ERGODICITY244

Although the notion of ergodicity can be applied to any statistic or combination

of statistics, e.g., the variogram itself, checking for ergodic conditions may not be

straightforward. To check convergence of a realization variogram requires knowledge

of fourth order moments, e.g., the fluctuation variance E{[γ(h) − γA(h)]2

}(p. 192

[90]), which are known only when the full spatial law is given by some analytical

model, e.g., the multivariate Gaussian model [134].

When one considers statistics based on any particular realization to infer the

model input statistics, one is implicitly assuming that the RF model considered is

ergodic in those statistics. In the practice of simulation, convergence is never fully

attained because realizations are simulated over finite fields, thus one should expect

fluctuations of the realization statistics from one realization to another. This appendix

provides an appreciation for the fluctuations that can be expected when the spatial

domain A is not large with respect to the range of correlation.

The statistics (histogram and variogram/covariance) of any one finite realization

will differ from the RF model or ensemble statistics. As the extent of the field gets

smaller the fluctuation of the realization statistics becomes greater.

Note that the important point is not the level of discretization. The size of the

field relative to the range of correlation is the important factor.

To illustrate this point consider a standard Gaussian RF model with a standard

normal histogram and an isotropic exponential covariance with effective range 25 grid

units (the a parameter is 25/3 = 8.3333):

C(h) =

[1 − exp(−h

a)

]

Two dimensional realizations, of varying size, are generated using the sgsim program

of GSLIB [40]. The fluctuation of the realization histogram and covariance/variogram

is observed for fields of different size. One hundred realizations of each size as specified

on Table B.1 were generated.

One hundred cumulative distribution functions for each of three selected grid sizes

(25, 100, and 250) are shown on Figure B.1. The one hundred isotropic variograms

for the same three selected grid sizes are shown on Figure B.2.

APPENDIX B. FLUCTUATIONS IN THE CDF AND VARIOGRAM DUE TO ERGODICITY245

Distributions (250 by 250 grid)

Reference N(0,1) Distribution

Rea

lizat

ion

Dis

trib

utio

ns

.00 .20 .40 .60 .80

.00

.20

.40

.60

.80

Distributions (100 by 100 grid)

Reference N(0,1) Distribution

Rea

lizat

ion

Dis

trib

utio

ns

.00 .20 .40 .60 .80

.00

.20

.40

.60

.80

Distributions (25 by 25 grid)

Reference N(0,1) Distribution

Rea

lizat

ion

Dis

trib

utio

ns

.00 .20 .40 .60 .80

.00

.20

.40

.60

.80

Figure B.1: P-P plots for 100 realizations of three different grid sizes.

APPENDIX B. FLUCTUATIONS IN THE CDF AND VARIOGRAM DUE TO ERGODICITY246

Variograms (250 by 250 grid)

Distance (grid units)

Var

iogr

am .0 5.0 10.0 15.0 20.0 25.0

.00

.40

.80

1.20

1.60

Variograms (100 by 100 grid)

Distance (grid units)

Var

iogr

am

.0 5.0 10.0 15.0 20.0 25.0 .00

.40

.80

1.20

1.60

Variograms (25 by 25 grid)

Distance (grid units)

Var

iogr

am

.0 5.0 10.0 15.0 20.0 25.0 .00

.40

.80

1.20

1.60

Figure B.2: Variograms for 100 realizations of three different grid sizes.

APPENDIX B. FLUCTUATIONS IN THE CDF AND VARIOGRAM DUE TO ERGODICITY247

Grid Size Relative Number of Nodes perSize Realization

500 20 250,000250 10 62,500150 6 22,500100 4 10,00075 3 5,62550 2 2,50025 1 625

Table B.1: The grid sizes considered to evaluate the ergodic fluctuations in the histogramand variogram. The size is given relative to the effective range (25).

As expected, the fluctuations in all three cases are centered around the model

statistics, i.e., there is no systematic bias. Also, as expected, there is significantly

more fluctuation for the 25 by 25 field than for the larger 250 by 250 field.

The fluctuation of the realization statistics from the model statistics could be

quantified by any measure of dispersion (e.g., the fluctuation variance as described

in Luster p. 192 [99]). One common, easily understood, measure of dispersion is the

mean absolute deviation:

mADcdf =1

nr

1

nc

nr∑i=1

nc∑j

|cdfmodel(j) − cdfrealization(i, j)| (B.1)

mADvariogram =1

nr

1

nlag

nr∑i=1

nlag∑j

|γmodel(j) − γrealization(i, j)| (B.2)

for nr realizations, nc cdf cutoffs, and nlag lags. In the present case there are 100

realizations (nr = 100), 9 deciles considered (nc = 9), and 24 lags (nlag = 24).

The mAD measure of spread was chosen over the more traditional variance measure

because it is expressed in the unit of the statistics being considered. The mAD

measure of dispersion for both the CDF and the variogram have been plotted versus

the relative size of the simulated field (relative to the effective range of correlation)

on Figures B.3 and B.4.

Comments:

APPENDIX B. FLUCTUATIONS IN THE CDF AND VARIOGRAM DUE TO ERGODICITY248

mAD of Simulated Variogram from Model Variogram

Size of Simulations (relative to range)

mA

D

.0 5.0 10.0 15.0 20.0 25.0 .000

.050

.100

.150

.200

Figure B.3: mAD of the Gaussian realization cdfs from the model cdfs.

mAD of Simulated Variogram from Model Variogram

Size of Simulations (relative to range)

mA

D

.0 5.0 10.0 15.0 20.0 25.0 .000

.050

.100

.150

.200

Figure B.4: mAD of the Gaussian realization variograms from the model variograms.

APPENDIX B. FLUCTUATIONS IN THE CDF AND VARIOGRAM DUE TO ERGODICITY249

• The fluctuations in the cdf and the variogram appear to decrease at the same

rate, e.g., the decrease in the mAD from a field the size of the range to that

10 times the range is 33% and 28% for the cdf and variogram respectively. It

is interesting to note that the bivariate statistics converge at the same rate or

faster than the univariate statistics.

• Note that the number of nodes is the size of the simulation squared, e.g., four

times the number of nodes must be simulated to double the size of the real-

izations. In fact, the increase is a function of dimension being considered; the

increase would be cubic for 3-D simulations.

• Fluctuations in the cdf can be removed by resetting the quantiles of the distri-

bution (see Isaaks and Srivastava [69], p. 469).

The multivariate Gaussian model is notable for its congenial properites (see Chap-

ters 2 and 4). It was conjectured that the fluctuations will be greater for any other

random function model. The median IK model (see Chapter 2), which may be used

to simulate realizations with the same univariate distribution and variogram, was

considered to test this conjecture. One hundred cumulative distribution functions

for the same three grid sizes (25, 100, and 250) are shown on Figure B.5. The one

hundred omnidirectional variograms for the same three selected grid sizes are shown

on Figure B.6.

The mAD dispersion of the realization cdfs from the model and the realization

variograms from the model are shown on Figures B.7 and B.8. These can be com-

pared to those for the Gaussian RF model shown on Figures B.3 and B.4 which are

reproduced in the gray dashed line on Figures B.7 and B.8.

The fluctuations in the cdf and the variogram appear significantly higher than

those for the Gaussian RF model. These ergodic fluctuations serve to expand the

sampled space of uncertainty and could be considered a good property (see Section

2.2).

APPENDIX B. FLUCTUATIONS IN THE CDF AND VARIOGRAM DUE TO ERGODICITY250

Distributions (250 by 250 SIS grid)

Reference N(0,1) Distribution

Rea

lizat

ion

Dis

trib

utio

ns

.00 .20 .40 .60 .80

.00

.20

.40

.60

.80

Distributions (100 by 100 SIS grid)

Reference N(0,1) Distribution

Rea

lizat

ion

Dis

trib

utio

ns

.00 .20 .40 .60 .80

.00

.20

.40

.60

.80

Distributions (25 by 25 SIS grid)

Reference N(0,1) Distribution

Rea

lizat

ion

Dis

trib

utio

ns

.00 .20 .40 .60 .80

.00

.20

.40

.60

.80

Figure B.5: P-P plots for 100 realizations of three different grid sizes.

APPENDIX B. FLUCTUATIONS IN THE CDF AND VARIOGRAM DUE TO ERGODICITY251

Variograms (250 by 250 SIS grid)

Distance (grid units)

Var

iogr

am .0 5.0 10.0 15.0 20.0 25.0

.00

.40

.80

1.20

1.60

Variograms (100 by 100 SIS grid)

Distance (grid units)

Var

iogr

am

.0 5.0 10.0 15.0 20.0 25.0 .00

.40

.80

1.20

1.60

Variograms (25 by 25 SIS grid)

Distance (grid units)

Var

iogr

am

.0 5.0 10.0 15.0 20.0 25.0 .00

.40

.80

1.20

1.60

Figure B.6: Variograms for 100 realizations of three different grid sizes.

APPENDIX B. FLUCTUATIONS IN THE CDF AND VARIOGRAM DUE TO ERGODICITY252

mAD of Simulated CDF from Model CDF

Size of Simulations (relative to range)

mA

D

.0 5.0 10.0 15.0 20.0 25.0 .000

.040

.080

.120

Figure B.7: mAD of the SIS realization cdfs from the model cdfs.

mAD of Simulated Variogram from Model Variogram

Size of Simulations (relative to range)

mA

D

.0 5.0 10.0 15.0 20.0 25.0 .000

.100

.200

.300

.400

Figure B.8: mAD of the SIS realization variograms from the model variograms.

Appendix C

A Detailed Look at Spatial Entropy

As of 1992 the vast majority of geostatistical estimation (prediction) and stochastic

simulation algorithms rely on a covariance model as the sole characteristic of the

spatial distribution of the attribute under study. The choice of the covariance im-

plicitly calls for a multivariate Gaussian model for either the attribute itself or for

its normal scores transform. The choice of a prior Gaussian model could be justified

on the basis that it is both analytically simple and it is a maximum entropy model,

i.e., a model that minimizes unwarranted prior information1. The Gaussian maxi-

mum entropy characteristic also entails maximization of spatial disorder (beyond the

imposed covariance) which may cause flow simulation results performed on multiple

stochastic images to be very similar; thus, the space of (posterior) uncertainty could

be too narrow entailing a misleading sense of safety. The ability of the sole covariance

to describe adequately spatial distributions for flow studies, and the assumption that

maximum spatial disorder amounts to either no information or a safe prior hypothesis

are questioned.

This appendix attempts to clarify the link between entropy and spatial disorder

and to provide, through some examples, an appreciation of the impact of entropy of

prior random function models on the resulting response distributions.

Although extensively used in thermodynamics and information theory [58, 74, 96,

120] the concept of entropy has only been recently introduced in geostatistics [24, 116].

1That is, information that is an artifact of the RF model and not true data.

253

APPENDIX C. A DETAILED LOOK AT SPATIAL ENTROPY 254

Entropy is a measure of the uncertainty of a prior distribution model; the principle

is to maximize the uncertainty (entropy) beyond the statistics that are considered

known. In other words, the prior random function model should account only for the

statistics deemed known. If those statistics reduce to the covariance function, it is

known that the maximum entropy random function model is Gaussian. Whenever

enough data are available to allow inference of a covariance model they also usually

provide valuable additional information; in which case the maximum entropy prior

random function model is no longer Gaussian.

A case-study will show that alternative prior random function models sharing the

same covariance model could lead to much larger (posterior) uncertainty in the re-

sponse variables of transport simulations, such as effective permeability, breakthrough

times, or oil/water sweepage.

C.1 Entropy of Continuous Distributions

Let fz = fz(z1, z2, . . . , zn) be the n-variate probability density function (pdf) of the

n random variables (RV’s) Zi, i = 1, . . . , n. The uncertainty associated to the distri-

bution fz is measured by its entropy defined as [119, 24]:

Hfz = E {−lnfz(z1, . . . , zn)} (C.1)

= −∫ +∞

−∞. . .

∫ +∞

−∞[lnfz(z1, . . . , zn)] fz(z1, . . . , zn)dz1, . . . , dzn

Since the posittive function −ln(y) increases as y decreases, the smaller the pdf value

fz(z1, . . . , zn) the larger its contribution to the entropy measure H . The greater the

uncertainty, the more spread and the smaller the pdf values fz, the larger the entropy.

Note that entropy is unit-free.

The logarithmic function ln(y) comes from thermodynamics and statistical physics

where it naturally arises from the Boltzmann distribution.

APPENDIX C. A DETAILED LOOK AT SPATIAL ENTROPY 255

Univariate pdf:

For n = 1, relation (C.1) defines the entropy of a univariate continuous distribution

with pdf f(z):

Hf = −∫ +∞

−∞[lnf(z)] f(z)dz (C.2)

Classical results include, Shannon and Jones [119, 74].

• The bounded pdf with maximum entropy is the uniform distribution, i.e., f(z) =1

b−afor z ∈ [a, b] and f(z) = 0 otherwise. The entropy is Hf = ln(b − a) which

decreases to −∞ as the interval [a, b] becomes narrower corresponding to greater

certainty.

• The maximum entropy unbounded pdf 2 with a fixed variance is the normal

(Gaussian) distribution independent of its mean or expected value.

Bivariate pdf:

For n = 2, or for two RV’s Z(u) and Z(u + h) separated by a given vector h, the

entropy associated to the bivariate pdf fh(z, z′) is:

Hf(h) = −∫ +∞

−∞

∫ +∞

−∞[lnfh(z, z′)] fh(z, z′)dzdz′ (C.3)

The previous univariate results extend to the bivariate case:

• For bounded RV’s, maximum entropy is obtained for the uniform or rectangular

bivariate pdf.

• Among all bivariate pdf’s sharing the same covariance function

C(h) = Cov{Z(u), Z(u + h)} or the same covariance matrix E{(Z − m)(Z −m)T}, the Gaussian pdf maximizes entropy. Thus, if the only prior structural

information (spatial statistics) retained is the covariance function C(h) the

maximum entropy RF model is the Gaussian model.

2In practice, most histograms are bounded and very few can be considered as normal. However,a normal score transform can be applied (p. 478 [90]), and the previous argument holds for thenormal pdf of the transforms.

APPENDIX C. A DETAILED LOOK AT SPATIAL ENTROPY 256

• Different bivariate pdf’s may share the same covariance C(h) and yet have

different entropy functions Hf(h).

C.2 Entropy of Discrete Distributions

The integral expressions (C.2) and (C.3) of entropy call for continuous pdf’s and are

usually only used for analytical developments. In most practical applications, the

variable Z(u) is either categorical (e.g., rock types) or, if continuous, is discretized

into a finite number of classes, e.g., the classes of its histogram.

Let Z be a discrete RV that can take K outcome values (or be valued in K

predefined classes) with probabilities pk, k = 1, . . . , K, such that∑K

k=1 pk = 1.

The entropy attached to this discrete probability set is defined as:

H = −K∑

k=1

[lnpk]pk ≥ 0 (C.4)

Similarly, consider the stationary discrete RF Z(u) with for stationary marginal

probabilities the set {pk, k = 1, . . . , K}, and for bivariate probabilities the set:

pk,k′(h) = Prob{Z(u) ∈ category k, Z(u + h) ∈ category k′}independent of u; k, k′ = 1, . . . , K (C.5)

Note that∑K

k′ pk,k′(h) = pk = Prob{Z(u) ∈ category k} for all h.

The entropy associated to that set of bivariate probabilities is defined as:

H(h) = −K∑

k=1

K∑k′=1

[lnpk,k′]pk,k′ ≥ 0 (C.6)

Recall: [lnp]p → 0, as p → 0.

Some additional remarks:

• The discrete case entropy is non-negative as compared to the continuous case

where the entropy can be as low as −∞ (because the density fZ(z) can be

greater than 1).

APPENDIX C. A DETAILED LOOK AT SPATIAL ENTROPY 257

• If the RV Z(u) is not naturally categorical but is made discrete through clas-

sification, the class definition impacts the entropy value as defined in relations

(C.4) or (C.6). Hence, before comparing the discrete entropies of originally con-

tinuous distributions, care should be taken to standardize the class definition.

• For lag h = 0 : pk,k′(0) = 0, ∀k �= k′, and: pkk(0) = pk. Thus, the definition (6)

entails:

H(0) = −K∑

k=1

[lnpk]pk ≡ H (C.7)

The univariate entropy represents the lower bound of the bivariate entropy, i.e.,

H(h) ≥ H(0), ∀h. This lower bound can be seen as the case corresponding to

perfect dependence: Z(u + h) ≡ Z(u).

• For lag h = +∞, the two very distant RV’s Z(u) and Z(u+h) are independent

from each other, hence:

pk,k′(+∞) = pkpk′, ∀k, k′

and from relation (6):

H(∞) = −K∑

k=1

K∑k′=1

[lnpk + lnpk′] pkpk′

= −K∑

k=1

lnpk

K∑k′=1

pk′ −K∑

k′=1

lnpk′K∑

k=1

pk = 2H(0)

H(∞) = 2H(0) (C.8)

Therefore, the bivariate entropy is bounded below by the univariate entropy H(0)

and above by twice the univariate entropy:

H(h) ∈ [H(0), 2H(0)] (C.9)

Thus, a standardized relative measure of bivariate entropy is:

HR(h) =H(h) − H(0)

H(0)∈ [0, 1] (C.10)

APPENDIX C. A DETAILED LOOK AT SPATIAL ENTROPY 258

This measure of spatial disorder is compared with the relative semi-variogram defined

as (p. 33 [90]):

γR(h) =1

2

E {[Z(u + h) − Z(u)]2}C(0)

=C(0) − C(h)

C(0)∈ [0, 1] (C.11)

with C(h) being the covariance function Cov{Z(u), Z(u + h)} and C(0) the corre-

sponding variance.

The relative bivariate entropy (C.10) allows comparison of the spatial distribution

of two discrete RF’s Z(u) and Y (u) with different histograms.

C.3 Spatial Entropy

Building on the similarity of the two expressions (C.10) and (C.11) one can define an

average bivariate entropy over a field A of measure |A| as:

HR(A, A) =1

|A|2∫

Adu

∫A

HR(u − u′)du′ ∈ [0, 1] (C.12)

In practice the field A would be discretized by n locations of coordinates ui,

i = 1, . . . , n, and the space integral (C.12) approximated by:

HR(A, A) ∼= 1

n2

n∑i=1

n∑j=1

HR(ui − uj) (C.13)

When A → ∞, HR(A, A) → 1, since HR(∞) → 1.

Expressions (C.12) and (C.13) are similar to the expressions defining the dispersion

(expected spatial) variance of Z(u) within the field A, (p. 67 [90]):

D2(0/A) = E{S2(0/A)} = γ(A, A) =1

|A|2∫

Adu

∫A

γ(u − u′)du′ (C.14)

with S2(0, A) being the randomization of the spatial variance

s2(0, A) =1

|A|∫

A[z(u) − mA]2du and mA =

1

|A|∫

Az(u)du

The average relative entropy measure HR(A, A), defined in (C.12) could be used as a

global measure of spatial entropy (disorder) over the field A. However, whether this

APPENDIX C. A DETAILED LOOK AT SPATIAL ENTROPY 259

measure will prove more useful than the little used dispersion variance (C.14) is yet

questionable.

Note that the measure HR(A, A) utilizes only bivariate entropy when a full mea-

sure of the spatial entropy should use all multivariate entropy functions. In practice,

multivariate pdf’s are inaccessible. Similarly, multivariate entropy functions are not

likely to be accessible.

C.4 Some Examples

The visual difference between the low and high spatial entropy cases in the following

examples is easily appreciated; however, it is more important to demonstrate that

there is a significant effect on the response variables. Such demonstration has been

given in [87] and in chapter 4.

C.4.1 A Discrete Variable Example

Figure C.1 shows two realizations of a 2-D bombing model, i.e., circles of a constant

diameter a with their centers located in space by a Poisson process. An indicator

I(u) is defined as 0 if the point at location u falls within a circle and 1 if it falls

outside. The variogram for this process can be calculated analytically as [102, 122].

γ(h) = p(1 − pCirca(|h|))

where, h is the separation vector of modulus |h|, a is the diameter of the circles, p is

the fraction outside the circles, (1 − p) is the fraction within the circles, p(1 − p) is

the variance of the indicator I(u), and Circa(h) is the circular variogram model:

Circa(h) =

2πa2

[|h|√

a2 − |h|2 + a2sin−1( |h|

a

)], |h| ≤ a

1, |h| ≥ a

This indicator variogram model γ(h) has the familiar p(1−p) variance form with the

exponent corresponding to the hyperspherical variogram model for the dimension of

the space being considered. In 1-D the exponent becomes the linear variogram up

APPENDIX C. A DETAILED LOOK AT SPATIAL ENTROPY 260

Figure C.1: Two realizations of a 2-D bombing model.

to the range a; in 2-D the exponent is the circular variogram model; and in 3-D the

exponent becomes the spherical variogram model.

All of the realizations shown hereafter feature p = 68%±1%. That is, all simulated

indicator realizations have about the same mean and variance.

Figure C.2 shows both the analytical model and the omnidirectional experimental

variograms computed from the two realizations shown on Figure C.1. Note that the

range of the variogram is 20 units (the diameter of the circles) and the size of the

simulated field is 200 units by 200 units, ensuring reasonable ergodicity.

The bombing model images have a very low entropy for the given variogram

model. It is conjectured that these images are minimum entropy realizations for this

variogram.

Realizations generated from a multiGaussian random function model have the

maximum spatial entropy for any given covariance model. A sequential Gaussian

simulation algorithm [40, 56] has been employed to create two realizations that match

the analytical variogram model of the previous 2-D bombing model process, see Fig-

ure C.3. The corresponding omnidirectional experimental variograms and the ana-

lytical model are shown on Figure C.4.

An alternative to the maximum entropy Gaussian RF model is the implicit an-

nealing RF model which remains poorly understood. Two realizations that match

the variogram of the 2-D bombing model process are shown on Figure C.5 and the

APPENDIX C. A DETAILED LOOK AT SPATIAL ENTROPY 261

Var

iogr

am

Distance

Variograms for Bombing Model Realizations

.0 10.0 20.0 30.0 40.0 50.0

.000

.050

.100

.150

.200

Figure C.2: The analytical variogram model and the variogram corresponding to the twoimages shown on Figure C.1.

Figure C.3: Two realizations of a multiGaussian random function model with the variogramof a 2-D bombing model process. For that variogram these are maximum spatial entropyrealizations.

APPENDIX C. A DETAILED LOOK AT SPATIAL ENTROPY 262

Var

iogr

am

Distance

Variograms for multiGaussian Realizations

.0 10.0 20.0 30.0 40.0 50.0

.000

.050

.100

.150

.200

Figure C.4: The analytical variogram model and the variogram corresponding to the twoimages shown on Figure C.3.

First Realization Second RealizationBombing Model 0.766 0.773MultiGaussian 0.930 0.935Annealing 0.887 0.886

Table C.1: The entropy in a unit vertical lag for the bombing model, the multiGaussianimages, and the annealing realizations shown on figures C.1, C.3, and C.5.

corresponding variograms are shown on Figure C.6.

The visual difference between the low (minimum?) entropy and the maximum

entropy realizations is quite dramatic, yet they essentially share the same variogram.

The bivariate entropy for a unit separation vector (lag) in the vertical direction

has been calculated to quantify the visual difference in the realizations generated by

the different methods, see table C.1.

APPENDIX C. A DETAILED LOOK AT SPATIAL ENTROPY 263

Figure C.5: Two realizations of a simulated annealing random function model with thevariogram of a 2-D bombing model process.

Var

iogr

am

Distance

Simulated Annealing Realizations

.0 10.0 20.0 30.0 40.0 50.0

.000

.050

.100

.150

.200

Figure C.6: The analytical variogram model and the variogram corresponding to the twoimages shown on Figure C.5.

APPENDIX C. A DETAILED LOOK AT SPATIAL ENTROPY 264

C.4.2 A Continuous Variable Example

The so-called “mosaic” model is a bivariate mixture of two binormal distributions

with common univariate Gaussian distribution. The first binormal distribution has a

correlation coefficient of 1 (i.e., all values are equal) and the second binormal distri-

bution has a correlation coefficient of 0 (i.e., all values are independent). A mixture of

these two Gaussian distributions with proportion p(h) yields the following “mosaic”

bivariate distribution [101, 77]:

Fh(z, z′) == Prob{Z(u) ≤ z; Z(u+h) ≤ z′} = p(h)F (min(z, z′))+[1 − p(h)] F (z)F (z′)

where z corresponds to Z(u) and z′ corresponds to Z(u+h). Thus, the mosaic model

corresponds to a random function model such that Z(u) and Z(u+h) are equal with

probability p(h) and independent with the complement probability [1 − p(h)].

For a given covariance, the mosaic model has low spatial entropy: its realizations

will appear as a mosaic of independent patches of constant values, see Figure C.7.

A second interesting property of the mosaic model is that indicator correlograms

do not tend towards a pure nugget effect as the cutoff departs symmetrically from

the median, as would be the case for all multiGaussian related models [77], includ-

ing the multilognormal model. In fact, the mosaic indicator correlograms and cross

correlograms are all equal, independently of the cutoff. Therefore, realizations of

this random function model may be obtained by median indicator kriging [77] and

simulation [83]. The median indicator approach corresponds to the case when the

variogram inferred from the median cutoff is used for all other cutoffs. Consequently,

the indicator kriging weights are identical for all cutoffs.

Two realizations of a 200 by 200 unit square area are shown on Figure C.7; they

correspond to a mosaic mixture of two multilognormal distributions. The common

univariate distribution is lognormal with a mean of 1.65 and a variance of 4.67. All

indicator variograms and the normal scores variogram are exponential with no nugget

effect and a practical range of 21 units (3 times the integral range 7):

γ(h) = 1 − e−|h|

7

Figure C.8 shows the analytical and the corresponding omnidirectional experimental

APPENDIX C. A DETAILED LOOK AT SPATIAL ENTROPY 265

.0

1.00

2.0

3.0

4.0

5.0

6.0

7.0

8.0

9.0

10.0

.0

1.00

2.0

3.0

4.0

5.0

6.0

7.0

8.0

9.0

10.0

Figure C.7: Two realizations of the mosaic model generated by median indicator simula-tion.

First Realization Second RealizationMosaic Model 3.480 3.460MultiGaussian 3.670 3.677

Table C.2: The entropy in a unit vertical lag for the mosaic model and multiGaussianimages shown on figures C.7 and C.9.

variograms for the normal scores variable.

Two maximum entropy counterparts with the same univariate lognormal distribu-

tion and normal scores variogram are shown on Figure C.9. These realizations were

generated with the sequential Gaussian simulation algorithm [83]. The analytical and

experimental variograms are shown on Figure C.10.

The significant visual difference between these two random function models may

be quantified by the bivariate entropy for a unit lag in the vertical direction shown

on table C.2.

C.5 Final Thoughts

The Gaussian RF model presents maximum spatial entropy and disorder beyond the

imposed covariance; this does not imply, however, that critical response variables

APPENDIX C. A DETAILED LOOK AT SPATIAL ENTROPY 266

Var

iogr

am

Distance

Normal Scores Variograms

.0 10.0 20.0 30.0 40.0 50.0

.00

.20

.40

.60

.80

1.00

1.20

Figure C.8: The analytical variogram model and the variogram corresponding to the twoimages shown on Figure C.7.

.0

1.00

2.0

3.0

4.0

5.0

6.0

7.0

8.0

9.0

10.0

.0

1.00

2.0

3.0

4.0

5.0

6.0

7.0

8.0

9.0

10.0

Figure C.9: Two realizations of a multilognormal random function model with the vari-ogram of the median indicator simulation realizations of Figure C.8. These are maximumspatial entropy realizations.

APPENDIX C. A DETAILED LOOK AT SPATIAL ENTROPY 267

Var

iogr

am

Distance

Normal Scores Variograms

.0 10.0 20.0 30.0 40.0 50.0

.00

.20

.40

.60

.80

1.00

1.20

Figure C.10: The analytical variogram model and the variogram corresponding to the twoimages shown on Figure C.9.

from some transfer function will present the same maximum entropy character. In

fact, for some transfer functions, the Gaussian maximum entropy model may entail a

too narrow and non-conservative assessment of output uncertainty (see Sections 4.5

and 5.2.2).

Appendix D

Kriging in a Finite Domain

Adopting a random function (RF) model {Z(u),u ∈ study area A} and using the

normal equations (kriging) for estimation amounts to assume that the study area A is

embedded within an infinite domain. At first glance, this assumption has no inherent

limitations since all locations outside A are of no interest and simply not considered.

There is an interesting and practically important consequence that is reflected in

the kriging weights assigned to data contiguously aligned along finite strings. This

appendix, which documents this feature of the normal equations or kriging, is relevant

since that feature affects the quality of realizations generated with the conventional

random function (RF) approach and is not documented in any known publication.

Recall that the posterior conditional cumulative distribution functions (ccdf’s)

needed for stochastic simulation are established by constructing linear combinations

of nearby data values (or specific non-linear transforms of the original data values,

see section 2.3). Simple kriging is the theoretically correct algorithm to arrive at the

weights for each linear combination1. In practice, the more robust ordinary kriging

is often used to filter the global stationary mean from the linear combination.

Figure D.1 shows two commonly encountered situations where finite strings of

contiguously aligned data are used in kriging. In the first case, when strings of

measurements taken along a borehole are truncated by geological or stratigraphic

1This is true for the multiGaussian and the indicator RF models - two of the most common RFmodels used in practice.

268

APPENDIX D. KRIGING IN A FINITE DOMAIN 269

Kriging Within Stratigraphic Limits

drillhole

drillholedrillhole

Kriging Within a Local Search Neighborhood

?

drillhole

drillhole

drillhole

Figure D.1: Two commonly encountered situations when finite strings of contiguouslyaligned data are used in kriging. In the top figure the shaded area represents a stratigraphiclayer of interest. The shaded area in the bottom figure represents the limits of a local searchneighborhood.

APPENDIX D. KRIGING IN A FINITE DOMAIN 270

boundaries, the study area A is clearly finite. In the second case, when strings of

data are truncated by the boundaries of a local search ellipsoid, the area A may be

infinite but the local neighborhood is not.

Figure D.2 illustrates the counter-intuitive weighting scheme that will result when

kriging with a finite string of data. The profile of the ordinary kriging weights is shown

next to the string. The variogram is a common spherical model with a range equal

to the length of the string. Note that the point being estimated is beyond the range

of the variogram and yet the implicit declustering of kriging causes the weights to

change considerably along the vertical extent of the string with the outermost samples

receiving a disproportionately large weight. Further note that the kriging weights will

remain unchanged as the point being estimated moves further away. Kriging yields

this type of weighting because of the implicit assumption that the data are within an

infinite domain - the outermost data inform the infinite half-space beyond the data

string and hence receive greater weights. As the relative nugget effect increases the

weight given to each end point decreases; that is, the central samples are considered

relatively less redundant.

In all three cases of Figure D.2 the point being estimated is beyond the range of

correlation. As the point being estimated gets closer to the string the declustering

becomes less important. This is illustrated on Figure D.3 where the point being

estimated is 1.0, 0.5, and 0.25 dimensionless units from the string. The proximity to

the central samples becomes more important than the declustering as the point being

estimated gets closer.

The artifact weighting is not as pronounced with simple kriging (SK). For ex-

ample, SK applied in all cases shown on Figure D.2 would give a weight of zero to

all data points since the point being kriged is beyond the range of correlation. The

declustering, and the extra weight assigned to the outer samples, becomes important

as the range of correlation increases. This is illustrated with a spherical variogram

with no nugget effect on Figure D.4. The outermost samples are weighted more as

the range is increased from one dimensionless unit (the length of the string) to four

units.

APPENDIX D. KRIGING IN A FINITE DOMAIN 271

Nugget = 0

Data Points Along a String (with Ordinary Kriging Weight)

Point Being Kriged

.338

.039

.037

.035

.034

.034

.034

.035

.037

.039

.338

Nugget = 20%

Data Points Along a String (with Ordinary Kriging Weight)

Point Being Kriged

.233

.108

.061

.043

.036

.035

.036

.043

.061

.108

.233

Nugget = 80%

Data Points Along a String (with Ordinary Kriging Weight)

Point Being Kriged

.118

.101

.089

.080

.075

.073

.075

.080

.089

.101

.118

Figure D.2: A typical situation where the point being kriged is one or more dimensionlessdistance units (one dimensionless unit is the length of the finite string) away from thestring, the spherical variogram has a range equal to one dimensionless unit, and there are11 equally spaced data points along the string.

APPENDIX D. KRIGING IN A FINITE DOMAIN 272

Distance = 1

Data Points Along a String (with Ordinary Kriging Weight)

Point Being Kriged

.233

.108

.061

.043

.036

.035

.036

.043

.061

.108

.233

Distance = 0.5

Data Points Along a String (with Ordinary Kriging Weight)

Point Being Kriged

.137

.084

.074

.078

.084

.087

.084

.078

.074

.084

.137

Distance = 0.25

Data Points Along a String (with Ordinary Kriging Weight)

Point Being Kriged

.057

.054

.072

.103

.138

.153

.138

.103

.072

.054

.057

Figure D.3: The kriging weights are seen to change considerably as the point being esti-mated nears the well. The kriging weights are illustrated when the point is 1.00, 0.50, and0.25 dimensionless units from the well. The nugget effect is 20% in all cases.

APPENDIX D. KRIGING IN A FINITE DOMAIN 273

Range = 1Distance=0.5

Simple Kriging Weights

Point Being Kriged

-.029

.022

.039

.054

.067

.071

.067

.054

.039

.022

-.029

Range = 2

Simple Kriging Weights

Point Being Kriged

.075

.045

.059

.073

.084

.088

.084

.073

.059

.045

.075

Range = 4

Simple Kriging Weights

Point Being Kriged

.135

.048

.061

.075

.086

.090

.086

.075

.061

.048

.135

Figure D.4: The simple kriging (SK) weights are shown for a spherical variogram model asthe range is increased from one dimensionless unit to four dimensionless units. The nuggeteffect is zero and the distance of the point to the string is 0.5 dimensionless units in allcases.

APPENDIX D. KRIGING IN A FINITE DOMAIN 274

Although this odd weighting behavior is theoretically valid, it is not what a practi-

tioner anticipates. In the case of a bounded stratigraphic horizon, there is no volume

outside the limits justifying overweighting of the end values. In the case of a limited

search neighborhood, there are data beyond the limits of the data strings but these

data need not be reflected by the end points of the search neighborhood. Moreover,

one goal of ordinary kriging is to re-estimate locally the mean; assigning more weight

to the end points of strings is contrary to that goal.

The artifacts caused by this overweighting is exacerbated when the data values

present a trend. In particular, a bias may ensue when the variable has relatively

high or low values near the top and bottom of the string (often the case in contact

controlled mineralization, in fining upwards sequences, . . . ).

Empirical Solutions

Four empirical solutions to this problem are proposed below. They are all based on

approaches to trick the normal equations into not assuming an infinite domain.

An Obvious Quick Fix:

One obvious solution is to use only two samples from any one string of data. In

this case, both samples are equally redundant and equally informative of the infinite

domain. The principle disadvantage of this approach is that this may result in too

few data to allow a reliable estimate at each unsampled location.

Extend The String:

The string can be extended by adding phantom data at each end and then removing

the weights assigned to the phantom data (personal communication with R.M. Sri-

vastava, June 1991). A larger ordinary kriging system with non-existent samples at

each end point is solved, then the weights assigned to the non-existent samples are

discarded and the remaining weights are restandardized to sum to 1.0. There are a

number of implementation problems with this approach, some of these being,

APPENDIX D. KRIGING IN A FINITE DOMAIN 275

• The “second” outer points may also receive a significant weight. Adding phan-

tom data (with the same volume support) may not entirely remove the artifact

weighting.

• The local direction vector of the string must be known so that the phantom

samples are assigned the correct location.

• Any given data configuration may contain multiple strings with different num-

bers of data in each string. Adding phantom data with the same support to the

end of each string would disproportionately weight the smaller strings. Ideally,

the phantom samples would have a variable support depending on the number

of samples in a string.

• An octant search could generate multiple strings from the same well. It would

not be straightforward to check all pathological cases of this situation. More-

over, missing samples in a string would further complicate the situation.

Use Simple Kriging:

The effect is not as pronounced with simple kriging; therefore, SK with the mean

determined by local estimation over a finite volume V of interest could be applied

(personal communication with A.G. Journel, January 1992). The idea is to proceed

stepwise:

Compute the local mean with ordinary kriging a finite volume V centered near

the point being estimated. Given n nearby data z(uα), α = 1, . . . , n, an estimate

of the local mean is written as:

m∗ =n∑

α=1

ναz(uα) (D.1)

with m∗ the local mean, z(uα), α = 1, . . . , n, the local data, and να, α = 1, . . . , n,

the weights given by the ordinary kriging system:

∑n

β=1 νβ(u)C(uβ − uα) + µoV = C(V,uα), α = 1, . . . , n∑nβ=1 νβ(u) = 1

(D.2)

APPENDIX D. KRIGING IN A FINITE DOMAIN 276

with C(V,uα) = 1|V |

∫V C(u − uα)du being the average V -data location covari-

ance.

Compute the local estimate with simple kriging using the local mean estimated

from the prior ordinary kriging. The estimate at location u is written as:

z∗(u) =n∑

α=1

ξα [z(uα) − m∗] + m∗ (D.3)

with ξα, α = 1, . . . , n the simple kriging weights given by the SK system:

n∑β=1

ξβ(u)C(uβ − uα) = C(u− uα), α = 1, . . . , n (D.4)

The final weight assigned to each of the local data may be obtained from expression

(D.3) above:

z∗(u) =n∑

α=1

ξα

z(uα) −

n∑β=1

νβz(uβ)

+

n∑β=1

νβz(uβ)

z∗(u) =n∑

α=1

ξαz(uα) −n∑

α=1

ξα

n∑β=1

νβz(uβ) +n∑

β=1

νβz(uβ)

z∗(u) =n∑

β=1

[ξβ + νβ −

n∑α=1

ξανβ

]z(uβ)

Thus, the final weight λα, α = 1, . . . , n, assigned to each datum is:

λα = ξα +

1 −

n∑β=1

ξβ

· να (D.5)

Note that the sum of the final weights∑n

α=1 is equal to one, i.e., the estimator

z∗(u) (D.3) has the same unbiasedness properties as ordinary kriging. Further,

note that as the sum of the SK weights ξα approaches 1.0 the effect of the prior

ordinary kriging of the mean is filtered from the estimate.

The hope with this approach is that the weights for estimating a finite block V will not

show the same artifact overweighting of the outermost points on the string. However,

the artifact overweighting is due to the declustering in the left hand side of the kriging

APPENDIX D. KRIGING IN A FINITE DOMAIN 277

system (D.2) and the final weights λα still show the artifact overweighting. This is

illustrated on Figure D.5 where a zero nugget effect spherical variogram with a range

equal to the length of the string has been used. The kriging weights for the block

mean, at the top, show a significant overweighting of the outermost samples. The

simple kriging weights, in the center of the figure, do not show the effect because of the

relatively short range and the importance of the right hand side closeness covariance

values C(u − uα), α = 1, . . . , n. The final weights, at the bottom of the figure, are a

combination of the two previous sets of weights.

The combination of ordinary kriging to estimate m∗ and then simple kriging (with

m∗) to estimate the local value does not entirely remove the extra weight given to

the end points. Note that the overweighting is significantly less with simple kriging.

Wrap the String:

Another idea is to wrap each finite string of data, i.e., connect the two end points when

building the declustering (left hand side) kriging matrix2 (developed in discussion with

A.G. Journel, 1991).

Consider ns contiguous data i = 1, . . . , ns, aligned in a string each separated from

its neighbors by a vector hs. The covariance between any two data points i and j

(i ≥ j) is,

Ci,j = C(khs) (D.6)

with: k = min {(j − i + ns), (i − j)}None of the data points are actually moved; the above calculation simply modifies

the data-data covariance values C(uα − uβ), α = 1, . . . , n, β = 1, . . . , n, in the left

hand side kriging matrix so that all points in a string are equally redundant 3.

Implementing this correction and repeating the cases shown on Figure D.2 yields

exactly the same result in all cases, i.e., an equal weight of 0.091 for all samples

regardless of the nugget effect. This makes sense since the point being kriged is beyond

2Wrapping the data is equivalent to the “circular stationary” decision adopted by certain tech-niques in statistics and geophysics.

3Interestingly, this correction imparts a perfect banding to the data-data covariance matrix. Thismay allow a faster numerical inversion.

APPENDIX D. KRIGING IN A FINITE DOMAIN 278

Ordinary Kriging for the Block Mean

Point Being Kriged

.236

.054

.058

.060

.061

.061

.061

.060

.058

.054

.236

Simple Kriging

Point Being Kriged

-.029

.022

.039

.054

.067

.071

.067

.054

.039

.022

-.029

Final Weights

Point Being Kriged

.118

.056

.075

.091

.105

.109

.105

.091

.075

.056

.118

Figure D.5: An illustration of the effect of ordinary block kriging to estimate the localmean followed by point simple kriging using that local mean. The variogram is a sphericalmodel with no nugget effect and a range equal to one dimensionless unit (the length of thedata string). The block V is one unit by two units and the point being estimated is 0.5 unitfrom the string.

APPENDIX D. KRIGING IN A FINITE DOMAIN 279

the range of correlation and all data points are now considered equally redundant.

Repeating the cases shown on Figure D.3, i.e., where the point being estimated gets

closer to the string yields the results shown on Figure D.6. The only thing changing

within Figure D.6 is the proximity of the data values to the location being estimated

(the right hand side covariance values C(u−uα), α = 1, . . . , n in the kriging system).

The screen effect, due to the proximity to the unknown, is seen to be more important.

The ad-hoc partial solution of “wrapping” continuous strings of data has been

adopted in other image processing and geophysical applications where the implicit

assumption of an infinite domain has the same effects.

It should be noted that the assumption of an infinite domain is a fundamental

part of kriging. The kriging weights always reflect this assumption; the effects are

more pronounced, however, when dealing with finite strings of contiguous data.

APPENDIX D. KRIGING IN A FINITE DOMAIN 280

Distance = 1

Data Points Along a String (with Kriging Weight)

Point Being Kriged

.091

.091

.091

.091

.091

.091

.091

.091

.091

.091

.091

Distance = 0.5

Data Points Along a String (with Kriging Weight)

Point Being Kriged

.002

.084

.120

.127

.115

.105

.115

.127

.120

.084

.002

Distance = 0.25

Data Points Along a String (with Kriging Weight)

Point Being Kriged

-.065

.070

.127

.149

.148

.141

.148

.149

.127

.070

-.065

Figure D.6: The same cases as Figure D.3 are shown on this figure. The difference isthat the declustering of kriging has been removed. The kriging weights are seen to changeconsiderably as the point being kriged nears the data.

Appendix E

Documentation

Many programs given in the GSLIB software [40] were developed and used throughout

the preparation of this dissertation. The GSLIB source code is widely available and

no part of it will be repeated. The same programming philosophy (including variable

names) used throughout the GSLIB software dominates the source code developed

for this dissertation. Furthermore, GSLIB programs were used for most graphics and

many of the examples and case studies. For these reasons it is suggested that persons

interested in reproducing or extending the results of this dissertation obtain a copy

of GSLIB [40].

This appendix documents two programs for simulation with simulated annealing.

The source code is not listed because of length considerations. The logical construc-

tion of the programs is discussed extensively in Chapter 3 and an overall flowchart

is given in Figure 3.2. The presentation in Chapter 3 and the following discussion

of the input parameter files will enable the interested reader to construct two similar

programs.

The first program sasimi allows the simulation of an integer-coded variable with

two-point statistics and multiple-point statistics using various simulated annealing

decision rules. The second program sasimr allows the simulation of a continuous

variable conditional to two-point statistics and well test-derived effective properties.

It is always possible to code a continuous variable as an integer-coded variable (with

some loss of precision) when a feature available with a different sasim program is

281

APPENDIX E. DOCUMENTATION 282

needed. The overall flowchart for both of these programs as well as a general descrip-

tion of how they work was presented in Chapter 3. This appendix presents the source

code and a detailed description of the input parameters.

.

E.1 Program sasimi:

The sasimi program is a general purpose stochastic simulation program for integer-

coded categorical variables that can be used to generate realizations from an initial

random realization or to post-process prior realizations.

The program allows conditioning to:

1. A variogram (computed from some realization or a model).

2. Two-point histogram with or without cross terms (computed from some real-

ization or a model).

3. Multiple-point non-centered indicator covariance functions (see section 2.3.2 and

Chapter 3, the indicator function at any location may defined in the traditional

way I(u) or as J(u) = 1 − I(u).

4. Multiple-point histogram with or without cross terms (see section 2.3.2 and

Chapter 3).

The parameters required by sasimi are illustrated on Figure E.1 and described below:

• nsim: the number of simulations requested.

• nx, ny, and nz: the number of nodes in the x, y, and z direction respectively.

• nr: the number of integer-coded classes to consider. The program sasimi con-

siders only integer coded categorical variables. Variables that are not naturally

categorical must be separated into a number of categories or classes.

APPENDIX E. DOCUMENTATION 283

HIGHSIM: Input Parameters *************************

START OF PARAMETERS:1 \Number of simulations25 25 1 2 \nx, ny, nz, nr (input classes)0 0 0 1 \control: var, tp, hic, hih1 1 1 1 \relative importance69069 \Random number seedhighsim.out \Output File for realization(s)highsim.sta \Output File for statisticshighsim.dbg \Output File for Debugging12 100 \Debug level, Reporting Interval

1 \starting image option../ibt/ibt01.io \ starting image filecondit.dat \Conditioning data file1 2 0 3 \ ixl, iyl, izl, ivrl4 1 \perturbation option and increm1.0e-21 25 \ target objective function1.0 0.1 62500 6250 3 0.0000001 \ SA Schedule:t0,lam,ka,k,e,del1.0 0.1 1000 \ TA Schedule:increment, lam

1 \VARIOGRAM specification optionshales.dat \ file for control (unless =2)2 \ number of directions1 0 0 4 \ ix, iy, iz, nlag0 1 0 4

1 \TWO POINT HIST optionshales.dat \ file for control (unless =2)2 \ number of directions1 0 0 4 1 \ ix, iy, iz, nlag,icross(0=no)0 1 0 4 1

1 \HIGH ORDER COVARIANCE optionshales.dat \ file for control (unless =2)2 \ number of covariances 4 1 0 \ n-pts, cutoff,(0=low,1=hi) 0 0 0 \ points, 1,...,n-points 1 0 0 \ points, 1,...,n-points 2 0 0 \ points, 1,...,n-points 3 0 0 \ points, 1,...,n-points 4 2 1 \ n-pts, cutoff,(0=low,1=hi) 0 0 0 \ points, 1,...,n-points 1 0 0 \ points, 1,...,n-points 2 0 0 \ points, 1,...,n-points 3 0 0 \ points, 1,...,n-points

1 \HIGH ORDER HISTOGRAM optionshales.dat \ file for control (unless =2) 4 1 \ n-pts, (0=direct,1=all) 0 0 0 \ points, 1,...,n-points 1 0 0 \ points, 1,...,n-points 2 0 0 \ points, 1,...,n-points 3 0 0 \ points, 1,...,n-points

Figure E.1: An example parameter file for sasimi.

APPENDIX E. DOCUMENTATION 284

• logv: a logical flag indicating whether variogram-type control statistics are to

be considered (1=yes, 0=no). If logv=1 then the variogram specification options

are actually used.

• logtp: a logical flag indicating whether two-point histogram control statistics

are to be considered (1=yes, 0=no). If logtp=1 then the two point specification

options are actually used.

• loghic: a logical flag indicating whether high order non-centered covariance

functions are to be considered (1=yes, 0=no). If loghic=1 then the high order

multiple point covariance function specification options are actually used.

• loghih: a logical flag indicating whether multiple-point histograms are to be

considered (1=yes, 0=no). If loghih=1 then the high order multiple point his-

togram specification options are actually used.

• relimp(): the relative importance of each of the four statistical measures (vari-

ogram, two-point histogram, multiple-point indicator covariance, and multiple-

point histogram). Only the relative magnitude of the numbers with respect

to each other is important. In most cases no preferential weighting need be

considered.

• seed: the random number seed. Large (> 103) odd integers are good choices.

• outfl: the output file. This file will contain the nsim final simulations.

• stafl: the file for the output statistics. This file will contain the input and out-

put control statistics (to check how close the control statistics are reproduced).

• dbgfl: the file for the debugging output. This file will contain the debugging

output (the echoed input parameters, warning messages, . . . ).

• idbg: the debugging level. This integer debugging level may be increased from

0 through 10 to achieve more detailed reports of the program execution.

APPENDIX E. DOCUMENTATION 285

• report: the reporting interval. After report perturbations the program will

report on the current status (objective function, . . . ).

• inputo: the input (or starting image) option. The allowable options:

1. a random image will be considered in all cases. The univariate distribution

will be taken from a control image.

2. a single starting image, to be used for all simulations, is read from the

input file inpfl (see below).

3. different starting images are to be sequentially read from the input file

inpfl (see below).

• inpfl: the input file with the starting pattern(s). This file is considered only if

inputo is set to 2 (only one grid is needed) or 3 (there should be nsim grids in

the file). This file should contain a three line header (as if the file were a GSLIB

[40] or Geo-EAS [47] file with one variable) and then the image written with a

Fortran (40i2) format cycling fastest on x, then y, then z. For example, the

fortran code to read the grid:

do iz=1,nz

do iy=1,ny

read(1,’(40i2)’) (val(ix,iy,iz),ix=1,nx)

enddo

enddo

Note that all input and output integer coded grids are written with this format.

• condfl: the conditioning data file. If this file exists, the data points and values

are imposed on all of the nsim simulations. The columns for the grid location

and data value are controlled the same as the GSLIB [40] format (see below).

• ixc, iyc, izc and ivrc: the column location for the x node index, the y node

index, the z node index, and the variable.

APPENDIX E. DOCUMENTATION 286

• perturbo: the option controlling how to perturb the image. The allowable

options:

1. the grid node values are perturbed and acceptance is based on the “sim-

ulated annealing” criteria (the schedule will be automatically established

within the program).

2. the grid node values are perturbed and acceptance is based on the “simu-

lated annealing” criteria (the annealing schedule is set by the user).

3. the grid node values are perturbed and acceptance is based on the “thresh-

old accepting” criteria (the schedule is set by the user).

4. the grid node values are perturbed and acceptance is based on the “MAP”

criteria, see Section 2.5.2.

• nperturb: the number of nodes to perturb before applying the acceptance rule.

Traditionally this is set to one; however, by setting this value greater than one

it allows a better ability to escape local minima.

• objtarg: target objective function. This is usually set very low, say, 0.000001.

• sas(): simulated annealing schedule used if perturbo=2.

– sas(1) = the initial temperature (typically 1.0).

– sas(2) = the reduction parameter (between 0.0 and 1.0).

– sas(3) = ka, the maximum to try before reducing the temperature.

– sas(4) = k, the maximum to accept before reducing the temperature.

– sas(5) = e, the maximum number of times that ka is reached without

reaching k.

• tas(): threshold accepting schedule used if perturbo=3.

– tas(1) = the initial temperature parameter (typically 1.0).

– tas(2) = the reduction parameter (between 0.0 and 1.0).

APPENDIX E. DOCUMENTATION 287

– tas(3) = the maximum number of times to reduce the temperature param-

eter.

• varspec: the control variogram specification option (only used if logv=1). The

allowable options:

1. the control variogram is taken from the control pattern.

2. the control variogram is taken from the initial image (for situations when

the goal is to perturb minimally the spatial configuration of a starting

image).

3. the control variogram values are directly input (e.g. from an analytical

model).

• varfl: the file containing the image for the control variogram. This file must

contain the variogram values if varspec=3. Note that sasimi attempts to read

the variogram model in the same format as produced by vmodel [40].

• ndirv: the number of directions to compute the control variogram.

• ixv, iyv, izv, and nlagv: control the direction and number of lags for each of

the ndirv directions. The direction is specified by the grid node offsets (ixv,iyv,

and izv) that must be considered to move one unit lag. There need not be the

same number of lags considered for all directions.

• tpspec: the control two-point histogram specification option (only used if

logtp=1). The allowable options:

1. the control two-point histogram is taken from the control pattern.

2. the control two-point histogram is taken from the initial image (for situa-

tions when the goal is to perturb minimally the spatial configuration of a

starting image).

3. the control two-point histogram values are directly input (e.g. from an

analytical model).

APPENDIX E. DOCUMENTATION 288

• tpfl: the file containing the image for the control two-point histogram. This

file must contain the two-point histogram values if tpspec=3.

• ndirtp: the number of directions to compute the control two-point histogram.

• ixtp, iytp, iztp, nlagtp, and crosstp: control the direction, number of lags,

and whether or not to include the “cross” terms for each of the ndirtp directions.

The direction is specified by the grid node offsets (ixtp,iytp, and iztp) that must

be considered to move one unit lag. The number of lags is specified differently for

each of the directions. Moreover, the flag crosstp is set to 1 if cross information

is to be included and to 0 if not.

• hicspec: the control high order indicator covariance specification option (only

used if loghic=1). The allowable options:

1. the control covariances are taken from the control pattern.

2. the control covariances are taken from the initial image.

3. the control covariances are directly input (perhaps from an analytical

model).

• hicfl: the file containing the image for the control high order statistics. This

file must contain the high order indicator covariances if hicspec=3.

• nhicov: the number of high order indicator covariances to consider.

• For each of the nhicov high order indicator covariances the following parameters

are needed:

– npthic: the number of control points for the indicator covariance.

– icuthi: the integer cutoff value (the critical class).

– indhic: an integer flag specifying the definition of the indicator. indhic=0

implies the traditional definition: I(u) = 1 if the code at u is less than

or equal to the cutoff value icuthi. indhic=1 implies the non-traditional

definition: J(u) = 1 if the code at u is greater than or equal to the cutoff

value icuthi.

APPENDIX E. DOCUMENTATION 289

– ixhic, iyhic, and izhic: the points that control the configuration of the

high order indicator covariance. Note that the first point should always be

0,0,0.

• hihspec: the control high order histogram specification option (only used if

loghih=1). The allowable options:

1. the control histogram values are taken from the control pattern.

2. the control histogram values are taken from the initial image.

3. the control histogram values are directly input (perhaps from an analytical

model).

• hihfl: the file containing the image for the control high order statistics. This

file must contain the high order indicator covariances if hihspec=3.

• npthih: the number of control points for the multiple-point histogram.

• indhih: an integer flag specifying whether or not to condition to the “cross”

classes or those involving more than one code. indhih=0 implies that only the

nr direct classes will be used for conditioning. indhih=1 implies that all classes

of the histogram will be used for conditioning.

• ixhih, iyhih, and izhih: the points that control the configuration of the high

order indicator histogram. Note that the first point should always be 0,0,0.

E.2 Program sasimr:

The sasimr program is a general purpose stochastic simulation program that can be

used to generate realizations of a continuous variable from an initial random realiza-

tion or to post-process prior realizations.

The program based on sasim of GSLIB [40] allows conditioning to:

1. a variogram (computed from some realization or a model).

2. well test-derived effective permeabilities.

APPENDIX E. DOCUMENTATION 290

The parameters required by sasimr are illustrated on Figure E.2 and described below:

• condfl: the conditioning data file. If this file exists, the data points will be

fixed in all of the nsim simulations. The columns for the grid location and data

value are controlled the same as the GSLIB [40] format (see below).

• ixc, iyc, izc and ivrc: the column location for the x node index, the y node

index, the z node index, and the variable.

• tmin and tmax: trimming limits to remove missing values.

• idist: flag specifying whether the univariate distribution is Gaussian (idist = 1)

or to be defined by data values (idist = 0).

• distfl: the file with the input univariate distribution (if idist = 0).

• ivr and iwt: the column location for the variable and an optional declustering

weight. If iwt ≤ 0 then all values will be given the same weight.

• zmin and zmax: the minimum and maximum data allowable data value. These

may be used in the back transformation procedure.

• ltail and ltpar specify extrapolation in the lower tail of the distribution: ltail =

1 implements linear interpolation to the lower limit zmin and ltail = 2 imple-

ments power model interpolation, with ω = ltpar, to the lower limit zmin.

• utail and utpar specify extrapolation in the upper tail of the distribution:

utail = 1 implements linear interpolation to the upper limit zmax, utail = 2

implements power model interpolation, with ω = utpar, to the upper limit zmax,

and utail = 4 implements hyperbolic model extrapolation with ω = utpar.

• outfl: the output grid is written to this file. The output file will contain the

results, cycling fastest on x, then y, then z, then simulation by simulation.

• varfl: this output file will contain the model variogram, and the variogram

of the final grid. If the two parts objective function is used the variogram of

APPENDIX E. DOCUMENTATION 291

Simulated Annealing Simulation ******************************

START OF PARAMETERS:welldata.dat \Conditioning Data (if any)1 2 0 3 \columns: x,y,z,vr-1.0e21 1.0e21 \data trimming limits1 \0=non parametric; 1=Gaussiannodata.dat \Non parametric Distribution1 0 \columns: vr,wt0.0 30.0 \minimum and maximum data values1 0.0 \Lower tail option and parameter1 45.0 \Upper tail option and parametersasimw.out \Output File for simulationsasimw.var \Output File for variogram10 50 \Debug level, Reporting Intervalsasimw.dbg \Output File for Debugging1 \Annealing schedule? (0=auto)0.2 0.1 50 25 10 0.00001 \Manual Schedule2 \1 or 2 part objective function112063 \Random number seed10 \Number of simulations51 0.0 20.0 \nx,xmn,xsiz51 0.0 20.0 \ny,ymn,ysiz1 0.0 10.0 \nz,zmn,zsiz55.0 \radius18 1 \ndir,nlag0 1 0 \ixl(i),iyl(i),izl(i)0 2 00 3 01 3 01 2 01 1 01 0 01 -1 01 -2 01 -3 02 2 02 1 02 0 02 -1 02 -2 03 1 03 0 03 -1 01 0.1 0 \nst, nugget, (0=renormalize)2 8.33 0.9 \it,aa,cc: STRUCTURE 10.0 0.0 0.0 1.0 1.0 \ang1,ang2,ang3,anis1,anis2:1 \number of wells25 25 1 8.194 0.55 \ixw,iyw,izw, Keff, omega0.33 5000.0 \well radius, max wt radius

Figure E.2: An example parameter file for sasimr.

APPENDIX E. DOCUMENTATION 292

both parts will be in the file. The format is the same as the output from the

variogram programs in chapter 2.

• idbg: an integer debugging level between 0 and 4. The larger the debugging

level the more information written out.

• dbgfl: the file for the debugging output.

• isas: the annealing schedule (next set of parameters) can be set explicitly or it

can be set automatically (0=automatic,1=then use the following:).

• sas(6): the annealing schedule: initial temperature, the reduction factor, the

maximum number of swaps at any one given temperature, the target number

of swaps, the stopping number, and a low objective function value indicating

convergence.

• part: a one part objective function can be used that considers the overall

variogram (part = 1), or a two parts objective function that separates the pairs

involving an original conditioning data can be used (part = 2).

• seed: random number seed.

• nsim: the number of simulations to generate.

• nx, xmn, xsiz: definition of the grid system (x axis).

• ny, ymn, ysiz: definition of the grid system (y axis).

• nz, zmn, zsiz: definition of the grid system (z axis).

• radius: the maximum isotropic distance separating a pair of values to be

swapped.

• ndir and nlag: parameters defining how many directions and lags to be con-

sidered in the objective function. These should be kept quite low (less than 2-3)

or else convergence could be very slow.

APPENDIX E. DOCUMENTATION 293

• ixl, iyl, and izl: the offsets in each coordinate direction to define the ndir

directions in the objective function.

• A complete 3-D variogram model must be specified with nst = the number of

variogram structures, c0 = the isotropic nugget effect; for each of the nst nested

structures one must define it the type of structure, aa the a parameter, cc the

c parameter, ang1, ang2, ang3, anis1 and anis2 the geometric anisotropy

parameters. A detailed description of these parameters is given in section II.3. It

is essential that the variance of the values of the initial random image match the

spatial (dispersion) variance implied by the variogram model. That dispersion

variance should be equal to the total sill if it exists (i.e., a power model has

not been used) and if the size of the field is much larger than the largest range

in the variogram model. Otherwise, the dispersion variance can be calculated

from traditional formulae ([90], p. 61-67).

• nwell: the number of well tests to be entered into the objective funciton. The

following parameters must be repeated nwell times:

• ixw, iyw, and izw: the grid block location of the well(s).

• keff and omega: the effective permeability and the averaging power (poten-

tially different for each well).

• wellrad and drainrad: the radius of the well (in feet) and the drainage radius

(in the units of the grid blocks, e.g., xsiz)

Bibliography

[1] E. Aarts and J. Korst. Simulated Annealing and Boltzmann Machines. John

Wiley & Sons, New York, NY, 1989.

[2] M. Abramovitz and I. Stegun, editors. Handbook of Mathematical Functions:

with Formulas, Graphs, and Mathematical Tables. Dover, New York, NY, 1972.

9th (revised) printing.

[3] Adobe Systems Incorporated. PostScript Language Reference Manual. Addison-

Wesley, Menlo Park, CA, 1985.

[4] Adobe Systems Incorporated. PostScript Language Tutorial and Cookbook.

Addison-Wesley, Menlo Park, CA, 1985.

[5] F. Alabert. The practice of fast conditional simulations through the LU decom-

position of the covariance matrix. Math Geology, 19(5):369–386, 1987.

[6] F. Alabert. Stochastic imaging of spatial distributions using hard and soft

information. Master’s thesis, Stanford University, Stanford, CA, 1987.

[7] F. Alabert. Constraining description of randomly heterogeneous reservoirs to

pressure test data: a Monte Carlo study. In SPE Annual Conference and Ex-

hibition, San Antonio. Society of Petroleum Engineers, October 1989.

[8] F. Alabert and G. J. Massonnat. Heterogeneity in a complex turbiditic reservoir:

Stochastic modelling of facies and petrophysical variability. In 65th Annual

Technical Conference and Exhibition, number 20604, pages 775–790. Society of

Petroleum Engineers, September 1990.

294

BIBLIOGRAPHY 295

[9] T. Anderson. An Introduction to Multivariate Statistical Analysis. John Wiley

& Sons, New York, NY, 1958.

[10] H. Andrews and B. Hunt. Digital Image Restoration. Prentice Hall, Englewood

Cliffs, NJ, 1989.

[11] K. Aziz and A. Settari. Petroleum Reservoir Simulation. Applied Science Pub-

lishers, New York, NY, 1979.

[12] P. Ballin. Quantile-preserving approximations of complex transfer functions: A

waterflood example. In Report 3, Stanford Center for Reservoir Forecasting,

Stanford, CA, May 1990.

[13] P. Ballin, A. Journel, and K. Aziz. Prediction of uncertainty in reservoir per-

formance forecasting. JCPT, 31(4), April 1992.

[14] M. Barnsley. Fractals Everywhere. Academic Press, San Diego, 1988.

[15] A. Basu and L. N. Frazer. Rapid determination of the critical temperature in

simulated annealing inversion. Science, 249:1409–1412, September 1990.

[16] S. Begg, R. Carter, and P. Dranfield. Assigning effective values to simulator

gridblock parameters for heterogeneous reservoirs. SPE Reservoir Engineering,

pages 455–463, November 1989.

[17] J. Berger. Statistical Decision Theory and Bayesian Analysis. Springer Verlag,

New York, NY, 1980.

[18] J. Besag. On the statistical analysis of dirty pictures. J.R. Statistical Society

B, 48(3):259–302, 1986.

[19] D. Bounds. New optimization methods from physics and biology. Nature,

329(17):215–219, September 1987.

[20] G. Box and G. Jenkins. Time Series Analysis forecasting and control. Holden-

Day, Oakland, CA, second edition, 1976.

BIBLIOGRAPHY 296

[21] J. Bridge and M. Leeder. A simulation model of alluvial stratigraphy. Sedimen-

tology, 26:617–644, 1979.

[22] P. Brooker. Kriging. Engineering and Mining Journal, 180(9):148–153, 1979.

[23] J. Butler Jr. Pumping Tests in Non-uniform Aquifers: a Determinis-

tic/Stochastic Analysis. PhD thesis, Stanford University, Stanford, CA, 1987.

[24] G. Christakos. A Bayesian/maximum-entropy view to the spatial estimation

problem. Math Geology, 22(7):763–777, 1990.

[25] J. Chu, W. Xu, H. Zhu, and A. Journel. The Amoco case study. In Report 4,

Stanford Center for Reservoir Forecasting, Stanford, CA, May 1991.

[26] The Complete PC Inc., 1983 Concourse Drive, San Jose, CA. The Complete

Page Scanner: User’s Guide Version 2.2, 1989.

[27] D. Cuthiell, S. Bachu, J. Kramers, and L. Yuan. Characterizing shale clast

heterogeneities and their effect on fluid flow. In Second International Symposium

on Reservoir Heterogeneities, June 1989.

[28] M. Dagbert, M. David, D. Crozel, and A. Desbarats. Computing variograms

in folded strata-controlled deposits. In G. Verly et al., editors, Geostatistics for

natural resources characterization, pages 71–89. Reidel, Dordrecht, Holland,

1984.

[29] M. David. Geostatistical Ore Reserve Estimation. Elsevier, Amsterdam, 1977.

[30] B. Davis. Indicator kriging as applied to an alluvial gold deposit. In G. Verly

et al., editors, Geostatistics for natural resources characterization, volume 1,

pages 337–348. Reidel, Dordrecht, Holland, 1984.

[31] J. Davis. Statistics and Data Analysis in Geology. John Wiley & Sons, New

York, NY, 2nd edition, 1986.

[32] M. Davis. Production of conditional simulations via the LU decomposition of

the covariance matrix. Math Geology, 19(2):91–98, 1987.

BIBLIOGRAPHY 297

[33] J. Denbigh. Entropy in Relation to Incomplete Knowledge. Cambridge Univer-

sity Press, New York, NY, 1985.

[34] A. Desbarats. Stochastic Modeling of Flow in Sand-Shale Sequences. PhD thesis,

Stanford University, Stanford, CA, 1987.

[35] C. Deutsch. A probabilistic approach to estimate effective absolute permeability.

Master’s thesis, Stanford University, Stanford, CA, 1987.

[36] C. Deutsch. Calculating effective absolute permeability in sandstone/shale se-

quences. SPE Formation Evaluation, pages 343–348, September 1989.

[37] C. Deutsch. DECLUS: A Fortran 77 program for determining optimum spatial

declustering weights. Computers & Geosciences, 15(3):325–332, 1989.

[38] C. Deutsch. Relative influence on flow performance of spatial heterogeneities

versus uncertainty in relative permeability curves. In Report 3, Stanford Center

for Reservoir Forecasting, Stanford, CA, May 1990.

[39] C. Deutsch and A. Journel. The application of simulated annealing to stochastic

reservoir modeling. Submitted to J. of. Pet. Tech., January 1991.

[40] C. Deutsch and A. Journel. GSLIB: Geostatistical Software Library. to be

published by Oxford University Press, New York, NY, 1992.

[41] C. Deutsch and R. Lewis. Advances in the practical implementation of indicator

geostatistics. In Proceedings of the 23rd International APCOM Symposium,

pages 133–148, Tucson, AZ, April 1992. Society of Mining Engineers.

[42] J. Doob. Stochastic Processes. John Wiley & Sons, New York, NY, 1953.

[43] P. Doyen. Porosity from seismic data: A geostatistical approach. Geophysics,

53(10):1263–1275, 1988.

[44] P. Doyen and T. Guidish. Seismic discrimination of lithology: A Bayesian

approach. In Geostatistics Symposium, Calgary, AB, May 1990.

BIBLIOGRAPHY 298

[45] O. Dubrule and C. Kostov. An interpolation method taking into account in-

equality constraints. Math Geology, 18(1):33–51, 1986.

[46] G. Dueck and T. Scheuer. Threshold accepting: A general purpose optimization

algorithm appearing superior to simulated annealing. Journal of Computational

Physics, 90:161–175, 1990.

[47] E. Englund and A. Sparks. Geo-EAS 1.2.1 User’s Guide, EPA Report # 60018-

91/008. EPA-EMSL, Las Vegas, NV, 1988.

[48] Exploration Consultants Limited, Highlands Farm, Greys Road, Henley-on-

Thames, Oxon, England. ECLIPSE Reference Manual, November 1984.

[49] C. Farmer. Numerical rocks. In J. Fayers and P. King, editors, The Mathematical

Generation of Reservoir Geology, New York, NY, 1991. Oxford University Press.

[50] L. Gandin. Objective Analysis of Meteorological Fields. Gidrometeorologich-

eskoe Izdatel’stvo (GIMEZ), Leningrad, 1963. Reprinted by Israel Program for

Scientific Translations, Jerusalem, 1965.

[51] I. Gelfand. Generalized random processes. volume 100, pages 853–856. Dokl.

Acad. Nauk., USSR, 1955.

[52] S. Geman and D. Geman. Stochastic relaxation, Gibbs distributions, and the

Bayesian restoration of images. IEEE Transactions on Pattern Analysis and

Machine Intelligence, PAMI-6(6):721–741, November 1984.

[53] R. Giordano, S. Salter, and K. Mohanty. The effect of permeability variations

on flow in porous media. SPE paper # 14365, 1985.

[54] A. Goldberger. Best linear unbiased prediction in the generalized linear regres-

sion model. JASA, 57:369–375, 1962.

[55] J. Gomez-Hernandez. A Stochastic Approach to the Simulation of Block Con-

ductivity Fields Conditioned upon Data Measured at a Smaller Scale. PhD

thesis, Stanford University, Stanford, CA, 1991.

BIBLIOGRAPHY 299

[56] J. Gomez-Hernandez and R. Srivastava. ISIM3D: An ANSI-C three dimensional

multiple indicator conditional simulation program. Computers & Geosciences,

16(4):395–440, 1990.

[57] D. Guerillot, J. Rudkiewicz, C. Ravenne, G. Renard, and A. Galli. An integrated

model for computer aided reservoir description: from outcrop study to fluid flow

simulations. In European Symposium on Improved Oil Recovery, April 1989.

[58] S. Gull and J. Skilling. The entropy of an image. In C. Smith and W. Gandy Jr.,

editors, Maximum Entropy and Bayesian Methods in Inverse Problems, pages

287–301. Reidel, Dordrecht, Holland, 1985.

[59] H. Haldorsen and E. Damsleth. Stochastic modeling. J. of Pet. Technology,

pages 404–412, April 1990.

[60] D. Han. Effects of Porosity and Clay Content on Acoustic Properties of Sand-

stones and Unconsolidated Sediments. PhD thesis, Stanford University, Stan-

ford, CA, 1986.

[61] T. Hewett. Fractal distributions of reservoir heterogeneity and their influence

on fluid transport. SPE paper # 15386, 1986.

[62] T. Hewett and R. Behrens. Conditional simulation of reservoir heterogeneity

with fractals. Formation Evaluation, pages 300–310, 1990.

[63] T. Hewett and R. Behrens. Considerations affecting the scaling of displacements

in heterogeneous permeability distributions. In SPE Annual Conference and

Exhibition, New Orleans, LA, number 20739, pages 223–244, New Orleans, LA,

September 1990. Society of Petroleum Engineers.

[64] T. Hewett and R. Behrens. Scaling laws in reservoir simulation and their use

in a hybrid finite difference/streamtube approach to simulating the effects of

permeability heterogeneity. Unpublished Research Report, 1990.

[65] M. Hohn. Geostatistics and Petroleum Geology. Van Nostrand, New York, NY,

1988.

BIBLIOGRAPHY 300

[66] R. Horne. Modern Well Test Analysis. Petroway Inc, 926 Bautista Court, Palo

Alto, CA, 94303, 1990.

[67] E. Isaaks. Indicator simulation: Application to the simulation of a high grade

uranium mineralization. In G. Verly et al., editors, Geostatistics for natural

resources characterization, pages 1057–1069. Reidel, Dordrecht, Holland, 1984.

[68] E. Isaaks. The Application of Monte Carlo Methods to the Analysis of Spatially

Correlated Data. PhD thesis, Stanford University, Stanford, CA, 1990.

[69] E. Isaaks and R. Srivastava. An Introduction to Applied Geostatistics. Oxford

University Press, New York, NY, 1989.

[70] E. Jaynes. Entropy and search theory. In C. Smith and W. Gandy Jr., editors,

Maximum Entropy and Bayesian Methods in Inverse Problems, pages 443–454.

Reidel, Dordrecht, Holland, 1985.

[71] E. Jaynes. Where do we go from here? In C. Smith and W. Gandy Jr., editors,

Maximum Entropy and Bayesian Methods in Inverse Problems, pages 21–58.

Reidel, Dordrecht, Holland, 1985.

[72] P. Johnson. The relationship between radius of drainage and cumulative pro-

duction. SPE Formation Evaluation, pages 267–270, March 1988.

[73] R. Johnson and D. Wichern. Applied Multivariate Statistical Analysis. Prentice

Hall, Englewood Cliffs, NJ, 1982.

[74] D. Jones. Elementary Information Theory. Clarendon Press, Oxford, 1979.

[75] T. Jones, D. Hamilton, and C. Johnson. Contouring of Geological Surfaces with

the Computer. Van Nostrand Reinhold, New York, NY, 1986.

[76] A. Journel. Geostatistics for conditional simulation of orebodies. Economic

Geology, 69:673–680, 1974.

[77] A. Journel. Non-parametric estimation of spatial distributions. Math Geology,

15(3):445–468, 1983.

BIBLIOGRAPHY 301

[78] A. Journel. Mad and conditional quantile estimators. In G. Verly et al., editors,

Geostatistics for natural resources characterization, volume 2, pages 915–934.

Reidel, Dordrecht, Holland, 1984.

[79] A. Journel. The place of non-parametric geostatistics. In G. Verly et al., editors,

Geostatistics for natural resources characterization, volume 1, pages 307–355.

Reidel, Dordrecht, Holland, 1984.

[80] A. Journel. Recoverable reserves estimation - the geostatistical approach. Min-

ing Engineering, pages 563–568, June 1985.

[81] A. Journel. Constrained interpolation and qualitative information. Math Geol-

ogy, 18(3):269–286, 1986.

[82] A. Journel. Geostatistics: Models and tools for the earth sciences. Math Geol-

ogy, 18(1):119–140, 1986.

[83] A. Journel. Fundamentals of Geostatistics in Five Lessons. Volume 8 Short

Course in Geology. American Geophysical Union, Washington, D.C., 1989.

[84] A. Journel. Petroleum geostatistics in North America. In Report 4, Stanford

Center for Reservoir Forecasting, Stanford, CA, May 1991.

[85] A. Journel and F. Alabert. Non-Gaussian data expansion in the earth sciences.

Terra Nova, 1:123–134, 1989.

[86] A. Journel and F. Alabert. New method for reservoir mapping. J. of Pet.

Technology, pages 212–218, February 1990.

[87] A. Journel and C. Deutsch. Entropy and spatial disorder. In Report 5, Stanford

Center for Reservoir Forecasting, Stanford, CA, May 1992.

[88] A. Journel, C. Deutsch, and A. Desbarats. Power averaging for block effective

permeability. In 56th California Regional Meeting, pages 329–334. Society of

Petroleum Engineers, April 1986.

BIBLIOGRAPHY 302

[89] A. Journel and J. Gomez-Hernandez. Stochastic imaging of the Wilmington

clastic sequence. SPE paper # 19857, 1989.

[90] A. Journel and C. J. Huijbregts. Mining Geostatistics. Academic Press, New

York, NY, 1978.

[91] A. Journel and D. Posa. Characteristic behavior and order relations for indicator

variograms. Math Geology, 22(8):1011–1025, 1990.

[92] A. Journel and H. Zhu. Integrating soft seismic data: Markov-Bayes updating,

an alternative to cokriging and traditional regression. In Report 3, Stanford

Center for Reservoir Forecasting, Stanford, CA, May 1990.

[93] S. Kirkpatrick, C. Gelatt Jr., and M. Vecchi. Optimization by simulated an-

nealing. Science, 220(4598):671–680, May 1983.

[94] G. Korvin. Axiomatic characterization of the general mixture rule. Geoexplo-

ration, 19:267–276, 1981.

[95] J. Kramers, S. Bachu, D. Cuthiell, M. Prentice, and L. Yuan. A multidisci-

plinary approach to reservoir characterization: the Provost Upper Mannville B

pool. Journal of Canadian Petroleum Technology, 28(3), May-June 1989.

[96] S. Kullback. Information Theory and Statistics. Dover, New York, NY, 1968.

[97] G. Kunkel. Graphic Design with PostScript. Scott, Foreman and Company,

Glenview, IL, 1990.

[98] D. Luenberger. Optimization by Vector Space Methods. John Wiley & Sons,

New York, NY, 1969.

[99] G. Luster. Raw Materials for Portland Cement: Applications of Conditional

Simulation of Coregionalization. PhD thesis, Stanford University, Stanford,

CA, 1985.

BIBLIOGRAPHY 303

[100] A. Marechal. Kriging seismic data in presence of faults. In G. Verly et al., edi-

tors, Geostatistics for natural resources characterization, pages 271–294. Reidel,

Dordrecht, Holland, 1984.

[101] A. Marechal. Recovery estimation: A review of models and methods. In G. Verly

et al., editors, Geostatistics for natural resources characterization, pages 385–

420. Reidel, Dordrecht, Holland, 1984.

[102] B. Matern. Spatial Variation, volume 36 of Lecture Notes in Statistics. Springer

Verlag, New York, NY, second edition, 1980. First edition published by Med-

delanden fran Statens Skogsforskningsinstitut, Band 49, No. 5, 1960.

[103] C. Matthews and D. Russell. Pressure Buildup and Flow Tests in Wells. Society

of Petroleum Engineers, New York, NY, 1967.

[104] E. D. McKee, editor. A Study of Global Sand Seas. United States Government

Printing Office, Washington, 1979. Geological Survey Professional Paper 1052.

[105] N. Metropolis, A. Rosenbluth, M. Rosenbluth, A. Teller, and E. Teller. Equation

of state calculations by fast computing machines. J. Chem. Phys., 21(6):1087–

1092, June 1953.

[106] L. Morrissey and G. Livingston. Methane emissions from Alaskan tundra: an

assessment of local variability. Journal of Geophysical Research, 1992. In Press.

[107] D. Oliver. The averaging process in permeability estimation from well test data.

Formation Evaluation, pages 319–324, September 1990.

[108] D. Oliver. Estimation of radial permeability distribution from well test data. In

SPE Annual Conference and Exhibition, New Orleans, LA, number 20555, pages

243–250, New Orleans, LA, September 1990. Society of Petroleum Engineers.

[109] H. Omre, K. Solna, and H. Tjelmeland. Calcite cementation description and

production consequences. SPE paper # 20607, 1990.

BIBLIOGRAPHY 304

[110] W. Press, B. Flannery, S. Teukolsky, and W. Vetterling. Numerical Recipes.

Cambridge University Press, New York, NY, 1986.

[111] H. Ramey Jr. Advances in practical well test analysis. In 65th Annual Techni-

cal Conference and Exhibition, pages 665–676. Society of Petroleum Engineers,

September 1990.

[112] B. Ripley. Spatial Statistics. John Wiley & Sons, New York, NY, 1981.

[113] B. Ripley. Statistical Inference for Spatial Processes. Cambridge University

Press, New York, NY, 1988.

[114] A. Rosa. Reservoir Description by Well Test Analysis Using Cyclic Flow Rate

Variations. PhD thesis, Stanford University, Stanford, CA, 1991.

[115] M. Rosenblatt. Remarks on a multivariate transformation. Annals of Mathe-

matical Statistics, 23(3):470–472, 1952.

[116] M. Rossi and D. Posa. Measuring departure from gaussian assumptions in

spatial processes. In Proceedings of the 23rd International APCOM Symposium,

pages 189–197, Tucson, AZ, April 1992. Society of Mining Engineers.

[117] D. Rothman. Nonlinear inversion, statistical mechanics, and residual statics

estimation. Geophysics, 50:2784–2796, 1985.

[118] P. Scholle and D. Spearing, editors. Sandstone Depositional Environments. The

American Association of Petroleum Geologists, Tulsa, Oklahoma, 1982.

[119] C. Shannon. A mathematical theory of communication. Bell System Technical

Journal, 27:379–623, 1948.

[120] J. Skilling and S. Gull. Algorithms and applications. In C. Smith and W. Gandy

Jr., editors, Maximum Entropy and Bayesian Methods in Inverse Problems,

pages 93–132. Reidel, Dordrecht, Holland, 1985.

[121] A. Soares. Geostatistical estimation of orebody geometry: Morphological krig-

ing. Math Geology, 22(7):787–802, 1990.

BIBLIOGRAPHY 305

[122] A. Solow. Kriging Under a Mixture Model. PhD thesis, Stanford University,

Stanford, CA, 1985.

[123] SPE. Society of Petroleum Engineers Publications Style Guide. SPE, P.O. Box

833836, Richardson, TX 75083-3836, 1987.

[124] R. Srivastava. Minimum variance or maximum profitability? CIM Bulletin,

80(901):63–68, 1987.

[125] R. Srivastava. INDSIM2D: An FSS International Training Tool. FSS Interna-

tional, Vancouver, Canada, 1990.

[126] D. Stoyan, W. Kendall, and J. Mecke. Stochastic Geometry and its Applications.

John Wiley & Sons, New York, NY, 1987.

[127] V. Suro-Perez. Generation of a turbiditic reservoir: the Boolean alternative. In

Report 4, Stanford, CA, May 1991. Stanford Center for Reservoir Forecasting.

[128] V. Suro-Perez. Indicator Kriging and Simulation Based on Principal Component

Analysis. PhD thesis, Stanford University, Stanford, CA, 1992.

[129] V. Suro-Perez and A. Journel. Indicator principal component kriging. Math

Geology, 23(5):759–788, 1991.

[130] P. Switzer. Some spatial statistics for the interpretation of satellite data. Bull.

Int. Statist. Inst., 50(2):962–967, 1983.

[131] D. Tetzlaff and J. Harbaugh. Simulating Clastic Sedimentation. Reinhold Van

Nostrand, New York, NY, 1989.

[132] H. van Poollen. A hard look at radius of drainage and stabilization-time equa-

tions. Oil and Gas Journal, pages 139–147, September 1964.

[133] V. Cerny. Thermodynamical approach to the travelling salesman problem: an

efficient simulation algorithm. Journal of Optimization Theory and Applica-

tions, 45:41–51, 1985.

BIBLIOGRAPHY 306

[134] G. Verly. Estimation of Spatial Point and Block Distributions: The MultiGaus-

sian Model. PhD thesis, Stanford University, Stanford, CA, 1984.

[135] K. Weber. Influence of common sedimentary structures on fluid flow in reservoir

models. J. of. Pet. Tech., pages 665–672, March 1982.

[136] K. Weber. How heterogeneity affects oil recovery. In L. Lake and H. Caroll,

editors, Reservoir Characterization, pages 487–584. Academic Press, 1986.

[137] D. Wolf. Geostatistics for mapping and risk analysis. In Report 4, Stanford,

CA, May 1991. Stanford Center for Reservoir Forecasting.

[138] H. Xiao. A description of the behavior of indicator variograms for a bivariate

normal distribution. Master’s thesis, Stanford University, Stanford, CA, 1985.

[139] A. Yaglom and M. Pinsker. Random processes with stationary increments of

order n. volume 90, pages 731–734. Dokl. Acad. Nauk., USSR, 1953.

[140] H. Zhu. Modeling Mixture of Spatial Distributions with Integration of Soft Data.

PhD thesis, Stanford University, Stanford, CA, 1991.

[141] H. Zhu and A. Journel. Mixture of populations. Math Geology, 23(4):647–671,

1991.