+ All Categories
Home > Documents > Emulation, Elicitation and Calibration

Emulation, Elicitation and Calibration

Date post: 20-Feb-2016
Category:
Upload: svein
View: 39 times
Download: 1 times
Share this document with a friend
Description:
Emulation, Elicitation and Calibration. UQ12 Minitutorial Presented by: Tony O’Hagan, Peter Challenor , Ian Vernon. Outline of the minitutorial. Three sessions of about 2 hours each Session 1: Monday, 2pm – 4pm, State C - PowerPoint PPT Presentation
67
Emulation, Elicitation and Calibration UQ12 Minitutorial Presented by: Tony O’Hagan, Peter Challenor, Ian Vernon UQ12 minitutorial - session 1 1
Transcript
Page 1: Emulation, Elicitation and Calibration

Emulation, Elicitation and Calibration

UQ12 MinitutorialPresented by: Tony O’Hagan, Peter Challenor, Ian Vernon

UQ12 minitutorial - session 1 1

Page 2: Emulation, Elicitation and Calibration

UQ12 minitutorial - session 1 2

Outline of the minitutorialThree sessions of about 2 hours each

• Session 1: Monday, 2pm – 4pm, State C• Overview of UQ ; total UQ; introduction to emulation; elicitation

• Session 2: Tuesday, 2pm – 4pm, State C• Building and using an emulator; sensitivity analysis

• Session 3: Wednesday, 2pm – 4pm, State C • Calibration and history matching; galaxy formation case study

Intended to introduce the applied maths/engineering UQ people to UQ methods developed in the statistics community

Page 3: Emulation, Elicitation and Calibration

Session 1

Introduction and elicitation

Page 4: Emulation, Elicitation and Calibration

Outline• Introduction

• UQ and Total UQ• Managing uncertainty• A brief case study• Emulators

• Elicitation• Elicitation principles• Elicitation practice

UQ12 minitutorial - session 1 4

Page 5: Emulation, Elicitation and Calibration

UQ and Total UQ

UQ12 minitutorial - session 1 5

Page 6: Emulation, Elicitation and Calibration

What is UQ?• Uncertainty quantification

• A term that seems to have been devised by engineers• Faced with uncertainty in some particular kinds of analyses

• Characterising how uncertainty about inputs to a complex computer model induces uncertainty about outputs

• Large body of work in engineering and applied maths• Uncertainty quantification

• What statisticians do!• And have always done

• In every field of application, for all kinds of analyses• In particular, statisticians have developed methods for

propagating and quantifying output uncertainty• And lots more relating to the use of complex simulation models

UQ12 minitutorial - session 1 6

Page 7: Emulation, Elicitation and Calibration

Simulators• In almost all fields of science, technology, industry and policy

making, people use mechanistic models• For understanding, prediction, control• Huge variety

• A model simulates a real-world, usually complex, phenomenon as a set of mathematical equations

• Models are usually implemented as computer programs• We will refer to a computer implementation of a model as a

simulator

UQ12 minitutorial - session 1 7

Page 8: Emulation, Elicitation and Calibration

Why worry about uncertainty?• Simulators are increasingly being used for decision-making

• Taking very seriously the implied claim that the simulator represents and predicts reality

• How accurate are model predictions?• There is growing concern about uncertainty in model outputs

• Particularly where simulator predictions are used to inform scientific debate or environmental policy

• Are their predictions robust enough for high stakes decision-making?

UQ12 minitutorial - session 1 8

Page 9: Emulation, Elicitation and Calibration

For instance …• Models for climate change produce different predictions

for the extent of global warming or other consequences• Which ones should we believe?• What error bounds should we put around these?• Are simulator differences consistent with the error bounds?

• Until we can answer such questions convincingly, why should anyone have faith in the science?

UQ12 minitutorial - session 1 9

Page 10: Emulation, Elicitation and Calibration

The simulator as a function• In order to talk about the uncertainty in model predictions we

need some simple notation

• Using computer language, a simulator takes a number of inputs and produces a number of outputs

• We can represent any output y as a functiony = f(x)

of a vector x of inputs

UQ12 minitutorial - session 1 10

Page 11: Emulation, Elicitation and Calibration

Where is the uncertainty?• How might the simulator output y = f(x) differ from the true

real-world value z that the simulator is supposed to predict?• Error in inputs x

• Initial values• Forcing inputs• Model parameters

• Error in model structure or solution• Wrong, inaccurate or incomplete science• Bugs, solution errors

UQ12 minitutorial - session 1 11

Page 12: Emulation, Elicitation and Calibration

Quantifying uncertainty

UQ12 minitutorial - session 1 12

• The ideal is to provide a probability distribution p(z) for the true real-world value• The centre of the distribution is a best estimate• Its spread shows how

much uncertainty about z is induced by uncertainties on the previous slide

• How do we get this?• Input uncertainty: characterise p(x), propagate through to p(y)• Structural uncertainty: characterise p(z–y)

Page 13: Emulation, Elicitation and Calibration

More uncertainties• It is important to recognise two more uncertainties that arise

when working with simulators

1. The act of propagating input uncertainty is imprecise• Approximations are made• Introducing additional code uncertainty

2. A key task in managing uncertainty is to use observations of the real world to tune or calibrate the model• We need to acknowledge uncertainty due to measurement

error

UQ12 minitutorial - session 1 13

Page 14: Emulation, Elicitation and Calibration

Code uncertainty – Monte Carlo• The simplest way to propagate uncertainty is Monte Carlo

• Take a large random sample of realisations from p(x)• Run the simulator at each sampled x to get a sample of outputs• This is a random sample from p(y)• E.g. sample mean estimates E(Y)

• Even with a very large sample, MC computations are not exact• Sample is an approximation of the population• Standard error of sample mean is population s.d. over root n• This is code uncertainty• MC has a built-in statistical quantification of code uncertainty

UQ12 minitutorial - session 1 14

Page 15: Emulation, Elicitation and Calibration

Code uncertainty – alternatives to MC• MC is impractical for simulators that require significant

resources, so other methods have been developed

• Polynomial chaos methods• PC expansions are always truncated• The truncation error is where the main code uncertainty lies

• Also in solving Galerkin equations

• Surrogate models (e.g. emulators)• Approximations to the true f(.)• Code uncertainty lies in the approximation error

UQ12 minitutorial - session 1 15

Page 16: Emulation, Elicitation and Calibration

How to quantify uncertainty• To quantify uncertainty in the true real world value that the

simulator is trying to predict we need the following steps• Quantify uncertainty in inputs, p(x)• Propagate to uncertainty in output, p(y)• Quantify and account for code uncertainty• Quantify and account for model discrepancy uncertainty

• Engineering/applied maths UQ apparently only deals with the second step• Ironically, this is the one step that doesn’t actually involve

quantifying uncertainty!

UQ12 minitutorial - session 1 16

Page 17: Emulation, Elicitation and Calibration

Total UQ• Here are my key demands

1. UQ for any quantity of interest must quantify all components of uncertainty

2. All UQ must be in the form of explicit, quantified probability distributions

3. All quantifications of uncertainty should be credible representations of what is, and is not, known

• None of this is easy but we should at least try• I call these aspirations the Total UQ Manifesto

UQ12 minitutorial - session 1 17

Page 18: Emulation, Elicitation and Calibration

Managing uncertainty

UQ12 minitutorial - session 1 18

Page 19: Emulation, Elicitation and Calibration

UQ is not enough• The presence of uncertainty creates several important tasks

• Engineering/applied maths UQ addresses only one of these• Managing uncertainty

• Uncertainty analysis – how much uncertainty do we have?• This is the basic UQ task

• Sensitivity analysis – which sources of uncertainty drive overall uncertainty, and how?

• Understanding the system, prioritising research• Calibration – how can we reduce uncertainty?

• Use of observations – Tuning, data assimilation, history matching, inverse problems

• Experimental design

UQ12 minitutorial - session 1 19

Page 20: Emulation, Elicitation and Calibration

• Decision-making under uncertainty – can we cope with uncertainty?

• Robust engineering design• Optimisation under uncertainty

UQ12 minitutorial - session 1 20

Page 21: Emulation, Elicitation and Calibration

MUCM• Managing Uncertainty in Complex Models

• Large 4-year UK research grant• June 2006 to September 2010• 7 postdoctoral research associates, 4 project PhD students• Objective to develop BACCO methods into a basic technology,

usable and widely applicable• MUCM2: New directions for MUCM

• Smaller 2-year grant to September 2012• Scoping and developing research proposals

UQ12 minitutorial - session 1 21

Page 22: Emulation, Elicitation and Calibration

Primary MUCM deliverables• Methodology and papers moving the technology forward

• Papers both in statistics and application area journals• The MUCM toolkit

• Documentation of the methods and how to use them• With emphasis on what is found to work reliably across a range of

modelling areas• Web-based

• Case studies• Three substantial case studies• Showcasing methods and best practice• Linked to toolkit

• Events• Workshops – conceptual and hands-on• Short courses• Conferences – UCM 2010 and UCM 2012 (July 2-4)

UQ12 minitutorial - session 1 22

Page 23: Emulation, Elicitation and Calibration

Focus on the • The toolkit is a ‘recipe book’

• The good sort that encourages you to experiment• There are recipes (procedures) but also lots of explanation of

concepts and discussion of choices• It is not a software package

• Software packages are great if they are in your favourite language

• But it probably wouldn’t be!• Packages are dangerous without basic understanding

• The purpose of the toolkit is to build that understanding• And it enables you to easily develop your own code

UQ12 minitutorial - session 1 23

Page 24: Emulation, Elicitation and Calibration

24

Resources• Introduction to emulators

• O'Hagan, A. (2006). Bayesian analysis of computer code outputs: a tutorial. Reliability Engineering and System Safety 91, 1290-1300.

• The MUCM website• http://mucm.ac.uk

• The MUCM toolkit• http://mucm.ac.uk/toolkit

• The UCM 2012 conference• http://mucm.ac.uk/UCM2012.html

UQ12 minitutorial - session 1

Page 25: Emulation, Elicitation and Calibration

This minitutorial• This minitutorial covers the key elements of Total UQ and

uncertainty management• Emulators

• Surrogate models that include quantification of code uncertainty• Brief outline in this session then details in session 2

• Elicitation• Tools for rigorous quantification of fundamental uncertainties• Introduction to this big field in this session

• Management tools• Sensitivity analysis in session 2• Calibration and history matching in session 3

UQ12 minitutorial - session 1 25

Page 26: Emulation, Elicitation and Calibration

A brief case study

Complex emulation and expert elicitation were essential components of this exercise

UQ12 minitutorial - session 1 26

Page 27: Emulation, Elicitation and Calibration

Example: UK carbon flux in 2000• Vegetation model predicts carbon exchange from each of 700

pixels over England & Wales in 2000• Principal output is Net Biosphere Production

• Accounting for uncertainty in inputs• Soil properties• Properties of different types of vegetation• Land usage

• Also code uncertainty• But not structural uncertainty

• Aggregated to England & Wales total• Allowing for correlations• Estimate 7.46 Mt C (± 0.54 Mt C)

UQ12 minitutorial - session 1 27

Page 28: Emulation, Elicitation and Calibration

Maps

UQ12 minitutorial - session 1 28

Mean NBP Standard deviation

Page 29: Emulation, Elicitation and Calibration

England & Wales aggregate

UQ12 minitutorial - session 1 29

PFTPlug-in estimate

(Mt C)Mean(Mt C)

Variance (Mt C2)

Grass 5.28 4.37 0.2453

Crop 0.85 0.43 0.0327

Deciduous 2.13 1.80 0.0221

Evergreen 0.80 0.86 0.0048

Covariances -0.0081

Total 9.06 7.46 0.2968

Page 30: Emulation, Elicitation and Calibration

Emulators

UQ12 minitutorial - session 1 30

Page 31: Emulation, Elicitation and Calibration

So far, so good, but• In principle, Total UQ is straightforward• In practice, there are many technical difficulties

• Formulating uncertainty on inputs• Elicitation of expert judgements

• Propagating input uncertainty• Modelling structural error• Anything involving observational

data!• The last two are intricately

linked• And computation

UQ12 minitutorial - session 1 31

Page 32: Emulation, Elicitation and Calibration

The problem of big models• Tasks like uncertainty propagation and calibration require us

to run the simulator many times• Uncertainty propagation

• Implicitly, we need to run f(x) at all possible x• Monte Carlo works by taking a sample of x from p(x)• Typically needs thousands of simulator runs

• Calibration• Traditionally done by searching x space for good fits to the data

• Both become impractical if the simulator takes more than a few seconds to run• 10,000 runs at 1 minute each takes a week of computer time• We need a more efficient technique

UQ12 minitutorial - session 1 32

Page 33: Emulation, Elicitation and Calibration

More efficient methods• This is what UQ theory is mostly about• Engineering/Applied Maths UQ

• Polynomial chaos expansions of random variables• Approximate by truncating

• Thereby build an expansion of outputs• Compute by Monte Carlo etc. using this surrogate

representation• Statistics UQ

• Gaussian process emulation of the simulator• A different kind of surrogate

• Propagate input uncertainty through surrogate• By Monte Carlo or analytically

UQ12 minitutorial - session 1 33

Page 34: Emulation, Elicitation and Calibration

Gaussian process representation• More efficient approach

• First work in early 1980s (DACE)• Represent the code as an unknown function

• f(.) becomes a random process• We generally represent it as a Gaussian process (GP)

• Or its second-order moment version• Training runs

• Run simulator for sample of x values• Condition GP on observed data• Typically requires many fewer runs than Monte Carlo

• And x values don’t need to be chosen randomly

UQ12 minitutorial - session 1 34

Page 35: Emulation, Elicitation and Calibration

Emulation• Analysis is completed by prior distributions for, and posterior

estimation of, hyperparameters

• The posterior distribution is known as an emulator of the computer simulator• Posterior mean estimates what the simulator would produce for

any untried x (prediction)• With uncertainty about that prediction given by posterior

variance• Correctly reproduces training data

• Gets its UQ right!• An essential requirement of credible quantification

UQ12 minitutorial - session 1 35

Page 36: Emulation, Elicitation and Calibration

2 code runs

UQ12 minitutorial - session 1 36

• Consider one input and one output• Emulator estimate interpolates data• Emulator uncertainty grows between data points

Page 37: Emulation, Elicitation and Calibration

3 code runs

UQ12 minitutorial - session 1 37

• Adding another point changes estimate and reduces uncertainty

Page 38: Emulation, Elicitation and Calibration

5 code runs

UQ12 minitutorial - session 1 38

• And so on

Page 39: Emulation, Elicitation and Calibration

Then what?• Given enough training data points we can in principle emulate

any simulator output accurately• So that posterior variance is small “everywhere”• Typically, this can be done with orders of magnitude fewer

model runs than traditional methods• At least in relatively low-dimensional problems

• Use the emulator to make inference about other things of interest• E.g. uncertainty analysis, sensitivity analysis, calibration

• The key feature that distinguishes an emulator from other kinds of surrogate• Code uncertainty is quantified naturally• And credibly

UQ12 minitutorial - session 1 39

Page 40: Emulation, Elicitation and Calibration

Elicitation principles

UQ12 minitutorial - session 1 40

Page 41: Emulation, Elicitation and Calibration

Where do probabilities come from?• Consider the probability distribution for a model input

• Like the hydraulic conductivity K in a geophysical model• Suppose we ask an expert, Mary

• Mary gives a probability distribution for K• We might be particularly interested in

one probability in that distribution• Like the probability that

K exceeds 10-3 (cm/sec)• Mary’s distribution says

Pr(K > 10-3) = 0.2

UQ12 minitutorial - session 1 41

Page 42: Emulation, Elicitation and Calibration

UQ12 minitutorial - session 1 42

How can K have probabilities?• Almost everyone learning probability is taught the frequency

interpretation• The probability of something is the long run relative frequency

with which it occurs in a very long sequence of repetitions• How can we have repetitions of K?

• It’s a one-off, and will only ever have one value• It’s that unique value we’re interested in

• Mary’s distribution can’t be a probability distribution in that sense

• So what do her probabilities actually mean?• And does she know?

Page 43: Emulation, Elicitation and Calibration

UQ12 minitutorial - session 1 43

Mary’s probabilities• Mary’s probability 0.3 that K > 10-3 is a judgement

• She thinks it’s more likely to be below 10-3 than above• So in principle she would bet even money on it• In fact she would bet $2 to win $1 (because 0.7 > 2/3)

• Her expectation of around 10-3.5 is a kind of best estimate• Not a long run average over many repetitions

• Her probabilities are an expression of her beliefs• They are personal judgements

• You or I would have different probabilities• We want her judgements because she’s the expert!

• We need a new definition of probability

Page 44: Emulation, Elicitation and Calibration

UQ12 minitutorial - session 1 44

Subjective probability• The probability of a proposition E is a measure of a person’s

degree of belief in the truth of E• If they are certain that E is true then Pr(E) = 1• If they are certain it is false then Pr(E) = 0• Otherwise Pr(E) lies between these two extremes

• Exercise

Page 45: Emulation, Elicitation and Calibration

UQ12 minitutorial - session 1 45

Subjective includes frequency• The frequency and subjective definitions of probability are

compatible• If the results of a very long sequence of repetitions are

available, they agree• Frequency probability equates to the long run frequency• All observers who accept the sequence as comprising

repetitions will assign that frequency as their (personal or subjective) probability for the next result in the sequence

• Subjective probability extends frequency probability• But also seamlessly covers propositions that are not repeatable• It’s also more controversial

Page 46: Emulation, Elicitation and Calibration

UQ12 minitutorial - session 1 46

It doesn’t include prejudice etc!• The word “subjective” has derogatory overtones

• Subjectivity should not admit prejudice, bias, superstition, wishful thinking, sloppy thinking, manipulation ...

• Subjective probabilities are judgements but they should be careful, honest, informed judgements• As “objective” as possible without ducking the issue• Using best practice

• Formal elicitation methods• Bayesian analysis

• Probability judgements go along with all the other judgements that a scientist necessarily makes• And should be argued for in the same careful, honest and

informed way

Page 47: Emulation, Elicitation and Calibration

UQ12 minitutorial - session 1 47

But people are poor probability judges• Our brains evolved to make quick decisions

• Heuristics are short-cut reasoning techniques• Allow us to make good judgements quickly in familiar situations

• Judgement of probability is not something that we evolved to do well

• The old heuristics now produce biases• Anchoring and adjustment• Availability • Representativeness• The range-frequency compromise• Overconfidence

Page 48: Emulation, Elicitation and Calibration

UQ12 minitutorial - session 1 48

Anchoring and adjustment• When asked to make two related judgements, the second is

affected by the first• The second is judged relative to the first• By adjustment away from the first judgement• The first is called the anchor

• Adjustment is typically inadequate• Second response too close to the first (anchor)• Anchoring can be strong even when obviously

not really relevant to the second question

Page 49: Emulation, Elicitation and Calibration

UQ12 minitutorial - session 1 49

• The probability of an event is judged more likely if we can quickly bring to mind instances of it• Things that are more memorable are deemed more probable

• High profile train accidents in the UK lead people to imagine rail travel is more risky than it really is

• My judgement of the risk of dying from a particular disease will be increased if I know (of) people who have the disease or have died from it

Availability

Page 50: Emulation, Elicitation and Calibration

UQ12 minitutorial - session 1 50

Representativeness• An event is considered more probable if the components of its

description fit together• Even when the juxtaposition of many components is

actually improbable• “Linda is 31, single, outspoken and very bright. She studied

philosophy at university and was deeply concerned with issues of discrimination and social justice. Is Linda …

• “A bank teller?• “A bank teller and active in the feminist movement?”

• The second is often judged more probable than the first• We are a story-telling species• This is also called the conjunction fallacy

Page 51: Emulation, Elicitation and Calibration

UQ12 minitutorial - session 1 51

??

Range-frequency compromise• Probability judgements are affected by how many alternatives

are presented• An example with a quantitative variable• Elicitation for X = cost of building project (in £k)• Ask for probabilities for ranges

• (0, 10), (10, 20 ), (20, 100)• (0, 10), (10, 20 ), (20, 50), (50, 100)

• Probabilities in first 2 ranges smaller in second case• Even if (20,100) is really unlikely

• We tend to spread probabilities evenly between whatever options we are given• May be a kind of anchoring

Page 52: Emulation, Elicitation and Calibration

UQ12 minitutorial - session 1 52

Overconfidence• It is generally said that experts are overconfident

• When asked to give 95% interval, say, then far fewer than 95% contain the true value

• Several possible explanations• Wish to demonstrate expertise• Anchoring to a central estimate• Difficulty of judging extreme events• Not thinking ‘outside the box’

• Expertise often consists of specialist heuristics• Situations we elicit judgements on are not typical

Page 53: Emulation, Elicitation and Calibration

UQ12 minitutorial - session 1 53

• Probably over-stated as a general phenomenon• Experts can be under-confident if afraid of consequences

• A matter of personality and feeling of security• Some evidence that people are not over-confident if asked for

intervals of moderate probability• E.g. 66% or 50%

• Evidence of over-confidence is not from real experts making judgements on serious questions

• Students and ‘almanac’ questions• Good elicitation practice needs to recognise these problems

• Answers depend on how the questions are posed• Protocol should avoid or minimise biases

Page 54: Emulation, Elicitation and Calibration

UQ12 minitutorial - session 1 54

Elicitation practice

Page 55: Emulation, Elicitation and Calibration

UQ12 minitutorial - session 1 55

Why elicit distributions?• Occasionally, we want expert opinion about a discrete

proposition• The Democrats will win the next US presidential election• There is, or has at one time been, life on Mars• Then a single probability needs to be elicited

• Mostly, though, we are interested in opinion about an uncertain quantity• The mean response of patients to a new drug• The increase in global temperature caused by a doubling of

atmospheric CO2

• Then we need to elicit a probability distribution• In fact, we are often interested in several quantities

Page 56: Emulation, Elicitation and Calibration

UQ12 minitutorial - session 1 56

Too many probabilities!• We’ll stick to a single quantity for now• One way to think of a distribution for a quantity X is as a set of

probabilities• Pr(X < x) for all possible x values• That’s a lot of probabilities to elicit!

• If we sat down to elicit them one by one, the interrogation would never finish!• And we’d have serious anchoring

problems!• We need a pragmatic approach

Page 57: Emulation, Elicitation and Calibration

UQ12 minitutorial - session 1 57

A pragmatic approach• Animal welfare – what proportion of a herd is diseased?

• X = incidence/1000 of a parasite• Expert says Pr(X < 10) = 0.4, Pr(X > 30) = 0.2

• Facilitator fits an inverse gamma distribution to the two given probabilities

• Check expert agrees Pr(X < 20) 0.68

• The usual approach has two steps1. Elicit a few probabilities or other ‘summaries’2. Fit a distribution to those summaries

0.20.40.4

0 10 30

Page 58: Emulation, Elicitation and Calibration

UQ12 minitutorial - session 1 58

Elicit a few summaries• We can just elicit a few probabilities

• As in the last example• Other possible ‘summaries’:

• Mean, median mode• Often expert is just asked for

an ‘estimate’• But that begs the question of

what kind of estimate• The mean and mode are not

recommended• Asking for the median is OK

(value with probability 0.5 on either side)

Mode = 6 Median = 13Mean = 30

Page 59: Emulation, Elicitation and Calibration

UQ12 minitutorial - session 1 59

• More summaries• Not variances!• Fixed probability intervals

• But not for extreme probability values• Shape

• E.g. Unimodal – or bimodal!!• Skewed

• Shape is important• Often overlooked• Should at least be checked by feedback (showing the fitted

distribution)

Page 60: Emulation, Elicitation and Calibration

UQ12 minitutorial - session 1 60

Then fit a distribution• Any convenient distribution

• As long as it fits the elicited summaries adequately• At this point, the choice should not matter

• The idea is that we have elicited enough• Any reasonable choice of distribution will be similar to any other

• Elicitation can never be exact• The elicited summaries are only approximate anyway

• If the choice does matter• i.e. different fitted distributions give different answers to the

problem for which we are doing the elicitation • then we need to elicit more summaries

Page 61: Emulation, Elicitation and Calibration

UQ12 minitutorial - session 1 61

The SHELF system• SHELF is a package of documents and simple software to aid

elicitation• General advice on conducting the elicitation• Templates for recording the elicitation

• Suitable for several different basic methods• Annotated versions of the templates with detailed guidance• Some R functions for fitting distributions and providing feedback

• SHELF is freely available and we welcome comments and suggestions for additions• Developed by Jeremy Oakley and myself

• R functions by Jeremy• http://tonyohagan.co.uk/shelf

Page 62: Emulation, Elicitation and Calibration

UQ12 minitutorial - session 1 62

Contents• SHELF Overview_v2.0• SHELF Pre-elicitation Briefing• SHELF Pre-elicitation Form + version with notes• SHELF 1 (Context) + version with notes• SHELF 2 (Distribution) Q + version with notes• SHELF 2 (Distribution) QP + version with notes• SHELF 2 (Distribution) R + version with notes• SHELF 2 (Distribution) RP + version with notes• SHELF 2 (Distribution) T + version with notes• SHELF 2 (Distribution) TP + version with notes• SHELF2 Distribution fitting instructions• shelf2.R

Page 63: Emulation, Elicitation and Calibration

UQ12 minitutorial - session 1 63

Page 64: Emulation, Elicitation and Calibration

UQ12 minitutorial - session 1 64

Page 65: Emulation, Elicitation and Calibration

UQ12 minitutorial - session 1 65

Example• We will now look at a hypothetical example of using SHELF

• Using a simplified form for a single expert

Page 66: Emulation, Elicitation and Calibration

Where to next• Two more sessions to come in this minitutorial• Tomorrow:

• Peter Challenor will present Session 2 on building and using emulators

• Wednesday• I will begin Session 3 on using observations of the real world

process – calibration and other tasks• Ian Vernon will finish with a case study on a grand scale –

history matching a galaxy formation model to observations of the universe

UQ12 minitutorial - session 1 66

Page 67: Emulation, Elicitation and Calibration

Another conference

• UCM 2012

• Still open for posterabstracts

• Early bird registrationdeadline 30th April

• http://mucm.ac.uk/ucm2012

67UQ12 minitutorial - session 1


Recommended