Model-based Synthesis. Tony O Hagan



Similar documents
Statistics Graduate Courses

PS 271B: Quantitative Methods II. Lecture Notes

Monte Carlo and Empirical Methods for Stochastic Inference (MASM11/FMS091)

Bayesian Statistical Analysis in Medical Research

11. Time series and dynamic linear models

Bayesian Machine Learning (ML): Modeling And Inference in Big Data. Zhuhua Cai Google, Rice University

Tutorial on Markov Chain Monte Carlo

Bayesian Statistics: Indian Buffet Process

Chapter 11 Monte Carlo Simulation

Monte Carlo-based statistical methods (MASM11/FMS091)

An Introduction to Using WinBUGS for Cost-Effectiveness Analyses in Health Economics

Latent variable and deep modeling with Gaussian processes; application to system identification. Andreas Damianou

Model Calibration with Open Source Software: R and Friends. Dr. Heiko Frings Mathematical Risk Consulting

Markov Chain Monte Carlo Simulation Made Simple

Gaussian Processes to Speed up Hamiltonian Monte Carlo

The Basics of Graphical Models

Probabilistic Models for Big Data. Alex Davies and Roger Frigola University of Cambridge 13th February 2014

Introduction to Markov Chain Monte Carlo

TEACHING SIMULATION WITH SPREADSHEETS

BayesX - Software for Bayesian Inference in Structured Additive Regression

CHAPTER 2 Estimating Probabilities

Introduction to Mobile Robotics Bayes Filter Particle Filter and Monte Carlo Localization

One-year reserve risk including a tail factor : closed formula and bootstrap approaches

Chapter 14 Managing Operational Risks with Bayesian Networks

Bayesian networks - Time-series models - Apache Spark & Scala

Towards running complex models on big data

SENSITIVITY ANALYSIS AND INFERENCE. Lecture 12

A Statistical Framework for Operational Infrasound Monitoring

How To Solve A Sequential Mca Problem

Copyright. Network and Protocol Simulation. What is simulation? What is simulation? What is simulation? What is simulation?

Handling attrition and non-response in longitudinal data

How To Solve A Save Problem With Time T

Current Standard: Mathematical Concepts and Applications Shape, Space, and Measurement- Primary

Inference on Phase-type Models via MCMC

Non Linear Dependence Structures: a Copula Opinion Approach in Portfolio Optimization

Earth Science & Environmental Science SOL

Applications of R Software in Bayesian Data Analysis

5 Directed acyclic graphs

Bayesian Statistics in One Hour. Patrick Lam

Bayesian Approaches to Handling Missing Data

Lecture 3: Linear methods for classification

Fairfield Public Schools

Data Modeling & Analysis Techniques. Probability & Statistics. Manfred Huber

Java Modules for Time Series Analysis

A Bayesian hierarchical surrogate outcome model for multiple sclerosis

Simulation and Risk Analysis

Monte Carlo-based statistical methods (MASM11/FMS091)

Program description for the Master s Degree Program in Mathematics and Finance

MCMC-Based Assessment of the Error Characteristics of a Surface-Based Combined Radar - Passive Microwave Cloud Property Retrieval

Evaluation of Machine Learning Techniques for Green Energy Prediction

Problem of Missing Data

Introduction to Engineering System Dynamics

Web-based Supplementary Materials for Bayesian Effect Estimation. Accounting for Adjustment Uncertainty by Chi Wang, Giovanni

Generalized linear models and software for network meta-analysis

Summary of Formulas and Concepts. Descriptive Statistics (Ch. 1-4)

CHAPTER 3 EXAMPLES: REGRESSION AND PATH ANALYSIS

Learning outcomes. Knowledge and understanding. Competence and skills

INTRUSION PREVENTION AND EXPERT SYSTEMS

STA 4273H: Statistical Machine Learning

HT2015: SC4 Statistical Data Mining and Machine Learning

Scheduling Home Health Care with Separating Benders Cuts in Decision Diagrams

Uncertainty of Power Production Predictions of Stationary Wind Farm Models

Statistics & Probability PhD Research. 15th November 2014

From CFD to computational finance (and back again?)

Stock Investing Using HUGIN Software

NEW YORK STATE TEACHER CERTIFICATION EXAMINATIONS

Better decision making under uncertain conditions using Monte Carlo Simulation

Dynamic Simulation and Supply Chain Management

Part III: Machine Learning. CS 188: Artificial Intelligence. Machine Learning This Set of Slides. Parameter Estimation. Estimation: Smoothing

Institute of Actuaries of India Subject CT3 Probability and Mathematical Statistics

Interactive simulation of an ash cloud of the volcano Grímsvötn

INDIRECT INFERENCE (prepared for: The New Palgrave Dictionary of Economics, Second Edition)

VI. Real Business Cycles Models

Basics of Statistical Machine Learning

COMBINING THE METHODS OF FORECASTING AND DECISION-MAKING TO OPTIMISE THE FINANCIAL PERFORMANCE OF SMALL ENTERPRISES

Two Topics in Parametric Integration Applied to Stochastic Simulation in Industrial Engineering

A Primer on Mathematical Statistics and Univariate Distributions; The Normal Distribution; The GLM with the Normal Distribution

Bayesian Analysis for the Social Sciences

An explicit link between Gaussian fields and Gaussian Markov random fields; the stochastic partial differential equation approach

From Control Loops to Software

South Carolina College- and Career-Ready (SCCCR) Probability and Statistics

More details on the inputs, functionality, and output can be found below.

Profit Forecast Model Using Monte Carlo Simulation in Excel

Deterministic Sampling-based Switching Kalman Filtering for Vehicle Tracking

Objections to Bayesian statistics

Neural Networks for Machine Learning. Lecture 13a The ups and downs of backpropagation

STRATEGIC CAPACITY PLANNING USING STOCK CONTROL MODEL

Bootstrapping Big Data

Transcription:

Model-based Synthesis Tony O Hagan

Stochastic models Synthesising evidence through a statistical model 2 Evidence Synthesis (Session 3), Helsinki, 28/10/11

Graphical modelling The kinds of models that we were creating in the metaanalysis section of the previous talk can all be considered in the broad framework of statistical modelling Many people find it useful to show the logic of such models graphically τ ε 1 θ ε 2 β x 1 x 2 A model like this is called a DAG (Direct Acyclic Graph) Some powerful computational algorithms operate on DAGs To compute posterior distributions for nodes of interest 3 Evidence Synthesis (Session 3), Helsinki, 28/10/11

DAGs A DAG is a graph with nodes representing different quantities Parameters, data, random effects And with arcs joining selected nodes Rules Showing that there is a link between them All arcs are directed, i.e. have an arrow showing direction Quantity at start node (parent) influences the one at the end node (child) τ A distribution is required for each node Conditional on all its parents ε 1 ε 2 θ Although some nodes may be deterministic No cycles are allowed x 1 x 2 A cycle is a sequence of arcs pointing round in a loop β 4 Evidence Synthesis (Session 3), Helsinki, 28/10/11

DAGs (contd.) Absence of an arc is as important as presence It says that these two nodes are conditionally independent given all their parents DAGs are sometimes called influence diagrams They may also be called expert systems Particularly a large graph relating to a specific complex task And where much of the structure (arcs and conditional distributions) comes from expert knowledge Rather than directly from data 5 Evidence Synthesis (Session 3), Helsinki, 28/10/11

A larger DAG (for protein biosynthesis) 6 Evidence Synthesis (Session 3), Helsinki, 28/10/11

Learning in a DAG The DAG implicitly specifies a large joint distribution across all the nodes If we learn the values of one or more nodes in the DAG, we can update the system to condition on these The result is a posterior joint distribution on the remaining nodes From which we can obtain marginal distributions of nodes of interest Bayes theorem is the simplest of all cases Prior and likelihood specify joint distribution for θ and x Bayes theorem conditions on observed x To give posterior distribution of θ given x θ x ε 7 Evidence Synthesis (Session 3), Helsinki, 28/10/11

Computation I mentioned that we have nice computational tools to handle this kind of synthesis Markov chain Monte Carlo Powerful tool for computing posterior distributions from general Bayesian models By drawing a large sample from the joint posterior distribution Model must be representable as a DAG Software: WinBUGS Message-passing algorithms Used in large expert system DAGs Very powerful when all nodes are discrete Methods are more patchy and complex for continuous nodes 8 Evidence Synthesis (Session 3), Helsinki, 28/10/11

Mechanistic models Synthesising evidence through a complex simulator 9 Evidence Synthesis (Session 3), Helsinki, 28/10/11

Health economic modelling In order to make decisions on the cost-effectiveness of medical treatments, health economists build models Evidence for effectiveness may come from clinical trials Possibly involving meta-analysis But with even bigger data gaps Trials look at short-term effects, with selected patient groups Health care providers need to look at long-term benefits for the general population Effectiveness needs to be on a common scale of health benefit Quality-adjusted life years Based on a lot of other evidence about how people value health And we need evidence about costs Unit costs and resource usage 10 Evidence Synthesis (Session 3), Helsinki, 28/10/11

Simulators Health economic models are instances of a wider phenomenon In almost all fields of science, technology, industry and policy making, people use mechanistic models to simulate complex real-world processes For understanding, prediction, control Usually implemented in computer codes Often very computationally intensive We ll call them simulators There is a growing realisation of the importance of uncertainty in simulator predictions Can we trust them? Without any quantification of output uncertainty, it s easy to dismiss them Slide 11

Examples Climate prediction Molecular dynamics Nuclear waste disposal Oil fields Engineering design Hydrology Slide 12

Simulators and synthesis Simulators synthesise a lot of different kinds of evidence Scientific understanding about how the outputs are driven by the inputs Often including empirical relationships derived from experimental evidence Knowledge about the inputs Obtained from a variety of sources Data (usually with gaps), expert judgements They are used to make predictions about the outputs Based on synthesis of this evidence But we need to be careful about all the sources of uncertainty 13 Evidence Synthesis (Session 3), Helsinki, 28/10/11

Sources of uncertainty A simulator takes inputs x and produces outputs y = f(x) How might y differ from the true real-world value z that the simulator is supposed to predict? Error in inputs x Initial values, forcing inputs, model parameters Error in model structure or solution Wrong, inaccurate or incomplete science Bugs, solution errors Model discrepancy a kind of gap Slide 14

Quantifying uncertainty The ideal is to provide a probability distribution p(z) for the true real-world value The centre of the distribution is a best estimate Its spread shows how much uncertainty about z is induced by uncertainties on the last slide How do we get this? Input uncertainty: characterise p(x), propagate through to p(y) For example, use Monte Carlo sampling Generate random sample of x values from p(x), run the model for each to get a random sample from p(y) Structural uncertainty: characterise p(z-y) Slide 15

Reducing uncertainty To reduce uncertainty, get more information! Informal more/better science Tighten p(x) through improved understanding Tighten p(z-y) through improved modelling or programming Formal using real-world data Calibration learn about model parameters Data assimilation learn about the state variables In a dynamic simulator Learn about structural error z-y Validation Slide 16

So far, so good In principle, all this is straightforward In practice, there are many technical difficulties Formulating uncertainty on inputs Elicitation of expert judgements Propagating input uncertainty Modelling structural error Anything involving observational data! The last two are intricately linked And computation Slide 17

The problem of big models Key tasks require us to run the simulator many times Uncertainty propagation Implicitly, we need to run f(x) at all possible x Monte Carlo works by taking a sample of x from p(x) Typically needs thousands of simulator runs Calibration Learn about uncertain inputs from observations of the real process Traditionally this is done by searching the x space for good fits to the data These tasks are impractical if the simulator takes more than a few seconds to run We need a more efficient technique Slide 18

Gaussian process representation More efficient approach First work in early 1980s Consider the simulator as an unknown function f(.) becomes a random process We represent it as a Gaussian process (GP) Conditional on hyperparameters Or its Bayes linear analogue Training runs Run simulator for sample of x values Condition GP on observed data Typically requires many fewer runs than MC And x values don t need to be chosen randomly Slide 19

Emulation Analysis is completed by prior distributions for, and posterior estimation of, hyperparameters The posterior distribution is known as an emulator of the simulator Posterior mean estimates what the simulator would produce for any untried x (prediction) With uncertainty about that prediction given by posterior variance Correctly reproduces training data Slide 20

2 code runs Consider one input and one output Emulator estimate interpolates data Emulator uncertainty grows between data points 10 dat2 5 0 0 1 2 3 x 4 5 6 Slide 21

3 code runs Adding another point changes estimate and reduces uncertainty 10 dat3 5 0 0 1 2 3 x 4 5 6 Slide 22

5 code runs And so on dat5 9 8 7 6 5 4 3 2 1 0 0 1 2 3 x 4 5 6 Slide 23

Then what? Given enough training data points we can emulate any simulator accurately So that posterior variance is small everywhere Typically, this can be done with orders of magnitude fewer simulator runs than traditional methods Use the emulator to make inference about other things of interest E.g. uncertainty analysis, calibration Conceptually very straightforward in the Bayesian framework But of course can be computationally hard Slide 24

Example: UK carbon flux in 2000 Vegetation simulator predicts carbon exchange from each of 700 pixels over England & Wales Principal output is Net Biosphere Production Sheffield Dynamic Global Vegetation Model (SDGVM) Accounting for uncertainty in inputs Soil properties Properties of different types of vegetation Propagate input uncertainty through the model Aggregated to England & Wales total Allowing for correlations Estimate 7.61 Mt C Std deviation 0.61 Mt C Slide 25

Maps Slide 26

Sensitivity analysis Map shows proportion of overall uncertainty in each pixel that is due to uncertainty in the vegetation parameters As opposed to soil parameters Contribution of vegetation uncertainty is largest in grasslands/moorlands Slide 27

England & Wales aggregate PFT Plug-in estimate (Mt C) Mean (Mt C) Variance (Mt C 2 ) Grass 5.28 4.65 0.323 Crop 0.85 0.50 0.038 Deciduous 2.13 1.69 0.009 Evergreen 0.80 0.78 0.001 Covariances 0.001 Total 9.06 7.61 0.372 Slide 28

Role of emulation Gaussian process emulation was crucial to the feasibility of this exercise Almost 3000 simulator runs for a single set of inputs Imagine this repeated hundreds or thousands of times for Monte Carlo And all that repeated to evaluate the sensitivity to each input group We emulated each PFT at a sample of 33 sites Typically 200 simulator runs for each Kriging to interpolate between sites Also equivalent to Gaussian process emulation Kennedy, M. C. et al (2008). Quantifying uncertainty in the biospheric carbon flux for England and Wales. Journal of the Royal Statistical Society A 171, 109-135. 29

Example: Nuclear accident Radiation was released after an accident at the Tomsk-7 chemical plant in 1993 Data comprise measurements of the deposition of ruthenium 106 at 695 locations obtained by aerial survey after the release The simulator is a simple Gaussian plume model for atmospheric dispersion Two calibration parameters Total release of 106 Ru (source term) Deposition velocity Slide 30

Data Slide 31

Calibration A small sample (N=10 to 25) of the 695 data points was used to calibrate the model Then the remaining observations were predicted and RMS prediction error of log-deposition computed Sample size N 10 15 20 25 Best fit calibration 0.82 0.79 0.76 0.66 Bayesian calibration 0.49 0.41 0.37 0.38 On a log scale, error of 0.7 corresponds to a factor of 2 Slide 32

Role of emulation The simulator in this case was fast No need to emulate it But emulation is effectively used for the model discrepancy Gaussian process representation Trained on N = 10, 15, 20, 25 data points Kennedy, M. C. and O'Hagan, A. (2001). Bayesian calibration of computer models (with discussion). Journal of the Royal Statistical Society B 63, 425-464. 33

Summary of session 3 Synthesis of many different kinds of evidence is usually conducted by building models Statistical models DAGs Nodes linked through conditional probability distributions Deterministic nodes allowed but usually involve simple equations Computation through MCMC or message-passing algorithms Mechanistic models Complex equations to model complex real-world processes Differential equations often involved Uncertainty on inputs and on model discrepancy Computation facilitated through emulation 34 Evidence Synthesis (Session 3), Helsinki, 28/10/11