An Introduction to Using WinBUGS for Cost-Effectiveness Analyses in Health Economics
|
|
|
- Emmeline Curtis
- 10 years ago
- Views:
Transcription
1 Slide 1 An Introduction to Using WinBUGS for Cost-Effectiveness Analyses in Health Economics Dr. Christian Asseburg Centre for Health Economics Part 1
2 Slide 2 Talk overview Foundations of Bayesian statistics Comparison between Frequentist and Bayesian approaches Calculations and computer implementations Example from health economics Questions and discussion
3 Slide 3 History of Bayesian Statistics The reverend Thomas Bayes ( ) proved a special case of what is now known as Bayes' Theorem. Pierre-Simon Laplace ( ) proved a more general version of Bayes' Theorem and used it for various applications. Revd. Thomas Bayes The relevance of Bayes' Theorem to statistics, however, was not appreciated until the 20th century. The Frequentist paradigm has been the mainstay of probability theory during the 19th and 20th century, with important contributions by e.g. Jerzy Neyman, Egon Pearson, John Venn, R.A. Fisher, and Richard von Mises. Frequentist tools such as hypothesis testing and confidence intervals have allowed many advances in statistics. Bayesian equivalents exist, but they often require more computations it was during the last two decades of increasing availability of computing resources that Bayesian statistics gained ground.
4 Slide 4 Bayes' Theorem Bayes' Theorem can be derived easily from the expression of the joint probability of two events A and B: Let p(a) denote the probability that event A will occur, let p(b) denote the probability that event B will occur, and let p(a,b) denote the probability that both of the events occur. Then p A, B = p A p B A = p B p A B Bayes' Theorem states simply that p B p A B p B A = p A
5 Slide 5 Priors and Posteriors (1) Of course, Bayes' Theorem as a way to relate the conditional probabilities of two events is valid both in Frequentist as well as in Bayesian statistics. However, in Bayesian statistics it is also applied to unknown parameters x directly: p B p A B p B A = p A p x p data x p x data = p data
6 Slide 6 Priors and Posteriors (2) Unknown parameter(s): x Data (known): data Probability of data given x: p(data x) Prior probability of x: p(x) Posterior probability of x: p(x data) p x p data x p x data = p data
7 Slide 7 Priors and Posteriors (2) Unknown parameter(s): x Data (known): data Probability of data given x: p(data x) Prior probability of x: p(x) Posterior probability of x: p(x data) Likelihood p x p data x p x data = p data
8 Slide 8 Priors and Posteriors (2) Unknown parameter(s): x Data (known): data Probability of data given x: p(data x) Prior probability of x: p(x) Posterior probability of x: p(x data) p x p data x p x data = p data The denominator is a constant and can usually be ignored.
9 Slide 9 Priors and Posteriors (3) Bayes' Theorem is thus used to combine data with a prior belief on an unknown quantity, resulting in a posterior belief on the unknown quantity. This approach has been compared to the task of learning in humans, where experience supports a constant updating of a person's belief system. Prior probability of x: p(x) Posterior probability of x: p(x data) p x p data x p x data = p data
10 Slide 10 Definition of Probability FREQUENTIST B AY E S I A N
11 Slide 11 Definition of Probability FREQUENTIST The probability of an event A occurring (or of a quantity taking a value in a given interval) is a frequency. Imagine many (hypothetical or actual) circumstances in which the data have been observed. The proportion of circumstances in which event A occurs (out of all circumstances) is the probability of A. This probability is objective. B AY E S I A N
12 Slide 12 Definition of Probability FREQUENTIST The probability of an event A occurring (or of a quantity taking a value in a given interval) is a frequency. Imagine many (hypothetical or actual) circumstances in which the data have been observed. The proportion of circumstances in which event A occurs (out of all circumstances) is the probability of A. This probability is objective. B AY E S I A N The probability of an event A occurring (or of a quantity taking a value in a given interval) is a degree of belief. The degree of belief in A may change when we are confronted with new data. The probability of A is a numerical representation of this degree of belief. If you and I (and everyone else) agree on the belief in event A, we define an objective probability, otherwise we define a subjective probability.
13 Slide 13 What is fixed, what is random? (1) FREQUENTIST B AY E S I A N
14 Slide 14 What is fixed, what is random? (1) FREQUENTIST There is a fixed, but unknown value for each parameter. The data are an instance of many possible data that could have been collected. A Frequentist statistician evaluates how likely the given data are according to different hypothetical values for the unknown quantities. Thus, statements about the probability of observing the data given different hypothetical parameter values are summarised in a confidence interval. B AY E S I A N
15 Slide 15 What is fixed, what is random? (1) FREQUENTIST There is a fixed, but unknown value for each parameter. The data are an instance of many possible data that could have been collected. A Frequentist statistician evaluates how likely the given data are according to different hypothetical values for the unknown quantities. Thus, statements about the probability of observing the data given different hypothetical parameter values are summarised in a confidence interval. B AY E S I A N The value for each parameter is unknown. The data are known, they have been observed. A Bayesian statistician evaluates how likely different values for the underlying quantities are, given the observed data. Thus, statements can be made about the probability of the unknown quantity taking a value in a certain credibility interval.
16 Slide 16 What is fixed, what is random? (2) FREQUENTIST A 95% confidence interval for a quantity x: B AY E S I A N
17 Slide 17 What is fixed, what is random? (2) FREQUENTIST A 95% confidence interval for a quantity x: If new data are collected many times and confidence intervals are calculated, then 95% of these confidence intervals contain the true value of x. B AY E S I A N
18 Slide 18 What is fixed, what is random? (2) FREQUENTIST A 95% confidence interval for a quantity x: If new data are collected many times and confidence intervals are calculated, then 95% of these confidence intervals contain the true value of x. B AY E S I A N A 95% credibility interval for a quantity x: The probability that the value of x lies between 2.5 and 4.7 is 95%, given the observed data and the prior belief.
19 Slide 19 Hypothesis testing FREQUENTIST Given two hypotheses, H0 and H1,... B AY E S I A N
20 Slide 20 Hypothesis testing FREQUENTIST Given two hypotheses, H0 and H1, calculate the probability of observing the data (or more extreme data) if H0 is true. If this probability is low (p value), reject H0. B AY E S I A N Given two hypotheses, H0 and H1,...
21 Slide 21 Hypothesis testing FREQUENTIST Given two hypotheses, H0 and H1, calculate the probability of observing the data (or more extreme data) if H0 is true. If this probability is low (p value), reject H0. B AY E S I A N Given two hypotheses, H0 and H1, calculate the probability of each of them, given the data and the priors. Favour the hypothesis that has the higher probability.
22 Slide 22 Hypothesis testing FREQUENTIST Given two hypotheses, H0 and H1, calculate the probability of observing the data (or more extreme data) if H0 is true. If this probability is low (p value), reject H0. Because a hypothesis is either true or false (this is just not known) and only the likelihood of observing the data is calculated, a Frequentist cannot assign a probability to each hypothesis. B AY E S I A N Given two hypotheses, H0 and H1, calculate the probability of each of them, given the data and the priors. Favour the hypothesis that has the higher probability. The probability of each of the hypotheses being true can be calculated. Relative statements (e.g. H0 is twice as likely as H1 ) can be made.
23 Slide 23 A Simple Example (1) In roulette, a spin of the wheel results in a red or a black number (or 0). In one hour, the roulette wheel resulted in 25 red and 15 black numbers. What is the probability z that this wheel gives a red number?
24 Slide 24 A Simple Example (1) In roulette, a spin of the wheel results in a red or a black number (or 0). In one hour, the roulette wheel resulted in 25 red and 15 black numbers. What is the probability z that this wheel gives a red number? FREQUENTIST The probability of observing 25 red and 15 black numbers can be described by a Binomial distribution with 25 successes out of 40. The sample proportion of success is 25/40, or Using the central limit theorem, an approximate confidence interval for a proportion can be found. The sampling distribution is summarised by its mean (0.625) and standard deviation (0.0765), and these are used to obtain a 95% confidence interval for the mean of a normal distribution. After correcting for the discrete nature of the data, the confidence interval for z is found: [0.46, 0.79].
25 Slide 25 A Simple Example (2) In roulette, a spin of the wheel results in a red or a black number (or 0). In one hour, the roulette wheel resulted in 25 red and 15 black numbers. What is the probability z that this wheel gives a red number? B AY E S I A N The probability of observing 25 red and 15 black numbers can be described by a Binomial distribution with 25 successes out of 40. The prior probability for z is assumed to be Beta(1,1). Bayes' Theorem is used to calculate the posterior probability of z. (See next slide) The 95% credibility interval for z is [0.47, 0.76].
26 Slide 26 A Simple Example (3) In roulette, a spin of the wheel results in a red or a black number (or 0). In one hour, the roulette wheel resulted in 25 red and 15 black numbers. What is the probability z that this wheel gives a red number? B AY E S I A N Bayes' theorem p z data p z p data z (The denominator of Bayes' Theorem, p(data), is a constant and can usually be ignored.) p(data z) = Binomial (25 out of 40 with prob. z) p(z) = Beta(1, 1)
27 Slide 27 A Simple Example (4) In roulette, a spin of the wheel results in a red or a black number (or 0). In one hour, the roulette wheel resulted in 25 red and 15 black numbers. What is the probability z that this wheel gives a red number? B AY E S I A N Bayes' theorem p z data p z p data z 1 40! p z data z 1 z z 25 1 z B 1,1 25! 40 25! p z data z 1 z So p(z data) = Beta(26,16), and the credibility interval can be calculated easily by looking up the cumulative probabilities.
28 Slide 28 A Simple Example (5) In roulette, a spin of the wheel results in a red or a black number (or 0). In one hour, the roulette wheel resulted in 25 red and 15 black numbers. What is the probability z that this wheel gives a red number? B AY E S I A N In this simple example, when the prior is from a particular family (Beta) and the likelihood of the data is also from a particular family (Binomial), the posterior likelihood also belongs to a particular family of distributions (Beta). The Beta prior and Binomial likelihood distribution are called conjugate. This is a special case usually the Bayesian posterior distributions cannot be calculated analytically, and numerical methods are required to approximate the posterior distribution.
29 Slide 29 A Simple Example (6) In roulette, a spin of the wheel results in a red or a black number (or 0). In one hour, the roulette wheel resulted in 25 red and 15 black numbers. What is the probability z that this wheel gives a red number? B AY E S I A N Different choices of prior distributions lead to different posterior distributions and thus to different credibility intervals. Prior Beta(1,1) Data 25 out of 40 Posterior Beta(26,16) 95% credibility interval [0.47, 0.76] Beta(50,50) 25 out of 40 Beta(75,65) [0.45, 0.62] Beta(26,16) 25 out of 40 Beta(51,31) [0.52, 0.72]
30 Slide 30 A Simple Example (7) Different choices of prior distributions lead to different posterior distributions and thus to different credibility intervals. Prior Beta(1,1) Data 25 out of 40 Posterior Beta(26,16) 95% credibility interval [0.47, 0.76]
31 Slide 31 A Simple Example (7) Different choices of prior distributions lead to different posterior distributions and thus to different credibility intervals. Prior Beta(1,1) Data 25 out of 40 Posterior Beta(26,16) 95% credibility interval [0.47, 0.76] Beta(50,50) 25 out of 40 Beta(75,65) [0.45, 0.62]
32 Slide 32 A Simple Example (7) Different choices of prior distributions lead to different posterior distributions and thus to different credibility intervals. Prior Beta(1,1) Data 25 out of 40 Posterior Beta(26,16) 95% credibility interval [0.47, 0.76] Beta(50,50) 25 out of 40 Beta(75,65) [0.45, 0.62] Beta(26,16) 25 out of 40 Beta(51,31) [0.52, 0.72]
33 Slide 33 Priors - again... Different choices of prior distributions lead to different posterior distributions and thus to different credibility intervals. Prior Beta(1,1) Data 25 out of 40 Posterior Beta(26,16) 95% credibility interval [0.47, 0.76] Beta(50,50) 25 out of 40 Beta(75,65) [0.45, 0.62] Beta(26,16) 25 out of 40 Beta(51,31) [0.52, 0.72] So how does one choose the right prior?
34 Slide 34 Controversy regarding priors There is no right prior. A good prior choice may be obvious, for example when earlier studies on a model quantity can be used. The influence of the prior on the model output can be minimised by choosing an uninformative prior or a reference prior. If different stakeholders are involved, whose prior opinions on a model quantity differ, each of them may propose a prior. The model can then be run in turn for each prior. Afterwards, it may be possible to reconcile the different posterior opinions. In general, if the prior choice makes a difference to the model's output, then more data should be collected. A good modelling application should either have an informative prior or be robust to prior choice.
35 Slide 35 Model Selection In Bayesian statistics, it is relatively straightforward to evaluate different explanations for a data-set (nested models or totally different models). The models are all evaluated simultaneously, together with additional parameters mi for the probabilities of each of the models. The posteriors for the parameters mi summarise how well each of the competing models fits the data. Depending on the model application, one most suitable model may be found, or predictions can be made from all models simultaneously, using the posterior values for mi as weights (model averaging). In Frequentist statistics, it is relatively easy to evaluate nested models but the evaluation of other competing models is not straightforward.
36 Slide 36 Summary FREQUENTIST B AY E S I A N Probability Frequency Belief Statements Probability of observing Probability of model quantity data Objectivity Result depends only on data Result depends on prior and data subjective Computation Often feasible Often complicated Flexibility Some applications reno intrinsic limitations quire normal or other simplifying assumptions Model selection Sometimes possible Straightforward
37 Slide 37 Interval any questions?... Good Bayesian text book that starts with a comparison of Bayesian and Frequentist methods: D'Agostini, G: Bayesian Reasoning in Data Analysis. World Scientific Publishing, Singapore, Why use Bayesian methods in health economics? E.g. B Luce, Y Shih, K Claxton: International Journal of Technology Assessment in Health Care 17/1, 2001, pp 1-5.
38 Slide 38 Calculating Bayesian Posteriors Models that can be solved analytically (such as the simple example before) are rare and require conjugacy. Most multi-dimensional models do not fall in this class. The problem is that, for each possible set of parameter values, Bayes' Theorem gives the posterior probability, but if the parametric form of the distribution cannot be recognised, there is no obvious method for calculating e.g. its mean value, or for sampling from it. Therefore, a Bayesian model usually requires numerical methods for calculating the posteriors of interest. Any algorithm that generates samples from a distribution that is defined by its probability density function could be used. p data p p data
39 Slide 39 Calculating Bayesian Posteriors p data p p data The most commonly used algorithms for sampling from the Bayesian posterior fall in two groups: Metropolis-Hastings: some algorithms in this class are Markov chain Monte Carlo (MCMC), e.g. Gibbs sampling or Reversible Jump. These algorithms work well when the posterior model space can be written as a product, such that factors correspond to subspaces. Sequential Importance Sampling. Very suitable for posteriors that can be written as products, such that factors correspond to individual data.
40 Slide 40 Calculating Bayesian Posteriors p data p p data The most commonly used algorithms for sampling from the Bayesian posterior fall in two groups: Metropolis-Hastings: some algorithms in this class are Markov chain Monte Carlo (MCMC), e.g. Gibbs sampling or Reversible Jump. p data p a p data a p b p data b Sequential Importance Sampling. p data p p data 1 p data 2...
41 Slide 41 Markov chain Monte Carlo (1) MCMC generates samples from the posterior space M by defining a chain C={x1, x2, x3,...} in M. At each step i in the chain, candidate values x* are generated randomly for each of the parameters. (The proposal distribution may depend on the current values, xi.) The posterior probabilities are calculated for both xi and x*. Depending on the likelihood of x* relative to xi, an acceptance probability is calculated, and the chain either moves to x* (xi+1=x*) or stays at its current value (xi+1=xi). Ergodic theory ensures that, in the limit, the distribution of the values of C converges to the posterior distribution of interest. The beginning of the chain is discarded because the initial values dominate it ( burn-in ).
42 Slide 42 Markov chain Monte Carlo (2) Illustration: At step i in the MCMC, the chain may jump to the candidate value θ* or stay at the current value θi. This depends on the posterior probabilities for these two points in parameter space.
43 Slide 43 Markov chain Monte Carlo (2) Illustration: At step i in the MCMC, the chain may jump to the candidate value θ* or stay at the current value θi. This depends on the posterior probabilities for these two points in parameter space. In the long run, the distribution of points in the chain approximates the posterior distribution.
44 Slide 44 Markov chain Monte Carlo (3) To sample from a multi-dimensional posterior (e.g. the posterior of a model with several unknown parameters), parameters can be grouped together (block sampling). Blocks are chosen such that calculations can be simplified. At each iteration, a new candidate is suggested for one block (and parameters in the other blocks retain their current value). The candidate values for that block are either accepted or rejected. Then the same is done for the next parameter block, etc. Example 1. Suggest a candidate for a. (In this example, a* is accepted.) 2. 3.
45 Slide 45 Markov chain Monte Carlo (3) To sample from a multi-dimensional posterior (e.g. the posterior of a model with several unknown parameters), parameters can be grouped together (block sampling). Blocks are chosen such that calculations can be simplified. At each iteration, a new candidate is suggested for one block (and parameters in the other blocks retain their current value). The candidate values for that block are either accepted or rejected. Then the same is done for the next parameter block, etc. Example 1. Suggest a candidate for a. (In this example, a* is accepted.) 2. Suggest a candidate for b. (In this example, b* is accepted.) 3.
46 Slide 46 Markov chain Monte Carlo (3) To sample from a multi-dimensional posterior (e.g. the posterior of a model with several unknown parameters), parameters can be grouped together (block sampling). Blocks are chosen such that calculations can be simplified. At each iteration, a new candidate is suggested for one block (and parameters in the other blocks retain their current value). The candidate values for that block are either accepted or rejected. Then the same is done for the next parameter block, etc. Example 1. Suggest a candidate for a. (In this example, a* is accepted.) 2. Suggest a candidate for b. (In this example, b* is accepted.) 3. The chain moves to [a*, b*].
47 Slide 47 Gibbs sampling Gibbs sampling is a special case of MCMC. Here, the posterior parameter space is divided into blocks of parameters, such that for each block, the conditional posterior probabilities are known. Then, at each step i in the chain, candidate values x* are generated randomly from the conditional posterior probability for each parameter block, given the current values of the other parameters in the model. Because x* is a draw from the conditional posterior probability, the calculation of the MCMC acceptance probability always gives 1. Thus, the chain always moves to x* (xi+1=x*). The sampler thus converges more quickly.
48 Slide 48 Difficulties with MCMC Unfortunately, the Markov chain Monte Carlo algorithms do not always work well, and some care is needed when checking for convergence to the posterior distribution of interest. The most common problems are: Bad mixing : The chain does not move well because the candidate acceptance rate is too low. Cause: The candidate generator often suggests candidates that are too unlikely compared to the current value. i
49 Slide 49 Difficulties with MCMC Unfortunately, the Markov chain Monte Carlo algorithms do not always work well, and some care is needed when checking for convergence to the posterior distribution of interest. The most common problems are: Bad mixing Trends in the chain: The exploration of posterior model space is slow and the chain seems to have a direction. Cause: The candidate generator suggests candidates too close to the current values.
50 Slide 50 Difficulties with MCMC Unfortunately, the Markov chain Monte Carlo algorithms do not always work well, and some care is needed when checking for convergence to the posterior distribution of interest. The most common problems are: Bad mixing Trends in the chain Poor coverage of posterior probability: The chain seems to mix well, but it is stuck at a local maximum of posterior probability. The samples thus do not exhaust the posterior model space. Cause: Inappropriate candidate generator.
51 Slide 51 Difficulties with MCMC Unfortunately, the Markov chain Monte Carlo algorithms do not always work well, and some care is needed when checking for convergence to the posterior distribution of interest. The most common problems are: Bad mixing Trends in the chain Poor coverage of posterior probability Because of these difficulties, generating samples from a Bayesian posterior requires a lot of attention to detail and can often not be fully automated. Diagnostic criteria exist to aid in detecting convergence and good mixing of the MCMC sampler.
52 Slide 52 Sequential Importance Sampling This algorithm is very suitable for data that can be obtained sequentially, for example to monitor an industrial process. SIS (usually implemented as a particle filter) can also be applied to more general problems. In SIS, the posterior distribution of interest is approximated by a swarm of particles, where each particle is one possible realisation of the model. For example, in a model with two parameter values, a and b, a particle could be the pair (a=4.5, b=-2). The posterior density function is split into factors, and at each step in the algorithm, all particles are resampled based on weights. These weights are derived from the factors that make up the pdf. For example, the first step might weight the particle sample according to the Bayesian prior. The second step might weight the updated set of particles according to the factor that corresponds to the first datum. The next resampling may take into account the next datum, etc, until the data are used up.
53 Slide 53 Sequential Importance Sampling This algorithm is very suitable for data that can be obtained sequentially, for example to monitor an industrial process. SIS (usually implemented as a particle filter) can also be applied to more general problems. p data p p data 1 p data 2... Prior a=2.5 a=3.1 a=-1 a=4 a=2.7 a=1.7...
54 Slide 54 Sequential Importance Sampling This algorithm is very suitable for data that can be obtained sequentially, for example to monitor an industrial process. SIS (usually implemented as a particle filter) can also be applied to more general problems. p data p p data 1 p data 2... a=2.5 a=3.1 a=-1 a=4 a=2.7 a= Weights due to the first datum
55 Slide 55 Sequential Importance Sampling This algorithm is very suitable for data that can be obtained sequentially, for example to monitor an industrial process. SIS (usually implemented as a particle filter) can also be applied to more general problems. p data p p data 1 p data 2... Weighted resampling... a=2.5 a=3.1 a=-1 a=4 a=2.7 a= a=-1 a=-1 a=-1 a=-1 a=1.7 a=1.7
56 Slide 56 Sequential Importance Sampling This algorithm is very suitable for data that can be obtained sequentially, for example to monitor an industrial process. SIS (usually implemented as a particle filter) can also be applied to more general problems. p data p p data 1 p data 2... again, calculate weights and resample... a=2.5 a=3.1 a=-1 a=4 a=2.7 a= a=-1 a=-1 a=-1 a=-1 a=1.7 a= a=-1 a=-1 a=-1 a=-1 a=-1 a=1.7
57 Slide 57 Sequential Importance Sampling This algorithm is very suitable for data that can be obtained sequentially, for example to monitor an industrial process. SIS (usually implemented as a particle filter) can also be applied to more general problems. p data p p data 1 p data 2... a=2.5 a=3.1 a=-1 a=4 a=2.7 a= a=-1 a=-1 a=-1 a=-1 a=1.7 a= a=-1 a=-1 a=-1 a=-1 a=-1 a=1.7 When all the data are used up, the final swarm of particles is a sample from the posterior distribution. Because of its sequential structure, SIS is often used with time-series data.
58 Slide 58 Difficulties with SIS The main problem with SIS is particle depletion: At each resampling step, the number of different particles is reduced, and no new particles are created. Because the number of particles is finite, eventually there are many identical particles. Different solutions have been suggested, usually based on randomly generating new particles at each step that are slightly different from the existing particles but not too different to break the ergodic properties of the sampler. Other methods are being explored this is an area of active research.
59 Slide 59 Comparison MCMC and SIS MCMC SIS Sampling Chain generates samples one by one All samples are generated at once Data Required from the start Can be added sequentially Computational cost 10,000's of iterations 10,000's of particles Challenges Convergence and mixing Particle depletion Uses Very versatile Live time-series
60 Slide 60 Implementations For MCMC, many ready-made implementations exist. A good place to start is the package OpenBUGS (ongoing development of WinBUGS), which implements the Gibbs and other samplers. With a familiar Windows interface and a very general symbolic language to specify models, OpenBUGS can solve most classes of Bayesian models. R offers several add-on packages with MCMC capabilities, as well as an interface to OpenBUGS, called BRugs. In terms of speed and efficiency, it may be best to hand-code the MCMC sampler directly in Fortran, C or another suitable language. For SIS, I am not aware of any ready-made packages, but there are ongoing developments.
61 Slide 61 Hands-on Example Here I demonstrate the use of OpenBUGS. I've made up this example but the basic approach carries through to real applications in health economics. 8 RCTs have been carried out to investigate the effectiveness of treatments A and B (observing the number of symptom-free patients after 1 year). Treatment A costs SEK 10,000, whereas treatment B costs SEK 14,000. QALY values are given by a probability distribution. The trial data is summarised as follows: na ra nb rb QALY symptom-free: Beta(9,1) QALY with symptoms: Beta(5,5)
62 Slide 62 Statistical model An evidence synthesis model is required to combine the information from the 8 RCTs. Let's choose a random-baselines, random-effects model. We model trial outcomes on the log-odds probability scale, with the treatment effect being additive on the log-odds scale. Letting i denote a trial, we have: Probability with treatment A (baseline): Log-odds treatment effect: Probability with treatment B: A logit pi = i ti B i logit p = i t i
63 Slide 63 Statistical model (2) Random-baselines, random-effects model on the log-odds scale A i Probability with treatment A (baseline): logit p = i ti Log-odds treatment effect: B logit pi = i t i Probability with treatment B: We need to relate the trial-specific parameters μi and ti to their underlying values Μ and T. On the log-odds scale, these are usually assumed to be normally distributed. 2 i ~ Norm M, M Random baseline: Random treatment effect: 2 T t i ~ Norm T,
64 Slide 64 Statistical model (3) Next, the model requires a sampling distribution: Given a set of values for the unknown parameters, how likely is an observed datum? The model yields a probability and we have binomial data, so the only sensible choice is A A A r i ~ Binom pi, ni B B B r i ~ Binom pi, ni By now we have 20 unknown parameters (8 ti, 8 μi, M, σm, T and σt). So far we have made arbitrary choices in model design we could just as well have chosen a fixed-effects model (with fewer parameters) or designed something more complicated.
65 Slide 65 Statistical model (4) Now, because this is a Bayesian model, we need priors for the unknown parameters M, σm, T and σt. If we already know something about the parameters we could add this knowledge as prior information. For example, there may be further information on the baseline probability of symptom-free days we could express this through the prior on M, if we consider the information relevant. Otherwise, we choose sensible priors that have little information, along the lines of: M and T lie in the real line, and we know little about it, so let's pick a Normal prior with mean 0 and large variance.
66 Slide 66 Statistical model (5) Model equations: logit pia = i B i logit p = i t i 2 M 2 T i ~ Norm M, Sampling distribution: A t i ~ Norm T, A A r i ~ Binom pi, ni Priors: M ~ Norm 0,10000 B B B r i ~ Binom pi, ni (log-odds probabilities) T ~ Norm 0,10000 M ~ Unif 0,2 T ~ Unif 0,2 (log-odds standard deviations)
67 Slide 67 Statistical model (6) Model equations: logit pia = i B i logit p = i t i 2 M 2 T i ~ Norm M, Sampling distribution: A t i ~ Norm T, A A r i ~ Binom pi, ni B B B r i ~ Binom pi, ni The model equations and the sampling distribution are common to the Frequentist and the Bayesian approaches. If you already have a Frequentist model, then you (should) already have specified these.
68 Slide 68 Statistical model (7) Priors do not occur in the Frequentist setting, so you probably have to make them up. In this example, the priors are meant to be uninformative, i.e. they are supposed to add no information to the result. It is good practice to test this by changing the priors a little bit and observing the impact on the results of your model. Priors: M ~ Norm 0,10000 (log-odds probabilities) T ~ Norm 0,10000 M ~ Unif 0,2 T ~ Unif 0,2 (log-odds standard deviation)
69 Slide 69 OpenBUGS Let us fit this Bayesian model using OpenBUGS. The OpenBUGS syntax is relatively straightforward and similar to R. model { for (i in 1:N) { logit(pa[i])<-mu[i] logit(pb[i])<-mu[i]+t[i] ra[i]~dbin(pa[i],na[i]) rb[i]~dbin(pb[i],nb[i]) mu[i]~dnorm(m,precm) t[i] ~dnorm(t,prect) } M~dnorm(0,0.0001) T~dnorm(0,0.0001) precm<-1/pow(sigmam,2) prect<-1/pow(sigmat,2) sigmam~dunif(0,2) sigmat~dunif(0,2) } A logit pi = i B i logit p = i t i A i B i A i B i A i B i 2 M r ~ Binom p, n r ~ Binom p, n i ~ Norm M, 2 t i ~ Norm T, T
70 Slide 70 OpenBUGS (2) The data are specified in a separate section so that they can be entered or changed easily. model { for (i in 1:N) { logit(pa[i])<-mu[i] logit(pb[i])<-mu[i]+t[i] ra[i]~dbin(pa[i],na[i]) rb[i]~dbin(pb[i],nb[i]) mu[i]~dnorm(m,precm) t[i] ~dnorm(t,prect) #data } list(n=8, M~dnorm(0,0.0001) na=c(120,15,84,398, T~dnorm(0,0.0001) ra=c( 65, 9,39,202, precm<-1/pow(sigmam,2) nb=c(120,16,45,402, prect<-1/pow(sigmat,2) rb=c( 81,15,29,270, sigmam~dunif(0,2) sigmat~dunif(0,2) } 80,40, 97,121), 45,17, 48, 63), 77,20,100,115), 52,12, 68, 80))
71 Slide 71 OpenBUGS (3) The OpenBUGS window can look like this.
72 Slide 72 OpenBUGS (4) In this example, OpenBUGS explores the model's posterior reasonably well. The three colours denote three chains that are run in parallel. Note that there is no evidence that initial values are influencing the chains. Also, each chain appears to wiggle quite well and the three chains overlap, indicating that they are exploring the same posterior space (as they should).
73 Slide 73 OpenBUGS (5) Here's a screenshot from another model, in which the sampler did not converge (just to give you an idea of what to look for...). Poor mixing: Three chains (which should all explore the same parameter space) are far from each other, and trends are evident.
74 Slide 74 Model convergence WinBUGS and OpenBUGS provide a few formal diagnostics to check for convergence and performance of the sampler, for example the BrooksGelman-Rubin diagram and plots of within-chain autocorrelation. High MC_error can also indicate convergence problems. To avoid convergence problems, bear in mind the following. 1. It is difficult to fit many unknown parameters to little data. 2. If you can exploit conjugate distributions, do so. 3. Sometimes you can get around convergence problems by re-writing your model equations without changing the underlying model. 4. If high within-chain autocorrelation is the only problem, you can thin the posterior samples and only keep every nth draw.
75 Slide 75 Example continued When you are satisfied with the posterior sampling, you can generate any desired summary statistic for your posterior. For example, here's an overview of the posterior means and credibility intervals. In this example, the treatment effect T is positive (on the log-odds scale), i.e. the treatment B has a higher probability of symptom-free.
76 Slide 76 Example continued When you are satisfied with the posterior sampling, you can generate any desired summary statistic for your posterior. MC_error is another indication for how well the sampler performed. The sampling error should be much smaller than the estimated posterior standard deviation.
77 Slide 77 Example continued When you are satisfied with the posterior sampling, you can generate any desired summary statistic for your posterior. The 95% credibility interval for the treatment efficacy parameter T is [0.481, ], i.e. treatment B has a very significant effect on the probability of symptom-freeness after 1 year.
78 Slide 78 Example continued But what is the probability P that treatment B is the cost-effective choice at a willingness-to-pay (WTP) of λ=sek 50,000? This probability P can be calculated from the model parameters as follows. Let's assume that the underlying baseline and treatment effect would apply to the target population, i.e. in the target population B A logit p =M and logit p =M T. We calculate the net benefit of the treatments (NB), using the costs C and utilities U. A A A A B B B B NB =[ p U free 1 p U symptoms ] C NB =[ p U free 1 p U symptoms ] C The probability P is given by B A P=Pr NB NB
79 Slide 79 Example continued But what is the probability P that treatment B is the cost-effective choice at a willingness-to-pay (WTP) of λ=sek 50,000? We can calculate this directly in WinBUGS, adding a few more lines of code. model {... logit(pa)<-m logit(pb)<-m+t Uf~dbeta(9,1) Us~dbeta(5,5) NBA<-(PA*Uf+(1-PA)*Us)*WTP-CA NBB<-(PB*Uf+(1-PB)*Us)*WTP-CB P<-step(NBB-NBA) } Numerically, we simply look at all the draws from the posterior and check which of them fulfill the condition. This proportion is the posterior probability P. There is no need for any further tests.
80 Slide 80 Example continued But what is the probability P that treatment B is the cost-effective choice at a willingness-to-pay (WTP) of λ=sek 50,000? Here's the posterior summary for our new quantities. There is a lot of overlap between the net benefits for treatment A (95%-CI [SEK 16240, 33200]) and B (95%-CI [SEK 15580, 31090]). Accordingly, the probability that treatment B is cost-effective is estimated at
81 Slide 81 Example continued But what is the probability P that treatment B is the cost-effective choice at a willingness-to-pay (WTP) of λ=sek 50,000? Here's the posterior summary for our new quantities. In this example, I defined the probability P by checking a condition, rather than as a stochastic variable with its own prior and probability density. This is why the confidence intervals for P do not mean anything.
82 Slide 82 Example continued OpenBUGS can also produce graphical output for all quantities of interest. NBA sample: NBB sample: P(NBB) E-4 P(NBA) E-5 For example, here are the posterior densities for the net benefits NBA and NBB. 1.0E+4 2.0E+4 3.0E+4-1.0E E+4 NBA 3.0E+4 NBB They both show quite wide distributions and their support on the x-axes overlaps substantially.
83 Slide 83 Bayesian decision theory With a probabilistic net benefit function, Bayesian decision theory can be applied to optimise management decisions. Bayesian models can thus directly feed in to management processes. In a Frequentist model, it is not generally possible to find a probability distribution for an unknown parameter because a Frequentist calculates the likelihood of observing the data and uses this to make inferences on the model parameters. Frequentist models do not offer an obvious way for calculating quantities that are derived from individual parameter values (such as the probability P or a net benefit) this makes them less amenable to management processes.
84 Slide 84 Summary Frequentist modelling centres on the likelihood function, i.e. how likely are the data given a particular model. Bayesian modelling centres on the probabilities of models and model parameters, by combining the likelihood of the data with prior probabilities of the unknown parameters. Both can be used equally well to fit models and to make inferences on model parameters. However, only Bayesian statistics is capable of assigning probabilities to model quantities. This makes it possible to calculate derived quantities and their uncertainties. Numerical methods for fitting Bayesian models require some care and experience.
Bayesian Statistics in One Hour. Patrick Lam
Bayesian Statistics in One Hour Patrick Lam Outline Introduction Bayesian Models Applications Missing Data Hierarchical Models Outline Introduction Bayesian Models Applications Missing Data Hierarchical
Model-based Synthesis. Tony O Hagan
Model-based Synthesis Tony O Hagan Stochastic models Synthesising evidence through a statistical model 2 Evidence Synthesis (Session 3), Helsinki, 28/10/11 Graphical modelling The kinds of models that
Introduction to Mobile Robotics Bayes Filter Particle Filter and Monte Carlo Localization
Introduction to Mobile Robotics Bayes Filter Particle Filter and Monte Carlo Localization Wolfram Burgard, Maren Bennewitz, Diego Tipaldi, Luciano Spinello 1 Motivation Recall: Discrete filter Discretize
Bayesian Analysis for the Social Sciences
Bayesian Analysis for the Social Sciences Simon Jackman Stanford University http://jackman.stanford.edu/bass November 9, 2012 Simon Jackman (Stanford) Bayesian Analysis for the Social Sciences November
THE USE OF STATISTICAL DISTRIBUTIONS TO MODEL CLAIMS IN MOTOR INSURANCE
THE USE OF STATISTICAL DISTRIBUTIONS TO MODEL CLAIMS IN MOTOR INSURANCE Batsirai Winmore Mazviona 1 Tafadzwa Chiduza 2 ABSTRACT In general insurance, companies need to use data on claims gathered from
Master s Theory Exam Spring 2006
Spring 2006 This exam contains 7 questions. You should attempt them all. Each question is divided into parts to help lead you through the material. You should attempt to complete as much of each problem
The HB. How Bayesian methods have changed the face of marketing research. Summer 2004
The HB How Bayesian methods have changed the face of marketing research. 20 Summer 2004 Reprinted with permission from Marketing Research, Summer 2004, published by the American Marketing Association.
BAYESIAN ANALYSIS OF AN AGGREGATE CLAIM MODEL USING VARIOUS LOSS DISTRIBUTIONS. by Claire Dudley
BAYESIAN ANALYSIS OF AN AGGREGATE CLAIM MODEL USING VARIOUS LOSS DISTRIBUTIONS by Claire Dudley A dissertation submitted for the award of the degree of Master of Science in Actuarial Management School
Likelihood: Frequentist vs Bayesian Reasoning
"PRINCIPLES OF PHYLOGENETICS: ECOLOGY AND EVOLUTION" Integrative Biology 200B University of California, Berkeley Spring 2009 N Hallinan Likelihood: Frequentist vs Bayesian Reasoning Stochastic odels and
STA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! [email protected]! http://www.cs.toronto.edu/~rsalakhu/ Lecture 6 Three Approaches to Classification Construct
Bayesian Statistical Analysis in Medical Research
Bayesian Statistical Analysis in Medical Research David Draper Department of Applied Mathematics and Statistics University of California, Santa Cruz [email protected] www.ams.ucsc.edu/ draper ROLE Steering
Bayesian Statistics: Indian Buffet Process
Bayesian Statistics: Indian Buffet Process Ilker Yildirim Department of Brain and Cognitive Sciences University of Rochester Rochester, NY 14627 August 2012 Reference: Most of the material in this note
Probabilistic Models for Big Data. Alex Davies and Roger Frigola University of Cambridge 13th February 2014
Probabilistic Models for Big Data Alex Davies and Roger Frigola University of Cambridge 13th February 2014 The State of Big Data Why probabilistic models for Big Data? 1. If you don t have to worry about
Service courses for graduate students in degree programs other than the MS or PhD programs in Biostatistics.
Course Catalog In order to be assured that all prerequisites are met, students must acquire a permission number from the education coordinator prior to enrolling in any Biostatistics course. Courses are
A Bayesian hierarchical surrogate outcome model for multiple sclerosis
A Bayesian hierarchical surrogate outcome model for multiple sclerosis 3 rd Annual ASA New Jersey Chapter / Bayer Statistics Workshop David Ohlssen (Novartis), Luca Pozzi and Heinz Schmidli (Novartis)
MAS2317/3317. Introduction to Bayesian Statistics. More revision material
MAS2317/3317 Introduction to Bayesian Statistics More revision material Dr. Lee Fawcett, 2014 2015 1 Section A style questions 1. Describe briefly the frequency, classical and Bayesian interpretations
How To Check For Differences In The One Way Anova
MINITAB ASSISTANT WHITE PAPER This paper explains the research conducted by Minitab statisticians to develop the methods and data checks used in the Assistant in Minitab 17 Statistical Software. One-Way
Parallelization Strategies for Multicore Data Analysis
Parallelization Strategies for Multicore Data Analysis Wei-Chen Chen 1 Russell Zaretzki 2 1 University of Tennessee, Dept of EEB 2 University of Tennessee, Dept. Statistics, Operations, and Management
Markov Chain Monte Carlo Simulation Made Simple
Markov Chain Monte Carlo Simulation Made Simple Alastair Smith Department of Politics New York University April2,2003 1 Markov Chain Monte Carlo (MCMC) simualtion is a powerful technique to perform numerical
CONTENTS OF DAY 2. II. Why Random Sampling is Important 9 A myth, an urban legend, and the real reason NOTES FOR SUMMER STATISTICS INSTITUTE COURSE
1 2 CONTENTS OF DAY 2 I. More Precise Definition of Simple Random Sample 3 Connection with independent random variables 3 Problems with small populations 8 II. Why Random Sampling is Important 9 A myth,
SENSITIVITY ANALYSIS AND INFERENCE. Lecture 12
This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike License. Your use of this material constitutes acceptance of that license and the conditions of use of materials on this
Modeling and Analysis of Call Center Arrival Data: A Bayesian Approach
Modeling and Analysis of Call Center Arrival Data: A Bayesian Approach Refik Soyer * Department of Management Science The George Washington University M. Murat Tarimcilar Department of Management Science
Inference of Probability Distributions for Trust and Security applications
Inference of Probability Distributions for Trust and Security applications Vladimiro Sassone Based on joint work with Mogens Nielsen & Catuscia Palamidessi Outline 2 Outline Motivations 2 Outline Motivations
CHAPTER 2 Estimating Probabilities
CHAPTER 2 Estimating Probabilities Machine Learning Copyright c 2016. Tom M. Mitchell. All rights reserved. *DRAFT OF January 24, 2016* *PLEASE DO NOT DISTRIBUTE WITHOUT AUTHOR S PERMISSION* This is a
PS 271B: Quantitative Methods II. Lecture Notes
PS 271B: Quantitative Methods II Lecture Notes Langche Zeng [email protected] The Empirical Research Process; Fundamental Methodological Issues 2 Theory; Data; Models/model selection; Estimation; Inference.
Dirichlet Processes A gentle tutorial
Dirichlet Processes A gentle tutorial SELECT Lab Meeting October 14, 2008 Khalid El-Arini Motivation We are given a data set, and are told that it was generated from a mixture of Gaussian distributions.
Institute of Actuaries of India Subject CT3 Probability and Mathematical Statistics
Institute of Actuaries of India Subject CT3 Probability and Mathematical Statistics For 2015 Examinations Aim The aim of the Probability and Mathematical Statistics subject is to provide a grounding in
Objections to Bayesian statistics
Bayesian Analysis (2008) 3, Number 3, pp. 445 450 Objections to Bayesian statistics Andrew Gelman Abstract. Bayesian inference is one of the more controversial approaches to statistics. The fundamental
Statistics Graduate Courses
Statistics Graduate Courses STAT 7002--Topics in Statistics-Biological/Physical/Mathematics (cr.arr.).organized study of selected topics. Subjects and earnable credit may vary from semester to semester.
Inference on Phase-type Models via MCMC
Inference on Phase-type Models via MCMC with application to networks of repairable redundant systems Louis JM Aslett and Simon P Wilson Trinity College Dublin 28 th June 202 Toy Example : Redundant Repairable
Bayesian Phylogeny and Measures of Branch Support
Bayesian Phylogeny and Measures of Branch Support Bayesian Statistics Imagine we have a bag containing 100 dice of which we know that 90 are fair and 10 are biased. The
Parameter estimation for nonlinear models: Numerical approaches to solving the inverse problem. Lecture 12 04/08/2008. Sven Zenker
Parameter estimation for nonlinear models: Numerical approaches to solving the inverse problem Lecture 12 04/08/2008 Sven Zenker Assignment no. 8 Correct setup of likelihood function One fixed set of observation
Using SAS PROC MCMC to Estimate and Evaluate Item Response Theory Models
Using SAS PROC MCMC to Estimate and Evaluate Item Response Theory Models Clement A Stone Abstract Interest in estimating item response theory (IRT) models using Bayesian methods has grown tremendously
Comparison of frequentist and Bayesian inference. Class 20, 18.05, Spring 2014 Jeremy Orloff and Jonathan Bloom
Comparison of frequentist and Bayesian inference. Class 20, 18.05, Spring 2014 Jeremy Orloff and Jonathan Bloom 1 Learning Goals 1. Be able to explain the difference between the p-value and a posterior
Information Theory and Coding Prof. S. N. Merchant Department of Electrical Engineering Indian Institute of Technology, Bombay
Information Theory and Coding Prof. S. N. Merchant Department of Electrical Engineering Indian Institute of Technology, Bombay Lecture - 17 Shannon-Fano-Elias Coding and Introduction to Arithmetic Coding
DECISION MAKING UNDER UNCERTAINTY:
DECISION MAKING UNDER UNCERTAINTY: Models and Choices Charles A. Holloway Stanford University TECHNISCHE HOCHSCHULE DARMSTADT Fachbereich 1 Gesamtbibliothek Betrtebswirtscrtaftslehre tnventar-nr. :...2>2&,...S'.?S7.
Applications of R Software in Bayesian Data Analysis
Article International Journal of Information Science and System, 2012, 1(1): 7-23 International Journal of Information Science and System Journal homepage: www.modernscientificpress.com/journals/ijinfosci.aspx
Part 2: One-parameter models
Part 2: One-parameter models Bernoilli/binomial models Return to iid Y 1,...,Y n Bin(1, θ). The sampling model/likelihood is p(y 1,...,y n θ) =θ P y i (1 θ) n P y i When combined with a prior p(θ), Bayes
Introduction to Markov Chain Monte Carlo
Introduction to Markov Chain Monte Carlo Monte Carlo: sample from a distribution to estimate the distribution to compute max, mean Markov Chain Monte Carlo: sampling using local information Generic problem
STT315 Chapter 4 Random Variables & Probability Distributions KM. Chapter 4.5, 6, 8 Probability Distributions for Continuous Random Variables
Chapter 4.5, 6, 8 Probability Distributions for Continuous Random Variables Discrete vs. continuous random variables Examples of continuous distributions o Uniform o Exponential o Normal Recall: A random
Solutions: Problems for Chapter 3. Solutions: Problems for Chapter 3
Problem A: You are dealt five cards from a standard deck. Are you more likely to be dealt two pairs or three of a kind? experiment: choose 5 cards at random from a standard deck Ω = {5-combinations of
Sample Size Designs to Assess Controls
Sample Size Designs to Assess Controls B. Ricky Rambharat, PhD, PStat Lead Statistician Office of the Comptroller of the Currency U.S. Department of the Treasury Washington, DC FCSM Research Conference
Gibbs Sampling and Online Learning Introduction
Statistical Techniques in Robotics (16-831, F14) Lecture#10(Tuesday, September 30) Gibbs Sampling and Online Learning Introduction Lecturer: Drew Bagnell Scribes: {Shichao Yang} 1 1 Sampling Samples are
Data Modeling & Analysis Techniques. Probability & Statistics. Manfred Huber 2011 1
Data Modeling & Analysis Techniques Probability & Statistics Manfred Huber 2011 1 Probability and Statistics Probability and statistics are often used interchangeably but are different, related fields
A Bootstrap Metropolis-Hastings Algorithm for Bayesian Analysis of Big Data
A Bootstrap Metropolis-Hastings Algorithm for Bayesian Analysis of Big Data Faming Liang University of Florida August 9, 2015 Abstract MCMC methods have proven to be a very powerful tool for analyzing
Basics of Statistical Machine Learning
CS761 Spring 2013 Advanced Machine Learning Basics of Statistical Machine Learning Lecturer: Xiaojin Zhu [email protected] Modern machine learning is rooted in statistics. You will find many familiar
Summary of Formulas and Concepts. Descriptive Statistics (Ch. 1-4)
Summary of Formulas and Concepts Descriptive Statistics (Ch. 1-4) Definitions Population: The complete set of numerical information on a particular quantity in which an investigator is interested. We assume
From the help desk: Bootstrapped standard errors
The Stata Journal (2003) 3, Number 1, pp. 71 80 From the help desk: Bootstrapped standard errors Weihua Guan Stata Corporation Abstract. Bootstrapping is a nonparametric approach for evaluating the distribution
11. Time series and dynamic linear models
11. Time series and dynamic linear models Objective To introduce the Bayesian approach to the modeling and forecasting of time series. Recommended reading West, M. and Harrison, J. (1997). models, (2 nd
More details on the inputs, functionality, and output can be found below.
Overview: The SMEEACT (Software for More Efficient, Ethical, and Affordable Clinical Trials) web interface (http://research.mdacc.tmc.edu/smeeactweb) implements a single analysis of a two-armed trial comparing
BayesX - Software for Bayesian Inference in Structured Additive Regression
BayesX - Software for Bayesian Inference in Structured Additive Regression Thomas Kneib Faculty of Mathematics and Economics, University of Ulm Department of Statistics, Ludwig-Maximilians-University Munich
Validation of Software for Bayesian Models using Posterior Quantiles. Samantha R. Cook Andrew Gelman Donald B. Rubin DRAFT
Validation of Software for Bayesian Models using Posterior Quantiles Samantha R. Cook Andrew Gelman Donald B. Rubin DRAFT Abstract We present a simulation-based method designed to establish that software
Tutorial on Markov Chain Monte Carlo
Tutorial on Markov Chain Monte Carlo Kenneth M. Hanson Los Alamos National Laboratory Presented at the 29 th International Workshop on Bayesian Inference and Maximum Entropy Methods in Science and Technology,
BASIC STATISTICAL METHODS FOR GENOMIC DATA ANALYSIS
BASIC STATISTICAL METHODS FOR GENOMIC DATA ANALYSIS SEEMA JAGGI Indian Agricultural Statistics Research Institute Library Avenue, New Delhi-110 012 [email protected] Genomics A genome is an organism s
Introduction to Monte Carlo. Astro 542 Princeton University Shirley Ho
Introduction to Monte Carlo Astro 542 Princeton University Shirley Ho Agenda Monte Carlo -- definition, examples Sampling Methods (Rejection, Metropolis, Metropolis-Hasting, Exact Sampling) Markov Chains
SAS Certificate Applied Statistics and SAS Programming
SAS Certificate Applied Statistics and SAS Programming SAS Certificate Applied Statistics and Advanced SAS Programming Brigham Young University Department of Statistics offers an Applied Statistics and
Bayesian Machine Learning (ML): Modeling And Inference in Big Data. Zhuhua Cai Google, Rice University [email protected]
Bayesian Machine Learning (ML): Modeling And Inference in Big Data Zhuhua Cai Google Rice University [email protected] 1 Syllabus Bayesian ML Concepts (Today) Bayesian ML on MapReduce (Next morning) Bayesian
Model Calibration with Open Source Software: R and Friends. Dr. Heiko Frings Mathematical Risk Consulting
Model with Open Source Software: and Friends Dr. Heiko Frings Mathematical isk Consulting Bern, 01.09.2011 Agenda in a Friends Model with & Friends o o o Overview First instance: An Extreme Value Example
Lecture 9: Bayesian hypothesis testing
Lecture 9: Bayesian hypothesis testing 5 November 27 In this lecture we ll learn about Bayesian hypothesis testing. 1 Introduction to Bayesian hypothesis testing Before we go into the details of Bayesian
1 Prior Probability and Posterior Probability
Math 541: Statistical Theory II Bayesian Approach to Parameter Estimation Lecturer: Songfeng Zheng 1 Prior Probability and Posterior Probability Consider now a problem of statistical inference in which
The Variability of P-Values. Summary
The Variability of P-Values Dennis D. Boos Department of Statistics North Carolina State University Raleigh, NC 27695-8203 [email protected] August 15, 2009 NC State Statistics Departement Tech Report
The Basics of Graphical Models
The Basics of Graphical Models David M. Blei Columbia University October 3, 2015 Introduction These notes follow Chapter 2 of An Introduction to Probabilistic Graphical Models by Michael Jordan. Many figures
Recursive Estimation
Recursive Estimation Raffaello D Andrea Spring 04 Problem Set : Bayes Theorem and Bayesian Tracking Last updated: March 8, 05 Notes: Notation: Unlessotherwisenoted,x, y,andz denoterandomvariables, f x
Dealing with Missing Data
Res. Lett. Inf. Math. Sci. (2002) 3, 153-160 Available online at http://www.massey.ac.nz/~wwiims/research/letters/ Dealing with Missing Data Judi Scheffer I.I.M.S. Quad A, Massey University, P.O. Box 102904
The result of the bayesian analysis is the probability distribution of every possible hypothesis H, given one real data set D. This prestatistical approach to our problem was the standard approach of Laplace
Fairfield Public Schools
Mathematics Fairfield Public Schools AP Statistics AP Statistics BOE Approved 04/08/2014 1 AP STATISTICS Critical Areas of Focus AP Statistics is a rigorous course that offers advanced students an opportunity
Handling attrition and non-response in longitudinal data
Longitudinal and Life Course Studies 2009 Volume 1 Issue 1 Pp 63-72 Handling attrition and non-response in longitudinal data Harvey Goldstein University of Bristol Correspondence. Professor H. Goldstein
Master s thesis tutorial: part III
for the Autonomous Compliant Research group Tinne De Laet, Wilm Decré, Diederik Verscheure Katholieke Universiteit Leuven, Department of Mechanical Engineering, PMA Division 30 oktober 2006 Outline General
Imputing Missing Data using SAS
ABSTRACT Paper 3295-2015 Imputing Missing Data using SAS Christopher Yim, California Polytechnic State University, San Luis Obispo Missing data is an unfortunate reality of statistics. However, there are
Government of Russian Federation. Faculty of Computer Science School of Data Analysis and Artificial Intelligence
Government of Russian Federation Federal State Autonomous Educational Institution of High Professional Education National Research University «Higher School of Economics» Faculty of Computer Science School
ECON 459 Game Theory. Lecture Notes Auctions. Luca Anderlini Spring 2015
ECON 459 Game Theory Lecture Notes Auctions Luca Anderlini Spring 2015 These notes have been used before. If you can still spot any errors or have any suggestions for improvement, please let me know. 1
Bootstrapping Big Data
Bootstrapping Big Data Ariel Kleiner Ameet Talwalkar Purnamrita Sarkar Michael I. Jordan Computer Science Division University of California, Berkeley {akleiner, ameet, psarkar, jordan}@eecs.berkeley.edu
Recall this chart that showed how most of our course would be organized:
Chapter 4 One-Way ANOVA Recall this chart that showed how most of our course would be organized: Explanatory Variable(s) Response Variable Methods Categorical Categorical Contingency Tables Categorical
Monte Carlo testing with Big Data
Monte Carlo testing with Big Data Patrick Rubin-Delanchy University of Bristol & Heilbronn Institute for Mathematical Research Joint work with: Axel Gandy (Imperial College London) with contributions from:
Centre for Central Banking Studies
Centre for Central Banking Studies Technical Handbook No. 4 Applied Bayesian econometrics for central bankers Andrew Blake and Haroon Mumtaz CCBS Technical Handbook No. 4 Applied Bayesian econometrics
Generalized linear models and software for network meta-analysis
Generalized linear models and software for network meta-analysis Sofia Dias & Gert van Valkenhoef Tufts University, Boston MA, USA, June 2012 Generalized linear model (GLM) framework Pairwise Meta-analysis
Journal of Statistical Software
JSS Journal of Statistical Software October 2014, Volume 61, Issue 7. http://www.jstatsoft.org/ WebBUGS: Conducting Bayesian Statistical Analysis Online Zhiyong Zhang University of Notre Dame Abstract
ESTIMATING COMPLETION RATES FROM SMALL SAMPLES USING BINOMIAL CONFIDENCE INTERVALS: COMPARISONS AND RECOMMENDATIONS
PROCEEDINGS of the HUMAN FACTORS AND ERGONOMICS SOCIETY 49th ANNUAL MEETING 200 2 ESTIMATING COMPLETION RATES FROM SMALL SAMPLES USING BINOMIAL CONFIDENCE INTERVALS: COMPARISONS AND RECOMMENDATIONS Jeff
Interpretation of Somers D under four simple models
Interpretation of Somers D under four simple models Roger B. Newson 03 September, 04 Introduction Somers D is an ordinal measure of association introduced by Somers (96)[9]. It can be defined in terms
1 Maximum likelihood estimation
COS 424: Interacting with Data Lecturer: David Blei Lecture #4 Scribes: Wei Ho, Michael Ye February 14, 2008 1 Maximum likelihood estimation 1.1 MLE of a Bernoulli random variable (coin flips) Given N
Tutorial 5: Hypothesis Testing
Tutorial 5: Hypothesis Testing Rob Nicholls [email protected] MRC LMB Statistics Course 2014 Contents 1 Introduction................................ 1 2 Testing distributional assumptions....................
Simulation Exercises to Reinforce the Foundations of Statistical Thinking in Online Classes
Simulation Exercises to Reinforce the Foundations of Statistical Thinking in Online Classes Simcha Pollack, Ph.D. St. John s University Tobin College of Business Queens, NY, 11439 [email protected]
How To Solve A Sequential Mca Problem
Monte Carlo-based statistical methods (MASM11/FMS091) Jimmy Olsson Centre for Mathematical Sciences Lund University, Sweden Lecture 6 Sequential Monte Carlo methods II February 3, 2012 Changes in HA1 Problem
Basic Bayesian Methods
6 Basic Bayesian Methods Mark E. Glickman and David A. van Dyk Summary In this chapter, we introduce the basics of Bayesian data analysis. The key ingredients to a Bayesian analysis are the likelihood
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION Introduction In the previous chapter, we explored a class of regression models having particularly simple analytical
Discrete Structures for Computer Science
Discrete Structures for Computer Science Adam J. Lee [email protected] 6111 Sennott Square Lecture #20: Bayes Theorem November 5, 2013 How can we incorporate prior knowledge? Sometimes we want to know
Nonparametric adaptive age replacement with a one-cycle criterion
Nonparametric adaptive age replacement with a one-cycle criterion P. Coolen-Schrijner, F.P.A. Coolen Department of Mathematical Sciences University of Durham, Durham, DH1 3LE, UK e-mail: [email protected]
Gaussian Processes to Speed up Hamiltonian Monte Carlo
Gaussian Processes to Speed up Hamiltonian Monte Carlo Matthieu Lê Murray, Iain http://videolectures.net/mlss09uk_murray_mcmc/ Rasmussen, Carl Edward. "Gaussian processes to speed up hybrid Monte Carlo
A Bayesian Antidote Against Strategy Sprawl
A Bayesian Antidote Against Strategy Sprawl Benjamin Scheibehenne ([email protected]) University of Basel, Missionsstrasse 62a 4055 Basel, Switzerland & Jörg Rieskamp ([email protected])
Lecture Note 1 Set and Probability Theory. MIT 14.30 Spring 2006 Herman Bennett
Lecture Note 1 Set and Probability Theory MIT 14.30 Spring 2006 Herman Bennett 1 Set Theory 1.1 Definitions and Theorems 1. Experiment: any action or process whose outcome is subject to uncertainty. 2.
An Introduction to Basic Statistics and Probability
An Introduction to Basic Statistics and Probability Shenek Heyward NCSU An Introduction to Basic Statistics and Probability p. 1/4 Outline Basic probability concepts Conditional probability Discrete Random
Probabilistic Methods for Time-Series Analysis
Probabilistic Methods for Time-Series Analysis 2 Contents 1 Analysis of Changepoint Models 1 1.1 Introduction................................ 1 1.1.1 Model and Notation....................... 2 1.1.2 Example:
Curriculum Map Statistics and Probability Honors (348) Saugus High School Saugus Public Schools 2009-2010
Curriculum Map Statistics and Probability Honors (348) Saugus High School Saugus Public Schools 2009-2010 Week 1 Week 2 14.0 Students organize and describe distributions of data by using a number of different
Statistics I for QBIC. Contents and Objectives. Chapters 1 7. Revised: August 2013
Statistics I for QBIC Text Book: Biostatistics, 10 th edition, by Daniel & Cross Contents and Objectives Chapters 1 7 Revised: August 2013 Chapter 1: Nature of Statistics (sections 1.1-1.6) Objectives
CHAPTER 3 EXAMPLES: REGRESSION AND PATH ANALYSIS
Examples: Regression And Path Analysis CHAPTER 3 EXAMPLES: REGRESSION AND PATH ANALYSIS Regression analysis with univariate or multivariate dependent variables is a standard procedure for modeling relationships
