Robert Piché Tampere University of Technology
|
|
- Bethany Hopkins
- 7 years ago
- Views:
Transcription
1 model fit: mu Statistical modelling with WinBUGS Robert Piché Tampere University of Technology diff sample: y/n !1 0 1 x Literature WinBUGS free download Read the first 15 pages of For a full intro to Bayesian statistics, take my course math.tut.fi/~piche/bayes My course is based on Antti Penttinen s course users.jyu.fi/~penttine/bayes09/
2 in this lesson we look at 6 basic statistical models inferring a proportion comparing proportions diff sample: inferring a mean comparing means linear regression model fit: mu slide 3 of 20 logistic regression y/n 0!1 0 1 x
3 inferring a proportion What proportion of the population of Ireland supports the Lisbon agreement? (let s call this:!) When we ask n Irish adults, s of them say that they are in favour. What can we conclude about!? 1. construct an observation model p( s! ) 2. construct a prior model p(!) 3. compute posterior p(! s ) with WinBUGS slide 4 of 20
4 inferring a proportion (2) Observation model Let yi = 1 if person i is in favour ( success ), otherwise yi = 0 assume: given!, the probability that case i is a success is!, i.e. p(y i θ)= assume: given!, the observations are statistically independent, so p(y 1:n θ)= θ (yi = 1) 1 θ (y i = 0) n! θ y i (1 θ) 1 y i = θ s (1 θ) n s, i=1 = θ y i (1 θ) 1 y i (y i {0,1}) where s = " n i=1 y i Thus, given!, s is a random variable with Binomial distribution s θ Binomial(θ,n) slide 5 of 20
5 inferring a proportion (3) The observation model ( likelihood ) tells us how we could generate random s, given! : s θ Binomial(θ,n) Inference is the inverse problem: given s, what is!? Bayesian inference The unknown proportion! is treated like a random variable Its probability density function (pdf) lives in [0, 1] The prior pdf ( before observation ) and posterior pdf are related by Bayes law: p(θ s)= p(s θ)p(θ) p(s θ)p(θ)dθ 2 p(!) prior posterior 20 p(! y) slide 6 of !
6 inferring a proportion (4) The prior, p(!) It is our state of belief about! before we make observation We can use any pdf that lives in [0, 1], e.g. Unif(0, 1) It s convenient to use Beta distribution, with 2 parameters ">0, #>0 x p(x) x E(x) mode(x) V(x) { } (µ σ ) "(α+β) Beta(α,β) (x µ)2 R µ µ σ "(α)"(β) xα 1 (1 x) β 1 α α 1 αβ [0,1] α+β α+β 2 (α+β) 2 (α+β+1) (λ) λ λ { } λ λ λ Beta(1, 1) = Unif small " and # give vague prior slide 7 of 20 plot the Beta pdf with Matlab s disttool
7 inferring a proportion (5) WinBUGS model model { s dbin(theta,n) theta dbeta(1,0.667) ypred } dbin(theta,1) data list(s=650,n=1000) initialisation list(theta=0.1) p(!) ! try dbeta(1,1) try s=65, n=100 try s=6500, n=10000 results node mean sd 2.5% median 97.5% theta slide 8 of 20 95% credibility interval is [0.6185,0.6802]
8 comparing proportions In a study of larynx cancer patients, s1 of the n1 patients who were treated with radiation therapy were cured, compared to s2 of the n2 patients who were treated with surgery. What can we say about!1 (success rate of radiation) vs!2 (surgery)? 1. construct an observation model p( s! ) 2. construct a prior model p(!) 3. compute posterior p(! s ) with WinBUGS 4. compute posterior probability that (!1 $!2 ) slide 9 of 20
9 comparing proportions (2) Observation model assume: given!=(!1,!2), the probability that a radiation therapy patient is cured is!1 and the probability that a surgery therapy patient is cured is!2 assume: given!, the observations are independent i.e. p(s 1,s 2 θ 1,θ 2 )=p(s 1 θ 1 )p(s 2 θ 2 ) s 1 θ 1 Binomial(θ 1,n 1 ) s 2 θ 2 Binomial(θ 2,n 2 ) The prior, p(!1,!2) independent vague slide 10 of 20 p(θ 1,θ 2 )=p(θ 1 )p(θ 2 ) θ 1 Beta(0.5,0.5) θ 2 Beta(0.5,0.5) plot these with dissttool
10 comparing proportions (3) WinBUGS model { s_1 ~ dbin(theta_1,n_1) s_2 ~ dbin(theta_2,n_2) theta_1 ~ dbeta(0.5,0.5) theta_2 ~ dbeta(0.5,0.5) diff <- theta_1-theta_2 P <- step(diff) } # data list(s_1=15,n_1=18,s_2=21,n_2=23) # initialisation list(theta_1=0.8,theta_2=0.8) variables that are deterministic functions of stochastic variables are specified with <- step(diff) = 1 if diff! 0, = 0 otherwise comments are indicated with # slide 11 of 20 results node mean sd 2.5% median 97.5% theta theta diff P diff sample:
11 inferring a mean In 1798, Henry Cavendish performed experiments to measure the specific density of the Earth (!). He repeated the experiment n times, obtaining results y1, y2,..., yn. What can we conclude about!? Observation model Assume that observation noise is zero-mean gaussian with precision ", and that noises are independent given " and! y i = µ + e i slide 12 of 20 e i µ,τ Normal(0,τ) p(y 1,...,y n µ,τ)= n i=1 p(y i µ,τ) precision = 1/variance
12 inferring a mean (2) The prior, p(",#) µ~gamma(10,2) Assume independence: p(µ, τ)=p(µ)p(τ) We can use any pdfs that live in [0,!) It s convenient to use Gamma distributions small " and # give vague prior x p(x) { x } E(x) β Gamma(α,β) µ σ α "(α) xα 1 e βx ( µ) α (0,#) R µ β (λ) λ λx 1 ( #) p(µ) 0 10 granite 2.5 µ! ~ Gamma(2.5,0.1) lead ore 7.5 p(!) slide 13 of 20 note: Matlab s Gamma parameters are A= ", B= 1/# "=25 #= !
13 WinBUGS inferring a mean (3) model { mu ~ dgamma(a_mu,b_mu) tau ~ dgamma(a_tau,b_tau) for (i in 1:n) { y[i] ~ dnorm(mu,tau) } } # data try y1 = 15.36, i.e. outlier try robust distribution y[i] ~ dt(mu,tau,4) list(y=c(5.36,5.29,5.58,5.65,5.57,5.53,5.62,5.29,5.44,5.34, 5.79,5.10,5.27,5.39,5.42,5.47,5.63,5.34,5.46,5.30, 5.78,5.68,5.85),n=23,a_mu=10,b_mu=2,a_tau=2.5,b_tau=0.1) # initialisation list(mu=5,tau=25) slide 14 of 20 results node mean sd 2.5% median 97.5% mu mu sample:
14 comparing means Cuckoo eggs found in m dunnock nests have diameters x1, x2,..., xn (mm). Cuckoo eggs found in n sedge warbler nests have diameters y1, y2,..., yn (mm). Do cuckoos lay bigger eggs in the nests of dunnocks than in the nests of sedge warblers? Observation model p(x 1,...,x n,y 1,...,y n µ x,τ x, µ y,τ y )= n i=1 p(x i µ x,τ x )p(y i µ y,τ y ) x i µ x,τ x Normal(µ x,τ x ), y i µ y,τ y Normal(µ y,τ y ) Prior p(µ x,τ x, µ y,τ y )=p(µ x )p(τ x )p(µ y )p(τ y ) µ x Gamma(0.22,.01), τ x Gamma(0.1,0.1) µ y Gamma(0.22,.01), τ y Gamma(0.1,0.1) plot these with dissttool slide 15 of 20
15 WinBUGS results comparing means (2) model { for(i in 1:m){ x[i] ~ dnorm(mu_x,tau_x) } for(i in 1:n){ y[i] ~ dnorm(mu_y,tau_y) } mu_x ~ dgamma(0.22,0.01) mu_y ~ dgamma(0.22,0.01) tau_x ~ dgamma(0.1,0.1) tau_y ~ dgamma(0.1,0.1) diff <- mu_x - mu_y P <- step(diff) } # data list(x=c(22,23.9,20.9,23.8,25,24,21.7,23.8,22.8,23.1),m=10, y=c(23.2,22,22.2,21.2,21.6,21.9,22,22.9,22.8),n=9) # init list(mu_x=22,mu_y=22,tau_x=1,tau_y=1) do the sizes of cuckoo eggs in dunnock nests have greater variance than those in sedge warbler nests? diff sample: 4500 slide 16 of 20 node mean sd 2.5% median 97.5% diff P
16 linear regression In1875, Scottish physicist James D. Forbes published a study relating data on the boiling temperature of water x1, x2,..., xn (deg F) and the atmospheric pressure y1, y2,..., yn (inches of Hg). If water boils at 190 deg F, what is the atmospheric pressure? Observation model p(y 1,...,y n µ 1,...,µ n,τ)= i y i µ i,τ Normal(µ i,τ) ln(µ i )=α + β(x i x) p(y i µ i,τ) physics predicts a straight line fit to ln(y) as a function of x slide 17 of 20 Prior p(α,β,τ)=p(α)p(β)p(τ) α Normal(0,10 6 ), β Normal(0,10 6 ) τ Gamma(0.001, 0.001)
17 slide 18 of 20 WinBUGS results linear regression (2) model { x_bar <- mean(x[ ]) for ( i in 1 : n ) { log(mu[i]) <- alpha+beta*(x[i]-x_bar) y[i] ~ dnorm(mu[i],tau) } alpha ~ dnorm( 0.0,1.E-6) beta ~ dnorm( 0.0,1.E-6) tau ~ dgamma(0.001,0.001) y190 <- exp(alpha+beta*(190-x_bar)) } # data model fit: mu list( x=c(210.8, 210.2, 208.4, 202.5, 200.6, 200.1, 199.5, 197, 196.4, 196.3, 195.6, 193.4, 193.6, 191.4, 191.1, 190.6, 189.5, 188.8, 188.5, 185.7, 186, 185.6, 184.1, 184.6, 184.1, 183.2, 182.4, 181.9, 181.9, , 180.6), P=c(29.211, , , , , , 23.03, , , , , 20.48, , , 19.49, , , , , , , , , , , , , , , , ), n=31) # init list(alpha=0,beta=0,tau=1) try y[i] ~ dt(mu[i],tau,4) node mean sd 2.5% median 97.5% y Inference > Compare node = mu other = y axis = x
18 logistic regression n1 lab mice are injected with a substance at log concentration x1, and y1 of them die. The experiment is repeated 4 times with different concentrations, yielding further data (n2, x2, y2), (n3, x3, y3), (n4, x4, y4). What dosage corresponds to a 50% chance of mortality? Observation model y i θ i Binomial(θ i,n i ) 1 1/2! slide 19 of 20 Prior logit(θ i ) log θ i 1 θ i α Normal(0,.001), = α + βx i 0 p(α,β)=p(α)p(β) β Normal(0,.001)!"/# plot prior with dissttool (note: Matlab s parametrization of Normal differs from WinBUGS usage) x
19 WinBUGS logistic regression (2) model { for (i in 1:nx) { logit(theta[i]) <- alpha + beta*x[i] y[i] ~ dbin(theta[i],n[i]) } alpha ~ dnorm(0.0,0.001) beta ~ dnorm(0.0,0.001) LD50 <- (logit(0.50)-alpha)/beta } # data list(y=c(0,1,3,5), n=c(5,5,5,5), x=c(-0.863,-0.296,-0.053,0.727), nx=4) # init list(alpha=0,beta=1) what log-concentration corresponds to a mortality probability of 1%? x i n i y i slide 20 of 20 results node mean sd 2.5% median 97.5% alpha beta LD y/n 0!1 0 1 x
1 Prior Probability and Posterior Probability
Math 541: Statistical Theory II Bayesian Approach to Parameter Estimation Lecturer: Songfeng Zheng 1 Prior Probability and Posterior Probability Consider now a problem of statistical inference in which
More informationPart 2: One-parameter models
Part 2: One-parameter models Bernoilli/binomial models Return to iid Y 1,...,Y n Bin(1, θ). The sampling model/likelihood is p(y 1,...,y n θ) =θ P y i (1 θ) n P y i When combined with a prior p(θ), Bayes
More informationModel-based Synthesis. Tony O Hagan
Model-based Synthesis Tony O Hagan Stochastic models Synthesising evidence through a statistical model 2 Evidence Synthesis (Session 3), Helsinki, 28/10/11 Graphical modelling The kinds of models that
More informationCHAPTER 2 Estimating Probabilities
CHAPTER 2 Estimating Probabilities Machine Learning Copyright c 2016. Tom M. Mitchell. All rights reserved. *DRAFT OF January 24, 2016* *PLEASE DO NOT DISTRIBUTE WITHOUT AUTHOR S PERMISSION* This is a
More informationPS 271B: Quantitative Methods II. Lecture Notes
PS 271B: Quantitative Methods II Lecture Notes Langche Zeng zeng@ucsd.edu The Empirical Research Process; Fundamental Methodological Issues 2 Theory; Data; Models/model selection; Estimation; Inference.
More informationSimple Linear Regression Inference
Simple Linear Regression Inference 1 Inference requirements The Normality assumption of the stochastic term e is needed for inference even if it is not a OLS requirement. Therefore we have: Interpretation
More informationBasics of Statistical Machine Learning
CS761 Spring 2013 Advanced Machine Learning Basics of Statistical Machine Learning Lecturer: Xiaojin Zhu jerryzhu@cs.wisc.edu Modern machine learning is rooted in statistics. You will find many familiar
More informationA Latent Variable Approach to Validate Credit Rating Systems using R
A Latent Variable Approach to Validate Credit Rating Systems using R Chicago, April 24, 2009 Bettina Grün a, Paul Hofmarcher a, Kurt Hornik a, Christoph Leitner a, Stefan Pichler a a WU Wien Grün/Hofmarcher/Hornik/Leitner/Pichler
More informationWhat is Statistics? Lecture 1. Introduction and probability review. Idea of parametric inference
0. 1. Introduction and probability review 1.1. What is Statistics? What is Statistics? Lecture 1. Introduction and probability review There are many definitions: I will use A set of principle and procedures
More informationBNG 202 Biomechanics Lab. Descriptive statistics and probability distributions I
BNG 202 Biomechanics Lab Descriptive statistics and probability distributions I Overview The overall goal of this short course in statistics is to provide an introduction to descriptive and inferential
More informationLinear regression methods for large n and streaming data
Linear regression methods for large n and streaming data Large n and small or moderate p is a fairly simple problem. The sufficient statistic for β in OLS (and ridge) is: The concept of sufficiency is
More informationNonparametric adaptive age replacement with a one-cycle criterion
Nonparametric adaptive age replacement with a one-cycle criterion P. Coolen-Schrijner, F.P.A. Coolen Department of Mathematical Sciences University of Durham, Durham, DH1 3LE, UK e-mail: Pauline.Schrijner@durham.ac.uk
More informationSampling via Moment Sharing: A New Framework for Distributed Bayesian Inference for Big Data
Sampling via Moment Sharing: A New Framework for Distributed Bayesian Inference for Big Data (Oxford) in collaboration with: Minjie Xu, Jun Zhu, Bo Zhang (Tsinghua) Balaji Lakshminarayanan (Gatsby) Bayesian
More informationStatistical Machine Learning
Statistical Machine Learning UoC Stats 37700, Winter quarter Lecture 4: classical linear and quadratic discriminants. 1 / 25 Linear separation For two classes in R d : simple idea: separate the classes
More informationThe Proportional Odds Model for Assessing Rater Agreement with Multiple Modalities
The Proportional Odds Model for Assessing Rater Agreement with Multiple Modalities Elizabeth Garrett-Mayer, PhD Assistant Professor Sidney Kimmel Comprehensive Cancer Center Johns Hopkins University 1
More informationChapter 11: r.m.s. error for regression
Chapter 11: r.m.s. error for regression Context................................................................... 2 Prediction error 3 r.m.s. error for the regression line...............................................
More informationUnivariate Regression
Univariate Regression Correlation and Regression The regression line summarizes the linear relationship between 2 variables Correlation coefficient, r, measures strength of relationship: the closer r is
More informationBayesian Analysis for the Social Sciences
Bayesian Analysis for the Social Sciences Simon Jackman Stanford University http://jackman.stanford.edu/bass November 9, 2012 Simon Jackman (Stanford) Bayesian Analysis for the Social Sciences November
More informationChristfried Webers. Canberra February June 2015
c Statistical Group and College of Engineering and Computer Science Canberra February June (Many figures from C. M. Bishop, "Pattern Recognition and ") 1of 829 c Part VIII Linear Classification 2 Logistic
More informationLikelihood: Frequentist vs Bayesian Reasoning
"PRINCIPLES OF PHYLOGENETICS: ECOLOGY AND EVOLUTION" Integrative Biology 200B University of California, Berkeley Spring 2009 N Hallinan Likelihood: Frequentist vs Bayesian Reasoning Stochastic odels and
More informationApplications of R Software in Bayesian Data Analysis
Article International Journal of Information Science and System, 2012, 1(1): 7-23 International Journal of Information Science and System Journal homepage: www.modernscientificpress.com/journals/ijinfosci.aspx
More informationPattern Analysis. Logistic Regression. 12. Mai 2009. Joachim Hornegger. Chair of Pattern Recognition Erlangen University
Pattern Analysis Logistic Regression 12. Mai 2009 Joachim Hornegger Chair of Pattern Recognition Erlangen University Pattern Analysis 2 / 43 1 Logistic Regression Posteriors and the Logistic Function Decision
More informationα α λ α = = λ λ α ψ = = α α α λ λ ψ α = + β = > θ θ β > β β θ θ θ β θ β γ θ β = γ θ > β > γ θ β γ = θ β = θ β = θ β = β θ = β β θ = = = β β θ = + α α α α α = = λ λ λ λ λ λ λ = λ λ α α α α λ ψ + α =
More informationUsing the Delta Method to Construct Confidence Intervals for Predicted Probabilities, Rates, and Discrete Changes
Using the Delta Method to Construct Confidence Intervals for Predicted Probabilities, Rates, Discrete Changes JunXuJ.ScottLong Indiana University August 22, 2005 The paper provides technical details on
More informationNote on the EM Algorithm in Linear Regression Model
International Mathematical Forum 4 2009 no. 38 1883-1889 Note on the M Algorithm in Linear Regression Model Ji-Xia Wang and Yu Miao College of Mathematics and Information Science Henan Normal University
More informationA Primer on Mathematical Statistics and Univariate Distributions; The Normal Distribution; The GLM with the Normal Distribution
A Primer on Mathematical Statistics and Univariate Distributions; The Normal Distribution; The GLM with the Normal Distribution PSYC 943 (930): Fundamentals of Multivariate Modeling Lecture 4: September
More informationModeling and Analysis of Call Center Arrival Data: A Bayesian Approach
Modeling and Analysis of Call Center Arrival Data: A Bayesian Approach Refik Soyer * Department of Management Science The George Washington University M. Murat Tarimcilar Department of Management Science
More informationI L L I N O I S UNIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN
Beckman HLM Reading Group: Questions, Answers and Examples Carolyn J. Anderson Department of Educational Psychology I L L I N O I S UNIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN Linear Algebra Slide 1 of
More informationChapter 13 Introduction to Nonlinear Regression( 非 線 性 迴 歸 )
Chapter 13 Introduction to Nonlinear Regression( 非 線 性 迴 歸 ) and Neural Networks( 類 神 經 網 路 ) 許 湘 伶 Applied Linear Regression Models (Kutner, Nachtsheim, Neter, Li) hsuhl (NUK) LR Chap 10 1 / 35 13 Examples
More informationAn Introduction to Using WinBUGS for Cost-Effectiveness Analyses in Health Economics
Slide 1 An Introduction to Using WinBUGS for Cost-Effectiveness Analyses in Health Economics Dr. Christian Asseburg Centre for Health Economics Part 1 Slide 2 Talk overview Foundations of Bayesian statistics
More informationMultinomial and Ordinal Logistic Regression
Multinomial and Ordinal Logistic Regression ME104: Linear Regression Analysis Kenneth Benoit August 22, 2012 Regression with categorical dependent variables When the dependent variable is categorical,
More informationMATH4427 Notebook 2 Spring 2016. 2 MATH4427 Notebook 2 3. 2.1 Definitions and Examples... 3. 2.2 Performance Measures for Estimators...
MATH4427 Notebook 2 Spring 2016 prepared by Professor Jenny Baglivo c Copyright 2009-2016 by Jenny A. Baglivo. All Rights Reserved. Contents 2 MATH4427 Notebook 2 3 2.1 Definitions and Examples...................................
More informationLinear Discrimination. Linear Discrimination. Linear Discrimination. Linearly Separable Systems Pairwise Separation. Steven J Zeil.
Steven J Zeil Old Dominion Univ. Fall 200 Discriminant-Based Classification Linearly Separable Systems Pairwise Separation 2 Posteriors 3 Logistic Discrimination 2 Discriminant-Based Classification Likelihood-based:
More informationSOCIETY OF ACTUARIES/CASUALTY ACTUARIAL SOCIETY EXAM C CONSTRUCTION AND EVALUATION OF ACTUARIAL MODELS EXAM C SAMPLE QUESTIONS
SOCIETY OF ACTUARIES/CASUALTY ACTUARIAL SOCIETY EXAM C CONSTRUCTION AND EVALUATION OF ACTUARIAL MODELS EXAM C SAMPLE QUESTIONS Copyright 005 by the Society of Actuaries and the Casualty Actuarial Society
More informationSTT315 Chapter 4 Random Variables & Probability Distributions KM. Chapter 4.5, 6, 8 Probability Distributions for Continuous Random Variables
Chapter 4.5, 6, 8 Probability Distributions for Continuous Random Variables Discrete vs. continuous random variables Examples of continuous distributions o Uniform o Exponential o Normal Recall: A random
More informationStatistics in Medicine Research Lecture Series CSMC Fall 2014
Catherine Bresee, MS Senior Biostatistician Biostatistics & Bioinformatics Research Institute Statistics in Medicine Research Lecture Series CSMC Fall 2014 Overview Review concept of statistical power
More informationMultivariate Normal Distribution
Multivariate Normal Distribution Lecture 4 July 21, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2 Lecture #4-7/21/2011 Slide 1 of 41 Last Time Matrices and vectors Eigenvalues
More informationUNIT I: RANDOM VARIABLES PART- A -TWO MARKS
UNIT I: RANDOM VARIABLES PART- A -TWO MARKS 1. Given the probability density function of a continuous random variable X as follows f(x) = 6x (1-x) 0
More informationAnalyzing Clinical Trial Data via the Bayesian Multiple Logistic Random Effects Model
Analyzing Clinical Trial Data via the Bayesian Multiple Logistic Random Effects Model Bartolucci, A.A 1, Singh, K.P 2 and Bae, S.J 2 1 Dept. of Biostatistics, University of Alabama at Birmingham, Birmingham,
More informationBayesian inference for population prediction of individuals without health insurance in Florida
Bayesian inference for population prediction of individuals without health insurance in Florida Neung Soo Ha 1 1 NISS 1 / 24 Outline Motivation Description of the Behavioral Risk Factor Surveillance System,
More informationPackage SHELF. February 5, 2016
Type Package Package SHELF February 5, 2016 Title Tools to Support the Sheffield Elicitation Framework (SHELF) Version 1.1.0 Date 2016-01-29 Author Jeremy Oakley Maintainer Jeremy Oakley
More informationInterpretation of Somers D under four simple models
Interpretation of Somers D under four simple models Roger B. Newson 03 September, 04 Introduction Somers D is an ordinal measure of association introduced by Somers (96)[9]. It can be defined in terms
More informationLinear Threshold Units
Linear Threshold Units w x hx (... w n x n w We assume that each feature x j and each weight w j is a real number (we will relax this later) We will study three different algorithms for learning linear
More informationErrata and updates for ASM Exam C/Exam 4 Manual (Sixteenth Edition) sorted by page
Errata for ASM Exam C/4 Study Manual (Sixteenth Edition) Sorted by Page 1 Errata and updates for ASM Exam C/Exam 4 Manual (Sixteenth Edition) sorted by page Practice exam 1:9, 1:22, 1:29, 9:5, and 10:8
More informationDongfeng Li. Autumn 2010
Autumn 2010 Chapter Contents Some statistics background; ; Comparing means and proportions; variance. Students should master the basic concepts, descriptive statistics measures and graphs, basic hypothesis
More informationLogistic Regression. Jia Li. Department of Statistics The Pennsylvania State University. Logistic Regression
Logistic Regression Department of Statistics The Pennsylvania State University Email: jiali@stat.psu.edu Logistic Regression Preserve linear classification boundaries. By the Bayes rule: Ĝ(x) = arg max
More informationNeural Networks Lesson 5 - Cluster Analysis
Neural Networks Lesson 5 - Cluster Analysis Prof. Michele Scarpiniti INFOCOM Dpt. - Sapienza University of Rome http://ispac.ing.uniroma1.it/scarpiniti/index.htm michele.scarpiniti@uniroma1.it Rome, 29
More informationComparison of frequentist and Bayesian inference. Class 20, 18.05, Spring 2014 Jeremy Orloff and Jonathan Bloom
Comparison of frequentist and Bayesian inference. Class 20, 18.05, Spring 2014 Jeremy Orloff and Jonathan Bloom 1 Learning Goals 1. Be able to explain the difference between the p-value and a posterior
More informationIntroduction to Mobile Robotics Bayes Filter Particle Filter and Monte Carlo Localization
Introduction to Mobile Robotics Bayes Filter Particle Filter and Monte Carlo Localization Wolfram Burgard, Maren Bennewitz, Diego Tipaldi, Luciano Spinello 1 Motivation Recall: Discrete filter Discretize
More informationAdaptive Design for Intra Patient Dose Escalation in Phase I Trials in Oncology
Adaptive Design for Intra Patient Dose Escalation in Phase I Trials in Oncology Jeremy M.G. Taylor Laura L. Fernandes University of Michigan, Ann Arbor 19th August, 2011 J.M.G. Taylor, L.L. Fernandes Adaptive
More informationBayesian Hierarchical Models: Practical Exercises
1 Bayesian Hierarchical Models: Practical Exercises You will be using WinBUGS 1.4.3 for these practicals. All the data and other files you will need for the practicals are provided in a zip file, which
More informationParallelization Strategies for Multicore Data Analysis
Parallelization Strategies for Multicore Data Analysis Wei-Chen Chen 1 Russell Zaretzki 2 1 University of Tennessee, Dept of EEB 2 University of Tennessee, Dept. Statistics, Operations, and Management
More informationLesson 1: Comparison of Population Means Part c: Comparison of Two- Means
Lesson : Comparison of Population Means Part c: Comparison of Two- Means Welcome to lesson c. This third lesson of lesson will discuss hypothesis testing for two independent means. Steps in Hypothesis
More informationPrinciple of Data Reduction
Chapter 6 Principle of Data Reduction 6.1 Introduction An experimenter uses the information in a sample X 1,..., X n to make inferences about an unknown parameter θ. If the sample size n is large, then
More informationBayesian Statistics in One Hour. Patrick Lam
Bayesian Statistics in One Hour Patrick Lam Outline Introduction Bayesian Models Applications Missing Data Hierarchical Models Outline Introduction Bayesian Models Applications Missing Data Hierarchical
More informationCSCI567 Machine Learning (Fall 2014)
CSCI567 Machine Learning (Fall 2014) Drs. Sha & Liu {feisha,yanliu.cs}@usc.edu September 22, 2014 Drs. Sha & Liu ({feisha,yanliu.cs}@usc.edu) CSCI567 Machine Learning (Fall 2014) September 22, 2014 1 /
More informationPredictive Modeling Techniques in Insurance
Predictive Modeling Techniques in Insurance Tuesday May 5, 2015 JF. Breton Application Engineer 2014 The MathWorks, Inc. 1 Opening Presenter: JF. Breton: 13 years of experience in predictive analytics
More informationThese slides follow closely the (English) course textbook Pattern Recognition and Machine Learning by Christopher Bishop
Music and Machine Learning (IFT6080 Winter 08) Prof. Douglas Eck, Université de Montréal These slides follow closely the (English) course textbook Pattern Recognition and Machine Learning by Christopher
More informationHYPOTHESIS TESTING: CONFIDENCE INTERVALS, T-TESTS, ANOVAS, AND REGRESSION
HYPOTHESIS TESTING: CONFIDENCE INTERVALS, T-TESTS, ANOVAS, AND REGRESSION HOD 2990 10 November 2010 Lecture Background This is a lightning speed summary of introductory statistical methods for senior undergraduate
More informationInstitute of Actuaries of India Subject CT3 Probability and Mathematical Statistics
Institute of Actuaries of India Subject CT3 Probability and Mathematical Statistics For 2015 Examinations Aim The aim of the Probability and Mathematical Statistics subject is to provide a grounding in
More information10-601. Machine Learning. http://www.cs.cmu.edu/afs/cs/academic/class/10601-f10/index.html
10-601 Machine Learning http://www.cs.cmu.edu/afs/cs/academic/class/10601-f10/index.html Course data All up-to-date info is on the course web page: http://www.cs.cmu.edu/afs/cs/academic/class/10601-f10/index.html
More informationFor a partition B 1,..., B n, where B i B j = for i. A = (A B 1 ) (A B 2 ),..., (A B n ) and thus. P (A) = P (A B i ) = P (A B i )P (B i )
Probability Review 15.075 Cynthia Rudin A probability space, defined by Kolmogorov (1903-1987) consists of: A set of outcomes S, e.g., for the roll of a die, S = {1, 2, 3, 4, 5, 6}, 1 1 2 1 6 for the roll
More informationNormal Distribution. Definition A continuous random variable has a normal distribution if its probability density. f ( y ) = 1.
Normal Distribution Definition A continuous random variable has a normal distribution if its probability density e -(y -µ Y ) 2 2 / 2 σ function can be written as for < y < as Y f ( y ) = 1 σ Y 2 π Notation:
More informationCCNY. BME I5100: Biomedical Signal Processing. Linear Discrimination. Lucas C. Parra Biomedical Engineering Department City College of New York
BME I5100: Biomedical Signal Processing Linear Discrimination Lucas C. Parra Biomedical Engineering Department CCNY 1 Schedule Week 1: Introduction Linear, stationary, normal - the stuff biology is not
More informationExamining credit card consumption pattern
Examining credit card consumption pattern Yuhao Fan (Economics Department, Washington University in St. Louis) Abstract: In this paper, I analyze the consumer s credit card consumption data from a commercial
More informationAuxiliary Variables in Mixture Modeling: 3-Step Approaches Using Mplus
Auxiliary Variables in Mixture Modeling: 3-Step Approaches Using Mplus Tihomir Asparouhov and Bengt Muthén Mplus Web Notes: No. 15 Version 8, August 5, 2014 1 Abstract This paper discusses alternatives
More informationProbability for Estimation (review)
Probability for Estimation (review) In general, we want to develop an estimator for systems of the form: x = f x, u + η(t); y = h x + ω(t); ggggg y, ffff x We will primarily focus on discrete time linear
More informationExample: Credit card default, we may be more interested in predicting the probabilty of a default than classifying individuals as default or not.
Statistical Learning: Chapter 4 Classification 4.1 Introduction Supervised learning with a categorical (Qualitative) response Notation: - Feature vector X, - qualitative response Y, taking values in C
More informationCHAPTER 6: Continuous Uniform Distribution: 6.1. Definition: The density function of the continuous random variable X on the interval [A, B] is.
Some Continuous Probability Distributions CHAPTER 6: Continuous Uniform Distribution: 6. Definition: The density function of the continuous random variable X on the interval [A, B] is B A A x B f(x; A,
More informationCS 688 Pattern Recognition Lecture 4. Linear Models for Classification
CS 688 Pattern Recognition Lecture 4 Linear Models for Classification Probabilistic generative models Probabilistic discriminative models 1 Generative Approach ( x ) p C k p( C k ) Ck p ( ) ( x Ck ) p(
More informationGenerating Random Numbers Variance Reduction Quasi-Monte Carlo. Simulation Methods. Leonid Kogan. MIT, Sloan. 15.450, Fall 2010
Simulation Methods Leonid Kogan MIT, Sloan 15.450, Fall 2010 c Leonid Kogan ( MIT, Sloan ) Simulation Methods 15.450, Fall 2010 1 / 35 Outline 1 Generating Random Numbers 2 Variance Reduction 3 Quasi-Monte
More informationstatistical learning; Bayesian learning; stochastic optimization; dynamic programming
INFORMS 2008 c 2008 INFORMS isbn 978-1-877640-23-0 doi 10.1287/educ.1080.0039 Optimal Learning Warren B. Powell and Peter Frazier Department of Operations Research and Financial Engineering, Princeton
More informationDetermining distribution parameters from quantiles
Determining distribution parameters from quantiles John D. Cook Department of Biostatistics The University of Texas M. D. Anderson Cancer Center P. O. Box 301402 Unit 1409 Houston, TX 77230-1402 USA cook@mderson.org
More informationPa8ern Recogni6on. and Machine Learning. Chapter 4: Linear Models for Classifica6on
Pa8ern Recogni6on and Machine Learning Chapter 4: Linear Models for Classifica6on Represen'ng the target values for classifica'on If there are only two classes, we typically use a single real valued output
More informationE3: PROBABILITY AND STATISTICS lecture notes
E3: PROBABILITY AND STATISTICS lecture notes 2 Contents 1 PROBABILITY THEORY 7 1.1 Experiments and random events............................ 7 1.2 Certain event. Impossible event............................
More informationInference of Probability Distributions for Trust and Security applications
Inference of Probability Distributions for Trust and Security applications Vladimiro Sassone Based on joint work with Mogens Nielsen & Catuscia Palamidessi Outline 2 Outline Motivations 2 Outline Motivations
More informationCentre for Central Banking Studies
Centre for Central Banking Studies Technical Handbook No. 4 Applied Bayesian econometrics for central bankers Andrew Blake and Haroon Mumtaz CCBS Technical Handbook No. 4 Applied Bayesian econometrics
More informationCourse 4 Examination Questions And Illustrative Solutions. November 2000
Course 4 Examination Questions And Illustrative Solutions Novemer 000 1. You fit an invertile first-order moving average model to a time series. The lag-one sample autocorrelation coefficient is 0.35.
More informationNonlinear Regression:
Zurich University of Applied Sciences School of Engineering IDP Institute of Data Analysis and Process Design Nonlinear Regression: A Powerful Tool With Considerable Complexity Half-Day : Improved Inference
More informationMaximum Likelihood Estimation
Math 541: Statistical Theory II Lecturer: Songfeng Zheng Maximum Likelihood Estimation 1 Maximum Likelihood Estimation Maximum likelihood is a relatively simple method of constructing an estimator for
More informationStatistical modelling with missing data using multiple imputation. Session 4: Sensitivity Analysis after Multiple Imputation
Statistical modelling with missing data using multiple imputation Session 4: Sensitivity Analysis after Multiple Imputation James Carpenter London School of Hygiene & Tropical Medicine Email: james.carpenter@lshtm.ac.uk
More informationProbabilistic Methods for Time-Series Analysis
Probabilistic Methods for Time-Series Analysis 2 Contents 1 Analysis of Changepoint Models 1 1.1 Introduction................................ 1 1.1.1 Model and Notation....................... 2 1.1.2 Example:
More informationSpatial Statistics Chapter 3 Basics of areal data and areal data modeling
Spatial Statistics Chapter 3 Basics of areal data and areal data modeling Recall areal data also known as lattice data are data Y (s), s D where D is a discrete index set. This usually corresponds to data
More informationA General Approach to Variance Estimation under Imputation for Missing Survey Data
A General Approach to Variance Estimation under Imputation for Missing Survey Data J.N.K. Rao Carleton University Ottawa, Canada 1 2 1 Joint work with J.K. Kim at Iowa State University. 2 Workshop on Survey
More informationMonte Carlo and Empirical Methods for Stochastic Inference (MASM11/FMS091)
Monte Carlo and Empirical Methods for Stochastic Inference (MASM11/FMS091) Magnus Wiktorsson Centre for Mathematical Sciences Lund University, Sweden Lecture 5 Sequential Monte Carlo methods I February
More informationLinear Models for Classification
Linear Models for Classification Sumeet Agarwal, EEL709 (Most figures from Bishop, PRML) Approaches to classification Discriminant function: Directly assigns each data point x to a particular class Ci
More informationBayesian Updating with Discrete Priors Class 11, 18.05, Spring 2014 Jeremy Orloff and Jonathan Bloom
1 Learning Goals Bayesian Updating with Discrete Priors Class 11, 18.05, Spring 2014 Jeremy Orloff and Jonathan Bloom 1. Be able to apply Bayes theorem to compute probabilities. 2. Be able to identify
More informationRecent Developments of Statistical Application in. Finance. Ruey S. Tsay. Graduate School of Business. The University of Chicago
Recent Developments of Statistical Application in Finance Ruey S. Tsay Graduate School of Business The University of Chicago Guanghua Conference, June 2004 Summary Focus on two parts: Applications in Finance:
More informationData Modeling & Analysis Techniques. Probability & Statistics. Manfred Huber 2011 1
Data Modeling & Analysis Techniques Probability & Statistics Manfred Huber 2011 1 Probability and Statistics Probability and statistics are often used interchangeably but are different, related fields
More informationCell Phone based Activity Detection using Markov Logic Network
Cell Phone based Activity Detection using Markov Logic Network Somdeb Sarkhel sxs104721@utdallas.edu 1 Introduction Mobile devices are becoming increasingly sophisticated and the latest generation of smart
More informationLecture 3: Linear methods for classification
Lecture 3: Linear methods for classification Rafael A. Irizarry and Hector Corrada Bravo February, 2010 Today we describe four specific algorithms useful for classification problems: linear regression,
More informationMonte Carlo-based statistical methods (MASM11/FMS091)
Monte Carlo-based statistical methods (MASM11/FMS091) Magnus Wiktorsson Centre for Mathematical Sciences Lund University, Sweden Lecture 6 Sequential Monte Carlo methods II February 7, 2014 M. Wiktorsson
More informationShort title: Measurement error in binary regression. T. Fearn 1, D.C. Hill 2 and S.C. Darby 2. of Oxford, Oxford, U.K.
Measurement error in the explanatory variable of a binary regression: regression calibration and integrated conditional likelihood in studies of residential radon and lung cancer Short title: Measurement
More informationA Statistical Framework for Operational Infrasound Monitoring
A Statistical Framework for Operational Infrasound Monitoring Stephen J. Arrowsmith Rod W. Whitaker LA-UR 11-03040 The views expressed here do not necessarily reflect the views of the United States Government,
More informationAn Introduction to Machine Learning
An Introduction to Machine Learning L5: Novelty Detection and Regression Alexander J. Smola Statistical Machine Learning Program Canberra, ACT 0200 Australia Alex.Smola@nicta.com.au Tata Institute, Pune,
More informationA Basic Introduction to Missing Data
John Fox Sociology 740 Winter 2014 Outline Why Missing Data Arise Why Missing Data Arise Global or unit non-response. In a survey, certain respondents may be unreachable or may refuse to participate. Item
More informationMore details on the inputs, functionality, and output can be found below.
Overview: The SMEEACT (Software for More Efficient, Ethical, and Affordable Clinical Trials) web interface (http://research.mdacc.tmc.edu/smeeactweb) implements a single analysis of a two-armed trial comparing
More informationAn Introduction to Basic Statistics and Probability
An Introduction to Basic Statistics and Probability Shenek Heyward NCSU An Introduction to Basic Statistics and Probability p. 1/4 Outline Basic probability concepts Conditional probability Discrete Random
More informationCS 2750 Machine Learning. Lecture 1. Machine Learning. http://www.cs.pitt.edu/~milos/courses/cs2750/ CS 2750 Machine Learning.
Lecture Machine Learning Milos Hauskrecht milos@cs.pitt.edu 539 Sennott Square, x5 http://www.cs.pitt.edu/~milos/courses/cs75/ Administration Instructor: Milos Hauskrecht milos@cs.pitt.edu 539 Sennott
More information