TIME SERIES ANALYSIS OF COMPOSITIONAL DATA USING A DYNAMIC LINEAR MODEL APPROACH
|
|
|
- Thomasine Hart
- 10 years ago
- Views:
Transcription
1 Joint Stascal Meengs - Bayesian Stascal Science TIME SERIES ANALYSIS OF COMPOSITIONAL DATA USING A DYNAMIC LINEAR MODEL APPROACH AMITABHA BHAUMIK, DIPAK K. DEY AND NALINI RAVISHANKER Department of Stascs, University of Conneccut, U-4120, 215 Glenbrook Road, Storrs, CT 06269, Abstract. Composional me series data comprises of mulvariate observaons that at each me point are essenally proporons of a whole quanty. This kind of data occurs frequently in many disciplines such as economics, geology and ecology. Usual mulvariate stascal procedures available in the literature are not applicable for the analysis of such data since they ignore the inherent constrained nature of these observaons as parts of a whole. This arcle describes new techniques for modeling composional me series data in a hierarchical Bayesian framework. Modified dynamic linear models are fit to composional data via Markov chain Monte Carlo techniques. The distribuon of the underlying errors is assumed to be a scale mixture of mulvariate normals of which the mulvariate normal, mulvariate t, mulvariate logisc, etc., are special cases. In parcular, mulvariate normal and Student-t error structures are considered and compared through predicve distribuons. The approach is illustrated on a data set. Keywords. Box-Cox transformaon; Composional data; Hierarchical Bayesian model; Scale mixture of normals distribuon; Student-t distribuon. 1. INTRODUCTION Composional me series occur in many applicaon areas including geology, manufacturing design, ecology, biology, etc. It is characterized by components which are posive and sum to one at each me point. In many cases, such composions also change over me. In general, a G-variate composional me series data consists of a G-dimensional vector of non-negave components W t1,..., W tg such that W t W tg = 1, for each t (Aitchison, 1982). Because of this unit sum constraint at each me t, the composion can be completely defined by any of the g = G 1 components so that the vector W t = (W t1,..., W tg ) lies in a g-dimensional simplex which is defined as S g t = {(W t1,..., W tg ) : W t1 0,..., W tg 0; W t W tg = 1}. (1) Stascal analysis of such constrained mulvariate me series in the presence of covariate informaon is very useful in pracce. However, the biggest hindrance to the analysis of composional data arises from the two inherent constraints menoned above. Standard techniques such as regression with vector autoregressive moving average (VARMA) errors, or Kalman filtering (Kalman, 1969), etc., are not directly applicable to raw composional me series data. To circumvent this problem, Aitchison s (1982) proposal was to transform data from the g- dimensional simplex to the g-dimensional real space R g, and to employ an addive error model. The most popular transformaon is the Addive Log Rao (ALR) transformaon, which is of the form Y = log(w /W tg ), i = 1,..., g. Another possibility is the Box-Cox transformaon (Box and Cox, 1964) which was used for cross-seconal composional data by Rayens and Srinivasan (1991a, b) in a frequenst setup and by Iyenger and Dey (1998) in a Bayesian framework. In both cases, the transformed variable may be assumed to follow a mulvariate normal distribuon. The Box-Cox transformaon is more general in the sense that the ALR transformaon is a special case. Literature is sparse on composional analysis of mulvariate me series incorporang covariate informaon. Quintana and West (1988) used an ALR transformaon with mulvariate dynamic linear models. A state space model based on the Dirichlet distribuon was developed by Grunwald, Raftery and Guttorp (1993). In this arcle, we analyze composional me series data after a Box- Cox transformaon. Inference is carried out in the Bayesian framework using a very rich class of scale mixture of mulvariate normals (SMMVN) for modeling the errors. This class of distribuons includes the mulvariate normal, Student-t, logisc, stable etc. (see Dey and Chen, 1998). In this arcle, we invesgate the performance of the mulvariate normal 226
2 Joint Stascal Meengs - Bayesian Stascal Science and mulvariate Student-t distribuons in parcular. Model performance is assessed via predicve distribuons; the Condional Predicve Ordinate (CPO) is computed and is used for model assessment. The format of the paper is as follows. In secon 2 we discuss the proposed model. Predicve distribuons and their Monte Carlo esmates are also discussed. Secon 3 presents an illustraon. It consists of composions of vehicles produced by, and other countries over the me period Secon 4 contains some concluding remarks. 2. A MODEL FOR TRANSFORMED COMPOSITIONS Let W t denote the G-dimensional composion at me t as defined in secon 1. We assume that W t depends on some unknown state of nature η t = (η t1,..., η tg ), a vector of true proporons. Due to the unit sum constraint, W t s are points on a g = G 1 dimensional simplex. The Box-Cox transformaon of W t transforms the vector to Y t in R g : ( W W ) λ 1 tg λ if λ 0 Y = (2) log( W W tg ) if λ = 0, where λ R is an unknown parameter known as the Box-Cox parameter, t = 1,..., T, i = 1,..., g and g = G 1. This transformaon is denoted in this paper as Y t = BC(W t, λ t ), as a special case of which we get ALR transformaon when λ = 0 for all i, t The Dynamic Linear Model Let X t denote the l -dimensional covariate vector at me t. The linear regression model for the g- dimensional me series Y t has the form Y t λ t = α t + X t β t + e t (3) where the unknown parameters α t and β t are g- dimensional vectors, λ t denotes the vector of Box- Cox parameters, and e t is the random error. This model is invesgated under the scale mixture of mulvariate normals (SMMVN) error distribuon. Suppose e t is normally distributed with mean 0 and unknown variance k(τ)v t, where k(τ) is a posive funcon of the mixing parameter τ, where τ has pdf g(τ). When k(τ) = τ, and τ has a degenerate distribuon at 1, it is well known that e t has a mulvariate normal distribuon with mean 0 and variance V t. When k(τ) = 1/τ, and the mixing density g(τ) is Gamma with both locaon and scale parameter equal to ν/2, the marginal distribuon of e t becomes Student-t, with locaon parameter 0, scale parameter τv t and degrees of freedom ν. While the mulvariate normal error model is simple to fit, the mulvariate Student-t distribuon handles extreme observaons better. The observaon equaon can be rewritten as Y t λ t = Z t η t + e t, (4) where Z t = (1 X t ) is a g l vector of covariates at me t and η t = (α t, β t ) is an l-dimensional vector, for l = l + 1. The essenal difference between the dynamic linear model and the stac linear model is that in the former case, the regression coefficients are not assumed to be constant, but may change with me. This dynamic feature is incorporated via the system equaon η t = Gη t 1 + w t, (5) where G is a known matrix, and we assume that p(w t W t ) is normal, i.e., w t N(0, W t ). (6) It is assumed that the unknown, me-dependent Box-Cox parameter follows the system equaon where p(u t U t ) is normal, i.e., λ t = Hλ t 1 + u t, (7) u t N(0, U t ). (8) Under this modeling framework, the system equaon and the equaon on the Box-Cox parameter can be reparametrized as η t = λ t = G t i w i (9) H t i u i, (10) where G t i and H t i are the t i th powers of G and H respecvely. Rewring the observaon equaon using these transformed parametric expressions, the model has the form 227
3 Joint Stascal Meengs - Bayesian Stascal Science Y t λ t = Z t G t i w i + e t (11) λ t = H t i u i, (12) where Y t λ t = BC(W t, λ t ). In addion to the usual linear model assumpons regarding the error terms, we also postulate that e t and u t are independent of w t. Let µ t = Z t t Gt i w i, Y=(Y 1,..., Y T ), V=(V 1,..., V T ), W=(W 1,..., W T ) and U=(U 1,..., U T ) Posterior Distribuon and Model Fit The exact likelihood has the form L(Y ; V, W, U, τ ) exp[ 1 2k(τ).., u t ) µ t ) V 1 t exp[ 1 2 u t T (BC(W t, u 1,... (BC(W t, u 1,..., u t ) µ t )] g(τ) U t 1 u t ] exp[ 1 2 w t W 1 t w t ] V t 1 2 U t 1 2 W t 1 2. In the special case when the mixing parameter has a Gamma distribuon with both the locaon and scale parameters ν/2 for known ν, the likelihood funcon becomes, L(Y ; V, W, U) [1 + (BC(W t, u 1, u 2,..., u t ) µ t ) V 1 t (BC(W t, u 1, u 2,..., u t ) µ t )/v] [(v+g)/2] exp[ 1 2 u t U t 1 u t ] exp[ 1 2 w t W t 1 w t ] V t 1 2 U t 1 2 W t 1 2. For the prior specificaon, we assume that p(w 1 ) and p(u 1 ) are N(a w, R w ) and N(a u, R u ) distribuons respecvely with pre-specified hyperparameters a w, a u, R w and R u. The prior specificaons on the covariance matrices are assumed to be of the form π(v t, W t, U t ) = π(v t )π(w t )π(u t ), (13) where π(v t ), π(w t ) and π(u t ) are inverse Wishart distribuons with known hyperparameters. A roune sensivity analysis may be carried out in order to invesgate the effect of these hyperparameters on the model fit. By Bayes theorem, the joint posterior density is then π(u, V, W, U, τ Y ) L(Y ; V, W, U, τ ) π(v t, W t, U t ) p(w t W t )p(u t U t ) t=2 p(w 1 )p(u 1 )g(τ). The resulng joint posterior is analycally intractable. The full condional distribuons of the parameters are proporonal to the joint posterior and do not have standard forms. We have employed a sampling based approach via Gibbs sampling (Gelfand and Smith, 1990), with the Metropolis Hasngs algorithm (Hasngs, 1970) using a Gaussian proposal. Maximum likelihood esmates of the parameters provide the inial values for the sampler. Let (η (s), λ (s) ), s = 1,..., B denote convergent samples generated from the joint posterior. The means and standard deviaons of these samples corresponding to each parameter provide summary features of the marginal posterior distribuons. Samples of esmated expected composion proporons can also be calculated at each me point for given covariates. For s = 1,..., B, these are η (s) = and for i = G η (s) [ λ (s) 1 + g Ŷ (s) + 1 [ λ (s) ] 1/λ (s) Ŷ (s) + 1 tg = [ g λ (s) Ŷ (s) + 1 ], i = 1,..., g (s) 1/λ ] 1/λ (s) (14) (15) where λ (s) and Ŷ (s) are calculated from the given model. Esmates of these expected proporons are obtained as the Monte Carlo averages of the expressions in (14) and (15) over the B samples. Convex credible regions for proporons can also be constructed. The first step is to generate a sample of B proporons at each me point t = 1,..., T. At me point t, the s th sample for the ith proporon is given by equaon (14) and (15). These 100(1 α) credible regions indicate how the shapes of the simplexes corresponding to the true proporons change over me. These regions are obtained using the SPlus funcon chull, which returns the indices of 228
4 Joint Stascal Meengs - Bayesian Stascal Science the points belonging to the hull. The peel opon permits peeling off from the convex hull successively unl either all the planar points are assigned or unl a user-specified limit is reached Cross Validaon In this arcle, model assessment is based on the cross-validaon predicve density. It is well known that the marginal density of f(y ) is equivalent to the set {f(y r Y (r) ) : r = 1, 2,..., T }, where f(y r Y (r) ) is the predicve distribuon of Y r when the r th observaon is deleted. The expression for f(y r Y (r) ) is f(y r Y (r) ) = f(y ) f(y (r) ) = f(y r η, Y (r), Z) π(η Y (r) )dη, (16) which is called the Condional Predicve Ordinate (CPO). Using the samples η (s), s = 1, 2,..., B generated from the posterior, the Monte Carlo esmate of CPO can be calculated as (Dey, Chang and Ray, 1996) CPO r = B ( B s=1 ) 1 1. (17) f(y r η (s), Y (r) ) For the purpose of model checking, the presence of many small CPO s cricizes the model. A useful summary stasc of the CPO s is the logarithm of the pseudo-bayes factor (Geisser and Eddy, 1979 and Dey, Chen and Chang, 1997), defined as P sbf = T log(cpo r ). (18) r=1 The advantage of P sbf over the Bayes factor (Gelfand and Dey, 1992) as a model assessment tool is that the latter is not well defined with improper priors, and is generally quite sensive to vague proper priors. The evaluaon of f(y r η (s), Y (r) ) is complicated in me series models. The calculaon is simplified as follows (Pai and Ravishanker, 1996): f(y r η (s), Y (r) ) f i (y i y i 1,..., y 1, η), (19) i=t and facilitates model assessment. 3. ILLUSTRATION We analyze trivariate composional data of motor vehicle producon, consisng of the number of motor vehicles produced by, and Other countries between 1947 and 1987 (see Grunwald, Raftery and Guttorp, 1993). Data on the movement of the US economy, measured by the percentage change in US Gross Naonal Product (GNP) is a covariate. Cumulave proporons Year Other Figure 1. Proporon of motor vehicle producon from three sources over me. Figure 1 shows the changes in the proporons of total motor vehicle producon in the three categories over me. Of the three sources, the change in the ese producon is most significant. accounted for around one third of the total producon in 1987 while their contribuon in 1947 was only around 2% of the total. The range of the US producon is (4796, 12900), indicang a producon growth over me, but with a sharp decrease in contribuon to total producon. Overall, the motor vehicle producon appears to be correlated with US growth rate, and there is evidence of some nonlinear relaonship. Visual examinaon of the scatter plots of producon from the three sources versus movement of the US economy suggests that whereas US producon is posively correlated with the covariate, the relaonship between the producon from and from and the covariate is only slightly posive, with a trace of nonlinear relaonship. Observed behavioral differences in the three sources indicate the usefulness of a composional me series analysis. Details from fitng the dynamic regression model which describes the effect of US growth rate on the Box-Cox transformed motor vehicle producon data from three different 229
5 Joint Stascal Meengs - Bayesian Stascal Science sources is presented here. Model 1 corresponds to the dynamic regression model with normal errors, while Model 2 corresponds to the dynamic regression model with Student-t errors. Details of the posterior esmates for t = 41 are given in Table I Coefficients of Intercept for Coefficients of US GNP for Coefficients of Intercept for Time = Coefficients of Intercept for Coefficients of US GNP for - - Coefficients of Intercept for Coefficients of Intercept for Figure 3. Convex hull for the intercept and US GNP coefficient under Model 1 The esmated convex credible regions at two selected me points for the true proporons corresponding to Model 1 are shown in Figure 2. It indicates that the relaonship between the proporon of vehicle producon in, and the do not change considerably. Figure 3 shows the 95% joint credible sets for the intercept and US GNP under Model 1 at three selected me points. Column 1 corresponds to and column 2 corresponds to. The figure indicates the presence of a slow change in the shape of the credible sets over me. We observed that this change is less severe for Model 2, corresponding to Student-t errors CONCLUDING REMARKS A regression model with a dynamic structure for parameter evoluon over me is used to analyze composional me series data. The inherent constraint in composional data is overcome by the Box-Cox transformaon which is a more general version of the ALR transformaon. In our modeling framework, we incorporate flexibility in two ways. The first is via a more general class of transformaons from the simplex, and the second is through the dynamic linear model framework. Our model thereby incorporates a possible change in the underlying mechanism. The Bayesian methodology facilitates parametric inference of the resulng complex model Coefficients of US GNP for It is clear that the Box-Cox transformed vehicle producon has significant negave relaonship with US economic growth rate. Values of P sbf are for the normal error model and for the Student-t error model, indicang that the normal error assumpon provides a better fit. This is substanated by a plot of the logarithm of CPO raos; all points in the plot of log(cpo from Model 1/CPO from Model 2) are above Model 2 Mean Stdev Coefficients of US GNP for λ λ α α β β 0.14 Coefficients of Intercept for - Parameter Coefficients of US GNP for Coefficients of US GNP for Table I Posterior Means and Standard Deviaons of Parameters under Normal and Student-t errors Model 1 Mean Stdev Time = Figure 2. Credible regions for true proporons from Model 1 230
6 Joint Stascal Meengs - Bayesian Stascal Science REFERENCES Aitchison, J. (1982) The Stascal Analysis of Computaonal Data with discussion. J. R. Stast. Soc. B, 44, Aitchison, J. (1986) The Stascal Analysis of Composional Data. London: Chapman and Hall. Box, G. E. P. and Cox, D. R. (1964) An Analysis of Transformaons. J. R. Stast. Soc. B, 26, Chen, M. H. and Dey, D. K. (1998) Bayesian modeling of correlated binary responses via scale mixture of mulvariate normal link funcons. Sankhya, 60, Dey, D. K. and Birmiwal, L. R. (1991) On Idenfying Mixing Density of Scale Mixtures of Normal Distribuons. Technical Report, Department of Stascs, University of Conneccut. Dey, D. K.; Chang, H. and Ray, S. C. (1996) A Bayesian Approach in Model Selecon for the Binary Response Data. Advances in Econometrics, 11, Dey, D. K.; Chen, M. H. and Chang, H. (1997) Bayesian approach for the nonlinear random effects models. Biometrics, 53, Gelfand, a, e. and Dey, D. K. (1994) Bayesian model choice: asymptocs and exact calculaons. J. R. Stast. Soc. B, 56, Geisser, S. (1993) Predicve Inference: An Introducon, Chapman and Hall: London. Grunwald, G. K.; Raftery, A. E. and Guttorp, P. (1993) Time Series of Connuous Proporons. J. R. Stast. Soc. B, 55, Hasngs, W. K. (1970) Monte Carlo sampling methods using Markov chains and their applicaons. Biometrika, 57, Iyenger, M. and Dey, D. K. (1998) Box-Cox transformaon in Bayesian analysis of composional data. Environmetrics, 9, Kalman, R. E. (1969) A New Approach to Linear Filtering and Predicng Problems. J. of Basic Engineering, 82, Pai, J. S. and Ravishanker, N. (1996) Bayesian Modelling of ARFIMA Processes by Markov Chain Monte Carlo Methods. J. of Forecasng, 15, Quintana, J. M. and West, M. (1988) Time Series Analysis of Composional Data. Bayesian Stascs : J. M. Bernardo, M. H. DeGroot, D. V. Lindley and A. F. M. Smith (Eds.), Oxford University Press. Rayens, W. S. and Srinivasan, C. (1991 a) Box- Cox Transformaons in the Analysis of Composional Data. J. of Chemometrics, 5, Rayens, W. S. and Srinivasan, C. (1991 b) Esmaon in Composional Data Analysis. J. of Chemometrics, 5, West, M. and Harrison, P. J. (1997) Bayesian Forecasng and Dynamic Models, Second edion, Springer: New York. Geisser, S. and Eddy, W. (1979) A predicve approach to model selecon. J. Amer. Stast. Assoc., 74, Gelfand, A. E.; Dey, D. K. and Chang, H. (1992) Model Determinaon using Predicve Distribuon with Implementaon via Sampling-Based Methods. Bayesian Stascs : J. M. Bernardo, J. O. Berger, A. P. Dawid and A. F. M. Smith (Eds.), Oxford University Press, Gelfand, A. E. and Smith, A. F. M. (1990) Sampling based approaches to calculang marginal densies. J. Amer. Stast. Assoc., 85,
Pa8ern Recogni6on. and Machine Learning. Chapter 4: Linear Models for Classifica6on
Pa8ern Recogni6on and Machine Learning Chapter 4: Linear Models for Classifica6on Represen'ng the target values for classifica'on If there are only two classes, we typically use a single real valued output
Missing Data. Katyn & Elena
Missing Data Katyn & Elena What to do with Missing Data Standard is complete case analysis/listwise dele;on ie. Delete cases with missing data so only complete cases are le> Two other popular op;ons: Mul;ple
STA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! [email protected]! http://www.cs.toronto.edu/~rsalakhu/ Lecture 6 Three Approaches to Classification Construct
11. Time series and dynamic linear models
11. Time series and dynamic linear models Objective To introduce the Bayesian approach to the modeling and forecasting of time series. Recommended reading West, M. and Harrison, J. (1997). models, (2 nd
Data Mining. Supervised Methods. Ciro Donalek [email protected]. Ay/Bi 199ab: Methods of Computa@onal Sciences hcp://esci101.blogspot.
Data Mining Supervised Methods Ciro Donalek [email protected] Supervised Methods Summary Ar@ficial Neural Networks Mul@layer Perceptron Support Vector Machines SoLwares Supervised Models: Supervised
Modeling and Analysis of Call Center Arrival Data: A Bayesian Approach
Modeling and Analysis of Call Center Arrival Data: A Bayesian Approach Refik Soyer * Department of Management Science The George Washington University M. Murat Tarimcilar Department of Management Science
Applications of R Software in Bayesian Data Analysis
Article International Journal of Information Science and System, 2012, 1(1): 7-23 International Journal of Information Science and System Journal homepage: www.modernscientificpress.com/journals/ijinfosci.aspx
Bayesian Statistics: Indian Buffet Process
Bayesian Statistics: Indian Buffet Process Ilker Yildirim Department of Brain and Cognitive Sciences University of Rochester Rochester, NY 14627 August 2012 Reference: Most of the material in this note
Spatial Statistics Chapter 3 Basics of areal data and areal data modeling
Spatial Statistics Chapter 3 Basics of areal data and areal data modeling Recall areal data also known as lattice data are data Y (s), s D where D is a discrete index set. This usually corresponds to data
Estimation and comparison of multiple change-point models
Journal of Econometrics 86 (1998) 221 241 Estimation and comparison of multiple change-point models Siddhartha Chib* John M. Olin School of Business, Washington University, 1 Brookings Drive, Campus Box
Marketing Mix Modelling and Big Data P. M Cain
1) Introduction Marketing Mix Modelling and Big Data P. M Cain Big data is generally defined in terms of the volume and variety of structured and unstructured information. Whereas structured data is stored
Tutorial on Markov Chain Monte Carlo
Tutorial on Markov Chain Monte Carlo Kenneth M. Hanson Los Alamos National Laboratory Presented at the 29 th International Workshop on Bayesian Inference and Maximum Entropy Methods in Science and Technology,
Centre for Central Banking Studies
Centre for Central Banking Studies Technical Handbook No. 4 Applied Bayesian econometrics for central bankers Andrew Blake and Haroon Mumtaz CCBS Technical Handbook No. 4 Applied Bayesian econometrics
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION Introduction In the previous chapter, we explored a class of regression models having particularly simple analytical
BayesX - Software for Bayesian Inference in Structured Additive Regression
BayesX - Software for Bayesian Inference in Structured Additive Regression Thomas Kneib Faculty of Mathematics and Economics, University of Ulm Department of Statistics, Ludwig-Maximilians-University Munich
Statistics Graduate Courses
Statistics Graduate Courses STAT 7002--Topics in Statistics-Biological/Physical/Mathematics (cr.arr.).organized study of selected topics. Subjects and earnable credit may vary from semester to semester.
Christfried Webers. Canberra February June 2015
c Statistical Group and College of Engineering and Computer Science Canberra February June (Many figures from C. M. Bishop, "Pattern Recognition and ") 1of 829 c Part VIII Linear Classification 2 Logistic
Linear regression methods for large n and streaming data
Linear regression methods for large n and streaming data Large n and small or moderate p is a fairly simple problem. The sufficient statistic for β in OLS (and ridge) is: The concept of sufficiency is
Bayesian Machine Learning (ML): Modeling And Inference in Big Data. Zhuhua Cai Google, Rice University [email protected]
Bayesian Machine Learning (ML): Modeling And Inference in Big Data Zhuhua Cai Google Rice University [email protected] 1 Syllabus Bayesian ML Concepts (Today) Bayesian ML on MapReduce (Next morning) Bayesian
Dynamic Linear Models with R
Giovanni Petris, Sonia Petrone, Patrizia Campagnoli Dynamic Linear Models with R SPIN Springer s internal project number, if known Monograph August 10, 2007 Springer Berlin Heidelberg NewYork Hong Kong
Handling attrition and non-response in longitudinal data
Longitudinal and Life Course Studies 2009 Volume 1 Issue 1 Pp 63-72 Handling attrition and non-response in longitudinal data Harvey Goldstein University of Bristol Correspondence. Professor H. Goldstein
SAS Software to Fit the Generalized Linear Model
SAS Software to Fit the Generalized Linear Model Gordon Johnston, SAS Institute Inc., Cary, NC Abstract In recent years, the class of generalized linear models has gained popularity as a statistical modeling
Big Data and Health Insurance Product Selec6on (and a few other applica6on) Jonathan Kolstad UC Berkeley and NBER
Big Data and Health Insurance Product Selec6on (and a few other applica6on) Jonathan Kolstad UC Berkeley and NBER Introduc6on Applica6ons of behavioral economics in health SeIng where behavioral assump6ons
ANALYTICAL TECHNIQUES FOR DATA VISUALIZATION
ANALYTICAL TECHNIQUES FOR DATA VISUALIZATION CSE 537 Ar@ficial Intelligence Professor Anita Wasilewska GROUP 2 TEAM MEMBERS: SAEED BOOR BOOR - 110564337 SHIH- YU TSAI - 110385129 HAN LI 110168054 SOURCES
Example: Credit card default, we may be more interested in predicting the probabilty of a default than classifying individuals as default or not.
Statistical Learning: Chapter 4 Classification 4.1 Introduction Supervised learning with a categorical (Qualitative) response Notation: - Feature vector X, - qualitative response Y, taking values in C
Chapter 12. The Costs of Produc4on
Chapter 12 The Costs of Produc4on Copyright 214 McGraw-Hill Education. All rights reserved. No reproduction or distribution without the prior written consent of McGraw-Hill Education. What will you learn
CHAPTER 3 EXAMPLES: REGRESSION AND PATH ANALYSIS
Examples: Regression And Path Analysis CHAPTER 3 EXAMPLES: REGRESSION AND PATH ANALYSIS Regression analysis with univariate or multivariate dependent variables is a standard procedure for modeling relationships
Analysis of Bayesian Dynamic Linear Models
Analysis of Bayesian Dynamic Linear Models Emily M. Casleton December 17, 2010 1 Introduction The main purpose of this project is to explore the Bayesian analysis of Dynamic Linear Models (DLMs). The main
Gaussian Conjugate Prior Cheat Sheet
Gaussian Conjugate Prior Cheat Sheet Tom SF Haines 1 Purpose This document contains notes on how to handle the multivariate Gaussian 1 in a Bayesian setting. It focuses on the conjugate prior, its Bayesian
Validation of Software for Bayesian Models using Posterior Quantiles. Samantha R. Cook Andrew Gelman Donald B. Rubin DRAFT
Validation of Software for Bayesian Models using Posterior Quantiles Samantha R. Cook Andrew Gelman Donald B. Rubin DRAFT Abstract We present a simulation-based method designed to establish that software
Linear Classification. Volker Tresp Summer 2015
Linear Classification Volker Tresp Summer 2015 1 Classification Classification is the central task of pattern recognition Sensors supply information about an object: to which class do the object belong
Understanding and Detec.ng Real- World Performance Bugs
Understanding and Detec.ng Real- World Performance Bugs Gouliang Jin, Linhai Song, Xiaoming Shi, Joel Scherpelz, and Shan Lu Presented by Cindy Rubio- González Feb 10 th, 2015 Mo.va.on Performance bugs
CHAPTER 9 EXAMPLES: MULTILEVEL MODELING WITH COMPLEX SURVEY DATA
Examples: Multilevel Modeling With Complex Survey Data CHAPTER 9 EXAMPLES: MULTILEVEL MODELING WITH COMPLEX SURVEY DATA Complex survey data refers to data obtained by stratification, cluster sampling and/or
Bayes and Naïve Bayes. cs534-machine Learning
Bayes and aïve Bayes cs534-machine Learning Bayes Classifier Generative model learns Prediction is made by and where This is often referred to as the Bayes Classifier, because of the use of the Bayes rule
Ensemble Methods. Adapted from slides by Todd Holloway h8p://abeau<fulwww.com/2007/11/23/ ensemble- machine- learning- tutorial/
Ensemble Methods Adapted from slides by Todd Holloway h8p://abeau
Optimizing Prediction with Hierarchical Models: Bayesian Clustering
1 Technical Report 06/93, (August 30, 1993). Presidencia de la Generalidad. Caballeros 9, 46001 - Valencia, Spain. Tel. (34)(6) 386.6138, Fax (34)(6) 386.3626, e-mail: [email protected] Optimizing Prediction
Parallelization Strategies for Multicore Data Analysis
Parallelization Strategies for Multicore Data Analysis Wei-Chen Chen 1 Russell Zaretzki 2 1 University of Tennessee, Dept of EEB 2 University of Tennessee, Dept. Statistics, Operations, and Management
Logistic Regression (1/24/13)
STA63/CBB540: Statistical methods in computational biology Logistic Regression (/24/3) Lecturer: Barbara Engelhardt Scribe: Dinesh Manandhar Introduction Logistic regression is model for regression used
HT2015: SC4 Statistical Data Mining and Machine Learning
HT2015: SC4 Statistical Data Mining and Machine Learning Dino Sejdinovic Department of Statistics Oxford http://www.stats.ox.ac.uk/~sejdinov/sdmml.html Bayesian Nonparametrics Parametric vs Nonparametric
CS 688 Pattern Recognition Lecture 4. Linear Models for Classification
CS 688 Pattern Recognition Lecture 4 Linear Models for Classification Probabilistic generative models Probabilistic discriminative models 1 Generative Approach ( x ) p C k p( C k ) Ck p ( ) ( x Ck ) p(
Statistics in Retail Finance. Chapter 6: Behavioural models
Statistics in Retail Finance 1 Overview > So far we have focussed mainly on application scorecards. In this chapter we shall look at behavioural models. We shall cover the following topics:- Behavioural
Nodes, Ties and Influence
Nodes, Ties and Influence Chapter 2 Chapter 2, Community Detec:on and Mining in Social Media. Lei Tang and Huan Liu, Morgan & Claypool, September, 2010. 1 IMPORTANCE OF NODES 2 Importance of Nodes Not
APPLIED MISSING DATA ANALYSIS
APPLIED MISSING DATA ANALYSIS Craig K. Enders Series Editor's Note by Todd D. little THE GUILFORD PRESS New York London Contents 1 An Introduction to Missing Data 1 1.1 Introduction 1 1.2 Chapter Overview
Tail-Dependence an Essential Factor for Correctly Measuring the Benefits of Diversification
Tail-Dependence an Essential Factor for Correctly Measuring the Benefits of Diversification Presented by Work done with Roland Bürgi and Roger Iles New Views on Extreme Events: Coupled Networks, Dragon
ECBDL 14: Evolu/onary Computa/on for Big Data and Big Learning Workshop July 13 th, 2014 Big Data Compe//on
ECBDL 14: Evolu/onary Computa/on for Big Data and Big Learning Workshop July 13 th, 2014 Big Data Compe//on Jaume Bacardit [email protected] The Interdisciplinary Compu/ng and Complex BioSystems
Statistical Machine Learning from Data
Samy Bengio Statistical Machine Learning from Data 1 Statistical Machine Learning from Data Gaussian Mixture Models Samy Bengio IDIAP Research Institute, Martigny, Switzerland, and Ecole Polytechnique
Component Ordering in Independent Component Analysis Based on Data Power
Component Ordering in Independent Component Analysis Based on Data Power Anne Hendrikse Raymond Veldhuis University of Twente University of Twente Fac. EEMCS, Signals and Systems Group Fac. EEMCS, Signals
1 Maximum likelihood estimation
COS 424: Interacting with Data Lecturer: David Blei Lecture #4 Scribes: Wei Ho, Michael Ye February 14, 2008 1 Maximum likelihood estimation 1.1 MLE of a Bernoulli random variable (coin flips) Given N
Gaussian Processes in Machine Learning
Gaussian Processes in Machine Learning Carl Edward Rasmussen Max Planck Institute for Biological Cybernetics, 72076 Tübingen, Germany [email protected] WWW home page: http://www.tuebingen.mpg.de/ carl
Service courses for graduate students in degree programs other than the MS or PhD programs in Biostatistics.
Course Catalog In order to be assured that all prerequisites are met, students must acquire a permission number from the education coordinator prior to enrolling in any Biostatistics course. Courses are
Department of Economics
Department of Economics On Testing for Diagonality of Large Dimensional Covariance Matrices George Kapetanios Working Paper No. 526 October 2004 ISSN 1473-0278 On Testing for Diagonality of Large Dimensional
STATISTICA Formula Guide: Logistic Regression. Table of Contents
: Table of Contents... 1 Overview of Model... 1 Dispersion... 2 Parameterization... 3 Sigma-Restricted Model... 3 Overparameterized Model... 4 Reference Coding... 4 Model Summary (Summary Tab)... 5 Summary
Validation of Software for Bayesian Models Using Posterior Quantiles
Validation of Software for Bayesian Models Using Posterior Quantiles Samantha R. COOK, Andrew GELMAN, and Donald B. RUBIN This article presents a simulation-based method designed to establish the computational
Applying MCMC Methods to Multi-level Models submitted by William J Browne for the degree of PhD of the University of Bath 1998 COPYRIGHT Attention is drawn tothefactthatcopyright of this thesis rests with
Lecture 9: Introduction to Pattern Analysis
Lecture 9: Introduction to Pattern Analysis g Features, patterns and classifiers g Components of a PR system g An example g Probability definitions g Bayes Theorem g Gaussian densities Features, patterns
Web-based Supplementary Materials for Bayesian Effect Estimation. Accounting for Adjustment Uncertainty by Chi Wang, Giovanni
1 Web-based Supplementary Materials for Bayesian Effect Estimation Accounting for Adjustment Uncertainty by Chi Wang, Giovanni Parmigiani, and Francesca Dominici In Web Appendix A, we provide detailed
Auxiliary Variables in Mixture Modeling: 3-Step Approaches Using Mplus
Auxiliary Variables in Mixture Modeling: 3-Step Approaches Using Mplus Tihomir Asparouhov and Bengt Muthén Mplus Web Notes: No. 15 Version 8, August 5, 2014 1 Abstract This paper discusses alternatives
Bayesian Statistics in One Hour. Patrick Lam
Bayesian Statistics in One Hour Patrick Lam Outline Introduction Bayesian Models Applications Missing Data Hierarchical Models Outline Introduction Bayesian Models Applications Missing Data Hierarchical
Least Squares Estimation
Least Squares Estimation SARA A VAN DE GEER Volume 2, pp 1041 1045 in Encyclopedia of Statistics in Behavioral Science ISBN-13: 978-0-470-86080-9 ISBN-10: 0-470-86080-4 Editors Brian S Everitt & David
Bayesian Methods for the Social and Behavioral Sciences
Bayesian Methods for the Social and Behavioral Sciences Jeff Gill Harvard University 2007 ICPSR First Session: June 25-July 20, 9-11 AM. Email: [email protected] TA: Yu-Sung Su ([email protected]).
DURATION ANALYSIS OF FLEET DYNAMICS
DURATION ANALYSIS OF FLEET DYNAMICS Garth Holloway, University of Reading, [email protected] David Tomberlin, NOAA Fisheries, [email protected] ABSTRACT Though long a standard technique
Markov Chain Monte Carlo Simulation Made Simple
Markov Chain Monte Carlo Simulation Made Simple Alastair Smith Department of Politics New York University April2,2003 1 Markov Chain Monte Carlo (MCMC) simualtion is a powerful technique to perform numerical
Multivariate Normal Distribution
Multivariate Normal Distribution Lecture 4 July 21, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2 Lecture #4-7/21/2011 Slide 1 of 41 Last Time Matrices and vectors Eigenvalues
December 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B. KITCHENS
December 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B KITCHENS The equation 1 Lines in two-dimensional space (1) 2x y = 3 describes a line in two-dimensional space The coefficients of x and y in the equation
Penalized Logistic Regression and Classification of Microarray Data
Penalized Logistic Regression and Classification of Microarray Data Milan, May 2003 Anestis Antoniadis Laboratoire IMAG-LMC University Joseph Fourier Grenoble, France Penalized Logistic Regression andclassification
INDIRECT INFERENCE (prepared for: The New Palgrave Dictionary of Economics, Second Edition)
INDIRECT INFERENCE (prepared for: The New Palgrave Dictionary of Economics, Second Edition) Abstract Indirect inference is a simulation-based method for estimating the parameters of economic models. Its
Longitudinal random effects models for genetic analysis of binary data with application to mastitis in dairy cattle
Genet. Sel. Evol. 35 (2003) 457 468 457 INRA, EDP Sciences, 2003 DOI: 10.1051/gse:2003034 Original article Longitudinal random effects models for genetic analysis of binary data with application to mastitis
A SURVEY ON CONTINUOUS ELLIPTICAL VECTOR DISTRIBUTIONS
A SURVEY ON CONTINUOUS ELLIPTICAL VECTOR DISTRIBUTIONS Eusebio GÓMEZ, Miguel A. GÓMEZ-VILLEGAS and J. Miguel MARÍN Abstract In this paper it is taken up a revision and characterization of the class of
The Chinese Restaurant Process
COS 597C: Bayesian nonparametrics Lecturer: David Blei Lecture # 1 Scribes: Peter Frazier, Indraneel Mukherjee September 21, 2007 In this first lecture, we begin by introducing the Chinese Restaurant Process.
Non Linear Dependence Structures: a Copula Opinion Approach in Portfolio Optimization
Non Linear Dependence Structures: a Copula Opinion Approach in Portfolio Optimization Jean- Damien Villiers ESSEC Business School Master of Sciences in Management Grande Ecole September 2013 1 Non Linear
Introduction to General and Generalized Linear Models
Introduction to General and Generalized Linear Models General Linear Models - part I Henrik Madsen Poul Thyregod Informatics and Mathematical Modelling Technical University of Denmark DK-2800 Kgs. Lyngby
Review Jeopardy. Blue vs. Orange. Review Jeopardy
Review Jeopardy Blue vs. Orange Review Jeopardy Jeopardy Round Lectures 0-3 Jeopardy Round $200 How could I measure how far apart (i.e. how different) two observations, y 1 and y 2, are from each other?
Sampling for Bayesian computation with large datasets
Sampling for Bayesian computation with large datasets Zaiying Huang Andrew Gelman April 27, 2005 Abstract Multilevel models are extremely useful in handling large hierarchical datasets. However, computation
Applied Statistics. J. Blanchet and J. Wadsworth. Institute of Mathematics, Analysis, and Applications EPF Lausanne
Applied Statistics J. Blanchet and J. Wadsworth Institute of Mathematics, Analysis, and Applications EPF Lausanne An MSc Course for Applied Mathematicians, Fall 2012 Outline 1 Model Comparison 2 Model
15.062 Data Mining: Algorithms and Applications Matrix Math Review
.6 Data Mining: Algorithms and Applications Matrix Math Review The purpose of this document is to give a brief review of selected linear algebra concepts that will be useful for the course and to develop
Bayesian Time Series Analysis
Bayesian Time Series Analysis Mark Steel, University of Warwick Abstract This article describes the use of Bayesian methods in the statistical analysis of time series. The use of Markov chain Monte Carlo
Simple Linear Regression Inference
Simple Linear Regression Inference 1 Inference requirements The Normality assumption of the stochastic term e is needed for inference even if it is not a OLS requirement. Therefore we have: Interpretation
Bias in the Estimation of Mean Reversion in Continuous-Time Lévy Processes
Bias in the Estimation of Mean Reversion in Continuous-Time Lévy Processes Yong Bao a, Aman Ullah b, Yun Wang c, and Jun Yu d a Purdue University, IN, USA b University of California, Riverside, CA, USA
Bayesian Factorization Machines
Bayesian Factorization Machines Christoph Freudenthaler, Lars Schmidt-Thieme Information Systems & Machine Learning Lab University of Hildesheim 31141 Hildesheim {freudenthaler, schmidt-thieme}@ismll.de
Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model
Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model 1 September 004 A. Introduction and assumptions The classical normal linear regression model can be written
Probabilistic Methods for Time-Series Analysis
Probabilistic Methods for Time-Series Analysis 2 Contents 1 Analysis of Changepoint Models 1 1.1 Introduction................................ 1 1.1.1 Model and Notation....................... 2 1.1.2 Example:
A Bootstrap Metropolis-Hastings Algorithm for Bayesian Analysis of Big Data
A Bootstrap Metropolis-Hastings Algorithm for Bayesian Analysis of Big Data Faming Liang University of Florida August 9, 2015 Abstract MCMC methods have proven to be a very powerful tool for analyzing
Lecture 3: Linear methods for classification
Lecture 3: Linear methods for classification Rafael A. Irizarry and Hector Corrada Bravo February, 2010 Today we describe four specific algorithms useful for classification problems: linear regression,
CHAPTER 8 EXAMPLES: MIXTURE MODELING WITH LONGITUDINAL DATA
Examples: Mixture Modeling With Longitudinal Data CHAPTER 8 EXAMPLES: MIXTURE MODELING WITH LONGITUDINAL DATA Mixture modeling refers to modeling with categorical latent variables that represent subpopulations
UW CSE Technical Report 03-06-01 Probabilistic Bilinear Models for Appearance-Based Vision
UW CSE Technical Report 03-06-01 Probabilistic Bilinear Models for Appearance-Based Vision D.B. Grimes A.P. Shon R.P.N. Rao Dept. of Computer Science and Engineering University of Washington Seattle, WA
