Dealing with uncertainty in Gaussian Bayesian networks from a regression perspective

Size: px
Start display at page:

Download "Dealing with uncertainty in Gaussian Bayesian networks from a regression perspective"

Transcription

1 Dealing with uncertainty in Gaussian ayesian networks from a regression perspective Miguel A. Gómez-Villegas a, Paloma Main a, Hilario Navarro b and Rosario Susi a a Universidad Complutense Madrid, Spain b Universidad Nacional de Educación a Distancia, Spain Abstract Some sensitivity analyses have been developed to evaluate the impact of uncertainty about the mean vector and the covariance matrix that specify the joint distribution of the variables in the nodes of a Gaussian ayesian network (GN). Nevertheless, uncertainty about the alternative conditional speci cation of GN based on the regression coe cients of each variable given its parents in the directed acyclic graph (DAG), has received low attention in the literature. In this line, we focus on evaluating the e ect of regression coe cients misspeci cation by means of the Kullback-Leibler (KL) divergence. Introduction GNs are de ned as ayesian networks (Ns) where the joint probability density of X = (X ; X ; :::; X p ) is a multivariate normal distribution N p (; ) with the p dimensional mean vector and the p p positive de nite covariance matrix using a directed acyclic graph (DAG) to represent the dependence structure of the variables. As in Ns, the joint density can be factorized using the conditional probability densities of X i (i = ; :::; p) given its parents in the DAG, pa(x i ) fx ; :::; X i g. hese are univariate normal distributions with densities Xi f(x i jpa(x i )) N(x i j i + b ji (x j j ); v i ) j= being i the marginal mean of X i, b ji the regression coe cients of X i given X j pa(x i ), and v i the conditional variance of X i given its parents in the DAG. Note that if b ji = then X j is not a parent of X i. he parameters of the joint distribution can be obtained from the previous conditional speci cation. More concretely, the means f i g are addresses: ma.gv@mat.ucm.es (M.A. Gómez-Villegas), pmain@mat.ucm.es (P. Main), hnavarro@ccia.uned.es (H. Navarro), rsusi@estad.ucm.es (R. Susi) obviously the elements of the p mean vector and = [(I p ) ] D(I p ) dimensional (Shachter and Kenley, 989) where D is a diagonal matrix D = diag(v) with the conditional variances v = (v ; :::; v p ) and a strictly u- pper triangular matrix with the regression coe- cients b ji ; j f; :::; i g. he speci cation based on (; ; D) is more manageable for experts because they only have to describe univariate distributions. Moreover, the DAG can be improved by adding the numerical values of the regression coe cient and conditional variance to the corresponding arc and node, respectively. Nevertheless, there can still be considerably uncertainty about parameters. Sensitivity analysis is an important phase of any modelling procedure. (Castillo and Kjærul, 3) performed a one-way sensitivity analysis investigating the impact of small changes in the network parameters and. Alternatively, (Gómez- Villegas, Main and Susi, 7) proposed a oneway global sensitivity analysis instead of considering local aspects as location and dispersion, over the network s output. Also, in (Gómez- Villegas, Main and Susi, 8) a n way sensitivity analysis is presented as a generalization of

2 the previous one both using the KL divergence to evaluate the impact of perturbations. It is well known the KL divergence is an nonsymmetric measure that evaluates the amount of information available to discriminate between two probability distributions. We have chosen it because we want to compare the global behaviors of two probability distributions. he directed KL divergence between the probability densities f(w) and f (w), de ned over the same domain is D KL (f j f) = Z f(w) ln f(w) f (w) dw : he expression for multivariate normal distributions is given by D KL (f j f) = = ln j j jj + tr + h I p + i ; where f and f are densities of normal distributions N p (; ) and N p ( ; ) respectively. In general, if X = (X ; X ; :::; X p ) is a random vector normally distributed with parameters (; ) where the covariance matrix is inaccurately speci ed, the e ect of a perturbation pp measured in terms of a directed Kullback- Leibler divergence can be expressed as follows D KL (fjf ) = [ln jj j + j + tr( ( + ))] = [ln ji p + j tr( )] 4 tr( ) = 4 jj jj F = = 4 i ( ) i i= being f the density function with the perturbed covariance matrix +, i (i = ; :::; p) each of the columns of, with the necessary restrictions to get symmetric and positive de nite matrices + and, jj jj F the Frobenius matrix norm and tr () the trace function. However, as the directed KL divergence D KL (f j f) can be interpreted as the information lost when f is used to approximate f, in the following sensitivity analyses f has to be the original model and f the perturbed one o- pposite to previously used divergence. Herein we focus on the repercussion of a misspeci ed pp matrix, while the rest of the conditional parameters are known. o evaluate perturbation e ects, the proper directed KL divergence is used in all the studied cases. he paper is organized as follows. In Section the problem is stated and analyzed for constant errors. In Section 3 random perturbations are considered; some examples illustrate the behavior of the proposed measure for both local and global uncertainty. Fixed misspeci cation of Let f be the density of a multivariate normal distribution N p (; ) with conditional parameters ; and D. Denoting by the matrix with additive perturbations on and f the corresponding perturbed density N p ;, where (I p ) D(Ip ) (see (Susi, Navarro, Main and Gómez-Villegas, 9)), the KL divergence for comparing two covariance matrices when the means are equal is h trace i p. DKL f j f = () It should be noted that as I p and I p are upper triangular matrices with diagonal entries equal to one then ln j j jj =. Now, given that and = Ip (Ip ) + (I p (Ip trace ) + D ) = = trace (I p ) ;

3 trace (I p ) = = trace (I p ) the divergence in () can be restored as D KL f j f = = [p trace (I p ) + +trace D p] = = tr D = = tr D. () Note that (I p ) is an upper triangular matrix with diagonal entries equal to one and (I p ) is an upper triangular matrix with diagonal entries zero. Under the same assumptions, if (i) denotes an (i )-dimensional vector of local errors in node i produced by an erroneous estimation or elicitation of the node i with its parents relationships, the perturbation matrix on, with only this error source is i (i) (i)i ;i = : A hus, denoting f ;i as the density with coe - cients matrix + ;i, the e ect on the joint distribution can be expressed by KL f ;i j f = h tr ;i ;id i = = Xi k= (i)k tr (i )k (i)k, (3) i = ; ::::; p, being (i )k the k-th column of the submatrix (i ) built with the rst i rows and columns of. he local perturbation divergence in (3) also may be written in the form: KL f ;i j f = Xi (i)k (i )k ; (i) = k= = (i) (i )(i) = = U (i ) (i) with U = (I p ) and the submatrix U (i ) determined by the rst i rows and columns of U. Returning to the measure of interest, DKL f j f in (), it can be immediately obtained that DKL f j f = KL f ;i j f = = v i (i) (i ) (i). (4) hen, the total e ect can be expressed as the sum of individual e ects and consequently, the global sensitivity analysis can be performed through local analyses of nodes. Some direct results may be useful in applications: If all the components in (i) are equal to, it follows KL f ;i j f = X X i i tk k= t= he possibly erroneous arc deletion from node j to node i having b ji 6=, would yield KL f ;i j f = b ji jj he e ect of arc inclusion from node j to node i introducing b ji is also KL f ;i j f = b ji jj hese last two cases describe the impact of the conditional and marginal variances for an uncertain knowledge of the exact model giving the distance between DAGs obtained by adding or removing arcs.

4 Figure : DAG with regression coe cients on the arcs Example. Let us consider the GN in Figure with parameters: = ; Using that D = diag (; ; ; ; 4; ; ) : = [(I 7 ) ] D(I 7 ) C A the covariance matrix is = 6 4 : A he divergences re ecting the e ect of each arc removal are shown in Figure. It is observed that divergence increases as the depth of the node grows with a maximum in the arc from X 6 to X 7. herefore, it gives us information about Figure : Measure of deviation for each arc removal the di erence between the original GN and the particular networks obtained by means of the cancellation of some regression coe cients. 3 he random case he expression (4) relates easily, global e ect to local errors e ects due to no interaction. Now, we are going to replace the hypothesis that is known by the assumption that it is a random matrix. he main aim is using the relation (4) to evaluate the impact of uncertainty in for a GN. oth this measure and its value for di erent nodes can be useful to point the most sensitive nodes so as to compare structures with respect to uncertainty sensitivity. If we suppose each vector (i) is distributed independent from the remaining errors as N i (; E i ) ; i = ; :::; p, while the common value v = v = = v p = v is known, each random variable (i) (i ) (i) is a quadratic form in normal variables with a chi-squared distribution with i degrees of freedom (i ), if and only if E i (i ) is a symmetric idempotent matrix of rank i (Rencher, ). Also, if Y is a random vector with mean vector, covariance matrix and M pp is a nonrandom matrix, then (Rencher, ) E(Y MY) = M + tr(m)

5 V ar(y MY) = 4 M+tr((M) ). (5) hus, it results that the mean of the random variable DKL f j f is E D KL f j f = v tr (i ) E i. If we express the covariance matrix E i terms of the di erence with (i ), that is in E i = (i ) + A i, then tr (i ) E i = tr I(i ) + (i ) A i = = i + tr (i ) A i. It follows immediately that for a positive semide nite matrix A i it could be denoted by E i "larger than" (i ) the minimum mean impact is obtained when A i = being k = p v (p ) a lower bound for the mean impact of all the matrices E i larger than (i ). Moreover, imposing E i = (i ) we could assure each summand is distributed as a v (i ) because E i (i ) = I i is a symmetric idempotent matrix of rank i. he lower bound k can be considered to de ne the mean relative sensitivity by E DKL f j f P p = tr (i )E i p k (p ), could be as- whenever E i "larger than" sumed. 3. Independent errors (i ) When the hypothesis of independent errors with common variance can be accepted, that is E i = I (i ) ; i = ; :::; p, using (5), the quadratic forms of each component have rst and second order moments given by: E (i) (i ) (i) = tr (i ) V ar (i) (i ) (i) = 4 tr[( (i ) ) ] Figure 3: Node divergence from node (red) to node 7 (yellow) and global divergence (black) Consequently, under the stated conditions, we obtain an increasing average e ect according to node depth in the network, independently of the network speci cation. Figure 3 illustrates the random behavior of Kullback-Leibler divergences displaying the empirical cumulative distribution function (ECDF) for samples from v (i) (i ) (i), i = ; :::; 7, as well as the random global e ect for the example discussed above. We have used a simulated sample of size ;, with v = and = 5, for each case. In this setting, a reasonable procedure to evaluate sensitivity to uncertainty in is to analyze the normalized ratio S (f) D KL f j f =, that can be interpreted as the distribution variation in terms of the uncertainty variation. Obviously, the random divergence DKL f j f changes with ; Figure 4 shows the ECDFs behavior for some values in Example, exhibiting an apparent dominance relation. Nevertheless, the mean as well as the variance of S (f) do not depend on ; more concretely

6 Example. For the GN in Example it is obtained E S (f) = 63 Using (6), the numerical results by nodes are ().8, (3).6, (4).4, (5).7, (6)., (7).76 Figure 4: Empirical cumulative distribution function of Kullback-Leibler divergences for GN in Example with di erent independent errors: " = :5 (black), (red), (green), 5 (dark blue), (light blue) E S (f) = P p = tr (i ) V ar S (f) = v P p tr (i ) Relying on the moments invariance we propose to evaluate the GN sensitivity to uncertainty in by E S (f) = v i= tr (i ) = = Xp ii (p v where ii denotes the variance of X i. hen, the relative contribution of each node to the total sensitivity measure will be given by tr (i P p i= ii(p i) : (6) his result enables us to classify nodes according to greatest contribution to global sensitivity. ) i); Here the most important values are the contribution of the nodes to the global mean normalized divergence, resulting a signi cantly large in uence for node (7) compared to the rest of nodes in the DAG. herefore, independent random errors in the regression coe cients of nodes () to (5) do not describe very di erent joint models, however, that is not the case for node (7) and some e ort has to be made to bring some additional information to asses the correct regression coe cient value. 4 Conclusions he factorization of the joint distribution in GN leads us to an additive decomposition of the Kullback-Leibler divergence. hen, for misspeci ed regression coe cients, the weight each node has in the global deviation of the initial structure can be determined. Modelling uncertainty with independent random errors provides a highly simpli ed analysis to achieve an uncertainty sensitivity measure de nition that can be easily handled. Acknowledgments his research has been supported by the Spanish Ministerio de Ciencia e Innovación, Grant MM 8.38, and partially by GR58/8- A, Métodos ayesianos by the SCH- UCM from Spain. References E. Castillo and U. Kjærul. 3. Sensitivity analysis in Gaussian ayesian networks using a symbolic-numerical technique. Reliability Engineering and System Safety, 79: M.A. Gómez-Villegas, P. Main, R. Susi. 7. Sensitivity Analysis in Gaussian ayesian Networks

7 Using a Divergence Measure. Communications in Statistics: heory and Methods, 36(3): M.A. Gómez-Villegas, P. Main and R. Susi. 8. Sensitivity of Gaussian ayesian Networks to Inaccuracies in heir Parameters. Proceedings of the 4th EuropeanWorkshop on Probabilistic Graphical Models, pages A. Rencher.. Linear Models in Statistics. New York : Wiley. R. Shachter and C. Kenley Gaussian in uence diagrams. Management Science, 35: R. Susi, H. Navarro, P. Main and M.A. Gómez- Villegas. Perturbing the structure in Gaussian ayesian networks. 9. Cuadernos de rabajo de la E.U. de Estadística, C/9:. _trabajo/c_9.pdf.

Chapter 3: The Multiple Linear Regression Model

Chapter 3: The Multiple Linear Regression Model Chapter 3: The Multiple Linear Regression Model Advanced Econometrics - HEC Lausanne Christophe Hurlin University of Orléans November 23, 2013 Christophe Hurlin (University of Orléans) Advanced Econometrics

More information

CAPM, Arbitrage, and Linear Factor Models

CAPM, Arbitrage, and Linear Factor Models CAPM, Arbitrage, and Linear Factor Models CAPM, Arbitrage, Linear Factor Models 1/ 41 Introduction We now assume all investors actually choose mean-variance e cient portfolios. By equating these investors

More information

15.062 Data Mining: Algorithms and Applications Matrix Math Review

15.062 Data Mining: Algorithms and Applications Matrix Math Review .6 Data Mining: Algorithms and Applications Matrix Math Review The purpose of this document is to give a brief review of selected linear algebra concepts that will be useful for the course and to develop

More information

Introduction to General and Generalized Linear Models

Introduction to General and Generalized Linear Models Introduction to General and Generalized Linear Models General Linear Models - part I Henrik Madsen Poul Thyregod Informatics and Mathematical Modelling Technical University of Denmark DK-2800 Kgs. Lyngby

More information

Least Squares Estimation

Least Squares Estimation Least Squares Estimation SARA A VAN DE GEER Volume 2, pp 1041 1045 in Encyclopedia of Statistics in Behavioral Science ISBN-13: 978-0-470-86080-9 ISBN-10: 0-470-86080-4 Editors Brian S Everitt & David

More information

Chapter 1. Vector autoregressions. 1.1 VARs and the identi cation problem

Chapter 1. Vector autoregressions. 1.1 VARs and the identi cation problem Chapter Vector autoregressions We begin by taking a look at the data of macroeconomics. A way to summarize the dynamics of macroeconomic data is to make use of vector autoregressions. VAR models have become

More information

A SURVEY ON CONTINUOUS ELLIPTICAL VECTOR DISTRIBUTIONS

A SURVEY ON CONTINUOUS ELLIPTICAL VECTOR DISTRIBUTIONS A SURVEY ON CONTINUOUS ELLIPTICAL VECTOR DISTRIBUTIONS Eusebio GÓMEZ, Miguel A. GÓMEZ-VILLEGAS and J. Miguel MARÍN Abstract In this paper it is taken up a revision and characterization of the class of

More information

Multivariate normal distribution and testing for means (see MKB Ch 3)

Multivariate normal distribution and testing for means (see MKB Ch 3) Multivariate normal distribution and testing for means (see MKB Ch 3) Where are we going? 2 One-sample t-test (univariate).................................................. 3 Two-sample t-test (univariate).................................................

More information

Theory of Errors and Least Squares Adjustment

Theory of Errors and Least Squares Adjustment Theory of Errors and Least Squares Adjustment Huaan Fan ISBN 91-7170-200-8 Royal Institute of Technology (KTH) Division of Geodesy and Geoinformatics 100 44 Stockholm Sweden August 2010 Theory of Errors

More information

Statistical Machine Learning

Statistical Machine Learning Statistical Machine Learning UoC Stats 37700, Winter quarter Lecture 4: classical linear and quadratic discriminants. 1 / 25 Linear separation For two classes in R d : simple idea: separate the classes

More information

Multivariate Analysis of Ecological Data

Multivariate Analysis of Ecological Data Multivariate Analysis of Ecological Data MICHAEL GREENACRE Professor of Statistics at the Pompeu Fabra University in Barcelona, Spain RAUL PRIMICERIO Associate Professor of Ecology, Evolutionary Biology

More information

Exact Nonparametric Tests for Comparing Means - A Personal Summary

Exact Nonparametric Tests for Comparing Means - A Personal Summary Exact Nonparametric Tests for Comparing Means - A Personal Summary Karl H. Schlag European University Institute 1 December 14, 2006 1 Economics Department, European University Institute. Via della Piazzuola

More information

Normalization and Mixed Degrees of Integration in Cointegrated Time Series Systems

Normalization and Mixed Degrees of Integration in Cointegrated Time Series Systems Normalization and Mixed Degrees of Integration in Cointegrated Time Series Systems Robert J. Rossana Department of Economics, 04 F/AB, Wayne State University, Detroit MI 480 E-Mail: r.j.rossana@wayne.edu

More information

1. Introduction to multivariate data

1. Introduction to multivariate data . Introduction to multivariate data. Books Chat eld, C. and A.J.Collins, Introduction to multivariate analysis. Chapman & Hall Krzanowski, W.J. Principles of multivariate analysis. Oxford.000 Johnson,

More information

1 Another method of estimation: least squares

1 Another method of estimation: least squares 1 Another method of estimation: least squares erm: -estim.tex, Dec8, 009: 6 p.m. (draft - typos/writos likely exist) Corrections, comments, suggestions welcome. 1.1 Least squares in general Assume Y i

More information

Quadratic forms Cochran s theorem, degrees of freedom, and all that

Quadratic forms Cochran s theorem, degrees of freedom, and all that Quadratic forms Cochran s theorem, degrees of freedom, and all that Dr. Frank Wood Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 1, Slide 1 Why We Care Cochran s theorem tells us

More information

3.1 Least squares in matrix form

3.1 Least squares in matrix form 118 3 Multiple Regression 3.1 Least squares in matrix form E Uses Appendix A.2 A.4, A.6, A.7. 3.1.1 Introduction More than one explanatory variable In the foregoing chapter we considered the simple regression

More information

y t by left multiplication with 1 (L) as y t = 1 (L) t =ª(L) t 2.5 Variance decomposition and innovation accounting Consider the VAR(p) model where

y t by left multiplication with 1 (L) as y t = 1 (L) t =ª(L) t 2.5 Variance decomposition and innovation accounting Consider the VAR(p) model where . Variance decomposition and innovation accounting Consider the VAR(p) model where (L)y t = t, (L) =I m L L p L p is the lag polynomial of order p with m m coe±cient matrices i, i =,...p. Provided that

More information

Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model

Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model 1 September 004 A. Introduction and assumptions The classical normal linear regression model can be written

More information

Portfolio selection based on upper and lower exponential possibility distributions

Portfolio selection based on upper and lower exponential possibility distributions European Journal of Operational Research 114 (1999) 115±126 Theory and Methodology Portfolio selection based on upper and lower exponential possibility distributions Hideo Tanaka *, Peijun Guo Department

More information

Metric Multidimensional Scaling (MDS): Analyzing Distance Matrices

Metric Multidimensional Scaling (MDS): Analyzing Distance Matrices Metric Multidimensional Scaling (MDS): Analyzing Distance Matrices Hervé Abdi 1 1 Overview Metric multidimensional scaling (MDS) transforms a distance matrix into a set of coordinates such that the (Euclidean)

More information

Eigenvalues, Eigenvectors, Matrix Factoring, and Principal Components

Eigenvalues, Eigenvectors, Matrix Factoring, and Principal Components Eigenvalues, Eigenvectors, Matrix Factoring, and Principal Components The eigenvalues and eigenvectors of a square matrix play a key role in some important operations in statistics. In particular, they

More information

Chapter 2. Dynamic panel data models

Chapter 2. Dynamic panel data models Chapter 2. Dynamic panel data models Master of Science in Economics - University of Geneva Christophe Hurlin, Université d Orléans Université d Orléans April 2010 Introduction De nition We now consider

More information

Mehtap Ergüven Abstract of Ph.D. Dissertation for the degree of PhD of Engineering in Informatics

Mehtap Ergüven Abstract of Ph.D. Dissertation for the degree of PhD of Engineering in Informatics INTERNATIONAL BLACK SEA UNIVERSITY COMPUTER TECHNOLOGIES AND ENGINEERING FACULTY ELABORATION OF AN ALGORITHM OF DETECTING TESTS DIMENSIONALITY Mehtap Ergüven Abstract of Ph.D. Dissertation for the degree

More information

Lecture 1: Systems of Linear Equations

Lecture 1: Systems of Linear Equations MTH Elementary Matrix Algebra Professor Chao Huang Department of Mathematics and Statistics Wright State University Lecture 1 Systems of Linear Equations ² Systems of two linear equations with two variables

More information

Part 2: Analysis of Relationship Between Two Variables

Part 2: Analysis of Relationship Between Two Variables Part 2: Analysis of Relationship Between Two Variables Linear Regression Linear correlation Significance Tests Multiple regression Linear Regression Y = a X + b Dependent Variable Independent Variable

More information

Representation of functions as power series

Representation of functions as power series Representation of functions as power series Dr. Philippe B. Laval Kennesaw State University November 9, 008 Abstract This document is a summary of the theory and techniques used to represent functions

More information

Variance Reduction. Pricing American Options. Monte Carlo Option Pricing. Delta and Common Random Numbers

Variance Reduction. Pricing American Options. Monte Carlo Option Pricing. Delta and Common Random Numbers Variance Reduction The statistical efficiency of Monte Carlo simulation can be measured by the variance of its output If this variance can be lowered without changing the expected value, fewer replications

More information

University of Saskatchewan Department of Economics Economics 414.3 Homework #1

University of Saskatchewan Department of Economics Economics 414.3 Homework #1 Homework #1 1. In 1900 GDP per capita in Japan (measured in 2000 dollars) was $1,433. In 2000 it was $26,375. (a) Calculate the growth rate of income per capita in Japan over this century. (b) Now suppose

More information

Least-Squares Intersection of Lines

Least-Squares Intersection of Lines Least-Squares Intersection of Lines Johannes Traa - UIUC 2013 This write-up derives the least-squares solution for the intersection of lines. In the general case, a set of lines will not intersect at a

More information

Risk Aversion. Expected value as a criterion for making decisions makes sense provided that C H A P T E R 2. 2.1 Risk Attitude

Risk Aversion. Expected value as a criterion for making decisions makes sense provided that C H A P T E R 2. 2.1 Risk Attitude C H A P T E R 2 Risk Aversion Expected value as a criterion for making decisions makes sense provided that the stakes at risk in the decision are small enough to \play the long run averages." The range

More information

Logistic Regression. Jia Li. Department of Statistics The Pennsylvania State University. Logistic Regression

Logistic Regression. Jia Li. Department of Statistics The Pennsylvania State University. Logistic Regression Logistic Regression Department of Statistics The Pennsylvania State University Email: jiali@stat.psu.edu Logistic Regression Preserve linear classification boundaries. By the Bayes rule: Ĝ(x) = arg max

More information

SF2940: Probability theory Lecture 8: Multivariate Normal Distribution

SF2940: Probability theory Lecture 8: Multivariate Normal Distribution SF2940: Probability theory Lecture 8: Multivariate Normal Distribution Timo Koski 24.09.2014 Timo Koski () Mathematisk statistik 24.09.2014 1 / 75 Learning outcomes Random vectors, mean vector, covariance

More information

Multivariate Statistical Inference and Applications

Multivariate Statistical Inference and Applications Multivariate Statistical Inference and Applications ALVIN C. RENCHER Department of Statistics Brigham Young University A Wiley-Interscience Publication JOHN WILEY & SONS, INC. New York Chichester Weinheim

More information

Java Modules for Time Series Analysis

Java Modules for Time Series Analysis Java Modules for Time Series Analysis Agenda Clustering Non-normal distributions Multifactor modeling Implied ratings Time series prediction 1. Clustering + Cluster 1 Synthetic Clustering + Time series

More information

SF2940: Probability theory Lecture 8: Multivariate Normal Distribution

SF2940: Probability theory Lecture 8: Multivariate Normal Distribution SF2940: Probability theory Lecture 8: Multivariate Normal Distribution Timo Koski 24.09.2015 Timo Koski Matematisk statistik 24.09.2015 1 / 1 Learning outcomes Random vectors, mean vector, covariance matrix,

More information

Partial Fractions Decomposition

Partial Fractions Decomposition Partial Fractions Decomposition Dr. Philippe B. Laval Kennesaw State University August 6, 008 Abstract This handout describes partial fractions decomposition and how it can be used when integrating rational

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.cs.toronto.edu/~rsalakhu/ Lecture 6 Three Approaches to Classification Construct

More information

Introduction to Matrix Algebra

Introduction to Matrix Algebra Psychology 7291: Multivariate Statistics (Carey) 8/27/98 Matrix Algebra - 1 Introduction to Matrix Algebra Definitions: A matrix is a collection of numbers ordered by rows and columns. It is customary

More information

Linear Threshold Units

Linear Threshold Units Linear Threshold Units w x hx (... w n x n w We assume that each feature x j and each weight w j is a real number (we will relax this later) We will study three different algorithms for learning linear

More information

Empirical Methods in Applied Economics

Empirical Methods in Applied Economics Empirical Methods in Applied Economics Jörn-Ste en Pischke LSE October 2005 1 Observational Studies and Regression 1.1 Conditional Randomization Again When we discussed experiments, we discussed already

More information

171:290 Model Selection Lecture II: The Akaike Information Criterion

171:290 Model Selection Lecture II: The Akaike Information Criterion 171:290 Model Selection Lecture II: The Akaike Information Criterion Department of Biostatistics Department of Statistics and Actuarial Science August 28, 2012 Introduction AIC, the Akaike Information

More information

Factor Analysis. Chapter 420. Introduction

Factor Analysis. Chapter 420. Introduction Chapter 420 Introduction (FA) is an exploratory technique applied to a set of observed variables that seeks to find underlying factors (subsets of variables) from which the observed variables were generated.

More information

Bias in the Estimation of Mean Reversion in Continuous-Time Lévy Processes

Bias in the Estimation of Mean Reversion in Continuous-Time Lévy Processes Bias in the Estimation of Mean Reversion in Continuous-Time Lévy Processes Yong Bao a, Aman Ullah b, Yun Wang c, and Jun Yu d a Purdue University, IN, USA b University of California, Riverside, CA, USA

More information

Predict the Popularity of YouTube Videos Using Early View Data

Predict the Popularity of YouTube Videos Using Early View Data 000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015 016 017 018 019 020 021 022 023 024 025 026 027 028 029 030 031 032 033 034 035 036 037 038 039 040 041 042 043 044 045 046 047 048 049 050

More information

Chapter 3 RANDOM VARIATE GENERATION

Chapter 3 RANDOM VARIATE GENERATION Chapter 3 RANDOM VARIATE GENERATION In order to do a Monte Carlo simulation either by hand or by computer, techniques must be developed for generating values of random variables having known distributions.

More information

Understanding the Impact of Weights Constraints in Portfolio Theory

Understanding the Impact of Weights Constraints in Portfolio Theory Understanding the Impact of Weights Constraints in Portfolio Theory Thierry Roncalli Research & Development Lyxor Asset Management, Paris thierry.roncalli@lyxor.com January 2010 Abstract In this article,

More information

Chapter 6: Multivariate Cointegration Analysis

Chapter 6: Multivariate Cointegration Analysis Chapter 6: Multivariate Cointegration Analysis 1 Contents: Lehrstuhl für Department Empirische of Wirtschaftsforschung Empirical Research and und Econometrics Ökonometrie VI. Multivariate Cointegration

More information

IDENTIFICATION IN A CLASS OF NONPARAMETRIC SIMULTANEOUS EQUATIONS MODELS. Steven T. Berry and Philip A. Haile. March 2011 Revised April 2011

IDENTIFICATION IN A CLASS OF NONPARAMETRIC SIMULTANEOUS EQUATIONS MODELS. Steven T. Berry and Philip A. Haile. March 2011 Revised April 2011 IDENTIFICATION IN A CLASS OF NONPARAMETRIC SIMULTANEOUS EQUATIONS MODELS By Steven T. Berry and Philip A. Haile March 2011 Revised April 2011 COWLES FOUNDATION DISCUSSION PAPER NO. 1787R COWLES FOUNDATION

More information

Factorization Theorems

Factorization Theorems Chapter 7 Factorization Theorems This chapter highlights a few of the many factorization theorems for matrices While some factorization results are relatively direct, others are iterative While some factorization

More information

Factor analysis. Angela Montanari

Factor analysis. Angela Montanari Factor analysis Angela Montanari 1 Introduction Factor analysis is a statistical model that allows to explain the correlations between a large number of observed correlated variables through a small number

More information

Investment and Financial Constraints: Empirical Evidence for Firms in Brazil and China

Investment and Financial Constraints: Empirical Evidence for Firms in Brazil and China Investment and Financial Constraints: Empirical Evidence for Firms in Brazil and China Stephen R. Bond Nu eld College and Department of Economics, University of Oxford and Institute for Fiscal Studies

More information

SPECIAL PERTURBATIONS UNCORRELATED TRACK PROCESSING

SPECIAL PERTURBATIONS UNCORRELATED TRACK PROCESSING AAS 07-228 SPECIAL PERTURBATIONS UNCORRELATED TRACK PROCESSING INTRODUCTION James G. Miller * Two historical uncorrelated track (UCT) processing approaches have been employed using general perturbations

More information

Module 3: Correlation and Covariance

Module 3: Correlation and Covariance Using Statistical Data to Make Decisions Module 3: Correlation and Covariance Tom Ilvento Dr. Mugdim Pašiƒ University of Delaware Sarajevo Graduate School of Business O ften our interest in data analysis

More information

1 Introduction to Matrices

1 Introduction to Matrices 1 Introduction to Matrices In this section, important definitions and results from matrix algebra that are useful in regression analysis are introduced. While all statements below regarding the columns

More information

Review of Bivariate Regression

Review of Bivariate Regression Review of Bivariate Regression A.Colin Cameron Department of Economics University of California - Davis accameron@ucdavis.edu October 27, 2006 Abstract This provides a review of material covered in an

More information

u = [ 2 4 5] has one row with three components (a 3 v = [2 4 5] has three rows separated by semicolons (a 3 w = 2:5 generates the row vector w = [ 2 3

u = [ 2 4 5] has one row with three components (a 3 v = [2 4 5] has three rows separated by semicolons (a 3 w = 2:5 generates the row vector w = [ 2 3 MATLAB Tutorial You need a small numb e r of basic commands to start using MATLAB. This short tutorial describes those fundamental commands. You need to create vectors and matrices, to change them, and

More information

Review Jeopardy. Blue vs. Orange. Review Jeopardy

Review Jeopardy. Blue vs. Orange. Review Jeopardy Review Jeopardy Blue vs. Orange Review Jeopardy Jeopardy Round Lectures 0-3 Jeopardy Round $200 How could I measure how far apart (i.e. how different) two observations, y 1 and y 2, are from each other?

More information

Example: Credit card default, we may be more interested in predicting the probabilty of a default than classifying individuals as default or not.

Example: Credit card default, we may be more interested in predicting the probabilty of a default than classifying individuals as default or not. Statistical Learning: Chapter 4 Classification 4.1 Introduction Supervised learning with a categorical (Qualitative) response Notation: - Feature vector X, - qualitative response Y, taking values in C

More information

Forecast covariances in the linear multiregression dynamic model.

Forecast covariances in the linear multiregression dynamic model. Forecast covariances in the linear multiregression dynamic model. Catriona M Queen, Ben J Wright and Casper J Albers The Open University, Milton Keynes, MK7 6AA, UK February 28, 2007 Abstract The linear

More information

Algebra 2 Chapter 1 Vocabulary. identity - A statement that equates two equivalent expressions.

Algebra 2 Chapter 1 Vocabulary. identity - A statement that equates two equivalent expressions. Chapter 1 Vocabulary identity - A statement that equates two equivalent expressions. verbal model- A word equation that represents a real-life problem. algebraic expression - An expression with variables.

More information

Summary of Formulas and Concepts. Descriptive Statistics (Ch. 1-4)

Summary of Formulas and Concepts. Descriptive Statistics (Ch. 1-4) Summary of Formulas and Concepts Descriptive Statistics (Ch. 1-4) Definitions Population: The complete set of numerical information on a particular quantity in which an investigator is interested. We assume

More information

Department of Economics

Department of Economics Department of Economics On Testing for Diagonality of Large Dimensional Covariance Matrices George Kapetanios Working Paper No. 526 October 2004 ISSN 1473-0278 On Testing for Diagonality of Large Dimensional

More information

How To Solve A Minimum Set Covering Problem (Mcp)

How To Solve A Minimum Set Covering Problem (Mcp) Measuring Rationality with the Minimum Cost of Revealed Preference Violations Mark Dean and Daniel Martin Online Appendices - Not for Publication 1 1 Algorithm for Solving the MASP In this online appendix

More information

DEGREES OF FREEDOM - SIMPLIFIED

DEGREES OF FREEDOM - SIMPLIFIED 1 Aust. J. Geod. Photogram. Surv. Nos 46 & 47 December, 1987. pp 57-68 In 009 I retyped this paper and changed symbols eg ˆo σ to VF for my students benefit DEGREES OF FREEDOM - SIMPLIFIED Bruce R. Harvey

More information

c 2008 Je rey A. Miron We have described the constraints that a consumer faces, i.e., discussed the budget constraint.

c 2008 Je rey A. Miron We have described the constraints that a consumer faces, i.e., discussed the budget constraint. Lecture 2b: Utility c 2008 Je rey A. Miron Outline: 1. Introduction 2. Utility: A De nition 3. Monotonic Transformations 4. Cardinal Utility 5. Constructing a Utility Function 6. Examples of Utility Functions

More information

The Optimality of Naive Bayes

The Optimality of Naive Bayes The Optimality of Naive Bayes Harry Zhang Faculty of Computer Science University of New Brunswick Fredericton, New Brunswick, Canada email: hzhang@unbca E3B 5A3 Abstract Naive Bayes is one of the most

More information

Panel Data Econometrics

Panel Data Econometrics Panel Data Econometrics Master of Science in Economics - University of Geneva Christophe Hurlin, Université d Orléans University of Orléans January 2010 De nition A longitudinal, or panel, data set is

More information

Lecture 5: Singular Value Decomposition SVD (1)

Lecture 5: Singular Value Decomposition SVD (1) EEM3L1: Numerical and Analytical Techniques Lecture 5: Singular Value Decomposition SVD (1) EE3L1, slide 1, Version 4: 25-Sep-02 Motivation for SVD (1) SVD = Singular Value Decomposition Consider the system

More information

The PageRank Citation Ranking: Bring Order to the Web

The PageRank Citation Ranking: Bring Order to the Web The PageRank Citation Ranking: Bring Order to the Web presented by: Xiaoxi Pang 25.Nov 2010 1 / 20 Outline Introduction A ranking for every page on the Web Implementation Convergence Properties Personalized

More information

Goodness of fit assessment of item response theory models

Goodness of fit assessment of item response theory models Goodness of fit assessment of item response theory models Alberto Maydeu Olivares University of Barcelona Madrid November 1, 014 Outline Introduction Overall goodness of fit testing Two examples Assessing

More information

Advanced Microeconomics

Advanced Microeconomics Advanced Microeconomics Ordinal preference theory Harald Wiese University of Leipzig Harald Wiese (University of Leipzig) Advanced Microeconomics 1 / 68 Part A. Basic decision and preference theory 1 Decisions

More information

Fitting Subject-specific Curves to Grouped Longitudinal Data

Fitting Subject-specific Curves to Grouped Longitudinal Data Fitting Subject-specific Curves to Grouped Longitudinal Data Djeundje, Viani Heriot-Watt University, Department of Actuarial Mathematics & Statistics Edinburgh, EH14 4AS, UK E-mail: vad5@hw.ac.uk Currie,

More information

Introduction to Multivariate Analysis

Introduction to Multivariate Analysis Introduction to Multivariate Analysis Lecture 1 August 24, 2005 Multivariate Analysis Lecture #1-8/24/2005 Slide 1 of 30 Today s Lecture Today s Lecture Syllabus and course overview Chapter 1 (a brief

More information

How to report the percentage of explained common variance in exploratory factor analysis

How to report the percentage of explained common variance in exploratory factor analysis UNIVERSITAT ROVIRA I VIRGILI How to report the percentage of explained common variance in exploratory factor analysis Tarragona 2013 Please reference this document as: Lorenzo-Seva, U. (2013). How to report

More information

Solution to Homework 2

Solution to Homework 2 Solution to Homework 2 Olena Bormashenko September 23, 2011 Section 1.4: 1(a)(b)(i)(k), 4, 5, 14; Section 1.5: 1(a)(b)(c)(d)(e)(n), 2(a)(c), 13, 16, 17, 18, 27 Section 1.4 1. Compute the following, if

More information

Lecture 3: Linear methods for classification

Lecture 3: Linear methods for classification Lecture 3: Linear methods for classification Rafael A. Irizarry and Hector Corrada Bravo February, 2010 Today we describe four specific algorithms useful for classification problems: linear regression,

More information

Chapter 4: Statistical Hypothesis Testing

Chapter 4: Statistical Hypothesis Testing Chapter 4: Statistical Hypothesis Testing Christophe Hurlin November 20, 2015 Christophe Hurlin () Advanced Econometrics - Master ESA November 20, 2015 1 / 225 Section 1 Introduction Christophe Hurlin

More information

Multiple Optimization Using the JMP Statistical Software Kodak Research Conference May 9, 2005

Multiple Optimization Using the JMP Statistical Software Kodak Research Conference May 9, 2005 Multiple Optimization Using the JMP Statistical Software Kodak Research Conference May 9, 2005 Philip J. Ramsey, Ph.D., Mia L. Stephens, MS, Marie Gaudard, Ph.D. North Haven Group, http://www.northhavengroup.com/

More information

Partial Least Squares (PLS) Regression.

Partial Least Squares (PLS) Regression. Partial Least Squares (PLS) Regression. Hervé Abdi 1 The University of Texas at Dallas Introduction Pls regression is a recent technique that generalizes and combines features from principal component

More information

Joint Exam 1/P Sample Exam 1

Joint Exam 1/P Sample Exam 1 Joint Exam 1/P Sample Exam 1 Take this practice exam under strict exam conditions: Set a timer for 3 hours; Do not stop the timer for restroom breaks; Do not look at your notes. If you believe a question

More information

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS 1. SYSTEMS OF EQUATIONS AND MATRICES 1.1. Representation of a linear system. The general system of m equations in n unknowns can be written a 11 x 1 + a 12 x 2 +

More information

160 CHAPTER 4. VECTOR SPACES

160 CHAPTER 4. VECTOR SPACES 160 CHAPTER 4. VECTOR SPACES 4. Rank and Nullity In this section, we look at relationships between the row space, column space, null space of a matrix and its transpose. We will derive fundamental results

More information

The p-norm generalization of the LMS algorithm for adaptive filtering

The p-norm generalization of the LMS algorithm for adaptive filtering The p-norm generalization of the LMS algorithm for adaptive filtering Jyrki Kivinen University of Helsinki Manfred Warmuth University of California, Santa Cruz Babak Hassibi California Institute of Technology

More information

General Framework for an Iterative Solution of Ax b. Jacobi s Method

General Framework for an Iterative Solution of Ax b. Jacobi s Method 2.6 Iterative Solutions of Linear Systems 143 2.6 Iterative Solutions of Linear Systems Consistent linear systems in real life are solved in one of two ways: by direct calculation (using a matrix factorization,

More information

Discussion Papers Department of Economics University of Copenhagen

Discussion Papers Department of Economics University of Copenhagen Discussion Papers Department of Economics University of Copenhagen No. 8-34 esting for Co-integration in Vector Autoregressions with Non-Stationary Volatility Giuseppe Cavaliere, Anders Rahbek, and A.

More information

A statistical test for forecast evaluation under a discrete loss function

A statistical test for forecast evaluation under a discrete loss function A statistical test for forecast evaluation under a discrete loss function Francisco J. Eransus, Alfonso Novales Departamento de Economia Cuantitativa Universidad Complutense December 2010 Abstract We propose

More information

Time Series Analysis

Time Series Analysis Time Series Analysis Identifying possible ARIMA models Andrés M. Alonso Carolina García-Martos Universidad Carlos III de Madrid Universidad Politécnica de Madrid June July, 2012 Alonso and García-Martos

More information

Handling missing data in large data sets. Agostino Di Ciaccio Dept. of Statistics University of Rome La Sapienza

Handling missing data in large data sets. Agostino Di Ciaccio Dept. of Statistics University of Rome La Sapienza Handling missing data in large data sets Agostino Di Ciaccio Dept. of Statistics University of Rome La Sapienza The problem Often in official statistics we have large data sets with many variables and

More information

A MULTIVARIATE OUTLIER DETECTION METHOD

A MULTIVARIATE OUTLIER DETECTION METHOD A MULTIVARIATE OUTLIER DETECTION METHOD P. Filzmoser Department of Statistics and Probability Theory Vienna, AUSTRIA e-mail: P.Filzmoser@tuwien.ac.at Abstract A method for the detection of multivariate

More information

NCSS Statistical Software Principal Components Regression. In ordinary least squares, the regression coefficients are estimated using the formula ( )

NCSS Statistical Software Principal Components Regression. In ordinary least squares, the regression coefficients are estimated using the formula ( ) Chapter 340 Principal Components Regression Introduction is a technique for analyzing multiple regression data that suffer from multicollinearity. When multicollinearity occurs, least squares estimates

More information

Chapter 1 Introduction. 1.1 Introduction

Chapter 1 Introduction. 1.1 Introduction Chapter 1 Introduction 1.1 Introduction 1 1.2 What Is a Monte Carlo Study? 2 1.2.1 Simulating the Rolling of Two Dice 2 1.3 Why Is Monte Carlo Simulation Often Necessary? 4 1.4 What Are Some Typical Situations

More information

Chapter 5: The Cointegrated VAR model

Chapter 5: The Cointegrated VAR model Chapter 5: The Cointegrated VAR model Katarina Juselius July 1, 2012 Katarina Juselius () Chapter 5: The Cointegrated VAR model July 1, 2012 1 / 41 An intuitive interpretation of the Pi matrix Consider

More information

Multivariate Analysis of Variance (MANOVA)

Multivariate Analysis of Variance (MANOVA) Multivariate Analysis of Variance (MANOVA) Aaron French, Marcelo Macedo, John Poulsen, Tyler Waterson and Angela Yu Keywords: MANCOVA, special cases, assumptions, further reading, computations Introduction

More information

STATISTICS AND DATA ANALYSIS IN GEOLOGY, 3rd ed. Clarificationof zonationprocedure described onpp. 238-239

STATISTICS AND DATA ANALYSIS IN GEOLOGY, 3rd ed. Clarificationof zonationprocedure described onpp. 238-239 STATISTICS AND DATA ANALYSIS IN GEOLOGY, 3rd ed. by John C. Davis Clarificationof zonationprocedure described onpp. 38-39 Because the notation used in this section (Eqs. 4.8 through 4.84) is inconsistent

More information

7.6 Approximation Errors and Simpson's Rule

7.6 Approximation Errors and Simpson's Rule WileyPLUS: Home Help Contact us Logout Hughes-Hallett, Calculus: Single and Multivariable, 4/e Calculus I, II, and Vector Calculus Reading content Integration 7.1. Integration by Substitution 7.2. Integration

More information

Applications to Data Smoothing and Image Processing I

Applications to Data Smoothing and Image Processing I Applications to Data Smoothing and Image Processing I MA 348 Kurt Bryan Signals and Images Let t denote time and consider a signal a(t) on some time interval, say t. We ll assume that the signal a(t) is

More information

SOLVING LINEAR SYSTEMS

SOLVING LINEAR SYSTEMS SOLVING LINEAR SYSTEMS Linear systems Ax = b occur widely in applied mathematics They occur as direct formulations of real world problems; but more often, they occur as a part of the numerical analysis

More information

A Comparison of Option Pricing Models

A Comparison of Option Pricing Models A Comparison of Option Pricing Models Ekrem Kilic 11.01.2005 Abstract Modeling a nonlinear pay o generating instrument is a challenging work. The models that are commonly used for pricing derivative might

More information

Chapter 4: Vector Autoregressive Models

Chapter 4: Vector Autoregressive Models Chapter 4: Vector Autoregressive Models 1 Contents: Lehrstuhl für Department Empirische of Wirtschaftsforschung Empirical Research and und Econometrics Ökonometrie IV.1 Vector Autoregressive Models (VAR)...

More information