Generating Valid 4 4 Correlation Matrices



Similar documents
Lecture 5 Principal Minors and the Hessian

DATA ANALYSIS II. Matrix Algorithms

Introduction to Matrix Algebra

SECOND DERIVATIVE TEST FOR CONSTRAINED EXTREMA

5 INTEGER LINEAR PROGRAMMING (ILP) E. Amaldi Fondamenti di R.O. Politecnico di Milano 1

QUALITY ENGINEERING PROGRAM

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS

Systems of Linear Equations

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m

STA 4273H: Statistical Machine Learning

Discrete Optimization

SPECTRAL POLYNOMIAL ALGORITHMS FOR COMPUTING BI-DIAGONAL REPRESENTATIONS FOR PHASE TYPE DISTRIBUTIONS AND MATRIX-EXPONENTIAL DISTRIBUTIONS

Spatial Statistics Chapter 3 Basics of areal data and areal data modeling

Linear Programming in Matrix Form

Factoring Algorithms

October 3rd, Linear Algebra & Properties of the Covariance Matrix

Globally Optimal Crowdsourcing Quality Management

DETERMINANTS IN THE KRONECKER PRODUCT OF MATRICES: THE INCIDENCE MATRIX OF A COMPLETE GRAPH

ANALYSIS, THEORY AND DESIGN OF LOGISTIC REGRESSION CLASSIFIERS USED FOR VERY LARGE SCALE DATA MINING

What is Linear Programming?

Statistical machine learning, high dimension and big data

Unit 18 Determinants

The Matrix Elements of a 3 3 Orthogonal Matrix Revisited

On the Interaction and Competition among Internet Service Providers

7 Gaussian Elimination and LU Factorization

Matrix Differentiation

A Branch and Bound Algorithm for Solving the Binary Bi-level Linear Programming Problem

Transportation Polytopes: a Twenty year Update

The Characteristic Polynomial

More than you wanted to know about quadratic forms

Determinants in the Kronecker product of matrices: The incidence matrix of a complete graph

Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh

Robust Geometric Programming is co-np hard

Classification of Cartan matrices

Linear Programming I

How to report the percentage of explained common variance in exploratory factor analysis

Extending the debate between Spearman and Wilson 1929: When do single variables optimally reproduce the common part of the observed covariances?

Chapter 6. Linear Programming: The Simplex Method. Introduction to the Big M Method. Section 4 Maximization and Minimization with Problem Constraints

OR topics in MRP-II. Mads Jepsens. OR topics in MRP-II p.1/25

Nonlinear Programming Methods.S2 Quadratic Programming

Introduction to Markov Chain Monte Carlo

1 Introduction. Linear Programming. Questions. A general optimization problem is of the form: choose x to. max f(x) subject to x S. where.

A characterization of trace zero symmetric nonnegative 5x5 matrices

Lecture 11: 0-1 Quadratic Program and Lower Bounds

A Brief Study of the Nurse Scheduling Problem (NSP)

Lecture 2: August 29. Linear Programming (part I)

NMR Measurement of T1-T2 Spectra with Partial Measurements using Compressive Sensing

Notes on Determinant

The Trip Scheduling Problem

Course Syllabus For Operations Management. Management Information Systems

Linear Programming for Optimization. Mark A. Schulze, Ph.D. Perceptive Scientific Instruments, Inc.

A Gradient Formula for Linear Chance Constraints Under Gaussian Distribution

SF2940: Probability theory Lecture 8: Multivariate Normal Distribution

Special Situations in the Simplex Algorithm

Multivariate Analysis of Variance (MANOVA): I. Theory

Duality of linear conic problems

Mehtap Ergüven Abstract of Ph.D. Dissertation for the degree of PhD of Engineering in Informatics

A note on companion matrices

Sensitivity analysis of utility based prices and risk-tolerance wealth processes

Chapter 5. Linear Inequalities and Linear Programming. Linear Programming in Two Dimensions: A Geometric Approach

Single item inventory control under periodic review and a minimum order quantity

Linear Programming Problems

a 11 x 1 + a 12 x a 1n x n = b 1 a 21 x 1 + a 22 x a 2n x n = b 2.

Lecture 3: Linear methods for classification

The Monte Carlo Framework, Examples from Finance and Generating Correlated Random Variables

December 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B. KITCHENS

The Master s Degree with Thesis Course Descriptions in Industrial Engineering

Solution of Linear Systems

Inner Product Spaces and Orthogonality

What Is School Mathematics?

Monte Carlo Methods in Finance

Production Planning Solution Techniques Part 1 MRP, MRP-II

Direct Methods for Solving Linear Systems. Matrix Factorization

FIXED CHARGE UNBALANCED TRANSPORTATION PROBLEM IN INVENTORY POOLING WITH MULTIPLE RETAILERS

Lecture 15 An Arithmetic Circuit Lowerbound and Flows in Graphs

1 Determinants and the Solvability of Linear Systems

LOGISTIC REGRESSION. Nitin R Patel. where the dependent variable, y, is binary (for convenience we often code these values as

A linear algebraic method for pricing temporary life annuities

How To Solve A Save Problem With Time T

Chapter 1 Introduction. 1.1 Introduction

Data Mining: Algorithms and Applications Matrix Math Review

Vector and Matrix Norms

4.5 Linear Dependence and Linear Independence

Hacking-proofness and Stability in a Model of Information Security Networks

IBM Research Report. A Pricing Problem under Monge Property

Solving polynomial least squares problems via semidefinite programming relaxations

Least-Squares Intersection of Lines

2.3 Convex Constrained Optimization Problems

Sample Size Planning for the Squared Multiple Correlation Coefficient: Accuracy in Parameter Estimation via Narrow Confidence Intervals

Mathematical finance and linear programming (optimization)

Proximal mapping via network optimization

Current Standard: Mathematical Concepts and Applications Shape, Space, and Measurement- Primary

24. The Branch and Bound Method

Question 2: How do you solve a matrix equation using the matrix inverse?

Compact Representations and Approximations for Compuation in Games

FUZZY CLUSTERING ANALYSIS OF DATA MINING: APPLICATION TO AN ACCIDENT MINING SYSTEM

Analysis of a Production/Inventory System with Multiple Retailers

MATLAB and Big Data: Illustrative Example

THE DYING FIBONACCI TREE. 1. Introduction. Consider a tree with two types of nodes, say A and B, and the following properties:

Revenue Management for Transportation Problems

Transcription:

Applied Mathematics E-Notes, 7(2007), 53-59 c ISSN 1607-2510 Available free at mirror sites of http://www.math.nthu.edu.tw/ amen/ Generating Valid 4 4 Correlation Matrices Mark Budden, Paul Hadavas, Lorrie Hoffman, Chris Pretz Received 10 March 2006 Abstract In this article, we provide an algorithm for generating valid 4 4 correlation matrices by creating bounds for the remaining three correlations that insure positive semidefiniteness once correlations between one variable and the other three are supplied. This is achieved by attending to the constraints placed on the leading principal minor determinants. Prior work in this area is restricted to 3 3 matrices or provides a means to investigate whether a 4 4 matrix is valid but does not offer a method of construction. We do offer such a method and supply a computer program to deliver a 4 4 positive semidefinite correlation matrix. 1 Introduction The level of sophistication and complexity of statistical inquiry has placed demands on researchers to increasingly employ simulation techniques. Frequently, distributions that reflect the essence of an application are assumed to be multivariate normal with a general prior knowledge of the correlation structure. More theoretical investigations such as those in the area of Bayesian analysis rely on Markov Chain Monte Carlo methods to investigate posterior distributions. Liechty, Liechty and Muller [4] encounter a challenge when conducting their simulations due to the awkward manner in which r ij is embedded in the likelihood (p. 6) noting that the requirement of positive semidefiniteness is the constraint that imposes an analysis dealing with truncated distributions. This leads them to develop a procedure that generates normalizing constants. This positive semidefinite matrix complication is encountered throughout the applied literature and sometimes leads to erroneous outcomes when simulation results are aggregated while using an invalid correlation matrix. The need to forecast demand for a group of products in order to realize savings by properly managing inventories is one such application requiring use of correlation matrices. Xu and Evers [10] offer a proof that partial pooling of customers can not lead to fewer resource requirements than complete pooling as was indicated in Tyagi and Das [9]. They further elucidate their position by showing that the error in the examples offered by these authors centered on the use of infeasible correlation matrices. Mathematics Subject Classifications: 65C05, 65C60. Department of Mathematics, Armstrong Atlantic State University, Savannah, GA 31419, USA The first author was supported in part by AASU Internal Grant #727188 Department of Statistics, University of Wyoming, Dept. 3332, 1000 E. University Ave., Laramie, WY 82071, USA 53

54 Generating Correlation Matrices Xu and Evers reveal that the matrices are not positive semidefinite and do so in the three variable case by revisiting boundary conditions offered by Marsaglia and Olkin [5] and via a computer program they developed to produce eigenvalues for the four variable case. Xu and Evers indicate that such an oversight is understandable because of the complex nature and interrelationships inherent in correlation structures particularly for non-statisticians like us ([10], p. 301). We contend that even statisticians and mathematicians find the problem of developing feasible correlation matrices to be challenging. The work done by Xu and Evers is helpful in that it prevents a researcher from offering an infeasible example or conducting a meaningless simulation due to nonpositive semidefinite correlation matrices. Their approach, though, is not constructive, as they say a correlation matrix is checked, (but) the determination of boundaries necessary for constructing feasible matrices is not performed ([10], p. 305). We offer such a method in the four variable case. The paper proceeds as follows. In the next section, previous results for 3 3 correlation matrices are explained. The generation of valid 4 4 correlation matrices is detailed in section 3. Finally, in section 4, we illustrate the utility of this constructive approach with an example from Xu and Evers [10]. Additionally, in section 4, other examples of valid 4 4 correlation matrices are provided along with directions for further research. The authors would like to thank Mortaza Jamshidian and Peter Rousseeuw for reading a preliminary draft of this paper and providing several helpful comments. 2 Review of 3 3 correlation matrices The method described in this section for generating 3 3 correlation matrices has been noted by many sources. Although the bounds described here were known earlier, Stanley and Wang [8] gave the first proof of the bounds in 1969. Other (geometric) proofs were subsequently given by Glass and Collins [1] and Leung and Lam [3]. Olkin [6] investigated the general question of how correlations are restricted when a submatrix of a correlation matrix is fixed. More recently, Rousseeuw and Molenberghs [7] used these bounds to investigate the shape and the volume of the set of valid 3 3 correlation matrices. Let x 1, x 2, and x 3 be random variables and let r ij denote the correlation coefficient for the variables x i and x j where i, j {1, 2, 3}. A3 3 correlation matrix is a positive semidefinite matrix of the form A = 1 r 12 r 13 r 12 1 r 23. r 13 r 23 1 To produce a valid correlation matrix, we begin by randomly picking the correlation coefficients 1 <r 12 < 1 and 1 <r 13 < 1. Here, and throughout the entire paper, we exclude the possibilities of having correlations of ±1 since such correlations indicate the redundancy of one variable where it is a perfect linear combination of another and can

Budden et al. 55 thus be eliminated from any analysis. Once these coefficients are selected, the possible range of values for r 23 can be described by considering the determinant This determinant is equal to 0 when deta = r 2 23 +(2r 12r 13 )r 23 +(1 r 2 12 r2 13 ) 0. r 23 = r 12 r 13 ± (1 r 2 12 )(1 r2 13 ) and it obtains its maximum value (for r 12 and r 13 fixed) of (1 r 2 12)(1 r 2 13) 1 when r 23 = r 12 r 13. Thus, A is a correlation matrix exactly when r 23 satisfies r 12 r 13 (1 r12 2 )(1 r2 13 ) r 23 r 12 r 13 + (1 r12 2 )(1 r2 13 ). (1) Again, we do not allow r 23 = ±1 for the reason described above. A special case to notice is when r 12 r 13 = 0. Without loss of generality, assume that r 12 = 0. The range of possible values of r 23 is given by 1 r13 2 r 23 1 r13 2. From this one sees that if r 12 =0=r 13, then r 23 can take on any possible correlation. 3 Algorithmic Generation of 4 4 Correlation Matrices We now consider a 4 4 matrix of the form A = 1 r 12 r 13 r 14 r 12 1 r 23 r 24 r 13 r 23 1 r 34 r 14 r 24 r 34 1 The correlations r 12, r 13, and r 14 can be randomly picked from the interval ( 1, 1). To find suitable ranges for the other correlations, we use the fact that a symmetric matrix is positive semidefinite if and only if all of its symmetric submatrices (including itself) have a nonnegative determinant (see. [2]). In other words, A is a correlation matrix if and only if deta 0 and every matrix of the form A ijk = 1 r ij r ik r ij 1 r jk r ik r jk 1. is a correlation matrix for i, j, k {1, 2, 3, 4} (no two of i, j, and k are equal).

56 Generating Correlation Matrices For each remaining correlation r jk {r 23,r 24,r 34 }, there are three limitations on the range of possible values. With the symmetry of the matrix, we use the fact that r ij = r ji for all i, j to simplify the notation used in the bounds for r 23,r 24,r 34. First, the bounds are restricted by det(a 1jk ) 0. From the previous section, we see that where jk jk r jk U (1) jk, := r 1jr 1k (1 r1j 2 )(1 r2 (1) 1k ) and U jk := r 1jr 1k + (1 r1j 2 )(1 r2 1k ). Next, we consider the restrictions imposed on the range of r jk by considering det(a ijk ) 0. Again, from the previous section, we have where L (2) jk L (2) jk r jk U (2) jk, := r ijr ik (1 rij 2 )(1 r2 (2) ik ) and U jk := r ijr ik + (1 rij 2 )(1 r2 ik ). Finally, we look at the restrictions imposed by det(a) 0. Noting that deta =1 r 2 12 r 2 13 r 2 14 r 2 23 r 2 24 r 2 34 + r 2 12r 2 34 + r 2 13r 2 24 + r 2 14r 2 23 +2r 23 r 24 r 34 +2r 12 r 13 r 23 +2r 12 r 14 r 24 +2r 13 r 14 r 34 2r 12 r 14 r 23 r 34 2r 13 r 14 r 23 r 24 2r 12 r 13 r 24 r 34, can be factored as a quadratic with respect to each r jk {r 23,r 24,r 34 }, we find the bounds L (3) jk r jk U (3) jk, where and L (3) jk := r jir ki + r 1j r 1k r 1j r 1i r ki r 1k r 1i r ji det(a 1ji ) det(a 1ki ) 1 r 2 1i U (3) jk := r jir ki + r 1j r 1k r 1j r 1i r ki r 1k r 1i r ji + det(a 1ji ) det(a 1ki ) 1 r 2 1i Now we describe the algorithm using the bounds from above. Once r 12, r 13, and r 14 are fixed, we choose r 23 from the range where L (2) 23 and L(3) 23 max{ 23,L(2) 23,L(3) 23 } r 23 min{u (1) 23,U(2) 23,U(3) 23 }, are minimized and U (2) 23 and U (3) 23 24 r 24 U (1) 24 and 34 r 34 U (1) 34. are maximized over the regions

Budden et al. 57 This guarantees the existence of some valid correlation matrix with the chosen values r 12, r 13, r 14, and r 23 since all of the conditions of positive semidefinite are met. A similar process is repeated for r 24. We choose r 24 from the range where L (2) 24 and L(3) 24 max{ 24,L(2) 24,L(3) 24 } r 24 min{u (1) 24,U(2) 24,U(3) 24 }, are minimized and U (2) 24 and U (3) 24 34 r 34 U (1) 34. are maximized over the region Finally, there is only one correlation left to bound. We choose r 34 from the range max{ 34,L(2) 34,L(3) 34 } r 34 min{u (1) 34,U(2) 34,U(3) 34 }. The constructive approach of this algorithm guarantees that any matrix formed in this way will be a valid correlation matrix. Furthermore, the ranges of possible values of each of the correlations r 23, r 24, and r 34, are the largest possible ranges so that there will exist some valid correlation matrix with any chosen value. In other words, any value chosen outside of the determined range will fail one of the necessary conditions for being valid. Thus, our algorithm provides both necessary and sufficient conditions for producing valid 4 4 correlation matrices. 4 Applications Having a method to create valid correlation matrices, we can now discuss the application presented by the resource allocation problem and present other examples that illustrate the difficulty of being able to guess proper correlation values. In offering a proof related to resource allocation, Xu and Evers [10] refute a prior claim by Tyagi and Das [9]. Xu and Evers explain that the error in logic was fostered by errors in the examples that inadvertently used invalid correlation matrices - ones that failed to be positive semidefinite. In the three variable case, Xu and Evers rely on boundary conditions offered by Marsaglia and Olkin [5] to establish that the matrix used by Tyagi and Das was not positive semidefinite. In the four variable case, they employ their own computer program developed to produce eigenvalues. The previous section contains a constructive algorithm that generates valid 4 4 correlation matrices. We have developed a computer program called validcor (2005) to accomplish this task. Our program delivers the remaining correlations after specifying those relating one variable to the other three. The program can be found in the general archive at Carnegie Mellon University s Statlab. http://lib.stat.cmu.edu To illustrate the utility of this constructive approach we use Tyagi and Das matrix 1.5.5.5.5 1.5.5.5.5 1.5.5.5.5 1

58 Generating Correlation Matrices and show that it would not be generated by our program. For this example we set r 12 =.5,r 13 =.5, and r 14 =.5. Feasible bounds found by our program for r 23 are ( 1,.5]. We chose r 23 =.5. Then, feasible bounds found for r 24 are [.5, 1). We chose r 24 =.5. Finally, feasible bounds found for r 34 are ( 1,.3333]. The r 34 value selected for the Tyagi and Das matrix was.5 and is outside the range that assures a positive semidefinite matrix. In addition to the use of the program showing how certain matrices are not valid correlation matrices, we provide the table below where each of the five rows correspond to generations of valid correlation matrices. The first three columns give the input parameters.the following columns give the bounds for the remaining correlations and the choices for r 23 and r 24 that led to the successive feasible bounds found by the computer program. Here, r 23 and r 24 represent those choices, respectively. No choice for r 34 is given as any value within the bounds will complete a valid matrix. Table 1: Examples of Valid Correlation Matrix Generations r 12 r 13 r 14 r 23 bounds r23 r 24 bounds r24 r 34 bounds.777.39.472 (.2766,.8827).88 (.1852,.9217).92 (.9835,.9957).981.397.961 (.2114,.5675).567 (.8895,.9964).996 (.6304,.6351).027.495.986 (.8552,.8819).881 (.1381,.1933).138 (.3441,.3462).801.913.6 (.4871,.9755).975 (.9553,.0017).002 (.2232,.2216).04.008.207 (.9988,.9995).754 (.9692,.9858).96 (.8716,.641) One interesting outcome to note from the table are the final bounds for the r 34 value. When both selections for r 23 and r 24 occur near their respective bounds, the interval for r 34 is severely limited. However, in the last example, a selection of r 23 not close to either bound allows r 34 to take on a wide range of possible values. This phenomenon along with its potential application to finding the hypervolume of possible 4 4 correlation matrices in 6-space will be explored in future research. Another direction will include the generalization of the algorithm to n n correlation matrices where n is greater than 4. Such a generalization is not trivial as the complexity of the bounds seems to grow exponentially with the size of the matrix. Most likely, the above algorithm will require a significant simplification before it can be extended to correlation matrices of arbitrary size. We save such an analysis for future study. References [1] G. Glass and J. Collins, Geometric proof of the restriction on the possible values of r xy when r xz and r yx are fixed, Educational and Psychological Measurement, 30(1970), 37-39.

Budden et al. 59 [2] R. Harris, A Primer of Multivariate Statistics, Lawrence Erlbaum Associates, Mahwah, NJ, 2001. [3] C.-K. Leung and K. Lam, A Note on the geometric representation of the correlation coefficients, The American Statistician, 29(1975), 128-130. [4] J.C. Liechty, M.W. Liechty, P. and Muller, Bayesian correlation estimation, Biometrika, 9(2004), 1-14. [5] G. Marsaglia and I. Olkin, Generating correlation-matrices, SIAM Journal on Scientific and Statistical Computing, 5(2)(1984), 470-475. [6] I. Olkin, Range restrictions for product-moment correlation matrices, Psychometrika, 46(1981), 469-472. [7] P. Rousseeuw and G. Molenberghs, The shape of correlation matrices, The American Statistician, 48(1994), 276-279. [8] J. Stanley and M. Wang, Restrictions on the possible values of r 12, given r 13 and r 23, Educational and Psychological Measurement, 29(1969), 579-581. [9] R. Tyagi and C. Das, Grouping customers for better allocation of resources to serve correlated demands, Computers and Operations Research, 26(1999), 1041-1058. [10] K. Xu and P. Evers, Managing single echelon inventories through demand aggregation and the feasibility of a correlation matrix, Computers and Operations Research, 30(2003), 297-308.