Lecture 7: Principal component analysis (PCA) What is PCA? Why use PCA?
|
|
- Lydia Hoover
- 7 years ago
- Views:
Transcription
1 Lecture 7: Princial comonent analysis (PCA) Rationale and use of PCA The underlying model (what is a rincial comonent anyway?) Eigenvectors and eigenvalues of the samle covariance matrix revisited! PCA scores and loadings The use of and rationale for rotations rthogonal and oblique rotations Comonent retention, significance, and reliability. L7. What is PCA? rom a set of Because the Z i s variables X, X,, X, (rincial comonents) we try and find are uncorrelated, they ( extract ) a set of measure different ordered indices Z, dimensions in the data. Z,, Z that are The hoe (sometimes uncorrelated and faint) is that most of the ordered in terms of variability in the original their variability: Var(Z ) set of variables will be > Var(Z ) > > Var(Z ) accounted for by c < comonents. L7. Why use PCA? PCA is generally used to reduce the number of variables considered in subsequent analyses, i.e. reduce the dimensionality of the data. Examles include: Reduce number of deendent variables in MANVA, mutivariate regression, correlation analysis, etc. Reduce number of indeendent variables (redictors) in regression analysis L7.3
2 Estimating rincial comonents The first rincial comonent is obtained by fitting (i.e. estimating the coefficients of) the linear function Z = a X j j which maximizes Var(Z ), subject to: a j = The second rincial comonent is obtained by fitting (i.e. estimating the coefficients of) the function Z = a X j j which maximizes Var(Z ), subject to: L7.4 a = Cov( Z Z j, ) = 0 Var( Z) Var( Z) Estimating rincial comonents (cont d) The third rincial comonent is obtained by fitting (i.e. estimating the coefficients of) the function Z = a X 3 3 j j which maximizes Var(Z 3 ), subject to: a3 j = as well as the additional constraints... Cov( Z, Z ) = 0 Cov( Z, Z ) = 0 Cov( Z, Z ) = 0 L7.5 and Var( Z3) Var( Z) Var( Z) 3 3 Estimating rincial comonents Estimation of the coefficients for each rincial comonent can be accomlished through several different methods (e.g. leastsquare estimation, maximum likelihood estimation, iterated rincial axis, etc.) The extracted rincial comonents may differ deending on the method of estimation. L7.6
3 The geometry of rincial comonents Princial comonents (Z i ) are linear functions of the original variables, and as such, define hyerlanes in the + - dimensional Z sace of Z and the original variables. Because the Z i s are uncorrelated, these lanes meet at right angles. X X Z X X L7.7 Multivariate variance: a geometric interretation Univariate variance is a measure of the volume occuied by samle oints in one dimension. Multivariate variance involving variables is the volume occuied by samle oints in an -dimensional sace. X Larger variance X ccuied volume X Smaller variance X L7.8 Multivariate variance: effects of correlations among variables X No correlation Correlations between airs of variables reduce the volume occuied by samle oints and hence, reduce the multivariate variance. ccuied volume X Positive correlation X Negative correlation X L7.9
4 C and the generalized multivariate variance L C = C N M Q P = 3 4 c o r = = 05. = cos θ, θ = 60 The determinant of the ss samle covariance matrix C is a generalized multivariate variance because area of a h arallelogram with sides θ s given by the individual standard deviations and s angle determined by the correlation between oosite h variables equals the sin 60 = = ; h = 3. hyotenuse determinant of C. Area = Base Height = 3, Area = C L7.0 Eigenvalues and eigenvectors of C No correlation Eigenvectors of the X covariance matrix C are orthogonal directed line segments that san the variation in the data, and the Positive X corresonding (unsigned) correlation eigenvalues are the length of these segments. X ξ so the roduct of the eigenvalues is the volume occuied by the data, i.e. the determinant of the covariance matrix. ξ X λ ξ ξ λ Negative correlation ξ ξ L7. The geometry of rincial comonents (cont d) The coefficients (a ij ) of the rincial comonents (Z i ) define vectors in the sace of coefficients. These vectors are the eigenvectors X (a i ) of the samle covariance matrix C, and the corresonding (unsigned) eigenvalues (λ i ) are the λ variances of each comonent, i.e. a Var(Z i )... 0 and the roduct of the eigenvalues is the volume occuied by the data, i.e. the - determinant of the covariance - 0 matrix. a X λ a a L7.
5 Another imortant relationshi! The sum of the eigenvalues of the covariance matrix C equals the sum of the diagonal elements of C, i.e. the trace of C. So, the sum of the variances of the rincial comonents equals the sum of the variances of the original variables. s = c C cm λ = i i= c c s m si i= c m cm sm = Tr( C) L7.3 Scale and the correlation matrix s = c C c m cm Since variables may be measured on different scales, and we want to cm cm sm eliminate scale effects, we ' X usually work with ik X c k ij X ik =, rij = standardized values so that sk sis j each variable is scaled to have zero mean and unit r r m variance. r r The samle covariance m C = R = matrix of standardized variables is the samle rm rm correlation matrix R. L7.4 c s Princial comonent scores Because rincial comonents are functions, we can lug in the values for each variable for each observation, and calculate a PC score for each observation and each rincial comonent. bservation X X a = S =. 07( 3. 7) (. 5) S S S = 0. 9( 3. 7) (. 5) =. 07(. 3) ( 0. ) = 0. 9(. 3) ( 0. ) L7.5
6 Princial comonent loadings Comonent loadings (L ij ) are the covariances (correlations for standardized values) of the original variables used in the PCA with the comonents, and are roortional to the comonent coefficients (a ij ). or each comonent, the (loading) for each variable summed over all variables equals the variance of the comonent. L = Cov( X, Z ) ij L = ka ij i= ij ij j i L = λ = Var( Z ) j j L7.6 More on loadings Sometimes comonents have variables with similar loadings, which form a natural grou. To assist in interretation, we may want to choose another comonent frame which emhasizes these differences among grous. ACTR() R () Loadings Variable Z Z Height Arm san Lower leg orearm Weight Uer thigh Chest width Chest girth actor lot REARM LWERLEG HEIGHT WEIGHT BITR CHESTGIR ACTR() L7.7 rthogonal rotations: varimax rthogonal (angle A WEIGHT BITR - CHESTGIR reserving): new (rotated) comonents are still uncorrelated ACTR().0 WEIGHT CHESTGIR BITR Varimax: rotation done so that each LWERLEG HEIGHT comonent loads high R () REARM on a small number of variables and low on - Varimax other variables (simlifies -.0 factors) ACTR() ACTR() R () C T.0 unrotated ACTR() REARM LWERLEG HEIGHT L7.8
7 rthogonal rotations: quartimax rthogonal (angle reserving): new (rotated) comonents are still uncorrelated Varimax: rotation done so that each variable loads mainly on one factor (simlified variables) ACTR() ACTR() R () R ().0 - unrotated REARM LWERLEG HEIGHT WEIGHT BITR CHESTGIR ACTR().0 WEIGHT CHESTGIR BITR - Varimax LWERLEG HEIGHT REARM ACTR() L7.9 rthogonal rotations: Equamax rthogonal (angle reserving): new (rotated) comonents are still uncorrelated Equamax: Combines varimax and quartimax. Number of variables that load highly on a factor and the number of factors needed to exlain the variable are otimized. - ACTR() ACTR() R () R () ACTR().0 unrotated Equamax LWERLEG HEIGHT WEIGHT CHESTGIR BITR REARM WEIGHT BITR CHESTGIR LWERLEG HEIGHT REARM ACTR() L7.0 blique rotations, e.g. blimin blique (non-angle reserving): new (rotated) comonents are now correlated Most reasonable when significant intercorrelations among factors exist. ACTR() R () ACTR() R () ACTR().0 CHESTGIR WEIGHT BITR - unrotated blimin REARM LWERLEG HEIGHT WEIGHT BITR CHESTGIR HEIGHT LWERLEG REARM ACTR() L7.
8 The consequences of rotation Unrotated comonents are () uncorrelated; () ordered in terms of decreasing variance (i.e., Var(Z ) > Var (Z ) > ). rthogonally rotated comonents are () still uncorrelated, but () need not be ordered in terms of decreasing variance (e.g. for Varimax rotation). bliquely rotated comonents are () correlated; () unordered (in general). L7. The rotated attern matrix for obliquely rotated factors The elements of the matrix are analogous to standardized artial regression coefficients from a multile regression analysis. So each element quantifies the imortance of the variable in question to the comonent, once the effects of other variables are controlled. Rotated Pattern Matrix (BLIMIN, Gamma = 000) HEIGHT REARM LWERLEG WEIGHT BITR CHESTGIR L7.3 The rotated structure matrix for obliquely rotated factors The elements of the rotated structure matrix are the simle correlations of the variable in question with the factor, i.e. the comonent loadings. or orthogonal factors, the factor attern and factor structure matrices are identical. Rotated Structure Matrix HEIGHT REARM LWERLEG WEIGHT BITR CHESTGIR L7.4
9 Which rotation is the best? bject: find the rotation which achieves the simlest structure among comonent loadings, thereby making interretation comaratively easy. Thurstone s criteria: for variables and m < comonents: () each comonent should have at least m nearzero loadings; () few comonents should have non-zero loadings on the same variable. L7.5 A final word on rotations You cannot say that any rotation is better than any other rotation from a statistical oint of view: all rotations are equally good statistically. Therefore, the choice among different rotations must be based on non-statistical grounds SAS STAT User s guide, Vol., L7.6 How many comonents to retain in subsequent analysis? Kaiser rule: retain only comonents with eigenvalues >. Scree test: lot eigenvalues against their ordinal numbers, retain all comonents in stee decent art of the curve. Retain as many factors as required to account for a secified amount of the total variance (e.g. 85%) Eigenvalue e u a l n v E ige Scree lot Kaiser threshold Number of actors L7.7
10 More on interretation: the significance of loadings Since loadings are correlation coefficients (r), we can test the null that each correlation equals zero. But analytic estimates of standard errors are often too small, esecially for rotated loadings. So, as a rule of thumb, use double the critical value to test significance. E.g., for N = 00, r(α = ) = 0.86, so significant factors have loadings greater than (0.86). L7.8 Comonent reliability: rules of thumb The absolute or N > 50, magnitude and number comonents with at of loadings are crucial least 0 loadings > for determining 0.40 are reliable. reliability Comonents with at least 4 loadings > 0.60 or with at least 3 loadings > 0.80 are reliable. L7.9 PCA: the rocedure. Calculate samle covariance matrix or correlation matrix. If all variables are on same scale, use samle covariance matrix, otherwise use correlation matrix.. Run PCA to extract unrotated comonents ( initial extraction ). 3. Decide which comonents to use in subsequent analysis based on Kaiser rule, Scree lots, etc. 4. Based on (3), rerun analysis using different orthogonal and oblique rotations and comare using factor lots ( follow-u extraction ) L7.30
11 PCA: the rocedure (cont d) 5. or obliquely rotated comonents, calculate correlations among comonents. Small correlations suggest that orthogonal rotations are reasonable. 6. Evaluate statistical significance of comonent loadings obtained from best rotation. 7. Check comonent reliability by redoing stes () - (6) with another (indeendent) data set, and comare the comonent loadings obtained from the two data sets. Are they close? L7.3
Principal Component Analysis
Principal Component Analysis Principle Component Analysis: A statistical technique used to examine the interrelations among a set of variables in order to identify the underlying structure of those variables.
More informationCommon factor analysis
Common factor analysis This is what people generally mean when they say "factor analysis" This family of techniques uses an estimate of common variance among the original variables to generate the factor
More informationFactor Analysis. Chapter 420. Introduction
Chapter 420 Introduction (FA) is an exploratory technique applied to a set of observed variables that seeks to find underlying factors (subsets of variables) from which the observed variables were generated.
More information4. There are no dependent variables specified... Instead, the model is: VAR 1. Or, in terms of basic measurement theory, we could model it as:
1 Neuendorf Factor Analysis Assumptions: 1. Metric (interval/ratio) data 2. Linearity (in the relationships among the variables--factors are linear constructions of the set of variables; the critical source
More informationFactor Analysis. Principal components factor analysis. Use of extracted factors in multivariate dependency models
Factor Analysis Principal components factor analysis Use of extracted factors in multivariate dependency models 2 KEY CONCEPTS ***** Factor Analysis Interdependency technique Assumptions of factor analysis
More informationA Multivariate Statistical Analysis of Stock Trends. Abstract
A Multivariate Statistical Analysis of Stock Trends Aril Kerby Alma College Alma, MI James Lawrence Miami University Oxford, OH Abstract Is there a method to redict the stock market? What factors determine
More informationPsychology 7291, Multivariate Analysis, Spring 2003. SAS PROC FACTOR: Suggestions on Use
: Suggestions on Use Background: Factor analysis requires several arbitrary decisions. The choices you make are the options that you must insert in the following SAS statements: PROC FACTOR METHOD=????
More informationReview Jeopardy. Blue vs. Orange. Review Jeopardy
Review Jeopardy Blue vs. Orange Review Jeopardy Jeopardy Round Lectures 0-3 Jeopardy Round $200 How could I measure how far apart (i.e. how different) two observations, y 1 and y 2, are from each other?
More informationA Brief Introduction to SPSS Factor Analysis
A Brief Introduction to SPSS Factor Analysis SPSS has a procedure that conducts exploratory factor analysis. Before launching into a step by step example of how to use this procedure, it is recommended
More informationExploratory Factor Analysis and Principal Components. Pekka Malo & Anton Frantsev 30E00500 Quantitative Empirical Research Spring 2016
and Principal Components Pekka Malo & Anton Frantsev 30E00500 Quantitative Empirical Research Spring 2016 Agenda Brief History and Introductory Example Factor Model Factor Equation Estimation of Loadings
More information2. Linearity (in relationships among the variables--factors are linear constructions of the set of variables) F 2 X 4 U 4
1 Neuendorf Factor Analysis Assumptions: 1. Metric (interval/ratio) data. Linearity (in relationships among the variables--factors are linear constructions of the set of variables) 3. Univariate and multivariate
More informationFactor Analysis. Advanced Financial Accounting II Åbo Akademi School of Business
Factor Analysis Advanced Financial Accounting II Åbo Akademi School of Business Factor analysis A statistical method used to describe variability among observed variables in terms of fewer unobserved variables
More informationIntroduction to Principal Components and FactorAnalysis
Introduction to Principal Components and FactorAnalysis Multivariate Analysis often starts out with data involving a substantial number of correlated variables. Principal Component Analysis (PCA) is a
More informationPrincipal Component Analysis
Principal Component Analysis ERS70D George Fernandez INTRODUCTION Analysis of multivariate data plays a key role in data analysis. Multivariate data consists of many different attributes or variables recorded
More informationMultivariate Analysis (Slides 13)
Multivariate Analysis (Slides 13) The final topic we consider is Factor Analysis. A Factor Analysis is a mathematical approach for attempting to explain the correlation between a large set of variables
More informationIntroduction to Principal Component Analysis: Stock Market Values
Chapter 10 Introduction to Principal Component Analysis: Stock Market Values The combination of some data and an aching desire for an answer does not ensure that a reasonable answer can be extracted from
More informationT-test & factor analysis
Parametric tests T-test & factor analysis Better than non parametric tests Stringent assumptions More strings attached Assumes population distribution of sample is normal Major problem Alternatives Continue
More informationExploratory Factor Analysis: rotation. Psychology 588: Covariance structure and factor models
Exploratory Factor Analysis: rotation Psychology 588: Covariance structure and factor models Rotational indeterminacy Given an initial (orthogonal) solution (i.e., Φ = I), there exist infinite pairs of
More informationDimensionality Reduction: Principal Components Analysis
Dimensionality Reduction: Principal Components Analysis In data mining one often encounters situations where there are a large number of variables in the database. In such situations it is very likely
More informationUnited Arab Emirates University College of Sciences Department of Mathematical Sciences HOMEWORK 1 SOLUTION. Section 10.1 Vectors in the Plane
United Arab Emirates University College of Sciences Deartment of Mathematical Sciences HOMEWORK 1 SOLUTION Section 10.1 Vectors in the Plane Calculus II for Engineering MATH 110 SECTION 0 CRN 510 :00 :00
More informationStatistics in Psychosocial Research Lecture 8 Factor Analysis I. Lecturer: Elizabeth Garrett-Mayer
This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike License. Your use of this material constitutes acceptance of that license and the conditions of use of materials on this
More informationIntroduction to Matrix Algebra
Psychology 7291: Multivariate Statistics (Carey) 8/27/98 Matrix Algebra - 1 Introduction to Matrix Algebra Definitions: A matrix is a collection of numbers ordered by rows and columns. It is customary
More informationChapter 7 Factor Analysis SPSS
Chapter 7 Factor Analysis SPSS Factor analysis attempts to identify underlying variables, or factors, that explain the pattern of correlations within a set of observed variables. Factor analysis is often
More informationFactor Analysis Example: SAS program (in blue) and output (in black) interleaved with comments (in red)
Factor Analysis Example: SAS program (in blue) and output (in black) interleaved with comments (in red) The following DATA procedure is to read input data. This will create a SAS dataset named CORRMATR
More informationCHAPTER 8 FACTOR EXTRACTION BY MATRIX FACTORING TECHNIQUES. From Exploratory Factor Analysis Ledyard R Tucker and Robert C.
CHAPTER 8 FACTOR EXTRACTION BY MATRIX FACTORING TECHNIQUES From Exploratory Factor Analysis Ledyard R Tucker and Robert C MacCallum 1997 180 CHAPTER 8 FACTOR EXTRACTION BY MATRIX FACTORING TECHNIQUES In
More informationFACTOR ANALYSIS NASC
FACTOR ANALYSIS NASC Factor Analysis A data reduction technique designed to represent a wide range of attributes on a smaller number of dimensions. Aim is to identify groups of variables which are relatively
More informationExploratory Factor Analysis Brian Habing - University of South Carolina - October 15, 2003
Exploratory Factor Analysis Brian Habing - University of South Carolina - October 15, 2003 FA is not worth the time necessary to understand it and carry it out. -Hills, 1977 Factor analysis should not
More informationExploratory Factor Analysis
Introduction Principal components: explain many variables using few new variables. Not many assumptions attached. Exploratory Factor Analysis Exploratory factor analysis: similar idea, but based on model.
More informationFactor analysis. Angela Montanari
Factor analysis Angela Montanari 1 Introduction Factor analysis is a statistical model that allows to explain the correlations between a large number of observed correlated variables through a small number
More informationOverview of Factor Analysis
Overview of Factor Analysis Jamie DeCoster Department of Psychology University of Alabama 348 Gordon Palmer Hall Box 870348 Tuscaloosa, AL 35487-0348 Phone: (205) 348-4431 Fax: (205) 348-8648 August 1,
More informationFACTOR ANALYSIS. Factor Analysis is similar to PCA in that it is a technique for studying the interrelationships among variables.
FACTOR ANALYSIS Introduction Factor Analysis is similar to PCA in that it is a technique for studying the interrelationships among variables Both methods differ from regression in that they don t have
More informationFactor Analysis. Sample StatFolio: factor analysis.sgp
STATGRAPHICS Rev. 1/10/005 Factor Analysis Summary The Factor Analysis procedure is designed to extract m common factors from a set of p quantitative variables X. In many situations, a small number of
More informationSteven M. Ho!and. Department of Geology, University of Georgia, Athens, GA 30602-2501
PRINCIPAL COMPONENTS ANALYSIS (PCA) Steven M. Ho!and Department of Geology, University of Georgia, Athens, GA 30602-2501 May 2008 Introduction Suppose we had measured two variables, length and width, and
More informationMultivariate Analysis
Table Of Contents Multivariate Analysis... 1 Overview... 1 Principal Components... 2 Factor Analysis... 5 Cluster Observations... 12 Cluster Variables... 17 Cluster K-Means... 20 Discriminant Analysis...
More informationA Brief Introduction to Factor Analysis
1. Introduction A Brief Introduction to Factor Analysis Factor analysis attempts to represent a set of observed variables X 1, X 2. X n in terms of a number of 'common' factors plus a factor which is unique
More informationStatistics for Business Decision Making
Statistics for Business Decision Making Faculty of Economics University of Siena 1 / 62 You should be able to: ˆ Summarize and uncover any patterns in a set of multivariate data using the (FM) ˆ Apply
More informationPrinciple Component Analysis and Partial Least Squares: Two Dimension Reduction Techniques for Regression
Principle Component Analysis and Partial Least Squares: Two Dimension Reduction Techniques for Regression Saikat Maitra and Jun Yan Abstract: Dimension reduction is one of the major tasks for multivariate
More informationFactor Rotations in Factor Analyses.
Factor Rotations in Factor Analyses. Hervé Abdi 1 The University of Texas at Dallas Introduction The different methods of factor analysis first extract a set a factors from a data set. These factors are
More informationExploratory Factor Analysis of Demographic Characteristics of Antenatal Clinic Attendees and their Association with HIV Risk
Doi:10.5901/mjss.2014.v5n20p303 Abstract Exploratory Factor Analysis of Demographic Characteristics of Antenatal Clinic Attendees and their Association with HIV Risk Wilbert Sibanda Philip D. Pretorius
More informationWhat is Rotating in Exploratory Factor Analysis?
A peer-reviewed electronic journal. Copyright is retained by the first or sole author, who grants right of first publication to the Practical Assessment, Research & Evaluation. Permission is granted to
More informationRachel J. Goldberg, Guideline Research/Atlanta, Inc., Duluth, GA
PROC FACTOR: How to Interpret the Output of a Real-World Example Rachel J. Goldberg, Guideline Research/Atlanta, Inc., Duluth, GA ABSTRACT THE METHOD This paper summarizes a real-world example of a factor
More informationSimilarity and Diagonalization. Similar Matrices
MATH022 Linear Algebra Brief lecture notes 48 Similarity and Diagonalization Similar Matrices Let A and B be n n matrices. We say that A is similar to B if there is an invertible n n matrix P such that
More informationMehtap Ergüven Abstract of Ph.D. Dissertation for the degree of PhD of Engineering in Informatics
INTERNATIONAL BLACK SEA UNIVERSITY COMPUTER TECHNOLOGIES AND ENGINEERING FACULTY ELABORATION OF AN ALGORITHM OF DETECTING TESTS DIMENSIONALITY Mehtap Ergüven Abstract of Ph.D. Dissertation for the degree
More informationHow to report the percentage of explained common variance in exploratory factor analysis
UNIVERSITAT ROVIRA I VIRGILI How to report the percentage of explained common variance in exploratory factor analysis Tarragona 2013 Please reference this document as: Lorenzo-Seva, U. (2013). How to report
More informationFactor Analysis. Factor Analysis
Factor Analysis Principal Components Analysis, e.g. of stock price movements, sometimes suggests that several variables may be responding to a small number of underlying forces. In the factor model, we
More informationCoordinate Transformation
Coordinate Transformation Coordinate Transformations In this chater, we exlore maings where a maing is a function that "mas" one set to another, usually in a way that reserves at least some of the underlyign
More information5.2 Customers Types for Grocery Shopping Scenario
------------------------------------------------------------------------------------------------------- CHAPTER 5: RESULTS AND ANALYSIS -------------------------------------------------------------------------------------------------------
More informationFactor Analysis Using SPSS
Psychology 305 p. 1 Factor Analysis Using SPSS Overview For this computer assignment, you will conduct a series of principal factor analyses to examine the factor structure of a new instrument developed
More informationBEHAVIOR BASED CREDIT CARD FRAUD DETECTION USING SUPPORT VECTOR MACHINES
BEHAVIOR BASED CREDIT CARD FRAUD DETECTION USING SUPPORT VECTOR MACHINES 123 CHAPTER 7 BEHAVIOR BASED CREDIT CARD FRAUD DETECTION USING SUPPORT VECTOR MACHINES 7.1 Introduction Even though using SVM presents
More informationLeast-Squares Intersection of Lines
Least-Squares Intersection of Lines Johannes Traa - UIUC 2013 This write-up derives the least-squares solution for the intersection of lines. In the general case, a set of lines will not intersect at a
More informationRisk and Return. Sample chapter. e r t u i o p a s d f CHAPTER CONTENTS LEARNING OBJECTIVES. Chapter 7
Chater 7 Risk and Return LEARNING OBJECTIVES After studying this chater you should be able to: e r t u i o a s d f understand how return and risk are defined and measured understand the concet of risk
More informationUsing Principal Components Analysis in Program Evaluation: Some Practical Considerations
http://evaluation.wmich.edu/jmde/ Articles Using Principal Components Analysis in Program Evaluation: Some Practical Considerations J. Thomas Kellow Assistant Professor of Research and Statistics Mercer
More informationExploratory Factor Analysis
Exploratory Factor Analysis Definition Exploratory factor analysis (EFA) is a procedure for learning the extent to which k observed variables might measure m abstract variables, wherein m is less than
More informationPythagorean Triples and Rational Points on the Unit Circle
Pythagorean Triles and Rational Points on the Unit Circle Solutions Below are samle solutions to the roblems osed. You may find that your solutions are different in form and you may have found atterns
More informationTo do a factor analysis, we need to select an extraction method and a rotation method. Hit the Extraction button to specify your extraction method.
Factor Analysis in SPSS To conduct a Factor Analysis, start from the Analyze menu. This procedure is intended to reduce the complexity in a set of data, so we choose Data Reduction from the menu. And the
More informationFactor Analysis: Statnotes, from North Carolina State University, Public Administration Program. Factor Analysis
Factor Analysis Overview Factor analysis is used to uncover the latent structure (dimensions) of a set of variables. It reduces attribute space from a larger number of variables to a smaller number of
More informationAnalysis of Crime Data using Principal Component Analysis: A case study of Katsina State
CBN Journal of Alied Statistics Vol. 3 No.2 39 Analysis of Crime Data using Princial Comonent Shehu U. Gulumbe 1, H.G. Dikko 2, and Yusuf Bello 3 This aer analyses Katsina State crime data which consists
More informationSPSS ADVANCED ANALYSIS WENDIANN SETHI SPRING 2011
SPSS ADVANCED ANALYSIS WENDIANN SETHI SPRING 2011 Statistical techniques to be covered Explore relationships among variables Correlation Regression/Multiple regression Logistic regression Factor analysis
More informationFactor Analysis Using SPSS
Factor Analysis Using SPSS The theory of factor analysis was described in your lecture, or read Field (2005) Chapter 15. Example Factor analysis is frequently used to develop questionnaires: after all
More informationFactor Analysis - 2 nd TUTORIAL
Factor Analysis - 2 nd TUTORIAL Subject marks File sub_marks.csv shows correlation coefficients between subject scores for a sample of 220 boys. sub_marks
More information9.2 User s Guide SAS/STAT. The FACTOR Procedure. (Book Excerpt) SAS Documentation
SAS/STAT 9.2 User s Guide The FACTOR Procedure (Book Excerpt) SAS Documentation This document is an individual chapter from SAS/STAT 9.2 User s Guide. The correct bibliographic citation for the complete
More informationA Solution Manual and Notes for: Exploratory Data Analysis with MATLAB by Wendy L. Martinez and Angel R. Martinez.
A Solution Manual and Notes for: Exploratory Data Analysis with MATLAB by Wendy L. Martinez and Angel R. Martinez. John L. Weatherwax May 7, 9 Introduction Here you ll find various notes and derivations
More informationPRINCIPAL COMPONENT ANALYSIS
1 Chapter 1 PRINCIPAL COMPONENT ANALYSIS Introduction: The Basics of Principal Component Analysis........................... 2 A Variable Reduction Procedure.......................................... 2
More informationThe impact of metadata implementation on webpage visibility in search engine results (Part II) q
Information Processing and Management 41 (2005) 691 715 www.elsevier.com/locate/inforoman The imact of metadata imlementation on webage visibility in search engine results (Part II) q Jin Zhang *, Alexandra
More informationChapter 6. Orthogonality
6.3 Orthogonal Matrices 1 Chapter 6. Orthogonality 6.3 Orthogonal Matrices Definition 6.4. An n n matrix A is orthogonal if A T A = I. Note. We will see that the columns of an orthogonal matrix must be
More informationMultivariate Normal Distribution
Multivariate Normal Distribution Lecture 4 July 21, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2 Lecture #4-7/21/2011 Slide 1 of 41 Last Time Matrices and vectors Eigenvalues
More information15.062 Data Mining: Algorithms and Applications Matrix Math Review
.6 Data Mining: Algorithms and Applications Matrix Math Review The purpose of this document is to give a brief review of selected linear algebra concepts that will be useful for the course and to develop
More informationQuestionnaire Evaluation with Factor Analysis and Cronbach s Alpha An Example
Questionnaire Evaluation with Factor Analysis and Cronbach s Alpha An Example - Melanie Hof - 1. Introduction The pleasure writers experience in writing considerably influences their motivation and consequently
More informationA Beginner s Guide to Factor Analysis: Focusing on Exploratory Factor Analysis
Tutorials in Quantitative Methods for Psychology 2013, Vol. 9(2), p. 79-94. A Beginner s Guide to Factor Analysis: Focusing on Exploratory Factor Analysis An Gie Yong and Sean Pearce University of Ottawa
More informationManifold Learning Examples PCA, LLE and ISOMAP
Manifold Learning Examples PCA, LLE and ISOMAP Dan Ventura October 14, 28 Abstract We try to give a helpful concrete example that demonstrates how to use PCA, LLE and Isomap, attempts to provide some intuition
More informationHow To Run Factor Analysis
Getting Started in Factor Analysis (using Stata 10) (ver. 1.5) Oscar Torres-Reyna Data Consultant otorres@princeton.edu http://dss.princeton.edu/training/ Factor analysis is used mostly for data reduction
More informationEXPLORATORY FACTOR ANALYSIS IN MPLUS, R AND SPSS. sigbert@wiwi.hu-berlin.de
EXPLORATORY FACTOR ANALYSIS IN MPLUS, R AND SPSS Sigbert Klinke 1,2 Andrija Mihoci 1,3 and Wolfgang Härdle 1,3 1 School of Business and Economics, Humboldt-Universität zu Berlin, Germany 2 Department of
More informationSTA 4107/5107. Chapter 3
STA 4107/5107 Chapter 3 Factor Analysis 1 Key Terms Please review and learn these terms. 2 What is Factor Analysis? Factor analysis is an interdependence technique (see chapter 1) that primarily uses metric
More informationby the matrix A results in a vector which is a reflection of the given
Eigenvalues & Eigenvectors Example Suppose Then So, geometrically, multiplying a vector in by the matrix A results in a vector which is a reflection of the given vector about the y-axis We observe that
More informationCanonical Correlation Analysis
Canonical Correlation Analysis Lecture 11 August 4, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2 Lecture #11-8/4/2011 Slide 1 of 39 Today s Lecture Canonical Correlation Analysis
More informationTopic 10: Factor Analysis
Topic 10: Factor Analysis Introduction Factor analysis is a statistical method used to describe variability among observed variables in terms of a potentially lower number of unobserved variables called
More informationLinear Algebra Review. Vectors
Linear Algebra Review By Tim K. Marks UCSD Borrows heavily from: Jana Kosecka kosecka@cs.gmu.edu http://cs.gmu.edu/~kosecka/cs682.html Virginia de Sa Cogsci 8F Linear Algebra review UCSD Vectors The length
More informationModule 3: Correlation and Covariance
Using Statistical Data to Make Decisions Module 3: Correlation and Covariance Tom Ilvento Dr. Mugdim Pašiƒ University of Delaware Sarajevo Graduate School of Business O ften our interest in data analysis
More informationFactor Analysis and Structural equation modelling
Factor Analysis and Structural equation modelling Herman Adèr Previously: Department Clinical Epidemiology and Biostatistics, VU University medical center, Amsterdam Stavanger July 4 13, 2006 Herman Adèr
More informationNotes on Orthogonal and Symmetric Matrices MENU, Winter 2013
Notes on Orthogonal and Symmetric Matrices MENU, Winter 201 These notes summarize the main properties and uses of orthogonal and symmetric matrices. We covered quite a bit of material regarding these topics,
More informationMultivariate Statistical Inference and Applications
Multivariate Statistical Inference and Applications ALVIN C. RENCHER Department of Statistics Brigham Young University A Wiley-Interscience Publication JOHN WILEY & SONS, INC. New York Chichester Weinheim
More informationThe president of a Fortune 500 firm wants to measure the firm s image.
4. Factor Analysis A related method to the PCA is the Factor Analysis (FA) with the crucial difference that in FA a statistical model is constructed to explain the interrelations (correlations) between
More informationOverview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model
Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model 1 September 004 A. Introduction and assumptions The classical normal linear regression model can be written
More informationA MOST PROBABLE POINT-BASED METHOD FOR RELIABILITY ANALYSIS, SENSITIVITY ANALYSIS AND DESIGN OPTIMIZATION
9 th ASCE Secialty Conference on Probabilistic Mechanics and Structural Reliability PMC2004 Abstract A MOST PROBABLE POINT-BASED METHOD FOR RELIABILITY ANALYSIS, SENSITIVITY ANALYSIS AND DESIGN OPTIMIZATION
More informationRegression III: Advanced Methods
Lecture 16: Generalized Additive Models Regression III: Advanced Methods Bill Jacoby Michigan State University http://polisci.msu.edu/jacoby/icpsr/regress3 Goals of the Lecture Introduce Additive Models
More informationPractical Considerations for Using Exploratory Factor Analysis in Educational Research
A peer-reviewed electronic journal. Copyright is retained by the first or sole author, who grants right of first publication to the Practical Assessment, Research & Evaluation. Permission is granted to
More informationAssignment 9; Due Friday, March 17
Assignment 9; Due Friday, March 17 24.4b: A icture of this set is shown below. Note that the set only contains oints on the lines; internal oints are missing. Below are choices for U and V. Notice that
More informationThe ith principal component (PC) is the line that follows the eigenvector associated with the ith largest eigenvalue.
More Principal Components Summary Principal Components (PCs) are associated with the eigenvectors of either the covariance or correlation matrix of the data. The ith principal component (PC) is the line
More informationEffect Sizes Based on Means
CHAPTER 4 Effect Sizes Based on Means Introduction Raw (unstardized) mean difference D Stardized mean difference, d g Resonse ratios INTRODUCTION When the studies reort means stard deviations, the referred
More informationMultivariate Analysis of Variance (MANOVA)
Multivariate Analysis of Variance (MANOVA) Aaron French, Marcelo Macedo, John Poulsen, Tyler Waterson and Angela Yu Keywords: MANCOVA, special cases, assumptions, further reading, computations Introduction
More informationAn important observation in supply chain management, known as the bullwhip effect,
Quantifying the Bullwhi Effect in a Simle Suly Chain: The Imact of Forecasting, Lead Times, and Information Frank Chen Zvi Drezner Jennifer K. Ryan David Simchi-Levi Decision Sciences Deartment, National
More informationLecture 9: Introduction to Pattern Analysis
Lecture 9: Introduction to Pattern Analysis g Features, patterns and classifiers g Components of a PR system g An example g Probability definitions g Bayes Theorem g Gaussian densities Features, patterns
More informationMeasuring relative phase between two waveforms using an oscilloscope
Measuring relative hase between two waveforms using an oscilloscoe Overview There are a number of ways to measure the hase difference between two voltage waveforms using an oscilloscoe. This document covers
More informationLinear algebra and the geometry of quadratic equations. Similarity transformations and orthogonal matrices
MATH 30 Differential Equations Spring 006 Linear algebra and the geometry of quadratic equations Similarity transformations and orthogonal matrices First, some things to recall from linear algebra Two
More informationEigenvalues, Eigenvectors, Matrix Factoring, and Principal Components
Eigenvalues, Eigenvectors, Matrix Factoring, and Principal Components The eigenvalues and eigenvectors of a square matrix play a key role in some important operations in statistics. In particular, they
More informationNCSS Statistical Software Principal Components Regression. In ordinary least squares, the regression coefficients are estimated using the formula ( )
Chapter 340 Principal Components Regression Introduction is a technique for analyzing multiple regression data that suffer from multicollinearity. When multicollinearity occurs, least squares estimates
More information13 MATH FACTS 101. 2 a = 1. 7. The elements of a vector have a graphical interpretation, which is particularly easy to see in two or three dimensions.
3 MATH FACTS 0 3 MATH FACTS 3. Vectors 3.. Definition We use the overhead arrow to denote a column vector, i.e., a linear segment with a direction. For example, in three-space, we write a vector in terms
More informationResearch Methodology: Tools
MSc Business Administration Research Methodology: Tools Applied Data Analysis (with SPSS) Lecture 02: Item Analysis / Scale Analysis / Factor Analysis February 2014 Prof. Dr. Jürg Schwarz Lic. phil. Heidi
More informationCOMP6053 lecture: Relationship between two variables: correlation, covariance and r-squared. jn2@ecs.soton.ac.uk
COMP6053 lecture: Relationship between two variables: correlation, covariance and r-squared jn2@ecs.soton.ac.uk Relationships between variables So far we have looked at ways of characterizing the distribution
More informationOrthogonal Diagonalization of Symmetric Matrices
MATH10212 Linear Algebra Brief lecture notes 57 Gram Schmidt Process enables us to find an orthogonal basis of a subspace. Let u 1,..., u k be a basis of a subspace V of R n. We begin the process of finding
More information