STA 4107/5107. Chapter 3

Size: px
Start display at page:

Download "STA 4107/5107. Chapter 3"

Transcription

1 STA 4107/5107 Chapter 3 Factor Analysis 1 Key Terms Please review and learn these terms. 2 What is Factor Analysis? Factor analysis is an interdependence technique (see chapter 1) that primarily uses metric variables, however, non-metric variables can be included as dummy variables. The basic idea is to describe the set of P variables X 1, X 2,, X P in our data as linear combinations of a smaller number of factors, and in the process get a better understanding of the relationships between our variables. Factor analysis is based on a statistical model, unlike principal components (which we will cover later). Mathematically, factor analysis can be represented thusly: x 1 = l 11 F 1 + l 12 F l 1m F m + ɛ 1 x 2 = l 21 F 1 + l 22 F l 2m F m + ɛ 2. x p = l p1 F 1 + l p2 F l pm F m + ɛ p Here, the X i are the original variables, the l ij are called the loadings and the F j are the common factors, and the ɛ i are the unique factors. In factor analysis, the loadings or weights are chosen to maximize the correlation between the variable and the factor. Our motive for studying factor analysis in this course is that we will extend the concept to structural equation modeling, which is also called confirmatory factor analysis near the end of the term. 3 Some History and Examples Factor analysis was developed at the turn of the 20 th century by psychologist Charles Spearman, who hypothesized that a person s score on a wide variety of tests of mental ability mathematical skill, vocabulary, other verbal skills, artistic skills, logical reasoning ability, etc. could all be explained by one underlying factor of general intelligence that he called g. From a data set of test scores of boys in preparatory school, he noticed that any two rows in the table of correlations were approximately proportional across the different variables. So for example, Classics and English in the table below have the ratios Upon observing this, he hypothesized that it was due to a common factor, g. 1

2 Classics French English Mathematics Pitch Music Classics French English Mathematics Pitch Music It was an interesting idea, but it turned out to be wrong. Today the College Board testing service operates a system based on the idea that there are at least three important factors of mental ability verbal, mathematical, and logical abilities and most psychologists agree that many other factors could be identified as well. 3.1 Examples 1. Consider various measures of the activity of the autonomic nervous system heart rate, blood pressure, etc. Psychologists have wanted to know whether, except for random fluctuation, all those measures move up and down together the activation hypothesis. Or do groups of autonomic measures move up and down together, but separate from other groups? Or are all the measures largely independent? An unpublished analysis by Richard Darlington at Cornell found that in one data set, at any rate, the data fitted the activation hypothesis quite well. 2. Suppose each of 500 people, who are all familiar with different kinds of automobiles, rates each of 20 automobile models on the question, How much would you like to own that kind of automobile? We could usefully ask about the number of dimensions on which the ratings differ. A one-factor theory would posit that people simply give the highest ratings to the most expensive models. A two-factor theory would posit that some people are most attracted to sporty models while others are most attracted to luxurious models. Three-factor and four-factor theories might add safety and reliability. Or instead of automobiles you might choose to study attitudes concerning foods, political policies, political candidates, or many other kinds of objects. 3. Rubenstein (1986) studied the nature of curiosity by analyzing the agreements of juniorhigh-school students with a large battery of statements such as I like to figure out how machinery works or I like to try new kinds of food. A factor analysis identified seven factors: three measuring enjoyment of problem-solving, learning, and reading; three measuring interests in natural sciences, art and music, and new experiences in general; and one indicating a relatively low interest in money. 2

3 4 Factor Analysis Decision Process We follow the six-stage model-building process introduced in Chapter 1. As we work through these stages we will be considering a data set, the US crime dataset, that contains variables measured on several cities in the US. These data are crime-related and demographic statistics for 47 US states in The data were collected from the FBI s Uniform Crime Report and other government agencies to determine how the variable crime rate depends on the other variables measured in the study, but we can use it as an example of looking for underlying patterns in the relationships between variables. There are 47 cases and 14 variables. R: Crime rate: number of offenses reported to police per million population Age: The number of males of age per 1000 population S: Indicator (or dummy) variable for Southern states (0 = No, 1 = Yes) Ed: Mean number of years of schooling x 10 for persons of age 25 or older Ex0: 1960 per capita expenditure on police by state and local government Ex1: 1959 per capita expenditure on police by state and local government LF: Labor force participation rate per 1000 civilian urban males age M: The number of males per 1000 females N: State population size in hundred thousands NW: The number of non-whites per 1000 population U1: Unemployment rate of urban males per 1000 of age U2: Unemployment rate of urban males per 1000 of age W: Median value of transferable goods and assets or family income in tens of dollars. X: The number of families per 1000 earning below 1/2 the median income We will see how well factor analysis identifies the underlying structure. What do you expect to happen? Stage 1: Objectives As with all statistical analyses, all the decisions need to be made with an emphasis on the research problem. Stage 1 focuses on carefully defining the objectives. Specifying the Unit of Analysis: Factor analysis can identify the structure of relationships either among variables or among respondents. Variables: factor analysis is applied to a correlation matrix of the variables. This is called R factor analysis and the approach is to identify the dimensions that are latent, or not initially observed. 3

4 Respondents: Here factor analysis is applied to the individual respondents based on their characteristics. This is called Q factor analysis and this method combines large numbers of cases into distinctly different groups. We will not consider this here, but will delve deeper when we get to cluster analysis. Achieving Data Summarization versus Data Reduction: Data summarization illucidates underlying relationships in an interpretable and understandable fashion. Data reduction uses the summarization as a variate instead of the original values in further data analyses. Variable Selection: As mentioned in the notes for chapter 1, the fact that we have statistical methods at our disposal that can extract patterns out of a large group of variables doesn t mean that we get to skip the hard part of research and simply measure everything we can think of and let the computer come up with the interesting ideas for us. The researcher should carefully specify the possible underlying factors of interest that could be identified with factor analysis. If the researcher doesn t actually have any ideas, it s unlikely that anything convincing will come of the analysis. Using Factor Analysis with Other Multivariate Techniques: Factor analysis can help simplify other procedures. Variables determined to be highly correlated and members of the same factor would be expected to have similar profiles of differences across groups in multivariate analysis of variance or discriminant analysis Highly correlated variables can cause problems in regression analyses, such as multiple regression and discriminant analysis. Including highly correlated predictor variables will not significantly improve predictive ability. Understanding the correlation structure can aid the researcher in building the regression model. 4.1 US Crime Example For this data set we are interested in the underlying structure between the variables, so we want an R factor analysis. Question: How many factors do you expect? What might they represent? We want both data summarization and reduction. We may want to see if the factors extracted can be used in predicting other variables taken on the States, such as population growth rate, violent crime rate, or economic growth. 4

5 Stage 2: Designing a Factor Analysis Obtain the Correlation Matrix for the Variables: The correlation matrix is the input for the factor analysis. This will be accomplished automatically when we perform proc factor in SAS. Variable Selection and Measurement Issues: What types of variables can be used in factor analysis? For the most part, we will prefer to use metric variables because correlation is well-defined, whereas non-metric variables are a bit trickier, but can be handled with the use of dummy variables. See the variable S, in our example. How many variables should be included? The researcher, as always, should measure no more and no fewer variables than he is actually interested in. Sample Size: The general rule is to have 5 times as many observations as variables. The reason for this is that the number of correlations increases as ( ) p 2, where p is the number of variables. US Crime Example We do not quite have the recommended sample size, but we expect fairly high correlation among several of the variables. We want to start out using all of them. Stage 3: Assumptions in Factor Analysis Conceptual Issues: The only hard assumption is that some underlying structure does exist. A factor analysis will always produce factors. It is up to the researcher to know whether the pattern being revealed is meaningful. Statistical Issues: Factor analysis does not have a large number of assumptions. We do not assume normality (except when testing the significance of the factors), or equal variances. The main issue, then, is the degree of correlation. We will consider some methods for assessing intercorrelation. If it is found that all correlations are low or all correlations are about the same, then no underlying structure is present and factor analysis is probably not appropriate Overall Measures of Intercorrelation: 1. visual inspection: if there are no correlations higher than 0.3, then factor analysis is unlikely to be helpful. SAS produces the partial correlation matrix. Partial correlations can also be considered. Partial correlation is the correlation between two variables after accounting for the effect of one or more other variables. For example, suppose an educator is interested in the correlation between the total scores from two tests, T 1 and T 2, given to 12th graders. Since both scores are most likely related to the student s socio-economic status, it could be informative to first remove the effect of SES from both T 1 and T 2 and then find the correlation between the adjusted scores. SAS will produce the anti-image correlation matrix 5

6 which is the matrix of negative partial correlations. Large partial or antiimage correlations are indicative of a data matrix not well-suited for factor analysis. 2. Bartlett Test of Sphericity: This is a significance test of correlation. SAS does not perform this test, so we will not consider it further. 3. Measure of sampling adequacy (MSA): This index ranges from 0 to 1, with 1 meaning that each variable is perfectly predicted by other variables. Above.80 is considered excellent. Below.50 is considered poor. If the MSA falls below.5, factor analysis is unlikely to be helpful. Variable-Specific Measures of Intercorrelation: The MSA index will be reported by SAS for each variable. Any variable that falls below.50 should be dropped. If more than one variable falls below.50, then the variable with the lowest MSA should be dropped first, the test run again and continue until all varibles with MSAs lower than.50 are dropped. US Crime Example SAS code for the various preliminary analyses is shown below. proc factor data=uscrime msa; var R Ed LF U1 U2 W X Age M N NW Ex0 Ex1; run; The output for the above code is shown below. The FACTOR Procedure Initial Factor Method: Principal Components Partial Correlations Controlling all other Variables R Ed LF U1 U2 W X R R Ed Ed LF LF U1 U U2 U W W X X Age Age M M N N NW NW Ex0 Ex Ex1 Ex So So Partial Correlations Controlling all other Variables Age M N NW Ex0 Ex1 So R R

7 Ed Ed LF LF U1 U U2 U W W X X Age Age M M N N NW NW Ex0 Ex Ex1 Ex So So Most of the partial correlations are low, indicating the dataset is a good candidate for factor analysis. Kaiser s Measure of Sampling Adequacy: Overall MSA = R Ed LF U1 U2 W X Kaiser s Measure of Sampling Adequacy: Overall MSA = Age M N NW Ex0 Ex1 So Both U1, the unemployment rate of urban males per 1000 of age and U2, the unemployment rate of urban males per 1000 of age 35-39, have MSAs below We should remove U1 first and then see if U2 is still below. This was done and both were removed due to low MSAs. The new output is shown below. Kaiser s Measure of Sampling Adequacy: Overall MSA = R Ed LF W X Age Kaiser s Measure of Sampling Adequacy: Overall MSA = M N NW Ex0 Ex1 So Prior Communality Estimates: ONE Everything looks good, so we will proceed with the analysis. 7

8 Stage 4: Deriving Factors and Assessing Overall Fit Selecting the Factor Extraction Method There are two main methods of extracting factors: Common Factor Analysis and Component Factor Analysis. Both methods deal with the variance of the variables slightly differently. Partitioning the Variance of a Variable: As mentioned in the introductory section, factor analysis breaks each variable into two parts, the common factor and the unique factor. It also breaks the variance into two parts, the communality and the specificity. Or in your text, the common variance and the specific variance. If we denote the communality by h 2 i and the specificity by u 2 i, for a given variable, the variance of the variable can be written V ar(x i ) = h 2 i + u2 i. Numerically, factor analysis tries to solve for the factor loadings l ij and the communalities h 2 i. What your text refers to as error variance is the remaining variance that is unaccounted for by the factor analysis. Common Factor Analysis versus Component Analysis: Component analysis is the preferred method when the goal of the research is data reduction, while common factor analysis is to identify latent relationships. Component Analysis: In component analysis, the basic idea is to find the first (ranked by the amount of variance explained) m uncorrelated factors that explain the greatest proportion of the variance in the variables. When the original variables are standardized, the factor loading l ij is the correlation between x i and F j. The factors themselves are derived from the first m principal components C j, that is F j = C j /(V arc j ) 1/2, so that the factors have unit variance. Hence, the correlation matrix for the factors has ones down the diagonal. Hence, we are factoring the total variance. Thus this approach focuses on explaining as much of the variance as possible so that the factors capture as much of the information as possible and are then useful in other analyses such as regression. Common Factor (also called iterated components) Analysis: Here the 1s in the diagonal of the correlation matrix are substituted with the communalities. With the communalities in the diagonal we are factoring the variance associated with the common factors. Thus this approach is selecting factors that maximize the total communality. Hence, this approach focuses on the underlying relationships. Common factor analysis is also call iterated components because the solution is solved iteratively. The steps are outlined below: 1. Find the initial communalities 2. Substitute the communalities for the diagonal elements in the correlation matrix. 3. Extract m principal components from the modified matrix. 4. Multiply the principal components coefficients by the standard deviation of the respective principal components to obtain the factor loadings. 5. Compute new communalities from the computed factor loadings. 8

9 6. Replace communalities in step 2 with these new communalities and repeat steps 3, 4, and Continue iterating, stopping when communalities do not change by more than a very small amount. We do not need to be too concerned with this, because SAS will perform it all for us. Algebraic Explanation of Component Factor Analysis In factor analysis, the variables usually are standardized x i = (X i X i )/S i so that they all have unit variance. That is V ar (x i ) = 1. Recall that V ar (x i ) = 1 = h 2 i + u 2 i the sum of the communality and the specificity. Component factor analysis starts with the principal components, linear combinations of the variables that are all mutually independent and orthogonal. That is, the j th principal component for p variables is given by: C j = a 1j x 1 + a 2j x a 1p x p It can be shown algebraically that the system of all m principal component equations can be inverted to give x 1 = a 11 C 1 + a 21 C 2 + a p1 C p. x p = a 1p C 1 + a 2p C 2 + a pp C p Now, since F j = C j V ar (C j ) 1/2 C j = F j V ar (C j ) 1/2 x i = a i1 V ar (C 1 ) 1/2 F a im V ar (C m ) 1/2 F m l ij = a ij V ar (C j ) 1/2 since x i = l i1 F 1 + l i2 F l im F m + ɛ i. 9

10 So, V ar (x i ) = V ar (l i1 F 1 + l i2 F l im F m + ɛ i ) = l 2 i1v ar (F 1 ) + l 2 i2v ar (F 2 ) + l 2 imv ar (F m ) + V ar (ɛ i ) = j l ij + u 2 i Recall that V ar (x i ) = 1 = h 2 i + u2 i = 1 h2 i = j l ij. The take-home message is that the factors are found by finding the principal components and the loadings are found by finding the correlations between the factors and the variables, and that component factor analysis is optimized to explain as much of the total variance as possible by way of principal components. US Crime Data As you may have already guessed from the SAS code, for these data, as will be the case most often, we will use the component method. The component method is the default method, makes fewer assumptions, whereas the common factor method often has several problems such as multiple solutions and unestimable communalites. Most of the time, the component method is the safer choice. Criteria for the Number of Factors to Extract There are several criteria for stopping the selection of factors. No one method is agreed upon to be superior to any other. 1. Latent Root: This approach chooses only those factors whose associated eigenvalues (latent roots) are greater than or equal to 1. This criterion is based on theoretical rationales developed using true population correlation coefficients. It appears to correctly estimate the number of factors when the communalities are high and the number of variables is not too large. With low communalities, then this criterion can be adjusted by setting the cut-off value for the latent roots to be equal to the average communality. 2. A priori: Here the number of factors is based on the investigators expert knowledge of the number of factors. 3. Percentage Variance Explained: This method also relies on expert knowledge and the goals of the researcher. The investigators decide before hand how much of the variance needs to be explained for the results to be meaningful, useful and achievable. 4. Scree Test: In this method, the eigenvalues are plotted against the number of factors and we choose the cut-off based on a sharp decrease in slope, indicating sharply decreasing improvement per additional factor. 5. Heterogeneity of Respondents: We use this method when certain variables do not load strongly until later factors. In general, the later the factor the less variance it explains. However, if certain variables we consider important do not load strongly in the first factors, we may want to retain later factors where these variables do load. 10

11 SAS outputs the eigenvectors and chooses the first few factors that all have eigenvalues greater than or equal to one. The FACTOR Procedure Initial Factor Method: Principal Components Eigenvalues of the Correlation Matrix: Total = 12 Average = 1 Eigenvalue Difference Proportion Cumulative factors will be retained by the MINEIGEN criterion. Stage 5: Interpreting the Factors The idea is to look where the high loadings are to see if certain groups of variables load high. Interpretation can sometimes be made easier with factor rotation. In factor rotation, the axes are rotated according to various maximization criteria. The basic idea here is to find rotations that produce loadings on the variables that maximally distinguish the variables from each other. Cross loadings are when a variable has loadings that are similar across different factors. If a variable has cross loadings, we may want to delete it. 1. Quartimax: this method tries to maximally differentiate loadings across rows. I find this method the easiest to interpret. However, it apparently fails often, usually creating the first factor with high loadings on most or all variables. 2. Varimax: this method maximizes the differentiation across columns. It apparently usually produces more stable estimates. 3. Equimax: this method simulaneously performs Quartimax and Varimax. SAS does not include this method in its toolbox. Our general approach will be to use both Quartimax and Varimax and compare the loadings and choose the best one, i.e., the one that produces the least and fewest cross loadings. SAS output for the rotated loadings, using both Quartimax and Varimax is shown below. 11

12 Quartimax Rotation The FACTOR Procedure Rotation Method: Quartimax Rotated Factor Pattern Factor1 Factor2 Factor3 R R Ed Ed LF LF W W X X Age Age M M N N NW NW Ex0 Ex Ex1 Ex So So

13 Variance Explained by Each Factor Factor1 Factor2 Factor Varimax Rotation The FACTOR Procedure Rotation Method: Varimax Rotated Factor Pattern Factor1 Factor2 Factor3 R R Ed Ed LF LF W W X X Age Age M M N N NW NW Ex0 Ex Ex1 Ex So So Variance Explained by Each Factor Factor1 Factor2 Factor Upon examination of both methods loadings we see that N, population size, has two almost equally-weighted, somewhat low, loadings. We may want to delete this variable and repeat the analysis. Both rotation methods produce comparable results no cross-loadings other than N. The factor patterns seem clearer for the Quartimax rotation, so we will choose this one. We also want to assess the communalities. Recall that the communalities are proportion of variance in each variable explained by the factor. If no factor explains a reasonable amount of the variance for a given variable, that variable is a candidate for deletion. A rule of thumb is that all variables having communalities less than 0.50 should be deleted unless there is a good reason not to. 13

14 Communalities Final Communality Estimates: Total = R Ed LF W X Age M N NW Ex0 Ex1 So All of our communalities are greater than 0.50, except that LF and N are somewhat low. We have three borderline reasons for deleting N, so we will. LF is borderline only in its communality but has a clear factor loading pattern, and we might be particularly interested in this variable, since the employment conditions in a city are fairly important in the overall quality of life. We will retain this variable. Final Factor Loadings Rotated Factor Pattern Factor1 Factor2 Factor3 R R Ed Ed LF LF W W X X Age Age M M NW NW Ex0 Ex Ex1 Ex So So Variance Explained by Each Factor Factor1 Factor2 Factor Final Communality Estimates: Total = R Ed LF W X Age M NW Ex0 Ex1 So

15 Now comes the more subjective step: Labeling the factors. The first step is to separate the variables according to the highest factor loading. The table below shows the group of variables that load highest on the first factor, and then the group that load highest on the second factor and likewise for the third factor. Our job now is to see what general label we might be able to give the three groups. Can you think of meaningful labels? variable Factor1 Factor2 Factor3 Ed W X Age NW So R Ex Ex LF M The first group includes: education; median value assets or family income; number of families per 1000 earning below 1/2 the median income; the number of males of age per 1000; the number of non-whites per 1000; and a dummy variable indicating whether the city is in the southern US. Is there a label we can give this group of variables? We might call this the high risk factor, since certain levels of all these variables are known to increase the risk of crime. The second group includes: crime rate; per capita expenditures on police in 1960; per capita expenditures on police in The second factor could be called the crime factor because it includes the crime rate and money spent on the police force. The third group includes: labor force participation per 1000 males of age 14-24; number of males per 1000 females. This factor could be called the male factor because it could be thought of as measuring the life satisfaction of the males in the city. Stage 6: Validation of Factor Analysis Use of a confirmatory perspective: We ll cover this in more detail when we get to structural equation modeling. Assessing Factor Structure Stability: This is a form of cross validation and involves splitting the sample in two and running the analysis for both halves to see how they 15

16 compare to each other and to the results of the analysis from the full data set. If the answers vary wildly, then the technique is not robust. Detecting influential observations: If any unusual observations are observed during exploratory analysis, then they should be removed to see how greatly they affect the results. If the results change drastically, then the researcher should consider removing the influential observation. Stage 7: Additional Uses of Factor Analysis Results We will skip this section. We are presumably not experts on these data and so our interpretation may seem a little shady. Rest assured that when you are intimately familiar with your data this process is much more fun and satisfying. However, many people feel that factor analysis is a little shady. This is pretty much an old view that is dying as the utility of the method for making sense of large numbers of variables in hugely complicated areas of study is becoming clear. This also explains the huge popularity of structural equation modeling SEM is seen as a more rigorous sort of factor analysis. 5 Appendix 5.1 References Afifi, Abdelmonem, Virginia Clark, Susanne May (2004) Computer-Aided Multivariate Analysis, 4 th ed. Chapman & Hall/CRC. 3. Darlington, Richard B., Sharon Weinberg, and Herbert Walberg (1973). Canonical variate analysis and related techniques. Review of Educational Research, Rubenstein, Amy S. (1986). An item-level analysis of questionnaire-type measures of intellectual curiosity. Cornell University Ph. D. thesis. 16

Factor Analysis. Principal components factor analysis. Use of extracted factors in multivariate dependency models

Factor Analysis. Principal components factor analysis. Use of extracted factors in multivariate dependency models Factor Analysis Principal components factor analysis Use of extracted factors in multivariate dependency models 2 KEY CONCEPTS ***** Factor Analysis Interdependency technique Assumptions of factor analysis

More information

FACTOR ANALYSIS NASC

FACTOR ANALYSIS NASC FACTOR ANALYSIS NASC Factor Analysis A data reduction technique designed to represent a wide range of attributes on a smaller number of dimensions. Aim is to identify groups of variables which are relatively

More information

Exploratory Factor Analysis

Exploratory Factor Analysis Introduction Principal components: explain many variables using few new variables. Not many assumptions attached. Exploratory Factor Analysis Exploratory factor analysis: similar idea, but based on model.

More information

Canonical Correlation Analysis

Canonical Correlation Analysis Canonical Correlation Analysis LEARNING OBJECTIVES Upon completing this chapter, you should be able to do the following: State the similarities and differences between multiple regression, factor analysis,

More information

4. There are no dependent variables specified... Instead, the model is: VAR 1. Or, in terms of basic measurement theory, we could model it as:

4. There are no dependent variables specified... Instead, the model is: VAR 1. Or, in terms of basic measurement theory, we could model it as: 1 Neuendorf Factor Analysis Assumptions: 1. Metric (interval/ratio) data 2. Linearity (in the relationships among the variables--factors are linear constructions of the set of variables; the critical source

More information

Factor Analysis Example: SAS program (in blue) and output (in black) interleaved with comments (in red)

Factor Analysis Example: SAS program (in blue) and output (in black) interleaved with comments (in red) Factor Analysis Example: SAS program (in blue) and output (in black) interleaved with comments (in red) The following DATA procedure is to read input data. This will create a SAS dataset named CORRMATR

More information

Factor Analysis. Advanced Financial Accounting II Åbo Akademi School of Business

Factor Analysis. Advanced Financial Accounting II Åbo Akademi School of Business Factor Analysis Advanced Financial Accounting II Åbo Akademi School of Business Factor analysis A statistical method used to describe variability among observed variables in terms of fewer unobserved variables

More information

Factor Analysis. Chapter 420. Introduction

Factor Analysis. Chapter 420. Introduction Chapter 420 Introduction (FA) is an exploratory technique applied to a set of observed variables that seeks to find underlying factors (subsets of variables) from which the observed variables were generated.

More information

FACTOR ANALYSIS. Factor Analysis is similar to PCA in that it is a technique for studying the interrelationships among variables.

FACTOR ANALYSIS. Factor Analysis is similar to PCA in that it is a technique for studying the interrelationships among variables. FACTOR ANALYSIS Introduction Factor Analysis is similar to PCA in that it is a technique for studying the interrelationships among variables Both methods differ from regression in that they don t have

More information

T-test & factor analysis

T-test & factor analysis Parametric tests T-test & factor analysis Better than non parametric tests Stringent assumptions More strings attached Assumes population distribution of sample is normal Major problem Alternatives Continue

More information

Exploratory Factor Analysis of Demographic Characteristics of Antenatal Clinic Attendees and their Association with HIV Risk

Exploratory Factor Analysis of Demographic Characteristics of Antenatal Clinic Attendees and their Association with HIV Risk Doi:10.5901/mjss.2014.v5n20p303 Abstract Exploratory Factor Analysis of Demographic Characteristics of Antenatal Clinic Attendees and their Association with HIV Risk Wilbert Sibanda Philip D. Pretorius

More information

Rachel J. Goldberg, Guideline Research/Atlanta, Inc., Duluth, GA

Rachel J. Goldberg, Guideline Research/Atlanta, Inc., Duluth, GA PROC FACTOR: How to Interpret the Output of a Real-World Example Rachel J. Goldberg, Guideline Research/Atlanta, Inc., Duluth, GA ABSTRACT THE METHOD This paper summarizes a real-world example of a factor

More information

Exploratory Factor Analysis Brian Habing - University of South Carolina - October 15, 2003

Exploratory Factor Analysis Brian Habing - University of South Carolina - October 15, 2003 Exploratory Factor Analysis Brian Habing - University of South Carolina - October 15, 2003 FA is not worth the time necessary to understand it and carry it out. -Hills, 1977 Factor analysis should not

More information

Common factor analysis

Common factor analysis Common factor analysis This is what people generally mean when they say "factor analysis" This family of techniques uses an estimate of common variance among the original variables to generate the factor

More information

Psychology 7291, Multivariate Analysis, Spring 2003. SAS PROC FACTOR: Suggestions on Use

Psychology 7291, Multivariate Analysis, Spring 2003. SAS PROC FACTOR: Suggestions on Use : Suggestions on Use Background: Factor analysis requires several arbitrary decisions. The choices you make are the options that you must insert in the following SAS statements: PROC FACTOR METHOD=????

More information

Exploratory Factor Analysis and Principal Components. Pekka Malo & Anton Frantsev 30E00500 Quantitative Empirical Research Spring 2016

Exploratory Factor Analysis and Principal Components. Pekka Malo & Anton Frantsev 30E00500 Quantitative Empirical Research Spring 2016 and Principal Components Pekka Malo & Anton Frantsev 30E00500 Quantitative Empirical Research Spring 2016 Agenda Brief History and Introductory Example Factor Model Factor Equation Estimation of Loadings

More information

Factor Analysis. Sample StatFolio: factor analysis.sgp

Factor Analysis. Sample StatFolio: factor analysis.sgp STATGRAPHICS Rev. 1/10/005 Factor Analysis Summary The Factor Analysis procedure is designed to extract m common factors from a set of p quantitative variables X. In many situations, a small number of

More information

4. Multiple Regression in Practice

4. Multiple Regression in Practice 30 Multiple Regression in Practice 4. Multiple Regression in Practice The preceding chapters have helped define the broad principles on which regression analysis is based. What features one should look

More information

Chapter 7 Factor Analysis SPSS

Chapter 7 Factor Analysis SPSS Chapter 7 Factor Analysis SPSS Factor analysis attempts to identify underlying variables, or factors, that explain the pattern of correlations within a set of observed variables. Factor analysis is often

More information

PRINCIPAL COMPONENT ANALYSIS

PRINCIPAL COMPONENT ANALYSIS 1 Chapter 1 PRINCIPAL COMPONENT ANALYSIS Introduction: The Basics of Principal Component Analysis........................... 2 A Variable Reduction Procedure.......................................... 2

More information

This chapter will demonstrate how to perform multiple linear regression with IBM SPSS

This chapter will demonstrate how to perform multiple linear regression with IBM SPSS CHAPTER 7B Multiple Regression: Statistical Methods Using IBM SPSS This chapter will demonstrate how to perform multiple linear regression with IBM SPSS first using the standard method and then using the

More information

Introduction to Principal Components and FactorAnalysis

Introduction to Principal Components and FactorAnalysis Introduction to Principal Components and FactorAnalysis Multivariate Analysis often starts out with data involving a substantial number of correlated variables. Principal Component Analysis (PCA) is a

More information

To do a factor analysis, we need to select an extraction method and a rotation method. Hit the Extraction button to specify your extraction method.

To do a factor analysis, we need to select an extraction method and a rotation method. Hit the Extraction button to specify your extraction method. Factor Analysis in SPSS To conduct a Factor Analysis, start from the Analyze menu. This procedure is intended to reduce the complexity in a set of data, so we choose Data Reduction from the menu. And the

More information

Factor Analysis. Factor Analysis

Factor Analysis. Factor Analysis Factor Analysis Principal Components Analysis, e.g. of stock price movements, sometimes suggests that several variables may be responding to a small number of underlying forces. In the factor model, we

More information

Pull and Push Factors of Migration: A Case Study in the Urban Area of Monywa Township, Myanmar

Pull and Push Factors of Migration: A Case Study in the Urban Area of Monywa Township, Myanmar Pull and Push Factors of Migration: A Case Study in the Urban Area of Monywa Township, Myanmar By Kyaing Kyaing Thet Abstract: Migration is a global phenomenon caused not only by economic factors, but

More information

2. Linearity (in relationships among the variables--factors are linear constructions of the set of variables) F 2 X 4 U 4

2. Linearity (in relationships among the variables--factors are linear constructions of the set of variables) F 2 X 4 U 4 1 Neuendorf Factor Analysis Assumptions: 1. Metric (interval/ratio) data. Linearity (in relationships among the variables--factors are linear constructions of the set of variables) 3. Univariate and multivariate

More information

Multivariate Analysis (Slides 13)

Multivariate Analysis (Slides 13) Multivariate Analysis (Slides 13) The final topic we consider is Factor Analysis. A Factor Analysis is a mathematical approach for attempting to explain the correlation between a large set of variables

More information

Association Between Variables

Association Between Variables Contents 11 Association Between Variables 767 11.1 Introduction............................ 767 11.1.1 Measure of Association................. 768 11.1.2 Chapter Summary.................... 769 11.2 Chi

More information

Exploratory Factor Analysis

Exploratory Factor Analysis Exploratory Factor Analysis Definition Exploratory factor analysis (EFA) is a procedure for learning the extent to which k observed variables might measure m abstract variables, wherein m is less than

More information

DATA ANALYSIS AND INTERPRETATION OF EMPLOYEES PERSPECTIVES ON HIGH ATTRITION

DATA ANALYSIS AND INTERPRETATION OF EMPLOYEES PERSPECTIVES ON HIGH ATTRITION DATA ANALYSIS AND INTERPRETATION OF EMPLOYEES PERSPECTIVES ON HIGH ATTRITION Analysis is the key element of any research as it is the reliable way to test the hypotheses framed by the investigator. This

More information

Factor analysis. Angela Montanari

Factor analysis. Angela Montanari Factor analysis Angela Montanari 1 Introduction Factor analysis is a statistical model that allows to explain the correlations between a large number of observed correlated variables through a small number

More information

Chapter Seven. Multiple regression An introduction to multiple regression Performing a multiple regression on SPSS

Chapter Seven. Multiple regression An introduction to multiple regression Performing a multiple regression on SPSS Chapter Seven Multiple regression An introduction to multiple regression Performing a multiple regression on SPSS Section : An introduction to multiple regression WHAT IS MULTIPLE REGRESSION? Multiple

More information

How to report the percentage of explained common variance in exploratory factor analysis

How to report the percentage of explained common variance in exploratory factor analysis UNIVERSITAT ROVIRA I VIRGILI How to report the percentage of explained common variance in exploratory factor analysis Tarragona 2013 Please reference this document as: Lorenzo-Seva, U. (2013). How to report

More information

Topic 10: Factor Analysis

Topic 10: Factor Analysis Topic 10: Factor Analysis Introduction Factor analysis is a statistical method used to describe variability among observed variables in terms of a potentially lower number of unobserved variables called

More information

Principal Component Analysis

Principal Component Analysis Principal Component Analysis Principle Component Analysis: A statistical technique used to examine the interrelations among a set of variables in order to identify the underlying structure of those variables.

More information

Additional sources Compilation of sources: http://lrs.ed.uiuc.edu/tseportal/datacollectionmethodologies/jin-tselink/tselink.htm

Additional sources Compilation of sources: http://lrs.ed.uiuc.edu/tseportal/datacollectionmethodologies/jin-tselink/tselink.htm Mgt 540 Research Methods Data Analysis 1 Additional sources Compilation of sources: http://lrs.ed.uiuc.edu/tseportal/datacollectionmethodologies/jin-tselink/tselink.htm http://web.utk.edu/~dap/random/order/start.htm

More information

5.2 Customers Types for Grocery Shopping Scenario

5.2 Customers Types for Grocery Shopping Scenario ------------------------------------------------------------------------------------------------------- CHAPTER 5: RESULTS AND ANALYSIS -------------------------------------------------------------------------------------------------------

More information

A Primer on Mathematical Statistics and Univariate Distributions; The Normal Distribution; The GLM with the Normal Distribution

A Primer on Mathematical Statistics and Univariate Distributions; The Normal Distribution; The GLM with the Normal Distribution A Primer on Mathematical Statistics and Univariate Distributions; The Normal Distribution; The GLM with the Normal Distribution PSYC 943 (930): Fundamentals of Multivariate Modeling Lecture 4: September

More information

Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model

Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model 1 September 004 A. Introduction and assumptions The classical normal linear regression model can be written

More information

SPSS Explore procedure

SPSS Explore procedure SPSS Explore procedure One useful function in SPSS is the Explore procedure, which will produce histograms, boxplots, stem-and-leaf plots and extensive descriptive statistics. To run the Explore procedure,

More information

Validation of the Core Self-Evaluations Scale research instrument in the conditions of Slovak Republic

Validation of the Core Self-Evaluations Scale research instrument in the conditions of Slovak Republic Validation of the Core Self-Evaluations Scale research instrument in the conditions of Slovak Republic Lenka Selecká, Jana Holienková Faculty of Arts, Department of psychology University of SS. Cyril and

More information

Module 3: Correlation and Covariance

Module 3: Correlation and Covariance Using Statistical Data to Make Decisions Module 3: Correlation and Covariance Tom Ilvento Dr. Mugdim Pašiƒ University of Delaware Sarajevo Graduate School of Business O ften our interest in data analysis

More information

A Brief Introduction to SPSS Factor Analysis

A Brief Introduction to SPSS Factor Analysis A Brief Introduction to SPSS Factor Analysis SPSS has a procedure that conducts exploratory factor analysis. Before launching into a step by step example of how to use this procedure, it is recommended

More information

The ith principal component (PC) is the line that follows the eigenvector associated with the ith largest eigenvalue.

The ith principal component (PC) is the line that follows the eigenvector associated with the ith largest eigenvalue. More Principal Components Summary Principal Components (PCs) are associated with the eigenvectors of either the covariance or correlation matrix of the data. The ith principal component (PC) is the line

More information

Introduction to Matrix Algebra

Introduction to Matrix Algebra Psychology 7291: Multivariate Statistics (Carey) 8/27/98 Matrix Algebra - 1 Introduction to Matrix Algebra Definitions: A matrix is a collection of numbers ordered by rows and columns. It is customary

More information

Chapter 1 Introduction. 1.1 Introduction

Chapter 1 Introduction. 1.1 Introduction Chapter 1 Introduction 1.1 Introduction 1 1.2 What Is a Monte Carlo Study? 2 1.2.1 Simulating the Rolling of Two Dice 2 1.3 Why Is Monte Carlo Simulation Often Necessary? 4 1.4 What Are Some Typical Situations

More information

INTRODUCTION TO MULTIPLE CORRELATION

INTRODUCTION TO MULTIPLE CORRELATION CHAPTER 13 INTRODUCTION TO MULTIPLE CORRELATION Chapter 12 introduced you to the concept of partialling and how partialling could assist you in better interpreting the relationship between two primary

More information

Chapter 10. Key Ideas Correlation, Correlation Coefficient (r),

Chapter 10. Key Ideas Correlation, Correlation Coefficient (r), Chapter 0 Key Ideas Correlation, Correlation Coefficient (r), Section 0-: Overview We have already explored the basics of describing single variable data sets. However, when two quantitative variables

More information

IBM SPSS Direct Marketing 23

IBM SPSS Direct Marketing 23 IBM SPSS Direct Marketing 23 Note Before using this information and the product it supports, read the information in Notices on page 25. Product Information This edition applies to version 23, release

More information

Overview of Factor Analysis

Overview of Factor Analysis Overview of Factor Analysis Jamie DeCoster Department of Psychology University of Alabama 348 Gordon Palmer Hall Box 870348 Tuscaloosa, AL 35487-0348 Phone: (205) 348-4431 Fax: (205) 348-8648 August 1,

More information

IBM SPSS Direct Marketing 22

IBM SPSS Direct Marketing 22 IBM SPSS Direct Marketing 22 Note Before using this information and the product it supports, read the information in Notices on page 25. Product Information This edition applies to version 22, release

More information

Data analysis process

Data analysis process Data analysis process Data collection and preparation Collect data Prepare codebook Set up structure of data Enter data Screen data for errors Exploration of data Descriptive Statistics Graphs Analysis

More information

NCSS Statistical Software Principal Components Regression. In ordinary least squares, the regression coefficients are estimated using the formula ( )

NCSS Statistical Software Principal Components Regression. In ordinary least squares, the regression coefficients are estimated using the formula ( ) Chapter 340 Principal Components Regression Introduction is a technique for analyzing multiple regression data that suffer from multicollinearity. When multicollinearity occurs, least squares estimates

More information

Factor Analysis: Statnotes, from North Carolina State University, Public Administration Program. Factor Analysis

Factor Analysis: Statnotes, from North Carolina State University, Public Administration Program. Factor Analysis Factor Analysis Overview Factor analysis is used to uncover the latent structure (dimensions) of a set of variables. It reduces attribute space from a larger number of variables to a smaller number of

More information

CALCULATIONS & STATISTICS

CALCULATIONS & STATISTICS CALCULATIONS & STATISTICS CALCULATION OF SCORES Conversion of 1-5 scale to 0-100 scores When you look at your report, you will notice that the scores are reported on a 0-100 scale, even though respondents

More information

UNDERSTANDING THE TWO-WAY ANOVA

UNDERSTANDING THE TWO-WAY ANOVA UNDERSTANDING THE e have seen how the one-way ANOVA can be used to compare two or more sample means in studies involving a single independent variable. This can be extended to two independent variables

More information

Introduction to Principal Component Analysis: Stock Market Values

Introduction to Principal Component Analysis: Stock Market Values Chapter 10 Introduction to Principal Component Analysis: Stock Market Values The combination of some data and an aching desire for an answer does not ensure that a reasonable answer can be extracted from

More information

Beef Demand: What is Driving the Market?

Beef Demand: What is Driving the Market? Beef Demand: What is Driving the Market? Ronald W. Ward Food and Economics Department University of Florida Demand is a term we here everyday. We know it is important but at the same time hard to explain.

More information

Facebook Friend Suggestion Eytan Daniyalzade and Tim Lipus

Facebook Friend Suggestion Eytan Daniyalzade and Tim Lipus Facebook Friend Suggestion Eytan Daniyalzade and Tim Lipus 1. Introduction Facebook is a social networking website with an open platform that enables developers to extract and utilize user information

More information

Multivariate Analysis of Variance (MANOVA): I. Theory

Multivariate Analysis of Variance (MANOVA): I. Theory Gregory Carey, 1998 MANOVA: I - 1 Multivariate Analysis of Variance (MANOVA): I. Theory Introduction The purpose of a t test is to assess the likelihood that the means for two groups are sampled from the

More information

Multivariate Analysis

Multivariate Analysis Table Of Contents Multivariate Analysis... 1 Overview... 1 Principal Components... 2 Factor Analysis... 5 Cluster Observations... 12 Cluster Variables... 17 Cluster K-Means... 20 Discriminant Analysis...

More information

How To Run Factor Analysis

How To Run Factor Analysis Getting Started in Factor Analysis (using Stata 10) (ver. 1.5) Oscar Torres-Reyna Data Consultant otorres@princeton.edu http://dss.princeton.edu/training/ Factor analysis is used mostly for data reduction

More information

by the matrix A results in a vector which is a reflection of the given

by the matrix A results in a vector which is a reflection of the given Eigenvalues & Eigenvectors Example Suppose Then So, geometrically, multiplying a vector in by the matrix A results in a vector which is a reflection of the given vector about the y-axis We observe that

More information

Statistics in Psychosocial Research Lecture 8 Factor Analysis I. Lecturer: Elizabeth Garrett-Mayer

Statistics in Psychosocial Research Lecture 8 Factor Analysis I. Lecturer: Elizabeth Garrett-Mayer This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike License. Your use of this material constitutes acceptance of that license and the conditions of use of materials on this

More information

Research Methodology: Tools

Research Methodology: Tools MSc Business Administration Research Methodology: Tools Applied Data Analysis (with SPSS) Lecture 02: Item Analysis / Scale Analysis / Factor Analysis February 2014 Prof. Dr. Jürg Schwarz Lic. phil. Heidi

More information

Session 7 Bivariate Data and Analysis

Session 7 Bivariate Data and Analysis Session 7 Bivariate Data and Analysis Key Terms for This Session Previously Introduced mean standard deviation New in This Session association bivariate analysis contingency table co-variation least squares

More information

Multiple Regression: What Is It?

Multiple Regression: What Is It? Multiple Regression Multiple Regression: What Is It? Multiple regression is a collection of techniques in which there are multiple predictors of varying kinds and a single outcome We are interested in

More information

Section 14 Simple Linear Regression: Introduction to Least Squares Regression

Section 14 Simple Linear Regression: Introduction to Least Squares Regression Slide 1 Section 14 Simple Linear Regression: Introduction to Least Squares Regression There are several different measures of statistical association used for understanding the quantitative relationship

More information

PARTIAL LEAST SQUARES IS TO LISREL AS PRINCIPAL COMPONENTS ANALYSIS IS TO COMMON FACTOR ANALYSIS. Wynne W. Chin University of Calgary, CANADA

PARTIAL LEAST SQUARES IS TO LISREL AS PRINCIPAL COMPONENTS ANALYSIS IS TO COMMON FACTOR ANALYSIS. Wynne W. Chin University of Calgary, CANADA PARTIAL LEAST SQUARES IS TO LISREL AS PRINCIPAL COMPONENTS ANALYSIS IS TO COMMON FACTOR ANALYSIS. Wynne W. Chin University of Calgary, CANADA ABSTRACT The decision of whether to use PLS instead of a covariance

More information

An Introduction to Path Analysis. nach 3

An Introduction to Path Analysis. nach 3 An Introduction to Path Analysis Developed by Sewall Wright, path analysis is a method employed to determine whether or not a multivariate set of nonexperimental data fits well with a particular (a priori)

More information

HYPOTHESIS TESTING: CONFIDENCE INTERVALS, T-TESTS, ANOVAS, AND REGRESSION

HYPOTHESIS TESTING: CONFIDENCE INTERVALS, T-TESTS, ANOVAS, AND REGRESSION HYPOTHESIS TESTING: CONFIDENCE INTERVALS, T-TESTS, ANOVAS, AND REGRESSION HOD 2990 10 November 2010 Lecture Background This is a lightning speed summary of introductory statistical methods for senior undergraduate

More information

II. DISTRIBUTIONS distribution normal distribution. standard scores

II. DISTRIBUTIONS distribution normal distribution. standard scores Appendix D Basic Measurement And Statistics The following information was developed by Steven Rothke, PhD, Department of Psychology, Rehabilitation Institute of Chicago (RIC) and expanded by Mary F. Schmidt,

More information

Reliability Analysis

Reliability Analysis Measures of Reliability Reliability Analysis Reliability: the fact that a scale should consistently reflect the construct it is measuring. One way to think of reliability is that other things being equal,

More information

Economics of Strategy (ECON 4550) Maymester 2015 Applications of Regression Analysis

Economics of Strategy (ECON 4550) Maymester 2015 Applications of Regression Analysis Economics of Strategy (ECON 4550) Maymester 015 Applications of Regression Analysis Reading: ACME Clinic (ECON 4550 Coursepak, Page 47) and Big Suzy s Snack Cakes (ECON 4550 Coursepak, Page 51) Definitions

More information

Descriptive Statistics

Descriptive Statistics Descriptive Statistics Primer Descriptive statistics Central tendency Variation Relative position Relationships Calculating descriptive statistics Descriptive Statistics Purpose to describe or summarize

More information

Notes on Orthogonal and Symmetric Matrices MENU, Winter 2013

Notes on Orthogonal and Symmetric Matrices MENU, Winter 2013 Notes on Orthogonal and Symmetric Matrices MENU, Winter 201 These notes summarize the main properties and uses of orthogonal and symmetric matrices. We covered quite a bit of material regarding these topics,

More information

Correlation key concepts:

Correlation key concepts: CORRELATION Correlation key concepts: Types of correlation Methods of studying correlation a) Scatter diagram b) Karl pearson s coefficient of correlation c) Spearman s Rank correlation coefficient d)

More information

Module 5: Multiple Regression Analysis

Module 5: Multiple Regression Analysis Using Statistical Data Using to Make Statistical Decisions: Data Multiple to Make Regression Decisions Analysis Page 1 Module 5: Multiple Regression Analysis Tom Ilvento, University of Delaware, College

More information

Simple Predictive Analytics Curtis Seare

Simple Predictive Analytics Curtis Seare Using Excel to Solve Business Problems: Simple Predictive Analytics Curtis Seare Copyright: Vault Analytics July 2010 Contents Section I: Background Information Why use Predictive Analytics? How to use

More information

Stepwise Regression. Chapter 311. Introduction. Variable Selection Procedures. Forward (Step-Up) Selection

Stepwise Regression. Chapter 311. Introduction. Variable Selection Procedures. Forward (Step-Up) Selection Chapter 311 Introduction Often, theory and experience give only general direction as to which of a pool of candidate variables (including transformed variables) should be included in the regression model.

More information

Multivariate Analysis of Variance (MANOVA)

Multivariate Analysis of Variance (MANOVA) Chapter 415 Multivariate Analysis of Variance (MANOVA) Introduction Multivariate analysis of variance (MANOVA) is an extension of common analysis of variance (ANOVA). In ANOVA, differences among various

More information

How to Get More Value from Your Survey Data

How to Get More Value from Your Survey Data Technical report How to Get More Value from Your Survey Data Discover four advanced analysis techniques that make survey research more effective Table of contents Introduction..............................................................2

More information

MISSING DATA TECHNIQUES WITH SAS. IDRE Statistical Consulting Group

MISSING DATA TECHNIQUES WITH SAS. IDRE Statistical Consulting Group MISSING DATA TECHNIQUES WITH SAS IDRE Statistical Consulting Group ROAD MAP FOR TODAY To discuss: 1. Commonly used techniques for handling missing data, focusing on multiple imputation 2. Issues that could

More information

Analysing Questionnaires using Minitab (for SPSS queries contact -) Graham.Currell@uwe.ac.uk

Analysing Questionnaires using Minitab (for SPSS queries contact -) Graham.Currell@uwe.ac.uk Analysing Questionnaires using Minitab (for SPSS queries contact -) Graham.Currell@uwe.ac.uk Structure As a starting point it is useful to consider a basic questionnaire as containing three main sections:

More information

Factor Analysis Using SPSS

Factor Analysis Using SPSS Psychology 305 p. 1 Factor Analysis Using SPSS Overview For this computer assignment, you will conduct a series of principal factor analyses to examine the factor structure of a new instrument developed

More information

SPSS ADVANCED ANALYSIS WENDIANN SETHI SPRING 2011

SPSS ADVANCED ANALYSIS WENDIANN SETHI SPRING 2011 SPSS ADVANCED ANALYSIS WENDIANN SETHI SPRING 2011 Statistical techniques to be covered Explore relationships among variables Correlation Regression/Multiple regression Logistic regression Factor analysis

More information

Simple Regression Theory II 2010 Samuel L. Baker

Simple Regression Theory II 2010 Samuel L. Baker SIMPLE REGRESSION THEORY II 1 Simple Regression Theory II 2010 Samuel L. Baker Assessing how good the regression equation is likely to be Assignment 1A gets into drawing inferences about how close the

More information

Multiple regression - Matrices

Multiple regression - Matrices Multiple regression - Matrices This handout will present various matrices which are substantively interesting and/or provide useful means of summarizing the data for analytical purposes. As we will see,

More information

15.062 Data Mining: Algorithms and Applications Matrix Math Review

15.062 Data Mining: Algorithms and Applications Matrix Math Review .6 Data Mining: Algorithms and Applications Matrix Math Review The purpose of this document is to give a brief review of selected linear algebra concepts that will be useful for the course and to develop

More information

Factor Analysis - 2 nd TUTORIAL

Factor Analysis - 2 nd TUTORIAL Factor Analysis - 2 nd TUTORIAL Subject marks File sub_marks.csv shows correlation coefficients between subject scores for a sample of 220 boys. sub_marks

More information

Linear Models in STATA and ANOVA

Linear Models in STATA and ANOVA Session 4 Linear Models in STATA and ANOVA Page Strengths of Linear Relationships 4-2 A Note on Non-Linear Relationships 4-4 Multiple Linear Regression 4-5 Removal of Variables 4-8 Independent Samples

More information

6.4 Normal Distribution

6.4 Normal Distribution Contents 6.4 Normal Distribution....................... 381 6.4.1 Characteristics of the Normal Distribution....... 381 6.4.2 The Standardized Normal Distribution......... 385 6.4.3 Meaning of Areas under

More information

Practical Considerations for Using Exploratory Factor Analysis in Educational Research

Practical Considerations for Using Exploratory Factor Analysis in Educational Research A peer-reviewed electronic journal. Copyright is retained by the first or sole author, who grants right of first publication to the Practical Assessment, Research & Evaluation. Permission is granted to

More information

Computer-Aided Multivariate Analysis

Computer-Aided Multivariate Analysis Computer-Aided Multivariate Analysis FOURTH EDITION Abdelmonem Af if i Virginia A. Clark and Susanne May CHAPMAN & HALL/CRC A CRC Press Company Boca Raton London New York Washington, D.C Contents Preface

More information

GMAC. Which Programs Have the Highest Validity: Identifying Characteristics that Affect Prediction of Success 1

GMAC. Which Programs Have the Highest Validity: Identifying Characteristics that Affect Prediction of Success 1 GMAC Which Programs Have the Highest Validity: Identifying Characteristics that Affect Prediction of Success 1 Eileen Talento-Miller & Lawrence M. Rudner GMAC Research Reports RR-05-03 August 23, 2005

More information

[1] Diagonal factorization

[1] Diagonal factorization 8.03 LA.6: Diagonalization and Orthogonal Matrices [ Diagonal factorization [2 Solving systems of first order differential equations [3 Symmetric and Orthonormal Matrices [ Diagonal factorization Recall:

More information

Premaster Statistics Tutorial 4 Full solutions

Premaster Statistics Tutorial 4 Full solutions Premaster Statistics Tutorial 4 Full solutions Regression analysis Q1 (based on Doane & Seward, 4/E, 12.7) a. Interpret the slope of the fitted regression = 125,000 + 150. b. What is the prediction for

More information

Imputing Missing Data using SAS

Imputing Missing Data using SAS ABSTRACT Paper 3295-2015 Imputing Missing Data using SAS Christopher Yim, California Polytechnic State University, San Luis Obispo Missing data is an unfortunate reality of statistics. However, there are

More information

CHAPTER 8 FACTOR EXTRACTION BY MATRIX FACTORING TECHNIQUES. From Exploratory Factor Analysis Ledyard R Tucker and Robert C.

CHAPTER 8 FACTOR EXTRACTION BY MATRIX FACTORING TECHNIQUES. From Exploratory Factor Analysis Ledyard R Tucker and Robert C. CHAPTER 8 FACTOR EXTRACTION BY MATRIX FACTORING TECHNIQUES From Exploratory Factor Analysis Ledyard R Tucker and Robert C MacCallum 1997 180 CHAPTER 8 FACTOR EXTRACTION BY MATRIX FACTORING TECHNIQUES In

More information

Multivariate Analysis of Variance. The general purpose of multivariate analysis of variance (MANOVA) is to determine

Multivariate Analysis of Variance. The general purpose of multivariate analysis of variance (MANOVA) is to determine 2 - Manova 4.3.05 25 Multivariate Analysis of Variance What Multivariate Analysis of Variance is The general purpose of multivariate analysis of variance (MANOVA) is to determine whether multiple levels

More information