Statistics without probability: the French way Multivariate Data Analysis: Duality Diagrams
|
|
|
- Dorcas Peters
- 10 years ago
- Views:
Transcription
1 Statistics without probability: the French way Multivariate Data Analysis: Duality Diagrams Susan Holmes Last Updated October 31st, 2007 jeihgfdcbabakl
2 Outline I. Historical Perspective: 1970 s. II. Maths but No Probability III. Special Geometrical Tricks: visualization. IV. Perturbation Methods V. Summary and References
3 I. Analyse des Données Part I Introduction: History
4 I. Analyse des Données Part I Introduction: History Thirty years ago...
5 I. Analyse des Données 1960 s 1970 s: Those were the days Here..
6 I. Analyse des Données 1960 s 1970 s: Those were the days Here..
7 I. Analyse des Données 1960 s 1970 s: Those were the days Here..
8 I. Analyse des Données 1960 s 1970 s: Those were the days There
9 I. Analyse des Données 1960 s 1970 s: Those were the days There
10 I. Analyse des Données EDA=English for Analyse des Données Resist the mathematicians bid to take over statistics (Tukey). Take away the control of the statistics jobs by the probabilists in France.
11 I. Analyse des Données EDA=English for Analyse des Données Resist the mathematicians bid to take over statistics (Tukey). Take away the control of the statistics jobs by the probabilists in France. (They failed).
12 I. Analyse des Données Abstraction Rules Bourbaki:
13 I. Analyse des Données Abstraction Rules Bourbaki: An overpowering triplet (Ω, A, P)
14 I. Analyse des Données Abstraction Rules Bourbaki: Creation of a new triplet (X, Q, D) and an algebraic and geometric framework.
15 Geometry and Algebra Part II Diagrams and Duality
16 Geometry and Algebra Geometry of Data Analysis: Rows and Columns
17 Geometry and Algebra A favorable review (Ramsay and de Leeuw) Book Review (Psychometrika, 1983) Quote: This remarkable book treats multivariate linear analysis in a manner that is with both distinctive and profoundly promising for future work in this field. With an approach that is strictly algebraic and geometric, it avoids almost all discussion of probabilistic notions, introduces a formalism that transcends matrix algebra, and offers a coherent treatment of topics not often found within the same volume. Finally, it achieves these things while remaining entirely accessible to nonmathematicians and including many excellent practical examples.
18 Geometry and Algebra In summary Introduction à l Analyse des Données offers a treatment of the multivariate linear model which is (a) metric and basis free, (b) offers a unified survey of both quantitative and certain qualitative procedures, (c) incorporates classical multidimensional scaling in a natural way, and (d) invites through its powerful formalism an extension in a number of valuable directions. We hope it will not be long before an English language counterpart appears.
19 Geometry and Algebra Geometry of Data Analysis: Rows and Columns
20 Geometry and Algebra Who spent time at the labs in the mid 1960 s? Shepard Tukey Mallows Kruskal
21 Geometry and Algebra Who spent time at the labs in the mid 1960 s? Shepard Tukey Mallows Kruskal and Benzecri
22 Geometry and Algebra A Geometrical Approach i. The data are p variables measured on n observations. ii. X with n rows (the observations) and p columns (the variables). iii. D n is an n n matrix of weights on the observations, which is most often diagonal. iv Symmetric definite positive matrix Q,often
23 Geometry and Algebra A Geometrical Approach i. The data are p variables measured on n observations. ii. X with n rows (the observations) and p columns (the variables). iii. D n is an n n matrix of weights on the observations, which is most often diagonal. iv Symmetric definite positive matrix Q,often σ Q = σ σ σp 2
24 Geometry and Algebra Euclidean Spaces These three matrices form the essential triplet (X, Q, D) defining a multivariate data analysis. Q and D define geometries or inner products in R p and R n, respectively, through x t Qy =< x, y > Q x, y R p x t Dy =< x, y > D x, y R n.
25 Geometry and Algebra An Algebraic Approach Q can be seen as a linear function from R p to R p = L(R p ), the space of scalar linear functions on R p. D can be seen as a linear function from R n to R n = L(R n ). Q R p R p X V X t D R n R n W
26 Geometry and Algebra An Algebraic Approach Duality diagram Q R p R p X V X t D R n R n i. Eigendecomposition of X t DXQ = VQ ii. Eigendecomposition of XQX t D = WD iii. W
27 Geometry and Algebra Notes (1) Suppose we have data and inner products defined by Q and D : (x, y) R p R p x t Qy = < x, y > Q R (x, y) R n R n x t Dy = < x, y > D R. p x 2 Q =< x, x > Q= q j (x.j ) 2 x 2 D =< x, x > D= j=1 p p i (x i. ) 2 (2) We say an operator O is B-symmetric if < x, Oy > B =< Ox, y > B, or equivalently BO = O t B. The duality diagram is equivalent to (X, Q, D) such that X is n p. Escoufier (1977) defined as XQX t D = WD and X t DXQ = VQ as the characteristic operators of the diagram. j=1
28 Geometry and Algebra (3) V = X t DX will be the variance-covariance matrix, if X is centered with regards to D (X D1 n = 0).
29 Geometry and Algebra Transposable Data There is an important symmetry between the rows and columns of X in the diagram, and one can imagine situations where the role of observation or variable is not uniquely defined. For instance in microarray studies the genes can be considered either as variables or observations. This makes sense in many contemporary situations which evade the more classical notion of n observations seen as a random sample of a population. It is certainly not the case that the 30,000 probes are a sample of genes since these probes try to be an exhaustive set.
30 Geometry and Algebra Two Dual Geometries
31 Geometry and Algebra Properties of the Diagram Rank of the diagram: X, X t, VQ and WD all have the same rank. For Q and D symmetric matrices, VQ and WD are diagonalisable and have the same eigenvalues. λ 1 λ 2 λ 3... λ r 0 0. Eigendecomposition of the diagram: VQ is Q symmetric, thus we can find Z such that VQZ = ZΛ, Z t QZ = I p, where Λ = diag(λ 1, λ 2,..., λ p ). (1)
32 Geometry and Algebra Practical Computations Cholesky decompositions of Q and D, (symmetric and positive definite) H t H = Q and K t K = D. Use the singular value decomposition of KXH: KXH = UST t, with T t T = I p, U t U = I n, S diagonal. Then Z = (H 1 ) t T satisfies VQZ = ZΛ, Z t QZ = I p with Λ = S 2. The renormalized columns of Z, A = SZ are called the principal axes and satisfy: A t QA = Λ.
33 Geometry and Algebra Practical Computations Similarly, we can define L = K 1 U that satisfies WDL = LΛ, L t DL = I n, where Λ = diag(λ 1, λ 2,..., λ r, 0,..., 0). (2) C = LS is usually called the matrix of principal components. It is normed so that C t DC = Λ.
34 Geometry and Algebra Transition Formulæ: Of the four matrices Z, A, L and C we only have to compute one, all others are obtained by the transition formulæ provided by the duality property of the diagram: XQZ = LS = C X t DL = ZS = A
35 Geometry and Algebra French Features Inertia: Trace(VQ) = Trace(WD) (inertia in the sense of Huyghens inertia formula for instance). Huygens, C. (1657), De ratiociniis in ludo alea, printed in Exercitationium mathematicaram by F. van Schooten. Elsevirii, Leiden. n p i d 2 (x i, a) i=1 Inertia with regards to a point a of a cloud of p i -weighted points. PCA with Q = I p, D = 1 n I n, and the variables are centered, the inertia is the sum of the variances of all the variables. If the variables are standardized (Q is the diagonal matrix of inverse variances), then the inertia is the number of variables p. For correspondence analysis the inertia is the Chi-squared statistic.
36 Geometry and Algebra Dimension Reduction: Eckart-Young X [k] = US [k] V is the best k rank approximation to X.
37 Visualization Tools Geometric Interpretations of Statistical Quantities x = x D1 n ˆσ x 2 = i call X = (I 1 n D1 n)x p i (x i x) = x 2 D covariance cov(x, y) =< x, ỹ > D correlation r xy = < x, ỹ > D x D ỹ D = cos( x, ỹ)
38 Visualization Tools Quality of Representations The cosine again, cos(x, y) = <x,y> x y cos 2 α = ˆx 2 x 2 tells us how well x is represented by its projection.
39 Visualization Tools Inertia and Contributions In(X ) = X 2 = i p i x i. 2 = j q j x.j 2 = p l=1 λ l Contribution of an observation to the total inertia: p i x i. 2 Contribution of a variable to the total inertia: q j x.j 2 X 2 X 2
40 Visualization Tools Inertia and Contributions λ k v 2 kj Contribution of the kth axis to variable j: x j 2 D Contribution of variable j to the kth axis q j vkj 2. Contribution of the kth axis to observation i: λ k u 2 ik x i 2 Q Contribution of observation i to the kth axis p i u 2 ik.
41 Visualization Tools Comparing Two Diagrams: the RV coefficient Many problems can be rephrased in terms of comparison of two duality diagrams or put more simply, two characterizing operators, built from two triplets, usually with one of the triplets being a response or having constraints imposed on it. Most often what is done is to compare two such diagrams, and try to get one to match the other in some optimal way. To compare two symmetric operators, there is either a vector covariance as inner product covv (O 1, O 2 ) = Tr(O 1 O 2 ) =< O 1, O 2 > or a vector correlation [Escoufier, 1977] RV (O 1, O 2 ) = Tr(O 1 O 2 ) Tr(O t 1 O 1 )tr(o t 2 O 2). If we were to compare the two triplets ( ( X n 1, 1, 1 n I n) and Yn 1, 1, 1 n I ) n we would have RV = ρ 2.
42 Visualization Tools PCA: Special case PCA can be seen as finding the matrix Y which maximizes the RV coefficient between characterizing operators, that is, between (X n p, Q, D) and (Y n q, I, D), under the constraint that Y be of rank q < p. RV ( XQX t D, YY t D ) = Tr (XQX t DYY t D) Tr (XQX t D) 2 Tr (YY t D) 2.
43 Visualization Tools This maximum is attained where Y is chosen as the first q eigenvectors of XQX t D normed so that Y t DY = Λ q. The maximum RV is RVmax = q i=1 λ2 i p i=1 λ2 i Of course, classical PCA has D = 1 n I, Q = I, but the extra flexibility is often useful. We define the distance between triplets (X, Q, D) and (Z, Q, M) where Z is also n p, as the distance deduced from the RV inner product between operators XQX t D and ZMZ t D..
44 Visualization Tools One Diagram to replace Two Diagrams Canonical correlation analysis was introduced by Hotelling[Hotelling, 1936] to find the common structure in two sets of variables X 1 and X 2 measured on the same observations. This is equivalent to merging the two matrices columnwise to form a large matrix with n rows and p 1 + p 2 columns and taking as the weighting of the variables the matrix defined by the two diagonal blocks (X t 1 DX 1) 1 and (X t 2 DX 2) 1 Q = (X t 1 DX 1) (X t 2 DX 2) 1
45 Visualization Tools This analysis gives the same eigenvectors as the analysis of the triple (X t 2 DX 1, (X t 1 DX 1) 1, (X t 2 DX 2) 1 ), also known as the canonical X1 I p1 V 1 D R p 1 R p 1 X t 1 R n R n W 1 Q R p 1+p 2 R p 1+p 2 [X 1 ;X 2 ] V X2 I p2 V 2 D D R n [X 1 ;X 2 ] t R n R p 2 R p 2 W X t 2 R n R n W 2
46 Visualization Tools correlation analysis of X 1 and X 2.
47 Visualization Tools Circle of Correlations
48 Visualization Tools Circle of Correlations for Score data mec vec alg ana sta
49 Pertubative Methods Part IV Perturbation for Validation and Discovery
50 Pertubative Methods Internal and External Methods Cross Validation: Leave one variable out, or leave one observation out, or both. Bootstrap, bootstrap rows or columns, partial bootstrap, where can we compare them? Convex Hulls. Procrustes solutions for data cubes (STATIS).
51 Pertubative Methods Successful Perturbative Method for Non-hierarchical Clustering Dynamical Clusters: Edwin Diday, 1970, [Diday, 1973] Repeated k-means with fixed class sizes. Choose a set of k nuclei (usually from the data). Partition the data as the nearest neighbors to each of the k points. For each partition define its centroid. Iterate the above 2 steps until convergence. This process gives a set of clusters. Organize these clusters according to sets of strong forms the ones that were always together (or mostly) together.
52 Conclusions Part V Conclusions: Data Analysis and Data Integration
53 Conclusions Computer Intensive Data Analysis Today i. Interactive. ii. Iteration. iii. Nonparametric. iv Heterogeneous Data. v Kernel Methods.
54 Conclusions Computer Intensive Data Analysis Today i. Interactive. ii. Iteration. iii. Nonparametric. iv Heterogeneous Data. v Kernel Methods. No more in statistics departments at Universities in France All in institutes, INRA, INRIA, INSERM,ENSET,ENSAEE,...à suivre...
55 Part VI V. References
56 Reading Few references in English explaining the duality/operator point of view. H. 2006, Multivariate Data Analysis: the French way.[holmes, 2006] French: Escoufier [Escoufier, 1987, Escoufier, 1977]. Fréderique Glaçon s PhD thesis [Glaçon, 1981] (in French) on data cubes. Masters level textbooks on the subject for many details and examples: Brigitte Escofier and Jérôme Pagès [Escofier and Pagès, 1998] do not delve into the Duality Diagram [Lebart et al., 2000] is one of the broader books on multivariate analyses Cailliez and Pagès [Cailliez and Pages., 1976] is hard to find, but was the first textbook completely based on the diagram approach, as was the case in the earlier literature they use transposed matrices. Stability studies: [Holmes, 1985],[Lebart et al., 2000].
57 Software The methods described in this article are all available in the form of R packages which I recommend. ade4 [Chessel et al., 2004] However, a complete understanding of the duality diagram terminology and philosophy is necessary. One of the most important features in all the dudi.* functions is that when the argument scannf is at its default value TRUE, the first step imposed on the user is the perusal of the scree plot of eigenvalues. vegan ecological community
58 Functions Principal Components Analysis (PCA) is available in prcomp and princomp in the standard package stats as pca in vegan and as dudi.pca in ade4. Two versions of PCAIV are available, one is called Redundancy Analysis (RDA) and is available as rda in vegan and pcaiv in ade4. Correspondence Analysis (CA) is available in cca in vegan and as dudi.coa in ade4. Discriminant analysis is available as lda in stats, as discrimin in ade4 Canonical Correlation Analysis is available in cancor in stats (Beware cca in ade4 is Canonical Correspondence Analysis). STATIS (Conjoint analysis of several tables) is available in ade4.
59 Cailliez, F. and Pages., J. P. (1976). Introduction à l analyse des donnés. SMASH, Paris. Chessel, D., Dufour, A. B., and Thioulouse., J. (2004). The ade4 package - i: One-table methods. R News, 4(1):5 10. Diday, E. (1973). The dynamic clusters method in nonhierarchical clustering. International Journal of Computer and Information Sciences, 2(1): Escofier, B. and Pagès, J. (1998). Analyse factorielles simples et multiples : Objectifs, méthodes et interprétation. Dunod. Escoufier, Y. (1977). Operators related to a data matrix.
60 In Barra, J. e. a., editor, Recent developments in Statistics., pages North Holland,. Escoufier, Y. (1987). The duality diagram: a means of better practical applications. In In Legendre, P. and Legendre, L., editors, Development. in numerical ecology., pages Glaçon, F. (1981). Analyse conjointe de plusieurs matrices de données. Comparaison de différentes méthodes. PhD thesis, Grenoble. Holmes, S. (1985). Outils Informatiques pour l Evaluation de la Pertinence d un Resultat en Analyse des Données. PhD thesis, Montpellier, USTL. Holmes, S. (2006). Multivariate analysis: The french way. Feschrift for David Freedman, IMS.
61 Hotelling, H. (1936). Relations between two sets of variables. Biometrika, 28: Lebart, L., Piron, M., and Morineau, A. (2000). Statistique exploratoire multidimensionnelle. Dunod, Paris, France.
62 Acknowledgements Chessel, Thioulouse, ADE4 team. Persi Diaconis. Elisabeth Purdom. Yves Escoufier. NSF-DMS award to SH.
Tutorial on Exploratory Data Analysis
Tutorial on Exploratory Data Analysis Julie Josse, François Husson, Sébastien Lê julie.josse at agrocampus-ouest.fr francois.husson at agrocampus-ouest.fr Applied Mathematics Department, Agrocampus Ouest
15.062 Data Mining: Algorithms and Applications Matrix Math Review
.6 Data Mining: Algorithms and Applications Matrix Math Review The purpose of this document is to give a brief review of selected linear algebra concepts that will be useful for the course and to develop
Journal of Statistical Software
JSS Journal of Statistical Software March 2008, Volume 25, Issue 1. http://www.jstatsoft.org/ FactoMineR: An R Package for Multivariate Analysis Sébastien Lê Agrocampus Rennes Julie Josse Agrocampus Rennes
How To Cluster
Data Clustering Dec 2nd, 2013 Kyrylo Bessonov Talk outline Introduction to clustering Types of clustering Supervised Unsupervised Similarity measures Main clustering algorithms k-means Hierarchical Main
Partial Least Squares (PLS) Regression.
Partial Least Squares (PLS) Regression. Hervé Abdi 1 The University of Texas at Dallas Introduction Pls regression is a recent technique that generalizes and combines features from principal component
How To Understand Multivariate Models
Neil H. Timm Applied Multivariate Analysis With 42 Figures Springer Contents Preface Acknowledgments List of Tables List of Figures vii ix xix xxiii 1 Introduction 1 1.1 Overview 1 1.2 Multivariate Models
Visualization of textual data: unfolding the Kohonen maps.
Visualization of textual data: unfolding the Kohonen maps. CNRS - GET - ENST 46 rue Barrault, 75013, Paris, France (e-mail: [email protected]) Ludovic Lebart Abstract. The Kohonen self organizing
Medical Information Management & Mining. You Chen Jan,15, 2013 [email protected]
Medical Information Management & Mining You Chen Jan,15, 2013 [email protected] 1 Trees Building Materials Trees cannot be used to build a house directly. How can we transform trees to building materials?
Mehtap Ergüven Abstract of Ph.D. Dissertation for the degree of PhD of Engineering in Informatics
INTERNATIONAL BLACK SEA UNIVERSITY COMPUTER TECHNOLOGIES AND ENGINEERING FACULTY ELABORATION OF AN ALGORITHM OF DETECTING TESTS DIMENSIONALITY Mehtap Ergüven Abstract of Ph.D. Dissertation for the degree
Manifold Learning Examples PCA, LLE and ISOMAP
Manifold Learning Examples PCA, LLE and ISOMAP Dan Ventura October 14, 28 Abstract We try to give a helpful concrete example that demonstrates how to use PCA, LLE and Isomap, attempts to provide some intuition
Exploratory data analysis for microarray data
Eploratory data analysis for microarray data Anja von Heydebreck Ma Planck Institute for Molecular Genetics, Dept. Computational Molecular Biology, Berlin, Germany [email protected] Visualization
CHAPTER 8 FACTOR EXTRACTION BY MATRIX FACTORING TECHNIQUES. From Exploratory Factor Analysis Ledyard R Tucker and Robert C.
CHAPTER 8 FACTOR EXTRACTION BY MATRIX FACTORING TECHNIQUES From Exploratory Factor Analysis Ledyard R Tucker and Robert C MacCallum 1997 180 CHAPTER 8 FACTOR EXTRACTION BY MATRIX FACTORING TECHNIQUES In
Introduction to Matrix Algebra
Psychology 7291: Multivariate Statistics (Carey) 8/27/98 Matrix Algebra - 1 Introduction to Matrix Algebra Definitions: A matrix is a collection of numbers ordered by rows and columns. It is customary
Nonlinear Iterative Partial Least Squares Method
Numerical Methods for Determining Principal Component Analysis Abstract Factors Béchu, S., Richard-Plouet, M., Fernandez, V., Walton, J., and Fairley, N. (2016) Developments in numerical treatments for
Linear Algebra Review. Vectors
Linear Algebra Review By Tim K. Marks UCSD Borrows heavily from: Jana Kosecka [email protected] http://cs.gmu.edu/~kosecka/cs682.html Virginia de Sa Cogsci 8F Linear Algebra review UCSD Vectors The length
Metric Multidimensional Scaling (MDS): Analyzing Distance Matrices
Metric Multidimensional Scaling (MDS): Analyzing Distance Matrices Hervé Abdi 1 1 Overview Metric multidimensional scaling (MDS) transforms a distance matrix into a set of coordinates such that the (Euclidean)
Multiple factor analysis: principal component analysis for multitable and multiblock data sets
Multiple factor analysis: principal component analysis for multitable and multiblock data sets Hervé Abdi 1, Lynne J. Williams 2 and Domininique Valentin 3 Multiple factor analysis (MFA, also called multiple
Review Jeopardy. Blue vs. Orange. Review Jeopardy
Review Jeopardy Blue vs. Orange Review Jeopardy Jeopardy Round Lectures 0-3 Jeopardy Round $200 How could I measure how far apart (i.e. how different) two observations, y 1 and y 2, are from each other?
Linear Algebra Notes for Marsden and Tromba Vector Calculus
Linear Algebra Notes for Marsden and Tromba Vector Calculus n-dimensional Euclidean Space and Matrices Definition of n space As was learned in Math b, a point in Euclidean three space can be thought of
Similarity and Diagonalization. Similar Matrices
MATH022 Linear Algebra Brief lecture notes 48 Similarity and Diagonalization Similar Matrices Let A and B be n n matrices. We say that A is similar to B if there is an invertible n n matrix P such that
Multivariate Analysis of Ecological Data
Multivariate Analysis of Ecological Data MICHAEL GREENACRE Professor of Statistics at the Pompeu Fabra University in Barcelona, Spain RAUL PRIMICERIO Associate Professor of Ecology, Evolutionary Biology
Steven M. Ho!and. Department of Geology, University of Georgia, Athens, GA 30602-2501
PRINCIPAL COMPONENTS ANALYSIS (PCA) Steven M. Ho!and Department of Geology, University of Georgia, Athens, GA 30602-2501 May 2008 Introduction Suppose we had measured two variables, length and width, and
Principle Component Analysis and Partial Least Squares: Two Dimension Reduction Techniques for Regression
Principle Component Analysis and Partial Least Squares: Two Dimension Reduction Techniques for Regression Saikat Maitra and Jun Yan Abstract: Dimension reduction is one of the major tasks for multivariate
Matrix Calculations: Applications of Eigenvalues and Eigenvectors; Inner Products
Matrix Calculations: Applications of Eigenvalues and Eigenvectors; Inner Products H. Geuvers Institute for Computing and Information Sciences Intelligent Systems Version: spring 2015 H. Geuvers Version:
LABEL PROPAGATION ON GRAPHS. SEMI-SUPERVISED LEARNING. ----Changsheng Liu 10-30-2014
LABEL PROPAGATION ON GRAPHS. SEMI-SUPERVISED LEARNING ----Changsheng Liu 10-30-2014 Agenda Semi Supervised Learning Topics in Semi Supervised Learning Label Propagation Local and global consistency Graph
APPM4720/5720: Fast algorithms for big data. Gunnar Martinsson The University of Colorado at Boulder
APPM4720/5720: Fast algorithms for big data Gunnar Martinsson The University of Colorado at Boulder Course objectives: The purpose of this course is to teach efficient algorithms for processing very large
Chapter 6. Orthogonality
6.3 Orthogonal Matrices 1 Chapter 6. Orthogonality 6.3 Orthogonal Matrices Definition 6.4. An n n matrix A is orthogonal if A T A = I. Note. We will see that the columns of an orthogonal matrix must be
Introduction: Overview of Kernel Methods
Introduction: Overview of Kernel Methods Statistical Data Analysis with Positive Definite Kernels Kenji Fukumizu Institute of Statistical Mathematics, ROIS Department of Statistical Science, Graduate University
ADE SOFTWARE: MULTIVARIATE ANALYSIS AND GRAPHICAL DISPLAY OF ENVIRONMENTAL DATA
ADE SOFTWARE: MULTIVARIATE ANALYSIS AND GRAPHICAL DISPLAY OF ENVIRONMENTAL DATA J. Thioulouse Laboratoire de Biométrie, Génétique et Biologie des Populations URA CNRS 243, Université Lyon 1 69622 Villeurbanne
13 MATH FACTS 101. 2 a = 1. 7. The elements of a vector have a graphical interpretation, which is particularly easy to see in two or three dimensions.
3 MATH FACTS 0 3 MATH FACTS 3. Vectors 3.. Definition We use the overhead arrow to denote a column vector, i.e., a linear segment with a direction. For example, in three-space, we write a vector in terms
Partial Least Square Regression PLS-Regression
Partial Least Square Regression Hervé Abdi 1 1 Overview PLS regression is a recent technique that generalizes and combines features from principal component analysis and multiple regression. Its goal is
Vector Spaces; the Space R n
Vector Spaces; the Space R n Vector Spaces A vector space (over the real numbers) is a set V of mathematical entities, called vectors, U, V, W, etc, in which an addition operation + is defined and in which
Factor analysis. Angela Montanari
Factor analysis Angela Montanari 1 Introduction Factor analysis is a statistical model that allows to explain the correlations between a large number of observed correlated variables through a small number
1 2 3 1 1 2 x = + x 2 + x 4 1 0 1
(d) If the vector b is the sum of the four columns of A, write down the complete solution to Ax = b. 1 2 3 1 1 2 x = + x 2 + x 4 1 0 0 1 0 1 2. (11 points) This problem finds the curve y = C + D 2 t which
Computer program review
Journal of Vegetation Science 16: 355-359, 2005 IAVS; Opulus Press Uppsala. - Ginkgo, a multivariate analysis package - 355 Computer program review Ginkgo, a multivariate analysis package Bouxin, Guy Haute
Multivariate Statistical Inference and Applications
Multivariate Statistical Inference and Applications ALVIN C. RENCHER Department of Statistics Brigham Young University A Wiley-Interscience Publication JOHN WILEY & SONS, INC. New York Chichester Weinheim
CITY UNIVERSITY LONDON. BEng Degree in Computer Systems Engineering Part II BSc Degree in Computer Systems Engineering Part III PART 2 EXAMINATION
No: CITY UNIVERSITY LONDON BEng Degree in Computer Systems Engineering Part II BSc Degree in Computer Systems Engineering Part III PART 2 EXAMINATION ENGINEERING MATHEMATICS 2 (resit) EX2005 Date: August
DATA MINING CLUSTER ANALYSIS: BASIC CONCEPTS
DATA MINING CLUSTER ANALYSIS: BASIC CONCEPTS 1 AND ALGORITHMS Chiara Renso KDD-LAB ISTI- CNR, Pisa, Italy WHAT IS CLUSTER ANALYSIS? Finding groups of objects such that the objects in a group will be similar
Dimensionality Reduction: Principal Components Analysis
Dimensionality Reduction: Principal Components Analysis In data mining one often encounters situations where there are a large number of variables in the database. In such situations it is very likely
Numerical Analysis Lecture Notes
Numerical Analysis Lecture Notes Peter J. Olver 5. Inner Products and Norms The norm of a vector is a measure of its size. Besides the familiar Euclidean norm based on the dot product, there are a number
Torgerson s Classical MDS derivation: 1: Determining Coordinates from Euclidean Distances
Torgerson s Classical MDS derivation: 1: Determining Coordinates from Euclidean Distances It is possible to construct a matrix X of Cartesian coordinates of points in Euclidean space when we know the Euclidean
ISOMETRIES OF R n KEITH CONRAD
ISOMETRIES OF R n KEITH CONRAD 1. Introduction An isometry of R n is a function h: R n R n that preserves the distance between vectors: h(v) h(w) = v w for all v and w in R n, where (x 1,..., x n ) = x
Lecture 3: Linear methods for classification
Lecture 3: Linear methods for classification Rafael A. Irizarry and Hector Corrada Bravo February, 2010 Today we describe four specific algorithms useful for classification problems: linear regression,
1 Introduction to Matrices
1 Introduction to Matrices In this section, important definitions and results from matrix algebra that are useful in regression analysis are introduced. While all statements below regarding the columns
DATA ANALYSIS II. Matrix Algorithms
DATA ANALYSIS II Matrix Algorithms Similarity Matrix Given a dataset D = {x i }, i=1,..,n consisting of n points in R d, let A denote the n n symmetric similarity matrix between the points, given as where
4.7. Canonical ordination
Université Laval Analyse multivariable - mars-avril 2008 1 4.7.1 Introduction 4.7. Canonical ordination The ordination methods reviewed above are meant to represent the variation of a data matrix in a
Applied Linear Algebra I Review page 1
Applied Linear Algebra Review 1 I. Determinants A. Definition of a determinant 1. Using sum a. Permutations i. Sign of a permutation ii. Cycle 2. Uniqueness of the determinant function in terms of properties
[1] Diagonal factorization
8.03 LA.6: Diagonalization and Orthogonal Matrices [ Diagonal factorization [2 Solving systems of first order differential equations [3 Symmetric and Orthonormal Matrices [ Diagonal factorization Recall:
Orthogonal Diagonalization of Symmetric Matrices
MATH10212 Linear Algebra Brief lecture notes 57 Gram Schmidt Process enables us to find an orthogonal basis of a subspace. Let u 1,..., u k be a basis of a subspace V of R n. We begin the process of finding
Multimedia Databases. Wolf-Tilo Balke Philipp Wille Institut für Informationssysteme Technische Universität Braunschweig http://www.ifis.cs.tu-bs.
Multimedia Databases Wolf-Tilo Balke Philipp Wille Institut für Informationssysteme Technische Universität Braunschweig http://www.ifis.cs.tu-bs.de 14 Previous Lecture 13 Indexes for Multimedia Data 13.1
Principal Component Analysis
Principal Component Analysis ERS70D George Fernandez INTRODUCTION Analysis of multivariate data plays a key role in data analysis. Multivariate data consists of many different attributes or variables recorded
Spazi vettoriali e misure di similaritá
Spazi vettoriali e misure di similaritá R. Basili Corso di Web Mining e Retrieval a.a. 2008-9 April 10, 2009 Outline Outline Spazi vettoriali a valori reali Operazioni tra vettori Indipendenza Lineare
Similar matrices and Jordan form
Similar matrices and Jordan form We ve nearly covered the entire heart of linear algebra once we ve finished singular value decompositions we ll have seen all the most central topics. A T A is positive
December 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B. KITCHENS
December 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B KITCHENS The equation 1 Lines in two-dimensional space (1) 2x y = 3 describes a line in two-dimensional space The coefficients of x and y in the equation
Bindel, Spring 2012 Intro to Scientific Computing (CS 3220) Week 3: Wednesday, Feb 8
Spaces and bases Week 3: Wednesday, Feb 8 I have two favorite vector spaces 1 : R n and the space P d of polynomials of degree at most d. For R n, we have a canonical basis: R n = span{e 1, e 2,..., e
Eigenvalues, Eigenvectors, Matrix Factoring, and Principal Components
Eigenvalues, Eigenvectors, Matrix Factoring, and Principal Components The eigenvalues and eigenvectors of a square matrix play a key role in some important operations in statistics. In particular, they
The ade4 package - I : One-table methods
Vol. 4/1, June 2004 5 The ade4 package - I : One-table methods by Daniel Chessel, Anne B Dufour and Jean Thioulouse Introduction This paper is a short summary of the main classes defined in the ade4 package
Factor Analysis. Chapter 420. Introduction
Chapter 420 Introduction (FA) is an exploratory technique applied to a set of observed variables that seeks to find underlying factors (subsets of variables) from which the observed variables were generated.
Mathematics Course 111: Algebra I Part IV: Vector Spaces
Mathematics Course 111: Algebra I Part IV: Vector Spaces D. R. Wilkins Academic Year 1996-7 9 Vector Spaces A vector space over some field K is an algebraic structure consisting of a set V on which are
How To Solve The Cluster Algorithm
Cluster Algorithms Adriano Cruz [email protected] 28 de outubro de 2013 Adriano Cruz [email protected] () Cluster Algorithms 28 de outubro de 2013 1 / 80 Summary 1 K-Means Adriano Cruz [email protected]
Didacticiel - Études de cas
1 Topic Linear Discriminant Analysis Data Mining Tools Comparison (Tanagra, R, SAS and SPSS). Linear discriminant analysis is a popular method in domains of statistics, machine learning and pattern recognition.
Effective Linear Discriminant Analysis for High Dimensional, Low Sample Size Data
Effective Linear Discriant Analysis for High Dimensional, Low Sample Size Data Zhihua Qiao, Lan Zhou and Jianhua Z. Huang Abstract In the so-called high dimensional, low sample size (HDLSS) settings, LDA
MULTIVARIATE DATA ANALYSIS WITH PCA, CA AND MS TORSTEN MADSEN 2007
MULTIVARIATE DATA ANALYSIS WITH PCA, CA AND MS TORSTEN MADSEN 2007 Archaeological material that we wish to analyse through formalised methods has to be described prior to analysis in a standardised, formalised
Linear algebra and the geometry of quadratic equations. Similarity transformations and orthogonal matrices
MATH 30 Differential Equations Spring 006 Linear algebra and the geometry of quadratic equations Similarity transformations and orthogonal matrices First, some things to recall from linear algebra Two
Linear Threshold Units
Linear Threshold Units w x hx (... w n x n w We assume that each feature x j and each weight w j is a real number (we will relax this later) We will study three different algorithms for learning linear
High-Dimensional Data Visualization by PCA and LDA
High-Dimensional Data Visualization by PCA and LDA Chaur-Chin Chen Department of Computer Science, National Tsing Hua University, Hsinchu, Taiwan Abbie Hsu Institute of Information Systems & Applications,
THREE DIMENSIONAL GEOMETRY
Chapter 8 THREE DIMENSIONAL GEOMETRY 8.1 Introduction In this chapter we present a vector algebra approach to three dimensional geometry. The aim is to present standard properties of lines and planes,
Linear Algebra and TI 89
Linear Algebra and TI 89 Abdul Hassen and Jay Schiffman This short manual is a quick guide to the use of TI89 for Linear Algebra. We do this in two sections. In the first section, we will go over the editing
Comparison of Non-linear Dimensionality Reduction Techniques for Classification with Gene Expression Microarray Data
CMPE 59H Comparison of Non-linear Dimensionality Reduction Techniques for Classification with Gene Expression Microarray Data Term Project Report Fatma Güney, Kübra Kalkan 1/15/2013 Keywords: Non-linear
QUALITY ENGINEERING PROGRAM
QUALITY ENGINEERING PROGRAM Production engineering deals with the practical engineering problems that occur in manufacturing planning, manufacturing processes and in the integration of the facilities and
Linear Algebra Done Wrong. Sergei Treil. Department of Mathematics, Brown University
Linear Algebra Done Wrong Sergei Treil Department of Mathematics, Brown University Copyright c Sergei Treil, 2004, 2009, 2011, 2014 Preface The title of the book sounds a bit mysterious. Why should anyone
MATH 551 - APPLIED MATRIX THEORY
MATH 55 - APPLIED MATRIX THEORY FINAL TEST: SAMPLE with SOLUTIONS (25 points NAME: PROBLEM (3 points A web of 5 pages is described by a directed graph whose matrix is given by A Do the following ( points
International comparisons of road safety using Singular Value Decomposition
International comparisons of road safety using Singular Value Decomposition Siem Oppe D-2001-9 International comparisons of road safety using Singular Value Decomposition D-2001-9 Siem Oppe Leidschendam,
Section 1.1. Introduction to R n
The Calculus of Functions of Several Variables Section. Introduction to R n Calculus is the study of functional relationships and how related quantities change with each other. In your first exposure to
Factorization Theorems
Chapter 7 Factorization Theorems This chapter highlights a few of the many factorization theorems for matrices While some factorization results are relatively direct, others are iterative While some factorization
Exploratory Factor Analysis and Principal Components. Pekka Malo & Anton Frantsev 30E00500 Quantitative Empirical Research Spring 2016
and Principal Components Pekka Malo & Anton Frantsev 30E00500 Quantitative Empirical Research Spring 2016 Agenda Brief History and Introductory Example Factor Model Factor Equation Estimation of Loadings
EM Clustering Approach for Multi-Dimensional Analysis of Big Data Set
EM Clustering Approach for Multi-Dimensional Analysis of Big Data Set Amhmed A. Bhih School of Electrical and Electronic Engineering Princy Johnson School of Electrical and Electronic Engineering Martin
The Singular Value Decomposition in Symmetric (Löwdin) Orthogonalization and Data Compression
The Singular Value Decomposition in Symmetric (Löwdin) Orthogonalization and Data Compression The SVD is the most generally applicable of the orthogonal-diagonal-orthogonal type matrix decompositions Every
Lecture 5: Singular Value Decomposition SVD (1)
EEM3L1: Numerical and Analytical Techniques Lecture 5: Singular Value Decomposition SVD (1) EE3L1, slide 1, Version 4: 25-Sep-02 Motivation for SVD (1) SVD = Singular Value Decomposition Consider the system
α = u v. In other words, Orthogonal Projection
Orthogonal Projection Given any nonzero vector v, it is possible to decompose an arbitrary vector u into a component that points in the direction of v and one that points in a direction orthogonal to v
SALEM COMMUNITY COLLEGE Carneys Point, New Jersey 08069 COURSE SYLLABUS COVER SHEET. Action Taken (Please Check One) New Course Initiated
SALEM COMMUNITY COLLEGE Carneys Point, New Jersey 08069 COURSE SYLLABUS COVER SHEET Course Title Course Number Department Linear Algebra Mathematics MAT-240 Action Taken (Please Check One) New Course Initiated
Design & Analysis of Ecological Data. Landscape of Statistical Methods...
Design & Analysis of Ecological Data Landscape of Statistical Methods: Part 3 Topics: 1. Multivariate statistics 2. Finding groups - cluster analysis 3. Testing/describing group differences 4. Unconstratined
Grassmann Algebra in Game Development. Eric Lengyel, PhD Terathon Software
Grassmann Algebra in Game Development Eric Lengyel, PhD Terathon Software Math used in 3D programming Dot / cross products, scalar triple product Planes as 4D vectors Homogeneous coordinates Plücker coordinates
Chapter 1. Vector autoregressions. 1.1 VARs and the identi cation problem
Chapter Vector autoregressions We begin by taking a look at the data of macroeconomics. A way to summarize the dynamics of macroeconomic data is to make use of vector autoregressions. VAR models have become
Soft Clustering with Projections: PCA, ICA, and Laplacian
1 Soft Clustering with Projections: PCA, ICA, and Laplacian David Gleich and Leonid Zhukov Abstract In this paper we present a comparison of three projection methods that use the eigenvectors of a matrix
Cluster Analysis. Isabel M. Rodrigues. Lisboa, 2014. Instituto Superior Técnico
Instituto Superior Técnico Lisboa, 2014 Introduction: Cluster analysis What is? Finding groups of objects such that the objects in a group will be similar (or related) to one another and different from
Common factor analysis
Common factor analysis This is what people generally mean when they say "factor analysis" This family of techniques uses an estimate of common variance among the original variables to generate the factor
by the matrix A results in a vector which is a reflection of the given
Eigenvalues & Eigenvectors Example Suppose Then So, geometrically, multiplying a vector in by the matrix A results in a vector which is a reflection of the given vector about the y-axis We observe that
LINEAR ALGEBRA W W L CHEN
LINEAR ALGEBRA W W L CHEN c W W L Chen, 1997, 2008 This chapter is available free to all individuals, on understanding that it is not to be used for financial gain, and may be downloaded and/or photocopied,
4. Matrix Methods for Analysis of Structure in Data Sets:
ATM 552 Notes: Matrix Methods: EOF, SVD, ETC. D.L.Hartmann Page 64 4. Matrix Methods for Analysis of Structure in Data Sets: Empirical Orthogonal Functions, Principal Component Analysis, Singular Value
Unsupervised and supervised dimension reduction: Algorithms and connections
Unsupervised and supervised dimension reduction: Algorithms and connections Jieping Ye Department of Computer Science and Engineering Evolutionary Functional Genomics Center The Biodesign Institute Arizona
Mean value theorem, Taylors Theorem, Maxima and Minima.
MA 001 Preparatory Mathematics I. Complex numbers as ordered pairs. Argand s diagram. Triangle inequality. De Moivre s Theorem. Algebra: Quadratic equations and express-ions. Permutations and Combinations.
Introduction to Principal Components and FactorAnalysis
Introduction to Principal Components and FactorAnalysis Multivariate Analysis often starts out with data involving a substantial number of correlated variables. Principal Component Analysis (PCA) is a
Multivariate Analysis of Variance (MANOVA): I. Theory
Gregory Carey, 1998 MANOVA: I - 1 Multivariate Analysis of Variance (MANOVA): I. Theory Introduction The purpose of a t test is to assess the likelihood that the means for two groups are sampled from the
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS 1. SYSTEMS OF EQUATIONS AND MATRICES 1.1. Representation of a linear system. The general system of m equations in n unknowns can be written a 11 x 1 + a 12 x 2 +
Principal components analysis
CS229 Lecture notes Andrew Ng Part XI Principal components analysis In our discussion of factor analysis, we gave a way to model data x R n as approximately lying in some k-dimension subspace, where k
Communication on the Grassmann Manifold: A Geometric Approach to the Noncoherent Multiple-Antenna Channel
IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 48, NO. 2, FEBRUARY 2002 359 Communication on the Grassmann Manifold: A Geometric Approach to the Noncoherent Multiple-Antenna Channel Lizhong Zheng, Student
Multivariate Normal Distribution
Multivariate Normal Distribution Lecture 4 July 21, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2 Lecture #4-7/21/2011 Slide 1 of 41 Last Time Matrices and vectors Eigenvalues
