Torgerson s Classical MDS derivation: 1: Determining Coordinates from Euclidean Distances



Similar documents
13 MATH FACTS a = The elements of a vector have a graphical interpretation, which is particularly easy to see in two or three dimensions.

1 Introduction to Matrices

9.4. The Scalar Product. Introduction. Prerequisites. Learning Style. Learning Outcomes

Figure 1.1 Vector A and Vector F

Physics 235 Chapter 1. Chapter 1 Matrices, Vectors, and Vector Calculus

Linear Algebra Notes for Marsden and Tromba Vector Calculus

Data Mining: Algorithms and Applications Matrix Math Review

A Primer on Index Notation

SOLVING LINEAR SYSTEMS

Section V.3: Dot Product

Lecture 2 Matrix Operations

1.3. DOT PRODUCT If θ is the angle (between 0 and π) between two non-zero vectors u and v,

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 2. x n. a 11 a 12 a 1n b 1 a 21 a 22 a 2n b 2 a 31 a 32 a 3n b 3. a m1 a m2 a mn b m

by the matrix A results in a vector which is a reflection of the given

Introduction to Matrix Algebra

Vector Math Computer Graphics Scott D. Anderson

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS

Systems of Linear Equations

Linear Algebra Review. Vectors

Vectors Math 122 Calculus III D Joyce, Fall 2012

Orthogonal Diagonalization of Symmetric Matrices

Abstract: We describe the beautiful LU factorization of a square matrix (or how to write Gaussian elimination in terms of matrix multiplication).

Adding vectors We can do arithmetic with vectors. We ll start with vector addition and related operations. Suppose you have two vectors

Section 9.1 Vectors in Two Dimensions

5.3 The Cross Product in R 3

Multidimensional data and factorial methods

Linear Algebra: Vectors

Matrix Differentiation

Similar matrices and Jordan form

Direct Methods for Solving Linear Systems. Matrix Factorization

Inner Product Spaces

17. Inner product spaces Definition Let V be a real vector space. An inner product on V is a function

3. INNER PRODUCT SPACES

Vector Spaces; the Space R n

Biggar High School Mathematics Department. National 5 Learning Intentions & Success Criteria: Assessing My Progress

MATH 423 Linear Algebra II Lecture 38: Generalized eigenvectors. Jordan canonical form (continued).

Lecture 2: Homogeneous Coordinates, Lines and Conics

Chapter 19. General Matrices. An n m matrix is an array. a 11 a 12 a 1m a 21 a 22 a 2m A = a n1 a n2 a nm. The matrix A has n row vectors

1 Determinants and the Solvability of Linear Systems

Introduction to Matrices for Engineers

Math 215 HW #6 Solutions

Rotation Matrices and Homogeneous Transformations

Least-Squares Intersection of Lines

Inner product. Definition of inner product

Mathematics Course 111: Algebra I Part IV: Vector Spaces

CHAPTER 8 FACTOR EXTRACTION BY MATRIX FACTORING TECHNIQUES. From Exploratory Factor Analysis Ledyard R Tucker and Robert C.

Lecture Notes 2: Matrices as Systems of Linear Equations

The Matrix Elements of a 3 3 Orthogonal Matrix Revisited

A linear combination is a sum of scalars times quantities. Such expressions arise quite frequently and have the form

A vector is a directed line segment used to represent a vector quantity.

Vector and Matrix Norms

Nonlinear Programming Methods.S2 Quadratic Programming

Introduction Assignment

DATA ANALYSIS II. Matrix Algorithms

3.1. Solving linear equations. Introduction. Prerequisites. Learning Outcomes. Learning Style

α = u v. In other words, Orthogonal Projection

Notes on Orthogonal and Symmetric Matrices MENU, Winter 2013

6. Vectors Scott Surgent (surgent@asu.edu)

Question 2: How do you solve a matrix equation using the matrix inverse?

6. Cholesky factorization

Lecture 5: Singular Value Decomposition SVD (1)

Mathematics Pre-Test Sample Questions A. { 11, 7} B. { 7,0,7} C. { 7, 7} D. { 11, 11}

Linearly Independent Sets and Linearly Dependent Sets

Mehtap Ergüven Abstract of Ph.D. Dissertation for the degree of PhD of Engineering in Informatics

F Matrix Calculus F 1

MAC Learning Objectives. Module 10. Polar Form of Complex Numbers. There are two major topics in this module:

5. Orthogonal matrices

7 Gaussian Elimination and LU Factorization

[1] Diagonal factorization

Logistic Regression. Jia Li. Department of Statistics The Pennsylvania State University. Logistic Regression

The Method of Partial Fractions Math 121 Calculus II Spring 2015

Continued Fractions and the Euclidean Algorithm

Solving Systems of Linear Equations

Using row reduction to calculate the inverse and the determinant of a square matrix

Section Inner Products and Norms

Elementary Matrices and The LU Factorization

Vectors. Objectives. Assessment. Assessment. Equations. Physics terms 5/15/14. State the definition and give examples of vector and scalar variables.

Lecture L3 - Vectors, Matrices and Coordinate Transformations

ECE 842 Report Implementation of Elliptic Curve Cryptography

NOTES ON LINEAR TRANSFORMATIONS

The Euclidean Algorithm

Review Jeopardy. Blue vs. Orange. Review Jeopardy

1 Lecture: Integration of rational functions by decomposition

Vectors VECTOR PRODUCT. Graham S McDonald. A Tutorial Module for learning about the vector product of two vectors. Table of contents Begin Tutorial

MATH 304 Linear Algebra Lecture 18: Rank and nullity of a matrix.

Solutions to Math 51 First Exam January 29, 2015

Vectors 2. The METRIC Project, Imperial College. Imperial College of Science Technology and Medicine, 1996.

Eigenvalues, Eigenvectors, Matrix Factoring, and Principal Components

Numerical Analysis Lecture Notes

v w is orthogonal to both v and w. the three vectors v, w and v w form a right-handed set of vectors.

Metric Multidimensional Scaling (MDS): Analyzing Distance Matrices

Unit 18 Determinants

Solving Systems of Linear Equations

Solution of Linear Systems

THREE DIMENSIONAL GEOMETRY

3. KINEMATICS IN TWO DIMENSIONS; VECTORS.

Similarity and Diagonalization. Similar Matrices

Structure of the Root Spaces for Simple Lie Algebras

TWO-DIMENSIONAL TRANSFORMATION

Geometry of Vectors. 1 Cartesian Coordinates. Carlo Tomasi

Transcription:

Torgerson s Classical MDS derivation: 1: Determining Coordinates from Euclidean Distances It is possible to construct a matrix X of Cartesian coordinates of points in Euclidean space when we know the Euclidean distances D amongst those points. To do this we take the following two steps: 1.Use the cosine law to convert D to a matrix B of scalar products..perform a singular value decomposition of B to obtain X. Cosine Law: For a triangle between points i, j and k in Euclidean space with sides d ij, d ik, and d jk, and angle θ jik between sides d ij and d ik : Cos θ jik = (d ij + d ik - d jk )/d ij d ik If we define b jik = (d ij + d ik - d jk )/ then b jik = d ij d ik Cos θ jik The element b jik is the scalar product (for points j and k relative to point i) of the vectors from point i to j, the vector from point i to k, and of Cos θ jik, the cosine of the angle between the two vectors. Singular Value Decomposition: The scalar products matrix B i formed from the elements b jik is a symmetric matrix which may be decomposed into singular values V (on diagonal) and singular vectors (column-wise) U such that B i = UVU. Furthermore, B i is positive semidefinite, so, the rank of B i (the number of positive eigen values) equals the dimensionality of D, and we may calculate X=UV 1/. Then B i = XX.

: Determining Coordinates from Dissimilarities If we have a matrix D of Euclidean distances amongst n points, we can use the two steps on the previous page to calculate the matrix X of the Cartesian coordinates of those points. If we have a matrix O of observed dissimilarities which approximate Euclidean distances we can use two similar steps to calculate X. 1.Calculate B: The first step is to convert O into B, a matrix of pseudo scalar products, by double centering O. Note: The superscript on O means we double center the matrix of squared dissimilarities. Double centering is subtracting the row and column means of the matrix from its elements, adding the grand mean and multiplying by -1/. If O are error-free Euclidean distances, then B calculated this way are error-free scalar products relative to the origin..singular Value Decomposition: We double center O to obtain B because (see the next page for Torgerson s derivation) we can then determine the coordinates X by the singular value decomposition B = UVU which then permits us to calculate X=UV 1/. Least Squares Properties of X: B is an indefinite matrix (it may have negative as well as zero or positive roots). It can be shown that if we define V r by selecting the r largest singular values, and the corresponding singular vectors U r and define: X r = U r V r 1/ then B r = X r X r is a least squares approximation to B.

Torgerson s Double Centering Derivation We begin with B i = XX, the scalar products given earlier, where B i is referred to an origin at point i. We translate the axes to the centroid of all points. This gives X *, a new set of Cartesian coordinates in the new system whose origin is at the centroid of the points. This translation is done by defining c = (1 X)/n a 1xr row vector of column means of X, where 1 is an nx1 column vector of n one s. We now define the centered X * by subtracting the column means from elements of X by the equation X * = X - 1c = X - (11 X)/n. We can now define the scalar products of the centered configuration as B * = X * X * = (X - 1c)(X - 1c) = (X - 1c)(X - c 1 ) = (X - 11 X/n)(X - X 11 /n) = (XX - XX 11 /n - 11 XX /n + 11 XX 11 /n ) = (B i - B i 11 /n - 11 B i /n + 11 B i 11 /n ) = (B i - B i. - B i. + B i..) where B i. is a matrix of (row or column) means of B i and where B i.. contains the grand mean of all elements of B i.

Torgerson s Double Centering Derivation (continued) The scalar version of the previous matrix equation allows us to see that an individual element of B * is defined as: n b 1 ij b ij -- b. n 1 n n n -- ik n b 1 = kj + ----- k k n b gh g h Noting that b jk = (d ij + d ik - d jk )/, substituting and simplifying gives: b ij n 1 1 -- d ij -- d n 1 n n n -- ik n d 1 = kj + ----- k k n d gh g h That is, the scalar products of the centered coordinates are -1/ times the squared distances with the row and column means removed and the grand mean added back in. As before, the singular value decomposition B * = U * V * U * allows us to define: X * =U * V *1/..

3: Torgerson s Scalar Product Algorithm for Classical (Metric Unweighted) MDS 1.Let o ij be the observed dissimilarity between stimuli i and j. d ij be the Euclidean distance between stimuli i and j. x i be the vector of coordinates of stimulus i on all dimensions. x. = 0 denote the origin. denote the distance from the origin to point i. d i..assume that the data are error-free Euclidean distances: =. 3.Note: o ij = d ij = ( x i x j )( x i x j )' = x i x i ' x i x j ' + x j x j ', and: = = x i x i ' x i x. ' + x. x. ' = x i x i '. 4.If we define then o i. d i. 1 b ij = -- ( o ij o i. o. j + o.. ) 1 b ij = -- [( x i x i ' x i x j ' + x j x j ' ) x i x i ' x j x j '] = x i x j. Thus we can solve for coordinates using the equality B = XX. 5.The algorithm is 1) Double center the data matrix O to obtain B (i.e., remove row and column means and add back in the grand mean of the squared data). ) Solve the singular value decomposition problem B = ULU. 3) Calculate X = UL 1. 6.This minimizes B XX., o ij d ij