Visualization of General Defined Space Data



Similar documents
Section 1.1. Introduction to R n

Solving Simultaneous Equations and Matrices

Data Mining: Algorithms and Applications Matrix Math Review

3. INNER PRODUCT SPACES

11.1. Objectives. Component Form of a Vector. Component Form of a Vector. Component Form of a Vector. Vectors and the Geometry of Space

1 VECTOR SPACES AND SUBSPACES

Linear Algebra Notes for Marsden and Tromba Vector Calculus

Linear Algebra Review. Vectors

Inner Product Spaces and Orthogonality

December 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B. KITCHENS

Manifold Learning Examples PCA, LLE and ISOMAP

Math 215 HW #6 Solutions

What are the place values to the left of the decimal point and their associated powers of ten?

State of Stress at Point

Nonlinear Iterative Partial Least Squares Method

1 Introduction to Matrices

Notes on the representational possibilities of projective quadrics in four dimensions

KEANSBURG SCHOOL DISTRICT KEANSBURG HIGH SCHOOL Mathematics Department. HSPA 10 Curriculum. September 2007

Mathematics Course 111: Algebra I Part IV: Vector Spaces

Dimensionality Reduction: Principal Components Analysis

Algebra 2 Chapter 1 Vocabulary. identity - A statement that equates two equivalent expressions.

Common Core Unit Summary Grades 6 to 8

Factoring Patterns in the Gaussian Plane

Algebra Unpacked Content For the new Common Core standards that will be effective in all North Carolina schools in the school year.

Prentice Hall Algebra Correlated to: Colorado P-12 Academic Standards for High School Mathematics, Adopted 12/2009

Linear Algebra: Vectors

Metrics on SO(3) and Inverse Kinematics

Direct Methods for Solving Linear Systems. Matrix Factorization

THREE DIMENSIONAL REPRESENTATION OF AMINO ACID CHARAC- TERISTICS

Thnkwell s Homeschool Precalculus Course Lesson Plan: 36 weeks

Geometry of Vectors. 1 Cartesian Coordinates. Carlo Tomasi

by the matrix A results in a vector which is a reflection of the given

x1 x 2 x 3 y 1 y 2 y 3 x 1 y 2 x 2 y 1 0.

Finite Element Formulation for Plates - Handout 3 -

Linear algebra and the geometry of quadratic equations. Similarity transformations and orthogonal matrices

Linear Programming for Optimization. Mark A. Schulze, Ph.D. Perceptive Scientific Instruments, Inc.

Bedford, Fowler: Statics. Chapter 4: System of Forces and Moments, Examples via TK Solver

THREE DIMENSIONAL GEOMETRY

Geometry and Measurement

Solving Systems of Linear Equations

MATH 304 Linear Algebra Lecture 9: Subspaces of vector spaces (continued). Span. Spanning set.

BookTOC.txt. 1. Functions, Graphs, and Models. Algebra Toolbox. Sets. The Real Numbers. Inequalities and Intervals on the Real Number Line

Georgia Standards of Excellence Curriculum Map. Mathematics. GSE 8 th Grade

Pennsylvania System of School Assessment

Vector Spaces; the Space R n

Current Standard: Mathematical Concepts and Applications Shape, Space, and Measurement- Primary

TWO-DIMENSIONAL TRANSFORMATION

α = u v. In other words, Orthogonal Projection

GRADES 7, 8, AND 9 BIG IDEAS

AP Physics - Vector Algrebra Tutorial

Year 9 set 1 Mathematics notes, to accompany the 9H book.

Lecture 1: Systems of Linear Equations

MATH 423 Linear Algebra II Lecture 38: Generalized eigenvectors. Jordan canonical form (continued).

with functions, expressions and equations which follow in units 3 and 4.

Numerical Analysis Lecture Notes

Definition 8.1 Two inequalities are equivalent if they have the same solution set. Add or Subtract the same value on both sides of the inequality.

ON TORI TRIANGULATIONS ASSOCIATED WITH TWO-DIMENSIONAL CONTINUED FRACTIONS OF CUBIC IRRATIONALITIES.

Metric Spaces. Chapter Metrics

Chapter 19. General Matrices. An n m matrix is an array. a 11 a 12 a 1m a 21 a 22 a 2m A = a n1 a n2 a nm. The matrix A has n row vectors

Precalculus REVERSE CORRELATION. Content Expectations for. Precalculus. Michigan CONTENT EXPECTATIONS FOR PRECALCULUS CHAPTER/LESSON TITLES

17. Inner product spaces Definition Let V be a real vector space. An inner product on V is a function

Vectors 2. The METRIC Project, Imperial College. Imperial College of Science Technology and Medicine, 1996.

Florida Math for College Readiness

Prentice Hall Mathematics: Algebra Correlated to: Utah Core Curriculum for Math, Intermediate Algebra (Secondary)

Topological Data Analysis Applications to Computer Vision

1 Sets and Set Notation.

Figure 1.1 Vector A and Vector F

Computer Graphics CS 543 Lecture 12 (Part 1) Curves. Prof Emmanuel Agu. Computer Science Dept. Worcester Polytechnic Institute (WPI)

Visualization by Linear Projections as Information Retrieval

Adding vectors We can do arithmetic with vectors. We ll start with vector addition and related operations. Suppose you have two vectors

Chapter 6. Orthogonality

Algebra Academic Content Standards Grade Eight and Grade Nine Ohio. Grade Eight. Number, Number Sense and Operations Standard

Recall that two vectors in are perpendicular or orthogonal provided that their dot

Course Syllabus For Operations Management. Management Information Systems

NEW YORK STATE TEACHER CERTIFICATION EXAMINATIONS

WORK SCHEDULE: MATHEMATICS 2007

Orthogonal Diagonalization of Symmetric Matrices

Algebra I Credit Recovery

DELAWARE MATHEMATICS CONTENT STANDARDS GRADES PAGE(S) WHERE TAUGHT (If submission is not a book, cite appropriate location(s))

Big Ideas in Mathematics

Introduction to Matrix Algebra

Visualization of Large Font Databases

Virtual Landmarks for the Internet

CS3220 Lecture Notes: QR factorization and orthogonal transformations

Estimated Pre Calculus Pacing Timeline

Vector Algebra II: Scalar and Vector Products

Linear Algebra I. Ronald van Luijk, 2012

LINEAR ALGEBRA. September 23, 2010

Math Review. for the Quantitative Reasoning Measure of the GRE revised General Test

4.3 Least Squares Approximations

Such As Statements, Kindergarten Grade 8

Solving Systems of Linear Equations

Investigating Area Under a Curve

Prerequisites: TSI Math Complete and high school Algebra II and geometry or MATH 0303.

Lectures notes on orthogonal matrices (with exercises) Linear Algebra II - Spring 2004 by D. Klain

IRREDUCIBLE OPERATOR SEMIGROUPS SUCH THAT AB AND BA ARE PROPORTIONAL. 1. Introduction

Math 312 Homework 1 Solutions

Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches

October 3rd, Linear Algebra & Properties of the Covariance Matrix

Essential Mathematics for Computer Graphics fast

Transcription:

International Journal of Computer Graphics & Animation (IJCGA) Vol.3, No.4, October 013 Visualization of General Defined Space Data John R Rankin La Trobe University, Australia Abstract A new algorithm is presented which determines the dimensionality and signature of a measured space. The algorithm generalizes the Map Maker s algorithm from D to n dimensions and works the same for D measured spaces as the Map Maker s algorithm but with better efficiency. The difficulty of generalizing the geometric approach of the Map Maker s algorithm from D to 3D and then to higher dimensions is avoided by using this new approach. The new algorithm preserves all distances of the distance matrix and also leads to a method for building the curved space as a subset of the N-1 dimensional embedding space. This algorithm has direct application to Scientific Visualization for data viewing and searching based on Computational Geometry. Categories and Subject Descriptors: I.3.3 Viewing algorithms, I.3.5 Computational Geometry and Object Modelling 1. Introduction The Map Maker s problem is to convert a measured space into a two-dimensional coordinatized space. This is well achieved by the Map Maker s algorithm [Ran13]. However the assumption in that process is that the distance matrix of the measured space corresponds at least roughly to a flat two-dimensional space. When this is true the Map Maker s algorithm is distance preserving. In general however this would not be the case and then the Map Maker s algorithm does not preserve all distances. A measured space is defined by a set of sites i = 1 to N and their measured separations d ij where d ij is the shortest distance between site i and site j. The set of NxN distances d ij is called the Distance Matrix. This matrix is symmetric and trace-free with zeros down the principal diagonal. Thus a set of known sites and their distance matrix constitutes what is called a defined (or measured) space. Research based on distance matrices has had wide applications such as in signal processing [PC9], pattern recognition [ZWZ08] and medical imaging [PE*09]. One of the foundational research efforts [LSB77] resulted in Lee s algorithm for projecting n-dimensional distance matrix data to two dimensions for visualization. Lee s algorithm does not determine the n-dimensional coordinates of the points involved but goes straight to a projection of those points in D. In the process of this projection most of the distance information is not preserved in the visualization. Other approaches such as LLE preserve distances just in the local neighbourhood of a chosen point [RS00, TSL00]. In the research presented here the coordinatization and projection processes are separated one is free to apply any desired projection transformation to the coordinatized data. Surprisingly, from the distance matrix d ij one can determine the dimensionality and signature of the space that it defines as well as the coordinates for the sites and the algorithm presented in this paper demonstrates this. DOI : 10.511/ijcga.013.340 15

International Journal of Computer Graphics & Animation (IJCGA) Vol.3, No.4, October 013 The algorithm presented here, called the Coordinatizator algorithm, does not require any special ordering of the sites before processing. The given ordering of the sites however determines the directions and orientations of all axes in the building of the n-dimensional space. Re-orderings of the sites therefore may be done for other reasons such as desired axial directions or reduction in numerical errors. The Coordinatizator algorithm coordinatizes by considering the sequence of simplices S k = P 1 P P k with vertices P i for i = 1 to k and setting up orthogonal axes for these simplices.. Coordinatizator Process The Coordinatizator algorithm uses the given sites to construct axes and produce coordinates for each of the sites. As in the Map Maker s algorithm, site 1 is taken as the origin so it is coordinatized as the point: P 1 = (0, 0, 0, 0...) The direction from site 1 to site is taken as the x-axis or the first axis so therefore site is coordinatized as the point: P = (x 1, 0, 0, 0...) and obviously x 1 = d 1. The triangle of site 1 to site to site 3 defines the x-y plane and thus the -axis (i.e. the y-axis which is perpendicular to the 1-axis which is the x-axis). Therefore P 3 is coordinatized as the point: P 3 = (x 31, x 3, 0, 0, 0...) The tetrahedron formed by the first 4 sites defines the third axial direction perpendicular to the first two. Thus site 4 is coordinatized as the point: P 4 = (x 41, x 4, x 43, 0, 0, 0...) and so forth in general: P i = (x i1, x i, x i3,, x ii-1,0, 0, 0, 0...) for i = 1 to N where N is the number of sites and the size of the Distance Matrix. This process of converting the Distance Matrix into a set of N points in n-dimensional space generally creates a non-euclidean space of n = N-1 dimensions. We add the points progressively by building up a simplex from 0 dimensions to n = N-1 dimensions. However it is possible that at some step the added point is degenerate and the dimension does not increase for that step in general the dimensionality of the measured space is bounded above by N-1, n N-1. It is interesting that every time another point say the k+1th point is added, we only have to consider k more edges since d ij = d ji. The Coordinatizator algorithm solves the following mathematical equations for the x ij given the d ij. Investigation into these equations found that the algebraic solution was possible for the general n-dimensional case. 16

International Journal of Computer Graphics & Animation (IJCGA) Vol.3, No.4, October 013 Given: P 1 = (0, 0, 0, 0...) P = (x 1, 0, 0, 0...) P 3 = (x 31, x 3, 0, 0, 0...) P 4 = (x 41, x 4, x 43, 0, 0, 0...) P 5 = (x 51, x 5, x 53, x 54, 0, 0, 0...) ::::: P i = (x i1, x i, x i3,, x ii-1,0, 0, 0, 0...) The unknowns x ij in the above equations can be solved for k = 1 to N by the following procedure. Every coordinatization is extended to higher dimensions by appending the additional required number of zeros in the n-tuple. For k = 1 we consider only point P 1 and coordinatize it by assign it to the 0 dimensional origin. For k = we bring in point P and get: d 1 = (P - P 1 )(P - P 1 ) = x 1 and without loss of generality we choose: x 1 = d 1 and thereby coordinatize P as the 1 dimensional point (x 1 ). For k = 3 we bring in point P 3 and coordinatize it by getting: d 13 = (P 3 - P 1 )(P 3 - P 1 ) = x 31 + x 3 d 3 = (P 3 - P )(P 3 - P ) = (x 31 -x 1 ) + x 3 Subtracting gives: d 13 - d 3 = x 1 x 31 - x 1 x 31 = (d 13 - d 3 + x 1 )/(x 1 ) and therefore Now if the term in the square root is negative then we have non-euclidean flat space. We will consider this type of situation later on. Here we will take the positive root without loss of generality since this will be the first point with a y component. For k = 4 we are adding in P 4 and we have these three equations to consider: d 14 = (P 4 - P 1 )(P 4 - P 1 ) = x 41 + x 4 + x 43 d 4 = (P 4 - P )(P 4 - P ) = (x 41 -x 1 ) + x 4 + x 43 d 34 = (P 4 - P 3 )(P 4 - P 3 ) = (x 41 -x 31 ) + (x 4 -x 3 ) + x 43 Therefore: d 14 - d 4 = x 1 x 41 - x 1 x 41 = (d 14 - d 4 + x 1 )/(x 1 ) Secondly d 4 - d 34 = (x 41 -x 1 ) - (x 41 -x 31 ) + x 3 x 4 - x 3 17

International Journal of Computer Graphics & Animation (IJCGA) Vol.3, No.4, October 013 x 4 = (d 4 - d 34 - (x 41 -x 1 ) + (x 41 -x 31 ) + x 3 )/(x 3 ) Finally we get: Now if the term in the square root is negative then we have non-euclidean flat space. We will consider this type of situation later on. Here we will take the positive root without loss of generality since this will be the first point with a z component (i.e. 3rd or kth component). For k = 5 we are adding in P 5 and we have these four equations to consider: d 15 = (P 5 - P 1 )(P 5 - P 1 ) = x 51 + x 5 + x 53 + x 54 d 5 = (P 5 - P )(P 5 - P ) = (x 51 -x 1 ) + x 5 + x 53 + x 54 d 35 = (P 5 - P 3 )(P 5 - P 3 ) = (x 51 -x 31 ) + (x 5 -x 3 ) + x 53 + x 54 d 45 = (P 5 - P 4 )(P 5 - P 4 ) = (x 51 -x 41 ) + (x 5 -x 4 ) + (x 53 -x 43 ) + x 54 Therefore: d 15 - d 5 = x 1 x 51 - x 1 x 51 = (d 15 - d 5 + x 1 )/(x 1 ) Secondly d 5 - d 35 = (x 51 -x 1 ) - (x 51 -x 31 ) + x 3 x 5 - x 3 x 5 = (d 5 - d 35 - (x 51 -x 1 ) + (x 51 -x 31 ) + x 3 )/(x 3 ) Thirdly d 35 - d 45 = (x 51 -x 31 ) - (x 51 -x 41 ) + (x 5 -x 3 ) - (x 5 -x 4 ) + x 43 x 53 - x 43 x 53 = (d 35 - d 45 - (x 51 -x 31 ) + (x 51 -x 41 ) - (x 5 -x 3 ) + (x 5 -x 4 ) + x 43 )/(x 43 ) Finally we get: Now if the term in the square root is negative then we have non-euclidean flat space. We will consider this type of situation later on. Here we will take the positive root without loss of generality since this will be the first point with a kth component. This forms a clear pattern for any value of k >= 1 and hence the algorithm can be worked for any value of N. 3. Imaginery Components The quantities under the square root signs in the equations above could be positive, zero or negative giving rise to three important cases to consider. If the quantity is positive we have an extra Euclidean dimension for housing the points. If the quantity is zero then we have sufficient dimensions already and an extra dimension is not (yet) required for embedding the points considered so far. If the quantity is negative then the Euclidean space becomes pseudo-euclidean [Wik13a] like Minkowski space in Special Relativity where time is an imaginary extra axial dimension for three-dimensional space. As in Relativity, the invariant quadratic form then is no 18

International Journal of Computer Graphics & Animation (IJCGA) Vol.3, No.4, October 013 longer positive definite and is properly termed a pseudo-metric though it is common to continue to call it a metric anyway). Sylvester s Theorem [Wik13b] states that the matrix for any quadratic form can be diagonalized by a suitable change of coordinates and with rescalings the diagonal elements are only {-1,0,1} and furthermore the number of +1s (called the positive index of inertia n + ) and the number of 0s (n 0 )and the number of -1s (called the negative index of inertia n - ) is invariant no matter what recoordinatization is chosen. Clearly n + + n 0 + n - = N-1 The dimensionality of the measured space is given by n = n + + n - The signature of the measured space is given by s = n + - n - Knowing the dimensionality, the signature and N is equivalent to knowing n +, n 0 and n - which the Coordinatization algorithm provides. The Corrdinatizator algorithm automatically determines these quantities by noting the sign of the quantities under the square root. The presentation of the algorithm below explicitly indicates these quantities. 4. Coding the Algorithm We have an outer loop for k = 1 to N. In each loop we determine k constants which are the nonzero components of P k. Computing the next component depends on the values of all previously computed components. Essentially we are constructing a new output NxN matrix called x from a given input NxN matrix called d. The input matrix cannot have negative values, must be symmetric and its principal diagonal should consist only of zeros. The output matrix is lowertriangular it also has zeros on its principal diagonal. In both cases the matrix indices run from 1 to N. So the algorithm looks like this: Inputs: The NxN distance matrix d[i,j] Outputs: The NxN matrix of point components x[i,j]. Initialize all x[i,j] = 0 -: for i = 1 to N for j = 1 to N x[i,j] = 0 PositiveIndex = 0 NegativeIndex = 0 for k = 1 to N l = k-1 for i = 1 to l-1 sum = Sq(d[i,k]) - Sq(d[i+1,k]) for j = 1 to i-1 sum += - Sq(x[k,j] - x[i,j]) sum += + Sq(x[k,j] - x[i+1,j]) sum += Sq(x[i+1,i]) x[k,i] = sum//x[i+1,i] 19

International Journal of Computer Graphics & Animation (IJCGA) Vol.3, No.4, October 013 sum = Sq(d[l,k]) for i = 1 to l-1 sum = sum - Sq(x[l,i]-x[k,i]) if sum > 0 PositiveIndex++ x[l,k] = SqRoot(sum) if sum < 0 NegativeIndex++ x[l,k] = SqRoot(-sum) 5. Curvature, Dimensionality and Degeneracy We will start by assuming that the N sites define a space of dimensionality n = N-1. If for any k in the algorithm we find x k+1k is zero or close to it then we will reduce the dimensionality n of the data space by 1. The Coordinatizator algorithm does not search for curvature in the space nor the dimensionality of the embedding space, so the defined space be either a flat n-dimensional Euclidean space or a flat n-dimensional pseudo-euclidean space. When there is degeneracy one has to check whether the negative is the right solution i.e. it gives the right distances to the other points before it according to the distance matrix as described in the Map Maker s algorithm. 6. Conclusions The Coordinatizator algorithm presented in this paper maps sites to an n-dimensional pseudo- Euclidean space preserving all distances of the given distance matrix. The algorithm returns the positive index of inertia and the negative index of inertia of the pseudo-metric of the pseudo- Euclidean space. If the negative index is zero then the space is pure Euclidean and the triangle inequality [Wik13c] is valid for all triplets of points within it. If the negative index is non-zero then the triangle inequality is reversed but only for the subspace of points spanned by the imaginary axial directions. Once this algorithm is applied to defined space data projection mappings can be applied to view the n-dimensional data on a flat screen. The Coordinatizator algorithm effectively maps one NxN lower triangular matrix (d ij ) to another NxN lower triangular matrix (x ij ) and this would allow us to reduce memory needs by storing data in a single square NxN real matrix. References [1]. [Fer57] Ferrar W, Algebra A Textbook of Determinants, Matrices and Algebraic Forms, Oxford University Press, nd edition, 1957, p 154. []. [LSB77] Lee R, Slagle J & Blum H, A Triangulation Method for the Sequential Mapping of Points from N-Space to Two-Space, 1977, IEEE Transaction on Computers, pp 88-9. [3]. [PC9] Potter L & Chiang D, Distance Matrices and Modified Cyclic Projections for Molecular Conformation, Acoustics, Speech and Signal Processing 199, Vol 4, pp 173-176. [4]. [PE*09] Pham T, Eisenblatter U, Golledge J, Baune B & Berger K, Segmentation of Medical Images Using Geo-theoretic Distance Matrix in Fuzzy Clustering, Image Processing (ICIP) 009, pp 3369-337. 0

International Journal of Computer Graphics & Animation (IJCGA) Vol.3, No.4, October 013 [5]. [Ran13] Rankin J, Map Making From Tables, International Journal of Computer Graphics and Animation, vol 3, no, 013, pp 11-19 [6]. [RS00] Rowels S & Saul L, Nonlinear Dimensionality Reduction by Locally Linear Embedding, Science, Vol 90, December 000, pp. 3-33. [7]. [Sle10] Sledge I, A Projection Transform for non-euclidean Relational Clustering, Neural Networks (IJCNN) 010, pp 1-8. [8]. [TSL00] Tenembaum J, de Silva V & Langford J, A Global Geometric Framework for Nonlinear Diemsionality Reduction, Science, Vol 90, December 000, pp. 3-33. [9]. [TT*1] Tzagkarakis G, Tsagkatakis G, Starck J & Tsakalides P, Compressive Video Classification in a Low-dimensional Manifold with Learned Distance Matrix, 0 th European Signal Processing Conference [EUSIPCO 01], 01, pp 155-159. [10]. [Wik13a] PseudoEuclidean spaces Wikipedia, 011 [11]. http://en.wikipedia.org/wiki/pseudo-euclidean_space [1]. [Wik13b] Sylvester s Law of Inertia, Wikipedia, 013 [13]. http://en.wikipedia.org/wiki/sylvester's_law_of_inertia [14]. [Wik13c] Triangle Inequality, Wikepedia [15]. http://en.wikipedia.org/wiki/triangle_inequality [16]. [YR13] Youldash M & Rankin J, Evaluation of a Geometric Approach to the Coordinatization of Measured Spaces, 8 th International Conference on Information Technology and Applications (ICITA 013), Sydney, Australia July 013. [17]. [ZWZ08] Zuo W, Wang K & Zhang D, Assembled Matrix Distance Metric for DPCA-based Face and Palmprint Recognition, Machine Learning and Cybernetics 008, Vol 8, pp4870-4875. 1