Point Cloud Segmentation via Constrained Nonlinear Least Squares Surface Normal Estimates
|
|
|
- Randell Sullivan
- 10 years ago
- Views:
Transcription
1 Point Cloud Segmentation via Constrained Nonlinear Least Squares Surface Normal Estimates Edward Castillo Radiation Oncology Department University of Texas MD Anderson Cancer Center, Houston TX Hongkai Zhao Department of Mathematics University of California, Irvine CA zhao Abstract We present a point cloud segmentation scheme based on estimated surface normals and local point connectivity, that operates on unstructured point cloud data. We can segment a point cloud into disconnected components as well as piecewise smooth components as needed. Given that the performance of the segmentation routine depends on the quality of the surface normal approximation, we also propose an improved surface normal approximation method based on recasting the popular principal component analysis formulation as a constrained least squares problem. The new approach is robust to singularities in the data, such as corners and edges, and also incorporates data denoising in a manner similar to planar moving least squares. 1. Introduction Two key processing tools for unstructured point cloud data are surface normal estimation and segmentation. Surface normal approximation is utilized by several applications such as reconstruction, estimating local feature size, computer aided design (CAD), and inside-outside queries, while segmentation is useful for efficient point cloud representation. We present a point cloud segmentation scheme that utilizes estimated surface normals and operates on raw point cloud data. The procedure is based on constructing an adjacency matrix representing the point connectivity of the full data set from unoriented surface normal estimates. The connected components present in the adjacency matrix constitute the initial segmentation, while the point connectivity information provides a means for orientating the surface normal estimates consistently, even for data sets containing one or multiple closed surfaces. A finer segmentation can then be derived from the initial connected component segmentation by classifying the oriented estimated normals according to their direction or their singularity and the local geometry information contained in the adjacency matrix. Naturally, the performance of the segmentation routine depends on the quality of the surface normal approximation. As such, we also propose an improved surface normal approximation method based on recasting the popular principal component analysis formulation as a constrained least squares problem. The new approach is robust to singularities in the data, such as sudden change in normals or edges, and also incorporates data denoising in a manner similar to planar moving least squares. 2. Least Squares Normal Estimation and PCA Principal component analysis (PCA) is a popular method for computing surface normal approximations from point cloud data [3]. Given a point cloud data set D = {x i } n i=1, the PCA surface normal approximation for a given data point p D is typically computed by first detering the K-nearest neighbors, x k D, of p. Given the K-neighbors, the approximate surface normal is then the eigenvector associated with the smallest eigenvalue of the symmetric positive semi-definite matrix P = K (x k p) T (x k p), (1) j=1 where p is the local data centroid, p = ( 1 K ) K j=1 x j. The PCA normal approximation, also referred to as total least squares [5], is accurate when the underlying surface 1
2 is smooth, but tends to smear across singularities, such as corners or intersecting planes. Figure 1 illustrates this phenomenon for a cube. The smearing is caused by the con- Figure 1. PCA normal estimation for a cube. Notice the smeared normals along the edges of the cube. tribution of data points from across the singularity to the PCA covariance matrix P. Simply reducing the weight of such contributions will therefore improve the PCA estimate. However, equation (1) alone does not provide much insight into detering which data point contributions should be dampened, and which should be considered valid. Fortunately, the PCA normal approximation can also be described as the solution to the following constrained least squares problem: η 1 2 V η 2 2 (2) s.t. η = 1, where the rows of the matrix V IR K 3 are the difference vectors v k = x k p. The equality constraint is necessary to ensure a nontrivial solution. After substituting η T η = 1 for the original equality constraint, the first order KKT conditions for problem 2 describe a standard eigenvalue problem: V T V η µη = 0, (3) Hence, the stationary points for problem 2 are the eigenpairs of P = V T V, where the matrix P is exactly the PCA covariance matrix in 1. The global imum is achieved at the smallest eigenvalue, and thus the solution to problem 2 is equivalent to the PCA approximation, contrary to the assertion in [2]. From the perspective of moving least squares, the PCA approximation is the normal associated with the plane, passing through the centroid, that best fits the local data in the least squares sense []. The advantage of formulation 2 over the standard PCA formulation is that the optimization problem gives insight into how to improve normal estimation near singularities, and also in the presence of noise. The objective function describes variance from orthogonality. In essence, problem 2, and consequently the PCA method, detere the normal approximation as the vector which is most orthogonal to each difference vector v k, by imizing the sum of squares of the inner products. Thus, in the simplest case where all v k live in the same plane, the value of the objective function V T η at the solution η is zero, and η is orthogonal to the plane. From this perspective, it is not necessary to compute the local centroid, since utilizing the difference vectors a k = x k p, as opposed to the vectors v k = x k p, yields the same result. Furthermore, the magnitude of the inner product between η and each difference vector quantifies the planar deviation of each local neighbor. Monitoring this quantity thus provides a means for detering which data points, if any, are contributing from across a singularity. 3. Normal Estimation via Nonlinear Least Squares The solution to problem (2) may be suspect near surface edges, due to erroneous contributions from data points across the singularity. Erroneous data contributions can be detected by exaing the magnitude of the orthogonality mismatch, a T k η, where a k = x k p. Assug noise free data, a large orthogonality mismatches corresponds to neighbors that deviate from the local least squares fitted plane. In order to reduce the contribution of points from across the singularity, we propose modifying formulation (2) by incorporating automatic weights, adaptive to orthogonality mismatch, to deflate such contributions. The result is the following constrained nonlinear least squares formulation: [ ] 2 1 K η 2 k=1 e λ(at k η)2 (a T k η) (4) s.t. η = 1. Traditional weighting terms place emphasis on proximity, and also serve as a window function which isolates the local neighbors. The weighting term e λ(at k η)2 adaptively deflates the contribution of terms with high orthogonality mismatch at a rate defined by the parameter λ. Naturally, setting λ to zero results in the original PCA linear least squares problem 2, with the exception that the difference vectors are taken from the data point p and not the centroid p. At first glance, formulation 4 appears to be significantly more difficult to solve than the small eigenvalue problem associated with the PCA approach. However, the equality constraint is easily absorbed into the objective function by representing the unknown surface normal η in spherical coordinates with magnitude set to unity: η(φ, θ) = cos(φ) cos(θ) cos(φ) sin(θ) sin(φ).
3 In addition, the spherical coordinate representation reduces the number of unknowns in the formulation to two. Thus, the simplified, unconstrained formulation is now: φ,θ 1 2 K k=1 [ e λ(at k η(φ,θ))2 (a T k η(φ, θ))] 2. (5) Applying formulation 4 to the box data results in significant improvement in the surface normal estimates, as illustrated by figure 2. Rearranging terms and taking advantage of the fact that η = 1, the problem reduces to a simple least squares problem for the scalar value t: t Aη t 2, (6) where the rows of A are the difference vectors a k = x k p, and the denoised location is given by the solution t as p = p + (t )η. As is the case for most denoising procedures, naively applying 6 to clean data will result in over smoothing and the loss of sharp surface features. However, given a bound T for the distance between p and the underlying surface, data fidelity can easily be incorporated into the formulation by way of linear inequality constraints. Specifically, t Aη t 2, (7) s.t. t < T. Figure 2. Nonlinear least squares normal estimation for a cube. Notice the improvement in the normal estimation along the edges of the cube compared to standard PCA (figure 1). 4. Incorporating Noise Robustness into Normal Estimation Noise in point cloud data is not uncommon and is a significant source of error for surface normal approximation. In addition, point cloud denoising is helpful to other applications, such as surface reconstruction. For a given noise polluted point p, the task of denoising can be viewed as finding the projection of the point p onto the unknown surface, a process often referred to as point cloud thinning. Developing methods for projecting points onto curves and surfaces is an active area of research with applications in computer graphics and vision (see [4]). Work has also been done for the problem of projecting points onto point cloud represented surfaces [1, 6], based on first detering a projection direction. The goal here is to incorporate point cloud denoising into the surface normal computation, an approach which can be thought off as point projection utilizing the estimated surface normal as the projection direction. Given an estimate η for the surface normal at p, we wish to detere the position along η, emanating from p, that imizes the orthogonality mismatch energy (formulation 5 with λ = 0): t K (x k ( p + tη)) T η. k=1 Single variable calculus dictates that the solution to problem 7 is achieved at either the unconstrained imum t unc = ( 1 K ) K i=1 at i η, or at one of the bounds T, T. For the typical case where η is not known, the noise in the data can lead to inaccurate PCA or NLSQ surface normal approximations. Noise robustness can be achieved within our least squares framework by coupling the nonlinear least squares formulation 5 with the denoising formulation 7: [ ] 2 1 K 2 k=1 e λ(at k η(φ,θ) t)2 (a T k η(φ, θ) t) (8) φ,θ,t s.t. t <= T. Figure 3. Nonlinear least squares normal estimation with noise robustness for a noisy cube. three percent random noise was added to the original cube. The estimated normals are still accurate. Given the simplicity of the objective function and linear constraints, problem 8 can be solved via any standard nonlinear least squares solver package, such as MINPACK or MATLAB s built in lsqnonlin routine. Our current implementation includes the two linear constraints into a global
4 formulation via log barrier functions: L 1 (t) = ln( h t h ) L 2 (t) = ln( h + t h ) Thus, the nonlinear least squares problem contains K + 2 equations and three unknowns. Problem 8 coupled with the log-barrier functions, comprise a standard, unconstrained, nonlinear least squares problem for which several high performance solver packages exist. Our current implementation utilizes the MINPACK software package, which is based on the Levenberg-Marquardt Quasi-Newton trust region algorithm, using exact values for the gradient of the objective function. The initial guess used for the optimization routine is computed by solving problems 5 and 7 for λ, h = 0, respectively. 5. Segmentation Based on Point Connectivity The overall segmentation strategy is two-fold. First, an initial segmentation is obtained from an adjacency matrix, constructed from unoriented surface normal estimates, describing the point connectivity for the entire data set. The final, finer segmentation is detered from the initial segmentation, the point connectivity information contained in the adjacency matrix, and the oriented surface normals estimates. Our strategy for the initial segmentation is built upon the assumption that one can obtain a reasonable approximation to the underlying unoriented surface normals (presumably by applying the methodology described in the previous section, but this is not necessary) at each point in the cloud. Based on the approximated normals and the k- nearest neighbor information, a symmetric adjacency matrix A is built such that A i,j = 1 if the data points i, j are connected, or in terms of graph theory, there exists an edge between the two points. A i,j = 0 otherwise. The initial segmentation then is given by the connected components of the undirected graph represented by A. Of course, the nature of this connected component segmentation is dictated by the criteria governing the placement of edges between points. Specifically, given points x i and x j, the corresponding adjacency matrix entry A i,j = 1 if both points are a k-nearest neighbor of each other, and if the line x j x i lies within a user defined angle θ (see figure 4) of the tangent plane associated with unoriented surface normal approximation for x i, and vice versa (see figure 5). In this way, point proximity, as well as local geometry are taken into account when detering point connectivity. Or put another way, neighboring points are connected if the local geometry estimate, given by the associated surface normal approximations, are compatible with location of x i and x j in IR 3. * θ Figure 4. Two dimensional representation of a data point and its tangent plane. Figure 5. The black point and blue point are connected to one another. The red point is not connected to either point. Based on this criteria, segmenting the point cloud according to the connected components of A can be considered an edge detection based scheme. However, a serendipitous characteristic of this methodology is its innate ability to account for the possibility of multiple closed surfaces, or bodies, within the data set. For a smoothly varying surface, such as a sphere, the corresponding adjacency matrix is comprised of one connected component. However, for the cylinder and city data depicted in figures 6 and 7, this is not the case. Consequently, many connected components (point cloud segments) are present, and the resulting segmentation breaks each data set into piecewise smooth segments. A finer segmentation can be obtained from the connected component segmentation by first traversing the connectivity information contained in A to orient the approximated surface normals in the same manner as the imum spanning tree method described in [3]. However, the benefit of employing the graph associated with A for the orientation is the ability to orient data sets containing more than one body. The orientation procedure is carried out on one individual connected component segment at a time, and is kept consis- * *
5 Figure 7. (Right) Point cloud representation of a city.(left) Connected component segmentation. Figure 8. Segmentation results for a sphere using the six outward normals for the faces of a cube as classification vectors. ness (in terms of angular difference) to one of the 6 classification vectors and then grouped together based on the connectivity described by A. Figure 6. (Right) Point cloud representation of a cylinder, with calculated normals. (Left) Connected component segmentation. 6. Results The full algorithm is as follows: Algorithm tent within a body by identifying points with nearest neighbors belonging to two connected component segments, and then propagating the information from one connected component segment to another. Given the oriented surface normal approximations, the data set is further segmented by classifying points with similar surface normals. Specifically, given a set of N classification vectors, each data point is categorized into one of N groups according to the similarity between the surface normal and the classification vectors. Consider the example of a sphere, with the classification vectors chosen to be the 6 outward normals for each face of a cube. Since the sphere is a smooth surface, there is only one connected component present in A (see 8). Assug properly orientated surface normals, each point is categorized according to their close- 1. Find the k-nearest neighbors of each data point. 2. Calculate the surface normal approximations. 3. Construct A by detering the edges between points. 4. Detere the initial segmentation by finding the connected components of A. 5. Orient the surface normal approximations. 6. Refine the segmentation according to N classification vectors. The main computational workload is represented by the calculation of the k-nearest neighbors. For the city examples shown in figure 7 containing 1.99 million points, the
6 k-nearest neighbor calculation (tree based sorting method with O(n log n) complexity) required 66 seconds on a Dell Precision M6400 laptop, while the normal calculations utilizing our nonlinear least squares required 39 seconds. The PCA normals required 24 seconds. [6] Y. Shen Liu, J. Claude Paul, J. Hai Yong, P. Qiang Yu, H. Zhang, J. Guang Sun, and K. Ramani. Automatic leastsquares projection of points onto point clouds with applications in reverse engineering. Computer-Aided Design, 38: , Conclusions and Future Work This paper presents a strategy for producing surface normals estimates for unstructured point cloud data based on a constrained nonlinear least squares formulation. The formulation is designed to be robust with respect to singularities in the data, such as surface edges, and incorporates data denoising in a fashion similar to planar moving least squares. Utilizing the surface normals, we also present a simple method for segmenting the point cloud data based on point connectivity. Refinement of the initial connected component segmentation, according to classification vectors and the oriented surface normals, allows for each new sub-segment to easily be approximated by a polynomial, the immediate benefit being a potentially simple procedure for point cloud compression. Rather than storing each data point, one need only store the coefficients of the polynomial, and perhaps a list of the points which lie more than some tolerance away from the approximating polynomial.furthermore, given the embarrassingly parallel nature of the procedure, such an algorithm could be implemented on a GPU (graphic processing unit) for high performance on large data sets. 8. Acknowledgments This material is based upon work supported by, or in part by, the U.S. Army Research Office under contract/grant number W911NF Edward Castillo was a part of the UC Irvine Department of Mathematics when this work was accomplished. References [1] P. Azariadis. Parameterization of clouds of unorganized points using dynamic base surfaces. Computer-Aided Design, 36: , [2] M. Gopi, S. Krishnan, and C. Silva. Surface reconstruction based on lower dimensional localized delaunay triangulation. In Proceedings of Eurographics, volume 19, [3] H. Hoppe, T. DeRose, T. Duchamp, J. McDonald, and W. Stuetzle. Surface reconstruction from unorganized points. In Proceedings of ACM Siggraph, pages 71 78, , 4 [4] S. Hu and J. Wallner. A second order algorithm for orhtogonal projection onto curves and surfaces. Computer Aided Graphic Design, 22(3): , [5] N. Mitra, A. Nguyen, and L. Guibas. Estimating surface normals in noisy point cloud data. In Special issue of International Journal of Computational Geometry and Applications, volume 14, pages ,
DATA ANALYSIS II. Matrix Algorithms
DATA ANALYSIS II Matrix Algorithms Similarity Matrix Given a dataset D = {x i }, i=1,..,n consisting of n points in R d, let A denote the n n symmetric similarity matrix between the points, given as where
Nonlinear Iterative Partial Least Squares Method
Numerical Methods for Determining Principal Component Analysis Abstract Factors Béchu, S., Richard-Plouet, M., Fernandez, V., Walton, J., and Fairley, N. (2016) Developments in numerical treatments for
Robust NURBS Surface Fitting from Unorganized 3D Point Clouds for Infrastructure As-Built Modeling
81 Robust NURBS Surface Fitting from Unorganized 3D Point Clouds for Infrastructure As-Built Modeling Andrey Dimitrov 1 and Mani Golparvar-Fard 2 1 Graduate Student, Depts of Civil Eng and Engineering
Least-Squares Intersection of Lines
Least-Squares Intersection of Lines Johannes Traa - UIUC 2013 This write-up derives the least-squares solution for the intersection of lines. In the general case, a set of lines will not intersect at a
1 Review of Least Squares Solutions to Overdetermined Systems
cs4: introduction to numerical analysis /9/0 Lecture 7: Rectangular Systems and Numerical Integration Instructor: Professor Amos Ron Scribes: Mark Cowlishaw, Nathanael Fillmore Review of Least Squares
Manifold Learning Examples PCA, LLE and ISOMAP
Manifold Learning Examples PCA, LLE and ISOMAP Dan Ventura October 14, 28 Abstract We try to give a helpful concrete example that demonstrates how to use PCA, LLE and Isomap, attempts to provide some intuition
Applications to Data Smoothing and Image Processing I
Applications to Data Smoothing and Image Processing I MA 348 Kurt Bryan Signals and Images Let t denote time and consider a signal a(t) on some time interval, say t. We ll assume that the signal a(t) is
From Scattered Samples to Smooth Surfaces
From Scattered Samples to Smooth Surfaces Kai Hormann 1 California Institute of Technology (a) (b) (c) (d) Figure 1: A point cloud with 4,100 scattered samples (a), its triangulation with 7,938 triangles
We can display an object on a monitor screen in three different computer-model forms: Wireframe model Surface Model Solid model
CHAPTER 4 CURVES 4.1 Introduction In order to understand the significance of curves, we should look into the types of model representations that are used in geometric modeling. Curves play a very significant
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION Introduction In the previous chapter, we explored a class of regression models having particularly simple analytical
A unified representation for interactive 3D modeling
A unified representation for interactive 3D modeling Dragan Tubić, Patrick Hébert, Jean-Daniel Deschênes and Denis Laurendeau Computer Vision and Systems Laboratory, University Laval, Québec, Canada [tdragan,hebert,laurendeau]@gel.ulaval.ca
Fast and Robust Normal Estimation for Point Clouds with Sharp Features
1/37 Fast and Robust Normal Estimation for Point Clouds with Sharp Features Alexandre Boulch & Renaud Marlet University Paris-Est, LIGM (UMR CNRS), Ecole des Ponts ParisTech Symposium on Geometry Processing
(Refer Slide Time: 1:42)
Introduction to Computer Graphics Dr. Prem Kalra Department of Computer Science and Engineering Indian Institute of Technology, Delhi Lecture - 10 Curves So today we are going to have a new topic. So far
H.Calculating Normal Vectors
Appendix H H.Calculating Normal Vectors This appendix describes how to calculate normal vectors for surfaces. You need to define normals to use the OpenGL lighting facility, which is described in Chapter
Section 1.1. Introduction to R n
The Calculus of Functions of Several Variables Section. Introduction to R n Calculus is the study of functional relationships and how related quantities change with each other. In your first exposure to
Machine Learning and Pattern Recognition Logistic Regression
Machine Learning and Pattern Recognition Logistic Regression Course Lecturer:Amos J Storkey Institute for Adaptive and Neural Computation School of Informatics University of Edinburgh Crichton Street,
Comparison of Non-linear Dimensionality Reduction Techniques for Classification with Gene Expression Microarray Data
CMPE 59H Comparison of Non-linear Dimensionality Reduction Techniques for Classification with Gene Expression Microarray Data Term Project Report Fatma Güney, Kübra Kalkan 1/15/2013 Keywords: Non-linear
Equations Involving Lines and Planes Standard equations for lines in space
Equations Involving Lines and Planes In this section we will collect various important formulas regarding equations of lines and planes in three dimensional space Reminder regarding notation: any quantity
Dimensionality Reduction: Principal Components Analysis
Dimensionality Reduction: Principal Components Analysis In data mining one often encounters situations where there are a large number of variables in the database. In such situations it is very likely
BEHAVIOR BASED CREDIT CARD FRAUD DETECTION USING SUPPORT VECTOR MACHINES
BEHAVIOR BASED CREDIT CARD FRAUD DETECTION USING SUPPORT VECTOR MACHINES 123 CHAPTER 7 BEHAVIOR BASED CREDIT CARD FRAUD DETECTION USING SUPPORT VECTOR MACHINES 7.1 Introduction Even though using SVM presents
Biggar High School Mathematics Department. National 5 Learning Intentions & Success Criteria: Assessing My Progress
Biggar High School Mathematics Department National 5 Learning Intentions & Success Criteria: Assessing My Progress Expressions & Formulae Topic Learning Intention Success Criteria I understand this Approximation
Introduction to Principal Components and FactorAnalysis
Introduction to Principal Components and FactorAnalysis Multivariate Analysis often starts out with data involving a substantial number of correlated variables. Principal Component Analysis (PCA) is a
Solving Simultaneous Equations and Matrices
Solving Simultaneous Equations and Matrices The following represents a systematic investigation for the steps used to solve two simultaneous linear equations in two unknowns. The motivation for considering
11.1. Objectives. Component Form of a Vector. Component Form of a Vector. Component Form of a Vector. Vectors and the Geometry of Space
11 Vectors and the Geometry of Space 11.1 Vectors in the Plane Copyright Cengage Learning. All rights reserved. Copyright Cengage Learning. All rights reserved. 2 Objectives! Write the component form of
How To Create A Surface From Points On A Computer With A Marching Cube
Surface Reconstruction from a Point Cloud with Normals Landon Boyd and Massih Khorvash Department of Computer Science University of British Columbia,2366 Main Mall Vancouver, BC, V6T1Z4, Canada {blandon,khorvash}@cs.ubc.ca
CS 5614: (Big) Data Management Systems. B. Aditya Prakash Lecture #18: Dimensionality Reduc7on
CS 5614: (Big) Data Management Systems B. Aditya Prakash Lecture #18: Dimensionality Reduc7on Dimensionality Reduc=on Assump=on: Data lies on or near a low d- dimensional subspace Axes of this subspace
Multimedia Databases. Wolf-Tilo Balke Philipp Wille Institut für Informationssysteme Technische Universität Braunschweig http://www.ifis.cs.tu-bs.
Multimedia Databases Wolf-Tilo Balke Philipp Wille Institut für Informationssysteme Technische Universität Braunschweig http://www.ifis.cs.tu-bs.de 14 Previous Lecture 13 Indexes for Multimedia Data 13.1
Segmentation of building models from dense 3D point-clouds
Segmentation of building models from dense 3D point-clouds Joachim Bauer, Konrad Karner, Konrad Schindler, Andreas Klaus, Christopher Zach VRVis Research Center for Virtual Reality and Visualization, Institute
Name Class. Date Section. Test Form A Chapter 11. Chapter 11 Test Bank 155
Chapter Test Bank 55 Test Form A Chapter Name Class Date Section. Find a unit vector in the direction of v if v is the vector from P,, 3 to Q,, 0. (a) 3i 3j 3k (b) i j k 3 i 3 j 3 k 3 i 3 j 3 k. Calculate
Social Media Mining. Network Measures
Klout Measures and Metrics 22 Why Do We Need Measures? Who are the central figures (influential individuals) in the network? What interaction patterns are common in friends? Who are the like-minded users
Review Jeopardy. Blue vs. Orange. Review Jeopardy
Review Jeopardy Blue vs. Orange Review Jeopardy Jeopardy Round Lectures 0-3 Jeopardy Round $200 How could I measure how far apart (i.e. how different) two observations, y 1 and y 2, are from each other?
Linear Threshold Units
Linear Threshold Units w x hx (... w n x n w We assume that each feature x j and each weight w j is a real number (we will relax this later) We will study three different algorithms for learning linear
GRADES 7, 8, AND 9 BIG IDEAS
Table 1: Strand A: BIG IDEAS: MATH: NUMBER Introduce perfect squares, square roots, and all applications Introduce rational numbers (positive and negative) Introduce the meaning of negative exponents for
The Scientific Data Mining Process
Chapter 4 The Scientific Data Mining Process When I use a word, Humpty Dumpty said, in rather a scornful tone, it means just what I choose it to mean neither more nor less. Lewis Carroll [87, p. 214] In
Introduction Epipolar Geometry Calibration Methods Further Readings. Stereo Camera Calibration
Stereo Camera Calibration Stereo Camera Calibration Stereo Camera Calibration Stereo Camera Calibration 12.10.2004 Overview Introduction Summary / Motivation Depth Perception Ambiguity of Correspondence
Algebra Unpacked Content For the new Common Core standards that will be effective in all North Carolina schools in the 2012-13 school year.
This document is designed to help North Carolina educators teach the Common Core (Standard Course of Study). NCDPI staff are continually updating and improving these tools to better serve teachers. Algebra
In mathematics, there are four attainment targets: using and applying mathematics; number and algebra; shape, space and measures, and handling data.
MATHEMATICS: THE LEVEL DESCRIPTIONS In mathematics, there are four attainment targets: using and applying mathematics; number and algebra; shape, space and measures, and handling data. Attainment target
Curves and Surfaces. Goals. How do we draw surfaces? How do we specify a surface? How do we approximate a surface?
Curves and Surfaces Parametric Representations Cubic Polynomial Forms Hermite Curves Bezier Curves and Surfaces [Angel 10.1-10.6] Goals How do we draw surfaces? Approximate with polygons Draw polygons
8. Linear least-squares
8. Linear least-squares EE13 (Fall 211-12) definition examples and applications solution of a least-squares problem, normal equations 8-1 Definition overdetermined linear equations if b range(a), cannot
Normal Estimation for Point Clouds: A Comparison Study for a Voronoi Based Method
Eurographics Symposium on Point-Based Graphics (2005) M. Pauly, M. Zwicker, (Editors) Normal Estimation for Point Clouds: A Comparison Study for a Voronoi Based Method Tamal K. Dey Gang Li Jian Sun The
Figure 2.1: Center of mass of four points.
Chapter 2 Bézier curves are named after their inventor, Dr. Pierre Bézier. Bézier was an engineer with the Renault car company and set out in the early 196 s to develop a curve formulation which would
13 MATH FACTS 101. 2 a = 1. 7. The elements of a vector have a graphical interpretation, which is particularly easy to see in two or three dimensions.
3 MATH FACTS 0 3 MATH FACTS 3. Vectors 3.. Definition We use the overhead arrow to denote a column vector, i.e., a linear segment with a direction. For example, in three-space, we write a vector in terms
Going Big in Data Dimensionality:
LUDWIG- MAXIMILIANS- UNIVERSITY MUNICH DEPARTMENT INSTITUTE FOR INFORMATICS DATABASE Going Big in Data Dimensionality: Challenges and Solutions for Mining High Dimensional Data Peer Kröger Lehrstuhl für
15.062 Data Mining: Algorithms and Applications Matrix Math Review
.6 Data Mining: Algorithms and Applications Matrix Math Review The purpose of this document is to give a brief review of selected linear algebra concepts that will be useful for the course and to develop
Computer Graphics. Geometric Modeling. Page 1. Copyright Gotsman, Elber, Barequet, Karni, Sheffer Computer Science - Technion. An Example.
An Example 2 3 4 Outline Objective: Develop methods and algorithms to mathematically model shape of real world objects Categories: Wire-Frame Representation Object is represented as as a set of points
MA 323 Geometric Modelling Course Notes: Day 02 Model Construction Problem
MA 323 Geometric Modelling Course Notes: Day 02 Model Construction Problem David L. Finn November 30th, 2004 In the next few days, we will introduce some of the basic problems in geometric modelling, and
Computer Graphics CS 543 Lecture 12 (Part 1) Curves. Prof Emmanuel Agu. Computer Science Dept. Worcester Polytechnic Institute (WPI)
Computer Graphics CS 54 Lecture 1 (Part 1) Curves Prof Emmanuel Agu Computer Science Dept. Worcester Polytechnic Institute (WPI) So Far Dealt with straight lines and flat surfaces Real world objects include
Metrics on SO(3) and Inverse Kinematics
Mathematical Foundations of Computer Graphics and Vision Metrics on SO(3) and Inverse Kinematics Luca Ballan Institute of Visual Computing Optimization on Manifolds Descent approach d is a ascent direction
CS 591.03 Introduction to Data Mining Instructor: Abdullah Mueen
CS 591.03 Introduction to Data Mining Instructor: Abdullah Mueen LECTURE 3: DATA TRANSFORMATION AND DIMENSIONALITY REDUCTION Chapter 3: Data Preprocessing Data Preprocessing: An Overview Data Quality Major
State of Stress at Point
State of Stress at Point Einstein Notation The basic idea of Einstein notation is that a covector and a vector can form a scalar: This is typically written as an explicit sum: According to this convention,
Linear Programming I
Linear Programming I November 30, 2003 1 Introduction In the VCR/guns/nuclear bombs/napkins/star wars/professors/butter/mice problem, the benevolent dictator, Bigus Piguinus, of south Antarctica penguins
Algebra 1 2008. Academic Content Standards Grade Eight and Grade Nine Ohio. Grade Eight. Number, Number Sense and Operations Standard
Academic Content Standards Grade Eight and Grade Nine Ohio Algebra 1 2008 Grade Eight STANDARDS Number, Number Sense and Operations Standard Number and Number Systems 1. Use scientific notation to express
P164 Tomographic Velocity Model Building Using Iterative Eigendecomposition
P164 Tomographic Velocity Model Building Using Iterative Eigendecomposition K. Osypov* (WesternGeco), D. Nichols (WesternGeco), M. Woodward (WesternGeco) & C.E. Yarman (WesternGeco) SUMMARY Tomographic
APPM4720/5720: Fast algorithms for big data. Gunnar Martinsson The University of Colorado at Boulder
APPM4720/5720: Fast algorithms for big data Gunnar Martinsson The University of Colorado at Boulder Course objectives: The purpose of this course is to teach efficient algorithms for processing very large
Compact Representations and Approximations for Compuation in Games
Compact Representations and Approximations for Compuation in Games Kevin Swersky April 23, 2008 Abstract Compact representations have recently been developed as a way of both encoding the strategic interactions
Mean Value Coordinates
Mean Value Coordinates Michael S. Floater Abstract: We derive a generalization of barycentric coordinates which allows a vertex in a planar triangulation to be expressed as a convex combination of its
Recall that two vectors in are perpendicular or orthogonal provided that their dot
Orthogonal Complements and Projections Recall that two vectors in are perpendicular or orthogonal provided that their dot product vanishes That is, if and only if Example 1 The vectors in are orthogonal
x1 x 2 x 3 y 1 y 2 y 3 x 1 y 2 x 2 y 1 0.
Cross product 1 Chapter 7 Cross product We are getting ready to study integration in several variables. Until now we have been doing only differential calculus. One outcome of this study will be our ability
Statistical Machine Learning
Statistical Machine Learning UoC Stats 37700, Winter quarter Lecture 4: classical linear and quadratic discriminants. 1 / 25 Linear separation For two classes in R d : simple idea: separate the classes
6 J - vector electric current density (A/m2 )
Determination of Antenna Radiation Fields Using Potential Functions Sources of Antenna Radiation Fields 6 J - vector electric current density (A/m2 ) M - vector magnetic current density (V/m 2 ) Some problems
Fast and Accurate Computation of Surface Normals from Range Images
Fast and Accurate Computation of Surface Normals from Range Images H. Badino, D. Huber, Y. Park and T. Kanade Abstract The fast and accurate computation of surface normals from a point cloud is a critical
ME 115(b): Solution to Homework #1
ME 115(b): Solution to Homework #1 Solution to Problem #1: To construct the hybrid Jacobian for a manipulator, you could either construct the body Jacobian, JST b, and then use the body-to-hybrid velocity
Solutions to old Exam 1 problems
Solutions to old Exam 1 problems Hi students! I am putting this old version of my review for the first midterm review, place and time to be announced. Check for updates on the web site as to which sections
How To Understand And Solve A Linear Programming Problem
At the end of the lesson, you should be able to: Chapter 2: Systems of Linear Equations and Matrices: 2.1: Solutions of Linear Systems by the Echelon Method Define linear systems, unique solution, inconsistent,
EM Clustering Approach for Multi-Dimensional Analysis of Big Data Set
EM Clustering Approach for Multi-Dimensional Analysis of Big Data Set Amhmed A. Bhih School of Electrical and Electronic Engineering Princy Johnson School of Electrical and Electronic Engineering Martin
A QUICK GUIDE TO THE FORMULAS OF MULTIVARIABLE CALCULUS
A QUIK GUIDE TO THE FOMULAS OF MULTIVAIABLE ALULUS ontents 1. Analytic Geometry 2 1.1. Definition of a Vector 2 1.2. Scalar Product 2 1.3. Properties of the Scalar Product 2 1.4. Length and Unit Vectors
the points are called control points approximating curve
Chapter 4 Spline Curves A spline curve is a mathematical representation for which it is easy to build an interface that will allow a user to design and control the shape of complex curves and surfaces.
Surface Normals and Tangent Planes
Surface Normals and Tangent Planes Normal and Tangent Planes to Level Surfaces Because the equation of a plane requires a point and a normal vector to the plane, nding the equation of a tangent plane to
THREE DIMENSIONAL GEOMETRY
Chapter 8 THREE DIMENSIONAL GEOMETRY 8.1 Introduction In this chapter we present a vector algebra approach to three dimensional geometry. The aim is to present standard properties of lines and planes,
Assessment. Presenter: Yupu Zhang, Guoliang Jin, Tuo Wang Computer Vision 2008 Fall
Automatic Photo Quality Assessment Presenter: Yupu Zhang, Guoliang Jin, Tuo Wang Computer Vision 2008 Fall Estimating i the photorealism of images: Distinguishing i i paintings from photographs h Florin
EECS 556 Image Processing W 09. Interpolation. Interpolation techniques B splines
EECS 556 Image Processing W 09 Interpolation Interpolation techniques B splines What is image processing? Image processing is the application of 2D signal processing methods to images Image representation
Common Core Unit Summary Grades 6 to 8
Common Core Unit Summary Grades 6 to 8 Grade 8: Unit 1: Congruence and Similarity- 8G1-8G5 rotations reflections and translations,( RRT=congruence) understand congruence of 2 d figures after RRT Dilations
December 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B. KITCHENS
December 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B KITCHENS The equation 1 Lines in two-dimensional space (1) 2x y = 3 describes a line in two-dimensional space The coefficients of x and y in the equation
Big Data: Rethinking Text Visualization
Big Data: Rethinking Text Visualization Dr. Anton Heijs [email protected] Treparel April 8, 2013 Abstract In this white paper we discuss text visualization approaches and how these are important
Parametric Curves. (Com S 477/577 Notes) Yan-Bin Jia. Oct 8, 2015
Parametric Curves (Com S 477/577 Notes) Yan-Bin Jia Oct 8, 2015 1 Introduction A curve in R 2 (or R 3 ) is a differentiable function α : [a,b] R 2 (or R 3 ). The initial point is α[a] and the final point
Mathematics (MAT) MAT 061 Basic Euclidean Geometry 3 Hours. MAT 051 Pre-Algebra 4 Hours
MAT 051 Pre-Algebra Mathematics (MAT) MAT 051 is designed as a review of the basic operations of arithmetic and an introduction to algebra. The student must earn a grade of C or in order to enroll in MAT
Review of Fundamental Mathematics
Review of Fundamental Mathematics As explained in the Preface and in Chapter 1 of your textbook, managerial economics applies microeconomic theory to business decision making. The decision-making tools
Mesh Discretization Error and Criteria for Accuracy of Finite Element Solutions
Mesh Discretization Error and Criteria for Accuracy of Finite Element Solutions Chandresh Shah Cummins, Inc. Abstract Any finite element analysis performed by an engineer is subject to several types of
How To Cluster
Data Clustering Dec 2nd, 2013 Kyrylo Bessonov Talk outline Introduction to clustering Types of clustering Supervised Unsupervised Similarity measures Main clustering algorithms k-means Hierarchical Main
Nonlinear Programming Methods.S2 Quadratic Programming
Nonlinear Programming Methods.S2 Quadratic Programming Operations Research Models and Methods Paul A. Jensen and Jonathan F. Bard A linearly constrained optimization problem with a quadratic objective
Classification of Fingerprints. Sarat C. Dass Department of Statistics & Probability
Classification of Fingerprints Sarat C. Dass Department of Statistics & Probability Fingerprint Classification Fingerprint classification is a coarse level partitioning of a fingerprint database into smaller
Part-Based Recognition
Part-Based Recognition Benedict Brown CS597D, Fall 2003 Princeton University CS 597D, Part-Based Recognition p. 1/32 Introduction Many objects are made up of parts It s presumably easier to identify simple
An Overview Of Software For Convex Optimization. Brian Borchers Department of Mathematics New Mexico Tech Socorro, NM 87801 borchers@nmt.
An Overview Of Software For Convex Optimization Brian Borchers Department of Mathematics New Mexico Tech Socorro, NM 87801 [email protected] In fact, the great watershed in optimization isn t between linearity
Subspace Analysis and Optimization for AAM Based Face Alignment
Subspace Analysis and Optimization for AAM Based Face Alignment Ming Zhao Chun Chen College of Computer Science Zhejiang University Hangzhou, 310027, P.R.China [email protected] Stan Z. Li Microsoft
Support Vector Machine (SVM)
Support Vector Machine (SVM) CE-725: Statistical Pattern Recognition Sharif University of Technology Spring 2013 Soleymani Outline Margin concept Hard-Margin SVM Soft-Margin SVM Dual Problems of Hard-Margin
3. INNER PRODUCT SPACES
. INNER PRODUCT SPACES.. Definition So far we have studied abstract vector spaces. These are a generalisation of the geometric spaces R and R. But these have more structure than just that of a vector space.
Blind Deconvolution of Barcodes via Dictionary Analysis and Wiener Filter of Barcode Subsections
Blind Deconvolution of Barcodes via Dictionary Analysis and Wiener Filter of Barcode Subsections Maximilian Hung, Bohyun B. Kim, Xiling Zhang August 17, 2013 Abstract While current systems already provide
The equivalence of logistic regression and maximum entropy models
The equivalence of logistic regression and maximum entropy models John Mount September 23, 20 Abstract As our colleague so aptly demonstrated ( http://www.win-vector.com/blog/20/09/the-simplerderivation-of-logistic-regression/
Surface Reconstruction from Point Clouds
Autonomous Systems Lab Prof. Roland Siegwart Bachelor-Thesis Surface Reconstruction from Point Clouds Spring Term 2010 Supervised by: Andreas Breitenmoser François Pomerleau Author: Inna Tishchenko Abstract
Module1. x 1000. y 800.
Module1 1 Welcome to the first module of the course. It is indeed an exciting event to share with you the subject that has lot to offer both from theoretical side and practical aspects. To begin with,
521493S Computer Graphics. Exercise 2 & course schedule change
521493S Computer Graphics Exercise 2 & course schedule change Course Schedule Change Lecture from Wednesday 31th of March is moved to Tuesday 30th of March at 16-18 in TS128 Question 2.1 Given two nonparallel,
Chapter 111. Texas Essential Knowledge and Skills for Mathematics. Subchapter B. Middle School
Middle School 111.B. Chapter 111. Texas Essential Knowledge and Skills for Mathematics Subchapter B. Middle School Statutory Authority: The provisions of this Subchapter B issued under the Texas Education
Optimization Modeling for Mining Engineers
Optimization Modeling for Mining Engineers Alexandra M. Newman Division of Economics and Business Slide 1 Colorado School of Mines Seminar Outline Linear Programming Integer Linear Programming Slide 2
Section 1.4. Lines, Planes, and Hyperplanes. The Calculus of Functions of Several Variables
The Calculus of Functions of Several Variables Section 1.4 Lines, Planes, Hyperplanes In this section we will add to our basic geometric understing of R n by studying lines planes. If we do this carefully,
4.3 Least Squares Approximations
18 Chapter. Orthogonality.3 Least Squares Approximations It often happens that Ax D b has no solution. The usual reason is: too many equations. The matrix has more rows than columns. There are more equations
L 2 : x = s + 1, y = s, z = 4s + 4. 3. Suppose that C has coordinates (x, y, z). Then from the vector equality AC = BD, one has
The line L through the points A and B is parallel to the vector AB = 3, 2, and has parametric equations x = 3t + 2, y = 2t +, z = t Therefore, the intersection point of the line with the plane should satisfy:
Arrangements And Duality
Arrangements And Duality 3.1 Introduction 3 Point configurations are tbe most basic structure we study in computational geometry. But what about configurations of more complicated shapes? For example,
