A Correlation-Based Approach to Robust Point Set Registration
|
|
|
- Donald Rich
- 10 years ago
- Views:
Transcription
1 A Correlation-Based Approach to Robust Point Set Registration Yanghai Tsin 1 and Takeo Kanade 2 1 Siemens Corporate Research, 755 College Road East, Princeton, NJ 854, USA 2 Carnegie Mellon University, 5 Forbes Ave., Pittsburgh, PA 15213, USA Abstract. Correlation is a very effective way to align intensity images. We etend the correlation technique to point set registration using a method we call kernel correlation. Kernel correlation is an affinity measure, and it is also a function of the point set entropy. We define the point set registration problem as finding the maimum kernel correlation configuration of the the two point sets to be registered. The new registration method has intuitive interpretations, simple to implement algorithm and easy to prove convergence property. Our method shows favorable performance when compared with the iterative closest point () and EM- methods. 1 Introduction Point set registration is among the most fundamental problems in vision research. It is widely used in areas such as range data fusion, medical image alignment, object localization, tracking, object recognition, just to name a few. One of the most effective methods in registration is correlation. In vision problems, correlation between two image patches has long been used for measuring the similarities between them. When studying discrete point sets, such as those returned by range sensors or feature detectors, however, we are given just the coordinates of a set of points. The definition of correlation is no longer directly applicable since we are given a set of geometric entities without any appearance information to compare. Nevertheless, the presence or absence of feature points themselves tell a lot more than the coordinates of individual points. They also present the structure implied by the point sets. The simplest way of capturing such structure is to treat the feature points as binary intensity images which have only values (absence) and 1 (presence). However, when noise presents, or when we have different sampling strategies in obtaining the two point sets, the binary images usually do not match. In the following we present a technique we call kernel correlation that etends the concept of correlation to point sets. We begin by introducing kernel correlation. T. Pajdla and J. Matas (Eds.): ECCV 24, LNCS 323, pp , 24. c Springer-Verlag Berlin Heidelberg 24
2 A Correlation-Based Approach to Robust Point Set Registration Kernel Correlation 2.1 Definitions Kernel correlation () is defined on three levels. First, it is defined on two points. Given two points i and j, their kernel correlation () is defined as ( i, j )= K(, i ) K(, j )d. (1) Here K(, i ) is a kernel function centered at the data point i. We limit ourselves to the symmetric, non-negative kernel functions that are usually used in the Parzen window density estimation [11], instead of the broader kernel definition in the machine learning community. Such kernels include the Gaussian kernel, Epanechnikov kernel, tri-cubic kernel, et. al. [7]. In the following we use the Gaussian kernel as an eample for its simplicity. The Gaussian kernel has the form: K G (, i )=(πσ 2 ) D/2 ep ( i 2 /σ 2 ). (2) Here y is the Euclidean distance between two vectors and y, and D is the dimension of the vectors. Because the kernel functions we adopt are symmetric, it s not surprising to see that the defined in (1) is a function of distance between the two points. For eample, the corresponding to the Gaussian kernel is, G ( i, j )=(2πσ 2 ) D/2 ep { i j 2 /2σ 2}. (3) s for other kernels can be shown to be functions of distance i j as well. For clarity of the presentation we will not list them here. But we will discuss their shared properties with the Gaussian kernel whenever necessary. Right side of (3) is known in the vision community as affinity or proimity [17]: a closeness measure. In this paper we show its utility in registration problems. Net we define the between a point and the whole set of points X, the Leave-one-out Kernel Correlation (LOO-), ( i, X )= ( i, j ). (4) j i According to (3) and (4), for the Gaussian case we have G ( i, X )=(2π) D/2 ep{ j i 2 /2σ 2 } (5) j i Leave-one-out defines the total affinity from a point to a point set. Finally, we etend the definition to a point set: the total sum of the LOO- of all the points k in the set, (X )= i ( i, X )=2 i j ( i, j ). (6) If the points in the set are close to each other, the value is large. In this sense of a point set is a compactness measure of the point set.
3 56 Y. Tsin and T. Kanade 2.2 Entropy Equivalence If we define the density of the point set X as the kernel density estimate: P () = N K(, i )/N, (7) i=1 and adopt the Renyi s Quadratic Entropy (RQE) [15] as, H rqe = log P () 2 d, (8) of the point set has a simple relationship with the entropy measure, (X ) C + ep{ H rqe }. (9) The above observation follows directly by epanding the P ()2 d term in the entropy definition. In fact, N 2 P () 2 d = K(, i ) 2 d +2 K(, i )K(, j )d i i j = C + (X ). Here we use the fact that K(, i) 2 d is a constant and the definition of (X ) (6). Note that the relationship does not assume any specific form of kernel function, as long as the integrals are defined. Thus the compactness measure of is linked to the compactness measure of entropy. A minimum entropy system is the one with the maimum affinity (minimum distance) between all pairs of points. The information theoretic compactness measure indeed has a geometric interpretation. We were brought to the attention of the independent work by Principe and Xu [13]. They epanded the RQE definition in the Gaussian case and defined the integral of the cross product terms as information potential. Their purpose for such decomposition is efficient evaluation of entropy and entropy gradients in the contet of information theoretic learning. In contrast, our goal is instead to configure a dynamic point set. 2.3 as an M-Estimator If there are just two points involved, maimum corresponds to minimum distance between them. However, when we are dealing with multiple points, it s not immediately obvious what is being optimized. For instance, in the Gaussian case we have (5). What does it mean to maimize? It turns out that in this case we are still minimizing the distance, but in the sense of M-estimators. In an M-estimator, instead of minimizing the usual sum of quadratic distances, E q = j ( i j ) 2, we are minimizing a robust version of the distance
4 A Correlation-Based Approach to Robust Point Set Registration 561 function E r = j g ( ( i j ) 2), where g is a robust function [8]. The advantage of changing from the quadratic distance function to the robust function is that local configuration of i is insensitive to remote points. To see this we compare the gradients of the above two functions. E q / i j ( i j ) (1) G ( i, X )/ i j ep( i j 2 /2σ 2 )( j i ). (11) The gradient term (1) is very sensitive to outliers in that any outlier point j can have arbitrarily large contribution to the gradient. Remember that the gradient is the direction (and magnitude in the quadratic function case) to update i.to minimize E q, estimation of i will be severely biased toward the outlier points. In the G case, however, there is a second term ep( i j 2 /2σ 2 ) that decays eponentially as a function of distance. Consequently, remote outliers will have no influence to local E r minimization. When we use kernels other than the Gaussian kernel, we can still have the M- estimator equivalence when maimizing. For eample, by using the Epanechnikov kernel, we implicitly embedded a line process [5] into the correlation process: Points beyond a certain distance don t contribute. Chen and Meer [3] also observed the equivalence of mode finding in a kernel density estimate and the M-estimators. The difference is that they are fitting parametric models to a set of static data (the projection pursuit eample), or clustering the static point set. The introduction of is to robustly configure dynamic point sets. 3 Kernel Correlation for Registration Given two finite size point sets, the model set M and the scene set S, our registration method is defined as finding the parameter θ of a transformation T to minimize the following cost function, COST (S, M,θ)= (s, T (m, θ)). (12) s S m M Notice that in the above equation each transformed model point m is interacting with all the scene points. We call (12) a multiply-linked registration cost function. This is in contrast to the algorithm, where each model point is connected to its nearest scene point only. It can be shown that, (S T (M,θ)) = (S)+(T (M,θ)) 2COST (S, M,θ). (13) (S) is independent of θ. Under rigid transformation, (T (M,θ)) is also constant. This is the case because is a function of Euclidean distances between pairs of points (e.g. (3)). Rigid transformation reconfigures the point set as a whole and preserves the Euclidean distances between all pairs of points. Thus (T (M,θ)) is invariant. As a result (S T (M,θ)) = C 2COST (S, M,θ).
5 562 Y. Tsin and T. Kanade Due to the equivalence of and entropy (Section 2.2), our registration method implies finding the minimum entropy configuration of the joint point set S T (M,θ) in the RQE sense. By denoting the kernel density estimates (KDE) as P M (, θ) = K(, s)/n, m M K(, T (m, θ))/n, P S () = s S we can show that the cost function is also proportional to the correlation of the two KDE s, COST (S, M,θ)= N 2 P M P S d. (14) 3.1 Convergence of a Registration Algorithm It s easy to show that the cost function (12) is bounded from below. If we use gradient descent based method to minimize the cost function such that the cost function is decreasing at each step, the convergence of the cost function to a fied point is guaranteed. Convergence properties for other registration methods, such as or EM- are usually difficult to study because their cost functions, defined on nearest neighbors, change from iteration to iteration as the point configuration evolves. In contrast, the registration function is defined globally and each step of minimization decreases the same cost function. 3.2 Accuracy of Registration We will empirically study the accuracy of our registration algorithm in Section 5. Here we will discuss one of the simplest cases to theoretically characterize the registration algorithm. Given a point set M and it s transformed version S = T (M,θ ), a registration method should satisfy what we call the minimum requirement for a registration algorithm. That is, θ should correspond to one of the global minima of the cost function. Although this requirement seems to be trivial, we will show in our eperiments that it is not met by other multiply-linked registration algorithms. Here we first give a proof that our registration algorithm meets the minimum requirement under rigid transformation. The etension to no-rigid motion is followed. We observe that ( ) N 2 (P M P S ) 2 d = N 2 PMd 2 + PSd 2 2 P M P S d (15) = C + (T (M,θ)) + (S)+2 COST(S, M,θ). (16) Here C is a constant due to values of a point with itself. (S) is independent of θ. As we discussed in the beginning of this section, (T (M,θ)) is also a constant under rigid transformation. Thus minimizing the left side of (15) is equivalent to minimizing our registration cost function. When θ = θ, P M and
6 A Correlation-Based Approach to Robust Point Set Registration 563 P S are eactly the same and the left side of (15) is zero, the global minimum. That is, θ corresponds to one of the global minima of the registration cost function. Note that this statement is independent of the kernel functions being chosen and the kernel scale, as long as the integrals in the proof are all defined. The registration framework can be etended to non-rigid transformations if we minimize a normalized cost function. By denoting the normalization term as I M =( P 2 M d)1/2 the normalized cost function is COST n = s S,m M (s, T (m, θ))/i M. (17) Similar to (16), we can show that N 2 (P M /I M P S /I S ) 2 d =2N 2 +2 COST n /I S, (18) where I S =( P 2 S d)1/2 is independent of θ. Given that S = T (M,θ ), θ will again be one of the global minima of the registration cost function (17), even under non-rigid transformations. 3.3 Discrete Approimation of the Registration Cost Function In practice we don t need to enumerate each pair of model and scene points in order to evaluate the cost function (12) or (17). We can use the discrete version of (14) to approimate the registration cost. That is, we compute two discrete KDE s, P M (, θ) and P S () at grid points, and use P M(, θ) P S () to approimate the scaled cost function. Compared to the or EM- methods, there is no nearest neighbor finding step involved in the registration, which can result in significant simplification in algorithm implementation. P S () plays the role of an affinity map in our algorithm. The affinity of a model point m to the scene points can be computed by correlating K(, m) with P S (). 4 Related Work We store the affinity information in a density estimate. This technique bears much resemblance to the registration methods based on distance transform (DT) [2]. However, there are some important differences. First, DT is known to be etremely sensitive to outliers and noise because a single point can have influence to a large area. The influence of each point in the case is local. Thus based registration can be robust to outliers. Second, our affinity map is usually sparse for usual point sets such as an edge map or a laser range scan, with most grid points having zero values. The affinity map can be efficiently stored in data structures such as an octree. In contrast, high resolution DT in 3D is very costly. This prompted Lavallée and Szeliski to approimate 3D DT using octree spline [1].
7 564 Y. Tsin and T. Kanade One elegant registration method based on DT is the partial Hausdorff distance registration [9]. By minimizing partial Hausdorff distance, a registration algorithm can have up to 5 percent breakdown point. The underlying robustness mechanism is the same as the least medium of squares (LMedS) algorithm [16] in robust regression. However, the registration depends on a single critical point in the data set and most information provided by other points are ignored. Compared to other registration methods such as and our proposed method, it is very sensitive to noise. Scott and Longuet-Higgins [17] eplored the possibility of finding correspondence by singular value decomposition (SVD) analysis of an affinity matri, whose elements are proportional to the Gaussian values. Their algorithm is known to be vulnerable to perturbations such as large rotations, outliers and noise. In addition, forming a large affinity matri for a large point set is costly. One of the most successful point set registration algorithms is the iterative closest point () algorithm [1,19]. A naive implementation of is not robust because the cost function is a quadratic function of distance. To be robust, line-process like outlier detection or M-estimator like cost functions have been suggested [19,4]. registration can be considered as multiply-linked and robust. The benefits of establishing multiple-links will become clear when we compare the two algorithms in Section 5. registration is mathematically most related to the EM- algorithm [6] and the SoftAssignment algorithm [14], which are also multiply-linked. For eample, at each step, EM- minimizes the following function: ep( T (m, θ) s 2 /σ 2 ) T (m, θ) s 2 /N (m, θ). (19) m M s S where N(m, θ) = s ep( T (m, θ) s 2 /σ 2 ) is a normalization term. In fact, the G cost function has the same gradient as EM-, ecept the normalization term. Due to these mathematical similarity, registration and EM- performs very similarly, ecept that EM- does not meet the minimum requirement of a registration algorithm: The eactly aligned point sets does not correspond to the global minimum of the EM- cost function. Depending on the point sets being registered and the kernel scale, the EM- (as well as SoftAssignment) algorithms can give biased registration even for clean data. This point will be demonstrated in our eperiments. For in-depth discussion on this topic, the reader is referred to our technical report [18] (pp ). In addition, the provides a framework for using different kernel functions for registration and its convergence proof does not rely on statistical methods such as EM. 5 Performance Evaluation We compare the registration algorithm with the and EM- algorithm in this section. We implemented two versions of the algorithm. The first one is a simple 2D Matlab version that uses the Matlab fminsearch function (Nelder- Mead simple method) for optimization. In this case the gradients are not eplicitly computed. The second implementation is a C++ 3D version that computes
8 A Correlation-Based Approach to Robust Point Set Registration 565 2D range data 3D model Fig. 1. The data set for performance evaluation. gradients and uses the variable metric method [12] for optimization. The 2D Matlab code is available at webpage: ytsin/reg/. The 3D registration algorithm used for comparison is developed independently by the CMU 3D vision group ( 3dvision/). Both the 3D and our 2D code in Matlab implemented the outlier thresholding mechanism [19]. For evaluation purpose, we use a 2D range data (the road-data, 277 points) acquired by a SICK LMS 221 laser scanner and a 3D bunny model (the bunny1 - data, 699 points) acquired by a Minolta Vivid 7 scanner. The models are shown in Figure 1. Etensive registration eperiments on thousands of 2D scans and some other 3D models can be found at our website. 5.1 Convergence Region We first test the convergence properties of, EM- and in 2D. Two copies of the road data are generated by adding different random noise. One of the copies is then rotated on its center of mass for a certain angle. We study the convergence performance of the three registration methods by registering rotated point sets at different angles. The results are shown in Figure 2. The leftmost plot shows the registration costs as functions of the rotation angle. Note that we allow full 2D Euclidean motions and the cost is a 3D function. We plot a 1D slice of the cost function for clarity. With a kernel scale of σ = 15, both EM- and have very smooth cost functions in the whole test range. In the cost function we see a lot of local minima, which correspond to the much smaller convergence region in the center plot of Figure 2. The plot shows average registration error between corresponding points after registration. The EM- has a little wider convergence region than the registration in this data set. However, we observed constantly larger registration error in the EM- case. Here we eperimentally demonstrate that EM- does not meet the minimum requirement for registration. The right plot shows the average registration error as a function of the kernel scale in the noiseless case. registration has zero error regardless of the kernel scale.
9 566 Y. Tsin and T. Kanade 1.8 Cost Function EM 4 3 Average Registration Error EM 4 3 Average Registration Error EM Angle Angle Kernel scale (σ) Fig. 2. 2D convergence study. The widest convergence region (with varying outlier detection threshold) is shown here. 16 K.C K.C (a) (b) Fig. 3. Venn diagram of the sets of successfully registered model-pairs. Numbers are sizes of the regions. (a) Random transformation. (b) Pairwise. We conduct two different convergence test on the 3D data. We draw 1 random θ (6D parameter space) samples from a uniform distribution. We transform the bunny1 data set using the random parameters and form 1 pairs for registration. The Venn diagram of the successful registrations is shown in Figure 3(a). The method has larger convergence region(79 versus 6, or 24 versus 5 when ecluding the easy cases for both ). Net, we study pairwise registration of 32 scans of the bunny model acquired from different views by the laser scanner. There are in total 496 pairs of point sets to be registered. We visually eamine each of the registration results. The Venn diagram for this eperiment is shown in Figure 3(b). Again, the method has larger success rate than (17 versus 69, or 48 versus 1 ecluding the easy cases). Our eperiments show that the registration method has larger convergence region. This is due to a smoothed cost function which enables an optimization algorithm to find a good registration more easily. The smoothness is provided by weighting the contributions of the multiple-links between a point and its neighbors. 5.2 Sensitivity to Noise For both the 2D and and 3D data, we use the same method to test how sensitive the registration methods are in the presence of noise perturbation. We generate slightly rotated versions of the same point set, and add zero mean random noise
10 A Correlation-Based Approach to Robust Point Set Registration Average Shift EM Noise Level (2D) Error STD EM 1 2 Noise Level (2D) Average Shift Noise Level (3D) Error STD 1 2 Noise Level (3D) Fig. 4. Sensitivity to noise tests Registration (2D) Registration (2D) registration (3D) registration (3D) Fig. 5. Robustness tests. Note the added outlier points in all cases. to both the reference model and the rotated model. At every noise level, we register 3 pairs of noise corrupted point sets. After registration we compute the average shift between corresponding points in the two registered point sets. If the point sets are registered well, the average shift should be close to zero because the added noise has zero mean. We can thus use the standard deviation of the average shifts over the 3 pairs as a measure of sensitivity to noise. We plot the standard deviation and average shift as a function of the noise level in Figure 4. The kernel scale is 5 for the 2D tests and 2 for the 3D tests. In both (2D and 3D) cases we observe that registration has smaller variance than. At the same time, the registration error is small. The superior capability of the technique can be eplained by its etended interaction with neighbors. considered weighted effects of points in a large neighborhood, instead of just its immediate nearest neighbor, thus achieving better ability to resist noise. 5.3 Robustness To test robustness, we register outlier corrupted point sets. We generate the reference model and the rotated model the same way as the previous section. Instead of corrupting the models with noise, we add 2% of outliers. The outlier points are randomly drawn from a uniform distribution. The corrupted point sets is illustrated in Figure 5. In both the 2D and 3D cases, we use the and methods to register 1 pairs of outlier corrupted data. Eamples of 2D registration final results are presented in the left two plots of Figure 5. For we tried three outlier-detection thresholds, 2, 5 and a concatenation of 2 and 5. The best of the three, by concatenating two
11 568 Y. Tsin and T. Kanade registrations with thresholds 2 and 5, correctly registered 43 out of 1 pairs. In contrast, registration robustly registered all 1 pairs. Eemplar 3D registration final results are demonstrated in the right two plots of Figure 5. The performance of the algorithm is beyond our epectation. It failed only in 8 pairs of the outlier corrupted data sets. Still, registration can achieve better robustness. Again, registered all 1 pairs without mistake, by using a large range of different kernel scales. In our eperiments with registration in 3D, we do observe failed cases when the scale is either too small (easily fell victim to outlier distractions) or too large (containing too many outliers in the neighborhood). Thus in the presence of outliers, proper scale selection is an important and open issue in our technique. In our eperiments there are two kinds of outliers. First, the points that have a large distance to all the model points. These points are taken care of by both the M-estimator mechanism of, and the distance thresholding of. Second, the points that fall in the neighborhood of a model point. These points can be very distracting to singly-linked methods such as. For, each point is connected to multiple points. As long as the percentage of outliers in the local neighborhood is small, their influence can be averaged out by contributions from other inlier points. Consequently, is capable of registering despite of these local distractions. 6 Conclusions In this paper we introduced a registration method by dynamically configuring point sets, whose fitness is measured by. is shown to be an M-estimator. is also equivalent to an entropy measure. based registration can be considered as a robust, multiply-linked. It has a built-in smoothing mechanism that makes it very important in dealing with noise and outlier corrupted data sets. We eperimentally demonstrated that it outperforms in terms of convergence region, robustness and resistance to noise, and it outperforms EM- in terms of registration accuracy. Kernel function selection is an interesting direction. The choice of kernels determines the underlying robust function to be used. We leave it to our future research. Acknowledgment. We thank Daniel Huber and Bob Chieh-Chih Wang from Carnegie Mellon for providing the 3D and 2D test data used in this work. References 1. P. J. Besl and N. D. McKay. A method for registration of 3-D shapes. IEEE TPAMI, 14(2): , February G. Borgefors. Hierarchical chamfer matching: A parametric edge matching algorithm. IEEE TPAMI, 1(6): , November 1988.
12 A Correlation-Based Approach to Robust Point Set Registration H. Chen and P. Meer. Robust computer vision through kernel density estimation. In (ECCV 2), pages Part I, Springer-Verlag, May A. Fitsgibbon. Robust registration of 2D and 3D point sets. In BMVC 1, S. Geman and D. Geman. Stochastic relaation, Gibbs distributions, and the Bayesian restoration of images. IEEE TPAMI, 6: , S. Granger and X. Pennec. Multi-scale EM-: A fast and robust approach for surface registration. In (ECCV 2), June T. Hastie, R. Tibshirani, and J. Friedman. The Elements of Statistical Learning, Data Mining, Inference and Prediction. Springer, P. J. Huber. Robust Statistics. John Wiley & Sons, New York, New York, D. P. Huttenlocher, G.A. Klanderman, and W.J. Rucklidge. Comparing images using the hausdorff distance. IEEE TPAMI, 15(9):85 863, September S. Lavallée and R. Szeliski. Recovering the position and orientation of free-form objects from image contours using 3-D distance maps. IEEE TPAMI, 17(4):378 39, April E. Parzen. On estimation of a probability density function and mode. Annals of Mathematical Statistics, 33(3): , W. H. Press, B. P. Flannery, S. A. Teukolsky, and W. T. Vetterling. Numerical Recipes in C: The Art of Scientific Computing. Cambridge University Press, Cambridge, England, second edition, J. Principe and D. Xu. Information-theoretic learning using Renyi s quadratic entropy. In First International Workshop on Independent Component Analysis (ICA 99), pages , A. Rangarajan, H. Chui, and F.L. Bookstein. The softassign procrustes matching algorithm. Information Processing in Medical Imaging, pages 29 42, A. Rényi. On measures of entropy and information. In Proceedings of the 4th Berkeley Symposium on Mathematics, Statistics and Probability, volume 1, pages University of California Press, P.J. Rousseeuw and A.M. Leroy. Robust Regression and Outlier Detection. Wiley- Interscience, New York, New York, G.L. Scott and H.C. Longuet-Higgins. An algorithm for associating the features of two images. Proceedings: Biological Sciences, 244(139):21 26, April Y. Tsin. Kernel correlation as an affinity measure in point-sampled vision problems. Techical Report, CMU-RI-3-36, Z. Zhang. Iterative point matching for registration of free-form curves and surfaces. IJCV, 13(2), 1994.
Probabilistic Latent Semantic Analysis (plsa)
Probabilistic Latent Semantic Analysis (plsa) SS 2008 Bayesian Networks Multimedia Computing, Universität Augsburg [email protected] www.multimedia-computing.{de,org} References
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION Introduction In the previous chapter, we explored a class of regression models having particularly simple analytical
Image Segmentation and Registration
Image Segmentation and Registration Dr. Christine Tanner ([email protected]) Computer Vision Laboratory, ETH Zürich Dr. Verena Kaynig, Machine Learning Laboratory, ETH Zürich Outline Segmentation
Robert Collins CSE598G. More on Mean-shift. R.Collins, CSE, PSU CSE598G Spring 2006
More on Mean-shift R.Collins, CSE, PSU Spring 2006 Recall: Kernel Density Estimation Given a set of data samples x i ; i=1...n Convolve with a kernel function H to generate a smooth function f(x) Equivalent
Accurate and robust image superresolution by neural processing of local image representations
Accurate and robust image superresolution by neural processing of local image representations Carlos Miravet 1,2 and Francisco B. Rodríguez 1 1 Grupo de Neurocomputación Biológica (GNB), Escuela Politécnica
Simultaneous Gamma Correction and Registration in the Frequency Domain
Simultaneous Gamma Correction and Registration in the Frequency Domain Alexander Wong [email protected] William Bishop [email protected] Department of Electrical and Computer Engineering University
Linear Threshold Units
Linear Threshold Units w x hx (... w n x n w We assume that each feature x j and each weight w j is a real number (we will relax this later) We will study three different algorithms for learning linear
ACCURACY ASSESSMENT OF BUILDING POINT CLOUDS AUTOMATICALLY GENERATED FROM IPHONE IMAGES
ACCURACY ASSESSMENT OF BUILDING POINT CLOUDS AUTOMATICALLY GENERATED FROM IPHONE IMAGES B. Sirmacek, R. Lindenbergh Delft University of Technology, Department of Geoscience and Remote Sensing, Stevinweg
Component Ordering in Independent Component Analysis Based on Data Power
Component Ordering in Independent Component Analysis Based on Data Power Anne Hendrikse Raymond Veldhuis University of Twente University of Twente Fac. EEMCS, Signals and Systems Group Fac. EEMCS, Signals
Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches
Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches PhD Thesis by Payam Birjandi Director: Prof. Mihai Datcu Problematic
Segmentation & Clustering
EECS 442 Computer vision Segmentation & Clustering Segmentation in human vision K-mean clustering Mean-shift Graph-cut Reading: Chapters 14 [FP] Some slides of this lectures are courtesy of prof F. Li,
Part-Based Recognition
Part-Based Recognition Benedict Brown CS597D, Fall 2003 Princeton University CS 597D, Part-Based Recognition p. 1/32 Introduction Many objects are made up of parts It s presumably easier to identify simple
Comparison of Non-linear Dimensionality Reduction Techniques for Classification with Gene Expression Microarray Data
CMPE 59H Comparison of Non-linear Dimensionality Reduction Techniques for Classification with Gene Expression Microarray Data Term Project Report Fatma Güney, Kübra Kalkan 1/15/2013 Keywords: Non-linear
A Study on SURF Algorithm and Real-Time Tracking Objects Using Optical Flow
, pp.233-237 http://dx.doi.org/10.14257/astl.2014.51.53 A Study on SURF Algorithm and Real-Time Tracking Objects Using Optical Flow Giwoo Kim 1, Hye-Youn Lim 1 and Dae-Seong Kang 1, 1 Department of electronices
Classification of Fingerprints. Sarat C. Dass Department of Statistics & Probability
Classification of Fingerprints Sarat C. Dass Department of Statistics & Probability Fingerprint Classification Fingerprint classification is a coarse level partitioning of a fingerprint database into smaller
Bildverarbeitung und Mustererkennung Image Processing and Pattern Recognition
Bildverarbeitung und Mustererkennung Image Processing and Pattern Recognition 1. Image Pre-Processing - Pixel Brightness Transformation - Geometric Transformation - Image Denoising 1 1. Image Pre-Processing
How To Register Point Sets
Non-rigid point set registration: Coherent Point Drift Andriy Myronenko Xubo Song Miguel Á. Carreira-Perpiñán Department of Computer Science and Electrical Engineering OGI School of Science and Engineering
A Learning Algorithm For Neural Network Ensembles
A Learning Algorithm For Neural Network Ensembles H. D. Navone, P. M. Granitto, P. F. Verdes and H. A. Ceccatto Instituto de Física Rosario (CONICET-UNR) Blvd. 27 de Febrero 210 Bis, 2000 Rosario. República
An Analysis on Density Based Clustering of Multi Dimensional Spatial Data
An Analysis on Density Based Clustering of Multi Dimensional Spatial Data K. Mumtaz 1 Assistant Professor, Department of MCA Vivekanandha Institute of Information and Management Studies, Tiruchengode,
An Introduction to Machine Learning
An Introduction to Machine Learning L5: Novelty Detection and Regression Alexander J. Smola Statistical Machine Learning Program Canberra, ACT 0200 Australia [email protected] Tata Institute, Pune,
Robotics. Chapter 25. Chapter 25 1
Robotics Chapter 25 Chapter 25 1 Outline Robots, Effectors, and Sensors Localization and Mapping Motion Planning Motor Control Chapter 25 2 Mobile Robots Chapter 25 3 Manipulators P R R R R R Configuration
Metrics on SO(3) and Inverse Kinematics
Mathematical Foundations of Computer Graphics and Vision Metrics on SO(3) and Inverse Kinematics Luca Ballan Institute of Visual Computing Optimization on Manifolds Descent approach d is a ascent direction
Vision based Vehicle Tracking using a high angle camera
Vision based Vehicle Tracking using a high angle camera Raúl Ignacio Ramos García Dule Shu [email protected] [email protected] Abstract A vehicle tracking and grouping algorithm is presented in this work
Polynomial Degree and Finite Differences
CONDENSED LESSON 7.1 Polynomial Degree and Finite Differences In this lesson you will learn the terminology associated with polynomials use the finite differences method to determine the degree of a polynomial
Convex Hull Probability Depth: first results
Conve Hull Probability Depth: first results Giovanni C. Porzio and Giancarlo Ragozini Abstract In this work, we present a new depth function, the conve hull probability depth, that is based on the conve
Local outlier detection in data forensics: data mining approach to flag unusual schools
Local outlier detection in data forensics: data mining approach to flag unusual schools Mayuko Simon Data Recognition Corporation Paper presented at the 2012 Conference on Statistical Detection of Potential
Mean-Shift Tracking with Random Sampling
1 Mean-Shift Tracking with Random Sampling Alex Po Leung, Shaogang Gong Department of Computer Science Queen Mary, University of London, London, E1 4NS Abstract In this work, boosting the efficiency of
A Learning Based Method for Super-Resolution of Low Resolution Images
A Learning Based Method for Super-Resolution of Low Resolution Images Emre Ugur June 1, 2004 [email protected] Abstract The main objective of this project is the study of a learning based method
Moving Least Squares Approximation
Chapter 7 Moving Least Squares Approimation An alternative to radial basis function interpolation and approimation is the so-called moving least squares method. As we will see below, in this method the
Statistics for Analysis of Experimental Data
Statistics for Analysis of Eperimental Data Catherine A. Peters Department of Civil and Environmental Engineering Princeton University Princeton, NJ 08544 Published as a chapter in the Environmental Engineering
Social Media Mining. Data Mining Essentials
Introduction Data production rate has been increased dramatically (Big Data) and we are able store much more data than before E.g., purchase data, social media data, mobile phone data Businesses and customers
A Complete Gradient Clustering Algorithm for Features Analysis of X-ray Images
A Complete Gradient Clustering Algorithm for Features Analysis of X-ray Images Małgorzata Charytanowicz, Jerzy Niewczas, Piotr A. Kowalski, Piotr Kulczycki, Szymon Łukasik, and Sławomir Żak Abstract Methods
Least Squares Estimation
Least Squares Estimation SARA A VAN DE GEER Volume 2, pp 1041 1045 in Encyclopedia of Statistics in Behavioral Science ISBN-13: 978-0-470-86080-9 ISBN-10: 0-470-86080-4 Editors Brian S Everitt & David
Supporting Online Material for
www.sciencemag.org/cgi/content/full/313/5786/504/dc1 Supporting Online Material for Reducing the Dimensionality of Data with Neural Networks G. E. Hinton* and R. R. Salakhutdinov *To whom correspondence
Data Mining Project Report. Document Clustering. Meryem Uzun-Per
Data Mining Project Report Document Clustering Meryem Uzun-Per 504112506 Table of Content Table of Content... 2 1. Project Definition... 3 2. Literature Survey... 3 3. Methods... 4 3.1. K-means algorithm...
The Scientific Data Mining Process
Chapter 4 The Scientific Data Mining Process When I use a word, Humpty Dumpty said, in rather a scornful tone, it means just what I choose it to mean neither more nor less. Lewis Carroll [87, p. 214] In
Bootstrapping Big Data
Bootstrapping Big Data Ariel Kleiner Ameet Talwalkar Purnamrita Sarkar Michael I. Jordan Computer Science Division University of California, Berkeley {akleiner, ameet, psarkar, jordan}@eecs.berkeley.edu
Segmentation of building models from dense 3D point-clouds
Segmentation of building models from dense 3D point-clouds Joachim Bauer, Konrad Karner, Konrad Schindler, Andreas Klaus, Christopher Zach VRVis Research Center for Virtual Reality and Visualization, Institute
An Iterative Image Registration Technique with an Application to Stereo Vision
An Iterative Image Registration Technique with an Application to Stereo Vision Bruce D. Lucas Takeo Kanade Computer Science Department Carnegie-Mellon University Pittsburgh, Pennsylvania 15213 Abstract
15.062 Data Mining: Algorithms and Applications Matrix Math Review
.6 Data Mining: Algorithms and Applications Matrix Math Review The purpose of this document is to give a brief review of selected linear algebra concepts that will be useful for the course and to develop
203.4770: Introduction to Machine Learning Dr. Rita Osadchy
203.4770: Introduction to Machine Learning Dr. Rita Osadchy 1 Outline 1. About the Course 2. What is Machine Learning? 3. Types of problems and Situations 4. ML Example 2 About the course Course Homepage:
Canny Edge Detection
Canny Edge Detection 09gr820 March 23, 2009 1 Introduction The purpose of edge detection in general is to significantly reduce the amount of data in an image, while preserving the structural properties
ASSESSMENT OF VISUALIZATION SOFTWARE FOR SUPPORT OF CONSTRUCTION SITE INSPECTION TASKS USING DATA COLLECTED FROM REALITY CAPTURE TECHNOLOGIES
ASSESSMENT OF VISUALIZATION SOFTWARE FOR SUPPORT OF CONSTRUCTION SITE INSPECTION TASKS USING DATA COLLECTED FROM REALITY CAPTURE TECHNOLOGIES ABSTRACT Chris Gordon 1, Burcu Akinci 2, Frank Boukamp 3, and
Wii Remote Calibration Using the Sensor Bar
Wii Remote Calibration Using the Sensor Bar Alparslan Yildiz Abdullah Akay Yusuf Sinan Akgul GIT Vision Lab - http://vision.gyte.edu.tr Gebze Institute of Technology Kocaeli, Turkey {yildiz, akay, akgul}@bilmuh.gyte.edu.tr
3D Model based Object Class Detection in An Arbitrary View
3D Model based Object Class Detection in An Arbitrary View Pingkun Yan, Saad M. Khan, Mubarak Shah School of Electrical Engineering and Computer Science University of Central Florida http://www.eecs.ucf.edu/
Local classification and local likelihoods
Local classification and local likelihoods November 18 k-nearest neighbors The idea of local regression can be extended to classification as well The simplest way of doing so is called nearest neighbor
Models of Cortical Maps II
CN510: Principles and Methods of Cognitive and Neural Modeling Models of Cortical Maps II Lecture 19 Instructor: Anatoli Gorchetchnikov dy dt The Network of Grossberg (1976) Ay B y f (
Palmprint Recognition. By Sree Rama Murthy kora Praveen Verma Yashwant Kashyap
Palmprint Recognition By Sree Rama Murthy kora Praveen Verma Yashwant Kashyap Palm print Palm Patterns are utilized in many applications: 1. To correlate palm patterns with medical disorders, e.g. genetic
Jiří Matas. Hough Transform
Hough Transform Jiří Matas Center for Machine Perception Department of Cybernetics, Faculty of Electrical Engineering Czech Technical University, Prague Many slides thanks to Kristen Grauman and Bastian
Applications to Data Smoothing and Image Processing I
Applications to Data Smoothing and Image Processing I MA 348 Kurt Bryan Signals and Images Let t denote time and consider a signal a(t) on some time interval, say t. We ll assume that the signal a(t) is
Epipolar Geometry. Readings: See Sections 10.1 and 15.6 of Forsyth and Ponce. Right Image. Left Image. e(p ) Epipolar Lines. e(q ) q R.
Epipolar Geometry We consider two perspective images of a scene as taken from a stereo pair of cameras (or equivalently, assume the scene is rigid and imaged with a single camera from two different locations).
CS 534: Computer Vision 3D Model-based recognition
CS 534: Computer Vision 3D Model-based recognition Ahmed Elgammal Dept of Computer Science CS 534 3D Model-based Vision - 1 High Level Vision Object Recognition: What it means? Two main recognition tasks:!
Cluster Analysis: Advanced Concepts
Cluster Analysis: Advanced Concepts and dalgorithms Dr. Hui Xiong Rutgers University Introduction to Data Mining 08/06/2006 1 Introduction to Data Mining 08/06/2006 1 Outline Prototype-based Fuzzy c-means
Support Vector Machine. Tutorial. (and Statistical Learning Theory)
Support Vector Machine (and Statistical Learning Theory) Tutorial Jason Weston NEC Labs America 4 Independence Way, Princeton, USA. [email protected] 1 Support Vector Machines: history SVMs introduced
Syllabus for MATH 191 MATH 191 Topics in Data Science: Algorithms and Mathematical Foundations Department of Mathematics, UCLA Fall Quarter 2015
Syllabus for MATH 191 MATH 191 Topics in Data Science: Algorithms and Mathematical Foundations Department of Mathematics, UCLA Fall Quarter 2015 Lecture: MWF: 1:00-1:50pm, GEOLOGY 4645 Instructor: Mihai
Assessment. Presenter: Yupu Zhang, Guoliang Jin, Tuo Wang Computer Vision 2008 Fall
Automatic Photo Quality Assessment Presenter: Yupu Zhang, Guoliang Jin, Tuo Wang Computer Vision 2008 Fall Estimating i the photorealism of images: Distinguishing i i paintings from photographs h Florin
PHOTOGRAMMETRIC TECHNIQUES FOR MEASUREMENTS IN WOODWORKING INDUSTRY
PHOTOGRAMMETRIC TECHNIQUES FOR MEASUREMENTS IN WOODWORKING INDUSTRY V. Knyaz a, *, Yu. Visilter, S. Zheltov a State Research Institute for Aviation System (GosNIIAS), 7, Victorenko str., Moscow, Russia
A Genetic Algorithm-Evolved 3D Point Cloud Descriptor
A Genetic Algorithm-Evolved 3D Point Cloud Descriptor Dominik Wȩgrzyn and Luís A. Alexandre IT - Instituto de Telecomunicações Dept. of Computer Science, Univ. Beira Interior, 6200-001 Covilhã, Portugal
High Quality Image Magnification using Cross-Scale Self-Similarity
High Quality Image Magnification using Cross-Scale Self-Similarity André Gooßen 1, Arne Ehlers 1, Thomas Pralow 2, Rolf-Rainer Grigat 1 1 Vision Systems, Hamburg University of Technology, D-21079 Hamburg
How To Understand The Theory Of Probability
Graduate Programs in Statistics Course Titles STAT 100 CALCULUS AND MATR IX ALGEBRA FOR STATISTICS. Differential and integral calculus; infinite series; matrix algebra STAT 195 INTRODUCTION TO MATHEMATICAL
Example: Credit card default, we may be more interested in predicting the probabilty of a default than classifying individuals as default or not.
Statistical Learning: Chapter 4 Classification 4.1 Introduction Supervised learning with a categorical (Qualitative) response Notation: - Feature vector X, - qualitative response Y, taking values in C
Adaptive Online Gradient Descent
Adaptive Online Gradient Descent Peter L Bartlett Division of Computer Science Department of Statistics UC Berkeley Berkeley, CA 94709 bartlett@csberkeleyedu Elad Hazan IBM Almaden Research Center 650
Visualization of Breast Cancer Data by SOM Component Planes
International Journal of Science and Technology Volume 3 No. 2, February, 2014 Visualization of Breast Cancer Data by SOM Component Planes P.Venkatesan. 1, M.Mullai 2 1 Department of Statistics,NIRT(Indian
Data Mining. Cluster Analysis: Advanced Concepts and Algorithms
Data Mining Cluster Analysis: Advanced Concepts and Algorithms Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 1 More Clustering Methods Prototype-based clustering Density-based clustering Graph-based
Character Image Patterns as Big Data
22 International Conference on Frontiers in Handwriting Recognition Character Image Patterns as Big Data Seiichi Uchida, Ryosuke Ishida, Akira Yoshida, Wenjie Cai, Yaokai Feng Kyushu University, Fukuoka,
New Ensemble Combination Scheme
New Ensemble Combination Scheme Namhyoung Kim, Youngdoo Son, and Jaewook Lee, Member, IEEE Abstract Recently many statistical learning techniques are successfully developed and used in several areas However,
Genetic algorithm for affine point pattern matching
Pattern Recognition Letters 24 (2003) 9 19 www.elsevier.com/locate/patrec Genetic algorithm for affine point pattern matching Lihua Zhang, Wenli Xu *, Cheng Chang Department of Automation, Tsinghua University,
Fast and Robust Normal Estimation for Point Clouds with Sharp Features
1/37 Fast and Robust Normal Estimation for Point Clouds with Sharp Features Alexandre Boulch & Renaud Marlet University Paris-Est, LIGM (UMR CNRS), Ecole des Ponts ParisTech Symposium on Geometry Processing
Face Model Fitting on Low Resolution Images
Face Model Fitting on Low Resolution Images Xiaoming Liu Peter H. Tu Frederick W. Wheeler Visualization and Computer Vision Lab General Electric Global Research Center Niskayuna, NY, 1239, USA {liux,tu,wheeler}@research.ge.com
INTRODUCTION TO ERRORS AND ERROR ANALYSIS
INTRODUCTION TO ERRORS AND ERROR ANALYSIS To many students and to the public in general, an error is something they have done wrong. However, in science, the word error means the uncertainty which accompanies
COS702; Assignment 6. Point Cloud Data Surface Interpolation University of Southern Missisippi Tyler Reese December 3, 2012
COS702; Assignment 6 Point Cloud Data Surface Interpolation University of Southern Missisippi Tyler Reese December 3, 2012 The Problem COS 702, Assignment 6: Given appropriate sets of Point Cloud data,
Subspace Analysis and Optimization for AAM Based Face Alignment
Subspace Analysis and Optimization for AAM Based Face Alignment Ming Zhao Chun Chen College of Computer Science Zhejiang University Hangzhou, 310027, P.R.China [email protected] Stan Z. Li Microsoft
3D Human Face Recognition Using Point Signature
3D Human Face Recognition Using Point Signature Chin-Seng Chua, Feng Han, Yeong-Khing Ho School of Electrical and Electronic Engineering Nanyang Technological University, Singapore 639798 [email protected]
Machine Learning Final Project Spam Email Filtering
Machine Learning Final Project Spam Email Filtering March 2013 Shahar Yifrah Guy Lev Table of Content 1. OVERVIEW... 3 2. DATASET... 3 2.1 SOURCE... 3 2.2 CREATION OF TRAINING AND TEST SETS... 4 2.3 FEATURE
DATA MINING CLUSTER ANALYSIS: BASIC CONCEPTS
DATA MINING CLUSTER ANALYSIS: BASIC CONCEPTS 1 AND ALGORITHMS Chiara Renso KDD-LAB ISTI- CNR, Pisa, Italy WHAT IS CLUSTER ANALYSIS? Finding groups of objects such that the objects in a group will be similar
Convolution. 1D Formula: 2D Formula: Example on the web: http://www.jhu.edu/~signals/convolve/
Basic Filters (7) Convolution/correlation/Linear filtering Gaussian filters Smoothing and noise reduction First derivatives of Gaussian Second derivative of Gaussian: Laplacian Oriented Gaussian filters
Practical Tour of Visual tracking. David Fleet and Allan Jepson January, 2006
Practical Tour of Visual tracking David Fleet and Allan Jepson January, 2006 Designing a Visual Tracker: What is the state? pose and motion (position, velocity, acceleration, ) shape (size, deformation,
Lecture 8: More Continuous Random Variables
Lecture 8: More Continuous Random Variables 26 September 2005 Last time: the eponential. Going from saying the density e λ, to f() λe λ, to the CDF F () e λ. Pictures of the pdf and CDF. Today: the Gaussian
siftservice.com - Turning a Computer Vision algorithm into a World Wide Web Service
siftservice.com - Turning a Computer Vision algorithm into a World Wide Web Service Ahmad Pahlavan Tafti 1, Hamid Hassannia 2, and Zeyun Yu 1 1 Department of Computer Science, University of Wisconsin -Milwaukee,
Clustering. Data Mining. Abraham Otero. Data Mining. Agenda
Clustering 1/46 Agenda Introduction Distance K-nearest neighbors Hierarchical clustering Quick reference 2/46 1 Introduction It seems logical that in a new situation we should act in a similar way as in
Enhancing the SNR of the Fiber Optic Rotation Sensor using the LMS Algorithm
1 Enhancing the SNR of the Fiber Optic Rotation Sensor using the LMS Algorithm Hani Mehrpouyan, Student Member, IEEE, Department of Electrical and Computer Engineering Queen s University, Kingston, Ontario,
An Overview of Knowledge Discovery Database and Data mining Techniques
An Overview of Knowledge Discovery Database and Data mining Techniques Priyadharsini.C 1, Dr. Antony Selvadoss Thanamani 2 M.Phil, Department of Computer Science, NGM College, Pollachi, Coimbatore, Tamilnadu,
How To Fix Out Of Focus And Blur Images With A Dynamic Template Matching Algorithm
IJSTE - International Journal of Science Technology & Engineering Volume 1 Issue 10 April 2015 ISSN (online): 2349-784X Image Estimation Algorithm for Out of Focus and Blur Images to Retrieve the Barcode
K-Means Cluster Analysis. Tan,Steinbach, Kumar Introduction to Data Mining 4/18/2004 1
K-Means Cluster Analsis Chapter 3 PPDM Class Tan,Steinbach, Kumar Introduction to Data Mining 4/18/4 1 What is Cluster Analsis? Finding groups of objects such that the objects in a group will be similar
Visualization by Linear Projections as Information Retrieval
Visualization by Linear Projections as Information Retrieval Jaakko Peltonen Helsinki University of Technology, Department of Information and Computer Science, P. O. Box 5400, FI-0015 TKK, Finland [email protected]
Fast Multipole Method for particle interactions: an open source parallel library component
Fast Multipole Method for particle interactions: an open source parallel library component F. A. Cruz 1,M.G.Knepley 2,andL.A.Barba 1 1 Department of Mathematics, University of Bristol, University Walk,
Face Recognition in Low-resolution Images by Using Local Zernike Moments
Proceedings of the International Conference on Machine Vision and Machine Learning Prague, Czech Republic, August14-15, 014 Paper No. 15 Face Recognition in Low-resolution Images by Using Local Zernie
Adaptive Equalization of binary encoded signals Using LMS Algorithm
SSRG International Journal of Electronics and Communication Engineering (SSRG-IJECE) volume issue7 Sep Adaptive Equalization of binary encoded signals Using LMS Algorithm Dr.K.Nagi Reddy Professor of ECE,NBKR
Path Tracking for a Miniature Robot
Path Tracking for a Miniature Robot By Martin Lundgren Excerpt from Master s thesis 003 Supervisor: Thomas Hellström Department of Computing Science Umeå University Sweden 1 Path Tracking Path tracking
Feature Point Selection using Structural Graph Matching for MLS based Image Registration
Feature Point Selection using Structural Graph Matching for MLS based Image Registration Hema P Menon Department of CSE Amrita Vishwa Vidyapeetham Coimbatore Tamil Nadu - 641 112, India K A Narayanankutty
Multi-Resolution Spin-Images
Multi-Resolution Spin-Images H. Quynh Dinh Department of Computer Science Stevens Institute of Technology [email protected] Steven Kropac Department of Computer Science Stevens Institute of Technology
Least Squares Approach for Initial Data Recovery in Dynamic
Computing and Visualiation in Science manuscript No. (will be inserted by the editor) Least Squares Approach for Initial Data Recovery in Dynamic Data-Driven Applications Simulations C. Douglas,2, Y. Efendiev
Machine Learning and Pattern Recognition Logistic Regression
Machine Learning and Pattern Recognition Logistic Regression Course Lecturer:Amos J Storkey Institute for Adaptive and Neural Computation School of Informatics University of Edinburgh Crichton Street,
Statistical Machine Learning
Statistical Machine Learning UoC Stats 37700, Winter quarter Lecture 4: classical linear and quadratic discriminants. 1 / 25 Linear separation For two classes in R d : simple idea: separate the classes
Pattern Analysis. Logistic Regression. 12. Mai 2009. Joachim Hornegger. Chair of Pattern Recognition Erlangen University
Pattern Analysis Logistic Regression 12. Mai 2009 Joachim Hornegger Chair of Pattern Recognition Erlangen University Pattern Analysis 2 / 43 1 Logistic Regression Posteriors and the Logistic Function Decision
