Camera geometry and image alignment
|
|
|
- Lorena Richards
- 9 years ago
- Views:
Transcription
1 Computer Vision and Machine Learning Winter School ENS Lyon 2010 Camera geometry and image alignment Josef Sivic INRIA, WILLOW, ENS/INRIA/CNRS UMR 8548 Laboratoire d Informatique, Ecole Normale Supérieure, Paris With slides from: O. Chum, K. Grauman, S. Lazebnik, B. Leibe, D. Lowe, J. Philbin, J. Ponce, D. Nister, C. Schmid, N. Snavely, A. Zisserman
2 Outline Part I - Camera geometry image formation Perspective projection Affine projection Projection of planes Part II - Image matching and recognition with local features Correspondence Semi-local and global geometric relations Robust estimation RANSAC and Hough Transform
3 Motivation: Stitching panoramas
4 Feature-based alignment outline
5 Feature-based alignment outline Extract features
6 Feature-based alignment outline Extract features Compute putative matches
7 Feature-based alignment outline Extract features Compute putative matches Loop: Hypothesize transformation T (small group of putative matches that are related by T)
8 Feature-based alignment outline Extract features Compute putative matches Loop: Hypothesize transformation T (small group of putative matches that are related by T) Verify transformation (search for other matches consistent with T)
9 Feature-based alignment outline Extract features Compute putative matches Loop: Hypothesize transformation T (small group of putative matches that are related by T) Verify transformation (search for other matches consistent with T)
10 2D transformation models Similarity (translation, scale, rotation) Affine Projective (homography) Why these transformations???
11 Camera geometry
12 They are formed by the projection of 3D objects. Images are two-dimensional patterns of brightness values.
13 Animal eye: a looonnng time ago. Photographic camera: Niepce, Pinhole perspective projection: Brunelleschi, XV th Century. Camera obscura: XVI th Century.
14
15 Pompei painting, 2000 years ago. Brunelleschi, 1415 Van Eyk, XIV th Century Massaccio s Trinity, 1425
16 Pinhole Perspective Equation Image plane (retina) Camera coordinate system World point Imaged point Camera center Principal axis P = (x,y,z) T P'= (x', y') T y' f ' O z y NOTE: z is always negative..
17 Affine projection models: Weak perspective projection is the magnification. When the scene relief is small compared its distance from the Camera, m can be taken constant: weak perspective projection.
18 Affine projection models: Orthographic projection When the camera is at a (roughly constant) distance from the scene, take m=1.
19 Strong perspective: Angles are not preserved The projections of parallel lines intersect at one point
20 From Zisserman & Hartley
21 Strong perspective: Angles are not preserved The projections of parallel lines intersect at one point Weak perspective: Angles are better preserved The projections of parallel lines are (almost) parallel
22 Beyond pinhole camera model: Geometric Distortion
23 Rectification
24 Radial Distortion Model
25 Perspective Projection Weak-Perspective Projection (Affine) Orthographic Projection (Affine) Common distortion model x,y: World coordinates x,y : Image coordinates f: pinhole-to-retina distance x,y: World coordinates x,y : Image coordinates m: magnification x,y: World coordinates x,y : Image coordinates x,y : Ideal image coordinates x,y : Actual image coordinates
26 Cameras and their parameters Images from M. Pollefeys
27 The Intrinsic Parameters of a Camera Units: k,l : pixel/m f : m α,β : pixel Normalized Image Coordinates Physical Image Coordinates
28 The Intrinsic Parameters of a Camera Calibration Matrix The Perspective Projection Equation
29 Notation Euclidean Geometry
30 Recall: Coordinate Changes and Rigid Transformations
31 The Extrinsic Parameters of a Camera W x u v = 1 m 12 m 12 m 13 m 14 m 21 m 22 m 23 m W y 24 z W z 1 m 31 m 32 m 33 m 34 1 W P = W x W y W z
32 Explicit Form of the Projection Matrix Note: M is only defined up to scale in this setting!!
33 Weak perspective (affine) camera z r = m 3 T P = const. u v = 1 z 1 r m 1 T m 2 T m 3 T m T 1 P /m T 3 P P = m T 2 P /m T 3 P m T 3 P /m T 3 P u = 1 T m 1 T P v z r m 2 u = a 11 a 12 a 13 b 1 v a 21 a 22 a 23 b 2 A 2 3 b 2 1 W x W y W z 1 =A W P + b
34 Re-cap: imaging and camera geometry (with a slight change of notation) perspective projection camera centre, image point and scene point are collinear an image point back projects to a ray in 3-space C x X image point scene point depth of the scene point is unknown camera centre image plane Slide credit: A. Zisserman
35 Slide credit: A. Zisserman
36 How a scene plane projects into an image?
37 Plane projective transformations Slide credit: A. Zisserman
38 Projective transformations continued This is the most general transformation between the world and image plane under imaging by a perspective camera. It is often only the 3 x 3 form of the matrix that is important in establishing properties of this transformation. A projective transformation is also called a ``homography'' and a ``collineation''. H has 8 degrees of freedom. Slide credit: A. Zisserman
39 Planes under affine projection x 1 x 2 x = a 11 a 12 a 13 b 1 y a 21 a 22 a 23 b = a 11 a 12 x + b 1 = A a 21 a 22 y b 2 2 P + b Points on a world plane map with a 2D affine geometric transformation (6 parameters)
40 Summary Affine projections induce affine transformations from planes onto their images. Perspective projections induce projective transformations from planes onto their images.
41 2D transformation models Similarity (translation, scale, rotation) Affine Projective (homography)
42 When is homography a valid transformation model?
43 Case I: Plane projective transformations Slide credit: A. Zisserman
44 Case II: Cameras rotating about their centre image plane 2 image plane 1 P = K [ I 0 ] P = K [ R 0 ] H = K R K^(-1) The two image planes are related by a homography H H depends only on the relation between the image planes and camera centre, C, not on the 3D structure Slide credit: A. Zisserman
45 Case II: Cameras rotating about their centre image plane 2 image plane 1 P = K[I 0] P'= K'[R 0] x = K[I 0] X = KX X = K 1 x 1 x'= K'[R 0] X = K'RX 1 x'= K'RK 1 x H Slide credit: A. Zisserman
46
47 Outline the rest of the lecture Part 2. Image matching and recognition with local features Correspondence Semi-local and global geometric relations Robust estimation RANSAC and Hough Transform
48 Image matching and recognition with local features The goal: establish correspondence between two or more images P x X P/ x' in correspondence if they are Image points x and x are X. projections of the same 3D scene point Images courtesy A. Zisserman
49 Example I: Wide baseline matching Establish correspondence between two (or more) images. Useful in visual geometry: Camera calibration, 3D reconstruction, Structure and motion estimation, Scale/affine invariant regions: SIFT, Harris-Laplace, etc.
50 Example II: Object recognition Establish correspondence between the target image and (multiple) images in the model database. Model database Target image [D. Lowe, 1999]
51 Example III: Visual search Given a query image, find images depicting the same place / object in a large unordered image collection. Find these landmarks...in these images and 1M more
52 Establish correspondence between the query image and all images from the database depicting the same object / scene. Query image Database image(s)
53 Why is it difficult? Want to establish correspondence despite possibly large changes in scale, viewpoint, lighting and partial occlusion Scale Viewpoint Lighting Occlusion and the image collection can be very large (e.g. 1M images)
54 Approach Pre-processing (last lecture): Detect local features. Extract descriptor for each feature. Matching: 1. Establish tentative (putative) correspondences based on local appearance of individual features (their descriptors). 2. Verify matches based on semi-local / global geometric relations.
55 Example I: Two images - Where is the Graffiti? object
56 Step 1. Establish tentative correspondence Establish tentative correspondences between object model image and target image by nearest neighbour matching on SIFT vectors Model (query) image 128D descriptor space Target image Need to solve some variant of the nearest neighbor problem for all feature vectors,, in the query image: where,, are features in the target image. Can take a long time if many target images are considered.
57 Problem with matching on local descriptors alone too much individual invariance each region can affine deform independently (by different amounts) Locally appearance can be ambiguous Solution: use semi-local and global spatial relations to verify matches.
58 Example I: Two images - Where is the Graffiti? Initial matches Nearest-neighbor search based on appearance descriptors alone. After spatial verification
59 Step 2: Spatial verification (now) a. Semi-local constraints Constraints on spatially close-by matches b. Global geometric relations Require a consistent global relationship between all matches
60 Semi-local constraints: Example I. neighbourhood consensus [Schmid&Mohr, PAMI 1997]
61 Semi-local constraints: Example I. neighbourhood consensus Original images Tentative matches [Schaffalitzky & Zisserman, CIVR 2004] After neighbourhood consensus
62 Semi-local constraints: Example II. Model image Matched image [Ferrari et al., IJCV 2005] Matched image
63 Geometric verification with global constraints All matches must be consistent with a global geometric relation / transformation. Need to simultaneously (i) estimate the geometric relation / transformation and (ii) the set of consistent matches Tentative matches Matches consistent with an affine transformation
64 Epipolar geometry (not considered here) In general, two views of a 3D scene are related by the epipolar constraint.? C epipole C / baseline A point in one view generates an epipolar line in the other view The corresponding point lies on this line. Slide credit: A. Zisserman
65 Epipolar geometry (not considered here) Epipolar geometry is a consequence of the coplanarity of the camera centres and scene point X x x / C C / The camera centres, corresponding points and scene point lie in a single plane, known as the epipolar plane Slide credit: A. Zisserman
66 Epipolar geometry (not considered here) Algebraically, the epipolar constraint can be expressed as X x x / where C C / x, x are homogeneous coordinates (3-vectors) of corresponding image points. F is a 3x3, rank 2 homogeneous matrix with 7 degrees of freedom, called the fundamental matrix. Slide credit: A. Zisserman
67 3D constraint: example (not considered here) Matches must be consistent with a 3D model 3 (out of 20) images used to build the 3D model Recovered 3D model Object recognized in a previously unseen pose Recovered pose [Lazebnik, Rothganger, Schmid, Ponce, CVPR 03]
68 3D constraint: example (not considered here) With a given 3D model (set of known X s) and a set of measured image points x, the goal is to find find camera matrix P and a set of geometrically consistent correspondences x X. P X x C
69 2D transformation models Similarity (translation, scale, rotation) Affine Projective (homography)
70 Example: estimating 2D affine transformation Simple fitting procedure (linear least squares) Approximates viewpoint changes for roughly planar objects and roughly orthographic cameras Can be used to initialize fitting for more complex models
71 Example: estimating 2D affine transformation Simple fitting procedure (linear least squares) Approximates viewpoint changes for roughly planar objects and roughly orthographic cameras Can be used to initialize fitting for more complex models Matches consistent with an affine transformation
72 Fitting an affine transformation Assume we know the correspondences, how do we get the transformation?
73 Fitting an affine transformation x i y i x i y i 0 1 m 1 m 2 m 3 m 4 t 1 t 2 x = i y i Linear system with six unknowns Each match gives us two linearly independent equations: need at least three to solve for the transformation parameters
74 Dealing with outliers The set of putative matches may contain a high percentage (e.g. 90%) of outliers How do we fit a geometric transformation to a small subset of all possible matches? Possible strategies: RANSAC Hough transform
75 Strategy 1: RANSAC RANSAC loop (Fischler & Bolles, 1981): Randomly select a seed group of matches Compute transformation from seed group Find inliers to this transformation If the number of inliers is sufficiently large, re-compute least-squares estimate of transformation on all of the inliers Keep the transformation with the largest number of inliers
76 Example: Robust line estimation - RANSAC Fit a line to 2D data containing outliers There are two problems 1. a line fit which minimizes perpendicular distance 2. a classification into inliers (valid points) and outliers Solution: use robust statistical estimation algorithm RANSAC (RANdom Sample Consensus) [Fishler & Bolles, 1981] Slide credit: A. Zisserman
77 RANSAC robust line estimation Repeat 1. Select random sample of 2 points 2. Compute the line through these points 3. Measure support (number of points within threshold distance of the line) Choose the line with the largest number of inliers Compute least squares fit of line to inliers (regression) Slide credit: A. Zisserman
78 Slide credit: O. Chum
79 Slide credit: O. Chum
80 Slide credit: O. Chum
81 Slide credit: O. Chum
82 Slide credit: O. Chum
83 Slide credit: O. Chum
84 Slide credit: O. Chum
85 Slide credit: O. Chum
86 Slide credit: O. Chum
87 Algorithm summary RANSAC robust estimation of 2D affine transformation Repeat 1. Select 3 point to point correspondences 2. Compute H (2x2 matrix) + t (2x1) vector for translation 3. Measure support (number of inliers within threshold distance, i.e. d 2 transfer < t) Choose the (H,t) with the largest number of inliers (Re-estimate (H,t) from all inliers)
88 How many samples? Number of samples N Choose N so that, with probability p, at least one random sample is free from outliers e.g.: > p=0.99 > outlier ratio: e Probability a randomly picked point is an inlier Probability of all points in a sample (of size s) are inliers Source: M. Pollefeys
89 How many samples? Number of samples N Choose N so that, with probability p, at least one random sample is free from outliers e.g.: > p=0.99 > outlier ratio: e Probability that all N samples (of size s) are corrupted (contain an outlier) Probability of at least one point in a sample (of size s) is an outlier proportion of outliers e s 5% 10% 20% 30% 40% 50% 90% e e e e e8 Source: M. Pollefeys
90 Example: line fitting p = 0.99 s =? e =? N =? Source: M. Pollefeys
91 Example: line fitting p = 0.99 s = 2 e = 2/10 = 0.2 N = 5 Compare with exhaustively trying all point pairs: 10 2 = 10*9 / 2 = 45 proportion of outliers e s 5% 10% 20% 30% 40% 50% 90% e e e e e8 Source: M. Pollefeys
92 How to reduce the number of samples needed? 1. Reduce the proportion of outliers. 2. Reduce the sample size use simpler model (e.g. similarity instead of affine tnf.) use local information (e.g. a region to region correspondence is equivalent to (up to) 3 point to point correspondences). Number of samples N Region to region correspondence proportion of outliers e s 5% 10% 20% 30% 40% 50% 90% e e e e e8
93 RANSAC (references) M. Fischler and R. Bolles, Random Sample Consensus: A Paradigm for Model Fitting with Applications to Image Analysis and Automated Cartography, Comm. ACM, 1981 R. Hartley and A. Zisserman, Multiple View Geometry in Computer Vision, 2 nd ed., Extensions: B. Tordoff and D. Murray, Guided Sampling and Consensus for Motion Estimation, ECCV 03 D. Nister, Preemptive RANSAC for Live Structure and Motion Estimation, ICCV 03 Chum, O.; Matas, J. and Obdrzalek, S.: Enhancing RANSAC by Generalized Model Optimization, ACCV 04 Chum, O.; and Matas, J.: Matching with PROSAC - Progressive Sample Consensus, CVPR 2005 Philbin, J., Chum, O., Isard, M., Sivic, J. and Zisserman, A.: Object retrieval with large vocabularies and fast spatial matching, CVPR 07 Chum, O. and Matas. J.: Optimal Randomized RANSAC, PAMI 08
94 Strategy 2: Hough Transform Origin: Detection of straight lines in cluttered images Can be generalized to arbitrary shapes Can extract feature groupings from cluttered images in linear time. Illustrate on extracting sets of local features consistent with a similarity transformation.
95 Hough transform for object recognition Suppose our features are scale- and rotation-covariant Then a single feature match provides an alignment hypothesis (translation, scale, orientation) model Target image David G. Lowe. Distinctive image features from scaleinvariant keypoints, IJCV 60 (2), pp , 2004.
96 Hough transform for object recognition Suppose our features are scale- and rotation-covariant Then a single feature match provides an alignment hypothesis (translation, scale, orientation) Of course, a hypothesis obtained from a single match is unreliable Solution: Coarsely quantize the transformation space. Let each match vote for its hypothesis in the quantized space. model David G. Lowe. Distinctive image features from scaleinvariant keypoints, IJCV 60 (2), pp , 2004.
97 Basic algorithm outline 1. Initialize accumulator H to all zeros 2. For each tentative match compute transformation hypothesis: tx, ty, s, θ H(tx,ty,s,θ) = H(tx,ty,s,θ) + 1 end end 3. Find all bins (tx,ty,s,θ) where H(tx,ty,s,θ) has at least three votes Correct matches will consistently vote for the same transformation while mismatches will spread votes. Cost: Linear scan through the matches (step 2), followed by a linear scan through the accumulator (step 3). tx H: 4D-accumulator array (only 2-d shown here) ty
98 Hough transform details (D. Lowe s system) Training phase: For each model feature, record 2D location, scale, and orientation of model (relative to normalized feature frame) Test phase: Let each match between a test and a model feature vote in a 4D Hough space Use broad bin sizes of 30 degrees for orientation, a factor of 2 for scale, and 0.25 times image size for location Vote for two closest bins in each dimension Find all bins with at least three votes and perform geometric verification Estimate least squares affine transformation Use stricter thresholds on transformation residual Search for additional features that agree with the alignment
99 Hough transform in object recognition (references) P.V.C. Hough, Machine Analysis of Bubble Chamber Pictures, Proc. Int. Conf. High Energy Accelerators and Instrumentation, 1959 D. Lowe, Distinctive image features from scale-invariant keypoints, IJCV 60 (2), H. Jegou, M. Douze, C. Schmid, Hamming embedding and weak geometric consistency for large scale image search, ECCV 2008 Extensions (object category detection): B. Leibe, A. Leonardis, and B. Schiele., Combined Object Categorization and Segmentation with an Implicit Shape Model, in ECCV'04 Workshop on Statistical Learning in Computer Vision, Prague, May S. Maji and J. Malik, Object Detection Using a Max-Margin Hough Tranform, CVPR 2009 A. Lehmann, B. Leibe, L. Van Gool. Fast PRISM: Branch and Bound Hough Transform for Object Class Detection, IJCV (to appear), O. Barinova, V. Lempitsky, P. Kohli, On the Detection of Multiple Object Instances using Hough Transforms, CVPR, 2010
100 Comparison Hough Transform Advantages Can handle high percentage of outliers (>95%) Extracts groupings from clutter in linear time Disadvantages Quantization issues Only practical for small number of dimensions (up to 4) Improvements available Probabilistic Extensions Continuous Voting Space Can be generalized to arbitrary shapes and objects [Leibe08] RANSAC Advantages General method suited to large range of problems Easy to implement Independent of number of dimensions Disadvantages Basic version only handles moderate number of outliers (<50%) Many variants available, e.g. PROSAC: Progressive RANSAC [Chum05] Preemptive RANSAC [Nister05] Slide credit: K. Grauman, B. Leibe
101 Beyond affine transformations What is the transformation between two views of a planar surface? What is the transformation between images from two cameras that share the same center?
102 Beyond affine transformations Homography: plane projective transformation (transformation taking a quad to another arbitrary quad)
103 Case II: Cameras rotating about their centre image plane 2 image plane 1 P = K [ I 0 ] P = K [ R 0 ] H = K R K^(-1) The two image planes are related by a homography H H depends only on the relation between the image planes and camera centre, C, not on the 3D structure
104 Fitting a homography Recall: homogenenous coordinates Converting to homogenenous image coordinates Converting from homogenenous image coordinates
105 Fitting a homography Recall: homogenenous coordinates Equation for homography: Converting to homogenenous image coordinates Converting from homogenenous image coordinates λ x y 1 = h 11 h 12 h 13 h 21 h 22 h 23 h 31 h 32 h 33 x y 1
106 Fitting a homography Equation for homography: x i λ i y i = 1 h 11 h 12 h 13 h 21 h 22 h 23 h 31 h 32 h 33 x i y i 1 λ i x i = H x i = h 1 T h 2 T h 3 T x i 9 entries, 8 degrees of freedom (scale is arbitrary) x i H x i = 0 3 equations, only 2 linearly independent
107 Direct linear transform H has 8 degrees of freedom (9 parameters, but scale is arbitrary) One match gives us two linearly independent equations Four matches needed for a minimal solution (null space of 8x9 matrix) More than four: homogeneous least squares
108 Application: Panorama stitching Images courtesy of A. Zisserman.
109 Recognizing panoramas Given contents of a camera memory card, automatically figure out which pictures go together and stitch them together into panoramas M. Brown and D. Lowe, Recognizing panoramas, ICCV 2003.
110 1. Estimate homography (RANSAC)
111 1. Estimate homography (RANSAC)
112 1. Estimate homography (RANSAC)
113 2. Find connected sets of images
114 2. Find connected sets of images
115 2. Find connected sets of images
116 3. Stitch and blend the panoramas
117 Results
118 M. Brown, D. Lowe, B. Hearn, J. Beis
Jiří Matas. Hough Transform
Hough Transform Jiří Matas Center for Machine Perception Department of Cybernetics, Faculty of Electrical Engineering Czech Technical University, Prague Many slides thanks to Kristen Grauman and Bastian
Epipolar Geometry. Readings: See Sections 10.1 and 15.6 of Forsyth and Ponce. Right Image. Left Image. e(p ) Epipolar Lines. e(q ) q R.
Epipolar Geometry We consider two perspective images of a scene as taken from a stereo pair of cameras (or equivalently, assume the scene is rigid and imaged with a single camera from two different locations).
CS 534: Computer Vision 3D Model-based recognition
CS 534: Computer Vision 3D Model-based recognition Ahmed Elgammal Dept of Computer Science CS 534 3D Model-based Vision - 1 High Level Vision Object Recognition: What it means? Two main recognition tasks:!
3D Model based Object Class Detection in An Arbitrary View
3D Model based Object Class Detection in An Arbitrary View Pingkun Yan, Saad M. Khan, Mubarak Shah School of Electrical Engineering and Computer Science University of Central Florida http://www.eecs.ucf.edu/
Build Panoramas on Android Phones
Build Panoramas on Android Phones Tao Chu, Bowen Meng, Zixuan Wang Stanford University, Stanford CA Abstract The purpose of this work is to implement panorama stitching from a sequence of photos taken
Geometric Camera Parameters
Geometric Camera Parameters What assumptions have we made so far? -All equations we have derived for far are written in the camera reference frames. -These equations are valid only when: () all distances
Recognition. Sanja Fidler CSC420: Intro to Image Understanding 1 / 28
Recognition Topics that we will try to cover: Indexing for fast retrieval (we still owe this one) History of recognition techniques Object classification Bag-of-words Spatial pyramids Neural Networks Object
Introduction Epipolar Geometry Calibration Methods Further Readings. Stereo Camera Calibration
Stereo Camera Calibration Stereo Camera Calibration Stereo Camera Calibration Stereo Camera Calibration 12.10.2004 Overview Introduction Summary / Motivation Depth Perception Ambiguity of Correspondence
Projective Geometry. Projective Geometry
Euclidean versus Euclidean geometry describes sapes as tey are Properties of objects tat are uncanged by rigid motions» Lengts» Angles» Parallelism Projective geometry describes objects as tey appear Lengts,
Segmentation of building models from dense 3D point-clouds
Segmentation of building models from dense 3D point-clouds Joachim Bauer, Konrad Karner, Konrad Schindler, Andreas Klaus, Christopher Zach VRVis Research Center for Virtual Reality and Visualization, Institute
Randomized Trees for Real-Time Keypoint Recognition
Randomized Trees for Real-Time Keypoint Recognition Vincent Lepetit Pascal Lagger Pascal Fua Computer Vision Laboratory École Polytechnique Fédérale de Lausanne (EPFL) 1015 Lausanne, Switzerland Email:
Lecture 2: Homogeneous Coordinates, Lines and Conics
Lecture 2: Homogeneous Coordinates, Lines and Conics 1 Homogeneous Coordinates In Lecture 1 we derived the camera equations λx = P X, (1) where x = (x 1, x 2, 1), X = (X 1, X 2, X 3, 1) and P is a 3 4
Cees Snoek. Machine. Humans. Multimedia Archives. Euvision Technologies The Netherlands. University of Amsterdam The Netherlands. Tree.
Visual search: what's next? Cees Snoek University of Amsterdam The Netherlands Euvision Technologies The Netherlands Problem statement US flag Tree Aircraft Humans Dog Smoking Building Basketball Table
Probabilistic Latent Semantic Analysis (plsa)
Probabilistic Latent Semantic Analysis (plsa) SS 2008 Bayesian Networks Multimedia Computing, Universität Augsburg [email protected] www.multimedia-computing.{de,org} References
Make and Model Recognition of Cars
Make and Model Recognition of Cars Sparta Cheung Department of Electrical and Computer Engineering University of California, San Diego 9500 Gilman Dr., La Jolla, CA 92093 [email protected] Alice Chu Department
Local features and matching. Image classification & object localization
Overview Instance level search Local features and matching Efficient visual recognition Image classification & object localization Category recognition Image classification: assigning a class label to
Relating Vanishing Points to Catadioptric Camera Calibration
Relating Vanishing Points to Catadioptric Camera Calibration Wenting Duan* a, Hui Zhang b, Nigel M. Allinson a a Laboratory of Vision Engineering, University of Lincoln, Brayford Pool, Lincoln, U.K. LN6
Avoiding confusing features in place recognition
Avoiding confusing features in place recognition Jan Knopp 1, Josef Sivic 2, Tomas Pajdla 3 1 VISICS, ESAT-PSI, K.U. Leuven, Belgium 2 INRIA, WILLOW, Laboratoire d Informatique de l Ecole Normale Superieure,
Least-Squares Intersection of Lines
Least-Squares Intersection of Lines Johannes Traa - UIUC 2013 This write-up derives the least-squares solution for the intersection of lines. In the general case, a set of lines will not intersect at a
Fast Matching of Binary Features
Fast Matching of Binary Features Marius Muja and David G. Lowe Laboratory for Computational Intelligence University of British Columbia, Vancouver, Canada {mariusm,lowe}@cs.ubc.ca Abstract There has been
The Geometry of Perspective Projection
The Geometry o Perspective Projection Pinhole camera and perspective projection - This is the simplest imaging device which, however, captures accurately the geometry o perspective projection. -Rays o
MetropoGIS: A City Modeling System DI Dr. Konrad KARNER, DI Andreas KLAUS, DI Joachim BAUER, DI Christopher ZACH
MetropoGIS: A City Modeling System DI Dr. Konrad KARNER, DI Andreas KLAUS, DI Joachim BAUER, DI Christopher ZACH VRVis Research Center for Virtual Reality and Visualization, Virtual Habitat, Inffeldgasse
MIFT: A Mirror Reflection Invariant Feature Descriptor
MIFT: A Mirror Reflection Invariant Feature Descriptor Xiaojie Guo, Xiaochun Cao, Jiawan Zhang, and Xuewei Li School of Computer Science and Technology Tianjin University, China {xguo,xcao,jwzhang,lixuewei}@tju.edu.cn
EXPERIMENTAL EVALUATION OF RELATIVE POSE ESTIMATION ALGORITHMS
EXPERIMENTAL EVALUATION OF RELATIVE POSE ESTIMATION ALGORITHMS Marcel Brückner, Ferid Bajramovic, Joachim Denzler Chair for Computer Vision, Friedrich-Schiller-University Jena, Ernst-Abbe-Platz, 7743 Jena,
FAST APPROXIMATE NEAREST NEIGHBORS WITH AUTOMATIC ALGORITHM CONFIGURATION
FAST APPROXIMATE NEAREST NEIGHBORS WITH AUTOMATIC ALGORITHM CONFIGURATION Marius Muja, David G. Lowe Computer Science Department, University of British Columbia, Vancouver, B.C., Canada [email protected],
Practical Tour of Visual tracking. David Fleet and Allan Jepson January, 2006
Practical Tour of Visual tracking David Fleet and Allan Jepson January, 2006 Designing a Visual Tracker: What is the state? pose and motion (position, velocity, acceleration, ) shape (size, deformation,
Part-Based Recognition
Part-Based Recognition Benedict Brown CS597D, Fall 2003 Princeton University CS 597D, Part-Based Recognition p. 1/32 Introduction Many objects are made up of parts It s presumably easier to identify simple
Feature Tracking and Optical Flow
02/09/12 Feature Tracking and Optical Flow Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem Many slides adapted from Lana Lazebnik, Silvio Saverse, who in turn adapted slides from Steve
Geometry: Unit 1 Vocabulary TERM DEFINITION GEOMETRIC FIGURE. Cannot be defined by using other figures.
Geometry: Unit 1 Vocabulary 1.1 Undefined terms Cannot be defined by using other figures. Point A specific location. It has no dimension and is represented by a dot. Line Plane A connected straight path.
TouchPaper - An Augmented Reality Application with Cloud-Based Image Recognition Service
TouchPaper - An Augmented Reality Application with Cloud-Based Image Recognition Service Feng Tang, Daniel R. Tretter, Qian Lin HP Laboratories HPL-2012-131R1 Keyword(s): image recognition; cloud service;
Automatic 3D Reconstruction via Object Detection and 3D Transformable Model Matching CS 269 Class Project Report
Automatic 3D Reconstruction via Object Detection and 3D Transformable Model Matching CS 69 Class Project Report Junhua Mao and Lunbo Xu University of California, Los Angeles [email protected] and lunbo
Performance Evaluation of RANSAC Family
CHOI, KIM, YU: PERFORMANCE EVALUATION OF RANSAC FAMILY 1 Performance Evaluation of RANSAC Family Sunglok Choi 1 [email protected] Taemin Kim 2 [email protected] Wonpil Yu 1 [email protected] 1 Robot Research
Projective Geometry: A Short Introduction. Lecture Notes Edmond Boyer
Projective Geometry: A Short Introduction Lecture Notes Edmond Boyer Contents 1 Introduction 2 11 Objective 2 12 Historical Background 3 13 Bibliography 4 2 Projective Spaces 5 21 Definitions 5 22 Properties
How To Analyze Ball Blur On A Ball Image
Single Image 3D Reconstruction of Ball Motion and Spin From Motion Blur An Experiment in Motion from Blur Giacomo Boracchi, Vincenzo Caglioti, Alessandro Giusti Objective From a single image, reconstruct:
521493S Computer Graphics. Exercise 2 & course schedule change
521493S Computer Graphics Exercise 2 & course schedule change Course Schedule Change Lecture from Wednesday 31th of March is moved to Tuesday 30th of March at 16-18 in TS128 Question 2.1 Given two nonparallel,
2-View Geometry. Mark Fiala Ryerson University [email protected]
CRV 2010 Tutorial Day 2-View Geometry Mark Fiala Ryerson University [email protected] 3-Vectors for image points and lines Mark Fiala 2010 2D Homogeneous Points Add 3 rd number to a 2D point on image
IMPLICIT SHAPE MODELS FOR OBJECT DETECTION IN 3D POINT CLOUDS
IMPLICIT SHAPE MODELS FOR OBJECT DETECTION IN 3D POINT CLOUDS Alexander Velizhev 1 (presenter) Roman Shapovalov 2 Konrad Schindler 3 1 Hexagon Technology Center, Heerbrugg, Switzerland 2 Graphics & Media
MVA ENS Cachan. Lecture 2: Logistic regression & intro to MIL Iasonas Kokkinos [email protected]
Machine Learning for Computer Vision 1 MVA ENS Cachan Lecture 2: Logistic regression & intro to MIL Iasonas Kokkinos [email protected] Department of Applied Mathematics Ecole Centrale Paris Galen
Combining Local Recognition Methods for Better Image Recognition
Combining Local Recognition Methods for Better Image Recognition Bart Lamiroy Patrick Gros y Sylvaine Picard INRIA CNRS DGA MOVI GRAVIR z IMAG INRIA RHÔNE-ALPES ZIRST 655, Av. de l Europe Montbonnot FRANCE
Image Segmentation and Registration
Image Segmentation and Registration Dr. Christine Tanner ([email protected]) Computer Vision Laboratory, ETH Zürich Dr. Verena Kaynig, Machine Learning Laboratory, ETH Zürich Outline Segmentation
More Local Structure Information for Make-Model Recognition
More Local Structure Information for Make-Model Recognition David Anthony Torres Dept. of Computer Science The University of California at San Diego La Jolla, CA 9093 Abstract An object classification
Segmentation & Clustering
EECS 442 Computer vision Segmentation & Clustering Segmentation in human vision K-mean clustering Mean-shift Graph-cut Reading: Chapters 14 [FP] Some slides of this lectures are courtesy of prof F. Li,
A System for Capturing High Resolution Images
A System for Capturing High Resolution Images G.Voyatzis, G.Angelopoulos, A.Bors and I.Pitas Department of Informatics University of Thessaloniki BOX 451, 54006 Thessaloniki GREECE e-mail: [email protected]
PHOTOGRAMMETRIC TECHNIQUES FOR MEASUREMENTS IN WOODWORKING INDUSTRY
PHOTOGRAMMETRIC TECHNIQUES FOR MEASUREMENTS IN WOODWORKING INDUSTRY V. Knyaz a, *, Yu. Visilter, S. Zheltov a State Research Institute for Aviation System (GosNIIAS), 7, Victorenko str., Moscow, Russia
Classifying Manipulation Primitives from Visual Data
Classifying Manipulation Primitives from Visual Data Sandy Huang and Dylan Hadfield-Menell Abstract One approach to learning from demonstrations in robotics is to make use of a classifier to predict if
The use of computer vision technologies to augment human monitoring of secure computing facilities
The use of computer vision technologies to augment human monitoring of secure computing facilities Marius Potgieter School of Information and Communication Technology Nelson Mandela Metropolitan University
Scale-Invariant Object Categorization using a Scale-Adaptive Mean-Shift Search
in DAGM 04 Pattern Recognition Symposium, Tübingen, Germany, Aug 2004. Scale-Invariant Object Categorization using a Scale-Adaptive Mean-Shift Search Bastian Leibe ½ and Bernt Schiele ½¾ ½ Perceptual Computing
Epipolar Geometry and Visual Servoing
Epipolar Geometry and Visual Servoing Domenico Prattichizzo joint with with Gian Luca Mariottini and Jacopo Piazzi www.dii.unisi.it/prattichizzo Robotics & Systems Lab University of Siena, Italy Scuoladi
Tracking in flussi video 3D. Ing. Samuele Salti
Seminari XXIII ciclo Tracking in flussi video 3D Ing. Tutors: Prof. Tullio Salmon Cinotti Prof. Luigi Di Stefano The Tracking problem Detection Object model, Track initiation, Track termination, Tracking
Geometric Transformation CS 211A
Geometric Transformation CS 211A What is transformation? Moving points (x,y) moves to (x+t, y+t) Can be in any dimension 2D Image warps 3D 3D Graphics and Vision Can also be considered as a movement to
THREE DIMENSIONAL GEOMETRY
Chapter 8 THREE DIMENSIONAL GEOMETRY 8.1 Introduction In this chapter we present a vector algebra approach to three dimensional geometry. The aim is to present standard properties of lines and planes,
ACCURACY ASSESSMENT OF BUILDING POINT CLOUDS AUTOMATICALLY GENERATED FROM IPHONE IMAGES
ACCURACY ASSESSMENT OF BUILDING POINT CLOUDS AUTOMATICALLY GENERATED FROM IPHONE IMAGES B. Sirmacek, R. Lindenbergh Delft University of Technology, Department of Geoscience and Remote Sensing, Stevinweg
Machine Learning and Data Mining. Regression Problem. (adapted from) Prof. Alexander Ihler
Machine Learning and Data Mining Regression Problem (adapted from) Prof. Alexander Ihler Overview Regression Problem Definition and define parameters ϴ. Prediction using ϴ as parameters Measure the error
Monash University Clayton s School of Information Technology CSE3313 Computer Graphics Sample Exam Questions 2007
Monash University Clayton s School of Information Technology CSE3313 Computer Graphics Questions 2007 INSTRUCTIONS: Answer all questions. Spend approximately 1 minute per mark. Question 1 30 Marks Total
Distinctive Image Features from Scale-Invariant Keypoints
Distinctive Image Features from Scale-Invariant Keypoints David G. Lowe Computer Science Department University of British Columbia Vancouver, B.C., Canada [email protected] January 5, 2004 Abstract This paper
Point Matching as a Classification Problem for Fast and Robust Object Pose Estimation
Point Matching as a Classification Problem for Fast and Robust Object Pose Estimation Vincent Lepetit Julien Pilet Pascal Fua Computer Vision Laboratory Swiss Federal Institute of Technology (EPFL) 1015
Object Recognition and Template Matching
Object Recognition and Template Matching Template Matching A template is a small image (sub-image) The goal is to find occurrences of this template in a larger image That is, you want to find matches of
Interactive Segmentation, Tracking, and Kinematic Modeling of Unknown 3D Articulated Objects
Interactive Segmentation, Tracking, and Kinematic Modeling of Unknown 3D Articulated Objects Dov Katz, Moslem Kazemi, J. Andrew Bagnell and Anthony Stentz 1 Abstract We present an interactive perceptual
Pennsylvania System of School Assessment
Pennsylvania System of School Assessment The Assessment Anchors, as defined by the Eligible Content, are organized into cohesive blueprints, each structured with a common labeling system that can be read
Classification of Fingerprints. Sarat C. Dass Department of Statistics & Probability
Classification of Fingerprints Sarat C. Dass Department of Statistics & Probability Fingerprint Classification Fingerprint classification is a coarse level partitioning of a fingerprint database into smaller
with functions, expressions and equations which follow in units 3 and 4.
Grade 8 Overview View unit yearlong overview here The unit design was created in line with the areas of focus for grade 8 Mathematics as identified by the Common Core State Standards and the PARCC Model
Chapter. 4 Mechanism Design and Analysis
Chapter. 4 Mechanism Design and Analysis 1 All mechanical devices containing moving parts are composed of some type of mechanism. A mechanism is a group of links interacting with each other through joints
Solution Guide III-C. 3D Vision. Building Vision for Business. MVTec Software GmbH
Solution Guide III-C 3D Vision MVTec Software GmbH Building Vision for Business Machine vision in 3D world coordinates, Version 10.0.4 All rights reserved. No part of this publication may be reproduced,
Solving Simultaneous Equations and Matrices
Solving Simultaneous Equations and Matrices The following represents a systematic investigation for the steps used to solve two simultaneous linear equations in two unknowns. The motivation for considering
Computer Graphics. Geometric Modeling. Page 1. Copyright Gotsman, Elber, Barequet, Karni, Sheffer Computer Science - Technion. An Example.
An Example 2 3 4 Outline Objective: Develop methods and algorithms to mathematically model shape of real world objects Categories: Wire-Frame Representation Object is represented as as a set of points
Optical Tracking Using Projective Invariant Marker Pattern Properties
Optical Tracking Using Projective Invariant Marker Pattern Properties Robert van Liere, Jurriaan D. Mulder Department of Information Systems Center for Mathematics and Computer Science Amsterdam, the Netherlands
3D Scanner using Line Laser. 1. Introduction. 2. Theory
. Introduction 3D Scanner using Line Laser Di Lu Electrical, Computer, and Systems Engineering Rensselaer Polytechnic Institute The goal of 3D reconstruction is to recover the 3D properties of a geometric
Georgia Standards of Excellence Curriculum Map. Mathematics. GSE 8 th Grade
Georgia Standards of Excellence Curriculum Map Mathematics GSE 8 th Grade These materials are for nonprofit educational purposes only. Any other use may constitute copyright infringement. GSE Eighth Grade
Object class recognition using unsupervised scale-invariant learning
Object class recognition using unsupervised scale-invariant learning Rob Fergus Pietro Perona Andrew Zisserman Oxford University California Institute of Technology Goal Recognition of object categories
A PHOTOGRAMMETRIC APPRAOCH FOR AUTOMATIC TRAFFIC ASSESSMENT USING CONVENTIONAL CCTV CAMERA
A PHOTOGRAMMETRIC APPRAOCH FOR AUTOMATIC TRAFFIC ASSESSMENT USING CONVENTIONAL CCTV CAMERA N. Zarrinpanjeh a, F. Dadrassjavan b, H. Fattahi c * a Islamic Azad University of Qazvin - [email protected]
VEHICLE LOCALISATION AND CLASSIFICATION IN URBAN CCTV STREAMS
VEHICLE LOCALISATION AND CLASSIFICATION IN URBAN CCTV STREAMS Norbert Buch 1, Mark Cracknell 2, James Orwell 1 and Sergio A. Velastin 1 1. Kingston University, Penrhyn Road, Kingston upon Thames, KT1 2EE,
How To Fuse A Point Cloud With A Laser And Image Data From A Pointcloud
REAL TIME 3D FUSION OF IMAGERY AND MOBILE LIDAR Paul Mrstik, Vice President Technology Kresimir Kusevic, R&D Engineer Terrapoint Inc. 140-1 Antares Dr. Ottawa, Ontario K2E 8C4 Canada [email protected]
INTRODUCTION TO RENDERING TECHNIQUES
INTRODUCTION TO RENDERING TECHNIQUES 22 Mar. 212 Yanir Kleiman What is 3D Graphics? Why 3D? Draw one frame at a time Model only once X 24 frames per second Color / texture only once 15, frames for a feature
Fast and Robust Normal Estimation for Point Clouds with Sharp Features
1/37 Fast and Robust Normal Estimation for Point Clouds with Sharp Features Alexandre Boulch & Renaud Marlet University Paris-Est, LIGM (UMR CNRS), Ecole des Ponts ParisTech Symposium on Geometry Processing
Simultaneous Gamma Correction and Registration in the Frequency Domain
Simultaneous Gamma Correction and Registration in the Frequency Domain Alexander Wong [email protected] William Bishop [email protected] Department of Electrical and Computer Engineering University
Automatic georeferencing of imagery from high-resolution, low-altitude, low-cost aerial platforms
Automatic georeferencing of imagery from high-resolution, low-altitude, low-cost aerial platforms Amanda Geniviva, Jason Faulring and Carl Salvaggio Rochester Institute of Technology, 54 Lomb Memorial
Equations Involving Lines and Planes Standard equations for lines in space
Equations Involving Lines and Planes In this section we will collect various important formulas regarding equations of lines and planes in three dimensional space Reminder regarding notation: any quantity
Projection Center Calibration for a Co-located Projector Camera System
Projection Center Calibration for a Co-located Camera System Toshiyuki Amano Department of Computer and Communication Science Faculty of Systems Engineering, Wakayama University Sakaedani 930, Wakayama,
Edge-based Template Matching and Tracking for Perspectively Distorted Planar Objects
Edge-based Template Matching and Tracking for Perspectively Distorted Planar Objects Andreas Hofhauser, Carsten Steger, and Nassir Navab TU München, Boltzmannstr. 3, 85748 Garching bei München, Germany
Lecture 9: Geometric map transformations. Cartographic Transformations
Cartographic Transformations Analytical and Computer Cartography Lecture 9: Geometric Map Transformations Attribute Data (e.g. classification) Locational properties (e.g. projection) Graphics (e.g. symbolization)
3D Object Recognition in Clutter with the Point Cloud Library
3D Object Recognition in Clutter with the Point Cloud Library Federico Tombari, Ph.D [email protected] University of Bologna Open Perception Data representations in PCL PCL can deal with both organized
REAL-TIME OBJECT DETECTION AND TRACKING FOR INDUSTRIAL APPLICATIONS
REAL-TIME OBJECT DETECTION AND TRACKING FOR INDUSTRIAL APPLICATIONS Selim Benhimane 1, Hesam Najafi 2, Matthias Grundmann 3, Ezio Malis 4, Yakup Genc 2, Nassir Navab 1 1 Department of Computer Science,
Lecture 2: Geometric Image Transformations
Lecture 2: Geometric Image Transformations Harvey Rhody Chester F. Carlson Center for Imaging Science Rochester Institute of Technology [email protected] September 8, 2005 Abstract Geometric transformations
Projective Reconstruction from Line Correspondences
Projective Reconstruction from Line Correspondences Richard I. Hartley G.E. CRD, Schenectady, NY, 12301. Abstract The paper gives a practical rapid algorithm for doing projective reconstruction of a scene
3D Human Face Recognition Using Point Signature
3D Human Face Recognition Using Point Signature Chin-Seng Chua, Feng Han, Yeong-Khing Ho School of Electrical and Electronic Engineering Nanyang Technological University, Singapore 639798 [email protected]
Common Core Unit Summary Grades 6 to 8
Common Core Unit Summary Grades 6 to 8 Grade 8: Unit 1: Congruence and Similarity- 8G1-8G5 rotations reflections and translations,( RRT=congruence) understand congruence of 2 d figures after RRT Dilations
