Towards Auto-calibration of Smart Phones Using Orientation Sensors
|
|
|
- Jerome Cross
- 10 years ago
- Views:
Transcription
1 2013 IEEE Conference on Computer Vision and Pattern Recognition Workshops Towards Auto-calibration of Smart Phones Using Orientation Sensors Philip Saponaro and Chandra Kambhamettu Video/Image Modeling and Synthesis (VIMS) Lab, Dept. of Computer and Information Science, University of Delaware Newark, Delaware Abstract In this paper, we address the problem of auto calibration of cameras which can rotate freely and change focal length, and we present an algorithm for finding the intrinsic parameters using only two images. We utilize orientation sensors found on many modern smart phones to help decompose the infinite homography into two equivalent upper triangular matrices based only on the intrinsic parameters. We account for small translations between views by calculating the homography based on correspondences on objects that are far away from the camera. We show results based on both real and synthetic data, and quantify the tolerance of our system to small translations and errors in the orientation sensors. Our results are comparable to other recent auto-calibration work while requiring only two images and being tolerant to some translation. 1. Introduction Smart phones with sophisticated cameras are now widespread. There are a wide variety of vision applications for smart phones; for example, distributed surveillance networks [2] and augmented reality systems [18]. Many of these applications require the intrinsic parameters of the camera the focal length, aspect ratio, skew, and principle point. However, due to the fact that many newer smart phones can change focus, zoom, and move freely, traditional off-line calibration techniques are not practical. Therefore, auto-calibration is needed to calibrate the camera on-line. Most Android phones made after 2010 use multiple sensors, such as accelerometers, gyroscopes, and compasses, to determine the orientation. Accelerometers record linear acceleration and gravity, and thus can give the orientation of the phone. However, the readings from the accelerometer can be quite noisy and change quickly with small hand jittering. Compasses are magnetic field sensors, and thus can be used to get the heading based on the Earth s magnetic field. However, any other magnetic field will interfere with the reading, causing inaccuracies. Gyroscopes sense angular velocity and give very smooth readings, but the readings drift over time, and cannot measure gravity [24]. The fusion of these sensors significantly improves the accuracy and usability of the orientation reading for real applications [27], with up to 0.03 accuracy. We developed our algorithms using the Samsung Galaxy S3 running Android Ice- Cream Sandwich (see Figure 1), which automatically handles the sensor fusion. This allows us to record the orientation of the phone along with an image, and more easily determine the intrinsic parameters of the camera. Our contributions are as follows. First, we implement an algorithm for finding the intrinsic parameters of a camera in the case of general rotation, zooming, and small translations using only two images. This algorithm is a modification of previous work by Junejo and Foroosh [13]. Secondly, we quantify how translations can affect the final results. We show that if correspondences are found on an object that is 40m away, then small translations, which are normally due to shaking or imperfect rotation about the camera center, are negligible. Finally, we quantify how errors in the orientation measurements affect the final calculated intrinsic parameters. The remainder of the paper is as follows. In Section 2, we describe related works and give context to our work. In Section 3, we describe our methods for finding the calibration parameters using only two images. In Section 4, we show the results of running our algorithm on real data, and compare the performance to other works. Finally, in Section 5, we conclude the paper and give recommendations for future work. 2. Background and Related Works Traditional calibration methods [26, 28] achieve high accuracy; however, these methods are offline and require physical calibration patterns to be used in the scene. Thus, they are not practical for use on cameras with dynamically changing intrinsic and extrinsic parameters, but can be used /13 $ IEEE DOI /CVPRW
2 Figure 1: Samsung Galaxy S3 running Android Ice-Cream Sandwich and Androsens app which shows results of sensor fusion (courtesy of Christian Schneider s Androsens app). for performance comparison. Auto-calibration, on the other hand, can deal with dynamically changing intrinsic and extrinsic parameters. Auto-calibration of a camera is calibration done with sequence of images instead of a calibration board, and can be performed online if necessary. Brown describes a method of calibrating a close range camera with bundle adjustment in [3, 4]. The first auto calibration using the Kruppa equations is described in [6], and handles the problem with constant intrinsic parameters of a freely moving camera. Other works focus on the problem when the camera can only translate [17] or when it can only rotate [10]. Finally, some works calibrate the camera with varying intrinsic parameters, specifically the focal length [13, 11, 12, 14, 19, 21]. Our work is related the most to [13], which in turn is built off the work in [5, 1]. Junejo and Foroosh [13] show that with only two images, four independent constraints can be found by decomposing the infinite homography using a series of Givens rotations. In our work, however, we do not need to calculate the Givens rotations, and instead use the orientation reading directly from the Android phone, which saves computation time. Unlike [13], we show how to deal with small translations that are expected from taking images while holding a phone. 3. Methods 3.1. Notation and General Methods Camera calibration is the process of finding the invertible 3x3 camera matrix K such that y Kx (1) where x is a 3-dimensional vector representing a 3D point relative to the camera, y is the corresponding 2D image point, and is equality within a scale factor. Specifically, K = λf s u 0 f v (2) which contains the 5 intrinsic camera parameters: f is the focal length, λ is the aspect ratio, s is the skew, and u, v are the principle point. Since most cameras generally use CCD (chargedcoupled device) or CMOS (complementary metal-oxidesemiconductor) type sensors, the transformation between imaged planes is 2D to 2D, which is called the plane induced homography [9]. This plane induced homography relates pixels in one image to pixels in the other image, and can be applied to an entire image to warp it. The infinite homography is a special type of plane induced homography which maps features of the plane at infinity of one view to another, and also can be used to calibrate a camera using image sequences, as in [15]. In general, we can write the infinite homography as H ij K i R 1 ij K 1 j = K i RijK T 1 j. (3) Our method is summarized as follows. First take two images and record their orientation values as Euler angles, which are used to construct R. Then calculate the infinite homography by obtaining image correspondences between the images using SIFT. Then, by assuming zero skew, we 21
3 can solve a nonlinear homogeneous system for the intrinsics. Finally, these intrinsics are further refined using bundle adjustment [25]. Note that if there is any translation, then the calculated homography will not be the infinite homography, and error will be introduced. We show that by limiting our correspondences to those on objects far away from the camera that this error is negligible (see section 4.1.1) Camera Calibration with Two Images General Rotation and Changing Focal Length As in [13], we need only two images to calibrate the camera. We rearrange Eq. 3 into an equivalency between two upper triangular matrices. This representation provides four constraints, which, together with the assumption that the skew s =0and that the principle point is constant, will allow us to solve for all the intrinsic parameters. Thus Eq. 3 becomes H 12 K 1 R T 12K 1 2. (4) Using language from aviation, yaw = α is the rotation around the z axis, pitch =β is the rotation around the y axis, and roll = γ is the rotation around the x axis. R 12 is defined as R 12 = R z (α)r y (β)r x (γ) cos α sin α 0 R z (α) = sin α cos α cos β 0 sinβ R y (β) = sin β 0 cosβ R z (γ) = 0 cosγ sin γ. (5) 0 sinγ cos γ Multiplying both sides of Eq. (4) on the left by K 1 1 and R 12 gives us R 12 K1 1 H12 K2 1. (6) We can rewrite this matrix as an upper triangular matrix and equate it to the inverse of the second camera matrix, with the assumption that the skew is zero. R 12 K1 1 H12 = c 1 1 c 2 c 3 λf 2 0 u λf 2 0 c 4 c f 2 v f 2 (7) 0 0 c The only unknowns left in this equality are the intrinsic parameters. Thus, each of the c i are first order polynomials in terms of only the intrinsic parameters. This equality leads to four nonlinear first-order homogeneous constraints on the intrinsic parameters: c 3 + uc 1 =0 c 5 + vc 4 =0 c 4 λc 1 =0 c 2 =0 (8) We assumed the skew is zero and the principle point is constant; therefore, these four equations are enough to solve for the four unknowns f,λ,u,v. Moreover, these equations do not rely on f 2, the focal length in the second image. These equations are first order; however, since they are nonlinear, it is possible to get multiple solutions. We take the solutions which yield positive intrinsic parameters. If the focal length of the second image, f 2, is desired, we simply switch the subscripts 1 and 2 in the above derivation since it is symmetric. The assumption that the principle point is constant is reasonable, since [20, 1] have shown performance is not hindered much by these assumptions. Moreover, bundle adjustment is used to refine the results, which can alleviate these assumptions. 4. Experiments and Results We performed several experiments to quantify the errors of our method, to compare to the method described in [13], and to quantify how inaccuracies in the measured orientation would affect the final calculated parameters. We ran both synthetic experiments and experiments on real data Experiments with Synthetic Data To measure the accuracy of our system in ideal circumstances, we created synthetic data. 3d points were generated along a quadric surface, and a synthetic camera with f = , λ =.9996, u = 600, v = 600 was placed at the world origin to project those points into a 2d image. A random rotation on each axis from [ 5, 5 ] was applied to the camera, then SIFT and RANSAC was run to compute a homography between the two images. Note that the focal length was held constant in this test. Table 1 shows the results. The average percent error from the ground truth was under 1% in all cases. We then perform a similiar experiment, but with differing focal lengths. The focal length was changed by a random number from [ 300, 300], as well as a random rotation 22
4 f λ u u Ground Truth Mean Std Dev Avg % Error Table 1: Results from synthetic tests without changing the focal length. f f 2 λ u u Ground Truth Mean Std Dev Avg % Error Table 2: Results from synthetic tests with changing the focal length. Figure 2: Images used in Table 3 for calculating intrinsic parameters. The camera was rotated about all 3 axes with minimal translation between pairs of images. on each axis from [ 5, 5 ] was applied. Those results are presented in Table The Effect of Translation Translations can be caused by the hand shaking while taking pictures, or by imperfect rotation about the camera center. According to Hartley and Zisserman [8], image points are related by the equation x 2 = K 2 R T 21K 1 1 x 1 + K 2t z = H 21x + K 2t z (9) where x 1 and x 2 are the image correspondences, t is the translation, and z is the depth measured from the first camera. Thus, if there is no translation between the two cameras, the homography we calculate with SIFT is the infinite homography. Furthermore, if we solve this equation using correspondences on objects that lie far from the camera (i.e. z ), the translation term can be negligible. To test this out, we generated synthetic data. The synthetic camera had the same parameters as in our first experiment: f = , λ =.9996, u = 600, v = 600. An image was created by placing a quadric surface 1m from the camera and applying the calibration matrix to the points on the surface. The camera was rotated and another image was generated. Then, in a loop, without changing the rotation, the camera was translated in small increments. A plot of the relative percent error vs. the translation is given in Figure 3. This experiment was repeated by placing the surface at 20m and 40m from the camera. In all three graphs, the error curves for the parameters are similar; however, note that the x-axis scale changes between the graphs, which shows that as the object moves further away, the error decreases. In fact, at 40m, the camera can move up to 8cm before there is an error of 7%. 23
5 f f 2 λ u v Mean Std Dev Std Dev [13] Ground Truth Percent Error Table 3: Results from calculating intrinsic parameters over all combinations of images. Figure 3: Effect of small translations on the relative percent error Experiments with Real Data All real data, i.e. orientations readings and images, were collected with the Samsung Galaxy S3 running Android (Ice-Cream Sandwich) at a resolution of 640x480. The images and corresponding orientations were sent to a desktop computer running 64-bit Windows 7, where the intrinsic parameters were calculated using Matlab Image feature correspondences were obtained using SIFT [16] and RANSAC [7] for outlier removal. For each pair of images, 150 correspondences were obtained. We measure the performance of the real data in two ways. The first way to measure the performance is by calibrating for the parameters from a planar checkerboard offline using Callab and Calde from the Institute of Robotics and Mechatronics in the German Aerospace Center [23, 22]. We took 15 images of the calibration board at different orientations while not changing the zoom of the camera. Using this method, we obtained sub-pixel RMS (Root Mean Square) error, and thus, we use this as the ground truth to compare to. The second way we measure performance is similar to [13]; we take a sequence of images and calculate the intrinsic parameters using all possible combinations between the images. Then we characterize the uncertainty by the standard deviation over all the calculated parameters. The results of these methods are compared to the results from [13], and are displayed in Table 3. As can be seen in the table, our results are comparable in error. The relatively high standard deviation for the v parameter was caused by 24
6 (a) Focal length error (b) Aspect ratio error. (c) Principle point u error (d) Principle point v error Figure 4: Results for image pair 1,2 with synthetic error added a single outlier. Moreover, the distance to the object was around 1m, which, according to Section 4.1.1, means the results are sensitive to translations Measuring the Sensitivity to Orientation Sensor Errors We test how tolerant our system is to orientation measurement errors. To do this, we take a single image pair and calculate the intrinsic parameters. Then, we introduce synthetic error to the recorded orientations and recalculate the intrinsic parameters. The error we introduce is from 2 to 2 in steps of See Figure 4 for a plot of the results. For each intrinsic parameter, Figure 4 plots the error introduced (x-axis) vs the relative difference from the original calculation with no error (y-axis). The error is introduced to the individual rotations yaw, pitch, and roll and then introduced to all three at the same time. 5. Conclusions and Future Work We described a method for calibrating a smartphone camera by taking two images at different rotations while tolerating small translations. We show how the amount of error in the smartphone orientation sensors and amount of translation affects the overall results. We demonstrated with synthetic data that our results are very close to the ground truth, and with real data our results are comparable to other state of the art works. Our work shows some advantages over other methods. First, since we are using measured rotation angles, our methods do not require us to calculate the rotation between camera positions. Secondly, we require only 2 images and no calibration board. This lends itself to future works which can calibrate a camera in real time. Also, in future work, we would like to remove effects of translation by directly calculating the infinite homography, H, by using vanishing points in the scene. This will allow the calibration to be effective in both indoor environments, which can be more sensitive to translations but have easy to find vanishing points, and outdoor environments. In conclusion this paper has 3 contributions. First, we implement an algorithm for finding the intrinsic parameters of a camera in the case of general rotation, zooming, and small translations using only two images. This algorithm is a modification of previous work by Junejo and Foroosh [13]. Secondly, we quantify how translations can affect the final results. We show that if correspondences are found on an object that is 40m away, then small translations, which are normally due to shaking or imperfect rotation about the camera center, are negligible. Finally, we quantify how errors in the orientation measurement affects the final calcu- 25
7 lated intrinsic parameters. References [1] L. Agapito, E. Hayman, and I. Reid. Self-calibration of rotating and zooming cameras. Int. J. Comput. Vision, 45(2): , Nov , 3 [2] P. Bolliger, M. Köhler, and K. Römer. Facet: towards a smart camera network of mobile phones. In Proceedings of the 1st international conference on Autonomic computing and communication systems, Autonomics 07, pages 17:1 17:10, ICST, Brussels, Belgium, Belgium, ICST (Institute for Computer Sciences, Social-Informatics and Telecommunications Engineering). 1 [3] D. C. Brown. Close-range camera calibration. PHO- TOGRAMMETRIC ENGINEERING, 37(8): , [4] D. C. Brown. The bundle adjustment progress and prospects. Int. Archives Photogrammetry, [5] L. de Agapito, R. Hartley, and E. Hayman. Linear selfcalibration of a rotating and zooming camera. In Computer Vision and Pattern Recognition, IEEE Computer Society Conference on., volume 1, pages 2 vol. (xxiii ), [6] O. Faugeras, Q. Luong, and S. Maybank. Camera selfcalibration: Theory and experiments. In G. Sandini, editor, Computer Vision ECCV 92, volume 588 of Lecture Notes in Computer Science, pages Springer Berlin Heidelberg, [7] M. A. Fischler and R. C. Bolles. Random sample consensus: a paradigm for model fitting with applications to image analysis and automated cartography. Commun. ACM, 24(6): , June [8] R. Hartley and A. Zisserman. Multiple View Geometry in Computer Vision. Cambridge University Press, New York, NY, USA, 2 edition, [9] R. Hartley and A. Zisserman. Multiple View Geometry in Computer Vision. Cambridge books online. Cambridge University Press, [10] R. I. Hartley. Self-calibration of stationary cameras. Int. J. Comput. Vision, 22(1):5 23, Feb [11] R. I. Hartley, E. Hayman, L. Agapito, and I. D. Reid. Camera calibration and the search for infinity. In Proc. 7th International Conference on Computer Vision, Kerkyra, Greece, pages , [12] A. Heyden and K. Astrom. Flexible calibration: minimal cases for auto-calibration. In Computer Vision, The Proceedings of the Seventh IEEE International Conference on, volume 1, pages vol.1, [13] I. N. Junejo and H. Foroosh. Optimizing ptz camera calibration from two images. Mach. Vision Appl., 23(2): , Mar , 2, 3, 5, 6 [14] F. Kahl, B. Triggs, and K. Astrom. Critical motions for auto-calibration when some intrinsic parameters can vary. J. Math. Imaging Vis., 13(2): , Oct [15] J.-S. Kim and I. S. Kweon. Infinite homography estimation using two arbitrary planar rectangles. In Proceedings of the 7th Asian conference on Computer Vision - Volume Part II, ACCV 06, pages 1 10, Berlin, Heidelberg, Springer- Verlag. 2 [16] D. Lowe. Object recognition from local scale-invariant features. In Computer Vision, The Proceedings of the Seventh IEEE International Conference on, volume 2, pages vol.2, [17] T. Moons, L. Van Gool, M. Proesmans, and E. Pauwels. Affine reconstruction from perspective image pairs with a relative object-camera translation in between. Pattern Analysis and Machine Intelligence, IEEE Transactions on, 18(1):77 83, jan [18] G. Papagiannakis, G. Singh, and N. Magnenat-Thalmann. A survey of mobile and wireless technologies for augmented reality systems. Computer Animation and Virtual Worlds, 19(1):3 22, [19] M. Pollefeys, R. Koch, and L. V. Gool. Self-calibration and metric reconstruction in spite of varying and unknown internal camera parameters. In INTERNATIONAL JOURNAL OF COMPUTER VISION, pages 7 25, [20] M. Pollefeys, R. Koch, and L. Van Gool. Self-calibration and metric reconstruction in spite of varying and unknown internal camera parameters. In Proceedings of the Sixth International Conference on Computer Vision, ICCV 98, pages 90, Washington, DC, USA, IEEE Computer Society. 3 [21] K.-T. Song and J.-C. Tai. Dynamic calibration of pan/tilt/zoom cameras for traffic monitoring. Trans. Sys. Man Cyber. Part B, 36(5): , Oct [22] K. Strobl and G. Hirzinger. Optimal hand-eye calibration. In Intelligent Robots and Systems, 2006 IEEE/RSJ International Conference on, pages , oct [23] K. H. Strobl, W. Sepp, S. Fuchs, C. Paredes, M. Smisek, and K. Arbter. DLR CalDe and DLR CalLab. 5 [24] D. Titterton, J. Weston, D. Titterton, and J. Weston. Strapdown Inertial Navigation Technology, 2nd Edition. Iee Radar Series. The Institution of Engineering and Technology, [25] B. Triggs, P. Mclauchlan, R. Hartley, and A. Fitzgibbon. Bundle adjustment a modern synthesis. In Vision Algorithms: Theory and Practice, LNCS, pages Springer Verlag, [26] R. Tsai. A versatile camera calibration technique for highaccuracy 3d machine vision metrology using off-the-shelf tv cameras and lenses. Robotics and Automation, IEEE Journal of, 3(4): , august [27] B. van Tonder and J. Wesson. The impact of sensor fusion on tilt interaction in a mobile map-based application. In Proceedings of the South African Institute of Computer Scientists and Information Technologists Conference on Knowledge, Innovation and Leadership in a Diverse, Multidisciplinary Environment, SAICSIT 11, pages , New York, NY, USA, ACM. 1 [28] Z. Zhang. A flexible new technique for camera calibration. Pattern Analysis and Machine Intelligence, IEEE Transactions on, 22(11): , nov
Wii Remote Calibration Using the Sensor Bar
Wii Remote Calibration Using the Sensor Bar Alparslan Yildiz Abdullah Akay Yusuf Sinan Akgul GIT Vision Lab - http://vision.gyte.edu.tr Gebze Institute of Technology Kocaeli, Turkey {yildiz, akay, akgul}@bilmuh.gyte.edu.tr
Relating Vanishing Points to Catadioptric Camera Calibration
Relating Vanishing Points to Catadioptric Camera Calibration Wenting Duan* a, Hui Zhang b, Nigel M. Allinson a a Laboratory of Vision Engineering, University of Lincoln, Brayford Pool, Lincoln, U.K. LN6
Build Panoramas on Android Phones
Build Panoramas on Android Phones Tao Chu, Bowen Meng, Zixuan Wang Stanford University, Stanford CA Abstract The purpose of this work is to implement panorama stitching from a sequence of photos taken
A Flexible New Technique for Camera Calibration
A Flexible New Technique for Camera Calibration Zhengyou Zhang December 2, 1998 (updated on December 14, 1998) (updated on March 25, 1999) (updated on Aug. 1, 22; a typo in Appendix B) (updated on Aug.
Can we calibrate a camera using an image of a flat, textureless Lambertian surface?
Can we calibrate a camera using an image of a flat, textureless Lambertian surface? Sing Bing Kang 1 and Richard Weiss 2 1 Cambridge Research Laboratory, Compaq Computer Corporation, One Kendall Sqr.,
521493S Computer Graphics. Exercise 2 & course schedule change
521493S Computer Graphics Exercise 2 & course schedule change Course Schedule Change Lecture from Wednesday 31th of March is moved to Tuesday 30th of March at 16-18 in TS128 Question 2.1 Given two nonparallel,
Essential Mathematics for Computer Graphics fast
John Vince Essential Mathematics for Computer Graphics fast Springer Contents 1. MATHEMATICS 1 Is mathematics difficult? 3 Who should read this book? 4 Aims and objectives of this book 4 Assumptions made
Lecture 2: Homogeneous Coordinates, Lines and Conics
Lecture 2: Homogeneous Coordinates, Lines and Conics 1 Homogeneous Coordinates In Lecture 1 we derived the camera equations λx = P X, (1) where x = (x 1, x 2, 1), X = (X 1, X 2, X 3, 1) and P is a 3 4
Sensor Fusion Mobile Platform Challenges and Future Directions Jim Steele VP of Engineering, Sensor Platforms, Inc.
Sensor Fusion Mobile Platform Challenges and Future Directions Jim Steele VP of Engineering, Sensor Platforms, Inc. Copyright Khronos Group 2012 Page 104 Copyright Khronos Group 2012 Page 105 How Many
EXPERIMENTAL EVALUATION OF RELATIVE POSE ESTIMATION ALGORITHMS
EXPERIMENTAL EVALUATION OF RELATIVE POSE ESTIMATION ALGORITHMS Marcel Brückner, Ferid Bajramovic, Joachim Denzler Chair for Computer Vision, Friedrich-Schiller-University Jena, Ernst-Abbe-Platz, 7743 Jena,
Solution Guide III-C. 3D Vision. Building Vision for Business. MVTec Software GmbH
Solution Guide III-C 3D Vision MVTec Software GmbH Building Vision for Business Machine vision in 3D world coordinates, Version 10.0.4 All rights reserved. No part of this publication may be reproduced,
A PHOTOGRAMMETRIC APPRAOCH FOR AUTOMATIC TRAFFIC ASSESSMENT USING CONVENTIONAL CCTV CAMERA
A PHOTOGRAMMETRIC APPRAOCH FOR AUTOMATIC TRAFFIC ASSESSMENT USING CONVENTIONAL CCTV CAMERA N. Zarrinpanjeh a, F. Dadrassjavan b, H. Fattahi c * a Islamic Azad University of Qazvin - [email protected]
Geometric Camera Parameters
Geometric Camera Parameters What assumptions have we made so far? -All equations we have derived for far are written in the camera reference frames. -These equations are valid only when: () all distances
CALIBRATION OF A ROBUST 2 DOF PATH MONITORING TOOL FOR INDUSTRIAL ROBOTS AND MACHINE TOOLS BASED ON PARALLEL KINEMATICS
CALIBRATION OF A ROBUST 2 DOF PATH MONITORING TOOL FOR INDUSTRIAL ROBOTS AND MACHINE TOOLS BASED ON PARALLEL KINEMATICS E. Batzies 1, M. Kreutzer 1, D. Leucht 2, V. Welker 2, O. Zirn 1 1 Mechatronics Research
Introduction Epipolar Geometry Calibration Methods Further Readings. Stereo Camera Calibration
Stereo Camera Calibration Stereo Camera Calibration Stereo Camera Calibration Stereo Camera Calibration 12.10.2004 Overview Introduction Summary / Motivation Depth Perception Ambiguity of Correspondence
Face Model Fitting on Low Resolution Images
Face Model Fitting on Low Resolution Images Xiaoming Liu Peter H. Tu Frederick W. Wheeler Visualization and Computer Vision Lab General Electric Global Research Center Niskayuna, NY, 1239, USA {liux,tu,wheeler}@research.ge.com
B4 Computational Geometry
3CG 2006 / B4 Computational Geometry David Murray [email protected] www.robots.o.ac.uk/ dwm/courses/3cg Michaelmas 2006 3CG 2006 2 / Overview Computational geometry is concerned with the derivation
Experiment #1, Analyze Data using Excel, Calculator and Graphs.
Physics 182 - Fall 2014 - Experiment #1 1 Experiment #1, Analyze Data using Excel, Calculator and Graphs. 1 Purpose (5 Points, Including Title. Points apply to your lab report.) Before we start measuring
11.1. Objectives. Component Form of a Vector. Component Form of a Vector. Component Form of a Vector. Vectors and the Geometry of Space
11 Vectors and the Geometry of Space 11.1 Vectors in the Plane Copyright Cengage Learning. All rights reserved. Copyright Cengage Learning. All rights reserved. 2 Objectives! Write the component form of
Tracking Moving Objects In Video Sequences Yiwei Wang, Robert E. Van Dyck, and John F. Doherty Department of Electrical Engineering The Pennsylvania State University University Park, PA16802 Abstract{Object
Epipolar Geometry. Readings: See Sections 10.1 and 15.6 of Forsyth and Ponce. Right Image. Left Image. e(p ) Epipolar Lines. e(q ) q R.
Epipolar Geometry We consider two perspective images of a scene as taken from a stereo pair of cameras (or equivalently, assume the scene is rigid and imaged with a single camera from two different locations).
Common Core Unit Summary Grades 6 to 8
Common Core Unit Summary Grades 6 to 8 Grade 8: Unit 1: Congruence and Similarity- 8G1-8G5 rotations reflections and translations,( RRT=congruence) understand congruence of 2 d figures after RRT Dilations
Projective Reconstruction from Line Correspondences
Projective Reconstruction from Line Correspondences Richard I. Hartley G.E. CRD, Schenectady, NY, 12301. Abstract The paper gives a practical rapid algorithm for doing projective reconstruction of a scene
Human-like Arm Motion Generation for Humanoid Robots Using Motion Capture Database
Human-like Arm Motion Generation for Humanoid Robots Using Motion Capture Database Seungsu Kim, ChangHwan Kim and Jong Hyeon Park School of Mechanical Engineering Hanyang University, Seoul, 133-791, Korea.
High-Resolution Multiscale Panoramic Mosaics from Pan-Tilt-Zoom Cameras
High-Resolution Multiscale Panoramic Mosaics from Pan-Tilt-Zoom Cameras Sudipta N. Sinha Marc Pollefeys Seon Joo Kim Department of Computer Science, University of North Carolina at Chapel Hill. Abstract
Segmentation of building models from dense 3D point-clouds
Segmentation of building models from dense 3D point-clouds Joachim Bauer, Konrad Karner, Konrad Schindler, Andreas Klaus, Christopher Zach VRVis Research Center for Virtual Reality and Visualization, Institute
Automated Pose Determination for Unrestrained, Non-anesthetized Small Animal Micro-SPECT and Micro-CT Imaging
Automated Pose Determination for Unrestrained, Non-anesthetized Small Animal Micro-SPECT and Micro-CT Imaging Shaun S. Gleason, Ph.D. 1, James Goddard, Ph.D. 1, Michael Paulus, Ph.D. 1, Ryan Kerekes, B.S.
Force/position control of a robotic system for transcranial magnetic stimulation
Force/position control of a robotic system for transcranial magnetic stimulation W.N. Wan Zakaria School of Mechanical and System Engineering Newcastle University Abstract To develop a force control scheme
CIS 536/636 Introduction to Computer Graphics. Kansas State University. CIS 536/636 Introduction to Computer Graphics
2 Lecture Outline Animation 2 of 3: Rotations, Quaternions Dynamics & Kinematics William H. Hsu Department of Computing and Information Sciences, KSU KSOL course pages: http://bit.ly/hgvxlh / http://bit.ly/evizre
Projection Center Calibration for a Co-located Projector Camera System
Projection Center Calibration for a Co-located Camera System Toshiyuki Amano Department of Computer and Communication Science Faculty of Systems Engineering, Wakayama University Sakaedani 930, Wakayama,
3D Model based Object Class Detection in An Arbitrary View
3D Model based Object Class Detection in An Arbitrary View Pingkun Yan, Saad M. Khan, Mubarak Shah School of Electrical Engineering and Computer Science University of Central Florida http://www.eecs.ucf.edu/
Enhancing the SNR of the Fiber Optic Rotation Sensor using the LMS Algorithm
1 Enhancing the SNR of the Fiber Optic Rotation Sensor using the LMS Algorithm Hani Mehrpouyan, Student Member, IEEE, Department of Electrical and Computer Engineering Queen s University, Kingston, Ontario,
Fast Optimal Three View Triangulation
Fast Optimal Three View Triangulation Martin Byröd, Klas Josephson and Kalle Åström {byrod, klasj, kalle}@maths.lth.se Center for Mathematical Sciences, Lund University, Lund, Sweden Abstract. We consider
Metrics on SO(3) and Inverse Kinematics
Mathematical Foundations of Computer Graphics and Vision Metrics on SO(3) and Inverse Kinematics Luca Ballan Institute of Visual Computing Optimization on Manifolds Descent approach d is a ascent direction
Make and Model Recognition of Cars
Make and Model Recognition of Cars Sparta Cheung Department of Electrical and Computer Engineering University of California, San Diego 9500 Gilman Dr., La Jolla, CA 92093 [email protected] Alice Chu Department
Removing Moving Objects from Point Cloud Scenes
1 Removing Moving Objects from Point Cloud Scenes Krystof Litomisky [email protected] Abstract. Three-dimensional simultaneous localization and mapping is a topic of significant interest in the research
Shape Measurement of a Sewer Pipe. Using a Mobile Robot with Computer Vision
International Journal of Advanced Robotic Systems ARTICLE Shape Measurement of a Sewer Pipe Using a Mobile Robot with Computer Vision Regular Paper Kikuhito Kawasue 1,* and Takayuki Komatsu 1 1 Department
Mechanics lecture 7 Moment of a force, torque, equilibrium of a body
G.1 EE1.el3 (EEE1023): Electronics III Mechanics lecture 7 Moment of a force, torque, equilibrium of a body Dr Philip Jackson http://www.ee.surrey.ac.uk/teaching/courses/ee1.el3/ G.2 Moments, torque and
Robot Perception Continued
Robot Perception Continued 1 Visual Perception Visual Odometry Reconstruction Recognition CS 685 11 Range Sensing strategies Active range sensors Ultrasound Laser range sensor Slides adopted from Siegwart
Solving Simultaneous Equations and Matrices
Solving Simultaneous Equations and Matrices The following represents a systematic investigation for the steps used to solve two simultaneous linear equations in two unknowns. The motivation for considering
Optimal Design of α-β-(γ) Filters
Optimal Design of --(γ) Filters Dirk Tenne Tarunraj Singh, Center for Multisource Information Fusion State University of New York at Buffalo Buffalo, NY 426 Abstract Optimal sets of the smoothing parameter
Lighting Estimation in Indoor Environments from Low-Quality Images
Lighting Estimation in Indoor Environments from Low-Quality Images Natalia Neverova, Damien Muselet, Alain Trémeau Laboratoire Hubert Curien UMR CNRS 5516, University Jean Monnet, Rue du Professeur Benoît
Epipolar Geometry and Visual Servoing
Epipolar Geometry and Visual Servoing Domenico Prattichizzo joint with with Gian Luca Mariottini and Jacopo Piazzi www.dii.unisi.it/prattichizzo Robotics & Systems Lab University of Siena, Italy Scuoladi
PHOTOGRAMMETRIC TECHNIQUES FOR MEASUREMENTS IN WOODWORKING INDUSTRY
PHOTOGRAMMETRIC TECHNIQUES FOR MEASUREMENTS IN WOODWORKING INDUSTRY V. Knyaz a, *, Yu. Visilter, S. Zheltov a State Research Institute for Aviation System (GosNIIAS), 7, Victorenko str., Moscow, Russia
An Iterative Image Registration Technique with an Application to Stereo Vision
An Iterative Image Registration Technique with an Application to Stereo Vision Bruce D. Lucas Takeo Kanade Computer Science Department Carnegie-Mellon University Pittsburgh, Pennsylvania 15213 Abstract
Classifying Manipulation Primitives from Visual Data
Classifying Manipulation Primitives from Visual Data Sandy Huang and Dylan Hadfield-Menell Abstract One approach to learning from demonstrations in robotics is to make use of a classifier to predict if
UAV Pose Estimation using POSIT Algorithm
International Journal of Digital ontent Technology and its Applications. Volume 5, Number 4, April 211 UAV Pose Estimation using POSIT Algorithm *1 M. He, 2. Ratanasawanya, 3 M. Mehrandezh, 4 R. Paranjape
We can display an object on a monitor screen in three different computer-model forms: Wireframe model Surface Model Solid model
CHAPTER 4 CURVES 4.1 Introduction In order to understand the significance of curves, we should look into the types of model representations that are used in geometric modeling. Curves play a very significant
Metric Measurements on a Plane from a Single Image
TR26-579, Dartmouth College, Computer Science Metric Measurements on a Plane from a Single Image Micah K. Johnson and Hany Farid Department of Computer Science Dartmouth College Hanover NH 3755 Abstract
Time Domain and Frequency Domain Techniques For Multi Shaker Time Waveform Replication
Time Domain and Frequency Domain Techniques For Multi Shaker Time Waveform Replication Thomas Reilly Data Physics Corporation 1741 Technology Drive, Suite 260 San Jose, CA 95110 (408) 216-8440 This paper
Subspace Analysis and Optimization for AAM Based Face Alignment
Subspace Analysis and Optimization for AAM Based Face Alignment Ming Zhao Chun Chen College of Computer Science Zhejiang University Hangzhou, 310027, P.R.China [email protected] Stan Z. Li Microsoft
REAL TIME TRAFFIC LIGHT CONTROL USING IMAGE PROCESSING
REAL TIME TRAFFIC LIGHT CONTROL USING IMAGE PROCESSING Ms.PALLAVI CHOUDEKAR Ajay Kumar Garg Engineering College, Department of electrical and electronics Ms.SAYANTI BANERJEE Ajay Kumar Garg Engineering
Flow Separation for Fast and Robust Stereo Odometry
Flow Separation for Fast and Robust Stereo Odometry Michael Kaess, Kai Ni and Frank Dellaert Abstract Separating sparse flow provides fast and robust stereo visual odometry that deals with nearly degenerate
Automatic Labeling of Lane Markings for Autonomous Vehicles
Automatic Labeling of Lane Markings for Autonomous Vehicles Jeffrey Kiske Stanford University 450 Serra Mall, Stanford, CA 94305 [email protected] 1. Introduction As autonomous vehicles become more popular,
Static Environment Recognition Using Omni-camera from a Moving Vehicle
Static Environment Recognition Using Omni-camera from a Moving Vehicle Teruko Yata, Chuck Thorpe Frank Dellaert The Robotics Institute Carnegie Mellon University Pittsburgh, PA 15213 USA College of Computing
3D Tranformations. CS 4620 Lecture 6. Cornell CS4620 Fall 2013 Lecture 6. 2013 Steve Marschner (with previous instructors James/Bala)
3D Tranformations CS 4620 Lecture 6 1 Translation 2 Translation 2 Translation 2 Translation 2 Scaling 3 Scaling 3 Scaling 3 Scaling 3 Rotation about z axis 4 Rotation about z axis 4 Rotation about x axis
Introduction to Computer Graphics
Introduction to Computer Graphics Torsten Möller TASC 8021 778-782-2215 [email protected] www.cs.sfu.ca/~torsten Today What is computer graphics? Contents of this course Syllabus Overview of course topics
15.062 Data Mining: Algorithms and Applications Matrix Math Review
.6 Data Mining: Algorithms and Applications Matrix Math Review The purpose of this document is to give a brief review of selected linear algebra concepts that will be useful for the course and to develop
Parameter identification of a linear single track vehicle model
Parameter identification of a linear single track vehicle model Edouard Davin D&C 2011.004 Traineeship report Coach: dr. Ir. I.J.M. Besselink Supervisors: prof. dr. H. Nijmeijer Eindhoven University of
Multiple View Geometry in Computer Vision
Multiple View Geometry in Computer Vision Information Meeting 15 Dec 2009 by Marianna Pronobis Content of the Meeting 1. Motivation & Objectives 2. What the course will be about (Stefan) 3. Content of
A Reliability Point and Kalman Filter-based Vehicle Tracking Technique
A Reliability Point and Kalman Filter-based Vehicle Tracing Technique Soo Siang Teoh and Thomas Bräunl Abstract This paper introduces a technique for tracing the movement of vehicles in consecutive video
Robotics. Lecture 3: Sensors. See course website http://www.doc.ic.ac.uk/~ajd/robotics/ for up to date information.
Robotics Lecture 3: Sensors See course website http://www.doc.ic.ac.uk/~ajd/robotics/ for up to date information. Andrew Davison Department of Computing Imperial College London Review: Locomotion Practical
Introduction. www.imagesystems.se
Product information Image Systems AB Main office: Ågatan 40, SE-582 22 Linköping Phone +46 13 200 100, fax +46 13 200 150 [email protected], Introduction Motion is the world leading software for advanced
High-accuracy ultrasound target localization for hand-eye calibration between optical tracking systems and three-dimensional ultrasound
High-accuracy ultrasound target localization for hand-eye calibration between optical tracking systems and three-dimensional ultrasound Ralf Bruder 1, Florian Griese 2, Floris Ernst 1, Achim Schweikard
3D Annotation and Manipulation of Medical Anatomical Structures
3D Annotation and Manipulation of Medical Anatomical Structures Dime Vitanovski, Christian Schaller, Dieter Hahn, Volker Daum, Joachim Hornegger Chair of Pattern Recognition, Martensstr. 3, 91058 Erlangen,
Origins of the Unusual Space Shuttle Quaternion Definition
47th AIAA Aerospace Sciences Meeting Including The New Horizons Forum and Aerospace Exposition 5-8 January 2009, Orlando, Florida AIAA 2009-43 Origins of the Unusual Space Shuttle Quaternion Definition
Dynamic composition of tracking primitives for interactive vision-guided navigation
Dynamic composition of tracking primitives for interactive vision-guided navigation Darius Burschka and Gregory Hager Johns Hopkins University, Baltimore, USA ABSTRACT We present a system architecture
CS 534: Computer Vision 3D Model-based recognition
CS 534: Computer Vision 3D Model-based recognition Ahmed Elgammal Dept of Computer Science CS 534 3D Model-based Vision - 1 High Level Vision Object Recognition: What it means? Two main recognition tasks:!
Monash University Clayton s School of Information Technology CSE3313 Computer Graphics Sample Exam Questions 2007
Monash University Clayton s School of Information Technology CSE3313 Computer Graphics Questions 2007 INSTRUCTIONS: Answer all questions. Spend approximately 1 minute per mark. Question 1 30 Marks Total
CS231M Project Report - Automated Real-Time Face Tracking and Blending
CS231M Project Report - Automated Real-Time Face Tracking and Blending Steven Lee, [email protected] June 6, 2015 1 Introduction Summary statement: The goal of this project is to create an Android
MetropoGIS: A City Modeling System DI Dr. Konrad KARNER, DI Andreas KLAUS, DI Joachim BAUER, DI Christopher ZACH
MetropoGIS: A City Modeling System DI Dr. Konrad KARNER, DI Andreas KLAUS, DI Joachim BAUER, DI Christopher ZACH VRVis Research Center for Virtual Reality and Visualization, Virtual Habitat, Inffeldgasse
Reflection and Refraction
Equipment Reflection and Refraction Acrylic block set, plane-concave-convex universal mirror, cork board, cork board stand, pins, flashlight, protractor, ruler, mirror worksheet, rectangular block worksheet,
Factoring Patterns in the Gaussian Plane
Factoring Patterns in the Gaussian Plane Steve Phelps Introduction This paper describes discoveries made at the Park City Mathematics Institute, 00, as well as some proofs. Before the summer I understood
Integer Computation of Image Orthorectification for High Speed Throughput
Integer Computation of Image Orthorectification for High Speed Throughput Paul Sundlie Joseph French Eric Balster Abstract This paper presents an integer-based approach to the orthorectification of aerial
Design-Simulation-Optimization Package for a Generic 6-DOF Manipulator with a Spherical Wrist
Design-Simulation-Optimization Package for a Generic 6-DOF Manipulator with a Spherical Wrist MHER GRIGORIAN, TAREK SOBH Department of Computer Science and Engineering, U. of Bridgeport, USA ABSTRACT Robot
SIMPLIFIED PERFORMANCE MODEL FOR HYBRID WIND DIESEL SYSTEMS. J. F. MANWELL, J. G. McGOWAN and U. ABDULWAHID
SIMPLIFIED PERFORMANCE MODEL FOR HYBRID WIND DIESEL SYSTEMS J. F. MANWELL, J. G. McGOWAN and U. ABDULWAHID Renewable Energy Laboratory Department of Mechanical and Industrial Engineering University of
Basic Principles of Inertial Navigation. Seminar on inertial navigation systems Tampere University of Technology
Basic Principles of Inertial Navigation Seminar on inertial navigation systems Tampere University of Technology 1 The five basic forms of navigation Pilotage, which essentially relies on recognizing landmarks
Tracking of Small Unmanned Aerial Vehicles
Tracking of Small Unmanned Aerial Vehicles Steven Krukowski Adrien Perkins Aeronautics and Astronautics Stanford University Stanford, CA 94305 Email: [email protected] Aeronautics and Astronautics Stanford
Lecture 2 Linear functions and examples
EE263 Autumn 2007-08 Stephen Boyd Lecture 2 Linear functions and examples linear equations and functions engineering examples interpretations 2 1 Linear equations consider system of linear equations y
A PAIR OF MEASURES OF ROTATIONAL ERROR FOR AXISYMMETRIC ROBOT END-EFFECTORS
A PAIR OF MEASURES OF ROTATIONAL ERROR FOR AXISYMMETRIC ROBOT END-EFFECTORS Sébastien Briot, Ilian A. Bonev Department of Automated Manufacturing Engineering École de technologie supérieure (ÉTS), Montreal,
Machine Learning and Data Mining. Regression Problem. (adapted from) Prof. Alexander Ihler
Machine Learning and Data Mining Regression Problem (adapted from) Prof. Alexander Ihler Overview Regression Problem Definition and define parameters ϴ. Prediction using ϴ as parameters Measure the error
CONTRIBUTIONS TO THE AUTOMATIC CONTROL OF AERIAL VEHICLES
1 / 23 CONTRIBUTIONS TO THE AUTOMATIC CONTROL OF AERIAL VEHICLES MINH DUC HUA 1 1 INRIA Sophia Antipolis, AROBAS team I3S-CNRS Sophia Antipolis, CONDOR team Project ANR SCUAV Supervisors: Pascal MORIN,
FREE FALL. Introduction. Reference Young and Freedman, University Physics, 12 th Edition: Chapter 2, section 2.5
Physics 161 FREE FALL Introduction This experiment is designed to study the motion of an object that is accelerated by the force of gravity. It also serves as an introduction to the data analysis capabilities
K.Prasanna NIT Rourkela,India Summer Internship NUS
K.Prasanna NIT Rourkela,India Summer Internship NUS LOCALIZATION Vehicle localization is an important topic as it finds its application in a multitude of fields like autonomous navigation and tracking
A DECISION TREE BASED PEDOMETER AND ITS IMPLEMENTATION ON THE ANDROID PLATFORM
A DECISION TREE BASED PEDOMETER AND ITS IMPLEMENTATION ON THE ANDROID PLATFORM ABSTRACT Juanying Lin, Leanne Chan and Hong Yan Department of Electronic Engineering, City University of Hong Kong, Hong Kong,
How does the Kinect work? John MacCormick
How does the Kinect work? John MacCormick Xbox demo Laptop demo The Kinect uses structured light and machine learning Inferring body position is a two-stage process: first compute a depth map (using structured
Accurate and robust image superresolution by neural processing of local image representations
Accurate and robust image superresolution by neural processing of local image representations Carlos Miravet 1,2 and Francisco B. Rodríguez 1 1 Grupo de Neurocomputación Biológica (GNB), Escuela Politécnica
Building an Advanced Invariant Real-Time Human Tracking System
UDC 004.41 Building an Advanced Invariant Real-Time Human Tracking System Fayez Idris 1, Mazen Abu_Zaher 2, Rashad J. Rasras 3, and Ibrahiem M. M. El Emary 4 1 School of Informatics and Computing, German-Jordanian
Projective Geometry. Projective Geometry
Euclidean versus Euclidean geometry describes sapes as tey are Properties of objects tat are uncanged by rigid motions» Lengts» Angles» Parallelism Projective geometry describes objects as tey appear Lengts,
