VOLUMNECT - Measuring Volumes with Kinect T M
|
|
|
- Kerry Marshall
- 9 years ago
- Views:
Transcription
1 VOLUMNECT - Measuring Volumes with Kinect T M Beatriz Quintino Ferreira a, Miguel Griné a, Duarte Gameiro a, João Paulo Costeira a,b and Beatriz Sousa Santos c,d a DEEC, Instituto Superior Técnico, Lisboa, Portugal b Instituto de Sistemas e Robótica, Instituto Superior Técnico, Lisboa, Portugal c Dep. Electrónica, Telecomunicações e Informática, Universidade de Aveiro, Aveiro, Portugal d Instituto de Engenharia Electrónica e Telemática de Aveiro, Aveiro, Portugal ABSTRACT This article presents a solution to volume measurement object packing using 3D cameras (such as the Microsoft Kinect T M ). We target application scenarios, such as warehouses or distribution and logistics companies, where it is important to promptly compute package volumes, yet high accuracy is not pivotal. Our application automatically detects cuboid objects using the depth camera data and computes their volume and sorting it allowing space optimization. The proposed methodology applies to a point cloud simple computer vision and image processing methods, as connected components, morphological operations and Harris corner detector, producing encouraging results, namely an accuracy in volume measurement of 8mm. Aspects that can be further improved are identified; nevertheless, the current solution is already promising turning out to be cost effective for the envisaged scenarios. Keywords: volume measurement, Microsoft Kinect T M, 3D depth camera 1. INTRODUCTION This article presents an approach to volume measurement using Microsoft Kinect T M, which has several application scenarios such as warehouses or distribution and logistics companies, where it is important to promptly compute package volumes. For these environments, instead of using the standardized and pricey laser based technology (e.g. laser-range finders which cost several thousands of dollars), we propose a low-cost solution using the widely available plus quite affordable Microsoft Kinect T M (with price approximately 100$). In such situations, where the laser high accuracy (around 5mm) is not mandatory, we can automatically detect boxy objects from the depth camera data and compute their volume, paving the way for further space optimization. The proposed methodology, based on simple computer vision and image processing methods 1, provides a new and cost-effective solution for the industry. Among the several methods used to process the Kinect T M depth camera information, we highlight: RANSAC ( RANdom SAmple Consensus 2 ) which allows to find the planes contained in the scene; connected components 3 in order to segment the object faces; morphological operations 4 to filter the faces and, also, the Harris corner detector 5 to find possible vertices. The attained results are promising, with accuracy close to the laser based volume measurement commercial systems. The remainder of the paper is organized as follows. Section 2 includes a brief description of the 3D camera device used (Kinect T M ), as well as the camera calibration performed. In Section 3 the system work-flow and the methods used to compute the volume of the regular objects are described. The different tests performed and results obtained are presented in Section 4. Lastly, in Section 5 some conclusions are drawn and future work directions are discussed. Beatriz Quintino Ferreira: [email protected] Miguel Griné: [email protected] Duarte Gameiro: [email protected]
2 2. KINECT T M AND CALIBRATION Kinect T M is a motion sensing input device by Microsoft R. 6 This device features a RGB camera and depth camera (which comprises an infrared laser projector and a depth sensor) at a very affordable price (approximately, 100$). The recent availability of low-cost Kinect T M sensor provides a valid alternative to other available sensors, namely laser-range finders and stereo-vision systems. The Kinect T M depth sensor brings a fast and relatively high resolution solution for depth sensing, while being considerably cheaper than the aforementioned alternatives. Nevertheless, it has some limitations, specifically, this device was designed to work indoors due to light conditions. Moreover, its range is up to 4m and can not acquire valid data at distances smaller than 0.8m. Figure 1 shows its main components, namely the RGB camera and depth sensors, a multi-array microphone and a motorized tilt. Figure 1. Microsft R Kinect T M : components (adapted from Standalone.png) An important previous step, before using the 3D depth camera data to compute volumes, concerns the camera calibration. The linear camera model which was used to calibrate the Kinect T M camera is presented below (in eq. 1): x1 X x2 = K [ R T ] Y Z (1) x3 1 where K is the intrinsic parameters matrix, while R and T are the extrinsic parameters (rotational and translational matrices, respectively). The intrinsic parameters are the conversion factors from meters to pixels and the coordinates of the principal point in the new frame. These parameters are denoted as intrinsic parameters of the camera since they only depend on the camera and are independent of its position and orientation. If we choose a 3D frame that is neither attached to the camera nor located at the optical center of the camera, in other words, with an arbitrary location and orientation, the new coordinates are related to the camera coordinates by a rigid body transformation (a rotation, followed by a translation). Thus extrinsic parameters are R, the rotational matrix, and T, the translational matrix. In our case, since the depth camera referential is the same that the world s referential, we considered that R is equal to identity matrix and T is zero. The process of calibrating the camera consists, basically, in estimating the camera parameters from an image of the calibration object with known geometry (for instance, a chess board), since it contains a set of easily detected points, which can be used as the calibration points. The camera parameters can be chosen in order to minimize the distance from the points projected by the camera and by the model. Since we have an optimization problem (which is non-convex), we applied a linear regression (least squares), using the pseudo-inverse method to solve it. We performed the calibration of Kinect T M as described, nevertheless, it is possible to find alternative, ready to use calibrations (e.g. 7 and 8 ).
3 3. VOLUME MEASUREMENT METHODS To achieve our goal of determining the volume of a cuboid shaped object, it is necessary to measure three edges in order to have the length, width and height of the box. Hence, our approach focuses on finding the inliers of the planes of the scene and, subsequently, the vertices that connect adjacent planes, so that it is possible to determine each edge length. The proposed application was developed in MATLAB R (R2009b). 9 Next, the system workflow is presented, in figure 2, illustrating the application high-level operation. The methods are further discussed, afterwards. Compute 3D coordinates Compute set of 3 orthogonal planes Compute length, width and height Determine planes of the image Obtain the vertices of each orthogonal plane Obtain angles between planes Compute the common vertex Calculate VOLUME Figure 2. Workflow of the developed system, allowing to compute volumes of cuboid objects Considering a known and clean environment with a cuboid object, for instance a box, both RGB and depth images are acquired from the cameras, as it can be seen in figures 3 and 4 (please note that the position of the box is very important, since the three edges to be computed have to be entirely visible; a solution to this constrain will be further discussed). Figure 3. RGB image Figure 4. Depth image One should note that this implies a previous calibration of the Kinect T M, using the camera model (as described in Section 2). From the acquisition step, the depth information is used to compute the 3D coordinates (X Y Z), using again the camera model, and a 3D point cloud representation is obtained. During this process, the invalid point readings caused by points out of range or lighting conditions are removed from the set (these points are depicted in red in figure 4). Invalid readings are retrieved by Kinect T M with a special code number, out of the range. With the obtained point cloud, the inliers of the planes shown in figure 5 can be determined applying the RANSAC method. 2
4 Figure 5. Point cloud with the planes inliers; representing different planes using different colors Having the inliers of each plane of the scene, considering a clean and known environment, we remove the two planes with more inliers which correspond to the background. Alternatively, in order to segment the object to be measured, we could perform background subtraction. In our first experimental setup (described in Section Results) we verify these assumptions, which in spite of being simplifying assumptions, we believe they are easily applicable in the envisaged scenarios. Once the planes of the object are delimited, the angle between each pair of neighboring planes is computed. From the obtained angles a set of three adjacent and orthogonal planes is selected so that we can determine the right edges to compute the volume. This selection is a very important step and it is performed using a clique problem approach. 10 The processing proceeds with the projection of each of the three orthogonal planes delimited into a 2D binary image, which will help the subsequent steps. This projection is performed using labels of connected components. 3 Due to some faults that occur during the capture with the camera, as mentioned in Section 1, morphological operations 4 are applied to filter these 2D images. We used mainly the dilation and closure operations, since the detected faces were irregular and showing some missing data points. This filtering is rather important for the next phase, which consists in applying the Harris corner detector 5 to every segmented face, in order to find possible vertices. Figure 6. Planes represented as binary images, with the detected corners depicted and the vertex common to the three orthogonal planes pointed out with an arrow The next step is to compute the corners which are of interest, in order to compute the lengths between them and, then, the final goal: the object volume. To find these special points we used the Harris corner detector; yet, the parameter fine-tuning proved arduous. We tested several values in order to achieve the best sensitivity (find the persued vertices), whereby the final parameters were a combination of a very small threshold with a large radius in non-maximal suppression. After applying the detector, it is crucial to determine which vertex, among all the corners found to be possible vertices, is the common vertex; in other words, we have to look for the vertex which belongs to the three orthogonal planes (emphasized in figure 6). Therefore, we are facing a minimization problem, since we are verifying which vertex has the smaller Euclidean distance to each plane. After that, the next step is to determine the vertex in the other end of each edge computing the maximum Euclidean distance
5 among the corners found for each couple of planes. Then, since we already have the four vertices we need, it is possible to measure the three lengths required to compute the volume of the box, as represented in figure 7. To do that, we just need to map the sets of two vertices (the common vertex of the three orthogonal planes plus the vertex on the opposite side of the edge which are in 2D binary images) into the real world coordinates (3D - point cloud), using the camera model. Figure 7. Common vertex and the vertices used to compute the length of the three edges These lengths are computed through Euclidean distances between each couple of vertices, and the object volume is, thereafter, readily calculated as the product of these lengths. When the common vertex for the three orthogonal planes is not found, it is not possible to calculate the volume. In such cases the application displays an error message asking the user to move the object so that this vertex can be computed. This allows to improve our solution robustness, since no inconsistent values are obtained as the final result. 4. RESULTS In order to evaluate the solution, accuracy and precision studies were conducted using data sets comprising several cuboid objects. The experimental setup used included a Kinect T M placed at about 1m distance from the object at a level slightly above the box, so that it is easy to obtain an adequate view of the three main faces of the object. This was intended to replicate the positioning in an real application scenario. Our data set comprised four boxes of different sizes and proportions, however, for all cases, the largest edge was smaller than 1m. Figures from 8 to 10 show the histograms corresponding to the three edges of a specific box, for which we performed 20 tests. The box real dimensions were 0.375m, 0.145m and 0.118m, length, width and height, respectively. As the histograms show, the maximum error is 0.008m for the length (2%), 0.007m for width (4.8%) and 0,018m for height (15%). These results were found to be representative considering all the performed tests. Figure 11 shows the volume computed, revealing that the maximum error is m 3 (15,6%), as the real volume was m 3. Figure 8. Histogram of 20 measurements of a box length Figure 9. Histogram of 20 measurements of a box width Figure 10. Histogram of 20 measurements of a box height
6 Figure 11. Histogram of 20 computation of a box volume 5. CONCLUSIONS AND FUTURE WORK This article presents a methodology to compute volumes of cuboid objects based on the data acquired by the depth camera of Microsoft Kinect T M. The volumes of the objects from our data set were successfully computed with an accuracy suggesting an adequate compromise between performance and cost regarding the identified application scenarios. Based on the experiments performed, we believe that better accuracy and precision could be obtained acting upon the following processing steps: improving the Harris corner detector parameters and exploring RGB information. Specifically, the RGB camera data could be used to help segmenting the objects (as faces from the same object tend to have a similar color palette distinct from the background), also allowing the segmentation of superimposed objects, as long as they have different colors. Another line of work would be improving the independence from the background, so that our solution would have a good performance in any environment, regardless the background (thus eliminating the restriction of a known and clean environment). Likewise, background subtraction could be applied using the RGB information. Moreover, a second experimental setup using an array of two or more depth cameras is envisaged. Provided that all cameras are calibrated and the transformation among them is known, this setup could produce more precise results and enable the usage in additional scenarios, since with more cameras a wider region is covered, thus allowing measuring larger volumes. REFERENCES 1. R. Szeliski, Computer Vision: Algorithms and Applications, draft Springer, M. A. Fischler and R. C. Bolles, Random sample consensus: A paradigm for model fitting with applications to image analysis and automated cartography, Comm. of the ACM 24(6), pp , H. Samet and M. Tamminen, Efficient component labeling of images of arbitrary dimension represented by linear bintrees, IEEE Transactions on Pattern Analysis and Machine Intelligence 10(4), pp , J. Serra and P. Soille, Mathematical Morphology and Its Applications to Image Processing, in Proceedings of the 2nd International Symposium on Mathematical Morphology (ISMM 94), C. Harris and M. Stephens, A combined corner and edge detector, in Proceedings of the 4th Alvey Vision Conference, pp , Kinect sensor. Accessed: November Kinect Calibration Toolbox. Accessed: November Kinect Calibration. kinect-calibration/. Accessed: November MATLAB - The Language of Technical Computing. Accessed: December R. E. Tarjan and A. E. Trojanowski, Finding a maximum independent set, SIAM Journal on Computing 6(3), pp , 1977.
Build Panoramas on Android Phones
Build Panoramas on Android Phones Tao Chu, Bowen Meng, Zixuan Wang Stanford University, Stanford CA Abstract The purpose of this work is to implement panorama stitching from a sequence of photos taken
3D Scanner using Line Laser. 1. Introduction. 2. Theory
. Introduction 3D Scanner using Line Laser Di Lu Electrical, Computer, and Systems Engineering Rensselaer Polytechnic Institute The goal of 3D reconstruction is to recover the 3D properties of a geometric
Face detection is a process of localizing and extracting the face region from the
Chapter 4 FACE NORMALIZATION 4.1 INTRODUCTION Face detection is a process of localizing and extracting the face region from the background. The detected face varies in rotation, brightness, size, etc.
A PHOTOGRAMMETRIC APPRAOCH FOR AUTOMATIC TRAFFIC ASSESSMENT USING CONVENTIONAL CCTV CAMERA
A PHOTOGRAMMETRIC APPRAOCH FOR AUTOMATIC TRAFFIC ASSESSMENT USING CONVENTIONAL CCTV CAMERA N. Zarrinpanjeh a, F. Dadrassjavan b, H. Fattahi c * a Islamic Azad University of Qazvin - [email protected]
Self-Calibrated Structured Light 3D Scanner Using Color Edge Pattern
Self-Calibrated Structured Light 3D Scanner Using Color Edge Pattern Samuel Kosolapov Department of Electrical Engineering Braude Academic College of Engineering Karmiel 21982, Israel e-mail: [email protected]
Classifying Manipulation Primitives from Visual Data
Classifying Manipulation Primitives from Visual Data Sandy Huang and Dylan Hadfield-Menell Abstract One approach to learning from demonstrations in robotics is to make use of a classifier to predict if
Solving Simultaneous Equations and Matrices
Solving Simultaneous Equations and Matrices The following represents a systematic investigation for the steps used to solve two simultaneous linear equations in two unknowns. The motivation for considering
Analecta Vol. 8, No. 2 ISSN 2064-7964
EXPERIMENTAL APPLICATIONS OF ARTIFICIAL NEURAL NETWORKS IN ENGINEERING PROCESSING SYSTEM S. Dadvandipour Institute of Information Engineering, University of Miskolc, Egyetemváros, 3515, Miskolc, Hungary,
Automatic Labeling of Lane Markings for Autonomous Vehicles
Automatic Labeling of Lane Markings for Autonomous Vehicles Jeffrey Kiske Stanford University 450 Serra Mall, Stanford, CA 94305 [email protected] 1. Introduction As autonomous vehicles become more popular,
Make and Model Recognition of Cars
Make and Model Recognition of Cars Sparta Cheung Department of Electrical and Computer Engineering University of California, San Diego 9500 Gilman Dr., La Jolla, CA 92093 [email protected] Alice Chu Department
Segmentation of building models from dense 3D point-clouds
Segmentation of building models from dense 3D point-clouds Joachim Bauer, Konrad Karner, Konrad Schindler, Andreas Klaus, Christopher Zach VRVis Research Center for Virtual Reality and Visualization, Institute
Locating and Decoding EAN-13 Barcodes from Images Captured by Digital Cameras
Locating and Decoding EAN-13 Barcodes from Images Captured by Digital Cameras W3A.5 Douglas Chai and Florian Hock Visual Information Processing Research Group School of Engineering and Mathematics Edith
Epipolar Geometry. Readings: See Sections 10.1 and 15.6 of Forsyth and Ponce. Right Image. Left Image. e(p ) Epipolar Lines. e(q ) q R.
Epipolar Geometry We consider two perspective images of a scene as taken from a stereo pair of cameras (or equivalently, assume the scene is rigid and imaged with a single camera from two different locations).
DINAMIC AND STATIC CENTRE OF PRESSURE MEASUREMENT ON THE FORCEPLATE. F. R. Soha, I. A. Szabó, M. Budai. Abstract
ACTA PHYSICA DEBRECINA XLVI, 143 (2012) DINAMIC AND STATIC CENTRE OF PRESSURE MEASUREMENT ON THE FORCEPLATE F. R. Soha, I. A. Szabó, M. Budai University of Debrecen, Department of Solid State Physics Abstract
Poker Vision: Playing Cards and Chips Identification based on Image Processing
Poker Vision: Playing Cards and Chips Identification based on Image Processing Paulo Martins 1, Luís Paulo Reis 2, and Luís Teófilo 2 1 DEEC Electrical Engineering Department 2 LIACC Artificial Intelligence
Wii Remote Calibration Using the Sensor Bar
Wii Remote Calibration Using the Sensor Bar Alparslan Yildiz Abdullah Akay Yusuf Sinan Akgul GIT Vision Lab - http://vision.gyte.edu.tr Gebze Institute of Technology Kocaeli, Turkey {yildiz, akay, akgul}@bilmuh.gyte.edu.tr
Mobile Robot FastSLAM with Xbox Kinect
Mobile Robot FastSLAM with Xbox Kinect Design Team Taylor Apgar, Sean Suri, Xiangdong Xi Design Advisor Prof. Greg Kowalski Abstract Mapping is an interesting and difficult problem in robotics. In order
Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches
Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches PhD Thesis by Payam Birjandi Director: Prof. Mihai Datcu Problematic
Solving Geometric Problems with the Rotating Calipers *
Solving Geometric Problems with the Rotating Calipers * Godfried Toussaint School of Computer Science McGill University Montreal, Quebec, Canada ABSTRACT Shamos [1] recently showed that the diameter of
Copyright 2011 Casa Software Ltd. www.casaxps.com. Centre of Mass
Centre of Mass A central theme in mathematical modelling is that of reducing complex problems to simpler, and hopefully, equivalent problems for which mathematical analysis is possible. The concept of
QUALITY TESTING OF WATER PUMP PULLEY USING IMAGE PROCESSING
QUALITY TESTING OF WATER PUMP PULLEY USING IMAGE PROCESSING MRS. A H. TIRMARE 1, MS.R.N.KULKARNI 2, MR. A R. BHOSALE 3 MR. C.S. MORE 4 MR.A.G.NIMBALKAR 5 1, 2 Assistant professor Bharati Vidyapeeth s college
REAL TIME TRAFFIC LIGHT CONTROL USING IMAGE PROCESSING
REAL TIME TRAFFIC LIGHT CONTROL USING IMAGE PROCESSING Ms.PALLAVI CHOUDEKAR Ajay Kumar Garg Engineering College, Department of electrical and electronics Ms.SAYANTI BANERJEE Ajay Kumar Garg Engineering
A Study on SURF Algorithm and Real-Time Tracking Objects Using Optical Flow
, pp.233-237 http://dx.doi.org/10.14257/astl.2014.51.53 A Study on SURF Algorithm and Real-Time Tracking Objects Using Optical Flow Giwoo Kim 1, Hye-Youn Lim 1 and Dae-Seong Kang 1, 1 Department of electronices
3D Drawing. Single Point Perspective with Diminishing Spaces
3D Drawing Single Point Perspective with Diminishing Spaces The following document helps describe the basic process for generating a 3D representation of a simple 2D plan. For this exercise we will be
Vision based Vehicle Tracking using a high angle camera
Vision based Vehicle Tracking using a high angle camera Raúl Ignacio Ramos García Dule Shu [email protected] [email protected] Abstract A vehicle tracking and grouping algorithm is presented in this work
Section 1.1. Introduction to R n
The Calculus of Functions of Several Variables Section. Introduction to R n Calculus is the study of functional relationships and how related quantities change with each other. In your first exposure to
Using MATLAB to Measure the Diameter of an Object within an Image
Using MATLAB to Measure the Diameter of an Object within an Image Keywords: MATLAB, Diameter, Image, Measure, Image Processing Toolbox Author: Matthew Wesolowski Date: November 14 th 2014 Executive Summary
Assessment. Presenter: Yupu Zhang, Guoliang Jin, Tuo Wang Computer Vision 2008 Fall
Automatic Photo Quality Assessment Presenter: Yupu Zhang, Guoliang Jin, Tuo Wang Computer Vision 2008 Fall Estimating i the photorealism of images: Distinguishing i i paintings from photographs h Florin
Big Ideas in Mathematics
Big Ideas in Mathematics which are important to all mathematics learning. (Adapted from the NCTM Curriculum Focal Points, 2006) The Mathematics Big Ideas are organized using the PA Mathematics Standards
C# Implementation of SLAM Using the Microsoft Kinect
C# Implementation of SLAM Using the Microsoft Kinect Richard Marron Advisor: Dr. Jason Janet 4/18/2012 Abstract A SLAM algorithm was developed in C# using the Microsoft Kinect and irobot Create. Important
USING THE XBOX KINECT TO DETECT FEATURES OF THE FLOOR SURFACE
USING THE XBOX KINECT TO DETECT FEATURES OF THE FLOOR SURFACE By STEPHANIE COCKRELL Submitted in partial fulfillment of the requirements For the degree of Master of Science Thesis Advisor: Gregory Lee
Geometric Camera Parameters
Geometric Camera Parameters What assumptions have we made so far? -All equations we have derived for far are written in the camera reference frames. -These equations are valid only when: () all distances
Chapter 19. General Matrices. An n m matrix is an array. a 11 a 12 a 1m a 21 a 22 a 2m A = a n1 a n2 a nm. The matrix A has n row vectors
Chapter 9. General Matrices An n m matrix is an array a a a m a a a m... = [a ij]. a n a n a nm The matrix A has n row vectors and m column vectors row i (A) = [a i, a i,..., a im ] R m a j a j a nj col
Robust and accurate global vision system for real time tracking of multiple mobile robots
Robust and accurate global vision system for real time tracking of multiple mobile robots Mišel Brezak Ivan Petrović Edouard Ivanjko Department of Control and Computer Engineering, Faculty of Electrical
Project 4: Camera as a Sensor, Life-Cycle Analysis, and Employee Training Program
Project 4: Camera as a Sensor, Life-Cycle Analysis, and Employee Training Program Team 7: Nathaniel Hunt, Chase Burbage, Siddhant Malani, Aubrey Faircloth, and Kurt Nolte March 15, 2013 Executive Summary:
Removing Moving Objects from Point Cloud Scenes
1 Removing Moving Objects from Point Cloud Scenes Krystof Litomisky [email protected] Abstract. Three-dimensional simultaneous localization and mapping is a topic of significant interest in the research
Basler. Line Scan Cameras
Basler Line Scan Cameras High-quality line scan technology meets a cost-effective GigE interface Real color support in a compact housing size Shading correction compensates for difficult lighting conditions
Shape Measurement of a Sewer Pipe. Using a Mobile Robot with Computer Vision
International Journal of Advanced Robotic Systems ARTICLE Shape Measurement of a Sewer Pipe Using a Mobile Robot with Computer Vision Regular Paper Kikuhito Kawasue 1,* and Takayuki Komatsu 1 1 Department
Advanced Methods for Pedestrian and Bicyclist Sensing
Advanced Methods for Pedestrian and Bicyclist Sensing Yinhai Wang PacTrans STAR Lab University of Washington Email: [email protected] Tel: 1-206-616-2696 For Exchange with University of Nevada Reno Sept. 25,
How does the Kinect work? John MacCormick
How does the Kinect work? John MacCormick Xbox demo Laptop demo The Kinect uses structured light and machine learning Inferring body position is a two-stage process: first compute a depth map (using structured
ECE 533 Project Report Ashish Dhawan Aditi R. Ganesan
Handwritten Signature Verification ECE 533 Project Report by Ashish Dhawan Aditi R. Ganesan Contents 1. Abstract 3. 2. Introduction 4. 3. Approach 6. 4. Pre-processing 8. 5. Feature Extraction 9. 6. Verification
Automated Recording of Lectures using the Microsoft Kinect
Automated Recording of Lectures using the Microsoft Kinect Daniel Sailer 1, Karin Weiß 2, Manuel Braun 3, Wilhelm Büchner Hochschule Ostendstraße 3 64319 Pfungstadt, Germany 1 [email protected] 2 [email protected]
Synthetic Sensing: Proximity / Distance Sensors
Synthetic Sensing: Proximity / Distance Sensors MediaRobotics Lab, February 2010 Proximity detection is dependent on the object of interest. One size does not fit all For non-contact distance measurement,
Visual Servoing Methodology for Selective Tree Pruning by Human-Robot Collaborative System
Ref: C0287 Visual Servoing Methodology for Selective Tree Pruning by Human-Robot Collaborative System Avital Bechar, Victor Bloch, Roee Finkelshtain, Sivan Levi, Aharon Hoffman, Haim Egozi and Ze ev Schmilovitch,
Charlesworth School Year Group Maths Targets
Charlesworth School Year Group Maths Targets Year One Maths Target Sheet Key Statement KS1 Maths Targets (Expected) These skills must be secure to move beyond expected. I can compare, describe and solve
Fixplot Instruction Manual. (data plotting program)
Fixplot Instruction Manual (data plotting program) MANUAL VERSION2 2004 1 1. Introduction The Fixplot program is a component program of Eyenal that allows the user to plot eye position data collected with
Reflection and Refraction
Equipment Reflection and Refraction Acrylic block set, plane-concave-convex universal mirror, cork board, cork board stand, pins, flashlight, protractor, ruler, mirror worksheet, rectangular block worksheet,
SYNTHESIZING FREE-VIEWPOINT IMAGES FROM MULTIPLE VIEW VIDEOS IN SOCCER STADIUM
SYNTHESIZING FREE-VIEWPOINT IMAGES FROM MULTIPLE VIEW VIDEOS IN SOCCER STADIUM Kunihiko Hayashi, Hideo Saito Department of Information and Computer Science, Keio University {hayashi,saito}@ozawa.ics.keio.ac.jp
Introduction Epipolar Geometry Calibration Methods Further Readings. Stereo Camera Calibration
Stereo Camera Calibration Stereo Camera Calibration Stereo Camera Calibration Stereo Camera Calibration 12.10.2004 Overview Introduction Summary / Motivation Depth Perception Ambiguity of Correspondence
A Reliability Point and Kalman Filter-based Vehicle Tracking Technique
A Reliability Point and Kalman Filter-based Vehicle Tracing Technique Soo Siang Teoh and Thomas Bräunl Abstract This paper introduces a technique for tracing the movement of vehicles in consecutive video
Spatio-Temporally Coherent 3D Animation Reconstruction from Multi-view RGB-D Images using Landmark Sampling
, March 13-15, 2013, Hong Kong Spatio-Temporally Coherent 3D Animation Reconstruction from Multi-view RGB-D Images using Landmark Sampling Naveed Ahmed Abstract We present a system for spatio-temporally
In: Proceedings of RECPAD 2002-12th Portuguese Conference on Pattern Recognition June 27th- 28th, 2002 Aveiro, Portugal
Paper Title: Generic Framework for Video Analysis Authors: Luís Filipe Tavares INESC Porto [email protected] Luís Teixeira INESC Porto, Universidade Católica Portuguesa [email protected] Luís Corte-Real
Automatic Detection of PCB Defects
IJIRST International Journal for Innovative Research in Science & Technology Volume 1 Issue 6 November 2014 ISSN (online): 2349-6010 Automatic Detection of PCB Defects Ashish Singh PG Student Vimal H.
A Genetic Algorithm-Evolved 3D Point Cloud Descriptor
A Genetic Algorithm-Evolved 3D Point Cloud Descriptor Dominik Wȩgrzyn and Luís A. Alexandre IT - Instituto de Telecomunicações Dept. of Computer Science, Univ. Beira Interior, 6200-001 Covilhã, Portugal
The process components and related data characteristics addressed in this document are:
TM Tech Notes Certainty 3D November 1, 2012 To: General Release From: Ted Knaak Certainty 3D, LLC Re: Structural Wall Monitoring (#1017) rev: A Introduction TopoDOT offers several tools designed specifically
Geometry Chapter 1. 1.1 Point (pt) 1.1 Coplanar (1.1) 1.1 Space (1.1) 1.2 Line Segment (seg) 1.2 Measure of a Segment
Geometry Chapter 1 Section Term 1.1 Point (pt) Definition A location. It is drawn as a dot, and named with a capital letter. It has no shape or size. undefined term 1.1 Line A line is made up of points
Shortest Inspection-Path. Queries in Simple Polygons
Shortest Inspection-Path Queries in Simple Polygons Christian Knauer, Günter Rote B 05-05 April 2005 Shortest Inspection-Path Queries in Simple Polygons Christian Knauer, Günter Rote Institut für Informatik,
A Simple Feature Extraction Technique of a Pattern By Hopfield Network
A Simple Feature Extraction Technique of a Pattern By Hopfield Network A.Nag!, S. Biswas *, D. Sarkar *, P.P. Sarkar *, B. Gupta **! Academy of Technology, Hoogly - 722 *USIC, University of Kalyani, Kalyani
Circle Object Recognition Based on Monocular Vision for Home Security Robot
Journal of Applied Science and Engineering, Vol. 16, No. 3, pp. 261 268 (2013) DOI: 10.6180/jase.2013.16.3.05 Circle Object Recognition Based on Monocular Vision for Home Security Robot Shih-An Li, Ching-Chang
High-accuracy ultrasound target localization for hand-eye calibration between optical tracking systems and three-dimensional ultrasound
High-accuracy ultrasound target localization for hand-eye calibration between optical tracking systems and three-dimensional ultrasound Ralf Bruder 1, Florian Griese 2, Floris Ernst 1, Achim Schweikard
Least-Squares Intersection of Lines
Least-Squares Intersection of Lines Johannes Traa - UIUC 2013 This write-up derives the least-squares solution for the intersection of lines. In the general case, a set of lines will not intersect at a
Largest Fixed-Aspect, Axis-Aligned Rectangle
Largest Fixed-Aspect, Axis-Aligned Rectangle David Eberly Geometric Tools, LLC http://www.geometrictools.com/ Copyright c 1998-2016. All Rights Reserved. Created: February 21, 2004 Last Modified: February
The Scientific Data Mining Process
Chapter 4 The Scientific Data Mining Process When I use a word, Humpty Dumpty said, in rather a scornful tone, it means just what I choose it to mean neither more nor less. Lewis Carroll [87, p. 214] In
TerraColor White Paper
TerraColor White Paper TerraColor is a simulated true color digital earth imagery product developed by Earthstar Geographics LLC. This product was built from imagery captured by the US Landsat 7 (ETM+)
ACCURACY ASSESSMENT OF BUILDING POINT CLOUDS AUTOMATICALLY GENERATED FROM IPHONE IMAGES
ACCURACY ASSESSMENT OF BUILDING POINT CLOUDS AUTOMATICALLY GENERATED FROM IPHONE IMAGES B. Sirmacek, R. Lindenbergh Delft University of Technology, Department of Geoscience and Remote Sensing, Stevinweg
Static Environment Recognition Using Omni-camera from a Moving Vehicle
Static Environment Recognition Using Omni-camera from a Moving Vehicle Teruko Yata, Chuck Thorpe Frank Dellaert The Robotics Institute Carnegie Mellon University Pittsburgh, PA 15213 USA College of Computing
Context-aware Library Management System using Augmented Reality
International Journal of Electronic and Electrical Engineering. ISSN 0974-2174 Volume 7, Number 9 (2014), pp. 923-929 International Research Publication House http://www.irphouse.com Context-aware Library
3D Drawing. Single Point Perspective with Diminishing Spaces
3D Drawing Single Point Perspective with Diminishing Spaces The following document helps describe the basic process for generating a 3D representation of a simple 2D plan. For this exercise we will be
Geometry: Unit 1 Vocabulary TERM DEFINITION GEOMETRIC FIGURE. Cannot be defined by using other figures.
Geometry: Unit 1 Vocabulary 1.1 Undefined terms Cannot be defined by using other figures. Point A specific location. It has no dimension and is represented by a dot. Line Plane A connected straight path.
Tracking devices. Important features. 6 Degrees of freedom. Mechanical devices. Types. Virtual Reality Technology and Programming
Tracking devices Virtual Reality Technology and Programming TNM053: Lecture 4: Tracking and I/O devices Referred to head-tracking many times Needed to get good stereo effect with parallax Essential for
. 0 1 10 2 100 11 1000 3 20 1 2 3 4 5 6 7 8 9
Introduction The purpose of this note is to find and study a method for determining and counting all the positive integer divisors of a positive integer Let N be a given positive integer We say d is a
Canny Edge Detection
Canny Edge Detection 09gr820 March 23, 2009 1 Introduction The purpose of edge detection in general is to significantly reduce the amount of data in an image, while preserving the structural properties
NCC-RANSAC: A Fast Plane Extraction Method for Navigating a Smart Cane for the Visually Impaired
NCC-RANSAC: A Fast Plane Extraction Method for Navigating a Smart Cane for the Visually Impaired X. Qian and C. Ye, Senior Member, IEEE Abstract This paper presents a new RANSAC based method for extracting
Laboratory Report Scoring and Cover Sheet
Laboratory Report Scoring and Cover Sheet Title of Lab _Newton s Laws Course and Lab Section Number: PHY 1103-100 Date _23 Sept 2014 Principle Investigator _Thomas Edison Co-Investigator _Nikola Tesla
Neural Network based Vehicle Classification for Intelligent Traffic Control
Neural Network based Vehicle Classification for Intelligent Traffic Control Saeid Fazli 1, Shahram Mohammadi 2, Morteza Rahmani 3 1,2,3 Electrical Engineering Department, Zanjan University, Zanjan, IRAN
An Iterative Image Registration Technique with an Application to Stereo Vision
An Iterative Image Registration Technique with an Application to Stereo Vision Bruce D. Lucas Takeo Kanade Computer Science Department Carnegie-Mellon University Pittsburgh, Pennsylvania 15213 Abstract
MD5-26 Stacking Blocks Pages 115 116
MD5-26 Stacking Blocks Pages 115 116 STANDARDS 5.MD.C.4 Goals Students will find the number of cubes in a rectangular stack and develop the formula length width height for the number of cubes in a stack.
ROBUST VEHICLE TRACKING IN VIDEO IMAGES BEING TAKEN FROM A HELICOPTER
ROBUST VEHICLE TRACKING IN VIDEO IMAGES BEING TAKEN FROM A HELICOPTER Fatemeh Karimi Nejadasl, Ben G.H. Gorte, and Serge P. Hoogendoorn Institute of Earth Observation and Space System, Delft University
2. igps Web Service Overview
Autonomous Mobile Robots Localization with Multiples igps Web Services Camilo Christo, Edwin Carvalho, Miguel Pedro Silva and Carlos Cardeira IDMEC/IST, Technical University of Lisbon Av. Rovisco Pais,
Determining optimal window size for texture feature extraction methods
IX Spanish Symposium on Pattern Recognition and Image Analysis, Castellon, Spain, May 2001, vol.2, 237-242, ISBN: 84-8021-351-5. Determining optimal window size for texture feature extraction methods Domènec
Automatic Building Facade Detection in Mobile Laser Scanner point Clouds
Automatic Building Facade Detection in Mobile Laser Scanner point Clouds NALANI HETTI ARACHCHIGE 1 & HANS-GERD MAAS 1 Abstract: Mobile Laser Scanner (MLS) has been increasingly used in the modeling of
Mouse Control using a Web Camera based on Colour Detection
Mouse Control using a Web Camera based on Colour Detection Abhik Banerjee 1, Abhirup Ghosh 2, Koustuvmoni Bharadwaj 3, Hemanta Saikia 4 1, 2, 3, 4 Department of Electronics & Communication Engineering,
A. OPENING POINT CLOUDS. (Notepad++ Text editor) (Cloud Compare Point cloud and mesh editor) (MeshLab Point cloud and mesh editor)
MeshLAB tutorial 1 A. OPENING POINT CLOUDS (Notepad++ Text editor) (Cloud Compare Point cloud and mesh editor) (MeshLab Point cloud and mesh editor) 2 OPENING POINT CLOUDS IN NOTEPAD ++ Let us understand
The Visual Internet of Things System Based on Depth Camera
The Visual Internet of Things System Based on Depth Camera Xucong Zhang 1, Xiaoyun Wang and Yingmin Jia Abstract The Visual Internet of Things is an important part of information technology. It is proposed
Rotation: Moment of Inertia and Torque
Rotation: Moment of Inertia and Torque Every time we push a door open or tighten a bolt using a wrench, we apply a force that results in a rotational motion about a fixed axis. Through experience we learn
Basic 3D reconstruction in Imaris 7.6.1
Basic 3D reconstruction in Imaris 7.6.1 Task The aim of this tutorial is to understand basic Imaris functionality by performing surface reconstruction of glia cells in culture, in order to visualize enclosed
Comparison of Non-linear Dimensionality Reduction Techniques for Classification with Gene Expression Microarray Data
CMPE 59H Comparison of Non-linear Dimensionality Reduction Techniques for Classification with Gene Expression Microarray Data Term Project Report Fatma Güney, Kübra Kalkan 1/15/2013 Keywords: Non-linear
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS Systems of Equations and Matrices Representation of a linear system The general system of m equations in n unknowns can be written a x + a 2 x 2 + + a n x n b a
Template-based Eye and Mouth Detection for 3D Video Conferencing
Template-based Eye and Mouth Detection for 3D Video Conferencing Jürgen Rurainsky and Peter Eisert Fraunhofer Institute for Telecommunications - Heinrich-Hertz-Institute, Image Processing Department, Einsteinufer
CS 534: Computer Vision 3D Model-based recognition
CS 534: Computer Vision 3D Model-based recognition Ahmed Elgammal Dept of Computer Science CS 534 3D Model-based Vision - 1 High Level Vision Object Recognition: What it means? Two main recognition tasks:!
Introduction to the TI-Nspire CX
Introduction to the TI-Nspire CX Activity Overview: In this activity, you will become familiar with the layout of the TI-Nspire CX. Step 1: Locate the Touchpad. The Touchpad is used to navigate the cursor
An Energy-Based Vehicle Tracking System using Principal Component Analysis and Unsupervised ART Network
Proceedings of the 8th WSEAS Int. Conf. on ARTIFICIAL INTELLIGENCE, KNOWLEDGE ENGINEERING & DATA BASES (AIKED '9) ISSN: 179-519 435 ISBN: 978-96-474-51-2 An Energy-Based Vehicle Tracking System using Principal
Cloud tracking with optical flow for short-term solar forecasting
Cloud tracking with optical flow for short-term solar forecasting Philip Wood-Bradley, José Zapata, John Pye Solar Thermal Group, Australian National University, Canberra, Australia Corresponding author:
THE MS KINECT USE FOR 3D MODELLING AND GAIT ANALYSIS IN THE MATLAB ENVIRONMENT
THE MS KINECT USE FOR 3D MODELLING AND GAIT ANALYSIS IN THE MATLAB ENVIRONMENT A. Procházka 1,O.Vyšata 1,2,M.Vališ 1,2, M. Yadollahi 1 1 Institute of Chemical Technology, Department of Computing and Control
Assessment of Microsoft Kinect Technology (Kinect for Xbox and Kinect for Windows) for Patient Monitoring during External Beam Radiotherapy
Assessment of Microsoft Kinect Technology (Kinect for Xbox and Kinect for Windows) for Patient Monitoring during External Beam Radiotherapy F. Tahavori, M. Alnowami, J. Jones, P. Elangovan, E. Donovan
Making Multiple Code Reading Easy. Webinar
Making Multiple Code Reading Easy Webinar Today s Agenda Introduction How DataMan Makes Multiple Code Reading Easy Multiple Code Reading Applications Product Demonstration Videos Q&A 2 Introduction Introduction
