1 Edge detection (Trucco, Chapt 4 AND Jain et al., Chapt 5) Definition of edges -Edges are significant local changes of intensity in an image. -Edges typically occur on the boundary between two different regions in an image. Goal of edge detection -Produce a line drawing of a scene from an image of that scene. - Important features can be extracted from the edges of an image (e.g., corners, lines, curves). -These features are used by higher-level computer vision algorithms (e.g., recognition).
2 -2- What causes intensity changes? -Various physical events cause intensity changes. -Geometric events *object boundary (discontinuity in depth and/or surface color and texture) *surface boundary (discontinuity in surface orientation and/or surface color and texture) -Non-geometric events *specularity (direct reflection of light, such as a mirror) *shadows (from other objects or from the same object) *inter-reflections Edge descriptors Edge normal: unit vector in the direction of maximum intensity change. Edge direction: unit vector to perpendicular to the edge normal. Edge position or center: the image position at which the edge is located. Edge strength: related to the local image contrast along the normal.
3 -3- Modeling intensity changes -Edges can be modeled according to their intensity profiles. Step edge: the image intensity abruptly changes from one value to one side of the discontinuity to a different value on the opposite side. Ramp edge: a step edge where the intensity change is not instantaneous but occur over afinite distance. Ridge edge: the image intensity abruptly changes value but then returns to the starting value within some short distance (generated usually by lines).
4 -4- Roof edge: a ridge edge where the intensity change is not instantaneous but occur over afinite distance (generated usually by the intersection of surfaces). The four steps of edge detection (1) Smoothing: suppress as much noise as possible, without destroying the true edges. (2) Enhancement: apply a filter to enhance the quality of the edges in the image (sharpening). (3) Detection: determine which edge pixels should be discarded as noise and which should be retained (usually, thresholding provides the criterion used for detection). (4) Localization: determine the exact location of an edge (sub-pixel resolution might be required for some applications, that is, estimate the location of an edge to better than the spacing between pixels). Edge thinning and linking are usually required in this step.
5 -5- Edge detection using derivatives -Calculus describes changes of continuous functions using derivatives. - An image is a 2D function, so operators describing edges are expressed using partial derivatives. -Points which lie on an edge can be detected by: (1) detecting local maxima or minima of the first derivative (2) detecting the zero-crossing of the second derivative
6 -6- Differencing 1D signals (see also Trucco, Appendix A.2) - To compute the derivative of a signal, we approximate the derivative by finite differences: Computing the 1st derivative: f (x) = lim h >0 f (x + h) f (x) h f (x + 1) f (x) (h=1) mask: [ 1 1] -Examples using the edge models and the mask [ 1 0 1] (centered about x):
7 -7- Computing the 2nd derivative: f (x) = lim h >0 f (x + h) f (x) h f (x + 1) f (x) = f (x + 2) 2 f (x + 1) + f (x) (h=1) -This approximation is centered about x + 1; by replacing x + 1by x we obtain: f (x) f (x + 1) 2 f (x) + f (x 1) mask: [ 1 2 1] f(x) f (x) = f(x+1) - f(x) (approximates f () at x+1/2) f (x) = f(x-1)-2f(x)+f(x+1) (approximates f () at x) zero-crossing -Examples using the edge models:
8 -8- Edge detection using the gradient Definition of the gradient -The gradient is a vector which has certain magnitude and direction: f = f x f y magn( f ) = ( f x )2 + ( f M x 2 + M y 2 y )2 = dir( f ) = tan 1 (M y /M x ) - To save computations, the magnitude of gradient is usually approximated by: Properties of the gradient magn( f ) M x + M y -The magnitude of gradient provides information about the strength of the edge. -The direction of gradient is always perpendicular to the direction of the edge (the edge direction is rotated with respect to the gradient direction by -90 degrees).
9 -9- Estimating the gradient with finite differences f x = lim h >0 f y = lim h >0 f (x + h, y) f (x, y) h f (x, y + h) f (x, y) h -The gradient can be approximated by finite differences: f x = f (x + h x, y) f (x, y) = f (x + 1, y) f (x, y), (h h x =1) y f y = f (x, y + h y) f (x, y) = f (x, y + 1) f (x, y), (h y =1) h y -Using pixel-coordinate notation (remember: j corresponds to the x direction and i to the negative y direction): x-h x x+h y+h y y-h f x = f (i, j + 1) f (i, j) f y = f (i 1, j) f (i, j) or f y = f (i, j) f (i + 1, j)
10 -10- Example Suppose we want to approximate the gradient magnitude at z 5 Z1 Z2 Z3 Z4 Z5 Z6 Z7 Z8 Z9 I x = z 6 z 5, I y = z 5 z 8 magn( I) = (z 6 z 5 ) 2 + (z 5 z 8 ) 2 -Wecan implement I x and I y using the following masks: (note: M x is the approximation at (i, j + 1/2) and M y is the approximation at (i + 1/2, j))
11 -11- The Roberts edge detector f x f y = f (i, j) f (i + 1, j + 1) = f (i + 1, j) f (i, j + 1) -This approximation can be implemented by the following masks: M x = M y = (note: M x and M y are is approximations at (i + 1/2, j + 1/2)) The Prewitt edge detector -Consider the arrangement of pixels about the pixel (i, j): a 0 a 7 a 6 a 1 [i, j] a 5 a 2 a 3 a 4 -The partial derivatives can be computed by: M x = (a 2 + ca 3 + a 4 ) (a 0 + ca 7 + a 6 ) M y = (a 6 + ca 5 + a 4 ) (a 0 + ca 1 + a 2 ) -The constant c implies the emphasis given topixels closer to the center of the mask. -Setting c = 1, we get the Prewitt operator: 1 M x = M y = (note: M x and M y are approximations at (i, j))
12 -12- The Sobel edge detector -Setting c = 2, we get the Sobel operator: 1 M x = M y = (note: M x and M y are approximations at (i, j)) Main steps in edge detection using masks (1) Smooth the input image ( ˆf (x, y) = f (x, y)*g(x, y)) (2) ˆf x = ˆf (x, y) *M x (x, y) (3) ˆf y = ˆf (x, y) *M y (x, y) (4) magn(x, y) = ˆf x + ˆf y (5) dir(x, y) = tan 1 ( ˆf y / ˆf x ) (6) If magn(x, y) >T,then possible edge point (an example using the Prewitt edge detector - don t divide by 2)
13 -13- (with noise filtering)
14 -14- (without noise filtering)
15 -15- Isotropic property of gradient magnitude -The magnitude of gradient is an isotropical operator (it detects edges in any direction!!)
16 -16- Some practical issues -The differential masks act as high-pass filters which tend to amplify noise. -Toreduce the effects of noise, the image needs to be smoothed first with a lowpass filter. (1) The noise suppression-localization tradeoff: a lar ger filter reduces noise, but worsens localization (i.e., it adds uncertainty to the location of the edge) and vice versa. -(2) How should we choose the threshold?
17 -17- -(3) Edge thinning and linking are required to obtain good contours. Criteria for optimal edge detection (1) Good detection: the optimal detector must minimize the probability of false positives (detecting spurious edges caused by noise), as well as that of false negatives (missing real edges). (2) Good localization: the edges detected must be as close as possible to the true edges. Single response constraint: the detector must return one point only for each true edge point; that is, minimize the number of local maxima around the true edge (created by noise).
18 -18- The Canny edge detector -This is probably the most widely used edge detector in computer vision. -Canny has shown that the first derivative of the Gaussian closely approximates the operator that optimizes the product of signal-to-noise ratio and localization. -His analysis is based on "step-edges" corrupted by "additive Gaussian noise". 1. Compute f x and f y Algorithm f x = x ( f * G) = f * x G = f * G x f y = y ( f * G) = f * y G = f * G y G(x, y) isthe Gaussian function G x (x, y) isthe derivate of G(x, y) with respect to x: G x (x, y) = x 2 G(x, y) G y (x, y) isthe derivate of G(x, y) with respect to y: G y (x, y) = y 2 G(x, y) 2. Compute the gradient magnitude 3. Apply non-maxima suppression. magn(i, j) = f 2 2 x + f y 4. Apply hysteresis thresholding/edge linking.
19 -19- Non-maxima suppression -Tofind the edge points, we need to find the local maxima of the gradient magnitude. -Broad ridges must be thinned so that only the magnitudes at the points of greatest local change remain. -All values along the direction of the gradient that are not peak values of a ridge are suppressed. magn(i1,j1) magn(i,j) magn(i2,j2) direction of gradient Algorithm For each pixel (x,y) do: if magn(i, j) <magn(i 1, j 1 )or magn(i, j) <magn(i 2, j 2 ) then I N (i, j) = 0 else I N (i, j) = magn(i, j)
20 -20- Hysteresis thresholding/edge Linking -The output of non-maxima suppression still contains the local maxima created by noise. -Can we get rid of them just by using a single threshold? *ifweset a low threshold, some noisy maxima will be accepted too. *ifweset a high threshold, true maxima might be missed (the value of true maxima will fluctuate above and below the threshold, fragmenting the edge). -Amore effective scheme is to use two thresholds: *alow threshold t l *ahigh threshold t h *usually, t h 2t l Algorithm 1. Produce two thresholded images I 1 (i, j) and I 2 (i, j). (note: since I 2 (i, j) was formed with a high threshold, it will contain fewer false edges but there might be gaps in the contours) 2. Link the edges in I 2 (i, j) into contours 2.1 Look in I 1 (i, j) when a gap is found. 2.2 By examining the 8 neighbors in I 1 (i, j), gather edge points from I 1 (i, j) until the gap has been bridged to an edge in I 2 (i, j). -The algorithm performs edge linking as a by-product of double-thresholding!!
23 -23- Edge detection using the second derivative - Edge points can be detected by finding the zero-crossings of the second derivative. -There are two operators in 2D that correspond to the second derivative: *Laplacian *Second directional derivative The Laplacian -Approximating 2 f : 2 f = 2 f x f y 2 2 f = f (i, j + 1) 2 f (i, j) + f (i, j 1) x2 2 f = f (i + 1, j) 2 f (i, j) + f (i 1, j) y2 2 f = 4 f (i, j) + f (i, j + 1) + f (i, j 1) + f (i + 1, j) + f (i 1, j)
24 -24- Example: Z1 Z2 Z3 Z4 Z5 Z6 Z7 Z8 Z9 2 f = 4z 5 + (z 2 + z 4 + z 6 + z 8 ) -The Laplacian can be implemented using the mask shown below: Example:
25 -25- Properties of the Laplacian -Itisanisotropic operator. -Itischeaper to implement (one mask only). -Itdoes not provide information about edge direction. -Itismore sensitive to noise (differentiates twice). The Laplacian-of-Gaussian (LOG) -Toreduce the noise effect, the image is first smoothed with a low-pass filter. -Inthe case of the LOG, the low-pass filter is chosen to be a Gaussian. G(x, y) = e x2 +y ( determines the degree of smoothing, mask size increases with ) -Itcan be shown that: 2 [ f (x, y) *G(x, y)] = 2 G(x, y) * f (x, y) 2 G(x, y) =( r2 4 2 )e r2 /2 2, (r 2 = x 2 + y 2 )
27 -27- Gradient vs LOG: a comparison -Gradient works well when the image contains sharp intensity transitions and low noise - Zero-crossings of LOG offer better localization, especially when the edges are not very sharp
28 -28- The second directional derivative -This is the second derivative computed in the direction of the gradient. 2 n 2 = f 2 x f xx + 2 f x f y f xy + f 2 y f yy f 2 x f 2 y Multiscale processing (scale space) -Aserious practical problem with any edge detector is the matter of choosing the scale of smoothing (e.g.,the value of using a Gaussian). -For many applications, it is desirable to be able to process an image at multiple scales. -Wedetermine which edges are most significant in terms of the range of scales over which they are observed to occur.
29 -29- (Canny edges at multiple scales of smoothing, =0.5, 1, 2, 4, 8, 16)
Basic Filters (7) Convolution/correlation/Linear filtering Gaussian filters Smoothing and noise reduction First derivatives of Gaussian Second derivative of Gaussian: Laplacian Oriented Gaussian filters
Today s Topics Lecture : LoG and DoG Filters Laplacian of Gaussian (LoG) Filter - useful for finding edges - also useful for finding blobs! approimation using Difference of Gaussian (DoG) Recall: First
Canny Edge Detection 09gr820 March 23, 2009 1 Introduction The purpose of edge detection in general is to significantly reduce the amount of data in an image, while preserving the structural properties
2D Image Processing Edge and Corner Detection Prof. Didier Stricker Dr. Alain Pagani Kaiserlautern University http://ags.cs.uni-kl.de/ DFKI Deutsches Forschungszentrum für Künstliche Intelligenz http://av.dfki.de
A New Image Edge Detection Method using Quality-based Clustering Bijay Neupane Zeyar Aung Wei Lee Woon Technical Report DNA #2012-01 April 2012 Data & Network Analytics Research Group (DNA) Computing and
Sharpening through spatial filtering Stefano Ferrari Università degli Studi di Milano email@example.com Elaborazione delle immagini (Image processing I) academic year 2011 2012 Sharpening The term
2.2. Creaseness operator 31 2.2 Creaseness operator Antonio López, a member of our group, has studied for his PhD dissertation the differential operators described in this section . He has compared
Image Segmentation Preview Segmentation subdivides an image to regions or objects Two basic properties of intensity values Discontinuity Edge detection Similarity Thresholding Region growing/splitting/merging
image processing images images: what are they? what is shown in this image? What is this? what is an image images describe the evolution of physical variables (intensity, color, reflectance, condutivity)
Machine Learning for Computer Vision 1 MVA ENS Cachan Lecture 2: Logistic regression & intro to MIL Iasonas Kokkinos Iasonas.firstname.lastname@example.org Department of Applied Mathematics Ecole Centrale Paris Galen
HOMEWORK SOLUTIONS All questions are from Vector Calculus, by Marsden and Tromba Question :..6 Let w = f(x, y) be a function of two variables, and let x = u + v, y = u v. Show that Solution. By the chain
Appendix H H.Calculating Normal Vectors This appendix describes how to calculate normal vectors for surfaces. You need to define normals to use the OpenGL lighting facility, which is described in Chapter
SOLUTIONS Problem. Find the critical points of the function f(x, y = 2x 3 3x 2 y 2x 2 3y 2 and determine their type i.e. local min/local max/saddle point. Are there any global min/max? Partial derivatives
Implementation of Canny Edge Detector of color images on CELL/B.E. Architecture. Chirag Gupta,Sumod Mohan K email@example.com, firstname.lastname@example.org Abstract In this project we propose a method to improve
One-Dimensional Processing for Edge Detection using Hilbert Transform P. Kiran Kumar, Sukhendu Das and B. Yegnanarayana Department of Computer Science and Engineering Indian Institute of Technology, Madras
Image Enhancement: Frequency domain methods The concept of filtering is easier to visualize in the frequency domain. Therefore, enhancement of image f ( m, n) can be done in the frequency domain, based
suggestive contours and abstracted shading Daniel Arias LINES AND SHADES Shape, volume, shades and texture in drawing lines as nonphotorealistic rendering technique Luis Caballero Suggestive contours Many
Math 241, Exam 1 Information. 9/24/12, LC 310, 11:15-12:05. Exam 1 will be based on: Sections 12.1-12.5, 14.1-14.3. The corresponding assigned homework problems (see http://www.math.sc.edu/ boylan/sccourses/241fa12/241.html)
A New Robust Algorithm for Video Text Extraction Pattern Recognition, vol. 36, no. 6, June 2003 Edward K. Wong and Minya Chen School of Electrical Engineering and Computer Science Kyungpook National Univ.
REAL TIME TRAFFIC LIGHT CONTROL USING IMAGE PROCESSING Ms.PALLAVI CHOUDEKAR Ajay Kumar Garg Engineering College, Department of electrical and electronics Ms.SAYANTI BANERJEE Ajay Kumar Garg Engineering
A CURVATURE SENSITIVE FILTER AND ITS APPLICATION IN MICROFOSSIL IMAGE CHARACTERISATION J. P. Oakley and R. T. Shann Department of Electrical Engineering University of Manchester Manchester M13 9PL U.K.
Classification of Fingerprints Sarat C. Dass Department of Statistics & Probability Fingerprint Classification Fingerprint classification is a coarse level partitioning of a fingerprint database into smaller
Section 26: Directional Derivatives and the Gradient Vector Recall that if f is a differentiable function of x and y and z = f(x, y), then the partial derivatives f x (x, y) and f y (x, y) give the rate
T H E U N I V E R S I T Y of T E X A S Convolution, Noise and Filters Philip Baldwin, Ph.D. Department of Biochemistry Response to an Entire Signal The response of a system with impulse response h(t) to
FACIAL EXPRESSION RECOGNITION BASED ON EDGE DETECTION Xiaoming CHEN and Wushan CHENG College of Mechanical Engineering, Shanghai University of Engineering Science, Shanghai 0160, China ABSTRACT Relational
Image Gradients Given a discrete image Á Òµ, consider the smoothed continuous image Üµ defined by Üµ Ü ¾ Ö µ Á Òµ Ü ¾ Ö µá µ (1) where Ü ¾ Ö Ô µ Ü ¾ Ý ¾. ½ ¾ ¾ Ö ¾ Ü ¾ ¾ Ö. Here Ü is the 2-norm for the
Optical Flow as a property of moving objects used for their registration Wolfgang Schulz Computer Vision Course Project York University Email:email@example.com 1. Introduction A soccer game is a real
QUALITY TESTING OF WATER PUMP PULLEY USING IMAGE PROCESSING MRS. A H. TIRMARE 1, MS.R.N.KULKARNI 2, MR. A R. BHOSALE 3 MR. C.S. MORE 4 MR.A.G.NIMBALKAR 5 1, 2 Assistant professor Bharati Vidyapeeth s college
Computational Foundations of Cognitive Science Lecture 15: Convolutions and Kernels Frank Keller School of Informatics University of Edinburgh firstname.lastname@example.org February 23, 2010 Frank Keller Computational
02/09/12 Feature Tracking and Optical Flow Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem Many slides adapted from Lana Lazebnik, Silvio Saverse, who in turn adapted slides from Steve
Arrowsmith: Automatic Archery Scorer Chanh Nguyen and Irving Lin Department of Computer Science, Stanford University ABSTRACT We present a method for automatically determining the score of a round of arrows
746A27 Remote Sensing and GIS Lecture 4 Digital image processing Chandan Roy Guest Lecturer Department of Computer and Information Science Linköping University Digital Image Processing Most of the common
Analysis of Face Recognition Based On Edge Detection Algorithm with Hardware Interfacing Pankaj Bhandari 1, Pankaj K Gupta 2, Karthik U.S Goutham Reddy 3,Jeeva.B 4 UG Students, Department of ECE, N.I.T
ISSN (Online): 1694-0784 ISSN (Print): 1694-0814 1 Signature Region of Interest using Auto cropping Bassam Al-Mahadeen 1, Mokhled S. AlTarawneh 2 and Islam H. AlTarawneh 2 1 Math. And Computer Department,
ADAPTIVE PROCEDURE FOR BOUNDARY DETECTION IN SATELLITE IMAGES Borislav Marinov, Veselina Gospodinova ABSTRACT The appropriate modification of adaptive procedure for generation of boundaries in multi-band
ColorCrack: Identifying Cracks in Glass James Max Kanter Massachusetts Institute of Technology 77 Massachusetts Ave Cambridge, MA 02139 email@example.com Figure 1: ColorCrack automatically identifies cracks
521466S Machine Vision Assignment #7 Hough transform Spring 2014 In this assignment we use the hough transform to extract lines from images. We use the standard (r, θ) parametrization of lines, lter the
Acta Polytechnica Hungarica Vol. 6, No. 1, 2009 Pathfinding Based on Edge Detection and Infrared Distance Measuring Sensor Bojan Kuljić, János Simon, Tibor Szakáll Polytechnical Engineering College Marka
Distinctive Image Features from Scale-Invariant Keypoints David G. Lowe Computer Science Department University of British Columbia Vancouver, B.C., Canada firstname.lastname@example.org January 5, 2004 Abstract This paper
. Introduction 3D Scanner using Line Laser Di Lu Electrical, Computer, and Systems Engineering Rensselaer Polytechnic Institute The goal of 3D reconstruction is to recover the 3D properties of a geometric
Lecture 2: 2D Fourier transforms and applications B14 Image Analysis Michaelmas 2014 A. Zisserman Fourier transforms and spatial frequencies in 2D Definition and meaning The Convolution Theorem Applications
J. Zhang and X. Yuan: Analysis of segmentation performance on super-resolved images. In Texture 2005: Proceedings of the th International Workshop on Texture Analysis and Synthesis, pp. 155 10, 2005. Analysis
Binary Image Analysis Segmentation produces homogenous regions each region has uniform gray-level each region is a binary image (0: background, 1: object or the reverse) more intensity values for overlapping
Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches PhD Thesis by Payam Birjandi Director: Prof. Mihai Datcu Problematic
, pp.233-237 http://dx.doi.org/10.14257/astl.2014.51.53 A Study on SURF Algorithm and Real-Time Tracking Objects Using Optical Flow Giwoo Kim 1, Hye-Youn Lim 1 and Dae-Seong Kang 1, 1 Department of electronices
An Iterative Image Registration Technique with an Application to Stereo Vision Bruce D. Lucas Takeo Kanade Computer Science Department Carnegie-Mellon University Pittsburgh, Pennsylvania 15213 Abstract
Precision edge detection with bayer pattern sensors Prof.Dr.-Ing.habil. Gerhard Linß Dr.-Ing. Peter Brückner Dr.-Ing. Martin Correns Folie 1 Agenda 1. Introduction 2. State of the art 3. Key aspects 1.
DIGITAL IMAGE PROCESSING AND ANALYSIS Human and Computer Vision Applications with CVIPtools SECOND EDITION SCOTT E UMBAUGH Uffi\ CRC Press Taylor &. Francis Group Boca Raton London New York CRC Press is
Image Synthesis Ambient Occlusion Ambient Occlusion (AO) Ambient Occlusion approximates the diffuse illumination of a surface based on its directly visible occluders Idea: Trace rays through the normal-oriented
SOBEL EDGE DETECTION METHOD FOR MATLAB Elif AYBAR Anadolu University, Porsuk Vocational School, 26410 Eskişehir ABSTRACT Sobel which is a popular edge detection method is considered in this work. There
COMPARISON OF OBJECT BASED AND PIXEL BASED CLASSIFICATION OF HIGH RESOLUTION SATELLITE IMAGES USING ARTIFICIAL NEURAL NETWORKS B.K. Mohan and S. N. Ladha Centre for Studies in Resources Engineering IIT
Lecture 14 Point Spread Function (PSF), Modulation Transfer Function (MTF), Signal-to-noise Ratio (SNR), Contrast-to-noise Ratio (CNR), and Receiver Operating Curves (ROC) Point Spread Function (PSF) Recollect
Vision based Vehicle Tracking using a high angle camera Raúl Ignacio Ramos García Dule Shu email@example.com firstname.lastname@example.org Abstract A vehicle tracking and grouping algorithm is presented in this work
Math a Lagrange Multipliers Spring, 009 The method of Lagrange multipliers allows us to maximize or minimize functions with the constraint that we only consider points on a certain surface To find critical
) Find the point of intersection of the lines x = t +, y = 3t + 4, z = 4t + 5, and x = 6s + 3, y = 5s +, z = 4s + 9, and then find the plane containing these two lines. Solution. Solve the system of equations
IMPLEMENTATION OF IMAGE PROCESSING IN REAL TIME CAR PARKING SYSTEM Ms.SAYANTI BANERJEE Ajay Kumar Garg Engineering College, Department of electrical and electronics Ms.PALLAVI CHOUDEKAR Ajay Kumar Garg
Image Segmentation and Registration Dr. Christine Tanner (email@example.com) Computer Vision Laboratory, ETH Zürich Dr. Verena Kaynig, Machine Learning Laboratory, ETH Zürich Outline Segmentation
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION Introduction In the previous chapter, we explored a class of regression models having particularly simple analytical
SECOND DERIVATIVE TEST FOR CONSTRAINED EXTREMA This handout presents the second derivative test for a local extrema of a Lagrange multiplier problem. The Section 1 presents a geometric motivation for the
Computer Graphics Course SS 2007 Antialiasing How to avoid spatial aliasing caused by an undersampling of the signal, i.e. the sampling frequency is not high enough to cover all details Supersampling -
IT 12 057 Examensarbete 30 hp November 2012 Stitching of X-ray Images Krishna Paudel Institutionen för informationsteknologi Department of Information Technology Abstract Stitching of X-ray Images Krishna
Poker Vision: Playing Cards and Chips Identification based on Image Processing Paulo Martins 1, Luís Paulo Reis 2, and Luís Teófilo 2 1 DEEC Electrical Engineering Department 2 LIACC Artificial Intelligence
Chapter 7 Div, grad, and curl 7.1 The operator and the gradient: Recall that the gradient of a differentiable scalar field ϕ on an open set D in R n is given by the formula: ( ϕ ϕ =, ϕ,..., ϕ. (7.1 x 1
Workshop Perceptual Effects of Filtering and Masking Introduction to Filtering and Masking The perception and correct identification of speech sounds as phonemes depends on the listener extracting various
A Template Matching Method for Recognition of Surveyor s Gestures in Images Artur Wilkowski Faculty of Geodesy and Cartography & Institute of Control and Computation Engineering Warsaw University of Technology
Card Matching Card Matching A Game of Matching Functions Description Give each group of students a packet of cards. Students work as a group to match the cards, by thinking about their card and what information
Microeconomic Theory: Basic Math Concepts Matt Van Essen University of Alabama Van Essen (U of A) Basic Math Concepts 1 / 66 Basic Math Concepts In this lecture we will review some basic mathematical concepts
CHAPTER 15 Moving Average Filters The moving average is the most common filter in DSP, mainly because it is the easiest digital filter to understand and use. In spite of its simplicity, the moving average
Single Depth Image Super Resolution and Denoising Using Coupled Dictionary Learning with Local Constraints and Shock Filtering Jun Xie 1, Cheng-Chuan Chou 2, Rogerio Feris 3, Ming-Ting Sun 1 1 University
Chapter 5 Filtering and Enhancing Images This chapter describes methods to enhance images for either human consumption or for further automatic operations. Perhaps we need to reduce noise in the image;
The Derivative and the Tangent Line Problem Calculus grew out of four major problems that European mathematicians were working on during the seventeenth century. 1. The tangent line problem 2. The velocity
Least-Squares Intersection of Lines Johannes Traa - UIUC 2013 This write-up derives the least-squares solution for the intersection of lines. In the general case, a set of lines will not intersect at a
The International Arab Journal of Information Technology, Vol. 1, No. 2, July 2004 221 A New Approach for Contrast Enhancement Using Sigmoid Function Naglaa Hassan 1&2 and Norio Akamatsu 1 1 Department
Practice Problems for Midterm () For each of the following, find and sketch the domain, find the range (unless otherwise indicated), and evaluate the function at the given point P : (a) f(x, y) = + 4 y,
Whiteboard It! Convert Whiteboard Content into an Electronic Document Zhengyou Zhang Li-wei He Microsoft Research Email: firstname.lastname@example.org, email@example.com Aug. 12, 2002 Abstract This ongoing project