Imageprocessing. Errors of measurements



Similar documents
Lectures 6&7: Image Enhancement

Digital Image Fundamentals. Selim Aksoy Department of Computer Engineering Bilkent University

Computational Optical Imaging - Optique Numerique. -- Deconvolution --

Bildverarbeitung und Mustererkennung Image Processing and Pattern Recognition

Introduction to image coding

3. Interpolation. Closing the Gaps of Discretization... Beyond Polynomials

Convolution. 1D Formula: 2D Formula: Example on the web:

Linear Filtering Part II

Lecture 14. Point Spread Function (PSF)

Aliasing, Image Sampling and Reconstruction

Solutions to Exam in Speech Signal Processing EN2300

Numerical Methods For Image Restoration

jorge s. marques image processing

The Image Deblurring Problem

EECS 556 Image Processing W 09. Interpolation. Interpolation techniques B splines

A BRIEF STUDY OF VARIOUS NOISE MODEL AND FILTERING TECHNIQUES

Digital Imaging and Multimedia. Filters. Ahmed Elgammal Dept. of Computer Science Rutgers University

Blind Deconvolution of Barcodes via Dictionary Analysis and Wiener Filter of Barcode Subsections

Point Lattices in Computer Graphics and Visualization how signal processing may help computer graphics

High Quality Image Magnification using Cross-Scale Self-Similarity

Computational Foundations of Cognitive Science

A System for Capturing High Resolution Images

Data Storage. Chapter 3. Objectives. 3-1 Data Types. Data Inside the Computer. After studying this chapter, students should be able to:

Comparison of different image compression formats. ECE 533 Project Report Paula Aguilera

Signal to Noise Instrumental Excel Assignment

Algebra I Vocabulary Cards

Image Compression through DCT and Huffman Coding Technique

E190Q Lecture 5 Autonomous Robot Navigation

IMPROVEMENT OF DIGITAL IMAGE RESOLUTION BY OVERSAMPLING

Common Core Unit Summary Grades 6 to 8

NEW YORK STATE TEACHER CERTIFICATION EXAMINATIONS

Data Storage 3.1. Foundations of Computer Science Cengage Learning

Resolution Enhancement of Photogrammetric Digital Images

Introduction to Medical Image Compression Using Wavelet Transform

Reading.. IMAGE COMPRESSION- I IMAGE COMPRESSION. Image compression. Data Redundancy. Lossy vs Lossless Compression. Chapter 8.

CM0340 SOLNS. Do not turn this page over until instructed to do so by the Senior Invigilator.

CHAPTER 3: DIGITAL IMAGING IN DIAGNOSTIC RADIOLOGY. 3.1 Basic Concepts of Digital Imaging

An Iterative Image Registration Technique with an Application to Stereo Vision

Using Image J to Measure the Brightness of Stars (Written by Do H. Kim)

ANALYSIS OF THE EFFECTIVENESS IN IMAGE COMPRESSION FOR CLOUD STORAGE FOR VARIOUS IMAGE FORMATS

Face detection is a process of localizing and extracting the face region from the

MMGD0203 Multimedia Design MMGD0203 MULTIMEDIA DESIGN. Chapter 3 Graphics and Animations

Linear Threshold Units

Epipolar Geometry. Readings: See Sections 10.1 and 15.6 of Forsyth and Ponce. Right Image. Left Image. e(p ) Epipolar Lines. e(q ) q R.

Admin stuff. 4 Image Pyramids. Spatial Domain. Projects. Fourier domain 2/26/2008. Fourier as a change of basis

Forensic Image Processing.

Applications to Data Smoothing and Image Processing I

Low-resolution Character Recognition by Video-based Super-resolution

3. INNER PRODUCT SPACES

DIGITAL IMAGE PROCESSING AND ANALYSIS


Robotics. Lecture 3: Sensors. See course website for up to date information.

High Performance GPU-based Preprocessing for Time-of-Flight Imaging in Medical Applications

Advanced Signal Processing and Digital Noise Reduction

Latest Results on High-Resolution Reconstruction from Video Sequences

Introduction to CCDs and CCD Data Calibration

Assessment. Presenter: Yupu Zhang, Guoliang Jin, Tuo Wang Computer Vision 2008 Fall

PERFORMANCE ANALYSIS OF HIGH RESOLUTION IMAGES USING INTERPOLATION TECHNIQUES IN MULTIMEDIA COMMUNICATION SYSTEM

Measuring Line Edge Roughness: Fluctuations in Uncertainty

Computer Networks and Internets, 5e Chapter 6 Information Sources and Signals. Introduction

Correlation and Convolution Class Notes for CMSC 426, Fall 2005 David Jacobs


Bayesian Image Super-Resolution

December 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B. KITCHENS

Texture. Chapter Introduction

An Evaluation of Irregularities of Milled Surfaces by the Wavelet Analysis

Algebra 1 Course Information

WAKING up without the sound of an alarm clock is a

A Study on SURF Algorithm and Real-Time Tracking Objects Using Optical Flow

Summary: Transformations. Lecture 14 Parameter Estimation Readings T&V Sec Parameter Estimation: Fitting Geometric Models

Robert Collins CSE598G. More on Mean-shift. R.Collins, CSE, PSU CSE598G Spring 2006

A PHOTOGRAMMETRIC APPRAOCH FOR AUTOMATIC TRAFFIC ASSESSMENT USING CONVENTIONAL CCTV CAMERA

The Role of Size Normalization on the Recognition Rate of Handwritten Numerals

Resolution Enhancement of images with Interpolation and DWT-SWT Wavelet Domain Components

1 Determinants and the Solvability of Linear Systems

dr hab. Paweł Strumiłło

B2.53-R3: COMPUTER GRAPHICS. NOTE: 1. There are TWO PARTS in this Module/Paper. PART ONE contains FOUR questions and PART TWO contains FIVE questions.

Video stabilization for high resolution images reconstruction

This unit will lay the groundwork for later units where the students will extend this knowledge to quadratic and exponential functions.

WHITE PAPER. Are More Pixels Better? Resolution Does it Really Matter?

Star Detection and Removal in Night Airglow Images

Analecta Vol. 8, No. 2 ISSN

Algebra Unpacked Content For the new Common Core standards that will be effective in all North Carolina schools in the school year.

Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches

Information, Entropy, and Coding

Adaptive Coded Aperture Photography

Sharpening through spatial filtering

Image Gradients. Given a discrete image Á Òµ, consider the smoothed continuous image ܵ defined by

Vector and Matrix Norms

Implementation of Digital Signal Processing: Some Background on GFSK Modulation


DYNAMIC RANGE IMPROVEMENT THROUGH MULTIPLE EXPOSURES. Mark A. Robertson, Sean Borman, and Robert L. Stevenson

Enhancing the SNR of the Fiber Optic Rotation Sensor using the LMS Algorithm

Automatic Restoration Algorithms for 35mm film

Detecting and measuring faint point sources with a CCD

Sachin Patel HOD I.T Department PCST, Indore, India. Parth Bhatt I.T Department, PCST, Indore, India. Ankit Shah CSE Department, KITE, Jaipur, India

Transcription:

Imageprocessing Errors of measurements Many source of errors affect observations, like CCD images: - Diffraction and seeing blur the image. - The imaging optics causes geometric distortions. - The imaging hardware itself produces some noise. The amount of noise is given by the signal to noise ratio. The ratio can be improved (increased) by repeating the obsevation several times and taking the average of the observations. - Discretisation of the signal, depending on the number of bits of the A/D converter, causes some error. This is a random error that does not affect the average of several measurements. Statistical methods can be applied to control the effect.

- Sensitivity of different cameras at different wavelengths is different. This is a systematic error that can be detected only after comparing the data with observations made eith other instruments. Observations made with the same camera can be systematically wrong but still compatible with each others. When comparing observations made with different instruments, each instrument must be calibrated separately. - The image may have reflections of bright objects (like the Moon). The problem is that different observations are incompatible, due to changing conditions. If a variable star is observed at different altitudes, the atmospheric perturbations are different. Observing comparison stars this error can be removed at least partly. - The object may be so bright that the image is saturated and no measurement value is obtained. It is only known that the brightness exceeds some limit.

- Other perturbations (cosmic rays, disturbances due to poor electric isolation etc.). - The data may also contain single deviating values (outliers). They can be erroneous observations or due to some external source. The reason for such deviations should be found out in order not to throw out exceptional but real observations that might indicate some interesting phenomenon. - Many observing methods involve their own specific reduction techniques to convert the raw data into actual values that can be analysed. An example in image processing is the removal of geometric distortions caused by the imaging system.

The purpose of image processing is to deal with these problems. It includes e.g. - data acquisition and storage - digitisation of an analogous signal - data compression - data representation - image restoration - geometric transforms - noise reduction - deconvolution - detail enhancing - image analysis - edge detection - object recognition - classification

Image representations An image is usually stored as a matrix, each element of which correponds to one pixel. Another possibility would be to store the coeddicients of a Fourier series. Both are special cases of a general representation. Let f be quadratically integrable function defined in some region S. We want to express the function in terms of some basis functions φ mn, m,n = 0,,... The basis functions are orthonormal, if φ mn (x,y)φ rs(x,y) dx dy = δ mr δ ns. S

We are looking for an expansion M m=0 N n=0 that will minimise the sampling error e 2 : e 2 = S M f(x,y) m=0 a mn φ mn (x,y) N n=0 a mn φ mn (x,y) 2 dx dy, This is satisfied when a mn = S f(x,y)φ mn(x,y) dx dy. The set of basis functions is complete, if for each quadratically integrable function the error e 0, when the number of terms is increased.

Standard representations: Basis functions are rectangles: { MN φ mn (x,y) = AB, ma jos x [ 0 muuten. The coefficient a mn is simply the mean value of f in the rectangle where φ mn 0. M, (m+)a M Fourier representasion: Basis functions are φ mn (x,y) = AB e 2πi(mx/A+ny/B), ) ja y [ nb N, (n+)b N ), where A and B are the width and height of the image, respectively. Optimal representation: Is there a basis that will make the error e 2 as small as possible. Minimising the error will lead to the eigenvalue problem R ff (x,x,y,y )φ m,n (x,y ) dx dy = γ mn φ mn (x,y). The optimal basis functions, known as Karhunen and Loève functions, are obtained as the solution of this equation.

Quantization Intensities are continuous real values, which must be mapped to a finite number of gray levels. Let the limiting intensities be z k, k =,2,...,K +. If the intensity is in the range [z k,z k+ ), the pixel will be given the value q k. If p(z) is the probability that the intensity is z, the quantization error is δ 2 q = K k= zk+ z k (z q k ) 2 p(z) dz.

Minimising this we get δ 2 q = (z k q k ) 2 p(z k ) (z k q k ) 2 p(z k ) = 0,k = 2,3,...,K, z k δq 2 zk+ = 2 (z q k )p(z) dz = 0,k =,2,...,K. q k z k or z k = q k + q k, 2 zk+ z q k = k zp(z) dz zk+ p(z) dz. z k If p(z) is constant, the limits will be equally spaced. In practice quantization is carried out directly by the hardware and the limits are equally spaced even if the result may not be optimal.

Compression Compression is needed to reduce storage space and also to speed up data transfer (important e.g. in communication with a space probe). Often neighbouring pixels are correlated, and space is wasted by storing data with low information content. There are many methods for compression: - Transformation methods express the image as a combination of suitable basis functions. - Coding of gray levels can be changed to adapt to the nonrandomness of the distribution (shorter codes are used for the most common values). - Use some fractal process to represent the image.

Noiseless methods keep all information of the original image; the image can be reconstructed exactly. Noisy methods keep only the essential features; usually the smallest details are lost. If the probability of gray level q i is p i, the information content of the image is H = K p i log 2 p i. i= This gives a lower limit for the number of bits per pixel that are needed to reconstruct the original image exactly. (Some methods may claim to give a very high compression ratio, but the amount of compression is always limited by the information content.) White noise (pixels not correlated at all) has the highest information content and cannot be compressed.

Transformation compression: Express the image in terms of basis functions and keep only the most important terms. Choosing the best base requires some information about the images. Karhunen and Loève transform (KL transform): If the image is expressed as a combination of the KL functions, the coefficients are independet. Finding the basis functions is a laborious eigenvalue problem. If the autocorrelation function is assumed to be known, expressions of the basis functions can be derived. Fourier transform: Easy to calculate, but usually to obtain the same image quality much more terms are needed than using the KL transform.

Hadamard transform: A very efficient transform; basis matrices contain only numbers + and (and a normalization constant). The basis matrices for an N N image, where N is a power of two, are φ (u,v) (m,n) = N ( )b(u,v,m,n), where b(u,v,m,n) = log 2 N k=0 (b k (u)b k (v) + b k (m)b k (n)), and b k (x) is the k th bit of the number x. Run length coding: Pixels are replaced by pairs (v,n), where v is the value of the pixel and n the number of successive pixels having the same value. If the image contains large areas with the same colour, a lot of space is saved. Huffman coding: Common values are represented by short codes and rare ones by longer codes.

Fractal coding: Natural objects rarely have simple geometric shapes, but have often a selfsimilar structure (a magnified detail looks similar to the whole). Fractal comprssion is based on iterated function systems (IFS, iterated function systems, ks. esim. Barnsley: Fractals everywhere, Academic Press 988.). IFS is a random walk at each point of which one of a given set of transforms is selected with a certain probability. Computationally heavy but can greatly compress complex images.

Wavelets Wavelets are sort of genaralisation of the Fourier transform, better adapted for nonperiodic and finite signals. They can also be used for representing data in compressed form. Wavelets are functions of one variable and two indices g x = ( ) x b g, a 0, a a where g satisfies C = 2π and F(g) is the Fourier transform of g. ξ F(g) dξ <,

Image formats eps (Encapsulated PostScript) Text, actually program code. Well suited for vector graphics, not as good for pixel images. Resolution independent and freely scalable. tif (Tagged Image format) Pixel image, not compressed, files are big. gif (CompuServe graphics interchange format) Compressed noisefree image. jpeg (Joint Photographic Experts Group file interchange format) Compressed noisy image. Accuracy suffers a little, but savings in space are considerable. mpeg (Motion Picture Experts Group file interchange format) Compression method for motion pictures. The colour of a given pixel does not usually change between successive frames. A lot of space is saved by storing only the changes. pdf (Portable Document Format) Standard mainly for publications; much more compact than eps.

The convert command in Linux can convert lots of image formats to other ones. 6646 Apr 5 3:40 forest.bmp ==> 524554 Apr 5 3:42 forest.tif 28378 Apr 5 3:4 forest.eps 755 Apr 5 3:40 forest.gif 2488 Apr 5 3:43 forest.jpg 826 Dec 7 997 color.ps ==> 2908547 Apr 5 3:59 color.tif 45466 Apr 5 4:00 color.bmp 55223 Apr 5 4:0 color.gif 3758 Apr 5 4:7 color.eps 47422 Apr 5 3:59 color.jpg

Geometric transforms A geometric transform maps a point (x,y) to another point (x,y ). In vector graphics points are usually expressed as column vectors with one additional element. Thus a point (x,y) on a plane is given as x = x y. Then transforms can be expressed as 3 3 matrices x y = M x y.

) translation M = 0 t x 0 t y 0 0. 2) scaling M = s x 0 0 0 s y 0 0 0. 3) rotation M = cos φ sinφ 0 sinφ cos φ 0. 0 0 These can be used for line drawings (vector graphics), NOT for pixel images.

Geometric transforms of pixel images ) Direct transform: x y M x y = Not good, because the mapping is not one-to-one: - some pixels may not be assigned any value - several pixels may get mapped to the same pixel 2) Inverse transform: for each (x,y ) find a pixel (x,y) that will be mapped to (x,y ) and assign its value to the pixel (x,y ). This is better, because will not leave any empty pixels. Still several pixels of the original image may be mapped to the same pixel, and some to none. x y.

3) Inverse interpolation: find (x,y) as before. Usually x and y are not integers; thus the value corresponding to the point (x,y) is interpolated using the values of the nearest neighboring pixels. This works quite well, particularly for moderatley continuous images. Sharp edges may become slightly blurred. 4) Reconstruction: express the original image in terms of some functions, and use the functions to calculate the values of the new pixels. Slow, but often gives the best results.

Filters A filter is usually a local transform f(x,y) f (x,y) = S f(x x,y y )g(x,y ) dx dy, where S is a small neighbourhood of the origin. Many filters have the form of a convolution; it is often possible to work either with the original image or its Fourier transform. Nonrecursive filter: all new values are evaluated using the original image only. Recursive filter: each pixel is replaced by a new value as soon as it has been computed. The transformed image can be stored in the space of the original one. Low-pass filter: remove all frequencies higher than a given limit. Most easily done using the Fourier transform. The filter has a smoothing effect. The Fourier transform of the rect function (a rectangular pulse) is a sinc function, whose nonzero values are not limited to a finite region. Thus smoothing the original image by convolving with the rect function does not completely remove high frequencies.

Low-pass filters: 9 5 6 2 High-pass filter: remove low frequencies, which has a sharpening effect. Corresponds to differentiation. E.g. the Laplace operator is a high-pass filter 2 f(i,j) = 2 xf(i,j) + 2 yf(i,j) = (f(i +,j) + f(i,j) + f(i,j + ) + f(i,j )) 4f(i,j). or 4

Assume the image has been blurred by adding a fraction ǫ of the sum of neighbouring pixels to each pixel: p(i,j) = f(i,j) + ǫn(i,j), where n(i,j) = f(i +,j) + f(i,j) + f(i,j + ) + f(i,j ). Subtract an expression proportional to the Laplace operator: p(i,j) k 2 p(i,j) = ( + 4k)p k(p(i +,j) + p(i,j) + p(i,j + ) + p(i,j )) = ( + 4k)f(i,j) + ( + 4k)ǫn(i,j) kn(i,j) kǫ(n(i +,j) + n(i,j) + n(i,j + ) + n(i,j )). If k and ǫ are small, the last term can be omitted. If k = ǫ/( 4ǫ), the other two terms vanish, and we get: p(i,j) k 2 p(i,j) = ( + 4k)f(i,j). Thus the original image has been restored (within a constant factor).

Median filter: Averaging reduces sharp spikes (like cosmic rays), but does not remove them; it only spreads their energy to a wider are. A better solution is to replace each pixel by the median of a small neighbourhood. This does not blur edges or affect areas where the gradient is monotonous.

Deconvolution Assume that the psf h is invarian w.r.t translations. The observed image p and the original one f are then related by p(x,y) = h(x x,y y )f(x,y ) dx dy + ν(x,y), where ν is the noise. Taking the Fourier transform of each side we get P(u,v) = H(u,v)F(u,v) + N(u,v), where H is the transfer function of the imaging system. In principle the original image can be solved: F(u,v) = P(u,v) H(u,v) N(u,v) H(u,v). Usually h is roughly a Gaussian profile, the Fourier transform of which is also a Gaussian profile, i.e. almost zero except in a smalle region.

If there is no nois, P = 0 when H = 0, and the ratio P/H is indefinite. When noise is added, H can in some places be much smaller than N, resulting in spikes. Rstoring the image with the filter M(u,v) = /H(u,v) will enhance the noise. A better filter is M(u,v) = { /H(u,v), if u 2 + v 2 r 2, otherwise, where r is sufficiently small to guarantee that H will remain nonzero. In the least squares sense the optimal filter is the Wiener filter: M(u,v) = H(u, v) H(u,v) 2 H(u,v) 2 + S νν (u,v)/s ff (u,v). If the noise is white noide, S νν is constant and approximately S νν (0,0) = R νν (x,y) dx dy. In practice, it is easiest just to experiment which value gives the most satisfactory result.

Another method is based on the concept of enthropy H f = f(m,n) ln f(m,n). m n If the total energy f tot = m f(m,n) remains constant, the enthropy is the smaller the more the values f(m,n) vary. Noise can also get negative values. Add a constant C to the noise so that ν(m,n) + C is always positive: H ν = (ν(m,n) + C) ln(ν(m,n) + C), m n In the maximum enthropy method we maximize the expression n H = H f + αh ν, where α is a constant describing the amount of smoothing. In addition, two constraints are needed:

) The total energy f tot must be conserved. 2) The observed image p must be of the form p(q,r) = m h(q m,r n)f(m,n) + ν(m,n), n i.e the convolution of the original image f and the psf h + noise. By solving this we get f and ν.