ROBUST COLOR JOINT MULTI-FRAME DEMOSAICING AND SUPER- RESOLUTION ALGORITHM

Similar documents
SUPER RESOLUTION FROM MULTIPLE LOW RESOLUTION IMAGES

Video stabilization for high resolution images reconstruction

A Learning Based Method for Super-Resolution of Low Resolution Images

Super-resolution Reconstruction Algorithm Based on Patch Similarity and Back-projection Modification

RESOLUTION IMPROVEMENT OF DIGITIZED IMAGES

Superresolution images reconstructed from aliased images

Low-resolution Character Recognition by Video-based Super-resolution

Accurate and robust image superresolution by neural processing of local image representations

Investigation of Color Aliasing of High Spatial Frequencies and Edges for Bayer-Pattern Sensors and Foveon X3 Direct Image Sensors

High Quality Image Deblurring Panchromatic Pixels

Super-resolution method based on edge feature for high resolution imaging

High Quality Image Magnification using Cross-Scale Self-Similarity

Bildverarbeitung und Mustererkennung Image Processing and Pattern Recognition

Redundant Wavelet Transform Based Image Super Resolution

Extracting a Good Quality Frontal Face Images from Low Resolution Video Sequences

A System for Capturing High Resolution Images

HANDS-FREE PC CONTROL CONTROLLING OF MOUSE CURSOR USING EYE MOVEMENT

Assessment of Camera Phone Distortion and Implications for Watermarking

Latest Results on High-Resolution Reconstruction from Video Sequences

WHITE PAPER. Are More Pixels Better? Resolution Does it Really Matter?

Scanners and How to Use Them

TVL - The True Measurement of Video Quality

High speed 3D capture for Configuration Management DOE SBIR Phase II Paul Banks

INTRODUCTION TO RENDERING TECHNIQUES

Resolution for Color photography

Ultra-High Resolution Digital Mosaics

Mean-Shift Tracking with Random Sampling

FCE: A Fast Content Expression for Server-based Computing

Automatic Labeling of Lane Markings for Autonomous Vehicles

Sub-pixel mapping: A comparison of techniques

Keysight Technologies Using Fine Resolution to Improve Thermal Images. Application Note

A Study on SURF Algorithm and Real-Time Tracking Objects Using Optical Flow

Bayesian Image Super-Resolution

Chapter 1 Simultaneous demosaicing and resolution enhancement from under-sampled image sequences

Robust and accurate global vision system for real time tracking of multiple mobile robots

Template-based Eye and Mouth Detection for 3D Video Conferencing

Super-Resolution Imaging Applied to Moving Targets in High Dynamic Scenes

Interpolation of RGB components in Bayer CFA images

Using visible SNR (vsnr) to compare image quality of pixel binning and digital resizing

Image Interpolation by Pixel Level Data-Dependent Triangulation

A NEW SUPER RESOLUTION TECHNIQUE FOR RANGE DATA. Valeria Garro, Pietro Zanuttigh, Guido M. Cortelazzo. University of Padova, Italy

High Resolution Images from Low Resolution Video Sequences

Sachin Patel HOD I.T Department PCST, Indore, India. Parth Bhatt I.T Department, PCST, Indore, India. Ankit Shah CSE Department, KITE, Jaipur, India

Visualization and Feature Extraction, FLOW Spring School 2016 Prof. Dr. Tino Weinkauf. Flow Visualization. Image-Based Methods (integration-based)

PERFORMANCE ANALYSIS OF HIGH RESOLUTION IMAGES USING INTERPOLATION TECHNIQUES IN MULTIMEDIA COMMUNICATION SYSTEM

Component Ordering in Independent Component Analysis Based on Data Power

A method of generating free-route walk-through animation using vehicle-borne video image

The Role of Size Normalization on the Recognition Rate of Handwritten Numerals

Simultaneous Gamma Correction and Registration in the Frequency Domain


A Proposal for OpenEXR Color Management

Face Recognition in Low-resolution Images by Using Local Zernike Moments

Performance Verification of Super-Resolution Image Reconstruction

HSI BASED COLOUR IMAGE EQUALIZATION USING ITERATIVE n th ROOT AND n th POWER

A Hybrid Artificial Intelligence System for Assistance in Remote Monitoring of Heart Patients

Low-resolution Image Processing based on FPGA

Adobe Marketing Cloud Sharpening images in Scene7 Publishing System and on Image Server

Measuring large areas by white light interferometry at the nanopositioning and nanomeasuring machine (NPMM)

Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches

Resolution Enhancement of Photogrammetric Digital Images

Build Panoramas on Android Phones

Image Compression through DCT and Huffman Coding Technique

The Trade-off between Image Resolution and Field of View: the Influence of Lens Selection

COMPACT GUIDE. Camera-Integrated Motion Analysis

NEIGHBORHOOD REGRESSION FOR EDGE-PRESERVING IMAGE SUPER-RESOLUTION. Yanghao Li, Jiaying Liu, Wenhan Yang, Zongming Guo

Parametric Comparison of H.264 with Existing Video Standards

ESE498. Intruder Detection System

Video Conferencing Display System Sizing and Location

THE CONTROL OF A ROBOT END-EFFECTOR USING PHOTOGRAMMETRY

Performance Analysis and Comparison of JM 15.1 and Intel IPP H.264 Encoder and Decoder

Lecture 12: Cameras and Geometry. CAP 5415 Fall 2010

3D SCANNING: A NEW APPROACH TOWARDS MODEL DEVELOPMENT IN ADVANCED MANUFACTURING SYSTEM

Single Depth Image Super Resolution and Denoising Using Coupled Dictionary Learning with Local Constraints and Shock Filtering

Tracking and Recognition in Sports Videos

To determine vertical angular frequency, we need to express vertical viewing angle in terms of and. 2tan. (degree). (1 pt)

Circle Object Recognition Based on Monocular Vision for Home Security Robot

A PHOTOGRAMMETRIC APPRAOCH FOR AUTOMATIC TRAFFIC ASSESSMENT USING CONVENTIONAL CCTV CAMERA

Tracking of Small Unmanned Aerial Vehicles

White Paper. "See" what is important

RESOLUTION MERGE OF 1: SCALE AERIAL PHOTOGRAPHS WITH LANDSAT 7 ETM IMAGERY

A Novel Method to Improve Resolution of Satellite Images Using DWT and Interpolation

Whitepaper. Image stabilization improving camera usability

Algorithms for the resizing of binary and grayscale images using a logical transform

Automatic 3D Mapping for Infrared Image Analysis

A Reliability Point and Kalman Filter-based Vehicle Tracking Technique

Synthetic Aperture Radar: Principles and Applications of AI in Automatic Target Recognition

Planetary Imaging Workshop Larry Owens

An Energy-Based Vehicle Tracking System using Principal Component Analysis and Unsupervised ART Network

Subspace Analysis and Optimization for AAM Based Face Alignment

Transcription:

ROBUST COLOR JOINT MULTI-FRAME DEMOSAICING AND SUPER- RESOLUTION ALGORITHM Theodor Heinze Hasso-Plattner-Institute for Software Systems Engineering Prof.-Dr.-Helmert-Str. 2-3, 14482 Potsdam, Germany theodor.heinze@hpi.uni-potsdam.de ABSTRACT This paper introduces a robust super-resolution algorithm for joint color multi-frame demosaicing. We show that our algorithm, although fast and simple, exhibits convincing results not only within the modeling assumptions, but also for real raw data series. The ultimate goal is its application to telemedical patient monitoring through mobile devices with limited computing power and low quality imaging devices. KEY WORDS Image Sequence Processing, Image Quality, Image Restauration, Digital Image Processing, Image Enhancement, Super-Resolution. 1. Introduction Recent developments in the field of Color Field Array (CFA) imaging sensors reveal a trend towards high frame rates while resolution of sensor / lens seems to converge towards physical limitations. Usage of mobile devices like smartphones, which often offer primitive but highframerate video functionality, spreads rapidly. Working in the field of telemedicine for infrastructurally underdeveloped regions [12], our current research considers improving image quality of simple imaging devices to allow for in-field telemedical assistance [11] and machine learning decision support. Imaging processors of current mobile devices use excessive noise reduction in the demosaicing process [2] to improve visual appeal of the data coming from tiny, noisy sensors. This leads to smearing of low contrast parts of the image thus awkwardly rendering important details like human skin, which is highly undesirable in the telemedical context. One approach to improving image quality is multiframe super-resolution [9]. The idea is to use information of multiple sequential frames to obtain one single frame of higher quality. Such algorithms have been introduced already [1] [7] [10], mostly with maximum a posteriori (MAP) estimation techniques. In [8] an algorithm based on the projections onto convex sets technique is proposed, which uses single-frame demosaicing before applying super-resolution steered by several constraint sets. In [4], the authors experiment with normalized convolution interpolation. The main contribution of this paper is a new multi-frame super-resolution algorithm which is highly adaptive to frame-rate, sensor sensitivity and computation time. We will show experimentally, that our algorithm performs well within modeling assumptions, as well as in real world conditions. Our further research will include optimizations of this algorithm towards rendering of human skin and tests with simple imaging devices. 2. Imaging Model When capturing a CFA image from the real world, we basically get only a distorted reflection of the reality. To take that into account, the image formation model we used for algorithm design and experiments is largely the same as described in [7] and [1]. Thus, to simulate CFA raw data capturing, we need to apply the following transformations to the source image: 1. Downsampling. 2. Blur. 3. CFA filtering. 4. Noise addition. Simulating capturing a series of raw data files means that step 1 (downsampling) needs also to be added a random The project on which this report is based was funded by the Ministry of Economy of Brandenburg and the EU under grant No. 80136574. This publication reflects the views only of the author.

shift to compensate for camera movement between the shots. One of the main ideas of super-resolution is that due to that camera movement, for every differentiable visible point of the real world, every captured image frame provides a unique description with slightly differing CFA color surrounding. This fact allows for resolution gain by jointly demosaicing multiple frames into one. determined. This information is then passed on the demosaicing algorithm. In our implementation, for speed and simplicity reasons we have only determined the mean frame shift information for x and y axes. For resolution increase, it is highly desirable to estimate shift between two frames not only as an integer value, but as precise as possible. There are many approaches like phase correlation and cross correlation [3] to estimate similarity between two images. Since for our demosaicing algorithm accuracy in shift estimation is crucial, we have implemented the following algorithm to estimate the exact shift between two frames: REPEAT r times Figure 1. Information gain of an image sequence. Figure 1 shows how a CFA sensor sees the world on the left. On the right, we overlapped 3 frames, which are slighty shifted to simulate camera movement. Now it is obvious, how much more exploitable demosaicing information an image sequence has to offer. Not only is there an overlapping of different color channels, the color patches are also much finer, which makes it possible to derive a higher resolution image. To make use of that, it is very important to align the image sequence as precisely as possible. The possible image quality gain depends on the number of the frames in the sequence, as well as on shift characeristics. If frames are not shifted at all, then any additional resolution in the joint demosaicing process can not be gained either. We will refer to the multiple raw CFA frames as source, while the goal image we will call the target. The target is often, but not necessarily, of higher resolution than the source frames. In fact, its resolution can be chosen freely within our algorithm. 3. Image Registration The main goal of image registration is to provide transformation information of one series of frames to fit one coordinate system. E.g. one of the frames in the series is picked to be the pivot frame. Usually, this is the frame with the least cumulative distance (mean square error sum of all RGB pixels) to the others. Then, for each pivot pixel p(i,j) transformation information (shifting, rotating, warping) towards each other frame is Pick p random pixels Find shift (integer value) based on cross correlation Compute mean value (floating point) from all r runs The idea is that the r semi-random runs will distribute around the real shift. Of course, estimation accuracy improves with increasing number of runs, while a small r might lead to a very inaccurate estimation. In our experiments we usually set r=100. 4. Joint Demosaicing Algorithm After image registration, joint demosaicing can be performed. The idea is to use the exact distance information between the target pixel and the corresponding source pixels projected onto the target plane to calculate a robust and reasonably precise estimate of the real value. Take a look at the following pseudocode: FOR each pixel (i,j) of target image t FOR each color channel {r,g,b} make a list of n neighbor source pixels from all source frames (value, distance) target value = weighted sum of the neighbor pixel values (closer distance means more weight)

For every target pixel, for each color channel a list with some nearest source pixels is made. Shift information is used to determine the nearest source pixels of a particular color for every source frame. Figure 4. Demosaicing problem: Pixel projection of source frames onto target plane. Figure 2. The nearest three red source pixels for target(i=5, j=4) located on frame 1 (pivot). Figure 4 illustrates the demosaicing problem. The goal is to estimate the value of the target pixel. We assume that light hitting the sensor follows a certain 2ddistribution and source pixel values correspond to the density of that distribution. Now the goal is to estimate that distribution for the target pixel area. For distribution estimation, there are plenty of known techniques, including machine learning. Since our algorithm needs to be fast, we went for a much simpler approach. It is easy to configure our algorithm towards speed or noise reduction. A low value for n would mean fast calculation, while a high number of contributing source pixels would mean better noise reduction. The formula is as follows: color value =!!!!!(!!!!! )!!!!!!!! Figure 3. The nearest three red source pixels for target(i=5, j=4) located on frame 2, which is slightly shifted towards the pivot frame. Figure 2 and 3 show how red source pixels are acquired from two source frames. The target pixel is depicted by a black spot and located at (i=5, j=4) in target coordinates. Here, the target resolution is twice the source resolution both vertically and horizontally. The nearest red source pixels are (2,1), (2,3) and (4,1) in the 1 st frame and (0,1), (2,1) and (2,3) in the second. The closer source pixels are located to the target, the more they contribute to its value. The distances to the target pixel are marked by black lines. with n number of source pixels per frame. f s number of frames. color value of source pixel. d distance of source pixel to target pixel. w(d) weight with which a source pixel contributes to the target pixel value. Function of distance.

Varying the w(d) term allows to configure the algorithm for more spatial resolution by giving the closest source pixel a dominating impact, while considering many source pixels to a nearly equal part will give heavy noise reduction at the cost of a slightly blurred image. Flexible involving of source pixels makes the algorithm very adaptive towards sensitivity, resolution, frame rate and processing speed issues. 5. Experiment Setup Our experiments consist of two main parts. First, we experiment with joint demosaicing of synthetically generated raw files. From one single high-resolution image ( real world ) we derive a series of randomly shifted, downsampled, blurred, CFA filtered images with Gaussian noise ( raw image series ). The goal for our algorithm is then to estimate the shift between the frames (image registration) and jointly demosaic them. In the second part, we show that our modelling assumptions hold very well for the real world. Therefore, we show joint demosaicing results we gained from a CFA raw image series shot with a Sony NEX-5 camera and compare them to single frame demosaicing results from a single frame a) JPEG straight from the camera and b) Adobe Camera Raw at default settings. To simulate a small and noisy sensor, Sony NEX-5 was set to very high sensitivity (ISO 12800 and ISO 3200 resp.) in the examples Hand and Color Theory. 6. Experimental Results Figure 7 shows a resolution chart ( Cornell University, Stephen H. Westin) synthetically decomposed into a series of raw images with random sub-pixel accuracy shift and then rendered bilinearly from one frame (bottom) and by multi-frame demosaicing with a pixel resolution increased with factor 2 (5 frames, mid). In this and the following examples, top image shows the overall view of the picture. Frame shift information has been determined by the algorithm described in the image registration section. The mean shift estimation error was 0.197 pixels, which means that the mean error in source/target pixel distance estimation was about 1/5 of a source pixel. There is some difference in resolution visible, our algorithm is also definitely more robust towards Moire and color fringing. Figure 8 shows a comparison with a frame rendered by Adobe Camera Raw (ACR). First pair exhibits a resolution advantage, especially when considering the rendering of the EU stars, from which ACR skipped one completely. Second pair shows Moire patterns in the ACR image and soft detail roll-off in the multi-frame rendering. Third pair, again, shows differency in treatment of fine detail. Figure 5. Synthetic Experiment Set-Up

Figure 6. Rendering of human skin. Left side: multiframe rendering (9 frames), Right side: jpeg from camera (Sony NEX-5) One important goal in the telemedical context is improved rendering of human skin. Here with Figure 6, we show that a high frame-rate might compensate for low image quality. Although the raw files contained extreme amount of image noise, our algorithm proved its robustness by accurately rendering important low contrast skin detail. Figure 9 shows the Color Theory chart, with kind permission from Paper Leaf Design (paperleaf.com). Multi-frame demosaicing (left side) vastly improves resolution and color rendition here. Since our prototype is implemented in Java and not optimized at all, we did not measure computing time systematically. On our 1-Ghz subnotebook it takes 20 seconds to jointly demosaic 5 source frames of 14.2 megapixels into a target with a resolution of also 14.2 megapixels and slow settings (n=9, 5x5 mask). Figure 7. Resolution chart rendering bilinearly (200% view, bottom) and multi-frame super-resolution (100%, mid).

Figure 8. Left side: multi-frame demosaicing (4 frames). Right Side: single frame rendered by Adobe Camera RAW (default settings) Figure 9. Color Theory : JPEG from Sony NEX-5 (ISO 3200) on the right versus multi-frame rendering (10 frames) on the left.

7. Conclusion and Future Work We have introduced a robust and adaptive random multiframe joint interpolation process. Experimental results show that it is well capable to extract additional information from image sequences when compared to single frames, even if they are processed by a state-of-theart raw-convertor. Also, its configurable processing speed and image noise adaptivity alleviate its application to mobile imaging. Our future research will include testing with mobile devices which feature limited imaging and computing capabilities. Also, we will research if the algorithm is appropriate for an application to livestreaming in telemedical context [6]. For further image quality improvement, we are currently experimenting with artificial neural networks, which have already been employed for demosaicing [5]. Our ultimate goal is to apply image sequence processing to telemedical decision support issues. References [1] M. Trimeche, Color Demosaicing Using Multi-Frame Super-Resolution. European Conference on Signal Processing 2008, Eusipco (2008): n. pag. [2] X. Li, B. K. Gunturk, & L. Zhang, Image Demosaicking: A Systematic Survey, Proc. of SPIE Electronic Imaging Conference, vol. 6822, 15 pages, 2008. [3] B. Zitová, J. Flusser, Image registration methods: a survey, Image and Vision Computing vol.21, 11 (2003), p. 977-1000 (2003). [4] P. Vandewalle, K. Krichane, D. Alleysson & S. Süsstrunk, Joint Demosaicing and Super-Resolution Imaging from a Set of Unregistered Aliased Images, Proc. IS&T/SPIE Electronic Imaging: Digital Photography III, Vol. 6502, 2007. [5] H. Z. Hel-Or & O. Kapah., Demosaicing using Artificial Neural Network. SPIE, p. 112-120, 2000. [6] Robinson M.D, Chiu S.J., Lo J.Y., Toth C.A., Izatt J.A, & Farsiu S, Novel Applications of Super-Resolution in Medical Imaging, Super-Resolution Imaging, Peyman Milanfar (Editor). (CRC Press, pp. 383-412, 2010). [7] S. Farsiu, M. Elad, & P. Milanfar, Multi-Frame Demosaicing and Super-Resolution of Color Images. IEEE Transactions on Image Processing, vol. 15, no. 1, pp. 141-159, 2006. [8] M. Gevrekci, B. K. Gunturk, & Y. Altunbasak, Restauration of bayer-sampled image sequences, Oxford University Press, Computer Journal, vol. 52, no. 1, pp. 1-14, January 2009. [9] S. C. Park, M. K. Park, & M. G. Kang, Superresolution image reconstruction: a technical overview, IEEE Signal Processing Magazine, Vol. 20, No. 3. (May 2003), pp. 21-36. [10] S. Farsiu, M. Elad, & P. Milanfar, Multi-Frame Demosaicing and Super-Resolution from Under-Sampled Color Images, IS&T/SPIE Symposium on Electronic Imaging 2004, CA. Volume 5299, pp. 222-233. [11] H. Greenspan, Super-Resolution in Medical Imaging, The Computer Journal (2009) 52(1): 43-63, 2008. [12] T. Heinze, R. Wierschke, A. Schacht, & M. von Lowis, A Hybrid Artificial Intelligence System for Assistance in Remote Monitoring of Heart Patients, Hybrid Artificial Intelligent Systems Proceedings (2) 2011, pp. 413-420, 2011.