Light Field Denoising and Upsampling Using Variations of the 4D Bilateral Filter

Size: px
Start display at page:

Download "Light Field Denoising and Upsampling Using Variations of the 4D Bilateral Filter"

Transcription

1 Light Field Denoising and Upsampling Using Variations of the 4D Bilateral Filter Mariana Barakchieva University of Bern, Switzerland Figure 1: Denoising of a light field. From left to right: the original light field (LF) central perspective, LF with added noise, denoised LF using the 4D Bilateral filter. Abstract This paper presents an effective and not complicated method for denoising and upsampling of light fields (LF). Denoising of a light field scene is an important problem, since sensor noise is often affecting image quality, especially in low-light conditions. Afterwards the denoised light field can be used for further processing, e.g. depth reconstruction and superresolution. Also, denoising a LF can be used to produce a clean single view, resulting in better signal-to-noise ratio (SNR) than any other single-image denoising method. Having in mind the multi-dimensional structure of the LF, it is also important to have a quick and efficient algorithm for downsampling and upsampling the LF. The method proposed here is based on the four-dimensional Bilateral filter. It solves both problems fast and outperforms other related algorithms. CR Categories: I.4 [Image Processing and Computer Vision]: ; [I.4.4]: Image Processing and Computer Vision Restoration; Keywords: light fields, denoising, upsampling, 4D filter, bilateral filter Links: DL PDF 1 Introduction When light fields were introduced to computer graphics some 17 years ago, the main application proposed was to render new views[levoy and Hanrahan 1996; Gortler et al. 1996]. Today, with the increase in computer speed, memory and bandwidth, the research interests in light fields has broadened, too. With two com- [email protected] mercial light field cameras on the market [Perwass and Wietzke 2012; Ng 2006], scientist look at re-devising traditional image processing techniques for light fields as well as inventing novel ideas that take into account the properties of the plenoptic function. Problems approached include multiperspective panoramas, synthetic aperture, refocusing, microscopy. Going further, Marc Levoy, a leading expert in computational photography, predicted that in less than 20 years from now, most consumer cameras will be light field cameras [Levoy 2006]. This paper solves two main problems - denoising and upsampling. Since many computer vision applications involve low-light condition of image capturing, having a simple and fast denoising algorithm is crucial. Unlike traditional denoising approaches, ours takes into account the four-dimensional structure of the light field (LF), removing the noise not of a single 2D perspective, but rather of the whole light field. The second issue is that algorithms for light fields processing usually involve high computational costs (since we add the angular domain of the LF). Thus, an alternative is to work on a low-resolution version of the light field and afterwards upsample the computed results. For both applications we use the four-dimensional Bilateral filter, and for the task of upsampling a reference high resolution LF is also needed. Before reviewing other works that solve the same or similar problems (Sec. 2), we propose a short introduction to light fields. Afterwards, the method we use is explained (Sec. 3) and results and comparisons are shown (Sec. 4). The final section proposes ideas for further development and concludes the paper. 1.1 From plenoptic function to light field In 1936, Arun Gershun first coined the term light field as the amount of light traveling in every direction through every point in space [Gershun 1939]. To formalize it using geometrical optics theory, the plenoptic function is defined. It is the radiance along all rays in a 3D space, and consists of five dimensions - the ray position (x,y,z) and direction (θ, φ). However, since radiance is constant along a ray (if there are no blockers), one dimension is redundant. Thus, we get to a 4D function that Marc Levoy named 4D light field and that is defined as radiance along rays in empty space [Levoy and Hanrahan 1996]. An intuitive way to visualize a light field is with the two-plane parametrization (see Fig. 2). Light rays are parametrized by the

2 signal-to-noise ratio (SNR)) and lower dynamic range [Liu et al. 2002]. Secondly, a lot of computer vision applications involve capturing images in imperfect and low-light conditions. A summary of several denoising algorithms follows below. (a) Figure 2: Two-plane parametrization of light fields: (a) the two planes, (u,v) and (s,t) are parallel and one can think of the camera image plane being parallel to (u,v) plane and the scene being behind (s,t) plane (b) the light field as a set of 2D images, each captured from a different observer position on the (u,v) plane. Figure 3: Scene image and corresponding EPI. Note that the EPI captures depth information - the lines with steeper slope are in the foreground in the scene (less depth value) [Chai et al. 2000]. intersection with the two parallel planes and each ray stores RGB radiance. From photographic point of view, the light field can be seen as a collection of images of the (s,t) plane (and any objects behind it), taken from each position on the (u,v) plane. Then one needs to define ray space: it is a coordinate system with v and t axes; thus, a ray from the scene is a point in ray space and all rays through one scene point form a line in ray space. The slope of a line in ray space is inversely related to depth in the original scene. Related to it is the epipolar plane image (EPI) - a 2D image, constructed by stacking one over the other image lines (fixed t coordinate) from all images along a row of the (u,v) plane (fixed v coordinate) (see Fig. 3). Next, we summarize works that deal with the problems of denoising and upsampling. 2 Related work 2.1 Denoising of light fields First, we will review works that solve the problem of light field denoising. While single-image denoising has developed for a while (a summary and evaluation of classical methods is provided in [Buades et al. 2005]), we could find only a few works that deal with multiple-image denoising. Indeed, this is an important issue because first of all, plenoptic cameras aim at a very high sensor resolution, thus dense and small sensor pixels, which makes them prone to noise; for example, all currently available light field cameras have a CMOS sensor, which generally suffers from high read noise and non-uniformity, resulting in more noisy images (lower (b) In [Zhang et al. 2009], multiple view denoising is conducted and the algorithm does not aim at denoising a light field but poses the problem as pinhole image denoising; for the solution several pinhole images from different prospectives are captured. Using a second input parameter - depth estimation, it groups similar patches in the different images, then models intensity-dependent noise and uses principal component analysis (PCA) and tensor analysis to remove that noise. The authors conducted experiments with images with added synthetic noise and scored highest compared to stateof-the-art single image denoising methods. The two main disadvantages of this algorithm are that it requires prior depth information (although the authors claim that even a low-quality depth map is enough) and outputs a single denoised image, rather than a denoised light field (loss of angular information). Next, Mitra and Veeraraghavan propose a patch based approach for solving light field processing tasks whose observation models are linear - including denoising, supperresolution, and refocusing [Mitra and Veeraraghavan 2012]. They use Gaussian Mixture Model (GMM) to model the light field patches. The inference algorithm proposed consists of extracting patches from observed data, then estimating disparity values (with a fast subspace projection), and at the end reconstructing the corresponding light field patches using Linear Minimum Mean Square Estimator (LMMSE) algorithm. For testing purposes, Gaussian distributed noise is added to a light field and is then removed effectively. However, no comparison with another denoising approach is presented. As mentioned in the paper, the algorithm is limited to diffuse scenes and small enough patches (otherwise depth discontinuities are possible). An efficient variational framework that takes into account the structure of light fields is presented in [Goldluecke and Wanner 2013]. They solve inverse problems on ray space, such as denoising, inpainting, and ray space labeling. This works by constructing convex priors for light fields that preserve the epipolar plane image structure and satisfy constraints related to object depth and occlusion ordering. In effect, they do regularization of vector-valued functions on ray space, while also respecting the light field geometry. Denoising is demonstrated for Raytrix plenoptic camera and synthetic light fields with added significant amount of Gaussian noise. The algorithm performs better than single-image denoising algorithms but no comparison is made to multiple-image denoising. The main disadvantage of this method again is that it requires a depth map estimation as an additional input. The denoising method proposed in [Dansereau et al. 2013] is similar to our method in the sense that denoising is conducted with a single linear filter. The main observation is that the light field of a diffusive scene has a 4D hyperfan-shaped frequency-domain region of support at the intersection of a dual-fan and a hypercone (see Fig. 4). Knowing that, they design a filter with appropriately shaped passband that filters out the noise. Experiments were conducted with plenoptic camera light fields and added different noise types. Comparison is made with competing methods, including synthetic focus, fan-shaped anti-aliasing filters, and different state-of-the-art nonlinear image and video techniques. Visually, the hyperfan outperforms the other methods but its disadvantage is that it does not work on non-lambertian and occluding scenes. That is because if there is occlusion, the planes in the light fields will be truncated, and in the case of non-lambertian surfaces, rays within the plane have different values; both of these conform to the hyperfan passband and thus filtering these areas might result in attenuation.

3 Figure 4: The maximum magnitude per frequency, calculated for six light fields, shows the hyperfan shape[dansereau et al. 2013]. Figure 6: Bilateral filtering[durand and Dorsey 2002]. 3 Method Figure 5: Mapping from input image to the image plane of the novel view; The point x is visible on the new image, while the point x is not - this is inferred by the depth map of the scene[wanner and Goldluecke 2013]. 2.2 Light fields upsampling We will review two papers that deal with upsampling of light fields. The work of Jarabo et al. incorporates downsapling and upsampling techniques as part of a framework for efficient propagation of light field edits [Adrian Jarabo and Gutierrez ]. First, they define a similarity metric, adapted to the context of light field editing, which models the affinity between pixels. Thus, downsampling is conducting by first mapping all pixels into affinity space, then recursively subdividing the space into clusters and getting a single representative for each cluster. After processing is done on the downsampled LF, they do upsampling with the Joint bilateral upsampling, using a guidance full-resolution LF. This is a very similar solution to the method described in this paper. However, since they project in affinity space, searching for a pixel neighbor is more computationally expensive. The method has two main disadvantages - first, it has complexity linear with the light field size, and second, it requires large memory for storing the clusters and the correspondence between pixels and clusters. The second paper describes upsampling in the spatial domain and generation of novel views in the angular domain, by solving a variational inverse problem [Wanner and Goldluecke 2013]. The algorithm is best explained in Fig. 5: having a depth map as additional input, the scene surface Σ can be inferred; then a transfer map τ i projects from the input image Ω i to the image plane Γ of the novel view. Scene ordering is preserved using a binary mask m i. The advantage of the described method is that it takes into account the scene geometry, but also requires a depth map as additional input and the output is a single super-resolved LF view, not a LF structure. Next, an explanation of the method adopted in this paper follows. Taken into account the high-dimensional structure of the light field, a robust filter would have to filter on all dimensions. One option would be to use the Gaussian filter. Since it is a separable filter, its extension to 4D is trivial. However, the Gaussian filter smooths also over the edges of the images. Thus, we worked on extending the Bilateral filter to four dimensions and looked at applications of it on light fields. First, since light field cameras capture a lot of redundant information, the 4D Bilateral filter could be used to elegantly reject unwanted noise, while at the same time preserve details and edges. Secondly, the 4D filter is also efficient in upsampling light field data, which is applicable in all computationally costly operations on light fields - like for example depth map computation. 3.1 Bilateral Filtering The Bilateral filter is a non-linear edge-preserving smoothing filter, first introduced by Tomasi and Manduchi in 1998 [Tomasi and Manduchi 1998]. It consists of two parts - spatial/domain and range filter kernel. J p = 1 I qf(p q)g(i p I q) (1) k p q where I is the input image and J is the output/filtered image; p and q are 2D pixel positions (center p and neighborhood q); f is the spatial filter kernel (2D Gaussian, controlled by spatial sigma - σ s); g is the range kernel, which acts as a penalty on the input difference (1D Gaussian function, controlled by range sigma - σ r); k p is normalization factor, sum of all weights f g (see Fig. 6). Edges are preserved since the bilateral filter f g decreases as the range distance and/or the spatial distance increases. Elevation of the two-dimensional Bilateral filter to higher dimensions is proposed in [Jayme C. Kosior 2007; M Mendrik 2011] for the purposes of denoising of 4D MR and CT data, where the signal is 4D and the time-intensity profiles need to be preserved. [Baek and Jacobs 2010] proposes methods for accelerating spatially varying high-dimensional Gaussian filters, and especially the bilateral filter. Formally, the 4D Bilateral filter does not differ much from the 2D one: now, p and q are 4D pixel coordinates, and the input and output images are 4D, too. Applied to the 4D light field with added white Gaussian noise, the filter cancels the noise and preserves details and edges. However, it works for only small noise quantities. Fig. 7 shows the decay of the denoising method as the noise standard deviation increases. Thus all experiments in Sec. 4 were made on LFs with noise σ =

4 Figure 7: The peak signal-to-noise ratio (PSNR) of the noisy and denoised LF as noise σ increases; also shown here is the PSNR of the central LF perspective - it does not differ much from the PSNR of the whole denoised LF, which is one of the main advantage of the method, proposed here. Figure 8: Joint bilateral upsampling method. The low-resolution image determines the spatial kernel and the high resolution image is used as guidance on the range kernel; the output is a highresolution solution. 3.2 Joint Bilateral Filtering An extended version of the 4D bilateral filter is used when smoothing an image without crossing strong edges in some other reference image. This is referred to as joint (or cross) bilateral filter. It was first introduced by [Eisemann and Durand 2004; Petschnigg et al. 2004] and used was used for combining images taken with and without a flash. Later, [Kopf et al. 2007] proposed upsampling a low-resolution image, using a high-resolution reference as another application of the joint bilateral filter. This is done by interpolating the low-resolution image in a manner that does not cross strong edges in the high-resolution reference image [Adams et al. 2011]. The only difference to Equation. 1 is that now the range filter is applied to a second guidance image, Î. J p = 1 k p q I qf(p q)g(îp Îq) (2) Joint Bilateral upsampling is used then to construct a high resolution solution from a full-resolution guidance image Î and a lowresolution solution S (computed from a downsampled version of the image). Ŝ p = 1 S qdown f(p down q down )g(îp Îq) (3) k p q down Note that the resolutions of the guidance image and the input solution are different, thus while q down takes only integer coordinates, p down can also be rational. An illustration of Equation. 3 is shown in Fig. 8. One of the applications of the Joint Bilateral upsampling on light fields is depth estimation, i.e. to upsample low-resolution depth map with guidance from the high resolution light field (see Fig. 9). Next, results and comparison with competing algorithms follows. 4 Results Denoising Experiments were conducted with both synthetic light fields (from the Heidelberg archive) and light fields captured with Figure 9: Depth estimation algorithm illustration. We first downsample the LF, compute its depth map and then upsample the depth map. the Lytro camera. A quantitative metric was used to evaluate the different denoising methods - peak-signal-to-noise-ratio (PSNR). For a 4D reference light field LF ref and its noisy counterpart LF noise, PSNR is defined using the mean-square error (MSE): MSE = 1 T SUV (LF ref LF noise) 2 t s u v PSNR = 20 log 10 MAX LFref MSE (4) where T, S, U, V is the size of the LF at each dimension, and MAX LFref is the maximum possible pixel value. PSNR is expressed in terms of the logarithmic decibel scale, where the closer the noisy image to the original noise-free image, the higher the PSNR is (for identical images, PSNR is infinity). In the case of synthetic LF, white Gaussian noise with standard deviation was artificially added to the original LF (see Fig. 10). This results in visually low noise in the image and SNR of 25 db. This is more noise than what is expected from the commercially available Lytro camera. The Lytro camera uses the CMOS image sensor Aptina MT9F002 1, which exhibits SNR of 35.5dB 2. Optimal parameters were found using brute force search. For LF denoising, the optimal spatial sigma σ s = 2 and σ r = Experiments with a synthetic LF Mona s room is shown in Fig. 10. Our method shows best qualitative results, i.e. highest PSNR. Also visually, denoising with the 4D Bilateral filter preserves details, which can be best seen in the detailed crops - the pattern on the leaves and the grainy structure on the purple letter are preserved. Also, the fact that our method denoises the LF and preserves consistency across

5 Figure 12: Top: superresolving a noisy LF, PSNR=26.74 db - noise is not filtered completely and artifacts are visible; Bottom: first denoising the LF and then superresolving it, PSNR= db - more smooth and coherent result. views is illustrated by calulating the PSNR of a single image - the central LF perspective. All other methods achieve the strongest denoising on the central perspective, while the 4D Bilateral filter equally denoises all perspectives. Experiments were also conducted with light fields, captured with the Lytro camera in low-light conditions (test data provided by [Dansereau et al. 2013]). First, the image gain was adjusted, which brings even more noise to the image, and afterwards the resulted LFs were denoised using the 4D Bilateral filter (see Fig. 11). Optimal parameters were found at σ s = 2 and σ r = A quantitative metric could not be adopted, since a reference, noise-free LF is lacking. Finally, another idea that we explored was to verify if when superresolving a LF, the little amount of noise is not cleared already. Thus, white Gaussian noise with σ = was added to a LF and then supperesolved using the method, described in [Wanner and Goldluecke 2012b]. The result can be seen on Fig. 12 (top). Then this is compared to first denoising the noisy LF with the 4D Bilateral filter and then superresolving the result - seen in Fig. 12 (bottom). Artifacts from the noise can be seen in the first solution, while the second solution is more coherent and has higher signal-to-noise ratio with about 10 db. Upsampling Two experiments were conducted with upsampling. First, to measure the speed and quality loss when upsampling with the Joint Bilateral filter, a LF was first downsampled and then upsampled, while measuring the execution time and also computing Figure 13: Time to upsample a LF (top) vs. quality loss (bottom); spatial scale factor of 1, 2, 4, 8, and 16 correspondingly, and angular factor of 2, for downsampling (top red) and upsampling (top blue). an error metric - the peak-signal-to-noise ratio as introduced earlier. To get a low-resolution LF, the spatial (ST) domain was downsampled by a certain spatial scale factor, while the angular (UV) domain was always downsampled by 2 (only odd perspectives taken). For upsampling, the optimal spatial sigma σ s = 0.5 and σ r varies from 0.01 to 0.1. Fig. 13 show the trade-off between computational time and quality loss as the spatial scale factor increases. Also, central perspective images from the upsampled LF for different scales are shown in Fig. 14. Finally, Fig. 15 shows an original angular perspective (a) and an upsampled one (b), aiming to illustrate the quality of the method when upsampling on the LF angular domain. As it can be seen, there is no visible quality degradation when upsampling the angular domain. Secondly, one application of the Joint Bilateral upsampling was explored - depth calculation. The idea is that since depth map calculation is an expensive operation, a low-resolution LF can be used and then upsampled. The time needed should be smaller and the quality loss insignificant. To test that, we used the depth map calculation method described in [Wanner and Goldluecke 2012a] and the CO- COLIB light field suite [Wanner and Goldluecke 2013]. The depth map calculation require two steps - obtaining EPI depth estimates, which is computationally fast, and then integrating these estimates into a consistent single depth map. The latter is very computationally expensive operation and is proportional on the resolution of the LF. For the synthetic LF with spatial resolution 768x768 pixels and angular of 9x9, the first step took about 74 seconds, and the second step 151 seconds per perspective (12231 seconds or more than 3 hours for the whole LF). Thus, assuming that the processing time changes linearly with the number of rays, downsampling the

6 Original LF central perspective LF with added white Gaussian noise, σ = D Gaussian PSNR: LF=25.40 db, image=25.40 db PSNR: LF=31.63 db, image=31.63 db 4D Gaussian 2D Bilateral [Goldluecke and Wanner 2013] PSNR: LF=33.61 db, image=33.62 db PSNR: LF=32.72 db, image=32.71 db PSNR: LF=32.76 db, image=34.03 db [Mitra and Veeraraghavan 2012] [Dansereau et al. 2013] 4D Bilateral (our method) PSNR: LF=32.56 db, image=36.62 db PSNR: LF= db, image=34.15 db PSNR: LF=34.17 db, image=34.22 db Figure 10: Denoising of a synthetic LF with various multiple-image denoising methods. PSNR on the whole LF is shown, as well as PSNR of the central LF perspective.

7 Figure 11: Denoising of Lytro light fields with the 4D Bilateral filter: first column - LF central perspective after gain adjustment, second column - denoised LF, central perspective. Ground truth Scale factor 1 Scale factor 2 Scale factor 4 Scale factor 8 Scale factor 16 PSNR=47.80 db PSNR=37.90dB PSNR=34.97 db PSNR=31.44 db PSNR=29.13 db Figure 14: Upsampled LF after downsampling with scale factors of 1, 2, 4, 8, and 16 correspondingly. Here a detail of the central LF perspective is shown, as well as the peak signal-to-noise ratio.

8 ACM, New York, NY, USA, SIGGRAPH ASIA 10, 169:1 169:10. BUADES, A., COLL, B., AND MOREL, J. M A review of image denoising algorithms, with a new one. Simul 4, CHAI, J.-X., TONG, X., CHAN, S.-C., AND SHUM, H.-Y Plenoptic sampling. In Proceedings of the 27th annual conference on Computer graphics and interactive techniques, ACM Press/Addison-Wesley Publishing Co., New York, NY, USA, SIGGRAPH 00, (a) Figure 15: Illustration of upsampling in the angular domain: (a) original LF perspective at (4,4); (b) upsampled LF perspective at (4,4). PSNR = db LF by a scale factor of 4 and an angular factor of 2 would speed up the depth map calculation by a factor of = 64 (theoretically resulting in about 191 seconds or 3 minutes). We need to add to that the time to do the downsampling and upsampling (468 seconds) and still we get execution time of 11 minutes, compared to 3 hours without upsampling. Fig. 16 summarizes the quality loss, compared to other upsampling methods. 5 Conclusion and further work This paper proposes a fast and easy method for light field denoising and upsampling. The main advantage of the method is that it does not require additional information such as depth map, and that is a plus especially with non-synthetic light fields, where computing the depth map is still not trivial. The method outperforms both visually and quantitatively other related methods for the case of low noise levels. However, as described in Sec. 4, this is also the noise level exhibited in current light fields cameras. As future work, optimizations could be performed on the algorithm, so as to speed up the Bilateral filtering. There is a lot of research on fast Bilateral filtering, a good example being the work of Baek et al. [Baek and Jacobs 2010]. Furthermore, applying the joint bilateral filter iteratively could be explored, as proposed in [Riemens et al. 2009]. To evaluate the quality of denoising of our algorithm on real LF (i.e. captured with a light field camera or camera array), Steins unbiased risk estimator (SURE) can be used, which can estimate the accuracy of the denoising algorithm, without the need of clean, noise-free image [Kishan and Seelamantula 2012]. Acknowledgements I would like to thank David and Clemens for their enormous help, constant support, and clever ideas. References ADAMS, A., LEVOY, M., GUIBAS, L., HOROWITZ, M., AND OF COMPUTER SCIENCE, S. U. D High-dimensional Gaussian Filtering for Computational Photography. Stanford University. ADRIAN JARABO, B. M., AND GUTIERREZ, D. Efficient propagation of light field edits. BAEK, J., AND JACOBS, D. E Accelerating spatially varying gaussian filters. In ACM SIGGRAPH Asia 2010 papers, (b) DANSEREAU, D. G., BONGIORNO, D. L., PIZARRO, O., AND WILLIAMS, S. B., Light field image denoising using a linear 4d frequency-hyperfan all-in-focus filter. DURAND, F., AND DORSEY, J Fast bilateral filtering for the display of high-dynamic-range images. In Proceedings of the 29th annual conference on Computer graphics and interactive techniques, ACM, New York, NY, USA, SIGGRAPH 02, EISEMANN, E., AND DURAND, F Flash photography enhancement via intrinsic relighting. In ACM SIGGRAPH 2004 Papers, ACM, New York, NY, USA, SIGGRAPH 04, GERSHUN, A The light field. J. Math. and Physics 18, GOLDLUECKE, B., AND WANNER, S The variational structure of disparity and regularization of 4d light fields. GORTLER, S. J., GRZESZCZUK, R., SZELISKI, R., AND COHEN, M. F The lumigraph. In Proceedings of the 23rd annual conference on Computer graphics and interactive techniques, ACM, New York, NY, USA, SIGGRAPH 96, JAYME C. KOSIOR, ROBERT K. KOSIOR, R. F Robust dynamic susceptibility contrast mr perfusion using 4d nonlinear noise filters. Journal of Magnetic Resonance Imaging KISHAN, H., AND SEELAMANTULA, C. S Sure-fast bilateral filters. In ICASSP, IEEE, KOPF, J., COHEN, M. F., LISCHINSKI, D., AND UYTTENDAELE, M Joint bilateral upsampling. ACM Trans. Graph. 26, 3 (July). LEVOY, M., AND HANRAHAN, P Light field rendering. In Proceedings of the 23rd annual conference on Computer graphics and interactive techniques, ACM, New York, NY, USA, SIG- GRAPH 96, LEVOY, M Light fields and computational imaging. Computer 39, 8, LIU, X., GAMAL, A. E., HOROWITZ, M. A., AND WANDELL, B. A., Cmos image sensors dynamic range and snr enhancement via statistical signal processing, June. M MENDRIK, EVERT-JAN VONKEN, B. G. H. W. D. J. A. R. T. V. S. E. J. S. M. A. V. M. P Tips bilateral noise reduction in 4d ct perfusion scans produces high-quality cerebral blood flow maps. Physics in Medicine and Biology 56, 13, MITRA, K., AND VEERARAGHAVAN, A Light field denoising, light field superresolution and stereo camera based refocussing using a gmm light field patch prior. In Computer Vision and Pattern Recognition Workshops (CVPRW), 2012 IEEE Computer Society Conference on, NG, R Digital light field photography. PhD thesis, Stanford, CA, USA. AAI

9 Ground truth Nearest neighbor Bilinear Joint Bilateral Figure 16: Comparison of different upsampling techniques for a spatial scale factor of 4. As it can be seen, the Joint Bilateral upsampling produces best visual results.

10 PERWASS, C., AND WIETZKE, L., Single lens 3d-camera with extended depth-of-field. PETSCHNIGG, G., SZELISKI, R., AGRAWALA, M., COHEN, M., HOPPE, H., AND TOYAMA, K Digital photography with flash and no-flash image pairs. In ACM SIGGRAPH 2004 Papers, ACM, New York, NY, USA, SIGGRAPH 04, RIEMENS, A. K., GANGWAL, O. P., BARENBRUG, B., AND BERRETTY, R.-P. M., Multistep joint bilateral depth upsampling. TOMASI, C., AND MANDUCHI, R Bilateral filtering for gray and color images. In Proceedings of the Sixth International Conference on Computer Vision, IEEE Computer Society, Washington, DC, USA, ICCV 98, 839. WANNER, S., AND GOLDLUECKE, B Globally consistent depth labeling of 4D lightfields. WANNER, S., AND GOLDLUECKE, B Spatial and angular variational super-resolution of 4d light fields. WANNER, S., AND GOLDLUECKE, B Variational light field analysis for disparity estimation and super-resolution. IEEE Transactions on Pattern Analysis and Machine Intelligence. ZHANG, L., VADDADI, S., JIN, H., AND NAYAR, S. K Multiple view image denoising. In Computer Vision and Pattern Recognition,

High Quality Image Magnification using Cross-Scale Self-Similarity

High Quality Image Magnification using Cross-Scale Self-Similarity High Quality Image Magnification using Cross-Scale Self-Similarity André Gooßen 1, Arne Ehlers 1, Thomas Pralow 2, Rolf-Rainer Grigat 1 1 Vision Systems, Hamburg University of Technology, D-21079 Hamburg

More information

High Performance GPU-based Preprocessing for Time-of-Flight Imaging in Medical Applications

High Performance GPU-based Preprocessing for Time-of-Flight Imaging in Medical Applications High Performance GPU-based Preprocessing for Time-of-Flight Imaging in Medical Applications Jakob Wasza 1, Sebastian Bauer 1, Joachim Hornegger 1,2 1 Pattern Recognition Lab, Friedrich-Alexander University

More information

Bildverarbeitung und Mustererkennung Image Processing and Pattern Recognition

Bildverarbeitung und Mustererkennung Image Processing and Pattern Recognition Bildverarbeitung und Mustererkennung Image Processing and Pattern Recognition 1. Image Pre-Processing - Pixel Brightness Transformation - Geometric Transformation - Image Denoising 1 1. Image Pre-Processing

More information

WHITE PAPER. Are More Pixels Better? www.basler-ipcam.com. Resolution Does it Really Matter?

WHITE PAPER. Are More Pixels Better? www.basler-ipcam.com. Resolution Does it Really Matter? WHITE PAPER www.basler-ipcam.com Are More Pixels Better? The most frequently asked question when buying a new digital security camera is, What resolution does the camera provide? The resolution is indeed

More information

Color Segmentation Based Depth Image Filtering

Color Segmentation Based Depth Image Filtering Color Segmentation Based Depth Image Filtering Michael Schmeing and Xiaoyi Jiang Department of Computer Science, University of Münster Einsteinstraße 62, 48149 Münster, Germany, {m.schmeing xjiang}@uni-muenster.de

More information

Optical Flow. Shenlong Wang CSC2541 Course Presentation Feb 2, 2016

Optical Flow. Shenlong Wang CSC2541 Course Presentation Feb 2, 2016 Optical Flow Shenlong Wang CSC2541 Course Presentation Feb 2, 2016 Outline Introduction Variation Models Feature Matching Methods End-to-end Learning based Methods Discussion Optical Flow Goal: Pixel motion

More information

High Quality Image Deblurring Panchromatic Pixels

High Quality Image Deblurring Panchromatic Pixels High Quality Image Deblurring Panchromatic Pixels ACM Transaction on Graphics vol. 31, No. 5, 2012 Sen Wang, Tingbo Hou, John Border, Hong Qin, and Rodney Miller Presented by Bong-Seok Choi School of Electrical

More information

Using visible SNR (vsnr) to compare image quality of pixel binning and digital resizing

Using visible SNR (vsnr) to compare image quality of pixel binning and digital resizing Using visible SNR (vsnr) to compare image quality of pixel binning and digital resizing Joyce Farrell a, Mike Okincha b, Manu Parmar ac, and Brian Wandell ac a Dept. of Electrical Engineering, Stanford

More information

Performance Verification of Super-Resolution Image Reconstruction

Performance Verification of Super-Resolution Image Reconstruction Performance Verification of Super-Resolution Image Reconstruction Masaki Sugie Department of Information Science, Kogakuin University Tokyo, Japan Email: [email protected] Seiichi Gohshi Department

More information

A Noise-Aware Filter for Real-Time Depth Upsampling

A Noise-Aware Filter for Real-Time Depth Upsampling A Noise-Aware Filter for Real-Time Depth Upsampling Derek Chan Hylke Buisman Christian Theobalt Sebastian Thrun Stanford University, USA Abstract. A new generation of active 3D range sensors, such as time-of-flight

More information

A Short Introduction to Computer Graphics

A Short Introduction to Computer Graphics A Short Introduction to Computer Graphics Frédo Durand MIT Laboratory for Computer Science 1 Introduction Chapter I: Basics Although computer graphics is a vast field that encompasses almost any graphical

More information

Accurate and robust image superresolution by neural processing of local image representations

Accurate and robust image superresolution by neural processing of local image representations Accurate and robust image superresolution by neural processing of local image representations Carlos Miravet 1,2 and Francisco B. Rodríguez 1 1 Grupo de Neurocomputación Biológica (GNB), Escuela Politécnica

More information

A NEW SUPER RESOLUTION TECHNIQUE FOR RANGE DATA. Valeria Garro, Pietro Zanuttigh, Guido M. Cortelazzo. University of Padova, Italy

A NEW SUPER RESOLUTION TECHNIQUE FOR RANGE DATA. Valeria Garro, Pietro Zanuttigh, Guido M. Cortelazzo. University of Padova, Italy A NEW SUPER RESOLUTION TECHNIQUE FOR RANGE DATA Valeria Garro, Pietro Zanuttigh, Guido M. Cortelazzo University of Padova, Italy ABSTRACT Current Time-of-Flight matrix sensors allow for the acquisition

More information

Theory and Methods of Lightfield Photography SIGGRAPH 2009

Theory and Methods of Lightfield Photography SIGGRAPH 2009 Theory and Methods of Lightfield Photography SIGGRAPH 2009 Todor Georgiev Adobe Systems [email protected] Andrew Lumsdaine Indiana University [email protected] 1 Web Page http://www.tgeorgiev.net/asia2009/

More information

A Novel Method to Improve Resolution of Satellite Images Using DWT and Interpolation

A Novel Method to Improve Resolution of Satellite Images Using DWT and Interpolation A Novel Method to Improve Resolution of Satellite Images Using DWT and Interpolation S.VENKATA RAMANA ¹, S. NARAYANA REDDY ² M.Tech student, Department of ECE, SVU college of Engineering, Tirupati, 517502,

More information

Highlight Removal by Illumination-Constrained Inpainting

Highlight Removal by Illumination-Constrained Inpainting Highlight Removal by Illumination-Constrained Inpainting Ping Tan Stephen Lin Long Quan Heung-Yeung Shum Microsoft Research, Asia Hong Kong University of Science and Technology Abstract We present a single-image

More information

Canny Edge Detection

Canny Edge Detection Canny Edge Detection 09gr820 March 23, 2009 1 Introduction The purpose of edge detection in general is to significantly reduce the amount of data in an image, while preserving the structural properties

More information

Redundant Wavelet Transform Based Image Super Resolution

Redundant Wavelet Transform Based Image Super Resolution Redundant Wavelet Transform Based Image Super Resolution Arti Sharma, Prof. Preety D Swami Department of Electronics &Telecommunication Samrat Ashok Technological Institute Vidisha Department of Electronics

More information

PIXEL-LEVEL IMAGE FUSION USING BROVEY TRANSFORME AND WAVELET TRANSFORM

PIXEL-LEVEL IMAGE FUSION USING BROVEY TRANSFORME AND WAVELET TRANSFORM PIXEL-LEVEL IMAGE FUSION USING BROVEY TRANSFORME AND WAVELET TRANSFORM Rohan Ashok Mandhare 1, Pragati Upadhyay 2,Sudha Gupta 3 ME Student, K.J.SOMIYA College of Engineering, Vidyavihar, Mumbai, Maharashtra,

More information

A Learning Based Method for Super-Resolution of Low Resolution Images

A Learning Based Method for Super-Resolution of Low Resolution Images A Learning Based Method for Super-Resolution of Low Resolution Images Emre Ugur June 1, 2004 [email protected] Abstract The main objective of this project is the study of a learning based method

More information

NEIGHBORHOOD REGRESSION FOR EDGE-PRESERVING IMAGE SUPER-RESOLUTION. Yanghao Li, Jiaying Liu, Wenhan Yang, Zongming Guo

NEIGHBORHOOD REGRESSION FOR EDGE-PRESERVING IMAGE SUPER-RESOLUTION. Yanghao Li, Jiaying Liu, Wenhan Yang, Zongming Guo NEIGHBORHOOD REGRESSION FOR EDGE-PRESERVING IMAGE SUPER-RESOLUTION Yanghao Li, Jiaying Liu, Wenhan Yang, Zongming Guo Institute of Computer Science and Technology, Peking University, Beijing, P.R.China,

More information

REAL TIME TRAFFIC LIGHT CONTROL USING IMAGE PROCESSING

REAL TIME TRAFFIC LIGHT CONTROL USING IMAGE PROCESSING REAL TIME TRAFFIC LIGHT CONTROL USING IMAGE PROCESSING Ms.PALLAVI CHOUDEKAR Ajay Kumar Garg Engineering College, Department of electrical and electronics Ms.SAYANTI BANERJEE Ajay Kumar Garg Engineering

More information

Sachin Patel HOD I.T Department PCST, Indore, India. Parth Bhatt I.T Department, PCST, Indore, India. Ankit Shah CSE Department, KITE, Jaipur, India

Sachin Patel HOD I.T Department PCST, Indore, India. Parth Bhatt I.T Department, PCST, Indore, India. Ankit Shah CSE Department, KITE, Jaipur, India Image Enhancement Using Various Interpolation Methods Parth Bhatt I.T Department, PCST, Indore, India Ankit Shah CSE Department, KITE, Jaipur, India Sachin Patel HOD I.T Department PCST, Indore, India

More information

Lighting Estimation in Indoor Environments from Low-Quality Images

Lighting Estimation in Indoor Environments from Low-Quality Images Lighting Estimation in Indoor Environments from Low-Quality Images Natalia Neverova, Damien Muselet, Alain Trémeau Laboratoire Hubert Curien UMR CNRS 5516, University Jean Monnet, Rue du Professeur Benoît

More information

Super-resolution Reconstruction Algorithm Based on Patch Similarity and Back-projection Modification

Super-resolution Reconstruction Algorithm Based on Patch Similarity and Back-projection Modification 1862 JOURNAL OF SOFTWARE, VOL 9, NO 7, JULY 214 Super-resolution Reconstruction Algorithm Based on Patch Similarity and Back-projection Modification Wei-long Chen Digital Media College, Sichuan Normal

More information

ROBUST COLOR JOINT MULTI-FRAME DEMOSAICING AND SUPER- RESOLUTION ALGORITHM

ROBUST COLOR JOINT MULTI-FRAME DEMOSAICING AND SUPER- RESOLUTION ALGORITHM ROBUST COLOR JOINT MULTI-FRAME DEMOSAICING AND SUPER- RESOLUTION ALGORITHM Theodor Heinze Hasso-Plattner-Institute for Software Systems Engineering Prof.-Dr.-Helmert-Str. 2-3, 14482 Potsdam, Germany [email protected]

More information

Lecture 14. Point Spread Function (PSF)

Lecture 14. Point Spread Function (PSF) Lecture 14 Point Spread Function (PSF), Modulation Transfer Function (MTF), Signal-to-noise Ratio (SNR), Contrast-to-noise Ratio (CNR), and Receiver Operating Curves (ROC) Point Spread Function (PSF) Recollect

More information

Low-resolution Character Recognition by Video-based Super-resolution

Low-resolution Character Recognition by Video-based Super-resolution 2009 10th International Conference on Document Analysis and Recognition Low-resolution Character Recognition by Video-based Super-resolution Ataru Ohkura 1, Daisuke Deguchi 1, Tomokazu Takahashi 2, Ichiro

More information

Assessment of Camera Phone Distortion and Implications for Watermarking

Assessment of Camera Phone Distortion and Implications for Watermarking Assessment of Camera Phone Distortion and Implications for Watermarking Aparna Gurijala, Alastair Reed and Eric Evans Digimarc Corporation, 9405 SW Gemini Drive, Beaverton, OR 97008, USA 1. INTRODUCTION

More information

HSI BASED COLOUR IMAGE EQUALIZATION USING ITERATIVE n th ROOT AND n th POWER

HSI BASED COLOUR IMAGE EQUALIZATION USING ITERATIVE n th ROOT AND n th POWER HSI BASED COLOUR IMAGE EQUALIZATION USING ITERATIVE n th ROOT AND n th POWER Gholamreza Anbarjafari icv Group, IMS Lab, Institute of Technology, University of Tartu, Tartu 50411, Estonia [email protected]

More information

Computed Tomography Resolution Enhancement by Integrating High-Resolution 2D X-Ray Images into the CT reconstruction

Computed Tomography Resolution Enhancement by Integrating High-Resolution 2D X-Ray Images into the CT reconstruction Digital Industrial Radiology and Computed Tomography (DIR 2015) 22-25 June 2015, Belgium, Ghent - www.ndt.net/app.dir2015 More Info at Open Access Database www.ndt.net/?id=18046 Computed Tomography Resolution

More information

Automatic Labeling of Lane Markings for Autonomous Vehicles

Automatic Labeling of Lane Markings for Autonomous Vehicles Automatic Labeling of Lane Markings for Autonomous Vehicles Jeffrey Kiske Stanford University 450 Serra Mall, Stanford, CA 94305 [email protected] 1. Introduction As autonomous vehicles become more popular,

More information

How To Fix Out Of Focus And Blur Images With A Dynamic Template Matching Algorithm

How To Fix Out Of Focus And Blur Images With A Dynamic Template Matching Algorithm IJSTE - International Journal of Science Technology & Engineering Volume 1 Issue 10 April 2015 ISSN (online): 2349-784X Image Estimation Algorithm for Out of Focus and Blur Images to Retrieve the Barcode

More information

Multivariate data visualization using shadow

Multivariate data visualization using shadow Proceedings of the IIEEJ Ima and Visual Computing Wor Kuching, Malaysia, Novembe Multivariate data visualization using shadow Zhongxiang ZHENG Suguru SAITO Tokyo Institute of Technology ABSTRACT When visualizing

More information

PERFORMANCE ANALYSIS OF HIGH RESOLUTION IMAGES USING INTERPOLATION TECHNIQUES IN MULTIMEDIA COMMUNICATION SYSTEM

PERFORMANCE ANALYSIS OF HIGH RESOLUTION IMAGES USING INTERPOLATION TECHNIQUES IN MULTIMEDIA COMMUNICATION SYSTEM PERFORMANCE ANALYSIS OF HIGH RESOLUTION IMAGES USING INTERPOLATION TECHNIQUES IN MULTIMEDIA COMMUNICATION SYSTEM Apurva Sinha 1, Mukesh kumar 2, A.K. Jaiswal 3, Rohini Saxena 4 Department of Electronics

More information

Common Core Unit Summary Grades 6 to 8

Common Core Unit Summary Grades 6 to 8 Common Core Unit Summary Grades 6 to 8 Grade 8: Unit 1: Congruence and Similarity- 8G1-8G5 rotations reflections and translations,( RRT=congruence) understand congruence of 2 d figures after RRT Dilations

More information

Face Model Fitting on Low Resolution Images

Face Model Fitting on Low Resolution Images Face Model Fitting on Low Resolution Images Xiaoming Liu Peter H. Tu Frederick W. Wheeler Visualization and Computer Vision Lab General Electric Global Research Center Niskayuna, NY, 1239, USA {liux,tu,wheeler}@research.ge.com

More information

Low-resolution Image Processing based on FPGA

Low-resolution Image Processing based on FPGA Abstract Research Journal of Recent Sciences ISSN 2277-2502. Low-resolution Image Processing based on FPGA Mahshid Aghania Kiau, Islamic Azad university of Karaj, IRAN Available online at: www.isca.in,

More information

PCL - SURFACE RECONSTRUCTION

PCL - SURFACE RECONSTRUCTION PCL - SURFACE RECONSTRUCTION TOYOTA CODE SPRINT Alexandru-Eugen Ichim Computer Graphics and Geometry Laboratory PROBLEM DESCRIPTION 1/2 3D revolution due to cheap RGB-D cameras (Asus Xtion & Microsoft

More information

Fast Image Labeling for Creating High-Resolution Panoramic Images on Mobile Devices

Fast Image Labeling for Creating High-Resolution Panoramic Images on Mobile Devices 2009 11th IEEE International Symposium on Multimedia Fast Image Labeling for Creating High-Resolution Panoramic Images on Mobile Devices Yingen Xiong and Kari Pulli Nokia Research Center Palo Alto, CA

More information

CHAPTER 3: DIGITAL IMAGING IN DIAGNOSTIC RADIOLOGY. 3.1 Basic Concepts of Digital Imaging

CHAPTER 3: DIGITAL IMAGING IN DIAGNOSTIC RADIOLOGY. 3.1 Basic Concepts of Digital Imaging Physics of Medical X-Ray Imaging (1) Chapter 3 CHAPTER 3: DIGITAL IMAGING IN DIAGNOSTIC RADIOLOGY 3.1 Basic Concepts of Digital Imaging Unlike conventional radiography that generates images on film through

More information

High-resolution Imaging System for Omnidirectional Illuminant Estimation

High-resolution Imaging System for Omnidirectional Illuminant Estimation High-resolution Imaging System for Omnidirectional Illuminant Estimation Shoji Tominaga*, Tsuyoshi Fukuda**, and Akira Kimachi** *Graduate School of Advanced Integration Science, Chiba University, Chiba

More information

Comparison of Non-linear Dimensionality Reduction Techniques for Classification with Gene Expression Microarray Data

Comparison of Non-linear Dimensionality Reduction Techniques for Classification with Gene Expression Microarray Data CMPE 59H Comparison of Non-linear Dimensionality Reduction Techniques for Classification with Gene Expression Microarray Data Term Project Report Fatma Güney, Kübra Kalkan 1/15/2013 Keywords: Non-linear

More information

Numerical Methods For Image Restoration

Numerical Methods For Image Restoration Numerical Methods For Image Restoration CIRAM Alessandro Lanza University of Bologna, Italy Faculty of Engineering CIRAM Outline 1. Image Restoration as an inverse problem 2. Image degradation models:

More information

Vision based Vehicle Tracking using a high angle camera

Vision based Vehicle Tracking using a high angle camera Vision based Vehicle Tracking using a high angle camera Raúl Ignacio Ramos García Dule Shu [email protected] [email protected] Abstract A vehicle tracking and grouping algorithm is presented in this work

More information

Assessment. Presenter: Yupu Zhang, Guoliang Jin, Tuo Wang Computer Vision 2008 Fall

Assessment. Presenter: Yupu Zhang, Guoliang Jin, Tuo Wang Computer Vision 2008 Fall Automatic Photo Quality Assessment Presenter: Yupu Zhang, Guoliang Jin, Tuo Wang Computer Vision 2008 Fall Estimating i the photorealism of images: Distinguishing i i paintings from photographs h Florin

More information

Video Camera Image Quality in Physical Electronic Security Systems

Video Camera Image Quality in Physical Electronic Security Systems Video Camera Image Quality in Physical Electronic Security Systems Video Camera Image Quality in Physical Electronic Security Systems In the second decade of the 21st century, annual revenue for the global

More information

Feature Tracking and Optical Flow

Feature Tracking and Optical Flow 02/09/12 Feature Tracking and Optical Flow Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem Many slides adapted from Lana Lazebnik, Silvio Saverse, who in turn adapted slides from Steve

More information

Advanced Computer Graphics. Rendering Equation. Matthias Teschner. Computer Science Department University of Freiburg

Advanced Computer Graphics. Rendering Equation. Matthias Teschner. Computer Science Department University of Freiburg Advanced Computer Graphics Rendering Equation Matthias Teschner Computer Science Department University of Freiburg Outline rendering equation Monte Carlo integration sampling of random variables University

More information

Analecta Vol. 8, No. 2 ISSN 2064-7964

Analecta Vol. 8, No. 2 ISSN 2064-7964 EXPERIMENTAL APPLICATIONS OF ARTIFICIAL NEURAL NETWORKS IN ENGINEERING PROCESSING SYSTEM S. Dadvandipour Institute of Information Engineering, University of Miskolc, Egyetemváros, 3515, Miskolc, Hungary,

More information

To determine vertical angular frequency, we need to express vertical viewing angle in terms of and. 2tan. (degree). (1 pt)

To determine vertical angular frequency, we need to express vertical viewing angle in terms of and. 2tan. (degree). (1 pt) Polytechnic University, Dept. Electrical and Computer Engineering EL6123 --- Video Processing, S12 (Prof. Yao Wang) Solution to Midterm Exam Closed Book, 1 sheet of notes (double sided) allowed 1. (5 pt)

More information

SUPER-RESOLUTION (SR) has been an active research

SUPER-RESOLUTION (SR) has been an active research 498 IEEE TRANSACTIONS ON MULTIMEDIA, VOL. 15, NO. 3, APRIL 2013 A Self-Learning Approach to Single Image Super-Resolution Min-Chun Yang and Yu-Chiang Frank Wang, Member, IEEE Abstract Learning-based approaches

More information

Superresolution images reconstructed from aliased images

Superresolution images reconstructed from aliased images Superresolution images reconstructed from aliased images Patrick Vandewalle, Sabine Süsstrunk and Martin Vetterli LCAV - School of Computer and Communication Sciences Ecole Polytechnique Fédérale de Lausanne

More information

Palmprint Recognition. By Sree Rama Murthy kora Praveen Verma Yashwant Kashyap

Palmprint Recognition. By Sree Rama Murthy kora Praveen Verma Yashwant Kashyap Palmprint Recognition By Sree Rama Murthy kora Praveen Verma Yashwant Kashyap Palm print Palm Patterns are utilized in many applications: 1. To correlate palm patterns with medical disorders, e.g. genetic

More information

Relating Vanishing Points to Catadioptric Camera Calibration

Relating Vanishing Points to Catadioptric Camera Calibration Relating Vanishing Points to Catadioptric Camera Calibration Wenting Duan* a, Hui Zhang b, Nigel M. Allinson a a Laboratory of Vision Engineering, University of Lincoln, Brayford Pool, Lincoln, U.K. LN6

More information

Image Compression through DCT and Huffman Coding Technique

Image Compression through DCT and Huffman Coding Technique International Journal of Current Engineering and Technology E-ISSN 2277 4106, P-ISSN 2347 5161 2015 INPRESSCO, All Rights Reserved Available at http://inpressco.com/category/ijcet Research Article Rahul

More information

Image Segmentation and Registration

Image Segmentation and Registration Image Segmentation and Registration Dr. Christine Tanner ([email protected]) Computer Vision Laboratory, ETH Zürich Dr. Verena Kaynig, Machine Learning Laboratory, ETH Zürich Outline Segmentation

More information

Study of the Human Eye Working Principle: An impressive high angular resolution system with simple array detectors

Study of the Human Eye Working Principle: An impressive high angular resolution system with simple array detectors Study of the Human Eye Working Principle: An impressive high angular resolution system with simple array detectors Diego Betancourt and Carlos del Río Antenna Group, Public University of Navarra, Campus

More information

Announcements. Active stereo with structured light. Project structured light patterns onto the object

Announcements. Active stereo with structured light. Project structured light patterns onto the object Announcements Active stereo with structured light Project 3 extension: Wednesday at noon Final project proposal extension: Friday at noon > consult with Steve, Rick, and/or Ian now! Project 2 artifact

More information

Segmentation of building models from dense 3D point-clouds

Segmentation of building models from dense 3D point-clouds Segmentation of building models from dense 3D point-clouds Joachim Bauer, Konrad Karner, Konrad Schindler, Andreas Klaus, Christopher Zach VRVis Research Center for Virtual Reality and Visualization, Institute

More information

DYNAMIC RANGE IMPROVEMENT THROUGH MULTIPLE EXPOSURES. Mark A. Robertson, Sean Borman, and Robert L. Stevenson

DYNAMIC RANGE IMPROVEMENT THROUGH MULTIPLE EXPOSURES. Mark A. Robertson, Sean Borman, and Robert L. Stevenson c 1999 IEEE. Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or

More information

Video stabilization for high resolution images reconstruction

Video stabilization for high resolution images reconstruction Advanced Project S9 Video stabilization for high resolution images reconstruction HIMMICH Youssef, KEROUANTON Thomas, PATIES Rémi, VILCHES José. Abstract Super-resolution reconstruction produces one or

More information

Mean-Shift Tracking with Random Sampling

Mean-Shift Tracking with Random Sampling 1 Mean-Shift Tracking with Random Sampling Alex Po Leung, Shaogang Gong Department of Computer Science Queen Mary, University of London, London, E1 4NS Abstract In this work, boosting the efficiency of

More information

Template-based Eye and Mouth Detection for 3D Video Conferencing

Template-based Eye and Mouth Detection for 3D Video Conferencing Template-based Eye and Mouth Detection for 3D Video Conferencing Jürgen Rurainsky and Peter Eisert Fraunhofer Institute for Telecommunications - Heinrich-Hertz-Institute, Image Processing Department, Einsteinufer

More information

Medical Image Processing on the GPU. Past, Present and Future. Anders Eklund, PhD Virginia Tech Carilion Research Institute [email protected].

Medical Image Processing on the GPU. Past, Present and Future. Anders Eklund, PhD Virginia Tech Carilion Research Institute andek@vtc.vt. Medical Image Processing on the GPU Past, Present and Future Anders Eklund, PhD Virginia Tech Carilion Research Institute [email protected] Outline Motivation why do we need GPUs? Past - how was GPU programming

More information

Files Used in this Tutorial

Files Used in this Tutorial Generate Point Clouds Tutorial This tutorial shows how to generate point clouds from IKONOS satellite stereo imagery. You will view the point clouds in the ENVI LiDAR Viewer. The estimated time to complete

More information

Colorado School of Mines Computer Vision Professor William Hoff

Colorado School of Mines Computer Vision Professor William Hoff Professor William Hoff Dept of Electrical Engineering &Computer Science http://inside.mines.edu/~whoff/ 1 Introduction to 2 What is? A process that produces from images of the external world a description

More information

product overview pco.edge family the most versatile scmos camera portfolio on the market pioneer in scmos image sensor technology

product overview pco.edge family the most versatile scmos camera portfolio on the market pioneer in scmos image sensor technology product overview family the most versatile scmos camera portfolio on the market pioneer in scmos image sensor technology scmos knowledge base scmos General Information PCO scmos cameras are a breakthrough

More information

For example, estimate the population of the United States as 3 times 10⁸ and the

For example, estimate the population of the United States as 3 times 10⁸ and the CCSS: Mathematics The Number System CCSS: Grade 8 8.NS.A. Know that there are numbers that are not rational, and approximate them by rational numbers. 8.NS.A.1. Understand informally that every number

More information

VEHICLE TRACKING USING ACOUSTIC AND VIDEO SENSORS

VEHICLE TRACKING USING ACOUSTIC AND VIDEO SENSORS VEHICLE TRACKING USING ACOUSTIC AND VIDEO SENSORS Aswin C Sankaranayanan, Qinfen Zheng, Rama Chellappa University of Maryland College Park, MD - 277 {aswch, qinfen, rama}@cfar.umd.edu Volkan Cevher, James

More information

Algorithms for the resizing of binary and grayscale images using a logical transform

Algorithms for the resizing of binary and grayscale images using a logical transform Algorithms for the resizing of binary and grayscale images using a logical transform Ethan E. Danahy* a, Sos S. Agaian b, Karen A. Panetta a a Dept. of Electrical and Computer Eng., Tufts University, 161

More information

Methodology for Emulating Self Organizing Maps for Visualization of Large Datasets

Methodology for Emulating Self Organizing Maps for Visualization of Large Datasets Methodology for Emulating Self Organizing Maps for Visualization of Large Datasets Macario O. Cordel II and Arnulfo P. Azcarraga College of Computer Studies *Corresponding Author: [email protected]

More information

SSIM Technique for Comparison of Images

SSIM Technique for Comparison of Images SSIM Technique for Comparison of Images Anil Wadhokar 1, Krupanshu Sakharikar 2, Sunil Wadhokar 3, Geeta Salunke 4 P.G. Student, Department of E&TC, GSMCOE Engineering College, Pune, Maharashtra, India

More information

Single Depth Image Super Resolution and Denoising Using Coupled Dictionary Learning with Local Constraints and Shock Filtering

Single Depth Image Super Resolution and Denoising Using Coupled Dictionary Learning with Local Constraints and Shock Filtering Single Depth Image Super Resolution and Denoising Using Coupled Dictionary Learning with Local Constraints and Shock Filtering Jun Xie 1, Cheng-Chuan Chou 2, Rogerio Feris 3, Ming-Ting Sun 1 1 University

More information

The Role of Size Normalization on the Recognition Rate of Handwritten Numerals

The Role of Size Normalization on the Recognition Rate of Handwritten Numerals The Role of Size Normalization on the Recognition Rate of Handwritten Numerals Chun Lei He, Ping Zhang, Jianxiong Dong, Ching Y. Suen, Tien D. Bui Centre for Pattern Recognition and Machine Intelligence,

More information

Visualization by Linear Projections as Information Retrieval

Visualization by Linear Projections as Information Retrieval Visualization by Linear Projections as Information Retrieval Jaakko Peltonen Helsinki University of Technology, Department of Information and Computer Science, P. O. Box 5400, FI-0015 TKK, Finland [email protected]

More information

Fast Digital Image Inpainting

Fast Digital Image Inpainting Appeared in the Proceedings of the International Conference on Visualization, Imaging and Image Processing (VIIP 2001), Marbella, Spain. September 3-5, 2001 Fast Digital Image Inpainting Manuel M. Oliveira

More information

Adaptive Coded Aperture Photography

Adaptive Coded Aperture Photography Adaptive Coded Aperture Photography Oliver Bimber, Haroon Qureshi, Daniel Danch Institute of Johannes Kepler University, Linz Anselm Grundhoefer Disney Research Zurich Max Grosse Bauhaus University Weimar

More information

Automatic Traffic Estimation Using Image Processing

Automatic Traffic Estimation Using Image Processing Automatic Traffic Estimation Using Image Processing Pejman Niksaz Science &Research Branch, Azad University of Yazd, Iran [email protected] Abstract As we know the population of city and number of

More information

NEW MEXICO Grade 6 MATHEMATICS STANDARDS

NEW MEXICO Grade 6 MATHEMATICS STANDARDS PROCESS STANDARDS To help New Mexico students achieve the Content Standards enumerated below, teachers are encouraged to base instruction on the following Process Standards: Problem Solving Build new mathematical

More information

Blind Deconvolution of Barcodes via Dictionary Analysis and Wiener Filter of Barcode Subsections

Blind Deconvolution of Barcodes via Dictionary Analysis and Wiener Filter of Barcode Subsections Blind Deconvolution of Barcodes via Dictionary Analysis and Wiener Filter of Barcode Subsections Maximilian Hung, Bohyun B. Kim, Xiling Zhang August 17, 2013 Abstract While current systems already provide

More information

A BRIEF STUDY OF VARIOUS NOISE MODEL AND FILTERING TECHNIQUES

A BRIEF STUDY OF VARIOUS NOISE MODEL AND FILTERING TECHNIQUES Volume 4, No. 4, April 2013 Journal of Global Research in Computer Science REVIEW ARTICLE Available Online at www.jgrcs.info A BRIEF STUDY OF VARIOUS NOISE MODEL AND FILTERING TECHNIQUES Priyanka Kamboj

More information

with functions, expressions and equations which follow in units 3 and 4.

with functions, expressions and equations which follow in units 3 and 4. Grade 8 Overview View unit yearlong overview here The unit design was created in line with the areas of focus for grade 8 Mathematics as identified by the Common Core State Standards and the PARCC Model

More information

Performance Level Descriptors Grade 6 Mathematics

Performance Level Descriptors Grade 6 Mathematics Performance Level Descriptors Grade 6 Mathematics Multiplying and Dividing with Fractions 6.NS.1-2 Grade 6 Math : Sub-Claim A The student solves problems involving the Major Content for grade/course with

More information

Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches

Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches PhD Thesis by Payam Birjandi Director: Prof. Mihai Datcu Problematic

More information

So which is the best?

So which is the best? Manifold Learning Techniques: So which is the best? Todd Wittman Math 8600: Geometric Data Analysis Instructor: Gilad Lerman Spring 2005 Note: This presentation does not contain information on LTSA, which

More information

The Scientific Data Mining Process

The Scientific Data Mining Process Chapter 4 The Scientific Data Mining Process When I use a word, Humpty Dumpty said, in rather a scornful tone, it means just what I choose it to mean neither more nor less. Lewis Carroll [87, p. 214] In

More information

Computational Optical Imaging - Optique Numerique. -- Deconvolution --

Computational Optical Imaging - Optique Numerique. -- Deconvolution -- Computational Optical Imaging - Optique Numerique -- Deconvolution -- Winter 2014 Ivo Ihrke Deconvolution Ivo Ihrke Outline Deconvolution Theory example 1D deconvolution Fourier method Algebraic method

More information

Simultaneous Gamma Correction and Registration in the Frequency Domain

Simultaneous Gamma Correction and Registration in the Frequency Domain Simultaneous Gamma Correction and Registration in the Frequency Domain Alexander Wong [email protected] William Bishop [email protected] Department of Electrical and Computer Engineering University

More information

Image Compression and Decompression using Adaptive Interpolation

Image Compression and Decompression using Adaptive Interpolation Image Compression and Decompression using Adaptive Interpolation SUNILBHOOSHAN 1,SHIPRASHARMA 2 Jaypee University of Information Technology 1 Electronicsand Communication EngineeringDepartment 2 ComputerScience

More information

MATLAB-based Applications for Image Processing and Image Quality Assessment Part II: Experimental Results

MATLAB-based Applications for Image Processing and Image Quality Assessment Part II: Experimental Results 154 L. KRASULA, M. KLÍMA, E. ROGARD, E. JEANBLANC, MATLAB BASED APPLICATIONS PART II: EXPERIMENTAL RESULTS MATLAB-based Applications for Image Processing and Image Quality Assessment Part II: Experimental

More information

Computer Graphics. Geometric Modeling. Page 1. Copyright Gotsman, Elber, Barequet, Karni, Sheffer Computer Science - Technion. An Example.

Computer Graphics. Geometric Modeling. Page 1. Copyright Gotsman, Elber, Barequet, Karni, Sheffer Computer Science - Technion. An Example. An Example 2 3 4 Outline Objective: Develop methods and algorithms to mathematically model shape of real world objects Categories: Wire-Frame Representation Object is represented as as a set of points

More information