Hole-Filling Method Using Depth Based In-Painting For View Synthesis in Free Viewpoint Television (FTV) and 3D Video



Similar documents
Adaptive Block Truncation Filter for MVC Depth Image Enhancement

A Novel Hole filling method based on Projection onto Convex Set in DIBR

Quality improving techniques for free-viewpoint DIBR

The Open University s repository of research publications and other research outputs

REPRESENTATION, CODING AND INTERACTIVE RENDERING OF HIGH- RESOLUTION PANORAMIC IMAGES AND VIDEO USING MPEG-4

Key Terms 2D-to-3D-conversion, depth image-based rendering, hole-filling, image warping, optimization.

Adaptive Image Warping Using Disparity Map-Based Rendering For Hole Prevention In 3D View Synthesis

Immersive Medien und 3D-Video

Study and Implementation of Video Compression Standards (H.264/AVC and Dirac)

Study and Implementation of Video Compression standards (H.264/AVC, Dirac)

Performance Analysis and Comparison of JM 15.1 and Intel IPP H.264 Encoder and Decoder

Color Segmentation Based Depth Image Filtering

Efficient Coding Unit and Prediction Unit Decision Algorithm for Multiview Video Coding

Parametric Comparison of H.264 with Existing Video Standards

Image Processing and Computer Graphics. Rendering Pipeline. Matthias Teschner. Computer Science Department University of Freiburg

Service-Oriented Visualization of Virtual 3D City Models

View Sequence Coding using Warping-based Image Alignment for Multi-view Video

Automatic Detection of PCB Defects

SSIM Technique for Comparison of Images

Introduction to Computer Graphics

MATLAB-based Applications for Image Processing and Image Quality Assessment Part II: Experimental Results

Video compression: Performance of available codec software

Text Localization & Segmentation in Images, Web Pages and Videos Media Mining I

Template-based Eye and Mouth Detection for 3D Video Conferencing

A Survey of Video Processing with Field Programmable Gate Arrays (FGPA)

A method of generating free-route walk-through animation using vehicle-borne video image

SYNTHESIZING FREE-VIEWPOINT IMAGES FROM MULTIPLE VIEW VIDEOS IN SOCCER STADIUM

COURSE GUIDE 3D Modelling and Animation Master's Degree in Digital Media Catholic University of Valencia

Silhouette completion in DIBR with the help of skeletal tracking

Consolidated Visualization of Enormous 3D Scan Point Clouds with Scanopy

A Short Introduction to Computer Graphics

A 3D-TV Approach Using Depth-Image-Based Rendering (DIBR)

Kapitel 12. 3D Television Based on a Stereoscopic View Synthesis Approach

CREATE A 3D MOVIE IN DIRECTOR

Real-Time Depth-Image-Based Rendering for 3DTV Using OpenCL

Vision based Vehicle Tracking using a high angle camera

Johannes Sametinger. C. Doppler Laboratory for Software Engineering Johannes Kepler University of Linz A-4040 Linz, Austria

A Prototype For Eye-Gaze Corrected

Detection and Restoration of Vertical Non-linear Scratches in Digitized Film Sequences

Performance Verification of Super-Resolution Image Reconstruction

A Remote Maintenance System with the use of Virtual Reality.

HSI BASED COLOUR IMAGE EQUALIZATION USING ITERATIVE n th ROOT AND n th POWER


Loss-resilient Coding of Texture and Depth for Free-viewpoint Video Conferencing

CUBE-MAP DATA STRUCTURE FOR INTERACTIVE GLOBAL ILLUMINATION COMPUTATION IN DYNAMIC DIFFUSE ENVIRONMENTS

White paper. HDTV (High Definition Television) and video surveillance

AUTOMATIC MULTI-VIEW TEXTURE MAPPING OF 3D SURFACE PROJECTIONS

PRODUCING DV VIDEO WITH PREMIERE & QUICKTIME

PHYSIOLOGICALLY-BASED DETECTION OF COMPUTER GENERATED FACES IN VIDEO

MeshLab and Arc3D: Photo-Reconstruction and Processing of 3D meshes

Image Processing Based Automatic Visual Inspection System for PCBs

Using Photorealistic RenderMan for High-Quality Direct Volume Rendering

Requirements Analysis Concepts & Principles. Instructor: Dr. Jerry Gao

Efficient Prediction Structures for Multi-view Video Coding

Skills Inventory: Art/Media Communications. 1. Pre-employment Training/Career Development. A. Formal; e.g., certificates. Date Description Location

How To Make A Texture Map Work Better On A Computer Graphics Card (Or Mac)

WHICH HD ENDOSCOPIC CAMERA?

An automatic system for sports analytics in multi-camera tennis videos

A Method of Caption Detection in News Video

IMPACT OF COMPRESSION ON THE VIDEO QUALITY

Immersive 3-D Video Conferencing: Challenges, Concepts, and Implementations

Bandwidth Adaptation for MPEG-4 Video Streaming over the Internet

Visibility optimization for data visualization: A Survey of Issues and Techniques

REMOTE RENDERING OF COMPUTER GAMES

Character Animation from 2D Pictures and 3D Motion Data ALEXANDER HORNUNG, ELLEN DEKKERS, and LEIF KOBBELT RWTH-Aachen University

Wavelet-Based Printed Circuit Board Inspection System

Very Low Frame-Rate Video Streaming For Face-to-Face Teleconference

Low-resolution Character Recognition by Video-based Super-resolution

A NEW SUPER RESOLUTION TECHNIQUE FOR RANGE DATA. Valeria Garro, Pietro Zanuttigh, Guido M. Cortelazzo. University of Padova, Italy

Video Coding Technologies and Standards: Now and Beyond

EE3414 Multimedia Communication Systems Part I

A Study on SURF Algorithm and Real-Time Tracking Objects Using Optical Flow

How To Improve Performance Of H.264/Avc With High Efficiency Video Coding (Hevc)

Subdivision Mapping: Making the Pieces Fit David P. Thaler, GIS Analyst A geographit White Paper, June 2009

High Quality Image Magnification using Cross-Scale Self-Similarity

PERFORMANCE ANALYSIS OF HIGH RESOLUTION IMAGES USING INTERPOLATION TECHNIQUES IN MULTIMEDIA COMMUNICATION SYSTEM

JPEG compression of monochrome 2D-barcode images using DCT coefficient distributions

Proposal for a Virtual 3D World Map

VIRTUAL VIDEO CONFERENCING USING 3D MODEL-ASSISTED IMAGE-BASED RENDERING

Internet of Things (IoT): A vision, architectural elements, and future directions

Example Chapter 08-Number 09: This example demonstrates some simple uses of common canned effects found in popular photo editors to stylize photos.

Interactive Visualization of Magnetic Fields

An Automatic Tool for Checking Consistency between Data Flow Diagrams (DFDs)

Super-resolution method based on edge feature for high resolution imaging

One-Way Pseudo Transparent Display

Efficient View-Dependent Image-Based Rendering with Projective Texture-Mapping

BUILDING TELEPRESENCE SYSTEMS: Translating Science Fiction Ideas into Reality

Transcription:

MISUBISHI ELECRIC RESEARCH LABORAORIES http://www.merl.com Hole-Filling Method Using Depth Based In-Painting For View Synthesis in Free Viewpoint elevision (FV) and 3D Video Kwan-Jung Oh, Sehoon Yea, Yo-Sung Ho R2009-025 June 2009 Abstract Depth image-based rendering (DIBR) is generally used to synthesize a virtual view in free viewpoint television (FV) and 3D video. One of the key techniques in DIBR is how to fill the holes caused by disocclusion regions and wrong depth values. In this paper, we propose a new hole-filling method using depth-based in-painting. he proposed method is designed by combining the depth-based-hole-filling and in-painting. From the experiments, we confirm that the proposed hole-filling method provides better rendering quality objectively and subjectively. Picture Coding Symposium 2009 his work may not be copied or reproduced in whole or in part for any commercial purpose. Permission to copy in whole or in part without payment of fee is granted for nonprofit educational and research purposes provided that all such whole or partial copies include the following: a notice that such copying is by permission of Mitsubishi Electric Research Laboratories, Inc.; an acknowledgment of the authors and individual contributions to the work; and all applicable portions of the copyright notice. Copying, reproduction, or republishing for any other purpose shall require a license with payment of fee to Mitsubishi Electric Research Laboratories, Inc. All rights reserved. Copyright c Mitsubishi Electric Research Laboratories, Inc., 2009 201 Broadway, Cambridge, Massachusetts 02139

MERLCoverPageSide2

HOLE-FILLING MEHOD USING DEPH BASED IN-PAINING FOR VIEW SYNHESIS IN FREE VIEWPOIN ELEVISION (FV) AND 3D VIDEO Kwan-Jung Oh, Sehoon Yea*, and Yo-Sung Ho School of Information and Mechatronics Gwangju Institute of Science and echnology (GIS) 261 Cheomdan-gwagiro, Buk-gu, Gwangju 500-712 Republic of Korea Mitsubishi Electric Research Laboratories (MERL)* 201 Broadway, Cambridge, MA 02139, USA* {kjoh81, hoyo}@gist.ac.kr, yea@merl.com* ABSRAC Depth image-based rendering (DIBR) is generally used to synthesize a virtual view in free viewpoint television (FV) and 3D video. One of the key techniques in DIBR is how to fill the holes caused by disocclusion regions and wrong depth values. In this paper, we propose a new hole-filling method using depth-based in-painting. he proposed method is designed by combining the depth-based holefilling and the in-painting. From the experiments, we confirm that the proposed hole-filling method provides better rendering quality objectively and subjectively. Index erms DIBR, View Synthesis, FV, 3D Video, Hole-filling, In-painting 1. INRODUCION Free viewpoint television (FV) and 3D video provide both realistic 3D impression and free view navigation of the scene to users and are now considered key technologies that could spur a next wave of multimedia experiences such as 3D cinema, 3D broadcasting, 3D displays, and 3D mobile services etc [1]-[3]. Among the key technical building blocks of the whole processing chain that constitutes such systems are coding and rendering. he role of efficient coding becomes much more important in these 3D systems due to the drastic potential increase in the volume of data. Some of the past researches and standardization efforts to address this issue include MPEG-2 multiview video profile (MVP) [4], MPEG-4 multiple auxiliary component (MAC) [5], and MPEG/JV multiview video coding (MVC) [6]. On the other hand, given the ever increasing diversity of forms of 3D services and displays, proper rendering of 3D views is indispensable that could meet the requirements that enable the key features of the application at hand. In other words, it becomes necessary to resample the views and resize each view depending on the number of views and resolutions required by the display, respectively. he case when there are more views to be rendered at the display than are actually coded such as FV, the resampling means generation of virtual views based upon the actual views. he problem of generating an arbitrary view of a 3D scene has been heavily addressed in the area of computer graphics. Among the techniques for rendering, image-based rendering (IBR) techniques [7] have drawn much attention lately for rendering real world scenes. hese techniques use image rather than geometry as primitives for rendering virtual views and often are classified into three categories depending on how much geometric information is used: rendering without geometry, with explicit geometry, and with implicit geometry. he 3D warping and layered-depthimages (LDI) belongs to the second category whereas view morphing and view interpolation belong to the third category. Recently, the moving picture experts group (MPEG) has initiated a new exploration experiment (EE) specifically targeted towards FV applications. Whereas the previous MPEG/JV standardization activities for MVC focused on improvement of compression efficiency for generic multiview coding scenarios, this new EE activities currently focus on depth estimation and rendering and reference softwares [8] are released. he reference software for rendering exploits an inpainting technique to fill the holes occurred during view synthesis. However, since the existing hole-filling technique does not consider the disocclusion property, it wrongly fills the holes located on the border between a foreground and a background from both a foreground and a background neighboring pixels even though it naturally fills the other holes. o complement this defect, we propose a new holefilling method using depth-based in-painting. he proposed hole-filling technique fills the holes only from neighborhood located on background by distinguishing the foreground and background using depth value.

2. OVERVIEW OF VIEW SYNHESIS he schematic diagram of a typical depth-based view synthesis system is shown in Fig. 1. he goal of such a system is to synthesize a virtual view from its neighboring views using the camera parameters, texture images, and depth images. normalized to (l/n, m/n, 1) and then represented as an integer-coordinate (U, V) in the virtual view. o avoid the errors like black-contours appeared in the 3D warped image as shown in Fig. 3 (a), the depth image is firstly warped and then median filtered to remove the blockcontour errors. Finally, the texture image is synthesized by calling the matched texture pixels from the reference view as shown in Fig. 3 (c). (a) (b) (c) Fig. 3. 3D warping without erroneous blanks: (a) 3D warped depth image, (b) median filtered depth, (c) 3D warped texture image Fig. 1. Depth image-based virtual view synthesis he three-dimensional image warping (3D warping) is a key technique in depth-based view synthesis. In 3D warping, pixels in reference image are back-projected to 3D spaces, and re-projected onto the target viewpoint as shown in Fig. 2. he boundary trace-errors around the big holes are generally caused by inaccuracy of the camera parameters and inaccurate boundary matching between texture images and depth images. o remove these visible errors we extend the holes boundaries by using image dilation as shown in Fig. 4. hese extended holes can be filled by the other 3D warped view and we can expect more natural synthesized view by removing this kind of errors. (x, y, z) d u,v (u, v) (U, V) (a) (b) Fig. 4. Hole extension: (a) before extension (b) after extension Reference view Fig. 2. General concept of 3D warping Virtual view Eq. (1) and (2) represent the back-projection and the re-projection processes, respectively. ( t (1) 1 x, y, z) = Rref Aref ( u, v,1) du, v + 1 ( l, m, n) = A R {( x, y, z) t } ver ver where A, R, and t are camera parameters and d represents the depth value of a point in the 3D space. he coordinate (u, v) located on the reference view is 3D warped to (U, V) on the virtual view. he coordinate (l, m, n) in (2) is ver ref (2) he next step is view blending to combine 3D warped views with the virtual view and the simplest way would be taking a weighted sum of two images as below: I V ( u, v) α I ( u, v) + (1 α) I ( u, v) = (3) L where I L and I R are the 3D warped reference texture images and I V is an image to be blended. Generally, the weighting factor α is calculated based on the baseline distance as in (4). t t V L α = (4) tv tl + tv tr where t is a translation vector for each view. he final step is the hole-filling for the remaining holes. R

3. PROPOSED HOLE-FILLING MEHOD Even though the view blending efficiently fills up most disocclusion regions, some holes still remain. In general, these kinds of remaining holes are caused by still remaining disocclusion regions and wrong depth values. Disocclusion regions are defined as areas that cannot be seen in the reference image, but exist in the synthesized one. Most of the existing hole-filling methods use image interpolation or in-painting techniques and fill up the remaining holes using neighboring pixels solely based upon geometrical distance. However, observe that it makes more sense to fill up the holes using the background pixels rather than the foreground ones as the disoccluded area usually belongs to the background by definition. herefore, we propose a depth based hole-filling algorithm which prefers the background pixels over the foreground ones in addition to considering the existing in-painting technique. he general in-painting problem is as follow [9]: the region to be in-painted Ω and its boundary Ω are defined and the pixel p belong to Ω would be in-painted by its neighboring region B ε (p) as shown in Fig. 5. known neighborhood Bε(P fg ) foreground region hole to be filled background region border known neighborhood Bε(P bg ) hole to be filled Replace Bε(P fg )to Bε(P bg ) Fig. 6 Manipulation of hole to have neighborhood only come from background o distinguish the foreground with background we use a corresponding depth data for blended image. For the horizontally opposite two depth pixels on hole, we regard the pixel having larger depth value as a foreground. Fig. 7 shows the procedure of the proposed hole-filling technique using depth based in-painting. known neighborhood Bε(P) P (a) (b) boundary Ω Fig. 5 General in-painting circumstance region Ω to be in-painted his concept is quite reasonable for common image inpainting but it should be changed to be applied to holefilling in view synthesis because Ω of a certain hole can be located on the both foreground and background. In this case, we replace the boundary region bordering on foreground to the background region located on the opposite side as depicted in (5). hat is, we intentionally manipulate the hole to have neighborhood only come from background as shown in Fig. 6. Now, since the holes only have the background pixels the in-painted image is more natural compared to the previous result. p fg Ω fg p bg Ω Bε ( p fg ) Bε ( pbg ) where fg and bg mean the foreground and background. bg (5) (c) (d) Fig. 7 In-painting procedure: (a) image with holes, (b) boundary region copy from background, (c) proposed depth based inpainting, (d) previous in-painting 4. EXPERIMENAL RESULS AND ANALYSIS We have tested the proposed algorithm on test sequences, Breakdancers and Ballet [10]. Among the 8 views, view 3 and view 5 are selected as reference views and view 4 is set as the virtual view. he proposed method is evaluated by existing objective evaluation measures, such as PSNR, SSIM [11], and VQM [12]. Whereas the larger value means the better quality in PSNR, the smaller value means the better quality for VQM. In case of SSIM, the closer value to 1 means the better quality. he experimental results for depth based in-painting are given in able 1 and able 2 and synthesized sample images in Fig. 8. he proposed depth-based in-painting shows the better results by filling up the remaining holes

using pixels located in background when the holes border on both foreground and background. able 1. Experimental results for Breakdancers Evaluation Measures Previous Proposed PSNR 31.7300 31.7484 SSIM 0.8381 0.8384 VQM 3.9973 3.9852 able 2. Experimental results for Ballet Evaluation Measures Previous Proposed PSNR 31.7773 32.4967 SSIM 0.8736 0.8740 VQM 2.6134 2.5131 5. CONCLUSIONS In this paper, we have proposed a hole-filling method using depth-based in-painting technique. he proposed method intentionally manipulates the holes to have neighborhood only come from background by considering depth value. And then, the existing in-painting technique is applied to fill up the holes. he effectiveness of the proposed method was confirmed by evaluating the quality of the synthesized image using various quality measures such as PSNR, SSIM, and VQM. We observed that the proposed method produced both subjectively and objectively better results compared with those by the current reference software being used in the MPEG FV/3D video standardization activities. ACKNOWLEDGMEN his work was supported in part by IRC through RBRC at GIS (IIA-2008-C1090-0801-0017). REFERENCES Fig. 8 Results of hole-filling: (top) previous in-painting, (bottom) proposed depth based in-painting As shown in experimental results and examples for synthesized images, the proposed depth-based in-painting shows subjectively and objectively better results for holefilling compared with the previous in-painting. [1] M. animoto, Overview of free viewpoint television, Proc. of Signal Processing: Image Communication, vol. 21, pp. 454 461, July 2006. [2] A. Smolic and P. Kauff, Interactive 3D video representation and coding technologies, Proc. of IEEE, Special Issue on Advances in Video Coding and Delivery. vol. 93, pp. 99 110, 2005. [3] F. Isgro, E. rucco, P. Kauff, and O. Schreer, 3D image processing in the future of immersive media, Proc. of IEEE ransactions on Circuits and Systems for Video echnology, vol.14, no. 3, pp. 288 303, 2004. [4] X. Chen and A. Luthra, MPEG-2 multiview profile and its application in 3D V, Proc. of SPIE-Multimedia Hardware Architectures, vol. 3021, pp. 212 223, 1997. [5] H.A. Karim, S. Worrall, A. H. Sadka, and A. M. Kondoz, 3D video compression using MPEG-4-multiple auxiliary component (MPEG4-MAC), Proc. of IEE 2nd International Conference on Visual Information Engineering (VIE2005), Apr. 2005. [6] ISO/IEC JC1/SC29/WG11, Survey of algorithms used for multi-view video coding (MVC), Doc. N6909, Jan. 2005. [7] S. C. Chan, H.Y. Shum, and K.. Ng, Image-based rendering and synthesis, Proc. of IEEE Signal Processing Magazines, pp. 22 33, Nov. 2007. [8] ISO/IEC JC1/SC29/WG11, Reference Softwares for Depth Estimation and View Synthesis, Doc. M15377, Apr. 2008. [9] A. elea, An image inpainting technique based on the fast marching method, Proc. of Journal of Graphics ools, vol.9, no.1, pp.25 36, 2004. [10] MSR 3D video Sequences. Available at: http://www.research. microsoft.com/vision/imagebasedrealities/3dvideodownload/. [11] Z. Wang, A. C. Bovik, H. R. Sheik, and E. P. Simoncelli, Image quality assessment: From error visibility to structural similarity, Proc. of IEEE ransaction on Image Processing vol. 13, no. 4, pp. 600 612, 2004. [12] F. Xiao, DC-based video quality evaluation, Final Project for EE392J Stanford Univ. 2000.