Coding for the Storage and Communication of. D. Tzovaras, N. Grammalidis, M. G. Strintzis, Aristotle University of Thessaloniki

Size: px
Start display at page:

Download "Coding for the Storage and Communication of. D. Tzovaras, N. Grammalidis, M. G. Strintzis, Aristotle University of Thessaloniki"

Transcription

1 Coding for the Storage and Communication of Visualisations of 3D Medical Data D. Tzovaras, N. Grammalidis, M. G. Strintzis, and S. Malassiotis Information Processing Laboratory Aristotle University of Thessaloniki Thessaloniki 54006, Greece phone: (+30-31) , fax: (+30-31) Abstract The transmission of the large store of information contained in 3D medical data sets through limited capacity channels, is a critical procedure in many telemedicine applications. In this paper, techniques are presented for the compression of visualizations of 3D image data for ecient storage and transmission. Methods are rst presented for the transmission of the 3D surface of the objects using contour following methods. Alternately, the visualisation at the receiver may be based on a series of depth maps corresponding to some motion of the object, specied by the medical observer. Depth maps may be transmitted by using depth map motion compensated prediction. Alternately, a wire-mesh model of the depth map maybe formed and transmitted by encoding the motion of its nodes. All these methods are used for the transmission of the 3D image with visualisation carried out at the receiver. Methods are also developed for ecient transmission of the images visualised at the encoder site. These methods allow remoteinteractive manipulation (rotation, translation, zoom) of the 3D objects, and may be implemented even if the receiver is a relatively simple and inexpensive workstation or a simple monitor. In all above cases, the coding of binocular views of the 3D scene is examined and recommendations are made for the implementation of coders of stereo views of 3D medical data. Subject Terms: 3D Data Compression Medical Data Visualisation Depth Map Transmission Telemedicine Applications. i

2 1 INTRODUCTION Data compression and coding is an essential processing task in medical image transmission and storage [1, 2, 3]. The transmission of the large store of information contained in medical data through limited capacity channels, is still an open problem [4, 5]. Also, this information must be coded for storage in Picture Archiving and Communication Systems in a way thatisecient, and also permits fast retrieval, especially for such operations as browsing or previewing the data. Coding and compression methods are necessary for the implementation of several medical data storage and transfer needes. These include methods for Coding of the whole 3D (volume) data set. Coding of the 3D object surface for visualisation at the receiver. Visualisation at the transmitter and coding of the visualised images. The entire 3D volume of the medical data is needed in applications such as stereotactic surgery planning and radiotherapy simulation. It is also needed for routine archival purposes. Very often, at least a portion of this data set must be coded in a lossless (reversible) way to fully retain the diagnostic information [1, 6, 7]. Methods for medical 3D volume compression and transmission, have beenwell investigated and their merits adequately discussed in the literature [1, 6, 7]. The present paper will concentrate on the remaining two general above methodologies, which have not been adequately analysed in the literature, even though they are very useful for the transmission of not only medical 3D data but also of other types of three dimensional information such asneededinquality control, virtual museum and teleshopping applications. Specifically, the present paper will investigate details of the implementation of these methods using alternative coding procedures and will experimentally assess the eciency of these implementations. First, methods for surface transmission will be examined, based on the lossless contourfollowing technique. Alternately, the receiver may use a depth map supplied by the encoder to visualise a still medical object, or if a depiction of a moving medical object is required, the receiver may rely on a series of depth maps corresponding to the required object motion. Such motion may for example be specied by the medical observer. Depth maps may be transmitted by using depth map motion compensated prediction. 1

3 In the above operations it is necessary to perform the nal visualisation of the data at the receiver site. This assumes the existence of a powerful workstation at the receiver site with the capacity of fast data visualisation. In practice it is very often preferable to have apowerful central workstation perform the visualisation and transmit the images to the peripheral stations in the form of images. This minimises the transmission delay and allows the viewing peripheral stations to be relatively simple and inexpensive \digital lightboxes". It also allows with minimum delay interactive manipulation of the data, with user initiated operations possible such as translation and rotation of the medical objects. To achieve this end, the present paperintroduces a special-purpose 3D object-based codec. This codec takes advantage of the fact that the motion of the object is user initiated and hence is completely known to the receiver. It also takes advantage of the knowledge at the transmitter site of the 3D shape of the object. Hence the only parameters needed to be transmitted for each group of pictures are : a) the initial image, b) a succession of depth maps. Depth compensation is used for the compression of the latter. Alternatively, the depths of the nodes of a wire mesh model of the object may be transmitted. The paper is organised as follows. In Section 2, the problem of coding of visualised data is presented, and solutions are proposed for the case where visualisation is performed at the receiver. Section 2.2 outlines a lossless technique for 3D surface compression, while in Section 2.3 novel ecient techniques for low bit rate coding of a sequence of depth maps corresponding to multiple views of the 3D scene, are described. An alternate depth map coding technique based on wire mesh modeling is then examined in Section 2.4. In Section 3, a solution is proposed for the case where visualisation is performed at the transmitter, using a novel object-based technique for motion compensation of the sequences of visualised images. Finally, Section 4 evaluates the use of the examined coding techniques for stereoscopic medical image viewing. Experimental results described in Section 5, demonstrate the performance of the proposed methods. 2 VISUALISATION AT THE DECODER SITE If a full-capacityworkstation is available at the decoder site, it will be sucient to transmit the 3D information: A. Using a 3D volume data form. 2

4 B. In 3D surface form. An alternative to the transmission of the whole 3D data set is the transmission of the object contour using methods which will be described later in this section. C. In the form of 2D snapshots of the depth maps. The decoder will then visualise the data. In the sequel the rendering process used for the visualisation of the medical data is presented as an introduction to the description of the proposed coding methods. 2.1 Rendering Process The term visualisation usually refers to any process that allows visual understanding and interpretation of 2D or 3D data. In medical imaging, visualisation is performed primarily using volume or surface rendering techniques. Volume rendering [4] is used to show the interior of a solid 3D object on a 2D image. This is very useful in many medical imaging applications where the boundaries between dierent tissues have to be indicated. In surface rendering [8, 9] the goal is the creation of realistic pictures of 3D objects. This is done by simulating the lighting of the scene, the surface properties (e.g. surface color), the camera-eye system, the position of the viewer and the object motion. In our experiments, visualisation of the Magnetic Resonance Imaging (MRI) data was performed by the 3DVIEWNIX software package [10]. This package provides a variety of manipulating and analysing tools for the visualisation and display ofmultidimensional image data. 2.2 Lossless 3D Surface Compression Using Contour Following The original 3D data usually consist of several slices of 2D data. If a single closed outer contour per slice is assumed, the 3D surface may beequivalently represented by stacking these 2D contours. In the contour following approach [13], an initial point on the contour is chosen and then the contour is traced in a clockwise manner, labeling the direction as we shift from one contour element to the next (see Fig. 1). The resulting data stream is then entropy coded using arithmetic coding techniques [14]. This is a lossless approach that also has the advantage of simplicity, leading to very fast implementation. 3

5 NW N NE W E...NE,E,SE,SE,E,NE,... SW S SE 2.3 Depth Map Coding Figure 1: Contour following. In several important medical applications, the method of 3D surface description using contour following is not easily applicable (for instance see Fig. 7). For example, while the test sequence \Head" (Fig. 6a) is described well by the methods discussed above, the sequence \Brain" is not easily amenable to such description, since it is extremely dicult to dene an enveloping contour for so complicated an object. In such cases a preferable description of the depth maps may be aorded directly by the ray-tracing used (Fig. 8). Specically, a number of snapshots of the depth maps, corresponding to certain angles of viewing the object, may be used for the depiction of the medical information that must be stored or transmitted. This series of snapshots may then be coded as though they were consecutive in time and their aggregate may be treated as a \moving" sequence. In fact, since motion parallax is a powerful aid to image understanding, their actual positioning in a moving sequence may be precisely what the medical observer requires [4]. The rotation back and forth for example of a 3D MRI image gives a very good understanding of medical detail. Ecient techniques must then be found for the coding and compression of the depth maps corresponding to multiple views of a particular 3D data set. If ~V t =(X t Y t Z t ) is the 3D vector denoting the position of a point in the 3D space at time instant t, the position of the same point at time t + 1 will be : ~ V t+1 = R ~ V t + ~ T (1) where R is the rotation matrix and ~ T is the translation vector. The general form of 4

6 equation (1) is : X t+1 Y t+1 Z t = 0 R 11 R 12 R 13 R 21 R 22 R 23 R 31 R 32 R C A 6 4 X t Y t Z t T x T y T z 1 C A : (2) The image formation process of the visualised images, is a perspective projection of a 3D object onto the image plane. The center of projection is located at the origin of the 3D world coordinate system, at a distance f from the image plane. If (X t Y t Z t ) T position vector of a point in the 3D space and (x t y t f) T is the is the position vector of its projection at the image plane, then the following relation exists between the image and the world coordinates at two consequent time instants: x t+1 = f X t+1 Z t+1 x t = f X t Z t y t+1 = f Y t+1 y t = f Y t : (3) Z t+1 Z t On replacing in (2), we obtain the following system of equations relating the image plane pixel coordinates (x t y t ) with the corresponding depth Z t and their evolution in time. If fr ij g ft x T y T z g are the object motion parameters: x t+1 Z t+1 = R 11 x t Z t + R 12 y t Z t + R 13 Z t + ft x y t+1 Z t+1 = R 21 x t Z t + R 22 y t Z t + R 23 Z t + ft y Z t+1 = R 31 x t Z t + R 32 y t Z t + R 33 Z t + ft z : (4a) (4b) (4c) The third of the above equations provides the new depth Z t+1 at time instant t +1 corresponding to pixel (x t y t ), while the rst two provide its new position (x t+1 y t+1 ). Some problems arise in the motion compensation procedure, due to the fact that x t+1 and y t+1 are oating point numbers. Thus the new position of the point (x t+1 y t+1 ) at time instant t + 1, lies outside the sampling grid of the subsequent frame. Therefore an interpolation procedure has to be adopted, in order to assign values at integer pixel locations. One such procedure, based on upsampling of the depth map Z t at time instant t was implemented. This technique led to satisfactory results that improve as the upsampling rate increases. With this scheme, only the object boundary information need to be transmitted for motion compensation. The object boundary information is coded losslessly using arithmetic coding techniques [14]. Further, to oset errors in the motion compensated depth map prediction, particularly those occuring in newly occluded or newly appearing areas, the error depth map is transmitted, using DCT and Human encoding. A block-diagram of the proposed depth map coder is shown in Figure 9. 5

7 2.4 Depth Map Transmission with Wire Mesh Modeling Alternately, depth information may be modeled using a wire mesh. We shall assume a surface model of the form z = S(x y ^P) (5) where ^P is a set of 3D \control" points ^P = f(xi y i z i ) i=1 ::: Ng that determine the shape of the surface and (x y z) are the coordinates of a 3D point. Consider a set of image points ^A = f(x i Y i ) i=1 ::: Ng and a triangulation ^T of ^A such as the triangles of ^T cover the image plane. An initial choice for ^A is the regular tesselation shown in Figure 2. Figure 2: Initial triangulation of the image plane. Then, we can model S by z = NX i=1 z i N i (x y) (6) 6

8 where N i (x y) are basis functions shown in Figure (x i,y i ) Figure 3: Basis functions It may be easily seen that (6) is a piecewise linear surface, consisting of adjoint triangular patches, which may be written in the form if (x y z)isapoint on the triangular patch z = z 1 g 1 (x y)+z 2 g 2 (x y)+z 3 g 3 (x y) (7) d P1 P 2 P 3 = f(x 1 y 1 z 1 ) (x 2 y 2 z 2 ) (x 3 y 3 z 3 )g. The functions g i (x y) are the barycentric coordinates of (x y z) relative to the triangle and they are given by g i (x y) =Area(A i )=Area( P1 P 2 P 3 ) (Figure 4). P 1 d A 3 A 2 P 2 A 1 P 3 Figure 4: Barycentric coordinates. This may be also written in the better known form z = ax + y + (8) where (a ) is the slope and the elevation of the corresponding triangular patch. The reconstruction of a surface from dense depth measurements may be eected by minimizing a functional of the form X M E 0 ( ^P )= i ~y i ^P ) ; ~zi ) i=1(s(~x 2 + E s (9) 7

9 where the rst term of (9) expresses condence to the data (~x i ~y i ~z i ), i =1 ::: M and E s is the function measuring departure from smoothness of S, given by E s = ZZ!(x y)(s 2 xx +2S 2 xy + Syy) 2 dx dy (10) where!(x y) is a binary function that equals zero if (x y) lies on a discontinuity. Taking derivatives of the surface function S(x y ^P) written in the form (8) and replacing in (10) we obtain, after simple manipulations E s = X e2e e l e f(a 1 e ; a2 e )2 +( 1 e ; 2 e )2 g (11) where E is the set of all non-boundary edges of the wireframe, l e is the length of an edge and (a 1 e 1) e (a2 e 2 e ) are the slopes of the triangular patches that have a common edge e. The binary variable e assumes the zero value if e crosses a discontinuity. The existence of the term E s guarantees that we obtain a smooth surface. In the case of the regular tesselation (11) becomes X E s = l2 (z i ; z a + z j ; z b ) 2 (12) A 2 e2e where l is the edge length, A is the area of each triangle, and P i P j P a P b are shown in Figure 5. Replacing (12),(7) in (9) P i P b e P a P j Figure 5: Edge e between two neighboring triangles. E 0 ( ^P )=ka ^P ; Bk 2 + kc ^P k 2 (13) where ( gj (~x A ij = i ~y i ) if (~x i ~y i ) inside triangle f(x j y j ) (x k y k ) (x l y l )g 0 otherwise i =1 ::: M j =1 ::: N 8

10 B i =~z i i =1 ::: M and C ki =1 C ka = ;1 C kj =1 C kb = ;1 C kl =0ifl=2fi j a bg k =1 ::: L where L is the number of non-boundary edges, and i j a b are the indices of the nodes that belong to triangles incident to the edge k as shown in Figure 5. Minimization of (13) occurs when : ^P =(A T A + C T C) ;1 A T B (14) Using (8), the depth z of any point onapatch can be expressed in terms of the depth information of the nodes of the wireframe and the X and Y coordinates of that point. Hence, full depth information will be available if only the depths of the nodes of the wireframe are transmitted. After wire mesh model initialization, motion compensated estimates may befound of the depth maps corresponding to subsequent time instants, by updating the position of the nodes of the wireframe only. Occlusions may then be detected and compensated by transmitting the depth map error information, as explained in the preceding section. Using this occlusion information, the wire mesh model is updated by adding new nodes at the new appearing areas, which occur at the boundary of the object. The wire mesh update information need also be transmitted at each time instant. 3 VISUALISATION AT THE ENCODER SITE In many practical applications, where the receiver must be a relatively simple and inexpensive workstation, visualisation at the receiver will require an inordinate amount of time. Then, if a powerful workstation is available at the database site or the diagnostic image source, the visualisation may be performed on this workstation and the resulting 2D images transmitted to the receiver. This will allow the receivers to be inexpensive personal computers, or even simple high-resolution monitors which may be dispersed throughout a hospital as \digital lightboxes". The need for visualisation at the transmitter or archive site also arises when the archive contains images of diagnostic value that the physician has decided to save in visualised format. The original 3D data may also be on le, but 9

11 the physician will normally want toavoid repeating the time-consuming segmentation and rendering procedure. The coding scheme to be used must allow remote interactive handling and manipulation (rotation, translation, zoom) of the 3D data by the user at the receiver site, using a simple signaling device such as a mouse or a joystick. 3.1 Special-Purpose 3D Object-Based Coding A coding scheme suitable for the coding of 3D medical images visualised at the encoder, may be based on the knowledge of the 3D motion eld. In fact, since the motion of the object is user-initiated, both encoder and decoder know the 3D motion of the object with precision. Thus, if a depth map of the object is also provided, the 2D motion eld may be obtained from the projection of the 3D motion and be used for interframe motion compensation. The production of the series of images (visualisations) corresponding to the prescribed 3D object motion may be accelerated using incremental visualization techniques such as reported in [25]. If the series of images to be transmitted are divided into MPEG-like groups of images, the initial image (\I-frame" in MPEG terminology) in each group can be transmitted using lossless coding, simple DPCM, DCT or wavelet-based coding [16, 17]. Interframe coding may beachieved using the known 3D motion parameters, on the basis of the analysis presented in Section 2.3. Specically, equations (4a),(4b),(4c) may be used for motion compensation, provided that the relation is found between the luminances of a pixel at its successive positions (x t y t )and(x t+1 y t+1 ). In theory this is possible. However, this relationship is an extremely complex function of the mechanics of the rendering procedure and depends on visualisation parameters such as the lighting conditions assumed, and surface reectabilitychosen, as well as the gradient of the surface at each point. A simple alternative to this very complicated construction is the working assumption of equal intensity intwo successive pixel positions: I(x t+1 y t+1 )=I(x t y t ) : (15) The resulting object-based 3D motion compensation scheme is dened by (4) and (15). Clearly, the scheme may be decoupled into two separate procedures : 10

12 Depth map evaluation using, Z t+1 = R 31 x t Z t + R 32 y t Z t + R 33 Z t + ft z (16) followed by motion compensated image intensity prediction using (15) and, x t+1 Z t+1 = R 11 x t Z t + R 12 y t Z t + R 13 Z t + ft x (17) y t+1 Z t+1 = R 21 x t Z t + R 22 y t Z t + R 23 Z t + ft y : (18) Alternately, the depth map Z t+1 may beevaluated by means other than (16), for example using the contour information discussed in section 2.2. In the experimental results presented in Section 5, it is assumed that (16) is used for depth map evaluation. The information contained in the depth map supplied to the receiver may be used to enable stereo viewing of the data (Section 4). A block-diagram of the proposed visualized image coding scheme is shown in Figure Transmission using a general-purpose object-based coder based on 3D motion compensation A general purpose object-based coder, for the coding of general scenes such as described in [20, 21, 22, 23, 24], may also be used for the coding of medical visualisations. This is of interest for the purpose of comparison with the special 3D medical image coder described above. The coder in [24] was implemented for this purpose. This coder supplies full depth information to the receiver. When such information is needed (e.g. for stereoscopic viewing applications), this coder is therefore advantageous compared to MPEG-2. The depth map is estimated by the coder, using for each time instant two visualisations of the object separated bya2 rotation about the vertical axis. This approximates a stereo view of the medical object and may be used to determine the approximate depth map. The motion eld is estimated using the sequence of visualisations. Specically, the generalpurpose coder works as follows [24]: a) Uniformly displaced regions are identied: Block-based estimation is used to identify an initial estimate of the 2D motion eld. Then a split-and-merge segmentation algorithm is used to identify the objects indicated by the uniformly displaced regions. 11

13 b) The 3D motion of each object is found: The 3D motion parameters (translation and rotation) are related to the 2D motion eld vectors d =(d x d y )by the equation: d x = f t z x ; X t z z ; XY w f x +(f + X2 )w f y ; Yw z d y = f t z y ; Y t z z ; (f + Y 2 )w f x + XY w (19) f y + Xw z where (X Y )isapoint in the current image and z is its corresponding depth. Also f is the focal length and (t x t y t z ),(w x w y w z ) are the 3D object translation and rotation parameters, respectively. Equations (19), are used to estimate the 3D motion of each object at two stages. In the rst stage, equations (19) are used for N of the initially estimated 2D vectors, forming a system of 2*N equations and 6 unknowns. With N 3 this system is overdetermined and can be solved using least-squares methods. In the second stage, improved estimation of model parameters is achieved by minimising the displaced frame dierence (DFD) between the object as seen at the current image, and the motion compensated reconstruction produced from the previous image, using the estimated 3D parameters. A steepest-descent technique is used to achieve this minimisation. c) The interframe coding scheme operates in two basic modes: object-based mode and block-based mode. In each image region the 3D motion parameters are rst estimated. If the resulting reconstruction error is smaller for a region than that of the 2D motion vectors, then the coder goes into object-based mode and the six 3D motion parameters are quantised and coded using PCM. We used a uniform scalar quantiser between the minimum and maximum values of the parameters. Otherwise the coder goes into blockbased mode, the region is divided into 8 8blocks and 2D motion compensation is used. A bitmap is also transmitted indicating the coding mode for each region. 4 CODING FOR STEREOSCOPIC VIEWING OF 3D MEDICAL DATA Depth perception, hence stereo viewing, is useful in many medical applications. If visualisation is performed at the decoder site, generation of a second (stereo) view will necessitate the completion of two separate rendering procedures. If visualisation is done at the encoder site, a second (right) channel image may be coded precisely as the rst (left) sequence. Alternately, the right channel image may be formed from the left image using disparity compensation. This is possible with all coding methods discussed in Sections 12

14 2 and 3 of this paper, because in all these methods the depth map is also transmitted. Therefore, at the decoder side the disparity eld is found from the depth map using the relation d = bf z (20) where b is the baseline, f is the focal length, d is the disparity and z is the depth. The right channel image is then formed from the left channel image and disparity compensation : I r (x y) =I l (x + d y) : (21) 5 EXPERIMENTAL RESULTS The methods in Sections 2-4 were tested using MRI data of the head (represented with 8 bits=pixel) forming 256 slices with a 1mm distance between slices. The resolution of each slice is pixels. A mouse driven segmentation procedure based on thresholding was used to extract from the original 3D data set \Head", the \Brain" data set. The sequence of depth maps corresponding to 2D views of the 3D data sets were generated from the original MRI slices. The depth maps corresponding to these data sets and the corresponding rendered views are illustrated in Fig Transmission of Contour and Depth Map Information To enable visualization at the receiver, the lossless contour-following method was used to transmit the 3D surface of the \Head" image. This required the transmission of bytes and resulted in the reconstruction shown in Figure 6b. Note that the bytes, necessary to transmit the entire 3D surface, correspond to an average of =( ) = 0:03807 bits=pixel when 180 depth maps (one every 2 degrees) have to be transmitted. Also, to enable the receiver to visualize a rotation of the medical object, a sequence of depth maps corresponding to 2D views of the 3D data set were generated from the original MRI slices. Each frame of the depth map sequence, corresponds to an object rotation by 2 degrees around the Y axis. An object-based 3D motion compensated prediction scheme for the interframe compression of the resulting sequence of depth maps is dened by equation (4c). The following information must be made available to the decoder: The previous depth map. 13

15 Denition of the objects in the scene (segmentation information). The error map corresponding to areas which were occluded in the previous depth map. The 3D object-based motion compensation method is integrated in an MPEG-like coding scheme for the transmission of a sequence of depth maps. In this scheme it is assumed that the rst depth map is intra coded using I-Frame coding techniques, while the subsequent frames are coded using motion compensation (P-Frames) : IPP...PPIPP... Each P-frame is predicted from the previous frame (P or I) in the sequence. The decoder which has already available a reconstruction of the previous frame, uses 3D objectbased motion compensation of the already available reconstruction of the previous frame, to obtain a motion-compensated approximation of the current frame. The error map corresponding to areas which were occluded in the previous depth map is then coded using DCT and Human coding. The block diagram of the resulting codec is shown in Fig 9. In the decoder, the rendering procedure described in Sec. 2.1 is used to visualise the reconstructed depth map information. Experimental results evaluating the performance of the proposed method before and after the coding of the prediction error images are given in Tables 1 and 2 respectively, relating the bits/pixel needed for the transmission of the depth maps to the quality of the images visualised on the basis of these depth maps. Furthermore, in Figure 10 the original and reconstructed depth maps of the frame 2 of the sequence \Head" are shown along with the rendered images corresponding to these depth maps. Alternately, if the wire mesh modeling technique is used for depth map coding the initial wireframe is computed using the techniques described in Section 2.4. The initial wireframe adapted to the rst frame of \Head" is shown in Figure 11. The wire mesh model parameters (coordinates of the 1112 nodes) were then quantised and coded using DPCM techniques, followed by arithmetic coding. We used uniform scalar quantisers between the minimum and maximumvalues of each model parameter. Depth information was then interpolated from the depths of the nodes of the wireframe. The depth map corresponding to the modeled rst frame of \Head" is shown in Figure 12a and the corresponding visualized image is illustrated in Figure 12b.. Motion compensation of the 14

16 nodes of the wireframe is then performed followed by depth error transmission. Table 3 shows the the bits/pixel needed for the transmission of the wire mesh versus the quality of the images visualised on the basis of the interpolated depth maps. 5.2 Transmission of Visualised Images The results of Section 3 were also tested to simulate the transmission of images visualized at the encoder. In the proposed special purpose object-based scheme, 3D motion compensation is possible for every point of an object in the scene, if the following information is made available to the decoder: A reconstruction of the previous frame in the visualised image sequence. The previous depth map. The error image, which includes the newly apparent, previously occluded regions and the errors due to the inaccuracy of assumption (15). An MPEG-like motion compensating scheme was implemented in this case as well. The 3D object-based motion-compensation method was integrated in an MPEG-like coding scheme for the transmission of a sequence of images. In this scheme, the rst visualised image (I-frame) is coded using DCT and Human techniques, while the subsequent frames are coded using 3D motion compensation (P-frames). Each P-frame is predicted from the previous frame (P or I) in the sequence. The decoder which has already available a reconstruction of the previous frame, uses 3D object-based motion compensation to obtain a motion-compensated approximation of the current frame. The number of P- frames between successive I-frames was chosen to be 6. The sequence of the corresponding depth maps which is necessary for motion compensation, is also coded using the adopted MPEG-like scheme. The prediction error images, may also be coded and transmitted. Note that in this scheme the error images involve not only intensity error images from equations (17),(18) but also depth-map errors from equation (16). These are coded using DCT and Human coding and transmitted. A block diagram of the resulting codec is shown in Fig. 13. Since the 3D motion eld is accurate, DCT and Human coding of the prediction errors produce near-lossless compression of the visualised images (Table 6). The small coding error is due to the accurate knowledge of the observer-initiated translation and 15

17 rotation parameters. By comparison of Tables 5 and 6 we conclude that the proposed method performs slightly better than MPEG-2 [15]. In the demonstration of the RACE II project DISTIMA, the evaluators failed to detect a signicant qualitative dierence between the two methods. Furthermore, the method proposed does not require motion estimation and hence is considerably simpler than MPEG-2 encoding/decoding. Finally, unlike MPEG, this coding method provides full depth map information to the receiver. 5.3 Transmission for Stereoscopic Viewing of Images Depth information is needed at the receiver site if stereoscopic viewing of the data is desired (Section 4). It is also useful in applications where the image data are to be stored and subsequently interactively manipulated in a 3D manner (rotation, 3D translation) at the receiver. For reasons of comparison, the original frame 2 of the \Brain" image, along with the reconstructed images using the 3D object coder, and the MPEG-2 method are shown in Figures 14(a,b,c). The performance of the general purpose stereoscopic coder was also tested on the \Brain" sequence. Tables 7 and 8 present the PSNR of the reconstructed left image and the number of the bytes that have tobesent to the decoder, with and without coding of the prediction error images respectively. The reconstructed frame 2 of the visualised image sequence is depicted in Figure 14d. As noted, the general-purpose object-based coder produces results inferior to those of both the special-purpose object-based coder and MPEG. Compared with MPEG however, the general-purpose object-based coder has the advantage of fully transmitting depth information which as noted, permits stereo viewing and 3D storage of the sequence and may otherwise be needed in medical use. With all above methods, disparity may be estimated from depth and disparity compensation may be used to generate the right image sequence from the left image sequence. For better quality stereo image, the error image resulting from the inaccuracy of (21) and from the presence in the right view of areas occluded in the left must be transmitted. This will result in a modest overhead in bitrate as shown in Table 9. 6 CONCLUSIONS In telemedicine applications, and also for ecient storage and retrieval of 3D images in Picture Archiving and Communication Systems, coding and compression methods are 16

18 necessary for the implementation of several data storage and transfer modes. include methods for, These Coding of the whole 3D (volume) data set. Coding of the 3D object surface for visualisation at the receiver. Visualisation at the transmitter and coding of the visualised images. For these needs, in summary, the paper discusses : afor visualisation at the decoder site, transmission of surface information using { the 3D contour method of Section 2.2 or, { the depth-map snapshot method of Section 2.3. { the depth-map snapshot method of Section 2.4 using wire-mesh modelling of the depth maps. b If rendering is performed at the encoder site, transmission of the resulting visualisation via the novel object-based coder of Section 3. Although this coder supplies full depth map information to the receiver, it was seen to perform slightly better than MPEG-2. cfor stereo viewing of images produced either at the encoder or the decoder site construction of the right image from the left channel images is proposed in Section 4. In fact, since the proposed monoscopic coders require depth map transmission, the disparity eld may be obtained and hence a stereo view may be composed from the monoscopic view by disparity compensation. In this way, an adequate stereo presentation may be obtained without additional computational cost or bitrate increase. For a better-quality stereo view, the disparity compensation error image must be transmitted, at a modest overhead in bitrate. 7 Acknowledgment This material is based on work supported by the RACE 2045-DISTIMA and the ACTS 092-PANORAMA projects. 17

19 References [1] P. Roos and M. A. Viergever, \Reversible 3D Decorrelation of Medical Images," IEEE Trans. on Medical Imaging, vol. 12, pp. 413{420, Sep [2] H. Musmann, P. Pirsch, and H. Grallert, \Advances in Picture Coding," Proc. IEEE, vol. 73, pp. 523{548, Apr [3] A. Netravali and B. Haskell, Digital Pictures: Representation and Compression. New Jersey: Plenum, [4] J. Udupa and G. Herman, 3D Imaging in Medicine. Boca Raton., FL: CRC Press, [5] S. H. Oguz, O. N. Gerek and A. E. Cetin, \Motion Compensated Prediction Based Algorithm for Medical Image Sequence Compression", Signal Processing : Image Communication, vol. 7, No. 1, pp , [6] L. Kondis, C. Chrysas, H. Sahinoglou, and M. G. Strintzis, \3D Medical Data Compression," in Proc. Int 'l. Workshop on Image Processing, (Budapest), June [7] H. Sahinoglou, S. Malassiotis, and M. G. Strintzis, \Lossless Coding and Visualization of 3D Medical Data With Lossy Preview Capability," Bioimaging, to appear, [8] A. Kaufman, A Tutorial on Volume Visualization. Los Alamitos, CA: IEEE Computer Society Press, [9] J. Udupa, H. Hung, and K. Chuang, \Surface and Volume Rendering in 3D Imaging : a comparison," Journal of Digital Imaging, vol. 4, pp. 196{214, [10] J. Udupa, R. Goncalves, K. Iyer, S. Narendula, D. Odhner, S. Samarasekera, and S. Sharma, \3DVIEWNIX : An open, transportable software system for the visualization and analysis of multidimensional, myltimodality, multiparametric images," in SPIEE Proceedings, vol. 1987, pp. 47{58, [11] R. Thoma and M.Bierling, \Motion Compensating Interpolation Considering Covered and Uncovered Background," Signal Processing : Image Communication, vol. 1, no. 1, pp. 191{212, [12] N. A. J.Weng and T. Huang, \Matching two perspective views," IEEE Trans. Pattern Anal. and Mach. Intell., vol. 14, pp. 806{825, August [13] J. A. Saghri and H. Freeman, \Computer processing of line drawing images," IEEE Trans. PAMI, pp. 533{539,

20 [14] R. Witten, I. Neal, and J. Cleary, \Arithmetic Coding for Data Compression," Communications of the ACM, vol. 30, pp. 520{540, Jun [15] MPEG, \MPEG Video Simulation Model Three," tech. rep., ISO/IEC JTC1/SC2/WG11 N0010, MPEG 90/041, July [16] M. G. Strintzis, \Optimal Biorthogonal Wavelet Bases for Signal Representation," IEEE Transactions on Signal Processing, vol. 44, no. 6, pp , June [17] S. Efstratiadis, D. Tzovaras, and M. G. Strintzis, \Hierarchical Partition Priority Wavelet Image Compression," IEEE Transactions on Image Processing, vol. 5, no. 7, pp , July [18] J. Weng, T. Huang, and N. Ahuja, \Motion and Structure from Two Perspective Views: Algorithms, Error Analysis and Error Estimation," IEEE Trans. Pattern Anal. and Mach. Intell., vol. 11, pp. 451{476, May [19] S. Malassiotis, and M. G. Strintzis, \Joint Motion/Disparity MAP Estimation for Stereo Image Sequences," accepted for publication, IEE Proceedings: Vision, Image and Signal Processing, [20] D. Tzovaras, N. Grammalidis, and M. G. Strintzis, \Object-Based Coding of Stereo Image Sequences Using Joint 3D Motion/Disparity Compensation," IEEE Transaction on Circuits & Systems for Video Technology, to appear, Vol. 7, Apr [21] N. Diehl, \Object-Oriented Motion Estimation and Segmentation in Image Sequences," Signal Processing : Image Communication, pp. 23{56, Feb [22] H. G. Musmann, M. Hoetter, and J. Ostermann, \Object-Oriented Analysis- Synthesis Coding of Moving Images," Signal Processing : Image Communication, pp. 117{138, Oct [23] J. Dugelay and D. Pele, \Motion and Disparity Analysis of a Stereoscopic Sequence, Application to 3DTV Coding," Signal Processing, pp. 1295{1298, Oct [24] N. Grammalidis, S. Malassiotis, D. Tzovaras, and M. G. Strintzis, \Stereo Image Sequence Coding based on 3D Motion Estimation and Compensation," Signal Processing : Image Communication, vol. 7, pp. 129{145, Aug [25] B. Gudmundson, and M. Rand, \Incremental Generation of Projections of CT - Volumes", Proc. 1st Conf. on Visualization in Biomedical Computing, N. F. Ezquerra, Editor, pp ,

21 Depth-Map Bit Rate PSNR Depth PSNR Corresponding Number (bits/pixel) Map (db) Visualised Image (db) Table 1: Depth Map Coding, \Head" : Bit-rate versus PSNR performance of the motion compensation method for depth map coding for : a) the depth map, b) the visualisation of the reconstructed depth map information. Depth Bit Rate Bit Rate Bit Rate PSNR PSNR Map Motion Comp. Error Encoding Total Depth Map Corresponding Number (bits/pixel) (bits/pixel) (bits/pixel) (db) Vis. Image (db) Table 2: Depth Map Coding + DCT Error Coding, \Head" : Bit-rate versus PSNR performance of the motion compensation method for depth map coding for : a) the depth map, b) the visualisation of the reconstructed depth map information. 20

22 Depth Bit Rate Bit Rate Bit Rate PSNR PSNR Map Motion Comp. Error Encoding Total Depth Map Corresponding Number (bits/pixel) (bits/pixel) (bits/pixel) (db) Vis. Image (db) Table 3: Wire Mesh Modeling of the Depth + DCT Error Coding, \Head" : Bit-rate versus PSNR performance of the wire mesh modeling method for depth map coding for : a) the depth map, b) the visualisation of the reconstructed depth map information. 21

23 Frame Bit Rate PSNR Number (bits/pixel) (db) Table 4: Visualised Image Coding using the special 3D object codec, \Brain" : Bit-rate versus PSNR performance of the 3D motion compensation method for visualised image coding, using the reconstructed depth maps (with DCT coding of the depth map error). Frame Bit Rate PSNR Number MPEG Vis. Image (bits/pixel) (db) Table 5: Visualised Image Coding using MPEG-2, \Brain" : Bit-rate versus PSNR performance of the MPEG-2 method for visualised image coding. 22

24 Frame Bit Rate Bit Rate Bit Rate PSNR Number Motion Comp. Error Encoding Total Vis. Image (bits/pixel) (bits/pixel) (bits/pixel) (db) Table 6: Visualised Image Coding using the special 3D object codec \Brain" : Bit-rate versus PSNR performance of the 3D motion compensation method for visualised image coding followed by DCT and Human coding of the error image and the error depth-map. 23

25 Frame Bit Rate PSNR Number (bits/pixel) (db) Table 7: Coding using the general-purpose coder, \Brain" : Bit-rate versus PSNR performance for visualised image coding using the general purpose object-based coder. Frame Bit Rate Bit Rate Bit Rate PSNR Number G-P Coder Error Encoding Total Vis. Image (bits/pixel) (bits/pixel) (bits/pixel) (db) Table 8: Coding using the general-purpose coder, \Brain" : Bit-rate versus PSNR performance for visualised image coding using the the general-purpose coder followed by DCT and Human coding of the error images. 24

26 Frame Bit Rate Bit Rate Bit Rate PSNR Number G-P Coder Error Encoding Total Vis. Image (bits/pixel) (bits/pixel) (bits/pixel) (db) Table 9: Coding Using Disparity Compensation + Error Coding, \Brain" : Bit-rate versus PSNR performance for stereoscopic viewing of the visualised image sequence followed by DCT and Human coding of the error images. 25

27 (a) (b) Figure 6: Performance of the contour following modeling technique. Original (a) and reconstructed (b) rendered views. 26

28 (a) (b) Figure 7: Performance of the contour following modeling technique. Original (a) and reconstructed (b) rendered views. 27

29 (a) (b) (c) (d) Figure 8: (a,b) Depth maps of the front view, (c,d) Rendered Views. 28

30 Current Depth Map Object Boundary Extraction Arithmetic Coder Previous Depth Map Previous Depth Map Reconstructed Depth Map 3 D Motion Compensation ENCODER Desired Motion Parameters DCT and Huffman Coding IDCT and Huffman Decoding DECODER 3 D Motion Compensation + Depth Map Estimate Figure 9: Block Diagram of the proposed Depth Map Coding Scheme. 29

31 (a) (b) (c) (d) Figure 10: a) Original depth map frame 2, b) reconstructed depth map using the depth map coding scheme, c) visualised image corresponding to the original depth map, d) visualized image corresponding to the reconstructed depth map. 30

32 Figure 11: Wire mesh adapted to the image \Head". (a) (b) Figure 12: (a) Wire mesh modeled depth map, (b) the corresponding visualized image. 31

33 Current Image Object Boundary Extraction Arithmetic Coder Previous Image Previous Image 3 D Motion Compensation Desired Motion Parameters Previous Depth Map 3 D Motion Compensation Reconstructed Image ENCODER Depth Map Coder DCT and Huffman Coding Depth Map Decoder IDCT and Huffman Decoding DECODER + Current Image Estimate Figure 13: Block Diagram of the proposed Visualised Image Coding Scheme. 32

34 (a) (b) (c) (d) Figure 14: a) Original visualised image frame 2, b) Reconstructed image using the special 3D object coder, c) Reconstructed image using the MPEG-II coder, d) Reconstructed image using the general purpose object-based coder. 33

Tracking Moving Objects In Video Sequences Yiwei Wang, Robert E. Van Dyck, and John F. Doherty Department of Electrical Engineering The Pennsylvania State University University Park, PA16802 Abstract{Object

More information

Peter Eisert, Thomas Wiegand and Bernd Girod. University of Erlangen-Nuremberg. Cauerstrasse 7, 91058 Erlangen, Germany

Peter Eisert, Thomas Wiegand and Bernd Girod. University of Erlangen-Nuremberg. Cauerstrasse 7, 91058 Erlangen, Germany RATE-DISTORTION-EFFICIENT VIDEO COMPRESSION USING A 3-D HEAD MODEL Peter Eisert, Thomas Wiegand and Bernd Girod Telecommunications Laboratory University of Erlangen-Nuremberg Cauerstrasse 7, 91058 Erlangen,

More information

Analyzing Facial Expressions for Virtual Conferencing

Analyzing Facial Expressions for Virtual Conferencing IEEE Computer Graphics & Applications, pp. 70-78, September 1998. Analyzing Facial Expressions for Virtual Conferencing Peter Eisert and Bernd Girod Telecommunications Laboratory, University of Erlangen,

More information

Module 8 VIDEO CODING STANDARDS. Version 2 ECE IIT, Kharagpur

Module 8 VIDEO CODING STANDARDS. Version 2 ECE IIT, Kharagpur Module 8 VIDEO CODING STANDARDS Version ECE IIT, Kharagpur Lesson H. andh.3 Standards Version ECE IIT, Kharagpur Lesson Objectives At the end of this lesson the students should be able to :. State the

More information

Figure 1: Relation between codec, data containers and compression algorithms.

Figure 1: Relation between codec, data containers and compression algorithms. Video Compression Djordje Mitrovic University of Edinburgh This document deals with the issues of video compression. The algorithm, which is used by the MPEG standards, will be elucidated upon in order

More information

Fast and Robust Moving Object Segmentation Technique for MPEG-4 Object-based Coding and Functionality Ju Guo, Jongwon Kim and C.-C. Jay Kuo Integrated Media Systems Center and Department of Electrical

More information

Performance Analysis and Comparison of JM 15.1 and Intel IPP H.264 Encoder and Decoder

Performance Analysis and Comparison of JM 15.1 and Intel IPP H.264 Encoder and Decoder Performance Analysis and Comparison of 15.1 and H.264 Encoder and Decoder K.V.Suchethan Swaroop and K.R.Rao, IEEE Fellow Department of Electrical Engineering, University of Texas at Arlington Arlington,

More information

Introduction to image coding

Introduction to image coding Introduction to image coding Image coding aims at reducing amount of data required for image representation, storage or transmission. This is achieved by removing redundant data from an image, i.e. by

More information

Parametric Comparison of H.264 with Existing Video Standards

Parametric Comparison of H.264 with Existing Video Standards Parametric Comparison of H.264 with Existing Video Standards Sumit Bhardwaj Department of Electronics and Communication Engineering Amity School of Engineering, Noida, Uttar Pradesh,INDIA Jyoti Bhardwaj

More information

http://www.springer.com/0-387-23402-0

http://www.springer.com/0-387-23402-0 http://www.springer.com/0-387-23402-0 Chapter 2 VISUAL DATA FORMATS 1. Image and Video Data Digital visual data is usually organised in rectangular arrays denoted as frames, the elements of these arrays

More information

High performance digital video servers: storage. Seungyup Paek and Shih-Fu Chang. Columbia University

High performance digital video servers: storage. Seungyup Paek and Shih-Fu Chang. Columbia University High performance digital video servers: storage and retrieval of compressed scalable video Seungyup Paek and Shih-Fu Chang Department of Electrical Engineering Columbia University New York, N.Y. 10027-6699,

More information

A review of compression methods for medical images in PACS

A review of compression methods for medical images in PACS International Journal of Medical Informatics 52 (1998) 159 165 A review of compression methods for medical images in PACS Michael G. Strintzis * Department of Electrical and Computer Engineering, Aristotle

More information

JPEG Image Compression by Using DCT

JPEG Image Compression by Using DCT International Journal of Computer Sciences and Engineering Open Access Research Paper Volume-4, Issue-4 E-ISSN: 2347-2693 JPEG Image Compression by Using DCT Sarika P. Bagal 1* and Vishal B. Raskar 2 1*

More information

Video compression: Performance of available codec software

Video compression: Performance of available codec software Video compression: Performance of available codec software Introduction. Digital Video A digital video is a collection of images presented sequentially to produce the effect of continuous motion. It takes

More information

White paper. H.264 video compression standard. New possibilities within video surveillance.

White paper. H.264 video compression standard. New possibilities within video surveillance. White paper H.264 video compression standard. New possibilities within video surveillance. Table of contents 1. Introduction 3 2. Development of H.264 3 3. How video compression works 4 4. H.264 profiles

More information

Transform-domain Wyner-Ziv Codec for Video

Transform-domain Wyner-Ziv Codec for Video Transform-domain Wyner-Ziv Codec for Video Anne Aaron, Shantanu Rane, Eric Setton, and Bernd Girod Information Systems Laboratory, Department of Electrical Engineering Stanford University 350 Serra Mall,

More information

Study and Implementation of Video Compression Standards (H.264/AVC and Dirac)

Study and Implementation of Video Compression Standards (H.264/AVC and Dirac) Project Proposal Study and Implementation of Video Compression Standards (H.264/AVC and Dirac) Sumedha Phatak-1000731131- sumedha.phatak@mavs.uta.edu Objective: A study, implementation and comparison of

More information

Video Coding Basics. Yao Wang Polytechnic University, Brooklyn, NY11201 yao@vision.poly.edu

Video Coding Basics. Yao Wang Polytechnic University, Brooklyn, NY11201 yao@vision.poly.edu Video Coding Basics Yao Wang Polytechnic University, Brooklyn, NY11201 yao@vision.poly.edu Outline Motivation for video coding Basic ideas in video coding Block diagram of a typical video codec Different

More information

A System for Capturing High Resolution Images

A System for Capturing High Resolution Images A System for Capturing High Resolution Images G.Voyatzis, G.Angelopoulos, A.Bors and I.Pitas Department of Informatics University of Thessaloniki BOX 451, 54006 Thessaloniki GREECE e-mail: pitas@zeus.csd.auth.gr

More information

Study and Implementation of Video Compression standards (H.264/AVC, Dirac)

Study and Implementation of Video Compression standards (H.264/AVC, Dirac) Study and Implementation of Video Compression standards (H.264/AVC, Dirac) EE 5359-Multimedia Processing- Spring 2012 Dr. K.R Rao By: Sumedha Phatak(1000731131) Objective A study, implementation and comparison

More information

MULTI-RESOLUTION PERCEPTUAL ENCODING FOR INTERACTIVE IMAGE SHARING IN REMOTE TELE-DIAGNOSTICS. Javed I. Khan and D. Y. Y. Yun

MULTI-RESOLUTION PERCEPTUAL ENCODING FOR INTERACTIVE IMAGE SHARING IN REMOTE TELE-DIAGNOSTICS. Javed I. Khan and D. Y. Y. Yun MULTI-RESOLUTION PERCEPTUAL ENCODING FOR INTERACTIVE IMAGE SHARING IN REMOTE TELE-DIAGNOSTICS Javed I. Khan and D. Y. Y. Yun Laboratories of Intelligent and Parallel Systems Department of Electrical Engineering,

More information

Vision based Vehicle Tracking using a high angle camera

Vision based Vehicle Tracking using a high angle camera Vision based Vehicle Tracking using a high angle camera Raúl Ignacio Ramos García Dule Shu gramos@clemson.edu dshu@clemson.edu Abstract A vehicle tracking and grouping algorithm is presented in this work

More information

How To Improve Performance Of The H264 Video Codec On A Video Card With A Motion Estimation Algorithm

How To Improve Performance Of The H264 Video Codec On A Video Card With A Motion Estimation Algorithm Implementation of H.264 Video Codec for Block Matching Algorithms Vivek Sinha 1, Dr. K. S. Geetha 2 1 Student of Master of Technology, Communication Systems, Department of ECE, R.V. College of Engineering,

More information

Understanding Compression Technologies for HD and Megapixel Surveillance

Understanding Compression Technologies for HD and Megapixel Surveillance When the security industry began the transition from using VHS tapes to hard disks for video surveillance storage, the question of how to compress and store video became a top consideration for video surveillance

More information

2695 P a g e. IV Semester M.Tech (DCN) SJCIT Chickballapur Karnataka India

2695 P a g e. IV Semester M.Tech (DCN) SJCIT Chickballapur Karnataka India Integrity Preservation and Privacy Protection for Digital Medical Images M.Krishna Rani Dr.S.Bhargavi IV Semester M.Tech (DCN) SJCIT Chickballapur Karnataka India Abstract- In medical treatments, the integrity

More information

MEDICAL IMAGE COMPRESSION USING HYBRID CODER WITH FUZZY EDGE DETECTION

MEDICAL IMAGE COMPRESSION USING HYBRID CODER WITH FUZZY EDGE DETECTION MEDICAL IMAGE COMPRESSION USING HYBRID CODER WITH FUZZY EDGE DETECTION K. Vidhya 1 and S. Shenbagadevi Department of Electrical & Communication Engineering, College of Engineering, Anna University, Chennai,

More information

Bandwidth Adaptation for MPEG-4 Video Streaming over the Internet

Bandwidth Adaptation for MPEG-4 Video Streaming over the Internet DICTA2002: Digital Image Computing Techniques and Applications, 21--22 January 2002, Melbourne, Australia Bandwidth Adaptation for MPEG-4 Video Streaming over the Internet K. Ramkishor James. P. Mammen

More information

Lossless Medical Image Compression using Redundancy Analysis

Lossless Medical Image Compression using Redundancy Analysis 5 Lossless Medical Image Compression using Redundancy Analysis Se-Kee Kil, Jong-Shill Lee, Dong-Fan Shen, Je-Goon Ryu, Eung-Hyuk Lee, Hong-Ki Min, Seung-Hong Hong Dept. of Electronic Eng., Inha University,

More information

View Sequence Coding using Warping-based Image Alignment for Multi-view Video

View Sequence Coding using Warping-based Image Alignment for Multi-view Video View Sequence Coding using Warping-based mage Alignment for Multi-view Video Yanwei Liu, Qingming Huang,, Wen Gao 3 nstitute of Computing Technology, Chinese Academy of Science, Beijing, China Graduate

More information

Efficient Coding Unit and Prediction Unit Decision Algorithm for Multiview Video Coding

Efficient Coding Unit and Prediction Unit Decision Algorithm for Multiview Video Coding JOURNAL OF ELECTRONIC SCIENCE AND TECHNOLOGY, VOL. 13, NO. 2, JUNE 2015 97 Efficient Coding Unit and Prediction Unit Decision Algorithm for Multiview Video Coding Wei-Hsiang Chang, Mei-Juan Chen, Gwo-Long

More information

FCE: A Fast Content Expression for Server-based Computing

FCE: A Fast Content Expression for Server-based Computing FCE: A Fast Content Expression for Server-based Computing Qiao Li Mentor Graphics Corporation 11 Ridder Park Drive San Jose, CA 95131, U.S.A. Email: qiao li@mentor.com Fei Li Department of Computer Science

More information

Video Coding Standards. Yao Wang Polytechnic University, Brooklyn, NY11201 yao@vision.poly.edu

Video Coding Standards. Yao Wang Polytechnic University, Brooklyn, NY11201 yao@vision.poly.edu Video Coding Standards Yao Wang Polytechnic University, Brooklyn, NY11201 yao@vision.poly.edu Yao Wang, 2003 EE4414: Video Coding Standards 2 Outline Overview of Standards and Their Applications ITU-T

More information

A Short Introduction to Computer Graphics

A Short Introduction to Computer Graphics A Short Introduction to Computer Graphics Frédo Durand MIT Laboratory for Computer Science 1 Introduction Chapter I: Basics Although computer graphics is a vast field that encompasses almost any graphical

More information

Very Low Frame-Rate Video Streaming For Face-to-Face Teleconference

Very Low Frame-Rate Video Streaming For Face-to-Face Teleconference Very Low Frame-Rate Video Streaming For Face-to-Face Teleconference Jue Wang, Michael F. Cohen Department of Electrical Engineering, University of Washington Microsoft Research Abstract Providing the best

More information

Viewer sees picture in 3D. Array of cameras Video multiplexer Autostereoscopic 3D display ❻ ❺ ❹ ❶ ❷❸

Viewer sees picture in 3D. Array of cameras Video multiplexer Autostereoscopic 3D display ❻ ❺ ❹ ❶ ❷❸ Autostereoscopic 3D Display in Laparoscopic Surgery N. A. Dodgson 1, N. E. Wiseman 1, S. R. Lang 1,D.C.Dunn 2,A.R.L.Travis 3 1 Computer Laboratory, Pembroke Street CB2 3QG 2 Department of Surgery, Addenbrooke's

More information

Introduzione alle Biblioteche Digitali Audio/Video

Introduzione alle Biblioteche Digitali Audio/Video Introduzione alle Biblioteche Digitali Audio/Video Biblioteche Digitali 1 Gestione del video Perchè è importante poter gestire biblioteche digitali di audiovisivi Caratteristiche specifiche dell audio/video

More information

Region of Interest Access with Three-Dimensional SBHP Algorithm CIPR Technical Report TR-2006-1

Region of Interest Access with Three-Dimensional SBHP Algorithm CIPR Technical Report TR-2006-1 Region of Interest Access with Three-Dimensional SBHP Algorithm CIPR Technical Report TR-2006-1 Ying Liu and William A. Pearlman January 2006 Center for Image Processing Research Rensselaer Polytechnic

More information

96 T. Kunkelmann, U. Horn

96 T. Kunkelmann, U. Horn Video Encryption Based on Data Partitioning and Scalable Coding - A Comparison Thomas Kunkelmann 1 and Uwe Horn 2 1 Darmstadt University oftechnology, Department of Computer Science Information Technology

More information

Video-Conferencing System

Video-Conferencing System Video-Conferencing System Evan Broder and C. Christoher Post Introductory Digital Systems Laboratory November 2, 2007 Abstract The goal of this project is to create a video/audio conferencing system. Video

More information

REPRESENTATION, CODING AND INTERACTIVE RENDERING OF HIGH- RESOLUTION PANORAMIC IMAGES AND VIDEO USING MPEG-4

REPRESENTATION, CODING AND INTERACTIVE RENDERING OF HIGH- RESOLUTION PANORAMIC IMAGES AND VIDEO USING MPEG-4 REPRESENTATION, CODING AND INTERACTIVE RENDERING OF HIGH- RESOLUTION PANORAMIC IMAGES AND VIDEO USING MPEG-4 S. Heymann, A. Smolic, K. Mueller, Y. Guo, J. Rurainsky, P. Eisert, T. Wiegand Fraunhofer Institute

More information

A HIGH PERFORMANCE SOFTWARE IMPLEMENTATION OF MPEG AUDIO ENCODER. Figure 1. Basic structure of an encoder.

A HIGH PERFORMANCE SOFTWARE IMPLEMENTATION OF MPEG AUDIO ENCODER. Figure 1. Basic structure of an encoder. A HIGH PERFORMANCE SOFTWARE IMPLEMENTATION OF MPEG AUDIO ENCODER Manoj Kumar 1 Mohammad Zubair 1 1 IBM T.J. Watson Research Center, Yorktown Hgts, NY, USA ABSTRACT The MPEG/Audio is a standard for both

More information

Image Compression through DCT and Huffman Coding Technique

Image Compression through DCT and Huffman Coding Technique International Journal of Current Engineering and Technology E-ISSN 2277 4106, P-ISSN 2347 5161 2015 INPRESSCO, All Rights Reserved Available at http://inpressco.com/category/ijcet Research Article Rahul

More information

Adaptive Block Truncation Filter for MVC Depth Image Enhancement

Adaptive Block Truncation Filter for MVC Depth Image Enhancement 2014 IEEE International Conference on Acoustic, Speech and Signal Processing (ICASSP) Adaptive Block Truncation Filter for MVC Depth Image Enhancement Xuyuan Xu 1, Lai-Man Po 1, Chun-Ho Cheung 2, Litong

More information

Fast Arithmetic Coding (FastAC) Implementations

Fast Arithmetic Coding (FastAC) Implementations Fast Arithmetic Coding (FastAC) Implementations Amir Said 1 Introduction This document describes our fast implementations of arithmetic coding, which achieve optimal compression and higher throughput by

More information

Multimedia Data Transmission over Wired/Wireless Networks

Multimedia Data Transmission over Wired/Wireless Networks Multimedia Data Transmission over Wired/Wireless Networks Bharat Bhargava Gang Ding, Xiaoxin Wu, Mohamed Hefeeda, Halima Ghafoor Purdue University Website: http://www.cs.purdue.edu/homes/bb E-mail: bb@cs.purdue.edu

More information

Video Encryption Exploiting Non-Standard 3D Data Arrangements. Stefan A. Kramatsch, Herbert Stögner, and Andreas Uhl uhl@cosy.sbg.ac.

Video Encryption Exploiting Non-Standard 3D Data Arrangements. Stefan A. Kramatsch, Herbert Stögner, and Andreas Uhl uhl@cosy.sbg.ac. Video Encryption Exploiting Non-Standard 3D Data Arrangements Stefan A. Kramatsch, Herbert Stögner, and Andreas Uhl uhl@cosy.sbg.ac.at Andreas Uhl 1 Carinthia Tech Institute & Salzburg University Outline

More information

A Tool for Multimedia Quality Assessment in NS3: QoE Monitor

A Tool for Multimedia Quality Assessment in NS3: QoE Monitor A Tool for Multimedia Quality Assessment in NS3: QoE Monitor D. Saladino, A. Paganelli, M. Casoni Department of Engineering Enzo Ferrari, University of Modena and Reggio Emilia via Vignolese 95, 41125

More information

ENG4BF3 Medical Image Processing. Image Visualization

ENG4BF3 Medical Image Processing. Image Visualization ENG4BF3 Medical Image Processing Image Visualization Visualization Methods Visualization of medical images is for the determination of the quantitative information about the properties of anatomic tissues

More information

Frsq: A Binary Image Coding Method

Frsq: A Binary Image Coding Method Frsq: A Binary Image Coding Method Peter L. Stanchev, William I. Grosky, John G. Geske Kettering University, Flint, MI 4854, {pstanche, jgeske}@kettering.edu University of Michigan-Dearborn, Dearborn,

More information

A comprehensive survey on various ETC techniques for secure Data transmission

A comprehensive survey on various ETC techniques for secure Data transmission A comprehensive survey on various ETC techniques for secure Data transmission Shaikh Nasreen 1, Prof. Suchita Wankhade 2 1, 2 Department of Computer Engineering 1, 2 Trinity College of Engineering and

More information

Overview of the Scalable Video Coding Extension of the H.264/AVC Standard

Overview of the Scalable Video Coding Extension of the H.264/AVC Standard To appear in IEEE Transactions on Circuits and Systems for Video Technology, September 2007. 1 Overview of the Scalable Video Coding Extension of the H.264/AVC Standard Heiko Schwarz, Detlev Marpe, Member,

More information

Quality Estimation for Scalable Video Codec. Presented by Ann Ukhanova (DTU Fotonik, Denmark) Kashaf Mazhar (KTH, Sweden)

Quality Estimation for Scalable Video Codec. Presented by Ann Ukhanova (DTU Fotonik, Denmark) Kashaf Mazhar (KTH, Sweden) Quality Estimation for Scalable Video Codec Presented by Ann Ukhanova (DTU Fotonik, Denmark) Kashaf Mazhar (KTH, Sweden) Purpose of scalable video coding Multiple video streams are needed for heterogeneous

More information

B2.53-R3: COMPUTER GRAPHICS. NOTE: 1. There are TWO PARTS in this Module/Paper. PART ONE contains FOUR questions and PART TWO contains FIVE questions.

B2.53-R3: COMPUTER GRAPHICS. NOTE: 1. There are TWO PARTS in this Module/Paper. PART ONE contains FOUR questions and PART TWO contains FIVE questions. B2.53-R3: COMPUTER GRAPHICS NOTE: 1. There are TWO PARTS in this Module/Paper. PART ONE contains FOUR questions and PART TWO contains FIVE questions. 2. PART ONE is to be answered in the TEAR-OFF ANSWER

More information

Template-based Eye and Mouth Detection for 3D Video Conferencing

Template-based Eye and Mouth Detection for 3D Video Conferencing Template-based Eye and Mouth Detection for 3D Video Conferencing Jürgen Rurainsky and Peter Eisert Fraunhofer Institute for Telecommunications - Heinrich-Hertz-Institute, Image Processing Department, Einsteinufer

More information

Department of Electrical and Computer Engineering, Northwestern University, e-mail: faggk,faisal,lkon,hongg@ece.nwu.edu,

Department of Electrical and Computer Engineering, Northwestern University, e-mail: faggk,faisal,lkon,hongg@ece.nwu.edu, ERROR RESILIENCE AND CONCEALMENT IN VIDEO CODING A. K. Katsaggelos, F. Ishtiaq, L. P. Kondi, M.-C. Hong, M. Banham +,J.Brailean + Department of Electrical and Computer Engineering, Northwestern University,

More information

EECS 556 Image Processing W 09. Interpolation. Interpolation techniques B splines

EECS 556 Image Processing W 09. Interpolation. Interpolation techniques B splines EECS 556 Image Processing W 09 Interpolation Interpolation techniques B splines What is image processing? Image processing is the application of 2D signal processing methods to images Image representation

More information

Clipping Plane. Overview Posterior

Clipping Plane. Overview Posterior Automated 3D Video Documentation for the Analysis of Medical Data S. Iserhardt-Bauer 1, C. Rezk-Salama 2, T. Ertl 1,P. Hastreiter 3,B.Tomandl 4, und K. Eberhardt 4 1 Visualization and Interactive Systems

More information

302 IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, VOL. 19, NO. 2, FEBRUARY 2009

302 IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, VOL. 19, NO. 2, FEBRUARY 2009 302 IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, VOL. 19, NO. 2, FEBRUARY 2009 Transactions Letters Fast Inter-Mode Decision in an H.264/AVC Encoder Using Mode and Lagrangian Cost Correlation

More information

TECHNICAL OVERVIEW OF VP8, AN OPEN SOURCE VIDEO CODEC FOR THE WEB

TECHNICAL OVERVIEW OF VP8, AN OPEN SOURCE VIDEO CODEC FOR THE WEB TECHNICAL OVERVIEW OF VP8, AN OPEN SOURCE VIDEO CODEC FOR THE WEB Jim Bankoski, Paul Wilkins, Yaowu Xu Google Inc. 1600 Amphitheatre Parkway, Mountain View, CA, USA {jimbankoski, paulwilkins, yaowu}@google.com

More information

Hybrid Lossless Compression Method For Binary Images

Hybrid Lossless Compression Method For Binary Images M.F. TALU AND İ. TÜRKOĞLU/ IU-JEEE Vol. 11(2), (2011), 1399-1405 Hybrid Lossless Compression Method For Binary Images M. Fatih TALU, İbrahim TÜRKOĞLU Inonu University, Dept. of Computer Engineering, Engineering

More information

Automatic Labeling of Lane Markings for Autonomous Vehicles

Automatic Labeling of Lane Markings for Autonomous Vehicles Automatic Labeling of Lane Markings for Autonomous Vehicles Jeffrey Kiske Stanford University 450 Serra Mall, Stanford, CA 94305 jkiske@stanford.edu 1. Introduction As autonomous vehicles become more popular,

More information

THE EMERGING JVT/H.26L VIDEO CODING STANDARD

THE EMERGING JVT/H.26L VIDEO CODING STANDARD THE EMERGING JVT/H.26L VIDEO CODING STANDARD H. Schwarz and T. Wiegand Heinrich Hertz Institute, Germany ABSTRACT JVT/H.26L is a current project of the ITU-T Video Coding Experts Group (VCEG) and the ISO/IEC

More information

National Technical University of Athens, 9 Iroon Polytechniou str, GR-15773 Athens, Greece. Email: ndroso@image.ntua.gr

National Technical University of Athens, 9 Iroon Polytechniou str, GR-15773 Athens, Greece. Email: ndroso@image.ntua.gr An Optical Camera Tracking System for Virtual Sets Applications Athanasios Drosopoulos, Yiannis Xirouhakis and Anastasios Delopoulos Computer Science Div., Dept. of Electrical and Computer Eng., National

More information

Conceptual Framework Strategies for Image Compression: A Review

Conceptual Framework Strategies for Image Compression: A Review International Journal of Computer Sciences and Engineering Open Access Review Paper Volume-4, Special Issue-1 E-ISSN: 2347-2693 Conceptual Framework Strategies for Image Compression: A Review Sumanta Lal

More information

Computer Graphics. Geometric Modeling. Page 1. Copyright Gotsman, Elber, Barequet, Karni, Sheffer Computer Science - Technion. An Example.

Computer Graphics. Geometric Modeling. Page 1. Copyright Gotsman, Elber, Barequet, Karni, Sheffer Computer Science - Technion. An Example. An Example 2 3 4 Outline Objective: Develop methods and algorithms to mathematically model shape of real world objects Categories: Wire-Frame Representation Object is represented as as a set of points

More information

Video Authentication for H.264/AVC using Digital Signature Standard and Secure Hash Algorithm

Video Authentication for H.264/AVC using Digital Signature Standard and Secure Hash Algorithm Video Authentication for H.264/AVC using Digital Signature Standard and Secure Hash Algorithm Nandakishore Ramaswamy Qualcomm Inc 5775 Morehouse Dr, Sam Diego, CA 92122. USA nandakishore@qualcomm.com K.

More information

JPEG compression of monochrome 2D-barcode images using DCT coefficient distributions

JPEG compression of monochrome 2D-barcode images using DCT coefficient distributions Edith Cowan University Research Online ECU Publications Pre. JPEG compression of monochrome D-barcode images using DCT coefficient distributions Keng Teong Tan Hong Kong Baptist University Douglas Chai

More information

MPEG-4 Natural Video Coding - An overview

MPEG-4 Natural Video Coding - An overview MPEG-4 Natural Video Coding - An overview Touradj Ebrahimi* and Caspar Horne** *Signal Processing Laboratory **Mediamatics Inc. Swiss Federal Institute of Technology EPFL 48430 Lakeview Blvd 1015 Lausanne,

More information

Overview: Video Coding Standards

Overview: Video Coding Standards Overview: Video Coding Standards Video coding standards: applications and common structure Relevant standards organizations ITU-T Rec. H.261 ITU-T Rec. H.263 ISO/IEC MPEG-1 ISO/IEC MPEG-2 ISO/IEC MPEG-4

More information

Tracking Groups of Pedestrians in Video Sequences

Tracking Groups of Pedestrians in Video Sequences Tracking Groups of Pedestrians in Video Sequences Jorge S. Marques Pedro M. Jorge Arnaldo J. Abrantes J. M. Lemos IST / ISR ISEL / IST ISEL INESC-ID / IST Lisbon, Portugal Lisbon, Portugal Lisbon, Portugal

More information

Efficient Motion Estimation by Fast Three Step Search Algorithms

Efficient Motion Estimation by Fast Three Step Search Algorithms Efficient Motion Estimation by Fast Three Step Search Algorithms Namrata Verma 1, Tejeshwari Sahu 2, Pallavi Sahu 3 Assistant professor, Dept. of Electronics & Telecommunication Engineering, BIT Raipur,

More information

Volume visualization I Elvins

Volume visualization I Elvins Volume visualization I Elvins 1 surface fitting algorithms marching cubes dividing cubes direct volume rendering algorithms ray casting, integration methods voxel projection, projected tetrahedra, splatting

More information

The Essence of Image and Video Compression 1E8: Introduction to Engineering Introduction to Image and Video Processing

The Essence of Image and Video Compression 1E8: Introduction to Engineering Introduction to Image and Video Processing The Essence of Image and Video Compression E8: Introduction to Engineering Introduction to Image and Video Processing Dr. Anil C. Kokaram, Electronic and Electrical Engineering Dept., Trinity College,

More information

International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, Vol. XXXIV-5/W10

International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, Vol. XXXIV-5/W10 Accurate 3D information extraction from large-scale data compressed image and the study of the optimum stereo imaging method Riichi NAGURA *, * Kanagawa Institute of Technology nagura@ele.kanagawa-it.ac.jp

More information

A Novel Method to Improve Resolution of Satellite Images Using DWT and Interpolation

A Novel Method to Improve Resolution of Satellite Images Using DWT and Interpolation A Novel Method to Improve Resolution of Satellite Images Using DWT and Interpolation S.VENKATA RAMANA ¹, S. NARAYANA REDDY ² M.Tech student, Department of ECE, SVU college of Engineering, Tirupati, 517502,

More information

December 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B. KITCHENS

December 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B. KITCHENS December 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B KITCHENS The equation 1 Lines in two-dimensional space (1) 2x y = 3 describes a line in two-dimensional space The coefficients of x and y in the equation

More information

Solving Simultaneous Equations and Matrices

Solving Simultaneous Equations and Matrices Solving Simultaneous Equations and Matrices The following represents a systematic investigation for the steps used to solve two simultaneous linear equations in two unknowns. The motivation for considering

More information

IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, VOL. 8, NO. 4, AUGUST 1998 369

IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, VOL. 8, NO. 4, AUGUST 1998 369 IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, VOL. 8, NO. 4, AUGUST 1998 369 Transactions Letters A Novel Unrestricted Center-Biased Diamond Search Algorithm for Block Motion Estimation

More information

COMPRESSION OF 3D MEDICAL IMAGE USING EDGE PRESERVATION TECHNIQUE

COMPRESSION OF 3D MEDICAL IMAGE USING EDGE PRESERVATION TECHNIQUE International Journal of Electronics and Computer Science Engineering 802 Available Online at www.ijecse.org ISSN: 2277-1956 COMPRESSION OF 3D MEDICAL IMAGE USING EDGE PRESERVATION TECHNIQUE Alagendran.B

More information

Geometric Camera Parameters

Geometric Camera Parameters Geometric Camera Parameters What assumptions have we made so far? -All equations we have derived for far are written in the camera reference frames. -These equations are valid only when: () all distances

More information

Color Segmentation Based Depth Image Filtering

Color Segmentation Based Depth Image Filtering Color Segmentation Based Depth Image Filtering Michael Schmeing and Xiaoyi Jiang Department of Computer Science, University of Münster Einsteinstraße 62, 48149 Münster, Germany, {m.schmeing xjiang}@uni-muenster.de

More information

Information, Entropy, and Coding

Information, Entropy, and Coding Chapter 8 Information, Entropy, and Coding 8. The Need for Data Compression To motivate the material in this chapter, we first consider various data sources and some estimates for the amount of data associated

More information

Copyright 2008 IEEE. Reprinted from IEEE Transactions on Multimedia 10, no. 8 (December 2008): 1671-1686.

Copyright 2008 IEEE. Reprinted from IEEE Transactions on Multimedia 10, no. 8 (December 2008): 1671-1686. Copyright 2008 IEEE. Reprinted from IEEE Transactions on Multimedia 10, no. 8 (December 2008): 1671-1686. This material is posted here with permission of the IEEE. Such permission of the IEEE does not

More information

Geo-referenced 3D Video as visualization and measurement tool for Cultural Heritage

Geo-referenced 3D Video as visualization and measurement tool for Cultural Heritage Geo-referenced 3D Video as visualization and measurement tool for Cultural Heritage Lazaros Sechidis, Vassilios Tsioukas, Petros Patias The Aristotle University of Thessaloniki, Department of Cadastre

More information

Multihypothesis Prediction using Decoder Side Motion Vector Derivation in Inter Frame Video Coding

Multihypothesis Prediction using Decoder Side Motion Vector Derivation in Inter Frame Video Coding Multihypothesis Prediction using Decoder Side Motion Vector Derivation in Inter Frame Video Coding Steffen Kamp, Johannes Ballé, and Mathias Wien Institut für Nachrichtentechnik, RWTH Aachen University,

More information

Introduction to Medical Image Compression Using Wavelet Transform

Introduction to Medical Image Compression Using Wavelet Transform National Taiwan University Graduate Institute of Communication Engineering Time Frequency Analysis and Wavelet Transform Term Paper Introduction to Medical Image Compression Using Wavelet Transform 李 自

More information

Application of Virtual Instrumentation for Sensor Network Monitoring

Application of Virtual Instrumentation for Sensor Network Monitoring Application of Virtual Instrumentation for Sensor etwor Monitoring COSTATI VOLOSECU VICTOR MALITA Department of Automatics and Applied Informatics Politehnica University of Timisoara Bd. V. Parvan nr.

More information

ANALYSIS OF THE EFFECTIVENESS IN IMAGE COMPRESSION FOR CLOUD STORAGE FOR VARIOUS IMAGE FORMATS

ANALYSIS OF THE EFFECTIVENESS IN IMAGE COMPRESSION FOR CLOUD STORAGE FOR VARIOUS IMAGE FORMATS ANALYSIS OF THE EFFECTIVENESS IN IMAGE COMPRESSION FOR CLOUD STORAGE FOR VARIOUS IMAGE FORMATS Dasaradha Ramaiah K. 1 and T. Venugopal 2 1 IT Department, BVRIT, Hyderabad, India 2 CSE Department, JNTUH,

More information

Oscillations of the Sending Window in Compound TCP

Oscillations of the Sending Window in Compound TCP Oscillations of the Sending Window in Compound TCP Alberto Blanc 1, Denis Collange 1, and Konstantin Avrachenkov 2 1 Orange Labs, 905 rue Albert Einstein, 06921 Sophia Antipolis, France 2 I.N.R.I.A. 2004

More information

Television Control by Hand Gestures

Television Control by Hand Gestures MITSUBISHI ELECTRIC RESEARCH LABORATORIES http://www.merl.com Television Control by Hand Gestures William T. Freeman, Craig D. Weissman TR94-24 December 1994 Abstract We study how a viewer can control

More information

How To Fix Out Of Focus And Blur Images With A Dynamic Template Matching Algorithm

How To Fix Out Of Focus And Blur Images With A Dynamic Template Matching Algorithm IJSTE - International Journal of Science Technology & Engineering Volume 1 Issue 10 April 2015 ISSN (online): 2349-784X Image Estimation Algorithm for Out of Focus and Blur Images to Retrieve the Barcode

More information

Compression techniques

Compression techniques Compression techniques David Bařina February 22, 2013 David Bařina Compression techniques February 22, 2013 1 / 37 Contents 1 Terminology 2 Simple techniques 3 Entropy coding 4 Dictionary methods 5 Conclusion

More information

Compression of parametric surfaces for efficient 3D model coding

Compression of parametric surfaces for efficient 3D model coding 1 Compression of parametric surfaces for efficient 3D model coding Diego Santa-Cruz a, Touradj Ebrahimi b Signal Processing Laboratory Swiss Federal Institute of Technology, Lausanne ABSTRACT In the field

More information

Accelerating Wavelet-Based Video Coding on Graphics Hardware

Accelerating Wavelet-Based Video Coding on Graphics Hardware Wladimir J. van der Laan, Andrei C. Jalba, and Jos B.T.M. Roerdink. Accelerating Wavelet-Based Video Coding on Graphics Hardware using CUDA. In Proc. 6th International Symposium on Image and Signal Processing

More information

Wireless Ultrasound Video Transmission for Stroke Risk Assessment: Quality Metrics and System Design

Wireless Ultrasound Video Transmission for Stroke Risk Assessment: Quality Metrics and System Design Wireless Ultrasound Video Transmission for Stroke Risk Assessment: Quality Metrics and System Design A. Panayides 1, M.S. Pattichis 2, C. S. Pattichis 1, C. P. Loizou 3, M. Pantziaris 4 1 A.Panayides and

More information

CHAPTER 2 LITERATURE REVIEW

CHAPTER 2 LITERATURE REVIEW 11 CHAPTER 2 LITERATURE REVIEW 2.1 INTRODUCTION Image compression is mainly used to reduce storage space, transmission time and bandwidth requirements. In the subsequent sections of this chapter, general

More information

Colour Image Encryption and Decryption by using Scan Approach

Colour Image Encryption and Decryption by using Scan Approach Colour Image Encryption and Decryption by using Scan Approach, Rinkee Gupta,Master of Engineering Scholar, Email: guptarinki.14@gmail.com Jaipal Bisht, Asst. Professor Radharaman Institute Of Technology

More information

MPEG-1 and MPEG-2 Digital Video Coding Standards

MPEG-1 and MPEG-2 Digital Video Coding Standards Please note that the page has been produced based on text and image material from a book in [sik] and may be subject to copyright restrictions from McGraw Hill Publishing Company. MPEG-1 and MPEG-2 Digital

More information

Full Interactive Functions in MPEG-based Video on Demand Systems

Full Interactive Functions in MPEG-based Video on Demand Systems Full nteractive Functions in MPEG-based Video on Demand Systems Kostas Psannis Marios Hadjinicolaou Dept of Electronic & Computer Engineering Brunel University Uxbridge, Middlesex, UB8 3PH,UK Argy Krikelis

More information

Lecture 2 Linear functions and examples

Lecture 2 Linear functions and examples EE263 Autumn 2007-08 Stephen Boyd Lecture 2 Linear functions and examples linear equations and functions engineering examples interpretations 2 1 Linear equations consider system of linear equations y

More information