Immersive 3-D Video Conferencing: Challenges, Concepts, and Implementations

Size: px
Start display at page:

Download "Immersive 3-D Video Conferencing: Challenges, Concepts, and Implementations"

Transcription

1 Proc. SPIE Visual Communications and Image Processing (VCIP), Lugano, Switzerland, July Immersive 3-D Video Conferencing: Challenges, Concepts, and Implementations Peter Eisert Image Processing Department Fraunhofer Institute for Telecommunications, Heinrich-Hertz-Institute Einsteinufer 37, D Berlin, Germany URL: eisert ABSTRACT In this paper, a next generation 3-D video conferencing system is presented that provides immersive telepresence and natural representation of all participants in a shared virtual meeting space. The system is based on the principle of a shared virtual table environment which guarantees correct eye contact and gesture reproduction and enhances the quality of human-centered communication. The virtual environment is modeled in MPEG-4 which also allows the seamless integration of explicit 3-D head models for a low-bandwidth connection to mobile users. In this case, facial expression and motion information is transmitted instead of video streams resulting in bit-rates of a few kbit/s per participant. Beside low bit-rates, the model-based approach enables new possibilities for image enhancements like digital make-up, digital dressing, or modification of scene lighting. Keywords: Immersive Video Conferencing, Tele-Collaboration, Shared Virtual Table Environment, Model- Based Coding, View Synthesis 1. INTRODUCTION The idea of video conferencing and video telephony has fascinated researchers for a long time and new devices have been developed as soon as new technologies became available. Already in 1927, a video phone conversation with life video transmission over telephone lines from Washington D.C. to an auditorium in Manhattan was demonstrated. The Bell picturephone shown in Fig. 1 used a mechanical Nipkow Disk for sampling the images and had a display of size 2 by 2.5 inches. Figure 1. Bell s picturephone system of Over time, devices got smaller while display sizes increased. In the sixties, a couple of different video phones were developed which enabled video transmission over telephone lines. The left hand side of Fig. 2 shows an example of such a picture phone. Unfortunately, these technologies never became well accepted due to costs, limited availability, and restricted gain in communication. First with the introduction of digital computers, cheap and powerful hardware and software devices have been developed and, today, video conferencing is rapidly advancing. Especially for decision-making in a wide range of applications, from world-wide operating companies, international research projects to global investment banking, the interest in tele-conferencing has increased in recent years.

2 Figure 2. Left: Picturephone from Right: Hydra system. Nevertheless, video conferencing today is limited in its support of natural human-centered communication. Body postures and subtle movements, gaze direction, room acoustics, and joint interactions are often misrepresented, lacking, or wholly absent. Eye contact is still a problem and also verbal and non-verbal communication between other participants can often not be followed. To address these limitations, several experimental systems have been proposed in the last few years. The Hydra system depicted on the right hand side of Fig. 2, e.g., uses three independent communication devices with three displays distributed on a desk to ensure a spatial and acoustical separation of the participants. Figure 3. Concept of an immersive shared virtual table video conferencing system. Beside the reproduction of spatial relations, large displays 1 and other benefits of tele-immersion are exploited in order to give the participants the impression of being present in a shared virtual table environment (SVTE) suggesting spatial proximity and enabling a higher degree of natural interaction and effective collaboration. In this sense, the participants should be able to make use of rich communication modalities as similar as possible to those used in a face-to-face meeting. The exact reproduction of gestures, eye contact, sound direction, and multiple simultaneous inter-participant conversations are examples of these requirements. The main objective is to enable conferees located in different geographical places to meet around a virtual table, appearing at each

3 station in such a way to create a convincing impression of presence as illustrated in Fig. 3. First approaches in this area were limited to strictly graphical environments using simple avatars for representing the remote conferees. A lot of systems have been proposed in this context during the last decade. One example for these research activities was the former Europe ACTS project COVEN that demonstrated the benefits of the SVTE concept by a networked business game VR application. 2 Later approaches were more and more focused on a seamless integration of 2-D video images or even 3-D video avatars into Collaborative Virtual Environments (CVE) to increase realism. 3, 4 Some of these approaches were driven by the MPEG-4 multimedia standard that offers new powerful coding and composition tools for such purposes. 5 Figure 4. Left: Setup of NTII at UNC. 6 Right: TELEPORT System. 7 An attractive SVTE approach is the one of tele-cubicles. 6 8 A common feature of these proposals is a special system set-up, where the participants are situated symmetrically around the shared table, with each conferee appearing on an own screen as illustrated in Fig. 4. Note that the symmetric geometry of this setup guarantees eye contact, gaze awareness and gesture reproduction. Thus, everybody in the session can observe who is talking to whom or who is pointing at what. In the next section, the Virtual Team User Environments (VIRTUE) system 9 is proposed that offers all benefits of the tele-cubicle approach and extends it towards an integration into shared virtual working spaces. A seamless transition between the real working desk in front of the display and the virtual conference table at the screen gives the user the impression of being part of an extended perception space. Moreover, the remote participants are rendered seamlessly und under correct perspective view into the virtual conference scene. Scene depth information obtained from stereo cameras and encoded and streamed using MPEG-4 10 enables view synthesis in a virtual 3-D scene. Replacing the depth maps by explicit 3-D computer graphics head-and-shoulder models leads to an enormous reduction in bit-rate for the transmission of scene data, since shape and texture data need not be updated for each frame. Such a model-based coding system, 11 which can be used as a fall-back solution for mobile devices connected to the Internet over low-bandwidth channels, is presented in Section VIRTUAL TEAM USER ENVIRONMENT In this section, we present an extension to the Virtual Team User Environments (VIRTUE) system 9, 12 that uses the state-of-the-art multimedia standard MPEG-4 for transmission and representation of scene and texture data. Due to usage of the MPEG-4 signaling protocol DMIF, a sophisticated 3-D video processing, and the MPEG-4 scene description language BIFS, the system architecture supports an arbitrary terminal setup, interoperability with other terminal configurations, and scalability in terms of the number of participants and it makes it open for future extensions and developments towards immersive tele-collaboration, multi-user terminals, and joint applications.

4 2.1. Shared Virtual Table Concept The basic idea of the shared virtual environment concept is to place 3-D video reproductions of a given number of participants at predefined positions in a shared virtual environment. For this purpose, the conferees are captured at each terminal by a multiple camera setup as shown in Fig. 5 and the desired 3-D video representation of the local person is extracted from the images. Figure 5. Left: Setup for a 3-party conference. Right: VIRTUE setup. Then, the 3-D video objects of all conferees are grouped virtually around the shared table. Ideally, this is done in an isotropic manner in order to obtain social and geometric symmetry. Hence, in the case of a 3-party conference, the participants form an equilateral triangle. In the case of four parties, it would be a square, and so on. Following such generic composition rules and knowing the number of participants, the same SVTE can be built at each terminal from previously loaded scene descriptions and the 3-D video streams. Figure 6. Rendering of the virtual 3-D conference scene. Based on this generic scene composition, individual views of the virtual conference environment can be rendered by using a virtual camera as illustrated in Fig. 6. Locally, the position of the virtual camera has to move coincidently with the current position of the conferee s head, which is permanently registered by a head tracker. Thus, supposing that the geometrical parameters of the multi-view capture device, the virtual scene, and the virtual camera are well fitted to each other, it is ensured that all conferees see the scene under the correct perspective view, even while changing their own viewing position. This geometrical coincidence provides all desired attributes mentioned in the introduction eye contact, gaze awareness, gesture reproduction, natural conference situation, and high amount of realism. In addition, the support of head motion parallax allows the conferees to change their viewing position in order to watch the scene from another perspective, to look behind objects, or to look at a previously occluded object.

5 2.2. Architecture of the System Fig. 7 outlines the system architecture of the 3-D video conferencing system. After multi-view capturing, the video frames are segmented to separate the person s silhouette from the background. As a result, the conferees are represented by arbitrarily shaped video objects, which can be integrated seamlessly into virtual environments. To extract depth information, disparity estimation is performed on the rectified video objects resulting in dense disparity maps. Both video objects and disparity maps are efficiently encoded using MPEG-4. Figure 7. Architecture of the3-d video conference system. The system concept takes advantage of several particular features of the MPEG-4 multimedia standard. MPEG-4 allows the encoding of arbitrarily shaped video objects and provides auxiliary alpha planes to transmit additional pixel information associated to the color data. These additional planes can be used for a joint transmission of disparity maps and video objects. After encoding, the packets are streamed to other participating terminals via RTP. Simultaneously, the terminal receives video streams from the other conference partners and decodes them with multiple MPEG-4 video decoders. The shaped video objects and audio data are synchronized to each other and integrated into the SVTE scene represented by the MPEG-4 scene description language BIFS. Finally, an MPEG-4 compositor is used for rendering purposes. This compositor is able to handle user events (head tracker or interaction with scene content) as well as scene updates sent by a server. The MPEG-4 video objects are three-dimensionally warped using image-based rendering techniques before they are integrated into the scene. Depending on the current input from the head tracker, the correct perspective view is calculated and the adapted view is inserted into the BIFS scene as a 2-D video Foreground Background Segmentation In the finally rendered scene, the participants appear as arbitrarily shaped video objects, seamlessly integrated into the virtual environment. This requires a segmentation of the moving person from the background which is assumed to remain rather static. Initially, the background is captured and a change detection scheme compares this reference image with the current video data and provides a segmentation mask. 5 The reference image is permanently updated to cope with slight changes of illumination or scene content. This baseline algorithm has been improved in speed and quality in order to meet the real-time constraints for full CCIR601 resolution and video frame rate. Its performance has further been improved by adding a fast and efficient shadow detection tool. 13 For a robust segmentation, this is particularly important, because shadows at the table can usually not be avoided, even under optimal illumination conditions. The effect of shadow detection on the segmentation result is shown in Fig. 8.

6 Figure 8. Left: Original frame. Middle: Segmented foreground without shadow detection. Right: with shadow detection Disparity Estimation and Depth Analysis To extract the required 3-D representation, the depth of the captured video object is analyzed by disparity matching on the basis of the rectified images. We have developed a new hybrid block- and pixel-recursive approach, which is able to compute the disparity fields in real-time on a state-of-the-art PC. 14 Apart from a considerable reduction of computational load, the algorithm leads to spatio-temporally consistent depth maps, particularly important for view synthesis, since temporal inconsistencies may cause annoying artifacts in the synthesized views. In order to deal with occlusions caused by hands, arms, or the head, several post-processing techniques are applied. In a first step, critical regions are detected by a consistency check between disparity fields for a left to right and a right to left match. Unreliable disparity values in occluded areas are filled efficiently by suitable extrapolation techniques. The segmentation of the hands and arms which cause severe depth discontinuities is further refined by tracking and segmenting the hands using motion and skin color information. Fig. 9 shows two rectified images, the disparity fields with occluded areas, and the final disparity map after post-processing. Figure and 2. Image: Rectified frames from the left and right camera. occluded areas. 5. Image: Disparity map after post-processing. 3. and 4. Image: Disparity fields with 2.5. Head Tracking The perspective view of the scene presented on the display depends on the viewer s position. This requires an accurate estimation of the viewer s position in the 3-D space, which is accomplished by the head-tracking module. The chosen approach is based on a skin color segmentation technique jointly with a facial feature tracker searching eye positions. Due to the restriction to video conference scenarios, several assumptions can be made to facilitate the tracking of the viewer s head. The obtained 2-D positions of the eyes in two images can then be used for an accurate calculation of the 3-D head position View Synthesis and Virtual Scene Composition The calculated head position and the analyzed depth of the transmitted video objects provide sufficient information to synthesize novel virtual views of the remote conferees. In order to meet real-time constraints 15, 16 and

7 efficient occlusion handling, a new view synthesis algorithm has been developed in the VIRTUE project which is able to take these issues into account. 17 Figure 10. Composed scene containing computer graphics and video objects. At the end of the processing chain, the virtual conference scene has to be composed and displayed onto the screen as depicted in Fig. 10. The virtual scene is represented by a number of polygons in 3-D space and encoded by the BIFS scene description language of MPEG-4. In this polygon-based scene representation the participants are substituted by 2-D rectangles positioned around the virtual table. In this sense, the synthesized novel views of the remote participants are treated as textures and are transferred directly to the graphics card D MODEL-BASED VIDEO CONFERENCING In the previous section, the participants are described by planar computer graphics objects. Onto these planes, the recorded video is projected after warping according to the estimated disparity maps. This has the advantage of very natural looking results, since the original video is exploited in the image-based view synthesis. On the other hand, video streams from each participant together with their corresponding disparity maps have to be transmitted to the client. Dependant on the number of participants, this results in rather high bit-rates even if the streams are encoded with MPEG-4. Current wireless channels are usually not able to handle such bit-rates which prevents the participation of users from mobile devices like PDAs or smart phones. Figure 11. Virtual Video Conferencing on a Mobile Device.

8 In this section, we propose an alternative model-based approach 11, 18, 19 that can be used as a fall-back solution for virtual conferencing systems which are connected to the Internet via low-bandwidth channels. Instead of transmitting texture and depth information of all conferees for each frame, 3-D models of the participants are created representing their appearance. The models are generated only once in an initial phase, encoded, and transmitted to all devices. During the video conference, motion and facial expressions of the conferees are tracked over time and only a small set of facial animation parameters has to be streamed over the network. Experiments have shown, that bit-rates of a few kbit/s per participant can be achieved which enables virtual video conferencing also on mobile devices as illustrated in Fig. 11. The decoder only has to move and deform the 3-D head-andshoulder models in the virtual conferencing scene and render it using computer graphics techniques. Rendering and the encoding and streaming of 3-D head models and facial animation parameters (FAPs) can be realized with MPEG-4 in a standard compliant way. 10 Thus, the compositor of the aforementioned immersive video conferencing system can incorporate these explicit models with only minor changes. The architecture of such a model-based coding component with image analysis and view synthesis is depicted in Fig. 12. Video Analysis Estimation of FAPs Coder Parameter entropy coding Channel about 1 kbit/s Parameter decoding Decoder Synthesis Animation and rendering of the head model Video Head Model Shape Texture Dynamics Figure 12. Model-based Video Codec. Besides the low bit-rate, the usage of explicit 3-D models for the people in the scene also has some other implications on the virtual conferencing system. A sophisticated real-time segmentation of the conferees can be avoided, since the computer models already provide an inherent segmentation. Moreover, view point transformations can easily be applied by simply moving the virtual camera, since all objects in the scene are represented by 3-D models and no image-based warping is required. Another problem in mobile virtual conferencing is an eventually unfavorable appearance or dressing of the conferee depended on his or her current location or situation. By using pre-stored 3-D models for transmission from the mobile device to the other participants, the desired look can be ensured even under awkward situations. This way, digital dressing and digital make-up can be easily applied Facial Expression Analysis The most challenging part of facial analysis and synthesis is the estimation of 3-D facial motion and deformation from two-dimensional images. Due to the loss of one dimension caused by the projection of the real world onto the image plane, this task can only be solved by exploiting additional knowledge of the objects in the scene. In particular, the way the objects move can often be restricted to a low number of degrees of freedom that can be described by a limited set of parameters. We have developed a new 3-D model-based method for the estimation of facial expressions that makes use of an explicit parameterized 3-D human head model describing shape, color, and motion constraints of an individual person. 20 This model information is jointly exploited with spatial and temporal intensity gradients of the images. Thus, the entire area of the image showing the object of interest is used instead of dealing with discrete feature points, resulting in a robust and highly accurate system for modelbased estimation of MPEG-4 facial animation parameters. A linear and computationally efficient algorithm can

9 be derived for different scenarios. The scheme is embedded in a hierarchical analysis-synthesis framework to avoid error accumulation in the long-term estimation Illumination Analysis For natural and uncontrolled video capture conditions, often occurring in mobile environments, scene lighting often varies over time. This illumination variability not only has a considerable influence on the visual appearance of the objects in the scene, but also on the performance of computer vision algorithms or video coding methods. The efficiency and robustness of these algorithms can be significantly improved by removing the undesired effects of changing illumination. In our facial analysis and synthesis system, we model also lighting properties in the scene and estimate them. 21 The explicit knowledge about light sources and reflection properties not only increases robustness of the facial expression analysis but also enables new possibilities for improvement or manipulation of lighting conditions. For example, variations in illumination can be removed in a post-processing step and the lighting conditions can be adjusted to the one in the virtual conference room. This is illustrated in Fig. 13, where the upper row refers to the original image sequence while the lower row depicts the corresponding frames after lighting change estimation and removal. Figure 13. Upper row: original sequence, lower row: corresponding illumination-compensated frames with constant lighting Model-based View Synthesis Synthesizing new views is accomplished by simply rendering the 3-D models in the virtual scene. MPEG-4 offers possibilities to describe head models as well as streaming of facial animation parameters. The color of the person is defined by a texture map which is created from the video frames. For the definition of the individual shape, multiple options are available in MPEG-4. We use our own generic triangle mesh which is adapted to the individual and defined using the scene description BIFS. A facial animation table (FAT) 10 specifies the mesh deformations and transforms which occur when changing a specific facial animation parameter. After the virtual scene, defined by the BIFS tree, is transmitted to all participants of the video conference, only facial animation parameters are sent for each frame after having been estimated from the camera images. Bit-rate measurements have shown, 20 that about 1 kbit/s is sufficient to stream these parameters. Fig. 14 shows an example for the proposed model-based video streaming system. The upper row depicts three frames from a head-and-shoulder video sequence of one participant in a conference. From this monocular image sequence, the 3-D head model shown in the middle row of Fig. 14 is created and facial expression analysis is performed. The estimated parameters are encoded and sent to the decoder which renders the models after

10 Figure 14. Upper row: sequence Peter, middle row: animated wireframe model, lower row: synthesized model frames generated by rendering the 3-D head model. applying local deformation and motion. The middle and lower row of Fig. 14 refer to the output of the decoder after rendering as a wire-frame or with a textured model. 4. CONCLUSIONS We have presented a new concept for immersive video conferencing using a shared virtual table environment (SVTE) which allows a very intensive communication and effective collaboration. It combines the benefits of former VR-based collaborative virtual environment approaches and video-based tele-cubicle systems with those of mixed reality applications. A seamless transition between the real and virtual conference table gives the user the impression of an extended perception space. The usage of 3-D video objects for the representation of remote conferees enables natural representation of gestures, eye contact and gaze. Due to the MPEG-4 multimedia standard, the system provides open and flexible system architectures. Beside arbitrarily shaped video objects, the MPEG-4 compositor allows the rendering of explicit head models enabling the usage of model-based coding techniques as an alternative for communication with devices connected via low-bandwidth channels. The concept is therefore scalable in the number of participants and channel bit-rates, and can be extended towards future applications of immersive tele-collaboration as well as auto-stereoscopic single- or multi-user displays. REFERENCES 1. M. Chen, Design of a virtual auditorium, in Proc. of ACM Multimedia, (Ottawa, Canada), Sep S. Benford, C. Greenhalgh, and D. Lloyd, Crowded collaborative virtual environments, in Proc. of the ACM SIGCHI Conference on Human Factors in Computing Systems, pp , (Atlanta, USA), Mar O. Ståhl, B. Serenius, and G. Taxén, Meetings for real - experiences from a series of VR-based project meetings, in Proc. Symposium on Virtual Reality Software and Technology, (London, UK), Dec

11 4. V. Rajan, S. Subramanian, D. Keenan, A. Johnson, D. Sandin, and T. DeFanti, A realistic video avatar system for networked virtual environments, in Proc. Immersive Projection Technology Symposium, (Orlando, USA), Mar S. Rauthenberg, A. Graffunder, U. Kowalik, and P. Kauff, Virtual shop and virtual meeting point - two prototype applications of interactive services using the new multimedia coding standard MPEG-4, in Proc. International Conference on Computer Communication, (Tokyo, Japan), Sep W. Chen, H. Towles, L. Nyland, G. Welch, and H. Fuchs, Toward a compelling sensation of telepresence: Demonstrating a portal to a distant (static) office, in Proc. Visualization 2000, pp , (Salt Lake City, USA), Oct S. J. Gibbs, C. Arapis, and C. Breiteneder, TELEPORT - Towards immersive copresence, Multimedia Systems, T. Aoki, W. Kustarto, N. Sakamoto, N. Suzuki, K. Saburi, and H. Yasuda, MONJUnoCHIE system: Videoconference system with eye contact for decision making, in Proc. International Workshop on Advanced Image Technology (IWAIT), P. Kauff and O. Schreer, Virtual team user environments - a step from tele-cubicles towards distributed tele-collaboration in mediated workspaces, in Proc. International Conference on Multimedia and Expo (ICME), (Lausanne, Switzerland), Aug ISO/IEC FDIS , Generic Coding of audio-visual objects: (MPEG-4 video), Final Draft International Standard, Document N2502, D. E. Pearson, Developments in model-based video coding, Proceedings of the IEEE 83, pp , June O. Schreer and P. Kauff, An immersive 3D video-conferencing system using shared virtual team user environments, in ACM Collaborative Environments, CVE 2002, (Bonn, Germany), O. Schreer, I. Feldmann, U. Gölz, and P. Kauff, Fast and robust shadow detection in videoconference applications, in Proc. of VIPromCom, 4th EURASIP IEEE International Symposium on Video Processing and Multimedia Communications, (Zadar, Croatia), June O. Schreer, N. Brandenburg, and P. Kauff, Real-time disparity analysis for applications in imersive teleconference scenarios - a comparative study, in Proc. ICIAP 01, (Palermo, Italy), Sep S. Laveau and O. D. Faugeras, 3D scene representation as a collection of images, in Proc. International Conference on Pattern Recognition (ICPR), pp , (Jerusalem, Israel), Oct S. Avidan and A. Shashua, Novel view synthesis in tensor space, in Proc. Computer Vision and Pattern Recognition, pp , (Puerto Rico), June B. J. Lei and E. A. Hendriks, Multi-step view synthesis with occlusion handling, in Proc. Vision, Modeling, and Visualization VMV 02, (Stuttgart, Germany), Nov W. J. Welsh, S. Searsby, and J. B. Waite, Model-based image coding, British Telecom Technology Journal 8, pp , Jul P. Eisert and B. Girod, Analyzing facial expressions for virtual conferencing, IEEE Computer Graphics and Applications 18, pp , Sep P. Eisert, Very Low Bit-Rate Video Coding Using 3-D Models. PhD thesis, University of Erlangen, Shaker Verlag, Aachen, Germany, P. Eisert and B. Girod, Model-based enhancement of lighting conditions in image sequences, in Proc. SPIE Visual Communications and Image Processing, VCIP-02, (San Jose, USA), Jan

VIRTUAL VIDEO CONFERENCING USING 3D MODEL-ASSISTED IMAGE-BASED RENDERING

VIRTUAL VIDEO CONFERENCING USING 3D MODEL-ASSISTED IMAGE-BASED RENDERING VIRTUAL VIDEO CONFERENCING USING 3D MODEL-ASSISTED IMAGE-BASED RENDERING Peter Eisert Fraunhofer Institute for Telecommunications, Heinrich-Hertz-Institute Image Processing Department Einsteinufer 37,

More information

Template-based Eye and Mouth Detection for 3D Video Conferencing

Template-based Eye and Mouth Detection for 3D Video Conferencing Template-based Eye and Mouth Detection for 3D Video Conferencing Jürgen Rurainsky and Peter Eisert Fraunhofer Institute for Telecommunications - Heinrich-Hertz-Institute, Image Processing Department, Einsteinufer

More information

VIRTUE The step towards immersive telepresence in virtual video-conference systems

VIRTUE The step towards immersive telepresence in virtual video-conference systems VIRTUE The step towards immersive telepresence in virtual video-conference systems Oliver SCHREER (HHI) 1 and Phil SHEPPARD (British Telecom) 2 1 Heinrich-Hertz-Institut, Einsteinufer 37, D-10587 Berlin,

More information

REPRESENTATION, CODING AND INTERACTIVE RENDERING OF HIGH- RESOLUTION PANORAMIC IMAGES AND VIDEO USING MPEG-4

REPRESENTATION, CODING AND INTERACTIVE RENDERING OF HIGH- RESOLUTION PANORAMIC IMAGES AND VIDEO USING MPEG-4 REPRESENTATION, CODING AND INTERACTIVE RENDERING OF HIGH- RESOLUTION PANORAMIC IMAGES AND VIDEO USING MPEG-4 S. Heymann, A. Smolic, K. Mueller, Y. Guo, J. Rurainsky, P. Eisert, T. Wiegand Fraunhofer Institute

More information

Peter Eisert, Thomas Wiegand and Bernd Girod. University of Erlangen-Nuremberg. Cauerstrasse 7, 91058 Erlangen, Germany

Peter Eisert, Thomas Wiegand and Bernd Girod. University of Erlangen-Nuremberg. Cauerstrasse 7, 91058 Erlangen, Germany RATE-DISTORTION-EFFICIENT VIDEO COMPRESSION USING A 3-D HEAD MODEL Peter Eisert, Thomas Wiegand and Bernd Girod Telecommunications Laboratory University of Erlangen-Nuremberg Cauerstrasse 7, 91058 Erlangen,

More information

REMOTE RENDERING OF COMPUTER GAMES

REMOTE RENDERING OF COMPUTER GAMES REMOTE RENDERING OF COMPUTER GAMES Peter Eisert, Philipp Fechteler Fraunhofer Institute for Telecommunications, Einsteinufer 37, D-10587 Berlin, Germany eisert@hhi.fraunhofer.de, philipp.fechteler@hhi.fraunhofer.de

More information

Immersive Medien und 3D-Video

Immersive Medien und 3D-Video Fraunhofer-Institut für Nachrichtentechnik Heinrich-Hertz-Institut Ralf Schäfer schaefer@hhi.de http://ip.hhi.de Immersive Medien und 3D-Video page 1 Outline Immersive Media Examples Interactive Media

More information

Analyzing Facial Expressions for Virtual Conferencing

Analyzing Facial Expressions for Virtual Conferencing IEEE Computer Graphics & Applications, pp. 70-78, September 1998. Analyzing Facial Expressions for Virtual Conferencing Peter Eisert and Bernd Girod Telecommunications Laboratory, University of Erlangen,

More information

Making Machines Understand Facial Motion & Expressions Like Humans Do

Making Machines Understand Facial Motion & Expressions Like Humans Do Making Machines Understand Facial Motion & Expressions Like Humans Do Ana C. Andrés del Valle & Jean-Luc Dugelay Multimedia Communications Dpt. Institut Eurécom 2229 route des Crêtes. BP 193. Sophia Antipolis.

More information

BUILDING TELEPRESENCE SYSTEMS: Translating Science Fiction Ideas into Reality

BUILDING TELEPRESENCE SYSTEMS: Translating Science Fiction Ideas into Reality BUILDING TELEPRESENCE SYSTEMS: Translating Science Fiction Ideas into Reality Henry Fuchs University of North Carolina at Chapel Hill (USA) and NSF Science and Technology Center for Computer Graphics and

More information

M3039 MPEG 97/ January 1998

M3039 MPEG 97/ January 1998 INTERNATIONAL ORGANISATION FOR STANDARDISATION ORGANISATION INTERNATIONALE DE NORMALISATION ISO/IEC JTC1/SC29/WG11 CODING OF MOVING PICTURES AND ASSOCIATED AUDIO INFORMATION ISO/IEC JTC1/SC29/WG11 M3039

More information

Talking Head: Synthetic Video Facial Animation in MPEG-4.

Talking Head: Synthetic Video Facial Animation in MPEG-4. Talking Head: Synthetic Video Facial Animation in MPEG-4. A. Fedorov, T. Firsova, V. Kuriakin, E. Martinova, K. Rodyushkin and V. Zhislina Intel Russian Research Center, Nizhni Novgorod, Russia Abstract

More information

Eye-contact in Multipoint Videoconferencing

Eye-contact in Multipoint Videoconferencing Eye-contact in Multipoint Videoconferencing Birgit Quante and Lothar Mühlbach Heinrich-Hertz-Institut für Nachrichtentechnik Berlin GmbH (HHI) Einsteinufer 37, D-15087 Berlin, Germany, http://www.hhi.de/

More information

Virtual Environments - Basics -

Virtual Environments - Basics - Virtual Environments - Basics - What Is Virtual Reality? A Web-Based Introduction Version 4 Draft 1, September, 1998 Jerry Isdale http://www.isdale.com/jerry/vr/whatisvr.html Virtual Environments allow

More information

Quality of Service Management for Teleteaching Applications Using the MPEG-4/DMIF

Quality of Service Management for Teleteaching Applications Using the MPEG-4/DMIF Quality of Service Management for Teleteaching Applications Using the MPEG-4/DMIF Gregor v. Bochmann and Zhen Yang University of Ottawa Presentation at the IDMS conference in Toulouse, October 1999 This

More information

How To Compress Video For Real Time Transmission

How To Compress Video For Real Time Transmission University of Edinburgh College of Science and Engineering School of Informatics Informatics Research Proposal supervised by Dr. Sethu Vijayakumar Optimized bandwidth usage for real-time remote surveillance

More information

White paper. H.264 video compression standard. New possibilities within video surveillance.

White paper. H.264 video compression standard. New possibilities within video surveillance. White paper H.264 video compression standard. New possibilities within video surveillance. Table of contents 1. Introduction 3 2. Development of H.264 3 3. How video compression works 4 4. H.264 profiles

More information

A Prototype For Eye-Gaze Corrected

A Prototype For Eye-Gaze Corrected A Prototype For Eye-Gaze Corrected Video Chat on Graphics Hardware Maarten Dumont, Steven Maesen, Sammy Rogmans and Philippe Bekaert Introduction Traditional webcam video chat: No eye contact. No extensive

More information

How to Send Video Images Through Internet

How to Send Video Images Through Internet Transmitting Video Images in XML Web Service Francisco Prieto, Antonio J. Sierra, María Carrión García Departamento de Ingeniería de Sistemas y Automática Área de Ingeniería Telemática Escuela Superior

More information

Conference interpreting with information and communication technologies experiences from the European Commission DG Interpretation

Conference interpreting with information and communication technologies experiences from the European Commission DG Interpretation Jose Esteban Causo, European Commission Conference interpreting with information and communication technologies experiences from the European Commission DG Interpretation 1 Introduction In the European

More information

FRAUNHOFER INSTITUTE FOR INTEGRATED CIRCUITS IIS AUDIO COMMUNICATION ENGINE RAISING THE BAR IN COMMUNICATION QUALITY

FRAUNHOFER INSTITUTE FOR INTEGRATED CIRCUITS IIS AUDIO COMMUNICATION ENGINE RAISING THE BAR IN COMMUNICATION QUALITY FRAUNHOFER INSTITUTE FOR INTEGRATED CIRCUITS IIS AUDIO COMMUNICATION ENGINE RAISING THE BAR IN COMMUNICATION QUALITY BENEFITS HIGHEST AUDIO QUALITY FOR NEXT GENERATION COMMU- NICATION SYSTEMS Communication

More information

Bandwidth Adaptation for MPEG-4 Video Streaming over the Internet

Bandwidth Adaptation for MPEG-4 Video Streaming over the Internet DICTA2002: Digital Image Computing Techniques and Applications, 21--22 January 2002, Melbourne, Australia Bandwidth Adaptation for MPEG-4 Video Streaming over the Internet K. Ramkishor James. P. Mammen

More information

Very Low Frame-Rate Video Streaming For Face-to-Face Teleconference

Very Low Frame-Rate Video Streaming For Face-to-Face Teleconference Very Low Frame-Rate Video Streaming For Face-to-Face Teleconference Jue Wang, Michael F. Cohen Department of Electrical Engineering, University of Washington Microsoft Research Abstract Providing the best

More information

Introduction to Computer Graphics

Introduction to Computer Graphics Introduction to Computer Graphics Torsten Möller TASC 8021 778-782-2215 torsten@sfu.ca www.cs.sfu.ca/~torsten Today What is computer graphics? Contents of this course Syllabus Overview of course topics

More information

An Animation Definition Interface Rapid Design of MPEG-4 Compliant Animated Faces and Bodies

An Animation Definition Interface Rapid Design of MPEG-4 Compliant Animated Faces and Bodies An Animation Definition Interface Rapid Design of MPEG-4 Compliant Animated Faces and Bodies Erich Haratsch, Technical University of Munich, erich@lis.e-tecknik.tu-muenchen.de Jörn Ostermann, AT&T Labs

More information

One-Way Pseudo Transparent Display

One-Way Pseudo Transparent Display One-Way Pseudo Transparent Display Andy Wu GVU Center Georgia Institute of Technology TSRB, 85 5th St. NW Atlanta, GA 30332 andywu@gatech.edu Ali Mazalek GVU Center Georgia Institute of Technology TSRB,

More information

Modelling 3D Avatar for Virtual Try on

Modelling 3D Avatar for Virtual Try on Modelling 3D Avatar for Virtual Try on NADIA MAGNENAT THALMANN DIRECTOR MIRALAB UNIVERSITY OF GENEVA DIRECTOR INSTITUTE FOR MEDIA INNOVATION, NTU, SINGAPORE WWW.MIRALAB.CH/ Creating Digital Humans Vertex

More information

Cloud-Empowered Multimedia Service: An Automatic Video Storytelling Tool

Cloud-Empowered Multimedia Service: An Automatic Video Storytelling Tool Cloud-Empowered Multimedia Service: An Automatic Video Storytelling Tool Joseph C. Tsai Foundation of Computer Science Lab. The University of Aizu Fukushima-ken, Japan jctsai@u-aizu.ac.jp Abstract Video

More information

ARIB STD-T64-C.S0042 v1.0 Circuit-Switched Video Conferencing Services

ARIB STD-T64-C.S0042 v1.0 Circuit-Switched Video Conferencing Services ARIB STD-T-C.S00 v.0 Circuit-Switched Video Conferencing Services Refer to "Industrial Property Rights (IPR)" in the preface of ARIB STD-T for Related Industrial Property Rights. Refer to "Notice" in the

More information

Text Localization & Segmentation in Images, Web Pages and Videos Media Mining I

Text Localization & Segmentation in Images, Web Pages and Videos Media Mining I Text Localization & Segmentation in Images, Web Pages and Videos Media Mining I Multimedia Computing, Universität Augsburg Rainer.Lienhart@informatik.uni-augsburg.de www.multimedia-computing.{de,org} PSNR_Y

More information

Bernice E. Rogowitz and Holly E. Rushmeier IBM TJ Watson Research Center, P.O. Box 704, Yorktown Heights, NY USA

Bernice E. Rogowitz and Holly E. Rushmeier IBM TJ Watson Research Center, P.O. Box 704, Yorktown Heights, NY USA Are Image Quality Metrics Adequate to Evaluate the Quality of Geometric Objects? Bernice E. Rogowitz and Holly E. Rushmeier IBM TJ Watson Research Center, P.O. Box 704, Yorktown Heights, NY USA ABSTRACT

More information

Communication Door: Real-Time Communication Middleware

Communication Door: Real-Time Communication Middleware Special Issue Advanced Technologies and Solutions toward Ubiquitous Network Society Communication Door: Real-Time Communication Middleware By Masahiro TABUCHI,* Kazuaki NAKAJIMA,* Akiko KAWAMOTO,* Takashi

More information

UHD als Format-agnostisches Produktionsverfahren und als eine Plattform für interaktives Video

UHD als Format-agnostisches Produktionsverfahren und als eine Plattform für interaktives Video UHD als Format-agnostisches Produktionsverfahren und als eine Plattform für interaktives Video, Einsteinufer 37, 10587 Berlin, www.hhi.fraunhofer.de 1 Immersive Telepresence Systems Immersive sound Immersive

More information

BACnet for Video Surveillance

BACnet for Video Surveillance The following article was published in ASHRAE Journal, October 2004. Copyright 2004 American Society of Heating, Refrigerating and Air-Conditioning Engineers, Inc. It is presented for educational purposes

More information

Study and Implementation of Video Compression Standards (H.264/AVC and Dirac)

Study and Implementation of Video Compression Standards (H.264/AVC and Dirac) Project Proposal Study and Implementation of Video Compression Standards (H.264/AVC and Dirac) Sumedha Phatak-1000731131- sumedha.phatak@mavs.uta.edu Objective: A study, implementation and comparison of

More information

Character Animation from 2D Pictures and 3D Motion Data ALEXANDER HORNUNG, ELLEN DEKKERS, and LEIF KOBBELT RWTH-Aachen University

Character Animation from 2D Pictures and 3D Motion Data ALEXANDER HORNUNG, ELLEN DEKKERS, and LEIF KOBBELT RWTH-Aachen University Character Animation from 2D Pictures and 3D Motion Data ALEXANDER HORNUNG, ELLEN DEKKERS, and LEIF KOBBELT RWTH-Aachen University Presented by: Harish CS-525 First presentation Abstract This article presents

More information

Disambiguation of Horizontal Direction for Video Conference Systems

Disambiguation of Horizontal Direction for Video Conference Systems Disambiguation of Horizontal Direction for Video Conference Systems Mabel Mengzi Zhang, Seth Rotkin, and Jürgen P. Schulze mabel.m.zhang@gmail.com, sethrotkin@gmail.com, jschulze@ucsd.edu University of

More information

Study and Implementation of Video Compression standards (H.264/AVC, Dirac)

Study and Implementation of Video Compression standards (H.264/AVC, Dirac) Study and Implementation of Video Compression standards (H.264/AVC, Dirac) EE 5359-Multimedia Processing- Spring 2012 Dr. K.R Rao By: Sumedha Phatak(1000731131) Objective A study, implementation and comparison

More information

3 Image-Based Photo Hulls. 2 Image-Based Visual Hulls. 3.1 Approach. 3.2 Photo-Consistency. Figure 1. View-dependent geometry.

3 Image-Based Photo Hulls. 2 Image-Based Visual Hulls. 3.1 Approach. 3.2 Photo-Consistency. Figure 1. View-dependent geometry. Image-Based Photo Hulls Greg Slabaugh, Ron Schafer Georgia Institute of Technology Center for Signal and Image Processing Atlanta, GA 30332 {slabaugh, rws}@ece.gatech.edu Mat Hans Hewlett-Packard Laboratories

More information

Color Segmentation Based Depth Image Filtering

Color Segmentation Based Depth Image Filtering Color Segmentation Based Depth Image Filtering Michael Schmeing and Xiaoyi Jiang Department of Computer Science, University of Münster Einsteinstraße 62, 48149 Münster, Germany, {m.schmeing xjiang}@uni-muenster.de

More information

False alarm in outdoor environments

False alarm in outdoor environments Accepted 1.0 Savantic letter 1(6) False alarm in outdoor environments Accepted 1.0 Savantic letter 2(6) Table of contents Revision history 3 References 3 1 Introduction 4 2 Pre-processing 4 3 Detection,

More information

A Survey of Video Processing with Field Programmable Gate Arrays (FGPA)

A Survey of Video Processing with Field Programmable Gate Arrays (FGPA) A Survey of Video Processing with Field Programmable Gate Arrays (FGPA) Heather Garnell Abstract This paper is a high-level, survey of recent developments in the area of video processing using reconfigurable

More information

MobiX3D: a player for displaying 3D content on mobile devices

MobiX3D: a player for displaying 3D content on mobile devices MobiX3D: a player for displaying 3D content on mobile devices Daniele Nadalutti, Luca Chittaro, Fabio Buttussi HCI Lab Dept. of Math and Computer Science University of Udine via delle Scienze, 206 33100

More information

Cisco TelePresence TX1300 Series

Cisco TelePresence TX1300 Series Data Sheet Cisco TelePresence TX1300 Series Product Overview The Cisco TelePresence TX1300 Series extends the experience pioneered by Cisco s fully immersive threescreen solutions to smaller general purpose

More information

CCTV & Video Surveillance over 10G ip

CCTV & Video Surveillance over 10G ip CCTV & Video Surveillance over 10G ip Background With the increase in data, research and development and corporate competition, many companies are realizing the need to not only protect their data, but

More information

Go to contents 18 3D Visualization of Building Services in Virtual Environment

Go to contents 18 3D Visualization of Building Services in Virtual Environment 3D Visualization of Building Services in Virtual Environment GRÖHN, Matti Gröhn; MANTERE, Markku; SAVIOJA, Lauri; TAKALA, Tapio Telecommunications Software and Multimedia Laboratory Department of Computer

More information

A General Framework for Tracking Objects in a Multi-Camera Environment

A General Framework for Tracking Objects in a Multi-Camera Environment A General Framework for Tracking Objects in a Multi-Camera Environment Karlene Nguyen, Gavin Yeung, Soheil Ghiasi, Majid Sarrafzadeh {karlene, gavin, soheil, majid}@cs.ucla.edu Abstract We present a framework

More information

MPEG-H Audio System for Broadcasting

MPEG-H Audio System for Broadcasting MPEG-H Audio System for Broadcasting ITU-R Workshop Topics on the Future of Audio in Broadcasting Jan Plogsties Challenges of a Changing Landscape Immersion Compelling sound experience through sound that

More information

Video Conferencing. Femi Alabi UNC-CH - Comp 523 November 22, 2010

Video Conferencing. Femi Alabi UNC-CH - Comp 523 November 22, 2010 Video Conferencing Femi Alabi UNC-CH - Comp 523 November 22, 2010 Introduction Videoconferencing What Is It? Videoconferencing is a method of communicating between two or more locations where sound, vision

More information

Internet Desktop Video Conferencing

Internet Desktop Video Conferencing Pekka Isto 13.11.1998 1(8) Internet Desktop Video Conferencing ABSTRACT: This is report outlines possible use of Internet desktop videoconferencing software in a distributed engineering project and presents

More information

Subjective evaluation of a 3D videoconferencing system

Subjective evaluation of a 3D videoconferencing system Subjective evaluation of a 3D videoconferencing system Hadi Rizek 1, Kjell Brunnström 1,3, Kun Wang 1,3,Börje Andrén 1 and Mathias Johanson 2 1 Dept. of NetLab: Visual Media Quality, Acreo Swedish ICT

More information

A Short Introduction to Computer Graphics

A Short Introduction to Computer Graphics A Short Introduction to Computer Graphics Frédo Durand MIT Laboratory for Computer Science 1 Introduction Chapter I: Basics Although computer graphics is a vast field that encompasses almost any graphical

More information

Disambiguation of Horizontal Direction for Video Conference Systems

Disambiguation of Horizontal Direction for Video Conference Systems Disambiguation of Horizontal Direction for Video Conference Systems Mabel Mengzi Zhang, Seth Rotkin, and Jürgen P. Schulze University of California San Diego 9500 Gilman Dr, La Jolla, CA 92093 {mabel.m.zhang,sethrotkin}@gmail.com,

More information

International Journal of Advanced Information in Arts, Science & Management Vol.2, No.2, December 2014

International Journal of Advanced Information in Arts, Science & Management Vol.2, No.2, December 2014 Efficient Attendance Management System Using Face Detection and Recognition Arun.A.V, Bhatath.S, Chethan.N, Manmohan.C.M, Hamsaveni M Department of Computer Science and Engineering, Vidya Vardhaka College

More information

Computer Graphics AACHEN AACHEN AACHEN AACHEN. Public Perception of CG. Computer Graphics Research. Methodological Approaches - - - - - - - - - -

Computer Graphics AACHEN AACHEN AACHEN AACHEN. Public Perception of CG. Computer Graphics Research. Methodological Approaches - - - - - - - - - - Public Perception of CG Games Computer Graphics Movies Computer Graphics Research algorithms & data structures fundamental continuous & discrete mathematics optimization schemes 3D reconstruction global

More information

VIDEO CHAT & CUTOUT. ADSC Research Highlight

VIDEO CHAT & CUTOUT. ADSC Research Highlight ADSC Research Highlight VIDEO CHAT & CUTOUT Researchers at the Advanced Digital Sciences Center (ADSC) have achieved a major advance in the state of the art for real-time, robust video object cutout, and

More information

Interactive Cards A game system in Augmented Reality

Interactive Cards A game system in Augmented Reality Interactive Cards A game system in Augmented Reality João Alexandre Coelho Ferreira, Instituto Superior Técnico Abstract: Augmented Reality can be used on innumerous topics, but the point of this work

More information

Parametric Comparison of H.264 with Existing Video Standards

Parametric Comparison of H.264 with Existing Video Standards Parametric Comparison of H.264 with Existing Video Standards Sumit Bhardwaj Department of Electronics and Communication Engineering Amity School of Engineering, Noida, Uttar Pradesh,INDIA Jyoti Bhardwaj

More information

A Cognitive Approach to Vision for a Mobile Robot

A Cognitive Approach to Vision for a Mobile Robot A Cognitive Approach to Vision for a Mobile Robot D. Paul Benjamin Christopher Funk Pace University, 1 Pace Plaza, New York, New York 10038, 212-346-1012 benjamin@pace.edu Damian Lyons Fordham University,

More information

WHITE PAPER Personal Telepresence: The Next Generation of Video Communication. www.vidyo.com 1.866.99.VIDYO

WHITE PAPER Personal Telepresence: The Next Generation of Video Communication. www.vidyo.com 1.866.99.VIDYO WHITE PAPER Personal Telepresence: The Next Generation of Video Communication www.vidyo.com 1.866.99.VIDYO 2009 Vidyo, Inc. All rights reserved. Vidyo is a registered trademark and VidyoConferencing, VidyoDesktop,

More information

Chapter 3 ATM and Multimedia Traffic

Chapter 3 ATM and Multimedia Traffic In the middle of the 1980, the telecommunications world started the design of a network technology that could act as a great unifier to support all digital services, including low-speed telephony and very

More information

Multimedia Data Transmission over Wired/Wireless Networks

Multimedia Data Transmission over Wired/Wireless Networks Multimedia Data Transmission over Wired/Wireless Networks Bharat Bhargava Gang Ding, Xiaoxin Wu, Mohamed Hefeeda, Halima Ghafoor Purdue University Website: http://www.cs.purdue.edu/homes/bb E-mail: bb@cs.purdue.edu

More information

ANIMATION a system for animation scene and contents creation, retrieval and display

ANIMATION a system for animation scene and contents creation, retrieval and display ANIMATION a system for animation scene and contents creation, retrieval and display Peter L. Stanchev Kettering University ABSTRACT There is an increasing interest in the computer animation. The most of

More information

A Realistic Video Avatar System for Networked Virtual Environments

A Realistic Video Avatar System for Networked Virtual Environments A Realistic Video Avatar System for Networked Virtual Environments Vivek Rajan, Satheesh Subramanian, Damin Keenan Andrew Johnson, Daniel Sandin, Thomas DeFanti Electronic Visualization Laboratory University

More information

Video Coding Technologies and Standards: Now and Beyond

Video Coding Technologies and Standards: Now and Beyond Hitachi Review Vol. 55 (Mar. 2006) 11 Video Coding Technologies and Standards: Now and Beyond Tomokazu Murakami Hiroaki Ito Muneaki Yamaguchi Yuichiro Nakaya, Ph.D. OVERVIEW: Video coding technology compresses

More information

Authors: Masahiro Watanabe*, Motoi Okuda**, Teruo Matsuzawa*** Speaker: Masahiro Watanabe**

Authors: Masahiro Watanabe*, Motoi Okuda**, Teruo Matsuzawa*** Speaker: Masahiro Watanabe** Visualization of the Blood flow and the Stress distribution with the Diagnostic Support System for Circulatory Disease in the Volume Communications Environment Authors: Masahiro Watanabe*, Motoi Okuda**,

More information

Video Collaboration & Application Sharing Product Overview

Video Collaboration & Application Sharing Product Overview . Video Collaboration & Application Sharing Product Overview Overview NPL s Collaborative Real-Time Information Sharing Platform (CRISP ) combines high quality video collaboration, remote application sharing

More information

Eye Contact in Leisure Video Conferencing. Annick Van der Hoest & Dr. Simon McCallum Gjøvik University College, Norway.

Eye Contact in Leisure Video Conferencing. Annick Van der Hoest & Dr. Simon McCallum Gjøvik University College, Norway. Eye Contact in Leisure Video Conferencing Annick Van der Hoest & Dr. Simon McCallum Gjøvik University College, Norway 19 November 2012 Abstract This paper presents systems which enable eye contact in leisure

More information

VRSPATIAL: DESIGNING SPATIAL MECHANISMS USING VIRTUAL REALITY

VRSPATIAL: DESIGNING SPATIAL MECHANISMS USING VIRTUAL REALITY Proceedings of DETC 02 ASME 2002 Design Technical Conferences and Computers and Information in Conference Montreal, Canada, September 29-October 2, 2002 DETC2002/ MECH-34377 VRSPATIAL: DESIGNING SPATIAL

More information

Implementation of Video Voice over IP in Local Area Network Campus Environment

Implementation of Video Voice over IP in Local Area Network Campus Environment Implementation of Video Voice over IP in Local Area Network Campus Environment Mohd Nazri Ismail Abstract--In this research, we propose an architectural solution to integrate the video voice over IP (V2oIP)

More information

Speed Performance Improvement of Vehicle Blob Tracking System

Speed Performance Improvement of Vehicle Blob Tracking System Speed Performance Improvement of Vehicle Blob Tracking System Sung Chun Lee and Ram Nevatia University of Southern California, Los Angeles, CA 90089, USA sungchun@usc.edu, nevatia@usc.edu Abstract. A speed

More information

Multiple Description Coding (MDC) and Scalable Coding (SC) for Multimedia

Multiple Description Coding (MDC) and Scalable Coding (SC) for Multimedia Multiple Description Coding (MDC) and Scalable Coding (SC) for Multimedia Gürkan Gür PhD. Candidate e-mail: gurgurka@boun.edu.tr Dept. Of Computer Eng. Boğaziçi University Istanbul/TR ( Currenty@UNITN)

More information

Effects of microphone arrangement on the accuracy of a spherical microphone array (SENZI) in acquiring high-definition 3D sound space information

Effects of microphone arrangement on the accuracy of a spherical microphone array (SENZI) in acquiring high-definition 3D sound space information November 1, 21 1:3 1 Effects of microphone arrangement on the accuracy of a spherical microphone array (SENZI) in acquiring high-definition 3D sound space information Shuichi Sakamoto 1, Jun ichi Kodama

More information

Topic Maps Visualization

Topic Maps Visualization Topic Maps Visualization Bénédicte Le Grand, Laboratoire d'informatique de Paris 6 Introduction Topic maps provide a bridge between the domains of knowledge representation and information management. Topics

More information

Interactive Multimedia Courses-1

Interactive Multimedia Courses-1 Interactive Multimedia Courses-1 IMM 110/Introduction to Digital Media An introduction to digital media for interactive multimedia through the study of state-of-the-art methods of creating digital media:

More information

Virtual Data Gloves : Interacting with Virtual Environments through Computer Vision

Virtual Data Gloves : Interacting with Virtual Environments through Computer Vision Virtual Data Gloves : Interacting with Virtual Environments through Computer Vision Richard Bowden (1), Tony Heap(2), Craig Hart(2) (1) Dept of M & ES (2) School of Computer Studies Brunel University University

More information

SYNTHESIZING FREE-VIEWPOINT IMAGES FROM MULTIPLE VIEW VIDEOS IN SOCCER STADIUM

SYNTHESIZING FREE-VIEWPOINT IMAGES FROM MULTIPLE VIEW VIDEOS IN SOCCER STADIUM SYNTHESIZING FREE-VIEWPOINT IMAGES FROM MULTIPLE VIEW VIDEOS IN SOCCER STADIUM Kunihiko Hayashi, Hideo Saito Department of Information and Computer Science, Keio University {hayashi,saito}@ozawa.ics.keio.ac.jp

More information

Design and implementation of IPv6 multicast based High-quality Videoconference Tool (HVCT) *

Design and implementation of IPv6 multicast based High-quality Videoconference Tool (HVCT) * Design and implementation of IPv6 multicast based High-quality conference Tool (HVCT) * Taewan You, Hosik Cho, Yanghee Choi School of Computer Science & Engineering Seoul National University Seoul, Korea

More information

Introduction. C 2009 John Wiley & Sons, Ltd

Introduction. C 2009 John Wiley & Sons, Ltd 1 Introduction The purpose of this text on stereo-based imaging is twofold: it is to give students of computer vision a thorough grounding in the image analysis and projective geometry techniques relevant

More information

A Tool for Multimedia Quality Assessment in NS3: QoE Monitor

A Tool for Multimedia Quality Assessment in NS3: QoE Monitor A Tool for Multimedia Quality Assessment in NS3: QoE Monitor D. Saladino, A. Paganelli, M. Casoni Department of Engineering Enzo Ferrari, University of Modena and Reggio Emilia via Vignolese 95, 41125

More information

Applications that Benefit from IPv6

Applications that Benefit from IPv6 Applications that Benefit from IPv6 Lawrence E. Hughes Chairman and CTO InfoWeapons, Inc. Relevant Characteristics of IPv6 Larger address space, flat address space restored Integrated support for Multicast,

More information

Proposal for a Virtual 3D World Map

Proposal for a Virtual 3D World Map Proposal for a Virtual 3D World Map Kostas Terzidis University of California at Los Angeles School of Arts and Architecture Los Angeles CA 90095-1467 ABSTRACT The development of a VRML scheme of a 3D world

More information

high-quality surround sound at stereo bit-rates

high-quality surround sound at stereo bit-rates FRAUNHOFER Institute For integrated circuits IIS MPEG Surround high-quality surround sound at stereo bit-rates Benefits exciting new next generation services MPEG Surround enables new services such as

More information

PHOTOGRAMMETRIC TECHNIQUES FOR MEASUREMENTS IN WOODWORKING INDUSTRY

PHOTOGRAMMETRIC TECHNIQUES FOR MEASUREMENTS IN WOODWORKING INDUSTRY PHOTOGRAMMETRIC TECHNIQUES FOR MEASUREMENTS IN WOODWORKING INDUSTRY V. Knyaz a, *, Yu. Visilter, S. Zheltov a State Research Institute for Aviation System (GosNIIAS), 7, Victorenko str., Moscow, Russia

More information

Basic Theory of Intermedia Composing with Sounds and Images

Basic Theory of Intermedia Composing with Sounds and Images (This article was written originally in 1997 as part of a larger text to accompany a Ph.D. thesis at the Music Department of the University of California, San Diego. It was published in "Monochord. De

More information

An Instructional Aid System for Driving Schools Based on Visual Simulation

An Instructional Aid System for Driving Schools Based on Visual Simulation An Instructional Aid System for Driving Schools Based on Visual Simulation Salvador Bayarri, Rafael Garcia, Pedro Valero, Ignacio Pareja, Institute of Traffic and Road Safety (INTRAS), Marcos Fernandez

More information

USING COMPUTER VISION IN SECURITY APPLICATIONS

USING COMPUTER VISION IN SECURITY APPLICATIONS USING COMPUTER VISION IN SECURITY APPLICATIONS Peter Peer, Borut Batagelj, Franc Solina University of Ljubljana, Faculty of Computer and Information Science Computer Vision Laboratory Tržaška 25, 1001

More information

Cisco Video Collaboration Guide. Select the Right Equipment and Vendor

Cisco Video Collaboration Guide. Select the Right Equipment and Vendor Cisco Video Collaboration Guide Select the Right Equipment and Vendor See How Video Collaboration Fits Your Needs Camera. Microphone. Monitor. Speaker. Codec. These are the five essential components that

More information

Cisco Telepresence Implementation for Telekom s Corporate Requirements

Cisco Telepresence Implementation for Telekom s Corporate Requirements Cisco Telepresence Implementation for Telekom s Corporate Requirements Zdravko Stafilov 1 1 Makedonski Telekom, Orce Nikolov bb, 1000 Skopje, Macedonia Abstract. The Cisco Telepresence system was the platform

More information

Product Specifications

Product Specifications Data Sheet Cisco TelePresence TX9000 Series Product Overview The Cisco TelePresence TX9000 Series (TX9000) is the new benchmark for immersive in-person collaboration. With modern design and advanced functions,

More information

3D Videocommunication

3D Videocommunication 3D Videocommunication Algorithms, concepts and real-time systems in human centred communication EDITED BY Oliver Schreer Fraunhofer Institute for Telecommunications Heinrich-Hertz-Institut, Berlin, Germany

More information

3D Client Software - Interactive, online and in real-time

3D Client Software - Interactive, online and in real-time 3D Client Software - Interactive, online and in real-time Dipl.Inform.Univ Peter Schickel CEO Bitmanagement Software Vice President Web3D Consortium, Mountain View, USA OGC/Web3D liaison manager Presentation

More information

CHAPTER 6 TEXTURE ANIMATION

CHAPTER 6 TEXTURE ANIMATION CHAPTER 6 TEXTURE ANIMATION 6.1. INTRODUCTION Animation is the creating of a timed sequence or series of graphic images or frames together to give the appearance of continuous movement. A collection of

More information

A method of generating free-route walk-through animation using vehicle-borne video image

A method of generating free-route walk-through animation using vehicle-borne video image A method of generating free-route walk-through animation using vehicle-borne video image Jun KUMAGAI* Ryosuke SHIBASAKI* *Graduate School of Frontier Sciences, Shibasaki lab. University of Tokyo 4-6-1

More information

H.264 Based Video Conferencing Solution

H.264 Based Video Conferencing Solution H.264 Based Video Conferencing Solution Overview and TMS320DM642 Digital Media Platform Implementation White Paper UB Video Inc. Suite 400, 1788 west 5 th Avenue Vancouver, British Columbia, Canada V6J

More information

Circle Object Recognition Based on Monocular Vision for Home Security Robot

Circle Object Recognition Based on Monocular Vision for Home Security Robot Journal of Applied Science and Engineering, Vol. 16, No. 3, pp. 261 268 (2013) DOI: 10.6180/jase.2013.16.3.05 Circle Object Recognition Based on Monocular Vision for Home Security Robot Shih-An Li, Ching-Chang

More information

Using Photorealistic RenderMan for High-Quality Direct Volume Rendering

Using Photorealistic RenderMan for High-Quality Direct Volume Rendering Using Photorealistic RenderMan for High-Quality Direct Volume Rendering Cyrus Jam cjam@sdsc.edu Mike Bailey mjb@sdsc.edu San Diego Supercomputer Center University of California San Diego Abstract With

More information

Video compression: Performance of available codec software

Video compression: Performance of available codec software Video compression: Performance of available codec software Introduction. Digital Video A digital video is a collection of images presented sequentially to produce the effect of continuous motion. It takes

More information

Internet Video Streaming and Cloud-based Multimedia Applications. Outline

Internet Video Streaming and Cloud-based Multimedia Applications. Outline Internet Video Streaming and Cloud-based Multimedia Applications Yifeng He, yhe@ee.ryerson.ca Ling Guan, lguan@ee.ryerson.ca 1 Outline Internet video streaming Overview Video coding Approaches for video

More information

The Flat Shape Everything around us is shaped

The Flat Shape Everything around us is shaped The Flat Shape Everything around us is shaped The shape is the external appearance of the bodies of nature: Objects, animals, buildings, humans. Each form has certain qualities that distinguish it from

More information