Immersive 3-D Video Conferencing: Challenges, Concepts, and Implementations
|
|
- Jonas Sanders
- 8 years ago
- Views:
Transcription
1 Proc. SPIE Visual Communications and Image Processing (VCIP), Lugano, Switzerland, July Immersive 3-D Video Conferencing: Challenges, Concepts, and Implementations Peter Eisert Image Processing Department Fraunhofer Institute for Telecommunications, Heinrich-Hertz-Institute Einsteinufer 37, D Berlin, Germany URL: eisert ABSTRACT In this paper, a next generation 3-D video conferencing system is presented that provides immersive telepresence and natural representation of all participants in a shared virtual meeting space. The system is based on the principle of a shared virtual table environment which guarantees correct eye contact and gesture reproduction and enhances the quality of human-centered communication. The virtual environment is modeled in MPEG-4 which also allows the seamless integration of explicit 3-D head models for a low-bandwidth connection to mobile users. In this case, facial expression and motion information is transmitted instead of video streams resulting in bit-rates of a few kbit/s per participant. Beside low bit-rates, the model-based approach enables new possibilities for image enhancements like digital make-up, digital dressing, or modification of scene lighting. Keywords: Immersive Video Conferencing, Tele-Collaboration, Shared Virtual Table Environment, Model- Based Coding, View Synthesis 1. INTRODUCTION The idea of video conferencing and video telephony has fascinated researchers for a long time and new devices have been developed as soon as new technologies became available. Already in 1927, a video phone conversation with life video transmission over telephone lines from Washington D.C. to an auditorium in Manhattan was demonstrated. The Bell picturephone shown in Fig. 1 used a mechanical Nipkow Disk for sampling the images and had a display of size 2 by 2.5 inches. Figure 1. Bell s picturephone system of Over time, devices got smaller while display sizes increased. In the sixties, a couple of different video phones were developed which enabled video transmission over telephone lines. The left hand side of Fig. 2 shows an example of such a picture phone. Unfortunately, these technologies never became well accepted due to costs, limited availability, and restricted gain in communication. First with the introduction of digital computers, cheap and powerful hardware and software devices have been developed and, today, video conferencing is rapidly advancing. Especially for decision-making in a wide range of applications, from world-wide operating companies, international research projects to global investment banking, the interest in tele-conferencing has increased in recent years.
2 Figure 2. Left: Picturephone from Right: Hydra system. Nevertheless, video conferencing today is limited in its support of natural human-centered communication. Body postures and subtle movements, gaze direction, room acoustics, and joint interactions are often misrepresented, lacking, or wholly absent. Eye contact is still a problem and also verbal and non-verbal communication between other participants can often not be followed. To address these limitations, several experimental systems have been proposed in the last few years. The Hydra system depicted on the right hand side of Fig. 2, e.g., uses three independent communication devices with three displays distributed on a desk to ensure a spatial and acoustical separation of the participants. Figure 3. Concept of an immersive shared virtual table video conferencing system. Beside the reproduction of spatial relations, large displays 1 and other benefits of tele-immersion are exploited in order to give the participants the impression of being present in a shared virtual table environment (SVTE) suggesting spatial proximity and enabling a higher degree of natural interaction and effective collaboration. In this sense, the participants should be able to make use of rich communication modalities as similar as possible to those used in a face-to-face meeting. The exact reproduction of gestures, eye contact, sound direction, and multiple simultaneous inter-participant conversations are examples of these requirements. The main objective is to enable conferees located in different geographical places to meet around a virtual table, appearing at each
3 station in such a way to create a convincing impression of presence as illustrated in Fig. 3. First approaches in this area were limited to strictly graphical environments using simple avatars for representing the remote conferees. A lot of systems have been proposed in this context during the last decade. One example for these research activities was the former Europe ACTS project COVEN that demonstrated the benefits of the SVTE concept by a networked business game VR application. 2 Later approaches were more and more focused on a seamless integration of 2-D video images or even 3-D video avatars into Collaborative Virtual Environments (CVE) to increase realism. 3, 4 Some of these approaches were driven by the MPEG-4 multimedia standard that offers new powerful coding and composition tools for such purposes. 5 Figure 4. Left: Setup of NTII at UNC. 6 Right: TELEPORT System. 7 An attractive SVTE approach is the one of tele-cubicles. 6 8 A common feature of these proposals is a special system set-up, where the participants are situated symmetrically around the shared table, with each conferee appearing on an own screen as illustrated in Fig. 4. Note that the symmetric geometry of this setup guarantees eye contact, gaze awareness and gesture reproduction. Thus, everybody in the session can observe who is talking to whom or who is pointing at what. In the next section, the Virtual Team User Environments (VIRTUE) system 9 is proposed that offers all benefits of the tele-cubicle approach and extends it towards an integration into shared virtual working spaces. A seamless transition between the real working desk in front of the display and the virtual conference table at the screen gives the user the impression of being part of an extended perception space. Moreover, the remote participants are rendered seamlessly und under correct perspective view into the virtual conference scene. Scene depth information obtained from stereo cameras and encoded and streamed using MPEG-4 10 enables view synthesis in a virtual 3-D scene. Replacing the depth maps by explicit 3-D computer graphics head-and-shoulder models leads to an enormous reduction in bit-rate for the transmission of scene data, since shape and texture data need not be updated for each frame. Such a model-based coding system, 11 which can be used as a fall-back solution for mobile devices connected to the Internet over low-bandwidth channels, is presented in Section VIRTUAL TEAM USER ENVIRONMENT In this section, we present an extension to the Virtual Team User Environments (VIRTUE) system 9, 12 that uses the state-of-the-art multimedia standard MPEG-4 for transmission and representation of scene and texture data. Due to usage of the MPEG-4 signaling protocol DMIF, a sophisticated 3-D video processing, and the MPEG-4 scene description language BIFS, the system architecture supports an arbitrary terminal setup, interoperability with other terminal configurations, and scalability in terms of the number of participants and it makes it open for future extensions and developments towards immersive tele-collaboration, multi-user terminals, and joint applications.
4 2.1. Shared Virtual Table Concept The basic idea of the shared virtual environment concept is to place 3-D video reproductions of a given number of participants at predefined positions in a shared virtual environment. For this purpose, the conferees are captured at each terminal by a multiple camera setup as shown in Fig. 5 and the desired 3-D video representation of the local person is extracted from the images. Figure 5. Left: Setup for a 3-party conference. Right: VIRTUE setup. Then, the 3-D video objects of all conferees are grouped virtually around the shared table. Ideally, this is done in an isotropic manner in order to obtain social and geometric symmetry. Hence, in the case of a 3-party conference, the participants form an equilateral triangle. In the case of four parties, it would be a square, and so on. Following such generic composition rules and knowing the number of participants, the same SVTE can be built at each terminal from previously loaded scene descriptions and the 3-D video streams. Figure 6. Rendering of the virtual 3-D conference scene. Based on this generic scene composition, individual views of the virtual conference environment can be rendered by using a virtual camera as illustrated in Fig. 6. Locally, the position of the virtual camera has to move coincidently with the current position of the conferee s head, which is permanently registered by a head tracker. Thus, supposing that the geometrical parameters of the multi-view capture device, the virtual scene, and the virtual camera are well fitted to each other, it is ensured that all conferees see the scene under the correct perspective view, even while changing their own viewing position. This geometrical coincidence provides all desired attributes mentioned in the introduction eye contact, gaze awareness, gesture reproduction, natural conference situation, and high amount of realism. In addition, the support of head motion parallax allows the conferees to change their viewing position in order to watch the scene from another perspective, to look behind objects, or to look at a previously occluded object.
5 2.2. Architecture of the System Fig. 7 outlines the system architecture of the 3-D video conferencing system. After multi-view capturing, the video frames are segmented to separate the person s silhouette from the background. As a result, the conferees are represented by arbitrarily shaped video objects, which can be integrated seamlessly into virtual environments. To extract depth information, disparity estimation is performed on the rectified video objects resulting in dense disparity maps. Both video objects and disparity maps are efficiently encoded using MPEG-4. Figure 7. Architecture of the3-d video conference system. The system concept takes advantage of several particular features of the MPEG-4 multimedia standard. MPEG-4 allows the encoding of arbitrarily shaped video objects and provides auxiliary alpha planes to transmit additional pixel information associated to the color data. These additional planes can be used for a joint transmission of disparity maps and video objects. After encoding, the packets are streamed to other participating terminals via RTP. Simultaneously, the terminal receives video streams from the other conference partners and decodes them with multiple MPEG-4 video decoders. The shaped video objects and audio data are synchronized to each other and integrated into the SVTE scene represented by the MPEG-4 scene description language BIFS. Finally, an MPEG-4 compositor is used for rendering purposes. This compositor is able to handle user events (head tracker or interaction with scene content) as well as scene updates sent by a server. The MPEG-4 video objects are three-dimensionally warped using image-based rendering techniques before they are integrated into the scene. Depending on the current input from the head tracker, the correct perspective view is calculated and the adapted view is inserted into the BIFS scene as a 2-D video Foreground Background Segmentation In the finally rendered scene, the participants appear as arbitrarily shaped video objects, seamlessly integrated into the virtual environment. This requires a segmentation of the moving person from the background which is assumed to remain rather static. Initially, the background is captured and a change detection scheme compares this reference image with the current video data and provides a segmentation mask. 5 The reference image is permanently updated to cope with slight changes of illumination or scene content. This baseline algorithm has been improved in speed and quality in order to meet the real-time constraints for full CCIR601 resolution and video frame rate. Its performance has further been improved by adding a fast and efficient shadow detection tool. 13 For a robust segmentation, this is particularly important, because shadows at the table can usually not be avoided, even under optimal illumination conditions. The effect of shadow detection on the segmentation result is shown in Fig. 8.
6 Figure 8. Left: Original frame. Middle: Segmented foreground without shadow detection. Right: with shadow detection Disparity Estimation and Depth Analysis To extract the required 3-D representation, the depth of the captured video object is analyzed by disparity matching on the basis of the rectified images. We have developed a new hybrid block- and pixel-recursive approach, which is able to compute the disparity fields in real-time on a state-of-the-art PC. 14 Apart from a considerable reduction of computational load, the algorithm leads to spatio-temporally consistent depth maps, particularly important for view synthesis, since temporal inconsistencies may cause annoying artifacts in the synthesized views. In order to deal with occlusions caused by hands, arms, or the head, several post-processing techniques are applied. In a first step, critical regions are detected by a consistency check between disparity fields for a left to right and a right to left match. Unreliable disparity values in occluded areas are filled efficiently by suitable extrapolation techniques. The segmentation of the hands and arms which cause severe depth discontinuities is further refined by tracking and segmenting the hands using motion and skin color information. Fig. 9 shows two rectified images, the disparity fields with occluded areas, and the final disparity map after post-processing. Figure and 2. Image: Rectified frames from the left and right camera. occluded areas. 5. Image: Disparity map after post-processing. 3. and 4. Image: Disparity fields with 2.5. Head Tracking The perspective view of the scene presented on the display depends on the viewer s position. This requires an accurate estimation of the viewer s position in the 3-D space, which is accomplished by the head-tracking module. The chosen approach is based on a skin color segmentation technique jointly with a facial feature tracker searching eye positions. Due to the restriction to video conference scenarios, several assumptions can be made to facilitate the tracking of the viewer s head. The obtained 2-D positions of the eyes in two images can then be used for an accurate calculation of the 3-D head position View Synthesis and Virtual Scene Composition The calculated head position and the analyzed depth of the transmitted video objects provide sufficient information to synthesize novel virtual views of the remote conferees. In order to meet real-time constraints 15, 16 and
7 efficient occlusion handling, a new view synthesis algorithm has been developed in the VIRTUE project which is able to take these issues into account. 17 Figure 10. Composed scene containing computer graphics and video objects. At the end of the processing chain, the virtual conference scene has to be composed and displayed onto the screen as depicted in Fig. 10. The virtual scene is represented by a number of polygons in 3-D space and encoded by the BIFS scene description language of MPEG-4. In this polygon-based scene representation the participants are substituted by 2-D rectangles positioned around the virtual table. In this sense, the synthesized novel views of the remote participants are treated as textures and are transferred directly to the graphics card D MODEL-BASED VIDEO CONFERENCING In the previous section, the participants are described by planar computer graphics objects. Onto these planes, the recorded video is projected after warping according to the estimated disparity maps. This has the advantage of very natural looking results, since the original video is exploited in the image-based view synthesis. On the other hand, video streams from each participant together with their corresponding disparity maps have to be transmitted to the client. Dependant on the number of participants, this results in rather high bit-rates even if the streams are encoded with MPEG-4. Current wireless channels are usually not able to handle such bit-rates which prevents the participation of users from mobile devices like PDAs or smart phones. Figure 11. Virtual Video Conferencing on a Mobile Device.
8 In this section, we propose an alternative model-based approach 11, 18, 19 that can be used as a fall-back solution for virtual conferencing systems which are connected to the Internet via low-bandwidth channels. Instead of transmitting texture and depth information of all conferees for each frame, 3-D models of the participants are created representing their appearance. The models are generated only once in an initial phase, encoded, and transmitted to all devices. During the video conference, motion and facial expressions of the conferees are tracked over time and only a small set of facial animation parameters has to be streamed over the network. Experiments have shown, that bit-rates of a few kbit/s per participant can be achieved which enables virtual video conferencing also on mobile devices as illustrated in Fig. 11. The decoder only has to move and deform the 3-D head-andshoulder models in the virtual conferencing scene and render it using computer graphics techniques. Rendering and the encoding and streaming of 3-D head models and facial animation parameters (FAPs) can be realized with MPEG-4 in a standard compliant way. 10 Thus, the compositor of the aforementioned immersive video conferencing system can incorporate these explicit models with only minor changes. The architecture of such a model-based coding component with image analysis and view synthesis is depicted in Fig. 12. Video Analysis Estimation of FAPs Coder Parameter entropy coding Channel about 1 kbit/s Parameter decoding Decoder Synthesis Animation and rendering of the head model Video Head Model Shape Texture Dynamics Figure 12. Model-based Video Codec. Besides the low bit-rate, the usage of explicit 3-D models for the people in the scene also has some other implications on the virtual conferencing system. A sophisticated real-time segmentation of the conferees can be avoided, since the computer models already provide an inherent segmentation. Moreover, view point transformations can easily be applied by simply moving the virtual camera, since all objects in the scene are represented by 3-D models and no image-based warping is required. Another problem in mobile virtual conferencing is an eventually unfavorable appearance or dressing of the conferee depended on his or her current location or situation. By using pre-stored 3-D models for transmission from the mobile device to the other participants, the desired look can be ensured even under awkward situations. This way, digital dressing and digital make-up can be easily applied Facial Expression Analysis The most challenging part of facial analysis and synthesis is the estimation of 3-D facial motion and deformation from two-dimensional images. Due to the loss of one dimension caused by the projection of the real world onto the image plane, this task can only be solved by exploiting additional knowledge of the objects in the scene. In particular, the way the objects move can often be restricted to a low number of degrees of freedom that can be described by a limited set of parameters. We have developed a new 3-D model-based method for the estimation of facial expressions that makes use of an explicit parameterized 3-D human head model describing shape, color, and motion constraints of an individual person. 20 This model information is jointly exploited with spatial and temporal intensity gradients of the images. Thus, the entire area of the image showing the object of interest is used instead of dealing with discrete feature points, resulting in a robust and highly accurate system for modelbased estimation of MPEG-4 facial animation parameters. A linear and computationally efficient algorithm can
9 be derived for different scenarios. The scheme is embedded in a hierarchical analysis-synthesis framework to avoid error accumulation in the long-term estimation Illumination Analysis For natural and uncontrolled video capture conditions, often occurring in mobile environments, scene lighting often varies over time. This illumination variability not only has a considerable influence on the visual appearance of the objects in the scene, but also on the performance of computer vision algorithms or video coding methods. The efficiency and robustness of these algorithms can be significantly improved by removing the undesired effects of changing illumination. In our facial analysis and synthesis system, we model also lighting properties in the scene and estimate them. 21 The explicit knowledge about light sources and reflection properties not only increases robustness of the facial expression analysis but also enables new possibilities for improvement or manipulation of lighting conditions. For example, variations in illumination can be removed in a post-processing step and the lighting conditions can be adjusted to the one in the virtual conference room. This is illustrated in Fig. 13, where the upper row refers to the original image sequence while the lower row depicts the corresponding frames after lighting change estimation and removal. Figure 13. Upper row: original sequence, lower row: corresponding illumination-compensated frames with constant lighting Model-based View Synthesis Synthesizing new views is accomplished by simply rendering the 3-D models in the virtual scene. MPEG-4 offers possibilities to describe head models as well as streaming of facial animation parameters. The color of the person is defined by a texture map which is created from the video frames. For the definition of the individual shape, multiple options are available in MPEG-4. We use our own generic triangle mesh which is adapted to the individual and defined using the scene description BIFS. A facial animation table (FAT) 10 specifies the mesh deformations and transforms which occur when changing a specific facial animation parameter. After the virtual scene, defined by the BIFS tree, is transmitted to all participants of the video conference, only facial animation parameters are sent for each frame after having been estimated from the camera images. Bit-rate measurements have shown, 20 that about 1 kbit/s is sufficient to stream these parameters. Fig. 14 shows an example for the proposed model-based video streaming system. The upper row depicts three frames from a head-and-shoulder video sequence of one participant in a conference. From this monocular image sequence, the 3-D head model shown in the middle row of Fig. 14 is created and facial expression analysis is performed. The estimated parameters are encoded and sent to the decoder which renders the models after
10 Figure 14. Upper row: sequence Peter, middle row: animated wireframe model, lower row: synthesized model frames generated by rendering the 3-D head model. applying local deformation and motion. The middle and lower row of Fig. 14 refer to the output of the decoder after rendering as a wire-frame or with a textured model. 4. CONCLUSIONS We have presented a new concept for immersive video conferencing using a shared virtual table environment (SVTE) which allows a very intensive communication and effective collaboration. It combines the benefits of former VR-based collaborative virtual environment approaches and video-based tele-cubicle systems with those of mixed reality applications. A seamless transition between the real and virtual conference table gives the user the impression of an extended perception space. The usage of 3-D video objects for the representation of remote conferees enables natural representation of gestures, eye contact and gaze. Due to the MPEG-4 multimedia standard, the system provides open and flexible system architectures. Beside arbitrarily shaped video objects, the MPEG-4 compositor allows the rendering of explicit head models enabling the usage of model-based coding techniques as an alternative for communication with devices connected via low-bandwidth channels. The concept is therefore scalable in the number of participants and channel bit-rates, and can be extended towards future applications of immersive tele-collaboration as well as auto-stereoscopic single- or multi-user displays. REFERENCES 1. M. Chen, Design of a virtual auditorium, in Proc. of ACM Multimedia, (Ottawa, Canada), Sep S. Benford, C. Greenhalgh, and D. Lloyd, Crowded collaborative virtual environments, in Proc. of the ACM SIGCHI Conference on Human Factors in Computing Systems, pp , (Atlanta, USA), Mar O. Ståhl, B. Serenius, and G. Taxén, Meetings for real - experiences from a series of VR-based project meetings, in Proc. Symposium on Virtual Reality Software and Technology, (London, UK), Dec
11 4. V. Rajan, S. Subramanian, D. Keenan, A. Johnson, D. Sandin, and T. DeFanti, A realistic video avatar system for networked virtual environments, in Proc. Immersive Projection Technology Symposium, (Orlando, USA), Mar S. Rauthenberg, A. Graffunder, U. Kowalik, and P. Kauff, Virtual shop and virtual meeting point - two prototype applications of interactive services using the new multimedia coding standard MPEG-4, in Proc. International Conference on Computer Communication, (Tokyo, Japan), Sep W. Chen, H. Towles, L. Nyland, G. Welch, and H. Fuchs, Toward a compelling sensation of telepresence: Demonstrating a portal to a distant (static) office, in Proc. Visualization 2000, pp , (Salt Lake City, USA), Oct S. J. Gibbs, C. Arapis, and C. Breiteneder, TELEPORT - Towards immersive copresence, Multimedia Systems, T. Aoki, W. Kustarto, N. Sakamoto, N. Suzuki, K. Saburi, and H. Yasuda, MONJUnoCHIE system: Videoconference system with eye contact for decision making, in Proc. International Workshop on Advanced Image Technology (IWAIT), P. Kauff and O. Schreer, Virtual team user environments - a step from tele-cubicles towards distributed tele-collaboration in mediated workspaces, in Proc. International Conference on Multimedia and Expo (ICME), (Lausanne, Switzerland), Aug ISO/IEC FDIS , Generic Coding of audio-visual objects: (MPEG-4 video), Final Draft International Standard, Document N2502, D. E. Pearson, Developments in model-based video coding, Proceedings of the IEEE 83, pp , June O. Schreer and P. Kauff, An immersive 3D video-conferencing system using shared virtual team user environments, in ACM Collaborative Environments, CVE 2002, (Bonn, Germany), O. Schreer, I. Feldmann, U. Gölz, and P. Kauff, Fast and robust shadow detection in videoconference applications, in Proc. of VIPromCom, 4th EURASIP IEEE International Symposium on Video Processing and Multimedia Communications, (Zadar, Croatia), June O. Schreer, N. Brandenburg, and P. Kauff, Real-time disparity analysis for applications in imersive teleconference scenarios - a comparative study, in Proc. ICIAP 01, (Palermo, Italy), Sep S. Laveau and O. D. Faugeras, 3D scene representation as a collection of images, in Proc. International Conference on Pattern Recognition (ICPR), pp , (Jerusalem, Israel), Oct S. Avidan and A. Shashua, Novel view synthesis in tensor space, in Proc. Computer Vision and Pattern Recognition, pp , (Puerto Rico), June B. J. Lei and E. A. Hendriks, Multi-step view synthesis with occlusion handling, in Proc. Vision, Modeling, and Visualization VMV 02, (Stuttgart, Germany), Nov W. J. Welsh, S. Searsby, and J. B. Waite, Model-based image coding, British Telecom Technology Journal 8, pp , Jul P. Eisert and B. Girod, Analyzing facial expressions for virtual conferencing, IEEE Computer Graphics and Applications 18, pp , Sep P. Eisert, Very Low Bit-Rate Video Coding Using 3-D Models. PhD thesis, University of Erlangen, Shaker Verlag, Aachen, Germany, P. Eisert and B. Girod, Model-based enhancement of lighting conditions in image sequences, in Proc. SPIE Visual Communications and Image Processing, VCIP-02, (San Jose, USA), Jan
VIRTUAL VIDEO CONFERENCING USING 3D MODEL-ASSISTED IMAGE-BASED RENDERING
VIRTUAL VIDEO CONFERENCING USING 3D MODEL-ASSISTED IMAGE-BASED RENDERING Peter Eisert Fraunhofer Institute for Telecommunications, Heinrich-Hertz-Institute Image Processing Department Einsteinufer 37,
More informationTemplate-based Eye and Mouth Detection for 3D Video Conferencing
Template-based Eye and Mouth Detection for 3D Video Conferencing Jürgen Rurainsky and Peter Eisert Fraunhofer Institute for Telecommunications - Heinrich-Hertz-Institute, Image Processing Department, Einsteinufer
More informationVIRTUE The step towards immersive telepresence in virtual video-conference systems
VIRTUE The step towards immersive telepresence in virtual video-conference systems Oliver SCHREER (HHI) 1 and Phil SHEPPARD (British Telecom) 2 1 Heinrich-Hertz-Institut, Einsteinufer 37, D-10587 Berlin,
More informationREPRESENTATION, CODING AND INTERACTIVE RENDERING OF HIGH- RESOLUTION PANORAMIC IMAGES AND VIDEO USING MPEG-4
REPRESENTATION, CODING AND INTERACTIVE RENDERING OF HIGH- RESOLUTION PANORAMIC IMAGES AND VIDEO USING MPEG-4 S. Heymann, A. Smolic, K. Mueller, Y. Guo, J. Rurainsky, P. Eisert, T. Wiegand Fraunhofer Institute
More informationPeter Eisert, Thomas Wiegand and Bernd Girod. University of Erlangen-Nuremberg. Cauerstrasse 7, 91058 Erlangen, Germany
RATE-DISTORTION-EFFICIENT VIDEO COMPRESSION USING A 3-D HEAD MODEL Peter Eisert, Thomas Wiegand and Bernd Girod Telecommunications Laboratory University of Erlangen-Nuremberg Cauerstrasse 7, 91058 Erlangen,
More informationREMOTE RENDERING OF COMPUTER GAMES
REMOTE RENDERING OF COMPUTER GAMES Peter Eisert, Philipp Fechteler Fraunhofer Institute for Telecommunications, Einsteinufer 37, D-10587 Berlin, Germany eisert@hhi.fraunhofer.de, philipp.fechteler@hhi.fraunhofer.de
More informationImmersive Medien und 3D-Video
Fraunhofer-Institut für Nachrichtentechnik Heinrich-Hertz-Institut Ralf Schäfer schaefer@hhi.de http://ip.hhi.de Immersive Medien und 3D-Video page 1 Outline Immersive Media Examples Interactive Media
More informationAnalyzing Facial Expressions for Virtual Conferencing
IEEE Computer Graphics & Applications, pp. 70-78, September 1998. Analyzing Facial Expressions for Virtual Conferencing Peter Eisert and Bernd Girod Telecommunications Laboratory, University of Erlangen,
More informationMaking Machines Understand Facial Motion & Expressions Like Humans Do
Making Machines Understand Facial Motion & Expressions Like Humans Do Ana C. Andrés del Valle & Jean-Luc Dugelay Multimedia Communications Dpt. Institut Eurécom 2229 route des Crêtes. BP 193. Sophia Antipolis.
More informationBUILDING TELEPRESENCE SYSTEMS: Translating Science Fiction Ideas into Reality
BUILDING TELEPRESENCE SYSTEMS: Translating Science Fiction Ideas into Reality Henry Fuchs University of North Carolina at Chapel Hill (USA) and NSF Science and Technology Center for Computer Graphics and
More informationM3039 MPEG 97/ January 1998
INTERNATIONAL ORGANISATION FOR STANDARDISATION ORGANISATION INTERNATIONALE DE NORMALISATION ISO/IEC JTC1/SC29/WG11 CODING OF MOVING PICTURES AND ASSOCIATED AUDIO INFORMATION ISO/IEC JTC1/SC29/WG11 M3039
More informationTalking Head: Synthetic Video Facial Animation in MPEG-4.
Talking Head: Synthetic Video Facial Animation in MPEG-4. A. Fedorov, T. Firsova, V. Kuriakin, E. Martinova, K. Rodyushkin and V. Zhislina Intel Russian Research Center, Nizhni Novgorod, Russia Abstract
More informationEye-contact in Multipoint Videoconferencing
Eye-contact in Multipoint Videoconferencing Birgit Quante and Lothar Mühlbach Heinrich-Hertz-Institut für Nachrichtentechnik Berlin GmbH (HHI) Einsteinufer 37, D-15087 Berlin, Germany, http://www.hhi.de/
More informationVirtual Environments - Basics -
Virtual Environments - Basics - What Is Virtual Reality? A Web-Based Introduction Version 4 Draft 1, September, 1998 Jerry Isdale http://www.isdale.com/jerry/vr/whatisvr.html Virtual Environments allow
More informationQuality of Service Management for Teleteaching Applications Using the MPEG-4/DMIF
Quality of Service Management for Teleteaching Applications Using the MPEG-4/DMIF Gregor v. Bochmann and Zhen Yang University of Ottawa Presentation at the IDMS conference in Toulouse, October 1999 This
More informationHow To Compress Video For Real Time Transmission
University of Edinburgh College of Science and Engineering School of Informatics Informatics Research Proposal supervised by Dr. Sethu Vijayakumar Optimized bandwidth usage for real-time remote surveillance
More informationWhite paper. H.264 video compression standard. New possibilities within video surveillance.
White paper H.264 video compression standard. New possibilities within video surveillance. Table of contents 1. Introduction 3 2. Development of H.264 3 3. How video compression works 4 4. H.264 profiles
More informationA Prototype For Eye-Gaze Corrected
A Prototype For Eye-Gaze Corrected Video Chat on Graphics Hardware Maarten Dumont, Steven Maesen, Sammy Rogmans and Philippe Bekaert Introduction Traditional webcam video chat: No eye contact. No extensive
More informationHow to Send Video Images Through Internet
Transmitting Video Images in XML Web Service Francisco Prieto, Antonio J. Sierra, María Carrión García Departamento de Ingeniería de Sistemas y Automática Área de Ingeniería Telemática Escuela Superior
More informationConference interpreting with information and communication technologies experiences from the European Commission DG Interpretation
Jose Esteban Causo, European Commission Conference interpreting with information and communication technologies experiences from the European Commission DG Interpretation 1 Introduction In the European
More informationFRAUNHOFER INSTITUTE FOR INTEGRATED CIRCUITS IIS AUDIO COMMUNICATION ENGINE RAISING THE BAR IN COMMUNICATION QUALITY
FRAUNHOFER INSTITUTE FOR INTEGRATED CIRCUITS IIS AUDIO COMMUNICATION ENGINE RAISING THE BAR IN COMMUNICATION QUALITY BENEFITS HIGHEST AUDIO QUALITY FOR NEXT GENERATION COMMU- NICATION SYSTEMS Communication
More informationBandwidth Adaptation for MPEG-4 Video Streaming over the Internet
DICTA2002: Digital Image Computing Techniques and Applications, 21--22 January 2002, Melbourne, Australia Bandwidth Adaptation for MPEG-4 Video Streaming over the Internet K. Ramkishor James. P. Mammen
More informationVery Low Frame-Rate Video Streaming For Face-to-Face Teleconference
Very Low Frame-Rate Video Streaming For Face-to-Face Teleconference Jue Wang, Michael F. Cohen Department of Electrical Engineering, University of Washington Microsoft Research Abstract Providing the best
More informationIntroduction to Computer Graphics
Introduction to Computer Graphics Torsten Möller TASC 8021 778-782-2215 torsten@sfu.ca www.cs.sfu.ca/~torsten Today What is computer graphics? Contents of this course Syllabus Overview of course topics
More informationAn Animation Definition Interface Rapid Design of MPEG-4 Compliant Animated Faces and Bodies
An Animation Definition Interface Rapid Design of MPEG-4 Compliant Animated Faces and Bodies Erich Haratsch, Technical University of Munich, erich@lis.e-tecknik.tu-muenchen.de Jörn Ostermann, AT&T Labs
More informationOne-Way Pseudo Transparent Display
One-Way Pseudo Transparent Display Andy Wu GVU Center Georgia Institute of Technology TSRB, 85 5th St. NW Atlanta, GA 30332 andywu@gatech.edu Ali Mazalek GVU Center Georgia Institute of Technology TSRB,
More informationModelling 3D Avatar for Virtual Try on
Modelling 3D Avatar for Virtual Try on NADIA MAGNENAT THALMANN DIRECTOR MIRALAB UNIVERSITY OF GENEVA DIRECTOR INSTITUTE FOR MEDIA INNOVATION, NTU, SINGAPORE WWW.MIRALAB.CH/ Creating Digital Humans Vertex
More informationCloud-Empowered Multimedia Service: An Automatic Video Storytelling Tool
Cloud-Empowered Multimedia Service: An Automatic Video Storytelling Tool Joseph C. Tsai Foundation of Computer Science Lab. The University of Aizu Fukushima-ken, Japan jctsai@u-aizu.ac.jp Abstract Video
More informationARIB STD-T64-C.S0042 v1.0 Circuit-Switched Video Conferencing Services
ARIB STD-T-C.S00 v.0 Circuit-Switched Video Conferencing Services Refer to "Industrial Property Rights (IPR)" in the preface of ARIB STD-T for Related Industrial Property Rights. Refer to "Notice" in the
More informationText Localization & Segmentation in Images, Web Pages and Videos Media Mining I
Text Localization & Segmentation in Images, Web Pages and Videos Media Mining I Multimedia Computing, Universität Augsburg Rainer.Lienhart@informatik.uni-augsburg.de www.multimedia-computing.{de,org} PSNR_Y
More informationBernice E. Rogowitz and Holly E. Rushmeier IBM TJ Watson Research Center, P.O. Box 704, Yorktown Heights, NY USA
Are Image Quality Metrics Adequate to Evaluate the Quality of Geometric Objects? Bernice E. Rogowitz and Holly E. Rushmeier IBM TJ Watson Research Center, P.O. Box 704, Yorktown Heights, NY USA ABSTRACT
More informationCommunication Door: Real-Time Communication Middleware
Special Issue Advanced Technologies and Solutions toward Ubiquitous Network Society Communication Door: Real-Time Communication Middleware By Masahiro TABUCHI,* Kazuaki NAKAJIMA,* Akiko KAWAMOTO,* Takashi
More informationUHD als Format-agnostisches Produktionsverfahren und als eine Plattform für interaktives Video
UHD als Format-agnostisches Produktionsverfahren und als eine Plattform für interaktives Video, Einsteinufer 37, 10587 Berlin, www.hhi.fraunhofer.de 1 Immersive Telepresence Systems Immersive sound Immersive
More informationBACnet for Video Surveillance
The following article was published in ASHRAE Journal, October 2004. Copyright 2004 American Society of Heating, Refrigerating and Air-Conditioning Engineers, Inc. It is presented for educational purposes
More informationStudy and Implementation of Video Compression Standards (H.264/AVC and Dirac)
Project Proposal Study and Implementation of Video Compression Standards (H.264/AVC and Dirac) Sumedha Phatak-1000731131- sumedha.phatak@mavs.uta.edu Objective: A study, implementation and comparison of
More informationCharacter Animation from 2D Pictures and 3D Motion Data ALEXANDER HORNUNG, ELLEN DEKKERS, and LEIF KOBBELT RWTH-Aachen University
Character Animation from 2D Pictures and 3D Motion Data ALEXANDER HORNUNG, ELLEN DEKKERS, and LEIF KOBBELT RWTH-Aachen University Presented by: Harish CS-525 First presentation Abstract This article presents
More informationDisambiguation of Horizontal Direction for Video Conference Systems
Disambiguation of Horizontal Direction for Video Conference Systems Mabel Mengzi Zhang, Seth Rotkin, and Jürgen P. Schulze mabel.m.zhang@gmail.com, sethrotkin@gmail.com, jschulze@ucsd.edu University of
More informationStudy and Implementation of Video Compression standards (H.264/AVC, Dirac)
Study and Implementation of Video Compression standards (H.264/AVC, Dirac) EE 5359-Multimedia Processing- Spring 2012 Dr. K.R Rao By: Sumedha Phatak(1000731131) Objective A study, implementation and comparison
More information3 Image-Based Photo Hulls. 2 Image-Based Visual Hulls. 3.1 Approach. 3.2 Photo-Consistency. Figure 1. View-dependent geometry.
Image-Based Photo Hulls Greg Slabaugh, Ron Schafer Georgia Institute of Technology Center for Signal and Image Processing Atlanta, GA 30332 {slabaugh, rws}@ece.gatech.edu Mat Hans Hewlett-Packard Laboratories
More informationColor Segmentation Based Depth Image Filtering
Color Segmentation Based Depth Image Filtering Michael Schmeing and Xiaoyi Jiang Department of Computer Science, University of Münster Einsteinstraße 62, 48149 Münster, Germany, {m.schmeing xjiang}@uni-muenster.de
More informationFalse alarm in outdoor environments
Accepted 1.0 Savantic letter 1(6) False alarm in outdoor environments Accepted 1.0 Savantic letter 2(6) Table of contents Revision history 3 References 3 1 Introduction 4 2 Pre-processing 4 3 Detection,
More informationA Survey of Video Processing with Field Programmable Gate Arrays (FGPA)
A Survey of Video Processing with Field Programmable Gate Arrays (FGPA) Heather Garnell Abstract This paper is a high-level, survey of recent developments in the area of video processing using reconfigurable
More informationMobiX3D: a player for displaying 3D content on mobile devices
MobiX3D: a player for displaying 3D content on mobile devices Daniele Nadalutti, Luca Chittaro, Fabio Buttussi HCI Lab Dept. of Math and Computer Science University of Udine via delle Scienze, 206 33100
More informationCisco TelePresence TX1300 Series
Data Sheet Cisco TelePresence TX1300 Series Product Overview The Cisco TelePresence TX1300 Series extends the experience pioneered by Cisco s fully immersive threescreen solutions to smaller general purpose
More informationCCTV & Video Surveillance over 10G ip
CCTV & Video Surveillance over 10G ip Background With the increase in data, research and development and corporate competition, many companies are realizing the need to not only protect their data, but
More informationGo to contents 18 3D Visualization of Building Services in Virtual Environment
3D Visualization of Building Services in Virtual Environment GRÖHN, Matti Gröhn; MANTERE, Markku; SAVIOJA, Lauri; TAKALA, Tapio Telecommunications Software and Multimedia Laboratory Department of Computer
More informationA General Framework for Tracking Objects in a Multi-Camera Environment
A General Framework for Tracking Objects in a Multi-Camera Environment Karlene Nguyen, Gavin Yeung, Soheil Ghiasi, Majid Sarrafzadeh {karlene, gavin, soheil, majid}@cs.ucla.edu Abstract We present a framework
More informationMPEG-H Audio System for Broadcasting
MPEG-H Audio System for Broadcasting ITU-R Workshop Topics on the Future of Audio in Broadcasting Jan Plogsties Challenges of a Changing Landscape Immersion Compelling sound experience through sound that
More informationVideo Conferencing. Femi Alabi UNC-CH - Comp 523 November 22, 2010
Video Conferencing Femi Alabi UNC-CH - Comp 523 November 22, 2010 Introduction Videoconferencing What Is It? Videoconferencing is a method of communicating between two or more locations where sound, vision
More informationInternet Desktop Video Conferencing
Pekka Isto 13.11.1998 1(8) Internet Desktop Video Conferencing ABSTRACT: This is report outlines possible use of Internet desktop videoconferencing software in a distributed engineering project and presents
More informationSubjective evaluation of a 3D videoconferencing system
Subjective evaluation of a 3D videoconferencing system Hadi Rizek 1, Kjell Brunnström 1,3, Kun Wang 1,3,Börje Andrén 1 and Mathias Johanson 2 1 Dept. of NetLab: Visual Media Quality, Acreo Swedish ICT
More informationA Short Introduction to Computer Graphics
A Short Introduction to Computer Graphics Frédo Durand MIT Laboratory for Computer Science 1 Introduction Chapter I: Basics Although computer graphics is a vast field that encompasses almost any graphical
More informationDisambiguation of Horizontal Direction for Video Conference Systems
Disambiguation of Horizontal Direction for Video Conference Systems Mabel Mengzi Zhang, Seth Rotkin, and Jürgen P. Schulze University of California San Diego 9500 Gilman Dr, La Jolla, CA 92093 {mabel.m.zhang,sethrotkin}@gmail.com,
More informationInternational Journal of Advanced Information in Arts, Science & Management Vol.2, No.2, December 2014
Efficient Attendance Management System Using Face Detection and Recognition Arun.A.V, Bhatath.S, Chethan.N, Manmohan.C.M, Hamsaveni M Department of Computer Science and Engineering, Vidya Vardhaka College
More informationComputer Graphics AACHEN AACHEN AACHEN AACHEN. Public Perception of CG. Computer Graphics Research. Methodological Approaches - - - - - - - - - -
Public Perception of CG Games Computer Graphics Movies Computer Graphics Research algorithms & data structures fundamental continuous & discrete mathematics optimization schemes 3D reconstruction global
More informationVIDEO CHAT & CUTOUT. ADSC Research Highlight
ADSC Research Highlight VIDEO CHAT & CUTOUT Researchers at the Advanced Digital Sciences Center (ADSC) have achieved a major advance in the state of the art for real-time, robust video object cutout, and
More informationInteractive Cards A game system in Augmented Reality
Interactive Cards A game system in Augmented Reality João Alexandre Coelho Ferreira, Instituto Superior Técnico Abstract: Augmented Reality can be used on innumerous topics, but the point of this work
More informationParametric Comparison of H.264 with Existing Video Standards
Parametric Comparison of H.264 with Existing Video Standards Sumit Bhardwaj Department of Electronics and Communication Engineering Amity School of Engineering, Noida, Uttar Pradesh,INDIA Jyoti Bhardwaj
More informationA Cognitive Approach to Vision for a Mobile Robot
A Cognitive Approach to Vision for a Mobile Robot D. Paul Benjamin Christopher Funk Pace University, 1 Pace Plaza, New York, New York 10038, 212-346-1012 benjamin@pace.edu Damian Lyons Fordham University,
More informationWHITE PAPER Personal Telepresence: The Next Generation of Video Communication. www.vidyo.com 1.866.99.VIDYO
WHITE PAPER Personal Telepresence: The Next Generation of Video Communication www.vidyo.com 1.866.99.VIDYO 2009 Vidyo, Inc. All rights reserved. Vidyo is a registered trademark and VidyoConferencing, VidyoDesktop,
More informationChapter 3 ATM and Multimedia Traffic
In the middle of the 1980, the telecommunications world started the design of a network technology that could act as a great unifier to support all digital services, including low-speed telephony and very
More informationMultimedia Data Transmission over Wired/Wireless Networks
Multimedia Data Transmission over Wired/Wireless Networks Bharat Bhargava Gang Ding, Xiaoxin Wu, Mohamed Hefeeda, Halima Ghafoor Purdue University Website: http://www.cs.purdue.edu/homes/bb E-mail: bb@cs.purdue.edu
More informationANIMATION a system for animation scene and contents creation, retrieval and display
ANIMATION a system for animation scene and contents creation, retrieval and display Peter L. Stanchev Kettering University ABSTRACT There is an increasing interest in the computer animation. The most of
More informationA Realistic Video Avatar System for Networked Virtual Environments
A Realistic Video Avatar System for Networked Virtual Environments Vivek Rajan, Satheesh Subramanian, Damin Keenan Andrew Johnson, Daniel Sandin, Thomas DeFanti Electronic Visualization Laboratory University
More informationVideo Coding Technologies and Standards: Now and Beyond
Hitachi Review Vol. 55 (Mar. 2006) 11 Video Coding Technologies and Standards: Now and Beyond Tomokazu Murakami Hiroaki Ito Muneaki Yamaguchi Yuichiro Nakaya, Ph.D. OVERVIEW: Video coding technology compresses
More informationAuthors: Masahiro Watanabe*, Motoi Okuda**, Teruo Matsuzawa*** Speaker: Masahiro Watanabe**
Visualization of the Blood flow and the Stress distribution with the Diagnostic Support System for Circulatory Disease in the Volume Communications Environment Authors: Masahiro Watanabe*, Motoi Okuda**,
More informationVideo Collaboration & Application Sharing Product Overview
. Video Collaboration & Application Sharing Product Overview Overview NPL s Collaborative Real-Time Information Sharing Platform (CRISP ) combines high quality video collaboration, remote application sharing
More informationEye Contact in Leisure Video Conferencing. Annick Van der Hoest & Dr. Simon McCallum Gjøvik University College, Norway.
Eye Contact in Leisure Video Conferencing Annick Van der Hoest & Dr. Simon McCallum Gjøvik University College, Norway 19 November 2012 Abstract This paper presents systems which enable eye contact in leisure
More informationVRSPATIAL: DESIGNING SPATIAL MECHANISMS USING VIRTUAL REALITY
Proceedings of DETC 02 ASME 2002 Design Technical Conferences and Computers and Information in Conference Montreal, Canada, September 29-October 2, 2002 DETC2002/ MECH-34377 VRSPATIAL: DESIGNING SPATIAL
More informationImplementation of Video Voice over IP in Local Area Network Campus Environment
Implementation of Video Voice over IP in Local Area Network Campus Environment Mohd Nazri Ismail Abstract--In this research, we propose an architectural solution to integrate the video voice over IP (V2oIP)
More informationSpeed Performance Improvement of Vehicle Blob Tracking System
Speed Performance Improvement of Vehicle Blob Tracking System Sung Chun Lee and Ram Nevatia University of Southern California, Los Angeles, CA 90089, USA sungchun@usc.edu, nevatia@usc.edu Abstract. A speed
More informationMultiple Description Coding (MDC) and Scalable Coding (SC) for Multimedia
Multiple Description Coding (MDC) and Scalable Coding (SC) for Multimedia Gürkan Gür PhD. Candidate e-mail: gurgurka@boun.edu.tr Dept. Of Computer Eng. Boğaziçi University Istanbul/TR ( Currenty@UNITN)
More informationEffects of microphone arrangement on the accuracy of a spherical microphone array (SENZI) in acquiring high-definition 3D sound space information
November 1, 21 1:3 1 Effects of microphone arrangement on the accuracy of a spherical microphone array (SENZI) in acquiring high-definition 3D sound space information Shuichi Sakamoto 1, Jun ichi Kodama
More informationTopic Maps Visualization
Topic Maps Visualization Bénédicte Le Grand, Laboratoire d'informatique de Paris 6 Introduction Topic maps provide a bridge between the domains of knowledge representation and information management. Topics
More informationInteractive Multimedia Courses-1
Interactive Multimedia Courses-1 IMM 110/Introduction to Digital Media An introduction to digital media for interactive multimedia through the study of state-of-the-art methods of creating digital media:
More informationVirtual Data Gloves : Interacting with Virtual Environments through Computer Vision
Virtual Data Gloves : Interacting with Virtual Environments through Computer Vision Richard Bowden (1), Tony Heap(2), Craig Hart(2) (1) Dept of M & ES (2) School of Computer Studies Brunel University University
More informationSYNTHESIZING FREE-VIEWPOINT IMAGES FROM MULTIPLE VIEW VIDEOS IN SOCCER STADIUM
SYNTHESIZING FREE-VIEWPOINT IMAGES FROM MULTIPLE VIEW VIDEOS IN SOCCER STADIUM Kunihiko Hayashi, Hideo Saito Department of Information and Computer Science, Keio University {hayashi,saito}@ozawa.ics.keio.ac.jp
More informationDesign and implementation of IPv6 multicast based High-quality Videoconference Tool (HVCT) *
Design and implementation of IPv6 multicast based High-quality conference Tool (HVCT) * Taewan You, Hosik Cho, Yanghee Choi School of Computer Science & Engineering Seoul National University Seoul, Korea
More informationIntroduction. C 2009 John Wiley & Sons, Ltd
1 Introduction The purpose of this text on stereo-based imaging is twofold: it is to give students of computer vision a thorough grounding in the image analysis and projective geometry techniques relevant
More informationA Tool for Multimedia Quality Assessment in NS3: QoE Monitor
A Tool for Multimedia Quality Assessment in NS3: QoE Monitor D. Saladino, A. Paganelli, M. Casoni Department of Engineering Enzo Ferrari, University of Modena and Reggio Emilia via Vignolese 95, 41125
More informationApplications that Benefit from IPv6
Applications that Benefit from IPv6 Lawrence E. Hughes Chairman and CTO InfoWeapons, Inc. Relevant Characteristics of IPv6 Larger address space, flat address space restored Integrated support for Multicast,
More informationProposal for a Virtual 3D World Map
Proposal for a Virtual 3D World Map Kostas Terzidis University of California at Los Angeles School of Arts and Architecture Los Angeles CA 90095-1467 ABSTRACT The development of a VRML scheme of a 3D world
More informationhigh-quality surround sound at stereo bit-rates
FRAUNHOFER Institute For integrated circuits IIS MPEG Surround high-quality surround sound at stereo bit-rates Benefits exciting new next generation services MPEG Surround enables new services such as
More informationPHOTOGRAMMETRIC TECHNIQUES FOR MEASUREMENTS IN WOODWORKING INDUSTRY
PHOTOGRAMMETRIC TECHNIQUES FOR MEASUREMENTS IN WOODWORKING INDUSTRY V. Knyaz a, *, Yu. Visilter, S. Zheltov a State Research Institute for Aviation System (GosNIIAS), 7, Victorenko str., Moscow, Russia
More informationBasic Theory of Intermedia Composing with Sounds and Images
(This article was written originally in 1997 as part of a larger text to accompany a Ph.D. thesis at the Music Department of the University of California, San Diego. It was published in "Monochord. De
More informationAn Instructional Aid System for Driving Schools Based on Visual Simulation
An Instructional Aid System for Driving Schools Based on Visual Simulation Salvador Bayarri, Rafael Garcia, Pedro Valero, Ignacio Pareja, Institute of Traffic and Road Safety (INTRAS), Marcos Fernandez
More informationUSING COMPUTER VISION IN SECURITY APPLICATIONS
USING COMPUTER VISION IN SECURITY APPLICATIONS Peter Peer, Borut Batagelj, Franc Solina University of Ljubljana, Faculty of Computer and Information Science Computer Vision Laboratory Tržaška 25, 1001
More informationCisco Video Collaboration Guide. Select the Right Equipment and Vendor
Cisco Video Collaboration Guide Select the Right Equipment and Vendor See How Video Collaboration Fits Your Needs Camera. Microphone. Monitor. Speaker. Codec. These are the five essential components that
More informationCisco Telepresence Implementation for Telekom s Corporate Requirements
Cisco Telepresence Implementation for Telekom s Corporate Requirements Zdravko Stafilov 1 1 Makedonski Telekom, Orce Nikolov bb, 1000 Skopje, Macedonia Abstract. The Cisco Telepresence system was the platform
More informationProduct Specifications
Data Sheet Cisco TelePresence TX9000 Series Product Overview The Cisco TelePresence TX9000 Series (TX9000) is the new benchmark for immersive in-person collaboration. With modern design and advanced functions,
More information3D Videocommunication
3D Videocommunication Algorithms, concepts and real-time systems in human centred communication EDITED BY Oliver Schreer Fraunhofer Institute for Telecommunications Heinrich-Hertz-Institut, Berlin, Germany
More information3D Client Software - Interactive, online and in real-time
3D Client Software - Interactive, online and in real-time Dipl.Inform.Univ Peter Schickel CEO Bitmanagement Software Vice President Web3D Consortium, Mountain View, USA OGC/Web3D liaison manager Presentation
More informationCHAPTER 6 TEXTURE ANIMATION
CHAPTER 6 TEXTURE ANIMATION 6.1. INTRODUCTION Animation is the creating of a timed sequence or series of graphic images or frames together to give the appearance of continuous movement. A collection of
More informationA method of generating free-route walk-through animation using vehicle-borne video image
A method of generating free-route walk-through animation using vehicle-borne video image Jun KUMAGAI* Ryosuke SHIBASAKI* *Graduate School of Frontier Sciences, Shibasaki lab. University of Tokyo 4-6-1
More informationH.264 Based Video Conferencing Solution
H.264 Based Video Conferencing Solution Overview and TMS320DM642 Digital Media Platform Implementation White Paper UB Video Inc. Suite 400, 1788 west 5 th Avenue Vancouver, British Columbia, Canada V6J
More informationCircle Object Recognition Based on Monocular Vision for Home Security Robot
Journal of Applied Science and Engineering, Vol. 16, No. 3, pp. 261 268 (2013) DOI: 10.6180/jase.2013.16.3.05 Circle Object Recognition Based on Monocular Vision for Home Security Robot Shih-An Li, Ching-Chang
More informationUsing Photorealistic RenderMan for High-Quality Direct Volume Rendering
Using Photorealistic RenderMan for High-Quality Direct Volume Rendering Cyrus Jam cjam@sdsc.edu Mike Bailey mjb@sdsc.edu San Diego Supercomputer Center University of California San Diego Abstract With
More informationVideo compression: Performance of available codec software
Video compression: Performance of available codec software Introduction. Digital Video A digital video is a collection of images presented sequentially to produce the effect of continuous motion. It takes
More informationInternet Video Streaming and Cloud-based Multimedia Applications. Outline
Internet Video Streaming and Cloud-based Multimedia Applications Yifeng He, yhe@ee.ryerson.ca Ling Guan, lguan@ee.ryerson.ca 1 Outline Internet video streaming Overview Video coding Approaches for video
More informationThe Flat Shape Everything around us is shaped
The Flat Shape Everything around us is shaped The shape is the external appearance of the bodies of nature: Objects, animals, buildings, humans. Each form has certain qualities that distinguish it from
More information