Piavca: A Framework for Heterogeneous Interactions with Virtual Characters
|
|
|
- Nigel Blake
- 9 years ago
- Views:
Transcription
1 Piavca: A Framework for Heterogeneous Interactions with Virtual Characters Marco Gillies 1 Xueni Pan 2 Mel Slater 3 1 Goldsmiths, University of London Department of Computing, London, UK [email protected] 2 University College London, Department of Computer Science, London UK 3 ICREA-Universitat de Barcelona, Barcelona, Spain. October 1, 2009 Abstract This paper presents a virtual character animation system for real time multimodal interaction in an immersive virtual reality setting. Human to human interaction is highly multimodal, involving features such as verbal language, tone of voice, facial expression, gestures and gaze. This multimodality means that, in order to simulate social interaction, our characters must be able to handle many different types of interaction, and many different types of animation, simultaneously. Our system is based on a model of animation that represents different types of animations as instantiations of an abstract function representation. This makes it easy to combine different types of animation. It also encourages the creation of behavior out of basic building blocks. making it easy to create and configure new beahviors for novel situations. The model has been implemented in Piavca, an open source character animation system. 1 Introduction Animated virtual humans are a vital part of many virtual environments today. Of particular interest are virtual human that we can interact with in 1
2 some approximation of social interaction. However, creating characters that we can interact with believably is an extremely complex problem. Part of this difficulty is that human interaction is highly multimodal. The most obvious modality is speech, but even this can be divided into the verbal content and non-verbal aspects of speech, such as tone of voice, whose function can be very subtle and complex. When we take into account bodily modalities we also have to deal with facial expression, gaze, gesture (both accompanying speech and giving feedback while listening), posture, body movements and touch. While non-immersive environments can ignore some of these modalities of interactions, the full power of an immersive virtual reality interaction with a character can only be achieved by modelling all (of most) of the modalities. This multimodality implies that there will be a great variety of ways of interacting with a virtual character. A human participant can use these different ways of interacting with a character: Verbal interaction, characters that can engage in conversation using a dialogue engine. Non-verbal aspects of speech, picking up features of the participants voice. What Slater[29] calls Body Centered Interaction. The participants movements are tracked, and their normal movements used to interact with a character. For example, the character may track the participants movements with their gaze and maintain a normal conversational distance to them. Control, the character might be an avatar which is being controlled by a participant, either through one of the above modalities or a more conventional user interface such as a mouse and keyboard or a joystick. Watching, not all of a characters behavior will be interactive, some will simply play back and the participant can observe it. The true complexity of the behavior of an interactive character is that most of these types of interactions are likely to be happening simultaneously. For example, a participant might be engaging in conversation with a character that is controlled by a chat-bot. The character nods in rhythm with variations in the participants speech amplitude and its gaze follows the participants position, while it walks around following the joystick movements made by another participant. Finally, its posture shifts occasionally, following a noninteractive algorithm. (Admittedly, this example is contrived, if the character 2
3 is an avatar under the control of a participant, then its speech will almost always be controlled by that participant. However, even given this constraint, many modalities of interaction are likely to appear simultaneously). This paper presents a method of creating characters that combine these very diverse forms of interaction. The diverse styles of interaction also imply diverse methods of generating behavior. This paper is mostly restricted to animation, but there are still many different styles. Some animation can be played back from preexisting data, whether it is from motion capture or hand animation. Some types of animation, such as gaze or lip synchronization are best generated on the fly, algorithmically, a process called procedural animation. Finally, most interactive animation is generated by transforming and combing clips of pre-existing animation data, to produce new animations. Another source of diversity is the different time scales of interactivity. A character s gaze has to respond instantly to changes in the humans position, while other body centered interactions have more variable time constraints. Speech interaction tends to be turn based, with long periods of non-interaction, interrupted by a change of speaker. Other modalities have no time constraints for interaction, or are not interactive at all. This flexibility in level of interactivity is also important because of the great difference in quality that exists between behavior that is generated on the fly and prerecorded behavior. This is particularly true with speech where it can be difficult to produce coherent speech for long periods with all but the best existing dialogue systems, and where synthesized speech falls far short of recorded speech quality. However, it is also true of animation, motion captured or hand animated animation is generally better than what can be produced in real time, even by transforming motion capture data. However, completely pre-recorded behavior reduces interactivity. The sense that we are truly interacting with a character is likely to be a strong contributor to presence and thus is vital. However, the illusion of interactivity can be maintained even if the behavior is not entirely interactive, as long as some elements are. For this reason, we believe it is important to balance the quality of pre-recorded elements. Behavior that consists of some pre-recorded clips, of audio and animation, should be used to ensure quality of the output. However, the sense of interactivity can be maintained to a degree if other aspects of behavior such as gaze and feedback, remain highly interactive. In turn this sense of interactivity can maintain a high level of presence. This flexibility also implies that the characters will need to be used in different ways and that different styles of interaction will need to be combined differently in different situations. Different types of behavior need to 3
4 be interacted with in different ways. Sometimes facial expressions need to be under the control of a human, at other times they must respond to the participants speech and at other times they can be random or scripted. The same factors apply to all the behavior types. Therefore it needs to be simple to create new interactive behaviors and combine them with others in a variety of different ways. This paper describes how to create a character capable of very heterogeneous forms of interaction. We first describe related work in this area. We then describe a functional model of animation that is designed for handling very diverse styles of animation, and its implementation with the Piavca open source character animation system. Finally we describe how it is used to create our heterogeneous character. 2 Related Work This work builds on a long tradition of research on expressive virtual characters[34]. This work has aimed at building animated characters that can autonomously produce the type of expressive non-verbal communication that humans naturally use in day to day interaction. This generally entails both an animation system and a higher level model for determining what behavior to produce in response to stimuli. Numerous general purpose systems have been produced notably the JACK system by Badler et al.[4]; the GRETA system by Pelachaud and colleagues[27]; work of Cassell et al.[7], and Guye-Vuilléme et al.[15]. Much of this work has concerned expression of emotions (for example the work of Gratch and Marsella[14]), but our work is closer to research that models use of non-verbal communication in face-to-face conversation, for example the work of Vilhjálmsson[32]. Each individual modality of expression is highly complex and many researchers have worked to create models of single modalities. Gaze is often linked to turn taking in conversation, for example the work of Lee, Badler and Badler[23], further developed by Vinayagamoorthy et al[33]. Facial expression has been extensively studied in the context of emotional expression, for example the work of Kshirsagar and Magnenat-Thalmann[21] or The Duy Bui [5]. Gestures are very closely related to speech[8] and have been modeled in a number of ways, from the highly data driven methods of Stone et al.[30] to the totally procedural methods of Kopp et al[19, 18]. Finally, posture is also an important modality of expression that has been used as a means of producing believable idling behaviour[9] or of expressing interpersonal relationships [12]. 4
5 One of the first systems that allows people to interact multimodally with a character using voice and gestures tracking was Thórisson s Gandalf system[31]. This work was later developed by Cassell s group into an interaction system with a full bodied character capable of complex non-verbal communication[6]. The Max system focus primarily on voice and gesture interaction[17]. Work by Maatman et al.[24] focused on the listening behaviour of a character, and like our work uses head tracking and voice input. These systems have demonstrated the power of multimodal face-to-face interaction with virtual characters. This paper shows how it is possible to rapidly create and customise such systems from basic building blocks. Our work also uses research in the area of data driven animation and motion editing. In particular the Motion Warping formulation of motion editing [35]. Our functional model is an excellent way of combining motion editing techniques, and we have implemented several within our framework. Examples include, interpolation based animation such as the work of Rose et al.[28], Principal Component Analysis based animation [1] and the Motion Graph data structure[3, 20, 22]. The work we present is a functional abstraction of character animation and behavior. In this sense it is similar to functional abstraction frameworks used in other domains. In particular there are a number of interesting abstraction frameworks used in virtual reality and graphics for example Figueroa et al. s InTML system[11] or Elliott et al. s TBAG system[10]. 3 A Functional Model of Animation Handling and combining many diverse methods of animation on a single character requires a single representation for all of them. At its most abstract, an animation can be viewed as a function of time to the state of a character: x = f (t) (1) where x is some representation of a characters state. The most common representation in body animation would be a skeletal one, in which the root of the character has a vector position p 0 and quaternion orientation q 0 and all of the joints have quaternion orientations (q i ): {p 0,q 0,q 1,...q n }. However, we do not restrict animations to this representation, for example joint orientations can also be represented using Grassia s exponential map representation[13] if it is more convenient for certain calculations. For facial animation the state can be represented as a set of weight values for morph targets or as positions or rotations of facial bones. The state can also 5
6 have more abstract representations, for example, the result of one animation function can be used as the output of another, as we shall see below. Most animations, will in fact take other parameters making this the general form of an animation function: x = f (θ,t) (2) We will now show how different types of animation can easily be represented in this framework. Keyframe or Motion Captured Animation Both hand animated data and motion captured animation are represented as a list of evenly or unevenly spaced keyframes. This can be represented as the following function: f (k,t) = interpolate(k τ(t),k τ(t)+1,t) Where k is the keyframe data and τ(t) is the keyframe time prior to t. interpolate can be any suitable interpolation function, we use cubic spline interpolation. Procedural Animation Procedural animation is the most general instantiation of the functional model, and can be represented as any function of the form (2). The parameters θ will depend heavily on the type of motion, for example, a gaze motion will have parameters the include the target of gaze and the length of gaze. Motion Transforms An animation function can be a transformation of another motion function g: x = f (g,θ,t) The most general form of transformation is a general motion warp [35], which consists of a timewarp, α, tranforming t and a space warp, β transforming the output of g: f (g,θ α,θ β,t) = β(θ β,g(α(θ α,t)) Simple examples include, spatial scaling 1 : f (g,s,t) = sg(t) ; temporal scaling: f (g,s,t) = g(st), and looping f (g,t) = g(t mod g ), where g is the length of g (in practice a more complex looping function is used to ensure 1 for quaternion animations multiplication is replaced by scaling the rotation angle by s 6
7 smooth transitions). A paricularly useful transform is a mask, which is used to select certain joint or morph targets to which to apply an animation, based on a mask, m: { g(t) i if m i = 1, f (g,m,t) i = 0 otherwise As mentioned above the parameters of a motion transform can themselves be other motion functions. For example, a general motion warp can be created by using a motion function, h as the warping parameter: f (g,h,t) = g(h(t)). Combining Animations Animations can also be combined together by functions of two other animations, which have the general form: f (g 1,g2,θ β,θ α1,θ α2,t) = β(θ β,g 1 (α 1 (θ α1,t)),g 1 (α 2 (θ α2,t))) Examples include addition 2 : f (g 1,g2,t) = g 1 (t) + g2(t); blending between animations: f (g 1,g2,λ,t) = λg 1 (t)+(1 λ)g2(t) and sequencing motions: { g 1 (t) if t < g 1, f (g 1,g2,t) = g 2 (t g 1 ) otherwise As with looped motions, in practice we would used a smoothed version of this function. There are also functions for combining multiple animations, for example blending between several animations; finite state machines that choose different animations based on their state; animations based on a Principal Component Analysis of other animations [1], and motion graphs described below. The power of the functional model of animation comes from the fact that all types of animation have the same form. The animation functions can be curried, pre-applying all parameters apart from t to get a new function of the form (1). Once in this form the user does not need to know anything about the type of animation. In particular the functions for transforming and combining animations have the same form as their inputs, making it possible to compose them. As we shall see arbitrary composition of functions for transforming and combining animations can be a powerful tool for creating complex transformations. This leads to a methodolgy of decomposing transformations into their most basic elements, which can then be reused by composing them in a number of different ways. This in turn makes it easy to author new behaviours, and variants of existing behaviours. 2 for quaternions, quaternion multiplication is used instead of addition 7
8 3.1 Implementation Before discussing how the functional model is used, a quick note on how it is implemented. The model has been implemented as part of Piavca (the Platform Independent Architecture for Virtual Characters and Avatars), an open source software framework for creating virtual characters (available at Piavcs is a generic API built in C++ that can be combined with different graphics engine. The current implementation uses Cal3d ( as a low level animation system providing functionality such as smooth skinning and morph targets while Piavca overrides the motion blending features. The renderer is based on Cal3d s OpenGL renderer. This system can be used with numerous graphics and virtual reality systems. Currently it is implemented on Dive ( XVR ( and we are working on an implementation for OpenSG ( Our implementation (in the C++ and Python languages) is object oriented. All of the functions are in fact function objects, and so can contain state. They all inherit from a single base class, Motion, which acts as an abstract representation of an animation function. Currying the functions is achieved by providing all parameters except t at initialization time, and making the function objects callable with t as a parameter. Users can configure and combine motion functions using a Python based scripting interface to Piavca, for maximum flexibility. Alternatively there is an XML based behavior definition language that allows users to create character behavior models without requiring programming skills. The behavior definition language exactly mirrors the functional model with tags for each possible function. 4 A Heterogeneous character system Figure 1 shows the type of heterogeneous interaction that is possible with our characters. The human participant s behavior is input with typical sensors for an immersive VR system, a microphone and head tracker. However, these two inputs are used in a variety of different ways by different behaviors. The head tracker is used to obtain the position of the participant in order to maintain an appropriate conversational distance (maintaining distance in conversation is called proxemics in the non-verbal communication literature). It is also used to detect when the participant shifts posture. The character s posture shifts are then synchronized with those of the participant, which is know to build rapport[16]. The position of the participant is 8
9 Head Tracker Microphone Proxemics Posture Shifts Gaze Head Nods/Feedback Speech Gestures Scripted Actions Facial Expression Lip Synchronization Blinking Human Controller Dialogue System Figure 1: Heterogeneous Interaction. also used by a gaze behavior to ensure the character is looking in the right place. The audio from the microphone is used to detect when the participant is talking. This is used by the gaze behavior so the character can look at the participant more when listening to him or her. The character also gives head nods and other feedback signals when the participant is speaking. All these behaviors happen automatically in response to the sensor input, using real time algorithms. The microphone audio is also used for speech interaction. Speech interaction is either controlled by a human controller, if the character is an avatar, or by a dialogue engine. When the character speaks a number of other behaviors are triggered. The character s lip movements will be synchronized to the speech and the character will gesture. The gaze behavior is also altered to take account of the fact that the character is speaking[2]. As well as triggering speech the character s controller can also trigger certain scripted actions and gestures. Apart from lip synchronization and gaze, the character s facial expression is independent of the participants behavior, consisting of occasional smiling and blinking. The character therefore has a wide range of styles of interaction, all happening simultaneously. These contain many different animation processes, both facial and bodily, that must be combined to create a single coherent animation. Figure 2 shows how this type of character can be implemented using our functional model. The example we give is of a human controlled character that is used in Wizard of Oz style experiments. A human operator controls certain aspects of the behavior, while others are automatic. The character s speech is controlled by a human being (the controller) selecting speech sequences from a library of possible utterances, while the character interacts with another person (the participant). This system has 3 inputs: the position of the participant, an audio signal of the voice of the participant and input 9
10 Figure 2: Implementing a Heterogeneous Interaction. from the person controlling the character, specifying speech utterances. The position input is a 3-vector whose value is obtained every frame from a head tracker on the participant. The voice signal is obtained from an ordinary microphone. For this application we simply threshold the audio value to detect whether the participant is speaking. The controller has a user interface with a number of buttons used to trigger speech utterances. Some behaviors of the character are not influenced by any of the inputs. For example, the character has a simple blinking behavior. This is a loop containing a blinking animation sequenced with a zero animation. A zero animation is simply an animation function that returns zero for all joint or facial expression values and in this case it is used to model the inter-blinking period. The length of the zero motion is varied every time around the loop to ensure the timings are not too repetitive. A facial animation loop, in which the character smiles occasionally, is implemented similarly. The head 10
11 tracker input is used in a number of ways. The first is for posture shifts. During conversation people tend to synchronize their movements, particularly movement such as posture shifts[16]. This synchronization is a strong sign of rapport between individuals. In order to simulate this we detect posture shifts by finding large changes in position. We then trigger a posture shifts. The characters posture is modeled as a finite state machine animation in which each state is a different possible posture. On a posture shift a new state is chosen at random. A finite state machine animation performs a smooth sequence between the animations associated with each state, ensuring smooth posture shifts. The head tracker is also used by the proxemics behavior. Proxmemics is the use of space in social interaction. For our characters this means maintaining a comfortable distance from and orientation to the participant. The relative distance and angle of the participant to the character are calculated from the tracker position. If they are too large or small the character turns to face the participant, or takes a step forward or backward. Again this behavior is modeled as a finite state machine, with a default state being the zero motion and a state for each motion direction. The final use for the position input is to control the gaze behavior. The position gives a target to look at. The audio input is used to detect when the participant is speaking and give feedback behavior. In this implementation the feedback consists of occasional nodding to give encouragement. This is implemented as a loop in the same way as blinking. The other major input is from the controller, who can issue commands to control the character s speech. The behavior consists of a number of multi-modal utterances that can be triggered using a graphical user interface. Multi-modal utterances are short scripted behaviors that combined speech (in this case audio files) with animation elements. For example, the audio is accompanied by facial animation for lip synchronization and also appropriate gestures. The scripts give the creators of the character very tight control of the character s behavior and potentially high quality behavior can be created. This comes at the possible cost of some interactivity, however, we believe that our methodology of combining more scripted elements with real time interaction can combine be benefits of both. Figure 3 shows some still frames from an interaction with our virtual character, the accompanying video shows the actual interaction. The type of character set up we have described is only one way of interacting. For example, the character could be total autonomous with utterances triggered from an AI chat-bot system, or the character s speech could be directly taken from the controllers own voice. In the second case, many elements such as gestures and lip synchronization would have to be generated auto- 11
12 Figure 3: A real and virtual human interacting in an immersive virtual environment matically to suit the speech. Our frameworks makes it easy to build new styles of interaction from existing components. 5 Conclusion This paper has presented a software framework for creating interactive virtual characters. The many different types of behavior involved with human social interaction imply a range of different styles of animation and interaction with a character. Our framework allows us to unify and combine these diverse methods using a single abstract function representation. It makes it easy to create new character systems by combining different behavior modules in different ways. The framework has been release as part of the open source project Piavca ( we encourage readers to try out the functionality. This framework has been used for virtual reality experiments at University College London. These have demonstrated that people respond to the characters in some way as if they are human. For example Pan and Slater 12
13 [26] conducted an experiment in which socially phobic male participants interacted with a virtual woman that engaged them in conversation that became increasingly intimate. The experimenters measured skin conductance level which demonstrated that the intimate conversation resulted in greater arousal. Interestingly, proxemic behavior played an important role in this. At one point the character s proxemic distance was decreased resulting in her moving closer to the participant. This produced the highest skin conductance levels in the experiment. These quantitative results were supported by the participants subjective reports, with many reporting strong emotions such as anxiety or even guilt at cheating on their partner with a virtual woman. An analysis of the body movements of the participants during the scenario [25] showed that they used more social non-verbal cues such as nodding or cocking their heads during the conversation than before it. Future work on the project will involve increasing the range of functionality and of possible applications. In a modular system such as ours it is easy to add functionality by either added new animation functions or combining existing ones in new ways. As the system is applied to different situations and styles of interaction new requirements will naturally emerge and therefore drive the development of new functionality. We are currently applying the system to more graphically realistic character and making greater use of motion capture, raising the level of realism that is possible. This greater realism will itself bring new requirements to our animation framework. The most important change we are currently planning is the addition of a graphical user interface for combining behavior functions. This will supplement the existing scripting interface and definition language, and provided a more accessible method of creating characters. acknowledgements We would like to thank the funders of this work: BT plc, the European Union FET project PRESENCIA (contract number 27731) and the Empathic Avatars project funded by the UK Engineering and Physical Sciences Research Council. We also would like to thank the members of the University College London Department of Computer Science Virtual Environments and Graphics Group. 13
14 References [1] M. Alexa and W. Müller. Representing animations by principal components. Computer Graphics Forum, 19(3): , Aug [2] M. Argyle and M. Cook. Gaze and Mutual Gaze. Cambridge University Press, [3] O. Arikan and D. A. Forsyth. Interactive motion generation from examples. ACM Transactions on Graphics, 21(3): , July [4] N. Badler, C. Philips, and B. Webber, editors. Simulating Humans: Computer Graphics, Animation and Control. Oxford University Press, [5] T. Bui, D. Heylen, and A. Nijholt. Combination of facial movements on a 3d talking head. In Computer Graphics International 2004, pages , June [6] J. Cassell, T. Bickmore, L. Campbell, K. Chang, H. Vilhjálmsson, and H. Yan. Embodiment in conversational interfaces: Rea. In ACM SIGCHI, pages ACM Press, [7] J. Cassell, T. Bickmore, L. Campbell, K. Chang, H. Vilhjálmsson, and H. Yan. Requirements for an architecture for embodied conversational characters. In Computer Animation and Simulation 99, Sept [8] J. Cassell and M. Stone. Living hand to mouth: Psychological theories about speech and gesture in interactive dialogue systems. In AAI 1999 Fall Symposium on Psychological Models of Communication in Collaborative Systems, pages 34 42, [9] A. Egges, T. Molet, and N. Magnenat-Thalmann. Personalised realtime idle motion synthesis. In 12th Pacific Conference on Computer Graphics and Applications, pages , Oct [10] C. Elliott, G. Schechter, R. Yeung, and S. Abi-Ezzi. Tbag: A high level framework for interactive, animated 3d graphics applications. In Proceedings of SIGGRAPH 94, Computer Graphics Proceedings, Annual Conference Series, pages , July [11] P. Figueroa, M. Green, and H. J. Hoover. Intml: A description language for vr applications. In Proceedings of Web3D 02, [12] M. Gillies and D. Ballin. A model of interpersonal attitude and posture generation. In T. Rist, R. Aylett, D. Ballin, and J. Rickel, editors, Fourth Workshop on Intelligent Virtual Agents, Kloster Irsee, Germany, September
15 [13] F. S. Grassia. Practical parameterization of rotations using the exponential map. Journal of Graphics Tools, 3(3):29 48, [14] J. Gratch and S. Marsella. Tears and fears: modeling emotions and emotional behaviors in synthetic agents. In AGENTS 01: Proceedings of the fifth international conference on Autonomous agents, pages , New York, NY, USA, ACM Press. [15] A. Guye-Vuilléme, T.K.Capin, I.S.Pandzic, N. Magnenat-Thalmann, and D.Thalmann. Non-verbal communication interface for collaborative virtual environments. The Virtual Reality Journal, 4:49 59, [16] A. Kendon. Movement coordination in social interaction. Acta Psychologica, 32:1 25, [17] S. Kopp, B. Jung, N. Lessmann, and I. Wachsmuth. Max a multimodal assistant in virtual reality construction. KI-Knstliche Intelligenz, 3(4):11 17, [18] S. Kopp, P. Tepper, and J. Cassell. Towards integrated microplanning of language and iconic gesture for multimodal output. In International Conference on Multimodal Interfaces (ICMI 04), pages ACM Press, [19] S. Kopp and I. Wachsmuth. Synthesizing multimodal utterances for conversational agents. The Journal Computer Animation and Virtual Worlds, 15(1):39 52, [20] L. Kovar, M. Gleicher, and F. Pighin. Motion graphs. ACM Transactions on Graphics, 21(3): , July [21] S. Kshirsagar and N. Magnenat-Thalmann. A multilayer personality model. In 2nd International Symposium on Smart Graphics, pages , June [22] J. Lee, J. Chai, P. S. A. Reitsma, J. K. Hodgins, and N. S. Pollard. Interactive control of avatars animated with human motion data. ACM Transactions on Graphics, 21(3): , July [23] S. P. Lee, J. B. Badler, and N. I. Badler. Eyes alive. ACM Transactions on Graphics, 21(3): , July [24] R. M. Maatman, J. Gratch, and S. Marsella. Natural behavior of a listening agent. In T. Panayiotopoulos, J. Gratch, R. Aylett, D. Ballin, P. Olivier, and T. Rist, editors, Intelligent Virtual Agents, 5th International Working Conference, Kos, Greece, September
16 [25] X. Pan, M. Gillies, and M. Slater. Male bodily responses during an interaction with a virtual woman. In H. Prendinger, J. C. Lester, and M. Ishizuka, editors, Intelligent Virtual Agents, 8th International Conference, IVA 2008, Tokyo, Japan, September 1-3, Proceedings, volume 5208 of Lecture Notes in Computer Science, pages Springer, [26] X. Pan and M. Slater. A preliminary study of shy males interacting with a virtual female. In Presence 2007: The 10th Annual International Workshop on Presence, pages , [27] I. Poggi, C. Pelachaud, F. De Rosis, V. Carofiglio, and B. De Carolis. Greta. a believable embodied conversational agent. In S. O. and Z. M., editors, Multimodal Intelligent Information Presentation (Text, Speech and Language Technology Vol. 27). Kluwer, [28] C. Rose, M. F. Cohen, and B. Bodenheimer. Verbs and adverbs: Multidimensional motion interpolation. IEEE Computer Graphics & Applications, 18(5):32 40, September - October [29] M. Slater and M. Usoh. Body centred interaction in immersive virtual environments. In N. M. Thalmann and D. Thalmann, editors, Artificial Life and Virtual Reality, pages John Wiley and Sons, [30] M. Stone, D. DeCarlo, I. Oh, C. Rodriguez, A. Stere, A. Lees, and C. Bregler. Speaking with hands: creating animated conversational characters from recordings of human performance. In SIGGRAPH 04: ACM SIGGRAPH 2004 Papers, pages , New York, NY, USA, ACM Press. [31] K. Thórisson. Real-time decision making in multimodal face-to-face communication. In second ACM international conference on autonomous agents, pages 16 23, [32] H. H. Vilhjálmsson. Augmenting online conversation through automated discourse tagging. In 6th annual minitrack on Persistent Conversation at the 38th Hawaii International Conference on System Sciences, [33] V. Vinayagamoorthy, M. Garau, A. Steed, and M. Slater. An eye gaze model for dyadic interaction in an immersive virtual environment: Practice and experience. Computer Graphics Forum, 23(1):1 12, Mar [34] V. Vinayagamoorthy, M. Gillies, A. Steed, E. Tanguy, X. Pan, C. Loscos, and M. Slater. Building expression into virtual characters. In Eurographics Conference State of the Art Reports,
17 [35] A. Witkin and Z. Popović. Motion warping. In ACM SIGGRAPH, pages ,
Comparing and evaluating real-time character engines for virtual environments. [email protected]
real-time character engines 1 Running head: REAL-TIME CHARACTER ENGINES Comparing and evaluating real-time character engines for virtual environments Marco Gillies 1, Bernhard Spanlang 2 1 Department of
Emotional Communicative Body Animation for Multiple Characters
Emotional Communicative Body Animation for Multiple Characters Arjan Egges, Nadia Magnenat-Thalmann MIRALab - University of Geneva 24, Rue General-Dufour, 1211 Geneva, Switzerland Telephone: +41 22 379
The Relation between Gaze Behavior and the Attribution of Emotion: An Empirical Study
The Relation between Gaze Behavior and the Attribution of Emotion: An Empirical Study Brent Lance and Stacy C. Marsella University of Southern California Information Sciences Institute 4676 Admiralty Way
Voice Driven Animation System
Voice Driven Animation System Zhijin Wang Department of Computer Science University of British Columbia Abstract The goal of this term project is to develop a voice driven animation system that could take
Motion Retargetting and Transition in Different Articulated Figures
Motion Retargetting and Transition in Different Articulated Figures Ming-Kai Hsieh Bing-Yu Chen Ming Ouhyoung National Taiwan University [email protected] [email protected] [email protected]
Nonverbal Communication Human Communication Lecture 26
Nonverbal Communication Human Communication Lecture 26 Mar-14-11 Human Communication 1 1 Nonverbal Communication NVC can be communicated through gestures and touch (Haptic communication), by body language
Background Animation Generator: Interactive Scene Design based on Motion Graph
Background Animation Generator: Interactive Scene Design based on Motion Graph Tse-Hsien Wang National Taiwan University [email protected] Bin-Yu Chen National Taiwan University [email protected]
Motion Capture Assisted Animation: Texturing and Synthesis
Motion Capture Assisted Animation: Texturing and Synthesis Katherine Pullen Stanford University Christoph Bregler Stanford University Abstract We discuss a method for creating animations that allows the
A Study of Immersive Game Contents System Design and Modeling for Virtual Reality Technology
, pp.411-418 http://dx.doi.org/10.14257/ijca.2014.7.10.38 A Study of Immersive Game Contents System Design and Modeling for Virtual Reality Technology Jung-Yoon Kim 1 and SangHun Nam 2 1 Graduate School
Virtual Patients: Assessment of Synthesized Versus Recorded Speech
Virtual Patients: Assessment of Synthesized Versus Recorded Speech Robert Dickerson 1, Kyle Johnsen 1, Andrew Raij 1, Benjamin Lok 1, Amy Stevens 2, Thomas Bernard 3, D. Scott Lind 3 1 Department of Computer
Principal Components of Expressive Speech Animation
Principal Components of Expressive Speech Animation Sumedha Kshirsagar, Tom Molet, Nadia Magnenat-Thalmann MIRALab CUI, University of Geneva 24 rue du General Dufour CH-1211 Geneva, Switzerland {sumedha,molet,thalmann}@miralab.unige.ch
Bachelor of Games and Virtual Worlds (Programming) Subject and Course Summaries
First Semester Development 1A On completion of this subject students will be able to apply basic programming and problem solving skills in a 3 rd generation object-oriented programming language (such as
An Instructional Aid System for Driving Schools Based on Visual Simulation
An Instructional Aid System for Driving Schools Based on Visual Simulation Salvador Bayarri, Rafael Garcia, Pedro Valero, Ignacio Pareja, Institute of Traffic and Road Safety (INTRAS), Marcos Fernandez
Fundamentals of Computer Animation
Fundamentals of Computer Animation Principles of Traditional Animation How to create maximum impact page 1 How to create maximum impact Early animators worked from scratch to analyze and improve upon silence
An Interactive method to control Computer Animation in an intuitive way.
An Interactive method to control Computer Animation in an intuitive way. Andrea Piscitello University of Illinois at Chicago 1200 W Harrison St, Chicago, IL [email protected] Ettore Trainiti University of
Motion Capture Assisted Animation: Texturing and Synthesis
Motion Capture Assisted Animation: Texturing and Synthesis Katherine Pullen Stanford University Christoph Bregler Stanford University Abstract We discuss a method for creating animations that allows the
Computer Animation in Future Technologies
Computer Animation in Future Technologies Nadia Magnenat Thalmann MIRALab, University of Geneva Daniel Thalmann Computer Graphics Lab, Swiss Federal Institute of Technology Abstract In this introductory
CS 4204 Computer Graphics
CS 4204 Computer Graphics Computer Animation Adapted from notes by Yong Cao Virginia Tech 1 Outline Principles of Animation Keyframe Animation Additional challenges in animation 2 Classic animation Luxo
Facial Expression Analysis and Synthesis
1. Research Team Facial Expression Analysis and Synthesis Project Leader: Other Faculty: Post Doc(s): Graduate Students: Undergraduate Students: Industrial Partner(s): Prof. Ulrich Neumann, IMSC and Computer
A Framework for Rendering, Simulation and Animation of Crowds
CEIG 09, San Sebastián, Sept. 9-11 (2009) C. Andújar and J. Lluch (Editors) A Framework for Rendering, Simulation and Animation of Crowds N. Pelechano, B. Spanlang, A. Beacco Universitat Politècnica de
Modelling 3D Avatar for Virtual Try on
Modelling 3D Avatar for Virtual Try on NADIA MAGNENAT THALMANN DIRECTOR MIRALAB UNIVERSITY OF GENEVA DIRECTOR INSTITUTE FOR MEDIA INNOVATION, NTU, SINGAPORE WWW.MIRALAB.CH/ Creating Digital Humans Vertex
Go to contents 18 3D Visualization of Building Services in Virtual Environment
3D Visualization of Building Services in Virtual Environment GRÖHN, Matti Gröhn; MANTERE, Markku; SAVIOJA, Lauri; TAKALA, Tapio Telecommunications Software and Multimedia Laboratory Department of Computer
INTERACTIVE 3D SIMULATION SYSTEM IN GAME ENGINE BASED COLLABORATIVE VIRTUAL ENVIRONMENT FOR ARCHITECTURAL DESIGN COMMUNICATION
INTERACTIVE 3D SIMULATION SYSTEM IN GAME ENGINE BASED COLLABORATIVE VIRTUAL ENVIRONMENT FOR ARCHITECTURAL DESIGN COMMUNICATION PRITTIPORN LOPKERD, PINTO JINUNTUYA Faculty of Architecture and Planning,
VRSPATIAL: DESIGNING SPATIAL MECHANISMS USING VIRTUAL REALITY
Proceedings of DETC 02 ASME 2002 Design Technical Conferences and Computers and Information in Conference Montreal, Canada, September 29-October 2, 2002 DETC2002/ MECH-34377 VRSPATIAL: DESIGNING SPATIAL
How do non-expert users exploit simultaneous inputs in multimodal interaction?
How do non-expert users exploit simultaneous inputs in multimodal interaction? Knut Kvale, John Rugelbak and Ingunn Amdal 1 Telenor R&D, Norway [email protected], [email protected], [email protected]
Interpersonal Communication Skills
Interpersonal Communication Skills Which one of the following traits is most predictive of success? 1. Intelligence 2. Education 3. Experience 4. Knowledge Not intelligence, education, experience, or knowledge
Project 2: Character Animation Due Date: Friday, March 10th, 11:59 PM
1 Introduction Project 2: Character Animation Due Date: Friday, March 10th, 11:59 PM The technique of motion capture, or using the recorded movements of a live actor to drive a virtual character, has recently
FET PRESENCE Brainstorm on November 13, 2003: some views.
FET PRESENCE Brainstorm on November 13, 2003: some views. Danielle PELE, Dominique PAVY Olivier GACHIGNARD, Christian BOUVILLE Le présent document contient des informations qui sont la propriété de France
Teaching Portfolio. Teaching Philosophy
Teaching Portfolio Teaching Philosophy Over the course of my education, I have had the privilege of interacting with some truly excellent teachers, who have shaped my knowledge, reasoning, and technical
Social Signal Processing Understanding Nonverbal Behavior in Human- Human Interactions
Social Signal Processing Understanding Nonverbal Behavior in Human- Human Interactions A.Vinciarelli University of Glasgow and Idiap Research Institute http://www.dcs.gla.ac.uk/~ vincia e- mail: [email protected]
Open-Source, Cross-Platform Java Tools Working Together on a Dialogue System
Open-Source, Cross-Platform Java Tools Working Together on a Dialogue System Oana NICOLAE Faculty of Mathematics and Computer Science, Department of Computer Science, University of Craiova, Romania [email protected]
animation animation shape specification as a function of time
animation animation shape specification as a function of time animation representation many ways to represent changes with time intent artistic motion physically-plausible motion efficiency control typically
Mocap in a 3D Pipeline
East Tennessee State University Digital Commons @ East Tennessee State University Undergraduate Honors Theses 5-2014 Mocap in a 3D Pipeline Logan T. Maides Follow this and additional works at: http://dc.etsu.edu/honors
Anime Studio Debut 10 Create Your Own Cartoons & Animations!
Anime Studio Debut 10 Create Your Own Cartoons & Animations! Anime Studio Debut is your complete animation program for creating 2D movies, cartoons, anime or cut out animations. This program is fun, easy,
Choosing My Avatar & the Psychology of Virtual Worlds: What Matters?
Kaleidoscope Volume 11 Article 89 July 2014 Choosing My Avatar & the Psychology of Virtual Worlds: What Matters? Jennifer Wu Follow this and additional works at: http://uknowledge.uky.edu/kaleidoscope
Course Overview. CSCI 480 Computer Graphics Lecture 1. Administrative Issues Modeling Animation Rendering OpenGL Programming [Angel Ch.
CSCI 480 Computer Graphics Lecture 1 Course Overview January 14, 2013 Jernej Barbic University of Southern California http://www-bcf.usc.edu/~jbarbic/cs480-s13/ Administrative Issues Modeling Animation
Standard Languages for Developing Multimodal Applications
Standard Languages for Developing Multimodal Applications James A. Larson Intel Corporation 16055 SW Walker Rd, #402, Beaverton, OR 97006 USA [email protected] Abstract The World Wide Web Consortium
Performance Driven Facial Animation Course Notes Example: Motion Retargeting
Performance Driven Facial Animation Course Notes Example: Motion Retargeting J.P. Lewis Stanford University Frédéric Pighin Industrial Light + Magic Introduction When done correctly, a digitally recorded
Interactive Multimedia Courses-1
Interactive Multimedia Courses-1 IMM 110/Introduction to Digital Media An introduction to digital media for interactive multimedia through the study of state-of-the-art methods of creating digital media:
Disappearing Computers, Social Actors and Embodied Agents
Disappearing Computers, Social Actors and Embodied Agents Anton Nijholt Department of Computer Science, University of Twente, Enschede, The Netherlands. [email protected] Abstract Presently, there
On Intuitive Dialogue-based Communication and Instinctive Dialogue Initiative
On Intuitive Dialogue-based Communication and Instinctive Dialogue Initiative Daniel Sonntag German Research Center for Artificial Intelligence 66123 Saarbrücken, Germany [email protected] Introduction AI
ACE: After Effects CS6
Adobe Training Services Exam Guide ACE: After Effects CS6 Adobe Training Services provides this exam guide to help prepare partners, customers, and consultants who are actively seeking accreditation as
Spatial Social Behavior in Second Life
Spatial Social Behavior in Second Life Doron Friedman 1,2, Anthony Steed 1, and Mel Slater 1,3 1 University College London, {d.friedman,a.steed,m.slater}@cs.ucl.ac.uk 2 The Interdisciplinary Center, Herzliya,
To Gesture or Not to Gesture: What is the Question?
University of Pennsylvania ScholarlyCommons Center for Human Modeling and Simulation Department of Computer & Information Science June 2000 To Gesture or Not to Gesture: What is the Question? Norman I.
Building Interactive Animations using VRML and Java
Building Interactive Animations using VRML and Java FABIANA SALDANHA TAMIOSSO 1,ALBERTO BARBOSA RAPOSO 1, LÉO PINI MAGALHÃES 1 2,IVAN LUIZ MARQUES RICARTE 1 1 State University of Campinas (UNICAMP) School
An Animation Definition Interface Rapid Design of MPEG-4 Compliant Animated Faces and Bodies
An Animation Definition Interface Rapid Design of MPEG-4 Compliant Animated Faces and Bodies Erich Haratsch, Technical University of Munich, [email protected] Jörn Ostermann, AT&T Labs
game development documentation game development documentation: concept document
topics: game design documents design document references: cisc3665 game design fall 2011 lecture # IV.1 game development documentation notes from: Game Design: Theory & Practice (2nd Edition), by Richard
Enable Location-based Services with a Tracking Framework
Enable Location-based Services with a Tracking Framework Mareike Kritzler University of Muenster, Institute for Geoinformatics, Weseler Str. 253, 48151 Münster, Germany [email protected] Abstract.
Virtual Environments - Basics -
Virtual Environments - Basics - What Is Virtual Reality? A Web-Based Introduction Version 4 Draft 1, September, 1998 Jerry Isdale http://www.isdale.com/jerry/vr/whatisvr.html Virtual Environments allow
Develop Computer Animation
Name: Block: A. Introduction 1. Animation simulation of movement created by rapidly displaying images or frames. Relies on persistence of vision the way our eyes retain images for a split second longer
Template-based Eye and Mouth Detection for 3D Video Conferencing
Template-based Eye and Mouth Detection for 3D Video Conferencing Jürgen Rurainsky and Peter Eisert Fraunhofer Institute for Telecommunications - Heinrich-Hertz-Institute, Image Processing Department, Einsteinufer
DESIGNING MPEG-4 FACIAL ANIMATION TABLES FOR WEB APPLICATIONS
DESIGNING MPEG-4 FACIAL ANIMATION TABLES FOR WEB APPLICATIONS STEPHANE GACHERY, NADIA MAGNENAT-THALMANN MIRALab - University of Geneva 22 Rue Général Dufour, CH-1211 GENEVA 4, SWITZERLAND Web: http://www.miralab.unige.ch
FACIAL RIGGING FOR 3D CHARACTER
FACIAL RIGGING FOR 3D CHARACTER Matahari Bhakti Nendya 1, Eko Mulyanto Yuniarno 2 and Samuel Gandang Gunanto 3 1,2 Department of Electrical Engineering, Institut Teknologi Sepuluh Nopember, Surabaya, Indonesia
Very Low Frame-Rate Video Streaming For Face-to-Face Teleconference
Very Low Frame-Rate Video Streaming For Face-to-Face Teleconference Jue Wang, Michael F. Cohen Department of Electrical Engineering, University of Washington Microsoft Research Abstract Providing the best
Technical Training Module ( 30 Days)
Annexure - I Technical Training Module ( 30 Days) Section 1 : Programmable Logic Controller (PLC) 1. Introduction to Programmable Logic Controller - A Brief History, Need and advantages of PLC, PLC configuration,
Approaching VR 2.0: Creating and Sharing Open Source Virtual Environments for Health Care and Research
Approaching VR 2.0: Creating and Sharing Open Source Virtual Environments for Health Care and Research Prof. G. Riva, Ph.D. Istituto Auxologico Italiano http://www.neurovr.org Applied Technology for Neuro-Psychology
Interactive solutions for experiential fashion marketing
Interactive solutions for experiential fashion marketing Stereoscape offers interactive solutions for experiential fashion marketing in events and retail environments. We combine the latest visualisation
BEAT: the Behavior Expression Animation Toolkit
BEAT: the Behavior Expression Animation Toolkit Justine Cassell MIT Media Laboratory 20 Ames St., E15-315 Cambridge, MA 02139 1-617-253-4899 [email protected] Hannes Högni Vilhjálmsson MIT Media Laboratory
Computational Modeling and Simulation for Learning an Automation Concept in Programming Course
Computational Modeling and Simulation for Learning an Automation Concept in Programming Course Yong Cheon Kim, Dai Young Kwon, and Won Gyu Lee Abstract Computational thinking is a fundamental skill for
Sensor Modeling for a Walking Robot Simulation. 1 Introduction
Sensor Modeling for a Walking Robot Simulation L. France, A. Girault, J-D. Gascuel, B. Espiau INRIA, Grenoble, FRANCE imagis, GRAVIR/IMAG, Grenoble, FRANCE Abstract This paper proposes models of short-range
Development and Application of a Distance Learning System by Using Virtual Reality Technology
Systems and Computers in Japan, Vol. 33, No. 4, 2002 Translated from Denshi Joho Tsushin Gakkai Ronbunshi, Vol. J83-D-I, No. 6, June 2000, pp. 619 6 26 Development and Application of a Distance Learning
Creating Scenes and Characters for Virtools in OpenFX
Creating Scenes and Characters for Virtools in OpenFX Scenes Scenes are straightforward: In Virtools use the Resources->Import File As->Scene menu command and select the.mfx (OpenFX model) file containing
HOW CAN I IMPROVE MY PUBLIC SPEAKING SKILLS? (To be used in conjunction with the assessment developed for Standard 4B.J and 5C.J)
(To be used in conjunction with the assessment developed for Standard 4B.J and 5C.J) Performance Standard 4A.J Each student will listen to a three to five minute speech given by another student and will
CS 378: Computer Game Technology
CS 378: Computer Game Technology http://www.cs.utexas.edu/~fussell/courses/cs378/ Spring 2013 University of Texas at Austin CS 378 Game Technology Don Fussell Instructor and TAs! Instructor: Don Fussell!
CATIA V5 Tutorials. Mechanism Design & Animation. Release 18. Nader G. Zamani. University of Windsor. Jonathan M. Weaver. University of Detroit Mercy
CATIA V5 Tutorials Mechanism Design & Animation Release 18 Nader G. Zamani University of Windsor Jonathan M. Weaver University of Detroit Mercy SDC PUBLICATIONS Schroff Development Corporation www.schroff.com
Spatial Pose Trees: Creating and Editing Motions Using a Hierarchy of Low Dimensional Control Spaces
Eurographics/ ACM SIGGRAPH Symposium on Computer Animation (2006), pp. 1 9 M.-P. Cani, J. O Brien (Editors) Spatial Pose Trees: Creating and Editing Motions Using a Hierarchy of Low Dimensional Control
Behavioral Animation Modeling in the Windows Environment
Behavioral Animation Modeling in the Windows Environment MARCELO COHEN 1 CARLA M. D. S. FREITAS 1 FLAVIO R. WAGNER 1 1 UFRGS - Universidade Federal do Rio Grande do Sul CPGCC - Curso de Pós Graduação em
Construct User Guide
Construct User Guide Contents Contents 1 1 Introduction 2 1.1 Construct Features..................................... 2 1.2 Speech Licenses....................................... 3 2 Scenario Management
Effective Use of Android Sensors Based on Visualization of Sensor Information
, pp.299-308 http://dx.doi.org/10.14257/ijmue.2015.10.9.31 Effective Use of Android Sensors Based on Visualization of Sensor Information Young Jae Lee Faculty of Smartmedia, Jeonju University, 303 Cheonjam-ro,
Effective Interface Design Using Face Detection for Augmented Reality Interaction of Smart Phone
Effective Interface Design Using Face Detection for Augmented Reality Interaction of Smart Phone Young Jae Lee Dept. of Multimedia, Jeonju University #45, Backma-Gil, Wansan-Gu,Jeonju, Jeonbul, 560-759,
SimFonIA Animation Tools V1.0. SCA Extension SimFonIA Character Animator
SimFonIA Animation Tools V1.0 SCA Extension SimFonIA Character Animator Bring life to your lectures Move forward with industrial design Combine illustrations with your presentations Convey your ideas to
Character Animation Tutorial
Character Animation Tutorial 1.Overview 2.Modelling 3.Texturing 5.Skeleton and IKs 4.Keys 5.Export the character and its animations 6.Load the character in Virtools 7.Material & texture tuning 8.Merge
