Gestures Body movements which convey meaningful information



Similar documents
Multimodal Interaction in Virtual Reality

Max, our Agent in the Virtual World

System Control. Virtuelle Realität Wintersemester 2007/08. Overview. Part 10: Virtuelle Realität. Prof. Bernhard Jung

SimFonIA Animation Tools V1.0. SCA Extension SimFonIA Character Animator

ANIMATION a system for animation scene and contents creation, retrieval and display

INFORMING A INFORMATION DISCOVERY TOOL FOR USING GESTURE

DATA VISUALIZATION GABRIEL PARODI STUDY MATERIAL: PRINCIPLES OF GEOGRAPHIC INFORMATION SYSTEMS AN INTRODUCTORY TEXTBOOK CHAPTER 7

Real-time Processing and Visualization of Massive Air-Traffic Data in Digital Landscapes

Face Locating and Tracking for Human{Computer Interaction. Carnegie Mellon University. Pittsburgh, PA 15213

Age Birth to Four Months Four to Eight Months

Communicating Agents Architecture with Applications in Multimodal Human Computer Interaction

HAND GESTURE BASEDOPERATINGSYSTEM CONTROL

Infant-Toddler Alignment. Preschool Alignment. HighScope Educational Research Foundation

OCPS Curriculum, Instruction, Assessment Alignment

Professor, D.Sc. (Tech.) Eugene Kovshov MSTU «STANKIN», Moscow, Russia

My Family FREE SAMPLE. This unit focuses on sequencing. These extension

Template-based Eye and Mouth Detection for 3D Video Conferencing

Teaching Methodology for 3D Animation

Multimedia Systems: Database Support

APPLYING VISION TO INTELLIGENT HUMAN-COMPUTER INTERACTION. Guangqi Ye

SYNTHETIC SIGNING FOR THE DEAF: esign

CPIT-285 Computer Graphics

Go to contents 18 3D Visualization of Building Services in Virtual Environment

Mouse Control using a Web Camera based on Colour Detection

Character Animation from 2D Pictures and 3D Motion Data ALEXANDER HORNUNG, ELLEN DEKKERS, and LEIF KOBBELT RWTH-Aachen University

Hand Analysis Tutorial Intel RealSense SDK 2014

How does the Kinect work? John MacCormick

Blender Notes. Introduction to Digital Modelling and Animation in Design Blender Tutorial - week 9 The Game Engine

Movement Phases in Signs and Co-Speech Gestures, and their Transcription by Human Coders

CLOUD HANDS Yun Shou

CITRIS Founding Corporate Members Meeting

Mobile Analytics for Emergency Response and Training

Part 21: Augmented Reality

Early Childhood Foundations to the Indiana Academic Standards

Colorado School of Mines Computer Vision Professor William Hoff

This is how quickly your brain forgets things

Data Storage 3.1. Foundations of Computer Science Cengage Learning

A Short Introduction to Computer Graphics

HamNoSys Hamburg Notation System for Sign Languages

Implementation of the Discrete Hidden Markov Model in Max / MSP Environment

Lesson 4. Temporal Management of Layers

The 3D rendering pipeline (our version for this class)

Understand the Sketcher workbench of CATIA V5.

What is Visualization? Information Visualization An Overview. Information Visualization. Definitions

Tracking devices. Important features. 6 Degrees of freedom. Mechanical devices. Types. Virtual Reality Technology and Programming

Vision-based Recognition of Gestures with Context. Jan Nikolaus Fritsch

Human-Computer Interaction: Input Devices

SOM-based Experience Representation for Dextrous Grasping

IDENTIFICATION OF HIGH-LEVEL OBJECT MANIPULATION OPERATIONS FROM MULTIMODAL INPUT

Perception of Light and Color

The Flat Shape Everything around us is shaped

Human-like Arm Motion Generation for Humanoid Robots Using Motion Capture Database

VIRTUAL TRIAL ROOM USING AUGMENTED REALITY

MIDLAND ISD ADVANCED PLACEMENT CURRICULUM STANDARDS AP ENVIRONMENTAL SCIENCE

Limits and Possibilities of Markerless Human Motion Estimation

B.A. ANIMATION & GRAPHIC DESIGN

Carpal Tunnel Pain. STRETCH YOUR: 1) Wrist Flexors 2) Wrist Extensors 3) Wrist Decompression 4) Neck (see other chapters for more examples)

Graduate Co-op Students Information Manual. Department of Computer Science. Faculty of Science. University of Regina

A Computer Vision System for Monitoring Production of Fast Food

2. Distributed Handwriting Recognition. Abstract. 1. Introduction

Graphic Design. Background: The part of an artwork that appears to be farthest from the viewer, or in the distance of the scene.

Musical Literacy. Clarifying Objectives. Musical Response

Working With Animation: Introduction to Flash

Effective Use of Android Sensors Based on Visualization of Sensor Information

Virtual CRASH 3.0 Staging a Car Crash

Quick Start Guide. Fundamentals of Systems Modeling Series. Systems Modeling with Enterprise Architect version 9.1. Version 1.

Alphabetic Knowledge / Exploring with Letters

ACE: After Effects CS6

No Equipment Agility/Core/Strength Program for Full Body No Equip Trainer: Rick Coe

Implementation of a Wireless Gesture Controlled Robotic Arm

The Creative Curriculum for Preschool: Objectives for Development & Learning

The REPERE Corpus : a multimodal corpus for person recognition

A Real Time Hand Tracking System for Interactive Applications

Visualization methods for patent data

Teaching Strategies GOLD Objectives for Development & Learning: Birth Through Kindergarten

Exploring Geometric Transformations in a Dynamic Environment Cheryll E. Crowe, Ph.D. Eastern Kentucky University

Course Title: Introduction to Video Game Design Board Approval Date: 4/15/13 Credit / Hours: 0.5credit

Automatic detection of motion sequences for motion analysis

Projection Center Calibration for a Co-located Projector Camera System

INSTRUCTOR WORKBOOK Quanser Robotics Package for Education for MATLAB /Simulink Users

SeaCloudDM: Massive Heterogeneous Sensor Data Management in the Internet of Things

STRAND: Number and Operations Algebra Geometry Measurement Data Analysis and Probability STANDARD:

Identity Guide. HHMI Identity Guidelines V 1.2 1

Robotics and Automation Blueprint

ABOUT THE FUNCTIONAL FITNESS TEST: ABOUT THE AUTHORS:

What is the most obvious difference between pipe flow and open channel flow????????????? (in terms of flow conditions and energy situation)

This week. CENG 732 Computer Animation. Challenges in Human Modeling. Basic Arm Model

GAZETRACKERrM: SOFTWARE DESIGNED TO FACILITATE EYE MOVEMENT ANALYSIS

Situated Visualization with Augmented Reality. Augmented Reality

Multi-modal Human-Computer Interaction. Attila Fazekas.

PART 1 Basic Setup. Section 1.1 Direct The Strokes 1.1.1

VISUAL RECOGNITION OF HAND POSTURES FOR INTERACTING WITH VIRTUAL ENVIRONMENTS

In mathematics, there are four attainment targets: using and applying mathematics; number and algebra; shape, space and measures, and handling data.

PIXEL-LEVEL IMAGE FUSION USING BROVEY TRANSFORME AND WAVELET TRANSFORM

Ohio Early Learning and Development Standards Domain: Language and Literacy Development

Design Elements & Principles

PRODUCT SHEET.

VISUAL ARTS VOCABULARY

encoding compression encryption

Adobe Illustrator CS5

Transcription:

Part 14: Interaction in VR: Speech and Gesture Input Virtuelle Realität Wintersemester 2006/07 Prof. Bernhard Jung Gestures Body movements which convey meaningful information Gesture types 1. emblematic gestures (symbols): Meaning defined by convention 2. deictic gestures: Action of pointing to an object or to a region 3. mimetic gestures: Actions imitating interaction with an object 4. iconic gestures: Describe shape or movement of a object 5. beats: Mark rhythm of speech Types 2-5 (often) accompanied by speech (co-verbal) 1

Gestures: Kinetic Structure McNeill, Levy & Pedelty (1990) Consistent Arm Arm Use Use and and Body Body Posture Posture Consistent Head Head Movement Gesture-Unit Gesture-Phrase Preparation Stroke Stroke Retraction Hold Hold (pre-stroke) Hold Hold (post-stroke) Gesture: uni-modal select (deictic) turn (mimetic) drag and drop 2

Gesture Recognition Data Glove Joint Angles 6DOF Tracker Hand Position Hand Model Classifier Classifiers Neural networks Decision trees Grammars atomic form elements are composed to 'gesture words' e.g. based on HamNoSys HamNoSys: Symbols for Body Parts Prillwitz et al. (1989) "Hamburg Notation System" 3

Some HamNoSys Symbols Prillwitz et al. (1989) "Hamburg Notation System" Symbol ASCII notation Description BSifinger basis shape index finger stretched <etc.> EFinA PalmL LocShoulder LocStretched MoveA MoveR... extended finger orientation ahead palm orientation left location shoulder height location stretched move hand ahead move hand right... ( ) [ ] PARALLEL SEQUENCE executed in parallel executed in sequence HamNoSys Parse Tree statischer Anteil konfiguration input dynamischer Anteil aktion handform handstellung lokation BrackSeqL aktion BrackSeqRq grundform koerperebene abstand aktion bewegung BSifinger LocShoulder LocStretched bewegung einfachebewegung einfachebewegung gerade fingeransatzrichtung handflaechenorientierung gerade MoveR EFinA PalmL MoveA 4

Iconic Gesture Recognition Timo Sowa, University of Bielefeld Object identification by way of iconic gestural descriptions Multimodal Interaction (Speech & Gesture) 5

Multimodal (Speech & Gesture) Interaction Put-That-There System (1980) MIT Media Room Spatial Data Management later: speech and static pointing gestures via Polhemus Tracker Timing of gestures and speech The gesture stroke is often marked by an abrupt stop which is correlated with accented words or syllables the stroke does not occur after an accented word but simultaneously or shortly before Nimm dieses Rohr, steck es da dran 0 1 2 3 => hypotheses for establishing correspondence between accented behaviors in speech and gesture channels 6

Timing in human face-to-face communication Multimodal Integration Two "logistic" problems to be solved (Srihari, 1995): Segmentation Problem How can a system be made to cope with open input? How can units be determined to be processed in one system cycle? Correspondence Problem How to determine cross-references between multiple modalities (speech/gesture)? 7

Correspondence Problem How to determine cross-references between gesture and speech? Importance of reconstructing temporal correspondence example: Put <gesture> this chair there! Specification of referent (presupposition: target loc. known) Put this chair there! Put this chair <gesture> there! Specification of target location (presupposition: referent known) VIENA - Virtual Environment VIENA Project, University of Bielefeld, 1996 Setting: Designing a virtual office environment (arrangement, color) Goal: Relief from technical detail by natural, situated interaction Graphics DB Renderer Modeler Viewing Observe changes Augmented Graphics DB Time-stamped scene descriptions (Geometry models, materials, object names, object types) Bookkeeping Adaptor Virtual Camera Plan & Physics Color Space Interpret Mediating agents Communicate changes P a r s e r Verbal input questions 8

Agents in the VIENA System VIENA Project, University of Bielefeld Space agent has "expert" knowledge: how to obtain current object positions how to calculate object transformations (transl, rot) that objects have to stand on something (not in the air) that an object can only be where there is space what "in front" means for a table or a desk, resp. which orientation an object is expected to have etc. Color agent has "expert" knowledge: Bookkeeper agent has knowledge about: how to obtain current object colors (r,g,b) how to identify objects by color ("the blue chair") how to calculate a color transformation (blue, lighter) by changing rgb vectors etc. geometric description of all scene details, even when changed "by hand" material descriptions of all scene objects previous scene descriptions and alterations etc. Multimodal Parsing VIENA Project, University of Bielefeld Examples: Speech and hand gesture move <gesture> forward move <gesture> that to the left put the bowl <gesture> there make <gesture> this chair green turn <gesture> right put <gesture> that <gesture> there put <gesture> this computer on <gesture> that table put <gesture> this computer on the blue desk 9

put this computer on the blue desk...make this chair green 10

VIENA Interface Agency Timed Input Agency motivated by work of Pöppel & Schwender (1993) Rough approach: record sensor data in small time cycles (here: 100 ms) integrate info from multiple channels in large time cycles (here: 2 sec) 11

Multimodal Parsing SGIM Project, University of Bielefeld Temporal ATN (Augmented Transition Network) Input: gesture and speech streams Output: logical form of multimodal input Multimodal Interaction in VR SGIM & Virtuelle Werkstatt Projects, University of Bielefeld 12