Hands Tracking from Frontal View for Vision-Based Gesture Recognition
|
|
|
- Robyn Woods
- 10 years ago
- Views:
Transcription
1 Hands Tracking from Frontal View for Vision-Based Gesture Recognition Jörg Zieren, Nils Unger, and Suat Akyol Chair of Technical Computer Science, Ahornst. 55, Aachen University (RWTH), Aachen, Germany Abstract. We present a system for tracking the hands of a user in a frontal camera view for gesture recognition purposes. The system uses multiple cues, incorporates tracing and prediction algorithms, and applies probabilistic inference to determine the trajectories of the hands reliably even in case of hand-face overlap. A method for assessing tracking quality is also introduced. Tests were performed with image sequences of 152 signs from German Sign Language, which have been segmented manually beforehand to oer a basis for quantitative evaluation. A hit rate of 81.1% was achieved on this material. 1 Introduction Vision-based hand gesture recognition is a popular research topic for humanmachine interaction. A common problem when working with monocular frontal view image sequences is the localization of the user's hands. This problem is usually ignored or simplied by special restrictions. For example in [2, 7, 10] recognition is based on properties that are computed for the whole input image rather than for hand regions only. This is disadvantageous when gestures merely dier in details of hand shape, since these constitute only a fraction of the image. In [5] the number and properties of all moving connected regions (motion blobs) are considered. This approach is intrinsically sensitive to motion originating from other objects and is therefore only applicable with static backgrounds. The system described in [12] performs explicit localization, but was not designed to yield correct positions in the case of hand and face overlap. In sign language applications, however, overlap is actually frequent. In summary we regard explicit hand localization as a prerequisite when: signs only dier in details of hand shape, the system has to be independent from the image background, overlap is likely to occur. Hand localization is primarily a tracking problem. Tracking algorithms which have been shown to work for faces fail because hand motion is fast and discontinuous and often disturbed by overlap. E.g. the mean shift tracker is unable to handle motion exceeding its maximum search extent [4].
2 Currently, the most promising research directions with regard to tracking hands for gesture recognition are probabilistic reasoning and multiple hypothesis testing [11, 9]. Additional stability can be gained from consideration of multiple visual cues [16, 6] and body models [3, 15]. Quantitative statements about the tracking quality are rarely found, although an immediate inuence on recognition results must be expected. We developed a tracking system that combines multiple visual cues, incorporates a mean shift tracker and a Kalman lter and applies probabilistic reasoning for nal inference. For testing purposes we recorded a total of 152 signs from German Sign Language in several variations. The image sequences (23 seconds, 25 fps, bpp) were segmented manually to have a basis for quantitative evaluation. Results were rated by a special tracking quality assessment method. At the above resolution the system runs approximately in real time on an 800MHz PC. 2 Experimental Setup and Basic Assumptions In our setup the framing dimensions are chosen to capture the signer's upper body. At the start and the end of each sign, both hands are outside or at the lower border of the image. The permanent presence of the face is essential, although it may be completely occluded by the hands. We regard the user's face position as given, since powerful face detection [13, 17] and face tracking methods are available [4]. We also presume that a user specic skin color distribution can be estimated from the face and a general skin color model [1] as source. Fig. 1. Left: Example view of experimental setup with manually segmented regions and hand trajectories ( pixel). Right: Corresponding skin probability map The only restriction for the user is to wear non-skin-colored clothes with long sleeves, which is a commonly stated requirement [11,16,6,3]. This allows to extract the skin regions which represent the hands and the face of the user in arbitrary environments as long as there is no other content similar in color and as long as illumination conditions are reasonable. Figure 1 shows an example view. It should be noted that a uniform white background and uniform black
3 clothing were used here for convenience, but are not necessary if the previously stated requirements are met. 3 Tracking Concept The basic idea of our tracker is to extract connected skin colored regions (blobs) from the image and assign the labels F (face), LH (left hand), and RH (right hand) to them. More than one label may be assigned to one blob, since overlapping objects form a single blob. Under the given conditions there can be at most three blobs, depending on whether the hands overlap the face, overlap each other, or are simply not inside the image. Since the face remains in the image, there is at least one blob. A virtual blob is placed below the lower border of the image, where the hands can enter and leave. Assigning a label to this virtual blob is equivalent to assuming a hand to be out of the image. Since a blob set is never unique, there are always multiple possible assignments. Each assignment is a hypothesis claiming a certain conguration of face and hands to be the source of the current blob set (see Fig. 2). There are always multiple hypotheses, the most likely of which is selected for further processing. Skin Color Based Segmentation Hypothesis Generation Hypothesis Evaluation F RH, LH Body Model Scores s 1 Continuity s 2 RH F Completeness s 3 + Total Score LH F,LH Face Overlap s 4 RH Blob Size s 5 Fig. 2. Generation and evaluation of hypotheses for extracted skin color blobs In order to nd the most likely one, all hypotheses are evaluated using the available information, and rated with a total score s IR 0. The total score is composed of several scores s i computed by distinct modules, each representing visual cues or high level knowledge about the tracking task. If a hypothesis is in complete agreement with the information considered in a module, it receives a
4 score of 0 (which is thus the maximum). With increasing deviation or disagreement, the score decreases accordingly. The subsequent sections describe these scoring modules. A module's score may consist of multiple subscores. Weights w i permit a variation of each module's inuence on the total score. Distances (measured between the objects' centers of gravity (COG)) and areas are normalized by the width and the area of the face bounding box, respecitvely, for independence of image resolution. 3.1 Body Model This module consists of two subcomponents that both aim at preventing a confusion of left and right hand. The rst one represents the assumption that the left (right) hand is typically found to the left (right) of the face. If a hypothesis violates this assumption, it receives a negative subscore which is proportional to the x coordinate dierence between the face and the hand, x LH ( x RH ). The second subcomponent contributes a negative subscore if the right hand is positioned left to the left hand, at a distance of x. The module's score s 1 is given by (1). s 1 = w 1a x LH w 1a x RH w 1b x (1) 3.2 Continuity This module rates the continuity of motion, considering initialization and dynamic conditions. Its score s 2 is the sum of two subscores s 2a and s 2b described below. Initialization Conditions. In the initial frame, hands may not overlap with the face (see Sect. 2); hypotheses violating this condition are disqualied by a subscore of. In all subsequent frames, the (dis)appearance of a hand is assumed most likely at the bottom border of the image. Thus, for every hand (dis)appearing at a distance of y from the bottom border (which is equivalent to a jump from (to) the virtual blob), a subscore s 2a = w 2a y is added. Dynamic Conditions. As soon as a hand has appeared in the image, a Kalman lter is initialized, using a 6-dimensional system state X consisting of position, velocity, and acceleration in both x and y direction (see (2)). A motion model of constant acceleration is assumed. The lter state is initialized to the detected hand's position, with velocity and acceleration set to zero. X = (x, ẋ, ẍ, y, ẏ, ÿ) T (2) In subsequent frames, the Kalman lter yields position predictions and receives the actual measurements for state updates (see Sect. 4). Depending on the distance d between the prediction and the position indicated by the hypothesis, a subscore s 2b is computed according to (3). d min can be set to allow for a deviation from the above motion model. If the hypothesis states that the hand has left the image, d is set to the prediction's distance from the lower border. { w2b d if d > d s 2b = min 0 otherwise (3)
5 3.3 Completeness The assumptions described in Sect. 2 ensure that the only skin colored objects in the image are the signer's face and hands. This means that each blob must be assigned at least one object. It is therefore an error if a hypothesis leaves a blob unassigned. In such a case, the output of the Completeness module is a score s 3 = ; otherwise, s 3 = Face Overlap The area of the face blob does not change signicantly in the given setup. A sudden increase can only be caused by an overlap; likewise, a sudden decrease is related to the cessation of an overlap. Therefore, unless the face is already overlapped, an observed face size change should be matched by the hypothesis in that it states the start/end of an overlap accordingly. Hence this module computes a score reecting the degree of match between observation and hypothesis as follows: w 4 w s 4 = 4 hypothesis expects size incr./decr., but none observed size incr./decr. observed, but none expected by hypothesis 2w 4 hypothesis expects size incr./decr., but opposite observed 0 otherwise (4) 3.5 Blob Size Since the size of both hand blobs may change quickly, the above scheme can only be applied to a face-hand-overlap, but not to a hand-hand-overlap. However, the size of a hand blob usually does not change more than 20% from one frame to another, making size a suitable property for telling hands apart. To this end, the Blob Size module computes the relative increase in area a from a hand's size in the previous and current frame (a n 1 and a n, respectively) as a = a n 1 a n 1. (5) This is only possible if the hand is visible and not overlapping in both the previous and the current frame, for in case of overlapping neither object's area can be determined accurately. The module's score s 5 is given by (6). s 5 = 4 Determination of Hand Position { w5 ( a 0.2) if a > otherwise (6) The hypothesis which received the highest score by the process described in Sect. 3 is considered correct and is processed further to yield measurements and estimations for the next frame (see e.g. Sect. 3.2).
6 Normally a hand's position is determined as the corresponding blob's COG. In case of overlap, however, this is inaccurate because the blob's border includes multiple objects. To obtain a more accurate position, the skin probability map is merged with a motion map to yield a combined map [1]. The CAMSHIFT algorithm [4] is then applied to this new map, with a search window of xed size and an initial position according to the Kalman prediction. The mean shift centers the search window on the nearest peak, which represents a skin colored and moving object. This technique is applied in case of hand-face overlap, since the hand is generally moving more than the face. It is less suitable for hand-hand overlaps, which show no reliable correlation between peaks and hand centers. Before processing the measurement, the Kalman lter's measurement noise matrix is set to reect the accuracy of the method used to determine the hand's position (high accuracy for skin color blob based segmentation (i.e. no overlap), low accuracy in case of the CAMSHIFT method (i.e. overlap)). The following gure shows the result of the described tracking scheme for the sign Evening from German Sign Language. Fig. 3. Tracking example. Left: Target trajectories for 10 variations (from manual segmentation). Right: Tracker output for 3 variations 5 Evaluation of Tracking Quality The tracker's output can be qualied as hit or miss using the reference target in the manually segmented data. Sensible denitions of a hit are (see Fig. 4): 1. Estimation lies inside a limited region around the target center. 2. Estimation lies within the border of the target object. Both denitions are valid, since subsequent processing steps might either require seed points which must be inside the object (e.g. region growing), or might be satised with points near the target center, regardless of whether the point is actually part of the object or not (e.g. active shape models). When the tracker misses the object, the displacement is quantied by the Euclidean distance to the target center. In case of a hit, a Mahalanobis distance
7 with regard to the object's main axes and their extent is used instead, resulting in elliptic lines of equal distance. This reects the fact that displacement in direction of the elongated axis is less severe than along the shorter axis. target center hit on object hit near center miss Fig.4. Denition of tracker hit and miss with regard to target object (grey area). The ellipse shows a line of constant distance for quantifying tracker accuracy The main axes are dened as the eigenvectors of the 2D shape and their extent is proportional to the square root of the eigenvalues λ 1 and λ 2. Considering these axes as origin for the coordinate system with the coordinates x and y allows to compute the Mahalanobis distance by ( ) x 2 ( ) y 2 d hit = c λ + 1 c λ c IR. (7) 2 Table 1 shows a selection of results for ve two-handed signs, one of which is free from overlap (Car), as well as the overall performance on the complete set of 152 signs. Here c = 2 was chosen and hits near center were limited to distances below or equal to 1. Only the right (dominant) hand was considered. (Note that on object and near center are independent properties of a hit, so the corresponding columns need not add up to the percentage of total hits.) Table 1. Hit statistics including Mahalanobis distance (for hits) and Euclidian distance (for misses; unit: pixels). Cells contain: Percentage; min/mean/max distance Sign Total Hits Hits on Object Hits Near Center Missed Evening 90.8; 0.0/0.5/ ; 0.0/0.5/ ; 0.0/0.4/ ; 16.2/22.3/30.4 Work 91.1; 0.0/0.2/ ; 0.0/0.2/ ; 0.0/0.2/ ; 18.0/47.9/60.1 Banana 65.7; 0.0/0.4/ ; 0.0/0.3/ ; 0.0/0.3/ ; 17.5/42.7/75.0 Computer 61.4; 0.0/0.7/ ; 0.0/0.7/ ; 0.0/0.4/ ; 13.0/27.4/66.4 Car 100; 0.0/0.0/ ; 0.0/0.0/ ; 0.0/0.0/ ; 0.0/0.0/0.0 (all) 81.1; 0.0/0.2/ ; 0.0/0.2/ ; 0.0/0.1/ ; 7.3/45.9/175.9 An analysis of the complete results identies two main causes for errors: 1. In case of hand-face overlap, the CAMSHIFT algorithm does not converge on the hand due to lack of motion, but rather on the face. 2. When hands overlap each other, the tracker uses for both hands the position given by the corresponding blob's COG, which deviates from each hand's individual center. Results for Eveningand Work are already very satisfactory, while Banana and Computer show a need for improvement. The complete hit statistics clearly identify overlap as the main problem. Sequences without overlap (e.g. Car) are almost always handled without error.
8 6 Outlook We expect to improve the system's performance by exploiting additional image cues, e.g. by template matching. Multiple hypothesis tracking and the handling of distractors are obvious but complex steps towards independence from the image background. This would also suggest a more sophisticated body model. Putting the system to use by extracting features and performing recognition tests will be the subject of further research. References 1. Akyol, S., Alvarado, P.: Finding Relevant Image Content for mobile Sign Language Recognition. In: Hamza, M.H. (ed.): IASTED International Conference- Signal Processing, Pattern Recognition and Applications (SPPRA), Rhodes (2001) Bobick, A.-F., Davis, J.-W.: The Representation and Recognition of Action Using Temporal Templates. IEEE PAMI 3:23 (2001) Bowden, R., Mitchell, T.A., Sahadi, M.: Non-linear statistical models for the 3D reconstruction of human pose and motion from monocular image sequences. Image and Vision Computing Journal 18 (2000) Bradski, G.R.: Computer vision face tracking for use in a perceptual user interface. Intel Technology Journal Q2 (1998) 5. Cutler, R., Turk, M.: View-based Interpretation of Real-time Optical Flow for Gesture Recognition. Proc. IEEE Conf. Face and Gesture Recognition (1998) Imagawa, K., Lu, S., Igi, S.: Color-Based Hands Tracking System for Sign Language Recognition. Proc. IEEE Conf. Face and Gesture Recognition (1998) Nagaya, S., Seki, S., Oka, R.: A Theoretical Consideration of Pattern Space Trajectory for Gesture Spotting Recognition. Proc. IEEE Conf. Face and Gesture Recognition (1996) Oliver, N., Pentland, A.: Lafter: Lips and face real-time tracker. Proc. IEEE Conf. Computer Vision Pattern Recognition (1997) Rasmussen, C., Hager, G.D.: Joint Probabilistic Techniques for Tracking Objects Using Visual Cues. Intl. Conf. Intelligent Robotic Systems (1998) no pagenumbers 10. Rigoll, G., Kosmala, A., Eickeler, S.: High Performance Real-Time Gesture Recognition Using Hidden Markov Models. In: Wachsmut, I., Fröhlich, M. (eds.): Gesture and Sign Language in Human-Computer Interaction. Springer (1998) Sherrah, J., Gong, S.: Tracking Discontinuous Motion Using Bayesian Inference. Proc. European Conf. on Computer Vision (2000) Starner, T., Pentland, A.: Visual Recognition of American Sign language Using Hidden Markov Models. Proc. IEEE Workshop Face and Gesture Recognition (1995) Viola, P., Jones, M.J.: Robust Real-time Object Detection. Technical Report CRL 2001/01, Cambridge Research Laboratory (2001) 14. Welch, G., Bishop, G.: An introduction to the Kalman Filter. Technical Report , Dept. of Computer Science, University of Chapel Hill (2001) 15. Wren, C., Azarbayejani, A., Darrell, T., Pentland, A.: Pnder: Real-time tracking of the human body. IEEE PAMI 7:19 (1997) Yang, M., Ahuja, N.: Extraction and Classication of Visual Motion Patterns for Hand Gesture Recognition. Proc. IEEE Conf. CVPR (1998) Yang, M., Kriegman, D.J., Ahuja, N.: Detecting faces in images: a survey. IEEE PAMI 1:24 (2002) 3458
Tracking People in 2 1 2 D. Ismail Haritaoglu, David Harwood and Larry S. Davis. University of Maryland. College Park, MD 20742, USA
W 4 S: A Real-Time System for Detecting and Tracking People in 2 1 2 D Ismail Haritaoglu, David Harwood and Larry S. Davis Computer Vision Laboratory University of Maryland College Park, MD 20742, USA
Practical Tour of Visual tracking. David Fleet and Allan Jepson January, 2006
Practical Tour of Visual tracking David Fleet and Allan Jepson January, 2006 Designing a Visual Tracker: What is the state? pose and motion (position, velocity, acceleration, ) shape (size, deformation,
Tracking Groups of Pedestrians in Video Sequences
Tracking Groups of Pedestrians in Video Sequences Jorge S. Marques Pedro M. Jorge Arnaldo J. Abrantes J. M. Lemos IST / ISR ISEL / IST ISEL INESC-ID / IST Lisbon, Portugal Lisbon, Portugal Lisbon, Portugal
1 Example of Time Series Analysis by SSA 1
1 Example of Time Series Analysis by SSA 1 Let us illustrate the 'Caterpillar'-SSA technique [1] by the example of time series analysis. Consider the time series FORT (monthly volumes of fortied wine sales
An Active Head Tracking System for Distance Education and Videoconferencing Applications
An Active Head Tracking System for Distance Education and Videoconferencing Applications Sami Huttunen and Janne Heikkilä Machine Vision Group Infotech Oulu and Department of Electrical and Information
Edge tracking for motion segmentation and depth ordering
Edge tracking for motion segmentation and depth ordering P. Smith, T. Drummond and R. Cipolla Department of Engineering University of Cambridge Cambridge CB2 1PZ,UK {pas1001 twd20 cipolla}@eng.cam.ac.uk
Deterministic Sampling-based Switching Kalman Filtering for Vehicle Tracking
Proceedings of the IEEE ITSC 2006 2006 IEEE Intelligent Transportation Systems Conference Toronto, Canada, September 17-20, 2006 WA4.1 Deterministic Sampling-based Switching Kalman Filtering for Vehicle
Tracking Moving Objects In Video Sequences Yiwei Wang, Robert E. Van Dyck, and John F. Doherty Department of Electrical Engineering The Pennsylvania State University University Park, PA16802 Abstract{Object
Mean-Shift Tracking with Random Sampling
1 Mean-Shift Tracking with Random Sampling Alex Po Leung, Shaogang Gong Department of Computer Science Queen Mary, University of London, London, E1 4NS Abstract In this work, boosting the efficiency of
A Real Time Hand Tracking System for Interactive Applications
A Real Time Hand Tracking System for Interactive Applications Siddharth Swarup Rautaray Indian Institute of Information Technology Allahabad ABSTRACT In vision based hand tracking systems color plays an
Vision based Vehicle Tracking using a high angle camera
Vision based Vehicle Tracking using a high angle camera Raúl Ignacio Ramos García Dule Shu [email protected] [email protected] Abstract A vehicle tracking and grouping algorithm is presented in this work
Component Ordering in Independent Component Analysis Based on Data Power
Component Ordering in Independent Component Analysis Based on Data Power Anne Hendrikse Raymond Veldhuis University of Twente University of Twente Fac. EEMCS, Signals and Systems Group Fac. EEMCS, Signals
Journal of Industrial Engineering Research. Adaptive sequence of Key Pose Detection for Human Action Recognition
IWNEST PUBLISHER Journal of Industrial Engineering Research (ISSN: 2077-4559) Journal home page: http://www.iwnest.com/aace/ Adaptive sequence of Key Pose Detection for Human Action Recognition 1 T. Sindhu
Efficient Background Subtraction and Shadow Removal Technique for Multiple Human object Tracking
ISSN: 2321-7782 (Online) Volume 1, Issue 7, December 2013 International Journal of Advance Research in Computer Science and Management Studies Research Paper Available online at: www.ijarcsms.com Efficient
A Reliability Point and Kalman Filter-based Vehicle Tracking Technique
A Reliability Point and Kalman Filter-based Vehicle Tracing Technique Soo Siang Teoh and Thomas Bräunl Abstract This paper introduces a technique for tracing the movement of vehicles in consecutive video
VISUAL RECOGNITION OF HAND POSTURES FOR INTERACTING WITH VIRTUAL ENVIRONMENTS
VISUAL RECOGNITION OF HAND POSTURES FOR INTERACTING WITH VIRTUAL ENVIRONMENTS Radu Daniel VATAVU Ştefan-Gheorghe PENTIUC "Stefan cel Mare" University of Suceava str.universitatii nr.13, RO-720229 Suceava
Template-based Eye and Mouth Detection for 3D Video Conferencing
Template-based Eye and Mouth Detection for 3D Video Conferencing Jürgen Rurainsky and Peter Eisert Fraunhofer Institute for Telecommunications - Heinrich-Hertz-Institute, Image Processing Department, Einsteinufer
Tracking and Recognition in Sports Videos
Tracking and Recognition in Sports Videos Mustafa Teke a, Masoud Sattari b a Graduate School of Informatics, Middle East Technical University, Ankara, Turkey [email protected] b Department of Computer
How To Analyze Ball Blur On A Ball Image
Single Image 3D Reconstruction of Ball Motion and Spin From Motion Blur An Experiment in Motion from Blur Giacomo Boracchi, Vincenzo Caglioti, Alessandro Giusti Objective From a single image, reconstruct:
Tracking And Object Classification For Automated Surveillance
Tracking And Object Classification For Automated Surveillance Omar Javed and Mubarak Shah Computer Vision ab, University of Central Florida, 4000 Central Florida Blvd, Orlando, Florida 32816, USA {ojaved,shah}@cs.ucf.edu
EM Clustering Approach for Multi-Dimensional Analysis of Big Data Set
EM Clustering Approach for Multi-Dimensional Analysis of Big Data Set Amhmed A. Bhih School of Electrical and Electronic Engineering Princy Johnson School of Electrical and Electronic Engineering Martin
An Energy-Based Vehicle Tracking System using Principal Component Analysis and Unsupervised ART Network
Proceedings of the 8th WSEAS Int. Conf. on ARTIFICIAL INTELLIGENCE, KNOWLEDGE ENGINEERING & DATA BASES (AIKED '9) ISSN: 179-519 435 ISBN: 978-96-474-51-2 An Energy-Based Vehicle Tracking System using Principal
VEHICLE LOCALISATION AND CLASSIFICATION IN URBAN CCTV STREAMS
VEHICLE LOCALISATION AND CLASSIFICATION IN URBAN CCTV STREAMS Norbert Buch 1, Mark Cracknell 2, James Orwell 1 and Sergio A. Velastin 1 1. Kingston University, Penrhyn Road, Kingston upon Thames, KT1 2EE,
Speed Performance Improvement of Vehicle Blob Tracking System
Speed Performance Improvement of Vehicle Blob Tracking System Sung Chun Lee and Ram Nevatia University of Southern California, Los Angeles, CA 90089, USA [email protected], [email protected] Abstract. A speed
Dimensionality Reduction: Principal Components Analysis
Dimensionality Reduction: Principal Components Analysis In data mining one often encounters situations where there are a large number of variables in the database. In such situations it is very likely
False alarm in outdoor environments
Accepted 1.0 Savantic letter 1(6) False alarm in outdoor environments Accepted 1.0 Savantic letter 2(6) Table of contents Revision history 3 References 3 1 Introduction 4 2 Pre-processing 4 3 Detection,
Subspace Analysis and Optimization for AAM Based Face Alignment
Subspace Analysis and Optimization for AAM Based Face Alignment Ming Zhao Chun Chen College of Computer Science Zhejiang University Hangzhou, 310027, P.R.China [email protected] Stan Z. Li Microsoft
Tracking performance evaluation on PETS 2015 Challenge datasets
Tracking performance evaluation on PETS 2015 Challenge datasets Tahir Nawaz, Jonathan Boyle, Longzhen Li and James Ferryman Computational Vision Group, School of Systems Engineering University of Reading,
Building an Advanced Invariant Real-Time Human Tracking System
UDC 004.41 Building an Advanced Invariant Real-Time Human Tracking System Fayez Idris 1, Mazen Abu_Zaher 2, Rashad J. Rasras 3, and Ibrahiem M. M. El Emary 4 1 School of Informatics and Computing, German-Jordanian
Detection and Restoration of Vertical Non-linear Scratches in Digitized Film Sequences
Detection and Restoration of Vertical Non-linear Scratches in Digitized Film Sequences Byoung-moon You 1, Kyung-tack Jung 2, Sang-kook Kim 2, and Doo-sung Hwang 3 1 L&Y Vision Technologies, Inc., Daejeon,
Real-Time Tracking of Pedestrians and Vehicles
Real-Time Tracking of Pedestrians and Vehicles N.T. Siebel and S.J. Maybank. Computational Vision Group Department of Computer Science The University of Reading Reading RG6 6AY, England Abstract We present
Analecta Vol. 8, No. 2 ISSN 2064-7964
EXPERIMENTAL APPLICATIONS OF ARTIFICIAL NEURAL NETWORKS IN ENGINEERING PROCESSING SYSTEM S. Dadvandipour Institute of Information Engineering, University of Miskolc, Egyetemváros, 3515, Miskolc, Hungary,
EFFICIENT VEHICLE TRACKING AND CLASSIFICATION FOR AN AUTOMATED TRAFFIC SURVEILLANCE SYSTEM
EFFICIENT VEHICLE TRACKING AND CLASSIFICATION FOR AN AUTOMATED TRAFFIC SURVEILLANCE SYSTEM Amol Ambardekar, Mircea Nicolescu, and George Bebis Department of Computer Science and Engineering University
A Movement Tracking Management Model with Kalman Filtering Global Optimization Techniques and Mahalanobis Distance
Loutraki, 21 26 October 2005 A Movement Tracking Management Model with ing Global Optimization Techniques and Raquel Ramos Pinho, João Manuel R. S. Tavares, Miguel Velhote Correia Laboratório de Óptica
Accurate and robust image superresolution by neural processing of local image representations
Accurate and robust image superresolution by neural processing of local image representations Carlos Miravet 1,2 and Francisco B. Rodríguez 1 1 Grupo de Neurocomputación Biológica (GNB), Escuela Politécnica
A video based real time fatigue detection system
DEPARTMENT OF APPLIED PHYSICS AND ELECTRONICS UMEÅ UNIVERISTY, SWEDEN DIGITAL MEDIA LAB A video based real time fatigue detection system Zhengrong Yao 1 Dept. Applied Physics and Electronics Umeå University
Vision-Based Blind Spot Detection Using Optical Flow
Vision-Based Blind Spot Detection Using Optical Flow M.A. Sotelo 1, J. Barriga 1, D. Fernández 1, I. Parra 1, J.E. Naranjo 2, M. Marrón 1, S. Alvarez 1, and M. Gavilán 1 1 Department of Electronics, University
Tracking of Small Unmanned Aerial Vehicles
Tracking of Small Unmanned Aerial Vehicles Steven Krukowski Adrien Perkins Aeronautics and Astronautics Stanford University Stanford, CA 94305 Email: [email protected] Aeronautics and Astronautics Stanford
Object tracking & Motion detection in video sequences
Introduction Object tracking & Motion detection in video sequences Recomended link: http://cmp.felk.cvut.cz/~hlavac/teachpresen/17compvision3d/41imagemotion.pdf 1 2 DYNAMIC SCENE ANALYSIS The input to
Static Environment Recognition Using Omni-camera from a Moving Vehicle
Static Environment Recognition Using Omni-camera from a Moving Vehicle Teruko Yata, Chuck Thorpe Frank Dellaert The Robotics Institute Carnegie Mellon University Pittsburgh, PA 15213 USA College of Computing
Low-resolution Character Recognition by Video-based Super-resolution
2009 10th International Conference on Document Analysis and Recognition Low-resolution Character Recognition by Video-based Super-resolution Ataru Ohkura 1, Daisuke Deguchi 1, Tomokazu Takahashi 2, Ichiro
Real-time pedestrian detection in FIR and grayscale images
Real-time pedestrian detection in FIR and grayscale images Dissertation zur Erlangung des Grades eines Doktor-Ingenieurs(Dr.-Ing.) an der Fakultät für Elektrotechnik und Informationstechnik der Ruhr-Universität
Real Time Target Tracking with Pan Tilt Zoom Camera
2009 Digital Image Computing: Techniques and Applications Real Time Target Tracking with Pan Tilt Zoom Camera Pankaj Kumar, Anthony Dick School of Computer Science The University of Adelaide Adelaide,
On-line Tracking Groups of Pedestrians with Bayesian Networks
On-line Tracking Groups of Pedestrians with Bayesian Networks Pedro M. Jorge ISEL / ISR [email protected] Jorge S. Marques IST / ISR [email protected] Arnaldo J. Abrantes ISEL [email protected] Abstract A
Wii Remote Calibration Using the Sensor Bar
Wii Remote Calibration Using the Sensor Bar Alparslan Yildiz Abdullah Akay Yusuf Sinan Akgul GIT Vision Lab - http://vision.gyte.edu.tr Gebze Institute of Technology Kocaeli, Turkey {yildiz, akay, akgul}@bilmuh.gyte.edu.tr
Vision-Based Human Tracking and Activity Recognition
Vision-Based Human Tracking and Activity Recognition Robert Bodor Bennett Jackson Nikolaos Papanikolopoulos AIRVL, Dept. of Computer Science and Engineering, University of Minnesota. Abstract-- The protection
Determining optimal window size for texture feature extraction methods
IX Spanish Symposium on Pattern Recognition and Image Analysis, Castellon, Spain, May 2001, vol.2, 237-242, ISBN: 84-8021-351-5. Determining optimal window size for texture feature extraction methods Domènec
Learning Visual Behavior for Gesture Analysis
Learning Visual Behavior for Gesture Analysis Andrew D. Wilson Aaron F. Bobick Perceptual Computing Group MIT Media Laboratory Cambridge, MA 239 drew, [email protected] Abstract A state-based method
Video Surveillance System for Security Applications
Video Surveillance System for Security Applications Vidya A.S. Department of CSE National Institute of Technology Calicut, Kerala, India V. K. Govindan Department of CSE National Institute of Technology
Tracking in flussi video 3D. Ing. Samuele Salti
Seminari XXIII ciclo Tracking in flussi video 3D Ing. Tutors: Prof. Tullio Salmon Cinotti Prof. Luigi Di Stefano The Tracking problem Detection Object model, Track initiation, Track termination, Tracking
A Cognitive Approach to Vision for a Mobile Robot
A Cognitive Approach to Vision for a Mobile Robot D. Paul Benjamin Christopher Funk Pace University, 1 Pace Plaza, New York, New York 10038, 212-346-1012 [email protected] Damian Lyons Fordham University,
NCSS Statistical Software Principal Components Regression. In ordinary least squares, the regression coefficients are estimated using the formula ( )
Chapter 340 Principal Components Regression Introduction is a technique for analyzing multiple regression data that suffer from multicollinearity. When multicollinearity occurs, least squares estimates
Colorado School of Mines Computer Vision Professor William Hoff
Professor William Hoff Dept of Electrical Engineering &Computer Science http://inside.mines.edu/~whoff/ 1 Introduction to 2 What is? A process that produces from images of the external world a description
VEHICLE TRACKING USING ACOUSTIC AND VIDEO SENSORS
VEHICLE TRACKING USING ACOUSTIC AND VIDEO SENSORS Aswin C Sankaranayanan, Qinfen Zheng, Rama Chellappa University of Maryland College Park, MD - 277 {aswch, qinfen, rama}@cfar.umd.edu Volkan Cevher, James
LOCAL SURFACE PATCH BASED TIME ATTENDANCE SYSTEM USING FACE. [email protected]
LOCAL SURFACE PATCH BASED TIME ATTENDANCE SYSTEM USING FACE 1 S.Manikandan, 2 S.Abirami, 2 R.Indumathi, 2 R.Nandhini, 2 T.Nanthini 1 Assistant Professor, VSA group of institution, Salem. 2 BE(ECE), VSA
OBJECT TRACKING USING LOG-POLAR TRANSFORMATION
OBJECT TRACKING USING LOG-POLAR TRANSFORMATION A Thesis Submitted to the Gradual Faculty of the Louisiana State University and Agricultural and Mechanical College in partial fulfillment of the requirements
Face Model Fitting on Low Resolution Images
Face Model Fitting on Low Resolution Images Xiaoming Liu Peter H. Tu Frederick W. Wheeler Visualization and Computer Vision Lab General Electric Global Research Center Niskayuna, NY, 1239, USA {liux,tu,wheeler}@research.ge.com
Face Recognition For Remote Database Backup System
Face Recognition For Remote Database Backup System Aniza Mohamed Din, Faudziah Ahmad, Mohamad Farhan Mohamad Mohsin, Ku Ruhana Ku-Mahamud, Mustafa Mufawak Theab 2 Graduate Department of Computer Science,UUM
The Scientific Data Mining Process
Chapter 4 The Scientific Data Mining Process When I use a word, Humpty Dumpty said, in rather a scornful tone, it means just what I choose it to mean neither more nor less. Lewis Carroll [87, p. 214] In
Monitoring Head/Eye Motion for Driver Alertness with One Camera
Monitoring Head/Eye Motion for Driver Alertness with One Camera Paul Smith, Mubarak Shah, and N. da Vitoria Lobo Computer Science, University of Central Florida, Orlando, FL 32816 rps43158,shah,niels @cs.ucf.edu
Limitations of Human Vision. What is computer vision? What is computer vision (cont d)?
What is computer vision? Limitations of Human Vision Slide 1 Computer vision (image understanding) is a discipline that studies how to reconstruct, interpret and understand a 3D scene from its 2D images
COLOR-BASED PRINTED CIRCUIT BOARD SOLDER SEGMENTATION
COLOR-BASED PRINTED CIRCUIT BOARD SOLDER SEGMENTATION Tz-Sheng Peng ( 彭 志 昇 ), Chiou-Shann Fuh ( 傅 楸 善 ) Dept. of Computer Science and Information Engineering, National Taiwan University E-mail: [email protected]
Robust Real-Time Face Detection
Robust Real-Time Face Detection International Journal of Computer Vision 57(2), 137 154, 2004 Paul Viola, Michael Jones 授 課 教 授 : 林 信 志 博 士 報 告 者 : 林 宸 宇 報 告 日 期 :96.12.18 Outline Introduction The Boost
International Journal of Advanced Information in Arts, Science & Management Vol.2, No.2, December 2014
Efficient Attendance Management System Using Face Detection and Recognition Arun.A.V, Bhatath.S, Chethan.N, Manmohan.C.M, Hamsaveni M Department of Computer Science and Engineering, Vidya Vardhaka College
Part-Based Recognition
Part-Based Recognition Benedict Brown CS597D, Fall 2003 Princeton University CS 597D, Part-Based Recognition p. 1/32 Introduction Many objects are made up of parts It s presumably easier to identify simple
Introduction to Logistic Regression
OpenStax-CNX module: m42090 1 Introduction to Logistic Regression Dan Calderon This work is produced by OpenStax-CNX and licensed under the Creative Commons Attribution License 3.0 Abstract Gives introduction
Interactive person re-identification in TV series
Interactive person re-identification in TV series Mika Fischer Hazım Kemal Ekenel Rainer Stiefelhagen CV:HCI lab, Karlsruhe Institute of Technology Adenauerring 2, 76131 Karlsruhe, Germany E-mail: {mika.fischer,ekenel,rainer.stiefelhagen}@kit.edu
Environmental Remote Sensing GEOG 2021
Environmental Remote Sensing GEOG 2021 Lecture 4 Image classification 2 Purpose categorising data data abstraction / simplification data interpretation mapping for land cover mapping use land cover class
Face detection is a process of localizing and extracting the face region from the
Chapter 4 FACE NORMALIZATION 4.1 INTRODUCTION Face detection is a process of localizing and extracting the face region from the background. The detected face varies in rotation, brightness, size, etc.
A Learning Based Method for Super-Resolution of Low Resolution Images
A Learning Based Method for Super-Resolution of Low Resolution Images Emre Ugur June 1, 2004 [email protected] Abstract The main objective of this project is the study of a learning based method
A NATURAL HAND GESTURE HUMAN COMPUTER INTERFACE USING CONTOUR SIGNATURES
A NATURAL HAND GESTURE HUMAN COMPUTER INTERFACE USING CONTOUR SIGNATURES Paulo Peixoto ISR - Institute of Systems and Robotics Department of Electrical and Computer Engineering University of Coimbra Polo
Local features and matching. Image classification & object localization
Overview Instance level search Local features and matching Efficient visual recognition Image classification & object localization Category recognition Image classification: assigning a class label to
3D Vehicle Extraction and Tracking from Multiple Viewpoints for Traffic Monitoring by using Probability Fusion Map
Electronic Letters on Computer Vision and Image Analysis 7(2):110-119, 2008 3D Vehicle Extraction and Tracking from Multiple Viewpoints for Traffic Monitoring by using Probability Fusion Map Zhencheng
Model-Based 3D Human Motion Capture Using Global-Local Particle Swarm Optimizations
Model-Based 3D Human Motion Capture Using Global-Local Particle Swarm Optimizations Tomasz Krzeszowski, Bogdan Kwolek, and Konrad Wojciechowski Abstract. We present an approach for tracking the articulated
Low-resolution Image Processing based on FPGA
Abstract Research Journal of Recent Sciences ISSN 2277-2502. Low-resolution Image Processing based on FPGA Mahshid Aghania Kiau, Islamic Azad university of Karaj, IRAN Available online at: www.isca.in,
Distributed Vision Processing in Smart Camera Networks
Distributed Vision Processing in Smart Camera Networks CVPR-07 Hamid Aghajan, Stanford University, USA François Berry, Univ. Blaise Pascal, France Horst Bischof, TU Graz, Austria Richard Kleihorst, NXP
Reconstructing 3D Pose and Motion from a Single Camera View
Reconstructing 3D Pose and Motion from a Single Camera View R Bowden, T A Mitchell and M Sarhadi Brunel University, Uxbridge Middlesex UB8 3PH [email protected] Abstract This paper presents a
Television Control by Hand Gestures
MITSUBISHI ELECTRIC RESEARCH LABORATORIES http://www.merl.com Television Control by Hand Gestures William T. Freeman, Craig D. Weissman TR94-24 December 1994 Abstract We study how a viewer can control
Two-Frame Motion Estimation Based on Polynomial Expansion
Two-Frame Motion Estimation Based on Polynomial Expansion Gunnar Farnebäck Computer Vision Laboratory, Linköping University, SE-581 83 Linköping, Sweden [email protected] http://www.isy.liu.se/cvl/ Abstract.
