An Active Head Tracking System for Distance Education and Videoconferencing Applications



Similar documents
Speed Performance Improvement of Vehicle Blob Tracking System

International Journal of Advanced Information in Arts, Science & Management Vol.2, No.2, December 2014

Mean-Shift Tracking with Random Sampling

Face Recognition in Low-resolution Images by Using Local Zernike Moments

Real Time Target Tracking with Pan Tilt Zoom Camera

A PHOTOGRAMMETRIC APPRAOCH FOR AUTOMATIC TRAFFIC ASSESSMENT USING CONVENTIONAL CCTV CAMERA

A System for Capturing High Resolution Images

LOCAL SURFACE PATCH BASED TIME ATTENDANCE SYSTEM USING FACE.

Parallelized Architecture of Multiple Classifiers for Face Detection

Automatic Maritime Surveillance with Visual Target Detection

2 Pelco Video Analytics

Whitepaper. Image stabilization improving camera usability

CS231M Project Report - Automated Real-Time Face Tracking and Blending

IR-Cut. Day/Night. Filter

2-Megapixel Sony Progressive CMOS Sensor with Super Wide Dynamic Range and High Frame Rate

VIDEO COMMUNICATION SYSTEM-TECHNICAL DOCUMENTATION. Tracking Camera (PCSA-CTG70/CTG70P) PCS-G70/G70P All

A Real Time Hand Tracking System for Interactive Applications

ivms-4200 Client Software Technical Specification v1.02

A Learning Based Method for Super-Resolution of Low Resolution Images

From Product Management Telephone Nuremberg

Intelligent Monitoring Software. IMZ-RS400 Series IMZ-RS401 IMZ-RS404 IMZ-RS409 IMZ-RS416 IMZ-RS432

Vision-Based Blind Spot Detection Using Optical Flow

Object Tracking System Using Motion Detection

BACnet for Video Surveillance

IP 67. IR-Cut. Day/Night. Filter

idisplay v.2.0 User Guide


ARTICLE. Which has better zoom: 18x or 36x?

HANDS-FREE PC CONTROL CONTROLLING OF MOUSE CURSOR USING EYE MOVEMENT

COMPACT GUIDE. Camera-Integrated Motion Analysis

ACTi Streaming Explorer User s Manual Ver

AXIS 262+ Network Video Recorder

Indoor Surveillance Security Robot with a Self-Propelled Patrolling Vehicle

Multiple Object Tracking Performance Metrics and Evaluation in a Smart Room Environment

AXIS Camera Station Quick Installation Guide

Robust Real-Time Face Detection

Professional Surveillance System User s Manual

Open-Set Face Recognition-based Visitor Interface System

Intelligent Monitoring Software

FACE RECOGNITION BASED ATTENDANCE MARKING SYSTEM

A&H Software House Inc. Web: Luxriot

OpenEXR Image Viewing Software

Distributed Vision Processing in Smart Camera Networks

Product Characteristics Page 2. Management & Administration Page 2. Real-Time Detections & Alerts Page 4. Video Search Page 6

A Robust Multiple Object Tracking for Sport Applications 1) Thomas Mauthner, Horst Bischof

Tracking and Recognition in Sports Videos

Tracking of Small Unmanned Aerial Vehicles

Using Real Time Computer Vision Algorithms in Automatic Attendance Management Systems

SCode CMS (Central Monitoring System) V3.5.1 (SP-501 and MP-9200)

AUTOMATED ATTENDANCE CAPTURE AND TRACKING SYSTEM

Environmental Remote Sensing GEOG 2021

IP Matrix MVC-FIPM. Installation and Operating Manual

Multi-view Intelligent Vehicle Surveillance System

Firmware Release Letter BVIP Firmware Table of Contents. 1. Introduction Disclaimer of Warranty Purpose 2 1.

A Method for Controlling Mouse Movement using a Real- Time Camera

False alarm in outdoor environments

Extracting a Good Quality Frontal Face Images from Low Resolution Video Sequences

Learning Detectors from Large Datasets for Object Retrieval in Video Surveillance

PHOTOGRAMMETRIC TECHNIQUES FOR MEASUREMENTS IN WOODWORKING INDUSTRY

A Real Time Driver s Eye Tracking Design Proposal for Detection of Fatigue Drowsiness

The Visual Internet of Things System Based on Depth Camera

USING COMPUTER VISION IN SECURITY APPLICATIONS

INTELLECT TM Software Package

Real time vehicle detection and tracking on multiple lanes

Face Model Fitting on Low Resolution Images

Designing and Embodiment of Software that Creates Middle Ware for Resource Management in Embedded System

A Study of Automatic License Plate Recognition Algorithms and Techniques

Boundless Security Systems, Inc.

Interactive person re-identification in TV series

Automated Monitoring System for Fall Detection in the Elderly

Development of an automated Red Light Violation Detection System (RLVDS) for Indian vehicles

ARTICLE. 10 reasons to switch to IP-based video

Professional Surveillance System User s Manual

Determining optimal window size for texture feature extraction methods

Video Analytics A New Standard

Tracking performance evaluation on PETS 2015 Challenge datasets

ivms-4200 Client Software Quick Start Guide V1.02

Building an Advanced Invariant Real-Time Human Tracking System

Service Pack Release Letter MIC Series 550 System Controller Table of Contents

OPERATION MANUAL. MV-410RGB Layout Editor. Version 2.1- higher

How does the Kinect work? John MacCormick

Journal of Chemical and Pharmaceutical Research, 2014, 6(5): Research Article

UNIVERSITY OF CENTRAL FLORIDA AT TRECVID Yun Zhai, Zeeshan Rasheed, Mubarak Shah

Video Tracking Software User s Manual. Version 1.0

ImagineWorldClient Client Management Software. User s Manual. (Revision-2)

FB-500A User s Manual

FPGA Implementation of Human Behavior Analysis Using Facial Image

Index Terms: Face Recognition, Face Detection, Monitoring, Attendance System, and System Access Control.

How To Create A Security System Based On Suspicious Behavior Detection

REAL TIME TRAFFIC LIGHT CONTROL USING IMAGE PROCESSING

CMS-DH CENTRAL MANAGEMENT SOFTWARE

802.3af. Build-in Speaker. PoE

Real-Time Tracking of Pedestrians and Vehicles

Transcription:

An Active Head Tracking System for Distance Education and Videoconferencing Applications Sami Huttunen and Janne Heikkilä Machine Vision Group Infotech Oulu and Department of Electrical and Information Engineering P.O. Box 4500, FIN-90014 University of Oulu, Finland {samhut, jth}@ee.oulu.fi Abstract We present a system for automatic head tracking with a single pan-tilt-zoom (PTZ) camera. In distance education the PTZ tracking system developed can be used to follow a teacher actively when s/he moves in the classroom. In other videoconferencing applications the system can be utilized to provide a close-up view of the person all the time. Since the color features used in tracking are selected and updated online, the system can adapt to changes rapidly. The information received from the tracking module is used to actively control the PTZ camera in order to keep the person in the camera view. In addition, the system implemented is able to recover from erroneous situations. Preliminary experiments indicate that the PTZ system can perform well under different lighting conditions and large scale changes. 1 Introduction Distance education and videoconferencing are nowadays widely used in universities and other schools, as well as in enterprises. Using wide-angle cameras, it is easy to keep the speaker in the field of view (FOV) all the time. Unfortunately, this approach results in a low resolution image, where the details can be blurry. To cope with this problem, the camera has to be steered automatically. In the lecture room, the lighting conditions can change rapidly due to, for example, slide changes. This is the main reason that only a few algorithms for active tracking are utilized in practice. In our system, a pan-tilt-zoom (PTZ) camera is designed to track the teacher when s/he walks at the front of the classroom. Tracking is carried out based on the information provided by the pan-tilt-zoom camera itself. In general, automatic PTZ tracking can be divided into two categories. The first option, also applied to this work, is to use mechanical PTZ cameras that are controlled through a certain command interface. The other way is to use virtual camera control based on panoramic video capturing. However, the virtual camera control based systems require that the camera s video resolution is higher than the required output resolution. A recent approach to track the teacher is introduced by Microsoft Research [12]. They present a hybrid speaker tracking scheme based on a single PTZ camera in an automated lecture capturing system. Since the camera s video resolution is usually higher than the required output resolution, they frame the output video as a sub-region of the camera s input video. That allows tracking the speaker both digitally and mechanically. According to their paper, digital tracking has the advantage of being smooth, and mechanical tracking can cover a wide area. The tracking is based on the motion information obtained from the camera image. The system introduced in [9] targets applications such as classroom lectures and videoconferencing. First wide-angle video is captured by stitching video images from multiple stationary cameras. After that, a region of interest (ROI) can be digitally cropped from the panoramic video. The ROI is located from motion and color information in the uncompressed domain and macroblock information in the compressed domain. As a result, the system simulates human controlled video recording. One example of active head tracking is [6] which utilizes both color and shape information. The shape of a head is assumed to be an ellipse and a model color histogram is acquired in advance. Then, in the first frame, the appropriate position and scale of the head is determined based on user input. In the following frames, the initial position is selected at the same position of the ellipse in the previous frame. The Mean Shift procedure is applied to make the ellipse position converge to the target center where the color histogram similarity to the model and previous one is maximized. The previous histogram in this case is a color histogram adaptively extracted from the result of the previous

Initialization Feature Selection Feature Update Head Tracking PTZ Control Figure 1. Overview of the PTZ tracking system. frame. A classroom or a meeting room environment is not the only place to apply active PTZ tracking. Especially visual surveillance solutions include some active tracking components. For example, the work done in [8] addresses the problem of continuous tracking of moving objects with a PTZ camera. Firstly, a set of good trackable features belonging to the selected target is extracted. Secondly, their system adopts a feature clustering method that is able to discriminate between features associated with the background and features associated with different moving objects. In another surveillance related work [4], the task of the two camera system is to continuously provide zoomed-in highresolution images of the face of a person present in the room. Usually the active tracking methods introduced are based on skin color. The work done in [3] employs the Bhattacharyya coefficient as a similarity measure between the color distribution of the face model and face candidates. A dual-mode implementation of the camera controller determines the pan, tilt, and zoom camera to switch between smooth pursuit and saccadic movements, as a function of the target presence in the fovea region. Another system [11] combines color-based face tracking and face detection to be able to track faces under varying pose. The tracking system implemented and described in this paper is mainly built on existing methods and algorithms which are brought together in a novel way to form a complete PTZ tracking solution. As a result of our work, we have a PTZ tracking system that is able to operate robustly under large scale changes and different lighting conditions. The active tracking methods mentioned above lack the ability to cope with scale changes and update color features in parallel with color-based tracking. Generally speaking, the most important thing is that the system developed does not require any manual control and is able to recover from erroneous situations. In our work, the initialization phase of tracking utilizes the human head shape and motion information. After the initialization step, the color features used in tracking can be selected and updated online. Finally, the location of the person s head is found and tracked by applying a modified version of the Continuously Adaptive Mean Shift (CAMSHIFT) [1] algorithm. On the basis of the tracking result, a PTZ camera is steered to keep the target in the camera view. If tracking failure occurs, the system returns to the initialization phase. Fig. 1 gives an overview of the system. This paper is organized as follows. Section 2 contains a review of the initialization and feature selection methods as well as the description of the feature update and tracking algorithms implemented. In Section 3, we describe our PTZ control hardware and methodology. Preliminary experiments are reported in Section 4. Finally, conclusions are given in Section 5. 2 Feature Selection and Tracking 2.1 Initialization In order to start tracking, the initial position of the target should be found. The main problem when using color information is how to get the object color model without a priori knowledge. Especially in the environments with changing lighting conditions it is nearly impossible to use any fixed color model acquired beforehand. To cope with this problem, some other non-color based method is needed for finding the original position. In this work the target is the speaker s head, which can be localized by utilizing shape and motion information. After the initial position of the head is acquired, the color based tracking method can select the features used. The object detector used in this work has been initially proposed by Viola [10] and improved by Lienhart [7]. The actual implementation of the detector is based on Intel Open Computer Vision Library (OpenCV) [5]. In this case, the size of the face samples the classifier has been trained on is 20x20. The implementation finds rectangular regions in the given image that are likely to contain faces and returns those regions as a sequence of rectangles. In order to be able to find the objects of interest at different sizes, the function scans the image several times on different scales, which is more efficient than resizing the image itself. Each time it considers overlapping regions in the image and applies the classifiers to the regions. After it has proceeded and collected the candidate rectangles (regions that passed the classifier cascade), it groups them and returns a sequence of average rectangles for each large enough group. Unfortunately the aforesaid face detector can occasionally give wrong results, which makes the initialization more difficult. Since the speaker is usually moving when giving a presentation, it is reasonable to use the head detector only inside the areas where motion has been detected. In addition, the head has to be detected in several successive frames, before the tracking phase can take place. This kind of approach decreases the possibility of a false alarm substantially.

Get next image Check feature update status Figure 2. Head detection example. Update ready? No Use previous features Yes Adopt new features In the current system, a simple video frame differencing method is utilized for detecting motion. Frame differencing is a computationally inexpensive way to reduce the size of the scanned region, thus ensuring real-time operation of the whole method. Fig. 2 shows an example head detection result from the method described above. Track the object Start new feature update process Figure 3. Overview of feature selection and update. 2.2 Feature Selection and Update When the location of the head has been found in the previous phase, the color-based tracking can take place. To guarantee robustness, the features used in tracking have to be updated on-line. Therefore the solution used in this pantilt-zoom tracking system relies on a modified method originally introduced in [2]. The basic hypothesis is that the features that best discriminate between object and background are also best for tracking the object. Provided with a set of seed features, log likelihood ratios of class conditional sample densities from object and background are computed. These densities are needed to form a new set of candidate features tailored to the local object/background discrimination task. Finally, a two-class variance ratio is used to rank these new features according to how well they separate sample distributions of object and background pixels, and only the top three are chosen. The set of seed candidate features is composed of linear combinations of camera R, G, B pixel values. As in the original method, in our system the integer weights for different color channels are between -2 and 2. By leaving out the coefficient groups which are linear combinations of each other, a pool of 49 features is left. In the original method described in [2], the feature update process is invoked only periodically. Unfortunately the update process is computationally heavy and therefore time consuming, which means that the object tracked can be lost during the update process. Since the feature update process can take a long time, a new approach to the problem is proposed. In our system, the features used are updated continuously as fast as possible without causing unwanted delays to tracking. The feature update process is run in parallel with the object tracking, as Fig. 3 illustrates. When the new features are available, they are utilized immediately. Such an approach guarantees that the tracking method can adapt to sudden changes occurring in the environment. To make tracking more robust, the original samples in the first frame are combined with the pixel samples from the current frame. 2.3 Head Tracking In our system, the size and location of the object changes during tracking due to both object movement and camera zooming. Therefore, the traditional Mean Shift cannot be utilized. The Continuously Adaptive Mean Shift (CAMSHIFT) algorithm instead has been proposed for tracking of head and face in a perceptual user interface [1]. In order to use the CAMSHIFT algorithm to track colored objects in a video scene, a probability distribution image of the desired color in the video frame must be created. Each pixel in the probability image represents a probability that the color of the pixel from an input frame belongs to the object. In the original CAMSHIFT algorithm the probability image is calculated using only a 1-D histogram of the hue component. Therefore the algorithm may fail in cases where hue alone cannot distinguish the target from the background. In the system implemented, the probability image is obtained from the scaled weight images which are generated for every incoming frame using the features selected. In the weight image object pixels contain positive values whereas background pixels contain negative values. Only the top three features selected are used to compute the weight im-

Image Figure 4. Tracked object and one of the corresponding weight images. Set calculation region at search window center but larger in size than the search window Use (X,Y) to set search window center Calculate weight image Find center of mass within the search window ages for the current frame. Fig. 4 illustrates how the weight image on the right side gives a good basis for tracking. The center and size of the head are found via the modified CAMSHIFT algorithm operating on every weight image independently, as shown in Fig. 5. The initial search window size and location for tracking are set using the results reported from the head detector. Since the number of features used is three, we get three weight images for one incoming frame. The results of the tracking algorithm for the different weight images are combined by calculating the average. The current size and location of the tracked object are used to set the size and location of the search window in the next video image. The process is then repeated for continuous tracking. To be able to recover from erroneous situations, the system has to detect tracking failures. In the current approach the target is reported lost when the size of the search window is smaller than a threshold value given in advance. Also the number of object pixels in the weight images has to be clearly above a predetermined threshold in order to continue tracking. 3 Pan-Tilt-Zoom Control 3.1 Hardware The basis of the system is a Sony EVI-D31 PTZ camera, which has been widely used especially in videoconferencing solutions. It offers a wide range of tunable parameters for PTZ control: Pan range: 200 deg in 1760 increments Tilt range: 50 deg in 600 increments Pan speed range: from 0 to 65 deg/sec in 24 discrete steps Tilt speed range: from 0 to 30 deg/sec in 20 discrete steps Zoom range: from f5.4 to f64.8 mm in 12 discrete steps Since there are many variables that can be adjusted, steering the camera can be done smoothly. This is one of the main reasons for selecting this particular camera model. Report X, Y and size Yes Center search window at the center of mass and find area under it Converged? Figure 5. Block diagram of the tracking algorithm. 3.2 Control Strategy and Software Adequate control of the PTZ camera is an essential part of the system implemented. The camera should react to movements of the target depending on how large those movements are. When the person is moving slowly, the camera should be able to follow smoothly. In the opposite situation, the camera control has to be fast to keep the speaker s head in the field of view (FOV). The software components of the PTZ control module are presented in Fig. 6. The user interface of the control system is developed in C++ and is only shortly described here. In practice, it includes buttons for starting and stopping the system, as well as it offers a way to configure the parameters of the system. To provide a communication channel between the hardware components of the system, there is a standard RS- 232C serial port connection between a regular Pentium IV PC and a Sony EVI-D31 camera. The control commands from the PC to the camera are sent using Sony s VISCA protocol. It is worth noting that tracking is not interrupted by the command execution. When the tracked target, the person s head in this case, moves to the boundary of the FOV, the PTZ camera is panned/tilted to keep the speaker inside the FOV. The current position and of the head in the image is received from the tracking method described in the previous section. To ensure smooth operation, the pan and tilt speeds of the camera are adjusted actively when the object is moving. In practice this means that the pan and tilt speeds are increased if the target is getting out of the FOV despite of the current camera movement. When the teacher stops moving, No

PC PTZ Camera User Interface Tracking Algorithm Camera Control Library Hardware Communication RS232 (VISCA) Command Interface Figure 6. PTZ control software. Figure 8. Tracking almost fails when the person stands up fast. the zoom level of the camera is set to keep the size of the head in the image the same all the time. The desired size of the object can be determined beforehand. Since the camera pan and tilt speeds are limited to a particular range and the tracking method can fail, there may be situations when the camera loses the person from the FOV. To solve this problem, a simple lost and found strategy is utilized. When the target is lost, the camera is stopped for five seconds. If the teacher does not appear in the view during this time, the camera zooms out and returns to the initial position waiting until the initialization method has detected the object again. 4 Experimental Results Comparing the performance of our system to other similar systems that have been developed is difficult. Every system has its own requirements for the hardware and the physical environment. It is also impossible to run all the possible systems at the same time to compare their real-time performance. However, to find out the performance of our system, some preliminary real-time tests have been conducted in a room where the lighting conditions are not uniform. The tests are run on a standard PC (Pentium IV at 3.0 GHz, 1 GB RAM). The video image size is CIF (352x288) captured by the aforementioned Sony EVI-D31 camera at 25fps. In the test sequence presented in this paper, a person is sitting in front of a computer screen at the beginning. First the person s head is detected when he is looking at the computer screen. After a while the person stands up and starts walking around the room. The camera is panned and tilted to keep him in the FOV. To visualize the operation of the system during the test, we selected some clips from the output video (Fig. 7(a)). When the person walks away from the camera, the camera zoom is adjusted to keep the size of the head constant. In the last situation the person has returned to the starting point, as shown in Fig. 7(b). As mentioned earlier, the lighting conditions in the test room are not stable. In addition, the PTZ camera used includes automatic gain and white balance control, which means that the color features used for tracking have to be updated constantly. During the test, approximately 10 frames pass on average before a feature update is ready. Since the execution time of the feature update process depends on the current size of the object tracked, the exact time between updates can vary. As we can see from the test sequence, the system can adapt to lighting condition changes. Also the PTZ control is able to follow the person without difficulties. However, it is possible that the serial port connection between the PC and the PTZ camera may cause unwanted delays. This problem appears clearly in the sample shown in Fig. 8 where the person stands up rapidly. In this situation in question, the tilt speed is not adjusted quickly enough due to buffering in the command sending and receiving. 5 Discussion and Conclusions In this paper, we have presented a PTZ tracking system which can be used to follow the speaker actively in different videoconferencing applications. The system detects the lecturer using head shape and motion information. The features applied to tracking are selected and updated online which makes the method used tolerant to lighting condition changes. The PTZ module keeps the person s head in the field of view by controlling the pan and tilt speeds as a response to the object s movements. In the future, the goal is to integrate the work presented in this paper to an automated distance education system which currently utilizes only fixed cameras. References [1] G. R. Bradski. Real time face and object tracking as a component of a perceptual user interface. In Fourth IEEE Workshop on Applications of Computer Vision (WACV 98), pages 214 219, October 1998. [2] R. T. Collins, Y. Liu, and M. Leordeanu. Online selection of discriminative tracking features. IEEE Transactions on Pattern Analysis and Machine Intelligence, 27(10):1631 1643, October 2005.

(a) (b) Figure 7. Tracking results. (a) The person leaves the table and starts moving around the room and (b) the person walks away from the camera and returns to the starting point. [3] D. Comaniciu and V. Ramesh. Robust detection and tracking of human faces with an active camera. In Third IEEE International Workshop on Visual Surveillance, pages 11 18, July 2000. [4] M. Greiffenhagen, D. Comaniciu, H. Niemann, and V. Ramesh. Design, analysis, and engineering of video monitoring systems: an approach and a case study. Proceedings of the IEEE, 89(10):1498 1517, October 2001. [5] Intel Corporation. Open computer vision library homepage. 0nline, 2006. http://www.intel.com/technology/computing/ opencv/index.htm. [6] D. Jeong, Y. K. Yang, D.-G. Kang, and J. B. Ra. Realtime head tracking based on color and shape information. In A. Said and J. G. Apostolopoulos, editors, Proceedings of SPIE: Image and Video Communications and Processing, volume 5685, pages 912 923, March 2005. [7] R. Lienhart and J. Maydt. An extended set of haar-like features for rapid object detection. In International Conference on Image Processing, volume 1, pages 900 903, September 2002. [8] C. Micheloni and G. L. Foresti. Zoom on target while tracking. In IEEE International Conference on Image Processing, volume 3, pages 117 120, September 2005. [9] X. Sun, J. Foote, D. Kimber, and B. S. Manjunath. Region of interest extraction and virtual camera control based on panoramic video capturing. IEEE Transactions on Multimedia, 7(5):981 990, October 2005. [10] P. Viola and M. Jones. Rapid object detection using a boosted cascade of simple features. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), volume 1, pages 511 518, 2001. [11] T. Yang, S. Z. Li, Q. Pan, J. Li, and C. Zhao. Reliable and fast tracking of faces under varying pose. In Seventh International Conference on Automatic Face and Gesture Recognition, pages 421 428, April 2006. [12] C. Zhang, Y. Rui, L. He, and M. Wallick. Hybrid speaker tracking in an automated lecture room. In International Conference on Multimedia and Expo (ICME), pages 81 84, July 2005.