Autonomous Monitoring of Cliff Nesting Seabirds using Computer Vision
|
|
|
- Hope Wiggins
- 10 years ago
- Views:
Transcription
1 Autonomous Monitoring of Cliff Nesting Seabirds using Computer Vision Patrick Dickinson 1, Robin Freeman 2, Sam Patrick 3 and Shaun Lawson 1 1 Dept. of Computing and Informatics University of Lincoln Lincoln UK [pdickinson, slawson]@lincoln.ac.uk 3 Dept. of Animal and Plant Science University of Sheffield Sheffield UK 2 Computational Ecology and Environmental Science Group Microsoft Research Cambridge UK Abstract In this paper we describe a proposed system for automatic visual monitoring of seabird populations. Image sequences of cliff face nesting sites are captured using time-lapse digital photography. We are developing image processing software which is designed to automatically interpret these images, determine the number of birds present, and monitor activity. We focus primarily on the the development of low-level image processing techniques to support this goal. We first describe our existing work in video processing, and show how it is suitable for this problem domain. Image samples from a particular nest site are presented, and used to describe the associated challenges. We conclude by showing how we intend to develop our work to construct a distributed system capable of simultaneously monitoring a number of sites in the same locality. ThereisrecentevidencethattheUKhasseenasustaineddeclineinthenumbersofitsbreeding seabirds. Indeed, the summer of 2008 has seen widespread speculation that the breeding cycle of large numbers of cliff nesting seabirds such as Guillemots, Razorbills and Kittiwakes has failed, with many pairs of birds abandoning nests for long periods of time. In the wider context, any variation in seabird populations is seen as an important biodiversity indicator. Accurate monitoring of seabird populations and aggregate behaviours of birds is therefore of crucial importance to gaining an understanding of the changing environmental picture. The monitoring of seabird populations using manual means (e.g. counting birds on cliff-faces) can be very labour intensive and often impractical: seabirds generally breed in harsh, remote environments which are exposed to the elements, resulting in inconsistent and infrequent data sampling as well as human error. In this paper we consider the use of computer vision as a method of automatically monitoring seabird populations. Our aim is to develop a system in which the output from a camera can be interpreted by a software system capable of automatically estimating the number of breeding birds on a cliff face, and determining whether an individual breeding bird is present at the nest site at the time of the image capture. This latter inference can be made since individual Guillemots, as
2 is typical of cliff nesting species, do not stray more than a short distance from their own nest site when present on the cliff face. The use of computer vision for automated surveillance has attracted significant attention from researchers in recent years, motivated largely by concerns over public safety, and the widespread installation of CCTV cameras in city centres and transportation infrastructure. Much of this work has been concerned with developing low-level image processing techniques which can identify and track moving objects. However, most of these techniques have been developed for use in urban and man-made environments and do not transpose well to video data collected in unconstrained natural environments. The focus of our work, then, is on the development of processing techniques which can reliably identify and tracking moving birds in natural environments. We start by reviewing some existing techniques, and show how they can be confounded by image artefacts which occur in natural scenes. We then describe some of our own previous work, in which we have sought to address some of these issues. We conclude by reviewing some recently acquired video data from a Guillemot nesting site in Skomer Island, off the coast of Wales. We consider what specific challenges are presented by data collected from this site, and how our work may be further developed to address them. 1 Vision based Surveillance Data captured from a static video camera comprises a sequence of digitised images, captured at a specific frame-rate: each image consists of a 2D array of intensity or colour values (pixels) sampled from the projection plane of the camera. The ultimate objective of an automated surveillance system is to take this raw image sequence and infer some kind of semantic information about an observed scene. This procedure normally involves several levels or layers of processing, such that each level produces a more abstract and informative representation of the scene. Almost all surveillance systems work in this way, and whilst there is variation in individual processing architectures, the majority follow a common pattern. 1.1 Foreground Segmentation The first processing stage is often referred to as background subtraction, or foreground segmentation, and involves identifying which pixels in an image correspond to moving objects of interest. In this respect, the term foreground is used to refer to moving objects (the part of the scene which is changing from one image to the next), and background refers to the invariant part of the scene. Naturally, it is the scene foreground which is usually of interest: once identified, foreground pixels are usually clustered to form object hypotheses, which may then be tracked from one image to the next. High level processes, such as behavioural modelling, may also be applied. The correct classification of foreground pixels is critical to system performance, since any errors will adversely impact on higher level processing. Whilst many methods have been proposed, robust foreground segmentation remains an elusive goal, and has attracted much attention from computer vision researchers. Most existing approaches to foreground segmentation involve independent modelling and classification of colour values recorded for each image pixel. A background model for a pixel is developed during an initial training phase. During this phase it is normally assumed that the majority of observed values for a pixel are generated by the scene background. Since pixels are classified independently, these are often referred to as per-pixel models.
3 In the simplest cases, exemplified by Hu et al. [4], a mean background colour value is learnt for each pixel from the training frames. A subsequent observed pixel value is classified as foreground if it is more than some threshold distance T from the mean. More sophisticated statistical pixel models have been proposed. For example, Park and Aggarwal s scheme [7] uses a Gaussian distribution in HSV colour space to model the background process for each pixel. Elgammal et al. [3] used a non-parametric kernel density estimate to describe each pixel background. 1.2 Stauffer and Grimson s Model The most widely adopted per-pixel model is that proposed by Stauffer and Grimson [8]. This parametric representation of a multi-modal pixel process is more compact, using an adaptive Mixture of Gaussian (MoG) to model observations of each pixel s process in RGB colour space. Under this MoG model the probability of observing a new colour value x i,t at pixel i and time t is given by: p(x i,t Θ i,t ) = K k=1 ωk i,t N ( ) µ k i,t,σ k i,t where N is the multivariate Gaussian probability distribution function, ωi,t k is the component weight, µ k i,t the mean, and Σk i,t the covariance matrix of the 3-dimensional colour component k at time t. The standard conditions for MoG models apply, such that the component weights for each pixel sum to one. The number of components is fixed, and typically between 3 and 5 are used. Components with the highest relative weightings are taken to represent the background. New observed pixel values are compared against the existing components, to determine if the new value is more likely to have been generated by the background or by a foreground object. The model is adaptive, such that gradual changes in lighting can be learnt. However, a systemic weakness of per-pixel models, including Stauffer and Grimson s, is that independent modelling and classification represents a significant simplification of the actual scene structure. In real-world environments this can lead to frequent misclassifications caused by a variety of artefacts, such as: 1 Colour similarity of overlapping foreground and background objects 2 Sudden lighting changes 3 Object shadows 4 Small movements of background objects 5 Camera movements: the camera is assumed static, but may be moved by wind or other processes In outdoor urban environments, some of these problems are more prevalent than others. For example, the issues of shadow detection and lighting changes has been well studied [10, 9]. 1.3 A Spatially Coherent Model Some authors, such as Migdal and Grimson [6], have attempted to reduce pixel misclassifications using Markov Random Field (MRF) based background models. This type of model places dependencies between adjacent pixels, such that a pixel is more likely to be classified as background if (1)
4 its neighbouring pixels are also background. Conversely, a foreground pixel is more likely to be found among other foreground pixels. MRFs are probabilistic graphical models in which each node represents a random variable and undirected edges between nodes represent dependencies. In the case of foreground segmentation schemes, nodes represent pixel labellings l i {foreground,background} and each node is connected to its immediate 4 or 8 pixel neighbourhood. The edges express the Markovian nature of the local node dependency: given its neighbours, each node is conditionally independent of the rest of the field. From a Bayesian perspective, given an observed image X = {x 1,...,x N }, and suitable observation likelihood functions, the aim is to estimate the image labelling l = {l 1,...,l N } which maximises the posterior probability: p(l X) p(x l)p(l) (2) in which p(l) represents a prior probability for a given labelling configuration, and is defined by the dependencies between adjacent pixels. p(x l) is the likelihood of the labelling determined independently for each pixel: Migdal and Grimson [6] simply use Stauffer and Grimson s per-pixel model to estimate the background likelihood. Maximum a posteriori (MAP) MRF labelling techniques seek to estimate a globally optimum image labelling by maximising the posterior of Equation (2). This may be formulated as an energy minimisation problem, and solved using Gibbs Sampling techniques. The advantage of the random field approach is that spatial dependencies can be easily expressed at pixel level and the MAP estimation results in a segmentation which is globally optimal. Whilst MRF based approaches are demonstrably superior to per-pixel models [2], their robustness is still challenged when large clusters of background pixels appear which have low likelihood under the per-pixel background model. This can arise where background objects move significantly, and may cause additional localised misclassifications. 1.4 Challenges in Natural Environments As we have mentioned previously, most research in vision based surveillance has been focussed on monitoring people and vehicles in urban environments. Consequently, work has been directed at optimising performance under these conditions. However, relatively little work has sought to apply such systems to the monitoring of wildlife in natural environments, where a quite different range of problems are presented. Application of per-pixel and MRF based techniques to video data captured from typical bird habitats reveals that pixel misclassifications are often caused by natural processes which act to modify the spatial structure of the scene background. For example: 1 Movement of trees and plants caused by wind 2 Movement of water 3 Camera movements caused by wind In each case, frequent false positive foreground classifications tend to obscure the true target. Currently, little work exists which addresses these problems, though these has been some recent interest in the elimination errors caused by moving foliage [1]. There has also been some recent interest in the use of automated surveillance for the monitoring of birds [5, 11], but the above issues have not yet been robustly addressed.
5 2 Our Approach to Segmentation We have recently proposed a method of foreground segmentation which is effective in reducing misclassifications caused by moving trees and water [2]. Rather use a pixel-based background model, we represent an observed scene as an adaptive mixture of Gaussians (MoG) in 5-dimensional feature space. The features used are the 2D spatial coordinates and 3 colour coordinates in Y UV format. The MoG is constructed with the intention that each component represents a homogeneous region, corresponding to a set of image pixels with similar spatial and colour characteristics. A pixel observation at time t is represented by a feature vector x t = [x,y,y,u,v] T. The probability distribution function for a model component j is given by: p(x t j) = e 1 2 (xt µ j,t )T (Σ j,t ) 1 (x t µ j,t ) (2π) D Σ j,t (3) where the parameters µ j,t and Σ j,t are the mean and covariance matrix of the j th component at time t, and the dimensionality D is 5. For K components the general mixture model conditions hold, such that the component weights sum to unity. The number of components, K, required to model the whole scene is estimated automatically from frame to frame, but normally we expect several hundred. Our premise is that the components of our model represent different background processes, such as different objects, or parts of objects in the scene. Each such process generates a subset of the pixels in an observed image. Thus, given a new observed image and a set of model parameters, we may determine which component is most likely to have generated a particular pixel x t. We thus seek to assign an observed pixel value to the component with the highest log likelihoods: C map = argmax j {log(p(x t j))+log(ω j,t )} (4) Model components represent both background and foreground regions of the scene, and each is explicitly labeled as l c {foreground,background}. Pixels are implicitly labeled according to the component to which they were assigned using Equation (4). Thus a pixel s classification is determined by the global characteristic of the component to which it is assigned, rather than it s own local characteristic. Within this framework we are able to specify the notions of scene foreground and background, and differentiate between them, at object level rather than pixel level. All model components are updated by the statistics of their assigned pixels. In general, background components are updated more slowly than foreground components, reflecting the expectation that foreground will exhibit more dynamic behaviour. The initialisation, assignments, and update procedures are described in more detail in the remainder of this section. Our approach expresses the spatial structure of the observed scene in the background model, and helps to reduce some types of misclassification by generating a spatially coherent segmentation. Further details are given in [2]. 3 Segmentation of Water Birds Our experiments [2] have shown that our background model is robust in scenes characterised by moving foliage and water. In contrast, we have also shown that the performance of both per-pixel and MRF-based models deteriorates drastically under the same conditions. As an example, we have compared our algorithm with Stauffer and Grimson s model, and Migdal and Grimson s model, on
6 a sequence showing a duck approaching water: example segmentations for a single image in this sequence are shown in Figure 1). The output from our algorithm is consistently better, showing considerably less false positive foreground pixels than either of the other two algorithms. Both of the other algorithms show high numbers of incorrectly labelled foreground pixels, likely to result in the incorrect identification of foreground objects. From these results, and others, we conclude that our algorithm is better adapted for use in natural bird habitats, where such processes are likely to occur. Figure 1: Foreground pixels classified in an image sequence showing a duck. From top to bottom, left to right: original image, using a per-pixel model, using an MRF model, using our model. 4 Planned Development of Our Work: The Skomer Island Colony Our current interest is in video foreground segmentation to automatically analyse image sequences captured from a Guillemot nesting site on Skomer Island, North wales. Our intention is to automatically determine the number of birds on a series of cliff faces, and to monitor the condition of specific nest sites. We are intending to utilise existing hardware, which has been set up to facilitate manual monitoring of the site. This includes digital cameras, and a wireless network. As a starting point, we have acquired image sequences of a particular cliff area, captured using a digital camera operated in a time-lapse capture mode. This sequence presents a significantly
7 different data profile than that normally encountered when using video capture equipment: firstly, the images are captured at seconds apart, and secondly they are of a relatively high resolution (3072 x 2304 pixels). Example images are shown in Figure 2. It should be noted that since these images are captured at a high resolution, low-level details are not visible in Figure 2. An enlarged section of an image is shown in Figure 3. Figure 2: Sample images captures captured from the Skomer breeding site. The images are also characterised by two important artefacts: significant camera shake, and the presence of moving water, both of which are likely to confound standard segmentation techniques. Our method (described in section 2) is more suitable for processing under these conditions; however, capture rate, resolution, and other features of the data necessitate some further development. We now describe modifications that we intend to make to our algorithm, and a proposed architecture by which we might collectively monitor several sites on the island. 4.1 Proposed Modifications to Our Algorithm Our algorithm (like Migdal and Grimson s) is computationally intensive, which makes real-time procesing at video frames rates (e.g. 25Hz) difficult, even at lower image resolution. To some extent this is exacerbated by the increased resolution of the Skomer images: we anticipate that segmentation of an individual image could take up to one minute on typical PC hardware. However, unlike video data, the time-lapse data is captured much more infrequently. Not only does this
8 Figure 3: Enlarged section of an image from the Skomer breeding site. facilitate our projected processing requirements, but potentially allows us in incorporate additional functionality to further improve quality. We could, for example, capture images every 5 minutes. The time between image frames does introduce one additional problem, however. Our segmentation method uses models of both the foreground and background to classify pixels. The foreground model relies on a relatively high frame rate, and assumes that the movement of individual objects is relatively small between successive images. We can no longer make this assumption with images captured minutes apart. Instead we propose to dispense with the foreground model (instead assuming a uniform colour distribution across the foreground), and develop a more sophisticated background model. In our experiments ([2]) we have identified that our model is effective at eliminating pixel mis-classifications caused by small background movements. However, the MRF model is also effective in enforcing spatial coherence. We thus propose to develop our model by combining our background representation (spatially distributed Gaussians) with an MRF framework. Each image pixel will then be classified by assignment to one of a number of possible labels: either the foreground, or to a component which is spatially local. The observation likelihood, p(x l) in Equation 2, for background labels will then be given by the colour distribution of those local Gaussian components. We believe that this will facilitate an effective segmentation using images which are captured minutes apart, and retain the robustness of our model to variations in the background. 4.2 Initialisation A remaining problem is that of model initialisation. Surveillance systems are often initialised under controlled conditions; however, this is not possible at the Skomer site. Consequently, we intend to construct a background model over a longer period of time, perhaps several hours or days, before
9 monitoring becomes active. Further investigation is required to determine a suitable initialisation period, which will depend largely on bird mobility. Lighting changes will also need to be accounted for when using an extended initialisation period. 4.3 Monitoring Functions Our proposed system is intended to provide two functions: estimation of bird populations over a cliff face area, and the monitoring of parent behaviour at specific nest sites. We anticipate that global monitoring can be effected completely autonomously. However, monitoring of individual nest sites may require the user to identify those sites manually in the image scene. It may be possible to infer the position of nest sites automatically from parent behaviour; however, this remains an open question. 4.4 Hardware Architecture Our current system is developed as a single stand-alone monitoring device, comprising a single camera. However, we wish to simultaneously monitor several cliff face scenes at the Skomer site, using several cameras, and provide an aggregated set of results in real-time. To this end, our aim is to develop a distributed system comprising a set of cameras connected connected across a local network, and single PC at which the results are made available. There is an existing wireless network at the Skomer site which we can use to faciltate this. The high bandwidth associated with transmitting video from multiple sources over a local network is often prohibitive. For this reason, a common approach is to operate low-level processing on-camera (using a smart camera ), and to transmit the results only. In this case, however, each camera will generate an image only every few minutes. This means that, with relatively low bandwidth, it is feasible to transmit complete images and perform all processing centrally. This significantly simplifies implementation, subject to the processing resources available to the central processing unit. Based on our estimates of one minute per frame, and a frame capture interval of 5 minutes, we might expect to be able to process 5 camera outputs on a single central machine in real-time. 4.5 Summary We have described a proposed method for using visual surveillance techniques to automatically a monitor a seabird population. In particular, we have considered the application of such techniques to a specific site on Skomer island, populated by Guillemots. We have described our previous work in low-level video image processing, and shown how we have successfully addressed specific problems associated with surveillance in natural environments. Examination of images collected from Skomer island have revealed a number of challenges associated with both image features, and the existing installed hardware. We have described how we intend to develop our existing work to support a distributed surveillance system capable of providing automated monitoring at this site.
10 References [1] G. Dalley, J. Migdal, and W. Grimson. Background subtraction for temporally irregular dynamic textures. In Proc. of IEEE Workshop on Applications of Computer Vision, Copper Mountain, CO, USA, [2] P. Dickinson. Monitoring the Vulnerable using Automated Visual Surveillance. PhD thesis, University of Lincoln, UK, [3] A. Elgammal, R. Duraiswami, D. Harwood, and L. Davis. Background and foreground modeling using nonparametric kernel density estimation for visual surveillance. Proceedings of the IEEE, 90(7): , July [4] W. Hu, X. Xiao, Z. Fu, D. Xie, T. Tan, and S. Maybank. A system for learning statistical motion patterns. IEEE Transaction on Pattern Analysis and Machine Intelligence, 28(9): , [5] A. Kembhavi, R. Farrell, Y. Lou, D. Jacobs, R. Duraiswami, and L. Davis. Traking down under: following the Satin Bowerbird. In Proc. of IEEE Workshop on Applications of Computer Vision, Copper Mountain, CO, USA, [6] J. Migdal and W. Grimson. Background subtraction using Markov thresholds. In Proc. of IEEE Workshop on Applications of Computer Vision / IEEE Workshop on Motion and Video Computing, volume 2, pages 58 65, Breckenridge, CO, USA, January [7] S. Park and J. Aggarwal. Segmentation and tracking of interacting human body parts under occlusion and shadowing. In Proc. of IEEE Workshop on Motion and Video Computing, pages , Orlando, FL, USA, [8] C. Stauffer and W.Grimson. Adaptive background mixture models for real-time tracking. In Proc. of IEEE Conference on Computer Vision and Pattern Recognition, volume 2, pages , Fort Collins, CO, USA, [9] Y. Tian, M. Lu, and A. Hampapur. Robust and efficient foreground analysis for real-time video surveillance. In Proc. of IEEE Conference on Computer Vision and Pattern Recognition, volume 1, pages , San Diego, CA, USA, [10] M. Xu and T. Ellis. Illumination-invariant motion detection using colour mixture models. In Proc. of British Machine Vision Conference, pages , Manchester, UK, [11] K. Younis, B. Aldawood, G. Abandah, S. Alsisan, and J. Igual. Automatic wireless monitoring of bird behavior. In Proc. of IASTED International Conference on Signal and Image Processing, pages , Hawaii, USA, 2008.
A Learning Based Method for Super-Resolution of Low Resolution Images
A Learning Based Method for Super-Resolution of Low Resolution Images Emre Ugur June 1, 2004 [email protected] Abstract The main objective of this project is the study of a learning based method
VEHICLE LOCALISATION AND CLASSIFICATION IN URBAN CCTV STREAMS
VEHICLE LOCALISATION AND CLASSIFICATION IN URBAN CCTV STREAMS Norbert Buch 1, Mark Cracknell 2, James Orwell 1 and Sergio A. Velastin 1 1. Kingston University, Penrhyn Road, Kingston upon Thames, KT1 2EE,
Real-Time Tracking of Pedestrians and Vehicles
Real-Time Tracking of Pedestrians and Vehicles N.T. Siebel and S.J. Maybank. Computational Vision Group Department of Computer Science The University of Reading Reading RG6 6AY, England Abstract We present
Edge tracking for motion segmentation and depth ordering
Edge tracking for motion segmentation and depth ordering P. Smith, T. Drummond and R. Cipolla Department of Engineering University of Cambridge Cambridge CB2 1PZ,UK {pas1001 twd20 cipolla}@eng.cam.ac.uk
Video Surveillance System for Security Applications
Video Surveillance System for Security Applications Vidya A.S. Department of CSE National Institute of Technology Calicut, Kerala, India V. K. Govindan Department of CSE National Institute of Technology
Tracking Groups of Pedestrians in Video Sequences
Tracking Groups of Pedestrians in Video Sequences Jorge S. Marques Pedro M. Jorge Arnaldo J. Abrantes J. M. Lemos IST / ISR ISEL / IST ISEL INESC-ID / IST Lisbon, Portugal Lisbon, Portugal Lisbon, Portugal
False alarm in outdoor environments
Accepted 1.0 Savantic letter 1(6) False alarm in outdoor environments Accepted 1.0 Savantic letter 2(6) Table of contents Revision history 3 References 3 1 Introduction 4 2 Pre-processing 4 3 Detection,
Vision based Vehicle Tracking using a high angle camera
Vision based Vehicle Tracking using a high angle camera Raúl Ignacio Ramos García Dule Shu [email protected] [email protected] Abstract A vehicle tracking and grouping algorithm is presented in this work
Spatio-Temporal Nonparametric Background Modeling and Subtraction
Spatio-Temporal Nonparametric Background Modeling and Subtraction Raviteja Vemulapalli and R. Aravind Department of Electrical engineering Indian Institute of Technology, Madras Background subtraction
Practical Tour of Visual tracking. David Fleet and Allan Jepson January, 2006
Practical Tour of Visual tracking David Fleet and Allan Jepson January, 2006 Designing a Visual Tracker: What is the state? pose and motion (position, velocity, acceleration, ) shape (size, deformation,
Bayesian Image Super-Resolution
Bayesian Image Super-Resolution Michael E. Tipping and Christopher M. Bishop Microsoft Research, Cambridge, U.K..................................................................... Published as: Bayesian
Tracking and Recognition in Sports Videos
Tracking and Recognition in Sports Videos Mustafa Teke a, Masoud Sattari b a Graduate School of Informatics, Middle East Technical University, Ankara, Turkey [email protected] b Department of Computer
Observing Human Behavior in Image Sequences: the Video Hermeneutics Challenge
Observing Human Behavior in Image Sequences: the Video Hermeneutics Challenge Pau Baiget, Jordi Gonzàlez Computer Vision Center, Dept. de Ciències de la Computació, Edifici O, Campus UAB, 08193 Bellaterra,
Tracking performance evaluation on PETS 2015 Challenge datasets
Tracking performance evaluation on PETS 2015 Challenge datasets Tahir Nawaz, Jonathan Boyle, Longzhen Li and James Ferryman Computational Vision Group, School of Systems Engineering University of Reading,
Traffic Flow Monitoring in Crowded Cities
Traffic Flow Monitoring in Crowded Cities John A. Quinn and Rose Nakibuule Faculty of Computing & I.T. Makerere University P.O. Box 7062, Kampala, Uganda {jquinn,rnakibuule}@cit.mak.ac.ug Abstract Traffic
Behavior Analysis in Crowded Environments. XiaogangWang Department of Electronic Engineering The Chinese University of Hong Kong June 25, 2011
Behavior Analysis in Crowded Environments XiaogangWang Department of Electronic Engineering The Chinese University of Hong Kong June 25, 2011 Behavior Analysis in Sparse Scenes Zelnik-Manor & Irani CVPR
How does the Kinect work? John MacCormick
How does the Kinect work? John MacCormick Xbox demo Laptop demo The Kinect uses structured light and machine learning Inferring body position is a two-stage process: first compute a depth map (using structured
Tracking And Object Classification For Automated Surveillance
Tracking And Object Classification For Automated Surveillance Omar Javed and Mubarak Shah Computer Vision ab, University of Central Florida, 4000 Central Florida Blvd, Orlando, Florida 32816, USA {ojaved,shah}@cs.ucf.edu
Environmental Remote Sensing GEOG 2021
Environmental Remote Sensing GEOG 2021 Lecture 4 Image classification 2 Purpose categorising data data abstraction / simplification data interpretation mapping for land cover mapping use land cover class
CCTV - Video Analytics for Traffic Management
CCTV - Video Analytics for Traffic Management Index Purpose Description Relevance for Large Scale Events Technologies Impacts Integration potential Implementation Best Cases and Examples 1 of 12 Purpose
The Delicate Art of Flower Classification
The Delicate Art of Flower Classification Paul Vicol Simon Fraser University University Burnaby, BC [email protected] Note: The following is my contribution to a group project for a graduate machine learning
The Scientific Data Mining Process
Chapter 4 The Scientific Data Mining Process When I use a word, Humpty Dumpty said, in rather a scornful tone, it means just what I choose it to mean neither more nor less. Lewis Carroll [87, p. 214] In
Automatic parameter regulation for a tracking system with an auto-critical function
Automatic parameter regulation for a tracking system with an auto-critical function Daniela Hall INRIA Rhône-Alpes, St. Ismier, France Email: [email protected] Abstract In this article we propose
A Reliability Point and Kalman Filter-based Vehicle Tracking Technique
A Reliability Point and Kalman Filter-based Vehicle Tracing Technique Soo Siang Teoh and Thomas Bräunl Abstract This paper introduces a technique for tracing the movement of vehicles in consecutive video
A ROBUST BACKGROUND REMOVAL ALGORTIHMS
A ROBUST BACKGROUND REMOVAL ALGORTIHMS USING FUZZY C-MEANS CLUSTERING ABSTRACT S.Lakshmi 1 and Dr.V.Sankaranarayanan 2 1 Jeppiaar Engineering College, Chennai [email protected] 2 Director, Crescent
Object Recognition. Selim Aksoy. Bilkent University [email protected]
Image Classification and Object Recognition Selim Aksoy Department of Computer Engineering Bilkent University [email protected] Image classification Image (scene) classification is a fundamental
CS 534: Computer Vision 3D Model-based recognition
CS 534: Computer Vision 3D Model-based recognition Ahmed Elgammal Dept of Computer Science CS 534 3D Model-based Vision - 1 High Level Vision Object Recognition: What it means? Two main recognition tasks:!
Intrusion Detection via Machine Learning for SCADA System Protection
Intrusion Detection via Machine Learning for SCADA System Protection S.L.P. Yasakethu Department of Computing, University of Surrey, Guildford, GU2 7XH, UK. [email protected] J. Jiang Department
Novel Probabilistic Methods for Visual Surveillance Applications
University of Pannonia Information Science and Technology PhD School Thesis Booklet Novel Probabilistic Methods for Visual Surveillance Applications Ákos Utasi Department of Electrical Engineering and
An Improved Adaptive Background Mixture Model for Realtime Tracking with Shadow Detection
In Proc. nd European Worshop on Advanced Video Based Surveillance Systems, AVBS0. Sept 00. VIDEO BASED SURVEILLACE SYSTEMS: Computer Vision and Distributed Processing, Kluwer Academic Publishers An Improved
Speed Performance Improvement of Vehicle Blob Tracking System
Speed Performance Improvement of Vehicle Blob Tracking System Sung Chun Lee and Ram Nevatia University of Southern California, Los Angeles, CA 90089, USA [email protected], [email protected] Abstract. A speed
Vehicle Tracking in Occlusion and Clutter
Vehicle Tracking in Occlusion and Clutter by KURTIS NORMAN MCBRIDE A thesis presented to the University of Waterloo in fulfilment of the thesis requirement for the degree of Master of Applied Science in
Course: Model, Learning, and Inference: Lecture 5
Course: Model, Learning, and Inference: Lecture 5 Alan Yuille Department of Statistics, UCLA Los Angeles, CA 90095 [email protected] Abstract Probability distributions on structured representation.
Bayesian networks - Time-series models - Apache Spark & Scala
Bayesian networks - Time-series models - Apache Spark & Scala Dr John Sandiford, CTO Bayes Server Data Science London Meetup - November 2014 1 Contents Introduction Bayesian networks Latent variables Anomaly
Journal of Industrial Engineering Research. Adaptive sequence of Key Pose Detection for Human Action Recognition
IWNEST PUBLISHER Journal of Industrial Engineering Research (ISSN: 2077-4559) Journal home page: http://www.iwnest.com/aace/ Adaptive sequence of Key Pose Detection for Human Action Recognition 1 T. Sindhu
Real Time Target Tracking with Pan Tilt Zoom Camera
2009 Digital Image Computing: Techniques and Applications Real Time Target Tracking with Pan Tilt Zoom Camera Pankaj Kumar, Anthony Dick School of Computer Science The University of Adelaide Adelaide,
Multi-view Intelligent Vehicle Surveillance System
Multi-view Intelligent Vehicle Surveillance System S. Denman, C. Fookes, J. Cook, C. Davoren, A. Mamic, G. Farquharson, D. Chen, B. Chen and S. Sridharan Image and Video Research Laboratory Queensland
An Automatic and Accurate Segmentation for High Resolution Satellite Image S.Saumya 1, D.V.Jiji Thanka Ligoshia 2
An Automatic and Accurate Segmentation for High Resolution Satellite Image S.Saumya 1, D.V.Jiji Thanka Ligoshia 2 Assistant Professor, Dept of ECE, Bethlahem Institute of Engineering, Karungal, Tamilnadu,
CS231M Project Report - Automated Real-Time Face Tracking and Blending
CS231M Project Report - Automated Real-Time Face Tracking and Blending Steven Lee, [email protected] June 6, 2015 1 Introduction Summary statement: The goal of this project is to create an Android
Neural Network based Vehicle Classification for Intelligent Traffic Control
Neural Network based Vehicle Classification for Intelligent Traffic Control Saeid Fazli 1, Shahram Mohammadi 2, Morteza Rahmani 3 1,2,3 Electrical Engineering Department, Zanjan University, Zanjan, IRAN
Design of Multi-camera Based Acts Monitoring System for Effective Remote Monitoring Control
보안공학연구논문지 (Journal of Security Engineering), 제 8권 제 3호 2011년 6월 Design of Multi-camera Based Acts Monitoring System for Effective Remote Monitoring Control Ji-Hoon Lim 1), Seoksoo Kim 2) Abstract With
Statistics Graduate Courses
Statistics Graduate Courses STAT 7002--Topics in Statistics-Biological/Physical/Mathematics (cr.arr.).organized study of selected topics. Subjects and earnable credit may vary from semester to semester.
A Real Time Hand Tracking System for Interactive Applications
A Real Time Hand Tracking System for Interactive Applications Siddharth Swarup Rautaray Indian Institute of Information Technology Allahabad ABSTRACT In vision based hand tracking systems color plays an
Colorado School of Mines Computer Vision Professor William Hoff
Professor William Hoff Dept of Electrical Engineering &Computer Science http://inside.mines.edu/~whoff/ 1 Introduction to 2 What is? A process that produces from images of the external world a description
Component Ordering in Independent Component Analysis Based on Data Power
Component Ordering in Independent Component Analysis Based on Data Power Anne Hendrikse Raymond Veldhuis University of Twente University of Twente Fac. EEMCS, Signals and Systems Group Fac. EEMCS, Signals
Mean-Shift Tracking with Random Sampling
1 Mean-Shift Tracking with Random Sampling Alex Po Leung, Shaogang Gong Department of Computer Science Queen Mary, University of London, London, E1 4NS Abstract In this work, boosting the efficiency of
Probabilistic Latent Semantic Analysis (plsa)
Probabilistic Latent Semantic Analysis (plsa) SS 2008 Bayesian Networks Multimedia Computing, Universität Augsburg [email protected] www.multimedia-computing.{de,org} References
Professor, D.Sc. (Tech.) Eugene Kovshov MSTU «STANKIN», Moscow, Russia
Professor, D.Sc. (Tech.) Eugene Kovshov MSTU «STANKIN», Moscow, Russia As of today, the issue of Big Data processing is still of high importance. Data flow is increasingly growing. Processing methods
Web 3.0 image search: a World First
Web 3.0 image search: a World First The digital age has provided a virtually free worldwide digital distribution infrastructure through the internet. Many areas of commerce, government and academia have
Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches
Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches PhD Thesis by Payam Birjandi Director: Prof. Mihai Datcu Problematic
A PHOTOGRAMMETRIC APPRAOCH FOR AUTOMATIC TRAFFIC ASSESSMENT USING CONVENTIONAL CCTV CAMERA
A PHOTOGRAMMETRIC APPRAOCH FOR AUTOMATIC TRAFFIC ASSESSMENT USING CONVENTIONAL CCTV CAMERA N. Zarrinpanjeh a, F. Dadrassjavan b, H. Fattahi c * a Islamic Azad University of Qazvin - [email protected]
Tracking in flussi video 3D. Ing. Samuele Salti
Seminari XXIII ciclo Tracking in flussi video 3D Ing. Tutors: Prof. Tullio Salmon Cinotti Prof. Luigi Di Stefano The Tracking problem Detection Object model, Track initiation, Track termination, Tracking
Machine Learning and Pattern Recognition Logistic Regression
Machine Learning and Pattern Recognition Logistic Regression Course Lecturer:Amos J Storkey Institute for Adaptive and Neural Computation School of Informatics University of Edinburgh Crichton Street,
The Big Data methodology in computer vision systems
The Big Data methodology in computer vision systems Popov S.B. Samara State Aerospace University, Image Processing Systems Institute, Russian Academy of Sciences Abstract. I consider the advantages of
Machine Learning and Data Analysis overview. Department of Cybernetics, Czech Technical University in Prague. http://ida.felk.cvut.
Machine Learning and Data Analysis overview Jiří Kléma Department of Cybernetics, Czech Technical University in Prague http://ida.felk.cvut.cz psyllabus Lecture Lecturer Content 1. J. Kléma Introduction,
Colour Image Segmentation Technique for Screen Printing
60 R.U. Hewage and D.U.J. Sonnadara Department of Physics, University of Colombo, Sri Lanka ABSTRACT Screen-printing is an industry with a large number of applications ranging from printing mobile phone
Finding people in repeated shots of the same scene
Finding people in repeated shots of the same scene Josef Sivic 1 C. Lawrence Zitnick Richard Szeliski 1 University of Oxford Microsoft Research Abstract The goal of this work is to find all occurrences
Robert Collins CSE598G. More on Mean-shift. R.Collins, CSE, PSU CSE598G Spring 2006
More on Mean-shift R.Collins, CSE, PSU Spring 2006 Recall: Kernel Density Estimation Given a set of data samples x i ; i=1...n Convolve with a kernel function H to generate a smooth function f(x) Equivalent
FLEXSYS Motion-based Traffic Analysis and Incident Detection
FLEXSYS Motion-based Traffic Analysis and Incident Detection Authors: Lixin Yang and Hichem Sahli, IBBT/VUB-ETRO Contents.1 Introduction......................................... 1.2 Traffic flow modelling
Machine Learning for Medical Image Analysis. A. Criminisi & the InnerEye team @ MSRC
Machine Learning for Medical Image Analysis A. Criminisi & the InnerEye team @ MSRC Medical image analysis the goal Automatic, semantic analysis and quantification of what observed in medical scans Brain
Low-resolution Character Recognition by Video-based Super-resolution
2009 10th International Conference on Document Analysis and Recognition Low-resolution Character Recognition by Video-based Super-resolution Ataru Ohkura 1, Daisuke Deguchi 1, Tomokazu Takahashi 2, Ichiro
Distributed Vision Processing in Smart Camera Networks
Distributed Vision Processing in Smart Camera Networks CVPR-07 Hamid Aghajan, Stanford University, USA François Berry, Univ. Blaise Pascal, France Horst Bischof, TU Graz, Austria Richard Kleihorst, NXP
Urban Vehicle Tracking using a Combined 3D Model Detector and Classifier
Urban Vehicle Tracing using a Combined 3D Model Detector and Classifier Norbert Buch, Fei Yin, James Orwell, Dimitrios Maris and Sergio A. Velastin Digital Imaging Research Centre, Kingston University,
Anastasios Dimou Researcher CERTH/ITI
Video Surveillance Systems: Current Systems and Future Trends Anastasios Dimou Researcher CERTH/ITI Video Surveillance Systems Definition: A video surveillance system monitors the behaviour, activities,
An Assessment of the Effectiveness of Segmentation Methods on Classification Performance
An Assessment of the Effectiveness of Segmentation Methods on Classification Performance Merve Yildiz 1, Taskin Kavzoglu 2, Ismail Colkesen 3, Emrehan K. Sahin Gebze Institute of Technology, Department
Face Model Fitting on Low Resolution Images
Face Model Fitting on Low Resolution Images Xiaoming Liu Peter H. Tu Frederick W. Wheeler Visualization and Computer Vision Lab General Electric Global Research Center Niskayuna, NY, 1239, USA {liux,tu,wheeler}@research.ge.com
The Optimality of Naive Bayes
The Optimality of Naive Bayes Harry Zhang Faculty of Computer Science University of New Brunswick Fredericton, New Brunswick, Canada email: hzhang@unbca E3B 5A3 Abstract Naive Bayes is one of the most
Lecture 9: Introduction to Pattern Analysis
Lecture 9: Introduction to Pattern Analysis g Features, patterns and classifiers g Components of a PR system g An example g Probability definitions g Bayes Theorem g Gaussian densities Features, patterns
Model-based Synthesis. Tony O Hagan
Model-based Synthesis Tony O Hagan Stochastic models Synthesising evidence through a statistical model 2 Evidence Synthesis (Session 3), Helsinki, 28/10/11 Graphical modelling The kinds of models that
Image Normalization for Illumination Compensation in Facial Images
Image Normalization for Illumination Compensation in Facial Images by Martin D. Levine, Maulin R. Gandhi, Jisnu Bhattacharyya Department of Electrical & Computer Engineering & Center for Intelligent Machines
An Energy-Based Vehicle Tracking System using Principal Component Analysis and Unsupervised ART Network
Proceedings of the 8th WSEAS Int. Conf. on ARTIFICIAL INTELLIGENCE, KNOWLEDGE ENGINEERING & DATA BASES (AIKED '9) ISSN: 179-519 435 ISBN: 978-96-474-51-2 An Energy-Based Vehicle Tracking System using Principal
Data Analytics at NICTA. Stephen Hardy National ICT Australia (NICTA) [email protected]
Data Analytics at NICTA Stephen Hardy National ICT Australia (NICTA) [email protected] NICTA Copyright 2013 Outline Big data = science! Data analytics at NICTA Discrete Finite Infinite Machine Learning
Data Mining Cluster Analysis: Basic Concepts and Algorithms. Lecture Notes for Chapter 8. Introduction to Data Mining
Data Mining Cluster Analysis: Basic Concepts and Algorithms Lecture Notes for Chapter 8 by Tan, Steinbach, Kumar 1 What is Cluster Analysis? Finding groups of objects such that the objects in a group will
Linear Threshold Units
Linear Threshold Units w x hx (... w n x n w We assume that each feature x j and each weight w j is a real number (we will relax this later) We will study three different algorithms for learning linear
EFFICIENT VEHICLE TRACKING AND CLASSIFICATION FOR AN AUTOMATED TRAFFIC SURVEILLANCE SYSTEM
EFFICIENT VEHICLE TRACKING AND CLASSIFICATION FOR AN AUTOMATED TRAFFIC SURVEILLANCE SYSTEM Amol Ambardekar, Mircea Nicolescu, and George Bebis Department of Computer Science and Engineering University
MACHINE LEARNING IN HIGH ENERGY PHYSICS
MACHINE LEARNING IN HIGH ENERGY PHYSICS LECTURE #1 Alex Rogozhnikov, 2015 INTRO NOTES 4 days two lectures, two practice seminars every day this is introductory track to machine learning kaggle competition!
Object tracking & Motion detection in video sequences
Introduction Object tracking & Motion detection in video sequences Recomended link: http://cmp.felk.cvut.cz/~hlavac/teachpresen/17compvision3d/41imagemotion.pdf 1 2 DYNAMIC SCENE ANALYSIS The input to
Basics of Statistical Machine Learning
CS761 Spring 2013 Advanced Machine Learning Basics of Statistical Machine Learning Lecturer: Xiaojin Zhu [email protected] Modern machine learning is rooted in statistics. You will find many familiar
COMPARISON OF OBJECT BASED AND PIXEL BASED CLASSIFICATION OF HIGH RESOLUTION SATELLITE IMAGES USING ARTIFICIAL NEURAL NETWORKS
COMPARISON OF OBJECT BASED AND PIXEL BASED CLASSIFICATION OF HIGH RESOLUTION SATELLITE IMAGES USING ARTIFICIAL NEURAL NETWORKS B.K. Mohan and S. N. Ladha Centre for Studies in Resources Engineering IIT
A Study on SURF Algorithm and Real-Time Tracking Objects Using Optical Flow
, pp.233-237 http://dx.doi.org/10.14257/astl.2014.51.53 A Study on SURF Algorithm and Real-Time Tracking Objects Using Optical Flow Giwoo Kim 1, Hye-Youn Lim 1 and Dae-Seong Kang 1, 1 Department of electronices
Extending Ethernet over 2-wire Copper Networks
Extending Ethernet over 2-wire Copper Networks Executive Summary To improve the efficiency of industrial automation systems, many operators are considering the possibility of IP upgrades to provide centralised
Building an Advanced Invariant Real-Time Human Tracking System
UDC 004.41 Building an Advanced Invariant Real-Time Human Tracking System Fayez Idris 1, Mazen Abu_Zaher 2, Rashad J. Rasras 3, and Ibrahiem M. M. El Emary 4 1 School of Informatics and Computing, German-Jordanian
Introduction to Deep Learning Variational Inference, Mean Field Theory
Introduction to Deep Learning Variational Inference, Mean Field Theory 1 Iasonas Kokkinos [email protected] Center for Visual Computing Ecole Centrale Paris Galen Group INRIA-Saclay Lecture 3: recap
Segmentation of building models from dense 3D point-clouds
Segmentation of building models from dense 3D point-clouds Joachim Bauer, Konrad Karner, Konrad Schindler, Andreas Klaus, Christopher Zach VRVis Research Center for Virtual Reality and Visualization, Institute
Object tracking in video scenes
A Seminar On Object tracking in video scenes Presented by Alok K. Watve M.Tech. IT 1st year Indian Institue of Technology, Kharagpur Under the guidance of Dr. Shamik Sural Assistant Professor School of
Tracking Moving Objects In Video Sequences Yiwei Wang, Robert E. Van Dyck, and John F. Doherty Department of Electrical Engineering The Pennsylvania State University University Park, PA16802 Abstract{Object
Classifying Manipulation Primitives from Visual Data
Classifying Manipulation Primitives from Visual Data Sandy Huang and Dylan Hadfield-Menell Abstract One approach to learning from demonstrations in robotics is to make use of a classifier to predict if
