Background Subtraction

Similar documents
EFFICIENT VEHICLE TRACKING AND CLASSIFICATION FOR AN AUTOMATED TRAFFIC SURVEILLANCE SYSTEM

Speed Performance Improvement of Vehicle Blob Tracking System

VSSN 06 Algorithm Competition

Video Surveillance System for Security Applications

A Study on SURF Algorithm and Real-Time Tracking Objects Using Optical Flow

Tracking and Recognition in Sports Videos

Vision based Vehicle Tracking using a high angle camera

Tracking And Object Classification For Automated Surveillance

VEHICLE LOCALISATION AND CLASSIFICATION IN URBAN CCTV STREAMS

A PHOTOGRAMMETRIC APPRAOCH FOR AUTOMATIC TRAFFIC ASSESSMENT USING CONVENTIONAL CCTV CAMERA

Environmental Remote Sensing GEOG 2021

Real-Time Tracking of Pedestrians and Vehicles

Spatio-Temporal Nonparametric Background Modeling and Subtraction

VECTORAL IMAGING THE NEW DIRECTION IN AUTOMATED OPTICAL INSPECTION

Towards License Plate Recognition: Comparying Moving Objects Segmentation Approaches

Object tracking & Motion detection in video sequences

Tracking of Small Unmanned Aerial Vehicles

Autonomous Monitoring of Cliff Nesting Seabirds using Computer Vision

Human behavior analysis from videos using optical flow

A Learning Based Method for Super-Resolution of Low Resolution Images

Neural Network based Vehicle Classification for Intelligent Traffic Control

Mean-Shift Tracking with Random Sampling

LIST OF CONTENTS CHAPTER CONTENT PAGE DECLARATION DEDICATION ACKNOWLEDGEMENTS ABSTRACT ABSTRAK

WAKING up without the sound of an alarm clock is a

A ROBUST BACKGROUND REMOVAL ALGORTIHMS

Automatic Traffic Estimation Using Image Processing

Multi-view Intelligent Vehicle Surveillance System

CCTV - Video Analytics for Traffic Management

An Intelligent Video Surveillance Framework for Remote Monitoring M.Sivarathinabala, S.Abirami

Vehicle Tracking in Occlusion and Clutter

Traffic Flow Monitoring in Crowded Cities

To determine vertical angular frequency, we need to express vertical viewing angle in terms of and. 2tan. (degree). (1 pt)

Video Analytics A New Standard

Decision Trees from large Databases: SLIQ

From Product Management Telephone Nuremberg

Real-Time Airport Security Checkpoint Surveillance Using a Camera Network

Object Recognition and Template Matching

Efficient Background Subtraction and Shadow Removal Technique for Multiple Human object Tracking

Traffic Monitoring Systems. Technology and sensors

CHAPTER 6 TEXTURE ANIMATION

Video Examples. Department of Electrical Engineering and Computer Science

Tracking Groups of Pedestrians in Video Sequences

3D Template-Based Single Camera Multiple Object Tracking

An Energy-Based Vehicle Tracking System using Principal Component Analysis and Unsupervised ART Network

Predict the Popularity of YouTube Videos Using Early View Data

Static Environment Recognition Using Omni-camera from a Moving Vehicle

Detecting and Tracking Moving Objects for Video Surveillance

Automatic parameter regulation for a tracking system with an auto-critical function

Colour Image Segmentation Technique for Screen Printing

OBJECT TRACKING USING LOG-POLAR TRANSFORMATION

Motion Activated Video Surveillance Using TI DSP

Bayesian networks - Time-series models - Apache Spark & Scala

Real Time Target Tracking with Pan Tilt Zoom Camera

Development of an automated Red Light Violation Detection System (RLVDS) for Indian vehicles

Practical Tour of Visual tracking. David Fleet and Allan Jepson January, 2006

Virtual Mouse Using a Webcam

The Scientific Data Mining Process

A Real Time Hand Tracking System for Interactive Applications

Tracking performance evaluation on PETS 2015 Challenge datasets

Real-time Traffic Congestion Detection Based on Video Analysis

The Big Data methodology in computer vision systems

University of Leeds SCHOOL OF COMPUTER STUDIES RESEARCH REPORT SERIES Report

Fall detection in the elderly by head tracking

A Highly Robust Vehicle Detection, Tracking and Speed Measurement Model for Intelligent Transport Systems

Analysis of kiva.com Microlending Service! Hoda Eydgahi Julia Ma Andy Bardagjy December 9, 2010 MAS.622j

ROBUST VEHICLE TRACKING IN VIDEO IMAGES BEING TAKEN FROM A HELICOPTER

Building an Advanced Invariant Real-Time Human Tracking System

Local features and matching. Image classification & object localization

A Reliability Point and Kalman Filter-based Vehicle Tracking Technique

Real-Time Background Estimation of Traffic Imagery Using Group-Based Histogram *

Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches

Introduction. Selim Aksoy. Bilkent University

System Architecture of the System. Input Real time Video. Background Subtraction. Moving Object Detection. Human tracking.

Implementation of Canny Edge Detector of color images on CELL/B.E. Architecture.

Computational Foundations of Cognitive Science

Linear Threshold Units

Anomaly Detection and Predictive Maintenance

A Computer Vision System on a Chip: a case study from the automotive domain

Deterministic Sampling-based Switching Kalman Filtering for Vehicle Tracking

False alarm in outdoor environments

FLEXSYS Motion-based Traffic Analysis and Incident Detection

Low-resolution Character Recognition by Video-based Super-resolution

Novel Probabilistic Methods for Visual Surveillance Applications

Supervised Feature Selection & Unsupervised Dimensionality Reduction

Optical Flow. Shenlong Wang CSC2541 Course Presentation Feb 2, 2016

Two-Frame Motion Estimation Based on Polynomial Expansion

E27 SPRING 2013 ZUCKER PROJECT 2 PROJECT 2 AUGMENTED REALITY GAMING SYSTEM

Assessment. Presenter: Yupu Zhang, Guoliang Jin, Tuo Wang Computer Vision 2008 Fall

VEHICLE TRACKING AND SPEED ESTIMATION SYSTEM CHAN CHIA YIK. Report submitted in partial fulfillment of the requirements

Journal of Industrial Engineering Research. Adaptive sequence of Key Pose Detection for Human Action Recognition

A Robust Multiple Object Tracking for Sport Applications 1) Thomas Mauthner, Horst Bischof

Design of Multi-camera Based Acts Monitoring System for Effective Remote Monitoring Control

EM Clustering Approach for Multi-Dimensional Analysis of Big Data Set

Lecture 8: Signal Detection and Noise Assumption

Analecta Vol. 8, No. 2 ISSN

The Advantages of Using a Fixed Stereo Vision sensor

Emotion Detection from Speech

APPM4720/5720: Fast algorithms for big data. Gunnar Martinsson The University of Colorado at Boulder

Feature Tracking and Optical Flow

Transcription:

Background Subtraction Birgi Tamersoy The University of Texas at Austin September 29 th, 2009

Background Subtraction Given an image (mostly likely to be a video frame), we want to identify the foreground objects in that image! Motivation In most cases, objects are of interest, not the scene. Makes our life easier: less processing costs, and less room for error.

Widely Used! Traffic monitoring (counting vehicles, detecting & tracking vehicles), Human action recognition (run, walk, jump, squat,...), Human-computer interaction ( human interface ), Object tracking (watched tennis lately?!?), And in many other cool applications of computer vision such as digital forensics. http://www.crime-scene-investigator.net/ DigitalRecording.html

Requirements A reliable and robust background subtraction algorithm should handle: Sudden or gradual illumination changes, High frequency, repetitive motion in the background (such as tree leaves, flags, waves,...), and Long-term scene changes (a car is parked for a month).

Simple Approach Image at time t: I (x, y, t) Background at time t: B(x, y, t) > Th 1. Estimate the background for time t. 2. Subtract the estimated background from the input frame. 3. Apply a threshold, Th, to the absolute difference to get the foreground mask. But, how can we estimate the background?

Frame Differencing Background is estimated to be the previous frame. Background subtraction equation then becomes: B(x, y, t) = I (x, y, t 1) I (x, y, t) I (x, y, t 1) > Th Depending on the object structure, speed, frame rate and global threshold, this approach may or may not be useful (usually not). > Th

Frame Differencing Th = 25 Th = 50 Th = 100 Th = 200

Mean Filter In this case the background is the mean of the previous n frames: B(x, y, t) = 1 n I (x, y, t) 1 n n 1 i=0 n 1 i=0 I (x, y, t i) I (x, y, t i) > Th For n = 10: Estimated Background Foreground Mask

Mean Filter For n = 20: Estimated Background Foreground Mask For n = 50: Estimated Background Foreground Mask

Median Filter Assuming that the background is more likely to appear in a scene, we can use the median of the previous n frames as the background model: For n = 10: B(x, y, t) = median{i (x, y, t i)} I (x, y, t) median{i (x, y, t i)} > Th where i {0,..., n 1}. Estimated Background Foreground Mask

Median Filter For n = 20: Estimated Background Foreground Mask For n = 50: Estimated Background Foreground Mask

Advantages vs. Shortcomings Advantages: Extremely easy to implement and use! All pretty fast. Corresponding background models are not constant, they change over time. Disadvantages: Accuracy of frame differencing depends on object speed and frame rate! Mean and median background models have relatively high memory requirements. In case of the mean background model, this can be handled by a running average: B(x, y, t) = t 1 t B(x, y, t 1) + 1 t I (x, y, t) or more generally: B(x, y, t) = (1 α)b(x, y, t 1) + αi (x, y, t) where α is the learning rate.

Advantages vs. Shortcomings Disadvantages: There is another major problem with these simple approaches: I (x, y, t) B(x, y, t) > Th 1. There is one global threshold, Th, for all pixels in the image. 2. And even a bigger problem: this threshold is not a function of t. So, these approaches will not give good results in the following conditions: if the background is bimodal, if the scene contains many, slowly moving objects (mean & median), if the objects are fast and frame rate is slow (frame differencing), and if general lighting conditions in the scene change with time!

The Paper on Background Subtraction Adaptive Background Mixture Models for Real-Time Tracking Chris Stauffer & W.E.L. Grimson

Motivation A robust background subtraction algorithm should handle: lighting changes, repetitive motions from clutter and long-term scene changes. Stauffer & Grimson

A Quick Reminder: Normal (Gaussian) Distribution Univariate: Multivariate: N (x µ, σ 2 ) = 1 (x µ)2 e 2σ 2 2πσ 2 N (x µ, Σ) = 1 (2π) D/2 1 Σ 1/2 e 1 2 (x µ)t Σ 1 (x µ) http://en.wikipedia.org/wiki/normal distribution

Algorithm Overview The values of a particular pixel is modeled as a mixture of adaptive Gaussians. Why mixture? Multiple surfaces appear in a pixel. Why adaptive? Lighting conditions change. At each iteration Gaussians are evaluated using a simple heuristic to determine which ones are mostly likely to correspond to the background. Pixels that do not match with the background Gaussians are classified as foreground. Foreground pixels are grouped using 2D connected component analysis.

Online Mixture Model At any time t, what is known about a particular pixel, (x 0, y 0 ), is its history: {X 1,..., X t } = {I (x 0, y 0, i) : 1 i t} This history is modeled by a mixture of K Gaussian distributions: P(X t ) = K i=1 ω i,t N (X t µ i,t, Σ i,t ) where N (X t µ it, Σ i,t ) = 1 1 e 1 (2π) D/2 Σ i,t 1/2 2 (Xt µ i,t) T Σ 1 What is the dimensionality of the Gaussian? i,t (Xt µ i,t)

Online Mixture Model If we assume gray scale images and set K = 5, history of a pixel will be something like this:

Model Adaptation An on-line K-means approximation is used to update the Gaussians. If a new pixel value, X t+1, can be matched to one of the existing Gaussians (within 2.5σ), that Gaussian s µ i,t+1 and σi,t+1 2 are updated as follows: µ i,t+1 = (1 ρ)µ i,t + ρx t+1 and σ 2 i,t+1 = (1 ρ)σ2 i,t + ρ(x t+1 µ i,t+1 ) 2 where ρ = αn (X t+1 µ i,t, σi,t 2 ) and α is a learning rate. Prior weights of all Gaussians are adjusted as follows: ω i,t+1 = (1 α)ω i,t + α(m i,t+1 ) where M i,t+1 = 1 for the matching Gaussian and M i,t+1 = 0 for all the others.

Model Adaptation If X t+1 do not match to any of the K existing Gaussians, the least probably distribution is replaced with a new one. Warning!!! Least probably in the ω/σ sense (will be explained). New distribution has µ t+1 = X t+1, a high variance and a low prior weight.

Background Model Estimation Heuristic: the Gaussians with the most supporting evidence and least variance should correspond to the background (Why?). The Gaussians are ordered by the value of ω/σ (high support & less variance will give a high value). Then simply the first B distributions are chosen as the background model: B = argmin b ( b i=1 ω i > T ) where T is minimum portion of the image which is expected to be background.

Background Model Estimation After background model estimation red distributions become the background model and black distributions are considered to be foreground.

Advantages vs. Shortcomings Advantages: A different threshold is selected for each pixel. These pixel-wise thresholds are adapting by time. Objects are allowed to become part of the background without destroying the existing background model. Provides fast recovery. Disadvantages: Cannot deal with sudden, drastic lighting changes! Initializing the Gaussians is important (median filtering). There are relatively many parameters, and they should be selected intelligently.

Does it get more complicated? Chen & Aggarwal: The likelihood of a pixel being covered or uncovered is decided by the relative coordinates of optical flow vector vertices in its neighborhood. Oliver et al.: Eigenbackgrounds and its variations. Seki et al.: Image variations at neighboring image blocks have strong correlation.

Example: A Simple & Effective Background Subtraction Approach Adaptive Background Mixture Model (Stauffer & Grimson) + 3D Connected Component Analysis (3 rd dimension: time) 3D connected component analysis incorporates both spatial and temporal information to the background model (by Goo et al.)!

Video Examples

Summary Simple background subtraction approaches such as frame differencing, mean and median filtering, are pretty fast. However, their global, constant thresholds make them insufficient for challenging real-world problems. Adaptive background mixture model approach can handle challenging situations: such as bimodal backgrounds, long-term scene changes and repetitive motions in the clutter. Adaptive background mixture model can further be improved by incorporating temporal information, or using some regional background subtraction approaches in conjunction with it.