T O B C A T C A S E G E O V I S A T DETECTIE E N B L U R R I N G V A N P E R S O N E N IN P A N O R A MISCHE BEELDEN



Similar documents
G E N E R A L A P P R O A CH: LO O K I N G F O R D O M I N A N T O R I E N T A T I O N I N I M A G E P A T C H E S

E27 SPRING 2013 ZUCKER PROJECT 2 PROJECT 2 AUGMENTED REALITY GAMING SYSTEM

Assessment. Presenter: Yupu Zhang, Guoliang Jin, Tuo Wang Computer Vision 2008 Fall

Computer Vision. Color image processing. 25 August 2014

Example: Credit card default, we may be more interested in predicting the probabilty of a default than classifying individuals as default or not.

Chapter 6. The stacking ensemble approach

Implementation of Canny Edge Detector of color images on CELL/B.E. Architecture.

Color Balancing Techniques

Tracking of Small Unmanned Aerial Vehicles

EMR-9 Quick Start Guide (00175)D 1

Cloud tracking with optical flow for short-term solar forecasting

Environmental Remote Sensing GEOG 2021

Lecture 6: Classification & Localization. boris. ginzburg@intel.com

Analecta Vol. 8, No. 2 ISSN

Computer Vision. License Plate Recognition Klaas Dijkstra - Jaap van de Loosdrecht

Convolution. 1D Formula: 2D Formula: Example on the web:

Traffic Prediction and Analysis using a Big Data and Visualisation Approach

Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches

Data Mining. Nonlinear Classification

MassArt Studio Foundation: Visual Language Digital Media Cookbook, Fall 2013

Tattoo Detection for Soft Biometric De-Identification Based on Convolutional NeuralNetworks

Visual Structure Analysis of Flow Charts in Patent Images

Scanners and How to Use Them

Visualization of 2D Domains

WHITE PAPER. Are More Pixels Better? Resolution Does it Really Matter?

Active Learning SVM for Blogs recommendation

CS1112 Spring 2014 Project 4. Objectives. 3 Pixelation for Identity Protection. due Thursday, 3/27, at 11pm

Robust Real-Time Face Detection

Machine Learning Final Project Spam Filtering

MGL Avionics. MapMaker 2. User guide

Multi-Zone Adjustment

BlipTrack. Case Studies. ANPR vs. Bluetooth. This case describes how data from ANPR cameras match corresponding Bluetooth sensor data.

BEHAVIOR BASED CREDIT CARD FRAUD DETECTION USING SUPPORT VECTOR MACHINES

Interpolation of RGB components in Bayer CFA images

Speed Performance Improvement of Vehicle Blob Tracking System

A Logistic Regression Approach to Ad Click Prediction

Object Recognition. Selim Aksoy. Bilkent University

VEHICLE LOCALISATION AND CLASSIFICATION IN URBAN CCTV STREAMS

Automated Monitoring System for Fall Detection in the Elderly

AngioSys 2.0 Image Analysis Software Manual

User Tutorial on Changing Frame Size, Window Size, and Screen Resolution for The Original Version of The Cancer-Rates.Info/NJ Application

Annual Giving Appeal Targeting & Effectiveness

Blind Deconvolution of Barcodes via Dictionary Analysis and Wiener Filter of Barcode Subsections

Tutorial for Tracker and Supporting Software By David Chandler

REAL TIME TRAFFIC LIGHT CONTROL USING IMAGE PROCESSING

THE BENEFITS OF SIGNAL GROUP ORIENTED CONTROL

Predicting the Stock Market with News Articles

Web-Scale Extraction of Structured Data Michael J. Cafarella, Jayant Madhavan & Alon Halevy

Colour Image Segmentation Technique for Screen Printing

Applying Machine Learning to Stock Market Trading Bryce Taylor

Neovision2 Performance Evaluation Protocol

Twelve. Figure 12.1: 3D Curved MPR Viewer Window

Lecture 9: Introduction to Pattern Analysis

Video Analytics A New Standard

Vision based Vehicle Tracking using a high angle camera

Using the ipad and Dropbox For Your Website Gallery Photos and Movie Clips

Supervised Classification workflow in ENVI 4.8 using WorldView-2 imagery

Open icon. The Select Layer To Add dialog opens. Click here to display

Mouse Control using a Web Camera based on Colour Detection

SAMPLE MIDTERM QUESTIONS

Otis Photo Lab Inkjet Printing Demo

SE05: Getting Started with Cognex DataMan Bar Code Readers - Hands On Lab Werner Solution Expo April 8 & 9

BCC Multi Stripe Wipe

Advanced Scanning Techniques

MobileMapper 6 White Paper

Logistic Regression for Spam Filtering

ALARM SYSTEM INSTALLATION GUIDE

Neural Network based Vehicle Classification for Intelligent Traffic Control

Face detection is a process of localizing and extracting the face region from the

Sentiment analysis using emoticons

Gas Dynamics Prof. T. M. Muruganandam Department of Aerospace Engineering Indian Institute of Technology, Madras. Module No - 12 Lecture No - 25

Template-based Eye and Mouth Detection for 3D Video Conferencing

Joomla Article Advanced Topics: Table Layouts

Interactive Flag Identification Using a Fuzzy-Neural Technique

Calculation of Minimum Distances. Minimum Distance to Means. Σi i = 1

Automatic Labeling of Lane Markings for Autonomous Vehicles

Each ievo reader needs its own network cable, as each reader has its own IP address. Ensure this has been taken into consideration

EFX Keying/Alpha plugins for After Effects

DATA LAYOUT AND LEVEL-OF-DETAIL CONTROL FOR FLOOD DATA VISUALIZATION

CCTV - Video Analytics for Traffic Management

Spatial data analysis: retrieval, (re)classification and measurement operations

Measurement of the effect of biomass burning aerosol on inhibition of cloud formation over the Amazon

CS231M Project Report - Automated Real-Time Face Tracking and Blending

UNIVERSITY OF OSLO. Faculty of Mathematics and Natural Sciences

Relative Permeability Measurement in Rock Fractures

Cardiac Tool User s Guide. Cardiac Tool. Version 1.0

3 hours One paper 70 Marks. Areas of Learning Theory

Image Optimization GUIDE

SESSION DEPENDENT DE-IDENTIFICATION OF ELECTRONIC MEDICAL RECORDS

T : Classification as Spam or Ham using Naive Bayes Classifier. Santosh Tirunagari :

Blood Vessel Classification into Arteries and Veins in Retinal Images

Data Visualization Techniques

Face Model Fitting on Low Resolution Images

Vision-Based Blind Spot Detection Using Optical Flow

Topographic Change Detection Using CloudCompare Version 1.0

Investigation of Color Aliasing of High Spatial Frequencies and Edges for Bayer-Pattern Sensors and Foveon X3 Direct Image Sensors

Galaxy Morphological Classification

Perception of Light and Color

The Role of Size Normalization on the Recognition Rate of Handwritten Numerals

Transcription:

T O B C A T C A S E G E O V I S A T DETECTIE E N B L U R R I N G V A N P E R S O N E N IN P A N O R A MISCHE BEELDEN Goal is to process 360 degree images and detect two object categories 1. Pedestrians, make sure they are blurred out so that privacy issues are no longer an actual issue. Would be good if this happens during the capturing of the data. 2. Number plates of cars, again to ensure privacy of people. Example of such a 360 degree panoramic view: There must be a possibility to go through the blurred images and apply an extra blurring manually if something was forgotten. FIRST TEST USING THE LATENTSVM BUILT-IN OPENCV DETECTOR Since there is a built in object detector with several object models the OpenCV code was tested first. However processing an image of 4880x2400 pixels is quite time consuming up to 5 minutes to match to model to the complete image. Some statistics, keeping in mind that the detection score should be > 0.3 (manually entered) 4800*2400 pixels all scale ranges 5 minutes 2400*1200 pixels all scale ranges 1.3 min We notice two things, the smaller the resolution, the faster the processing. However if the image is resized, than the actual cars contain less pixel information. In the images we can see that on the right a car detected at full resolution is lost at half resolution. This is also a downside of this technique. Going further to a resolution of 600*300 pixels results in not a single car being detected.

We applied the same detector but now used the pretrained pedestrian model. Processing times raised a bit higher since the model is smaller and thus more image patches are available for classification. Here we see two remarkable effects. At full resolution we have multiple false positive detections but we do succeed in detecting all available pedestrians. Lowering the resolution again leads to losing the pedestrians completely. Results can be seen below.

Our research group has an optimized C implementation of the algorithm. This was used to perform detections on the complete dataset available. If you would be interested in processing your data with these detectors that are not publicly available, then please mail to toon.goedeme@kuleuven.be. SECOND TEST USING THE EAVISE C-VERSION OF THE LATENTSVM DETECTOR AND PROCESSING THE RETRIEVED RESULTS A dataset of 450 images was manually annotated in order to get a ground truth dataset. Based on that data a scale location mapping was made. Again the bounding box center Y value is plotted against the height of the bounding box.

In the above relation the two blue borders define regions in which objects can occur. Going outside those borders is not smart since that are sky and car areas. The red line is a linear relation mapped onto the data with the green borders describing a -3sigma / +3sigma range around that linear mapping. We then processed the 450 images and retrieved the thresholded detections. As a threshold a score of 0 was used. The result is quite different and will be explained below. Here we see a mapping of every single detection in those 450 images. This includes true positive and false positive detections. However from our training data we know that it is stupid to perform detections outside of our boundaries. The larger spread of detections is mainly due to the fact that we are performing detections on several scale layers of the scale pyramid. Looking below, the green dots are actually the only points still interesting for us, so we can remove the others.

However, we now know that elements outside the green borders are quite impossible. The borders should capture about 98 percent of all the possible scale occurrences of pedestrians. It is safe to decide to remove all points that are not in those regions because they will not fulfil the following rules 1. It is a possible location in the image to have a pedestrian 2. The scale at that position is possible, so larger near the camera and smaller away from the camera So only the red dots are actually detections that should be kept. Using this we can see already that a large set of false positive detections can be removed and thus the accuracy of our algorithm will increase.

We applied the same steps on two images 1. Manual annotations 2. Standard model detection using optimized C-implementation

We immediately notice that false positive detections occur in the air, even if we put the score high enough. 3. Using the retrieved model relation between scale and position to filter detections

POSSIBLE EXPANSIONS TO THIS APPROACH We notice that several persons are missed due to being on a bicycle. Using a bicycle model we do retrieve these objects and we can define the persons above. The only downside of this extra approach is finding a decent threshold value to apply so that false positives can be filtered out without losing crucial detections. Again a space scale relation for bicycles could be built up here to reach a cleaner image and to classify each retrieved detection correctly. MAKING THE BORDERS MORE STRICTLY We know that when moving further away from the car, the variation in scales in the training data becomes clearly smaller. Therefore we want to redefine the borders.

By applying these new rules we get even better results and less false positive detections inside of the image database. A check with images shows us we detect all pedestrians without a flood of false positives. We even succeed in removing objects that still crawled through the previous boundaries.

But one of the most problematic items are linear structures with the correct size and location who still have a large score. Some examples: Thresholding even further is possible, losing these detections but giving raise to loosing good detections because of their lower score as seen below. First image has detection score of -0.36 and applying a more strict score threshold will result in losing the detection. The second image has a score of -0.46 but here the question rises if seeing their backs is actually still a privacy issue here. REMARK: a good discussion on what can be seen as a privacy invasion should be held here.

By adding the score > 0 threshold we keep only the green detections, which are actual pedestrians. Also an interface was built to calculate the smallest scale still classified as a pedestrian to get an insight in the limits of the model used. Final result pedestrians detected by using a smart approach!

However one element still results in a problem, being a yellow-blue traffic sign construction as seen here and it is consistent in several images: In order to solve this last problem, two approaches were implemented and shall be discussed below. SEGMENTING TRAFFIC POLES FOR A ROUNDABOUT POST DETECTION CLASSIFICATION BASED ON A SELF CONSTRUCTED RELATION Looking purely on the color distribution of such a pole, we can see the following properties Central yellow color region of the lower 2/3 Upper 1/3 has a blue sign Color separation can be optimized in the YUV or HSV space Combine this with eliminating the detection of a similar sign, see figure2. First collected a small set of testing data to learn some relations

Segmentation of yellow objects looking at different color spaces 1. Looking at the YUV space U space thresholding on darker regions V space thresholding on lighter regions IWT-TETRA TOBCAT (n 120135) 2. Looking at the HSV space S space thresholding on lighter regions H space thresholding on darker regions

3. Looking at the YMC color space Thresholding on the yellow channel finally gave the clearest result for thresholding the poles. After that we needed to define blue plates and try to segment those. This was done based on the normalized blue channel and to ensure that the top 50% has a certain amount of blue pixels. After combining this all we had the following results into two rules we could classify the poles 1. Yellow poles if percentage of yellow is crossed no person 2. Blue plates if a percentage in the top 50% region is crossed no person On the following set this gave a pretty decent result

Looking at the output We notice that most classifications are correctly except when people start wearing fluo jackets because that contains a large yellow component. A fix was made by switching to a learning technique called naïve bayes. CONCLUSION: it is possible to make a decent result on trial and error and searching for good thresholds. However computer vision has more advance techniques that can learn those relations themselves. Code can be found at USB Stick > TOBCAT > Software > Matlab > segment_trafic_signs.m POST DETECTION CLASSIFICATION BASED ON A NAÏVE BAYERS CLASSIFIER Code of this approach can be found at USB stick > TOBCAT > Software > cpp_windows > naive_bayes_model.cpp USB stick > TOBCAT > Software > cpp_windows > naive_bayes_detect.cpp Basically there are several steps that should be followed 1. Create a positive and negative set that resembles the data you want to separate. POSITIVE SET CONTAINING TRAFFIC SIGNS

NEGATIVE SET CONTAINING PROBLEM CASES 2. Also create a test set that will be used to see if it all works. TEST SET WITH CASES THAT SHOULD PASS 3. Select a set of features that can be used to represent the sample provided. For this we could use knowledge from the previous approach. YMC color space good respondance in Y and M color channel and bad reaction in the C color channel. Selection of regions inside the image that are specific to a traffic sign. See image below. 1/3 of the image above can be ignored horizontally. The rest is divided vertically in a region between 1/3 and 2/3. Inside the region the average Y, M and C channel value is calculated. 4. Push those features into a feature vector for each training image and supply a corresponding label set. Use that to train the classifier. 5. Then use the resulting classifier to decide on the test set. 1 = traffic sign / 0 = something else