Automatic 3D Reconstruction via Object Detection and 3D Transformable Model Matching CS 269 Class Project Report

Size: px
Start display at page:

Download "Automatic 3D Reconstruction via Object Detection and 3D Transformable Model Matching CS 269 Class Project Report"

Transcription

1 Automatic 3D Reconstruction via Object Detection and 3D Transformable Model Matching CS 69 Class Project Report Junhua Mao and Lunbo Xu University of California, Los Angeles and lunbo Abstract In this report, we present a system that can automatically detect and 3D reconstruct the target objects. After the reconstruction, users can edit the image by scaling, translating, rotating and deleting the target objects. The system consists of three parts. The first part is the detection part. We use HOG (Histogram of Gradient) template matching for detection and adopt sliding window strategy with multiple scales and positions. We allow a small deformation of the HOG template. Non-maximum suppression is applied to remove duplicate detected windows. The second part is the 3D transformable model matching part. We will search for the optimal 3D models that match the edge maps of the detected target objects. The third part is the image editing part, the users can rotate, translate and scale multiple target objects detected in the image using simple operations of their mouse. The whole system will reduce the effort the user need to take, especially in the situation that the user need to 3D reconstruct near-duplicate objects in multiple images. 1. Introduction 3D modeling of objects in an image has been an interesting research topic for computer graphic community for several years. One of the limitation of these 3D modeling algorithms is that they generally require the user to input the information of the objects in the image. It will take users lots of time and effort if they want to 3D reconstruct multiple duplicated objects in several images. At the same time, object detection algorithm has also been widely studied in computer vision community. Our motivation of this project is that we can utilize the state-of-the-art algorithms in object detection to find and 3D reconstruct the objects the user want automatically. In this section, we will briefly review the recent progress of object detection and 3D modeling. Then we will introduce the goal of our system, the three component of the system and detailed steps to build the system. Lots of the images features was proposed recently, such as the HOG (Histogram of Gradient) features [] and SIFT (Scale Invariant Feature Transform) features [5]. Both of the features utilize the edge information in the images and use histogram to achieve the invariance of image conditions. There are lots of advanced methods that are based on these two features. Specifically, the Deformable Part Based Model [3] uses several part filters and a root filter to detect object in the images. They extract HOG features with several scales and use them as the image feature for the part filters and root filter. They achieve the state-of-the-art results in several benchmark object detection databases, such as Pascal VOC. Most recently, the Deep Convolutional Neural Network reports the large performance improvement than the transitional computer vision algorithm given millions of training images [4]. Speaking of 3D modeling, traditional interfaces offer tedious mouse-based click and drag. but great advances in usability have been made in recent years since the focus point is shifted toward improving interaction. Sketch-based interfaces for modelling (SBIM) are a promising trend in user interaction. Extracting three dimensional objects from a s- ingle photo is still a great challenge.[6] With the assistance of user input sketch, several impressive works have been done. We list some of them here: Olsen & Samavati s [6] Image-Assisted Modelling system require user sketches object boundary and features on a single image. Shtof et al.[7] reconstruct 3D shapes while simultaneously inferring global mutual geosemantic relations between their parts. And the most impressive one 3-Sweep[1], also need user s help to implicitly decompose the object into simple parts, which are semantically meaningful. None of previous 3D reconstruction work is based on single image and, at the mean time, without user s help. On the basis of all these methods in computer vision and graphics, we want to build a system that can: 1. Detect 1

2 Sample test image HOG scale 1 (c) HOG scale 9 (d) HOG scale 18 Figure 1. Example of the template image. The visualization of HOG template the locations of the objects we want to find;. Match the 3D transformable object model to the detected objects; 3. Allow the users to Edit the image by rotating, scaling and translating the objects. The inputs of the users are a single template image of the object they want to find and its template 3D model. To achieve this goal, our system consists of three parts: the detection part (see Section ), the 3D reconstruction part (see Section 3) and the image editing part (see Section 4). For the detection part, we need to 1. extract image features of the template image;. detect the objects in the testing images; 3. estimate the initial 3D model of the detected objects; 4. remove the detected objects to get the background images. In the 3D reconstruction part, the system will match the transformable model to the edge maps of the detected objects and 3D reconstruct them. Finally, in the image editing part, we design a User Interface that can help the user to edit the image. We will introduce these parts in details. Figure. Example testing image. (c)(d) Visualization of HOG features of several scales.. Feature matching Because the size and position of the objects in the testing images are arbitrary, we proposed a deformable template feature matching method with sliding window strategy to detect the objects in the testing images. Specifically, we extract features from multiple scales (0 scales in total) and use sliding window approach to exhaustively search for each position position with each scales of the objects. We also allow a small deformation of the HOG template. We adopt Normalized Cross-Correlation as the similarity measurement. Figure shows a sample testing image and the visualization of its HOG features of several scales. After the sliding window strategy, we will get lots of candidate object bounding boxes (see Figure 3). But many of them are duplicate bounding boxes and we need to get the most accurate ones. Therefore, Non-Maximum Suppression is applied among the candidate bounding boxes that have IoU (Intersection-Over-Union) ratio larger than 0.5 with each other. The IoU ratio between two candidate. Object Detection Part.1. Image features There are some difference between the transitional object detection system and our system. For transitional object detection system, they typically have lots of training images and they allow the objects to have large visual variance in the images. On the contrast, for our system, we only have one training image (the template image) and we want to find near-duplicated objects in the testing images. Therefore, we choose HOG template as the features. For template image, the HOG feature size is 30 pixels. The feature dimension for one HOG cell is 31. Please see Figure 1 as an example. To further increase the discriminative power of our feature descriptor, we also extract Color Histogram. Specifically, we convert the images to HSV color space. The Hue channel, Saturation channel and Value channel are quantized into 8 bins, 4 bins and bins respectively. This leads to a feature with dimension 64. Figure 3. Candidate object bounding boxes after sliding window search. Detection results after applying Non-Maximum Suppression

3 Original image Background Figure 4. More detection results of our system under different image conditions and complex background. bounding boxes is calculated as the area of the intersection region of the two bounding boxes divided by the area of the union region of the two bounding boxes. We can get the final detection results after Non-Maximum Suppression (see Figure 3). In Figure 4, we show more detection results. Our system can successfully detect near-duplicated objects in complex background and different image conditions. (c) Edge maps and texture Figure 5. Example of the output of the detection part 3.1. Modeling.3. Interface of the first two part We model a cup by 4 faces(outside wall, inside wall, outside bottom and inside bottom) and it is applied with flat shading (Figure 6). The texture of each cup is extracted in the detection part, and it is applied to only the outside wall of the cup. The rest faces of the cup are simply painted white. In order to control a cup model, we need parameters to specify not only its shape and coordinates, but also its orientation. Therefore, we control the model by the following seven parameters (Figure 6): The output of the first part (detection part) will serve as the input of the second part (3D reconstruction part). The output includes: Original testing image Image without target objects (background image). We use Partial Differential Equation based image inprinting to refill the detected bounding boxes regions. Edge map for each detected object Distance map of every pixel to its nearest edge pixel. We use distance transform to calculate the distance map, which has O(W H) time complexity, where W and H are width and height of the testing image. This step is important to accelerate the optimization step in 3D reconstruction part. r0 : the radius of the bottom circle face; r1 : the radius of the top circle face; (x0, y0 ): the centre s coordinate on the drawing plane; < x, y, z >: a vector specify the orientation of the cup, which p also indicates that the height of the cup is h = x + y + z ; Texture of each detected object Estimated initial 3D model parameters (e.g. radius of the bottom and top circle of the cup, the orientation) The example outputs of the first part are shown in Figure 5 (Distance map and 3D model parameters are not presented). 3. 3D Reconstruction Part In this section, we will introduce how we use the output of the detection part to reconstruct the 3-D cup models, project them on the background image with estimated position, and optimize their position. Figure 6. The direction of normals on 4 faces indicates the cup is applied with flat shading. The model is controlled by 7 parameters. 3

4 Figure 9. Representative points are projected on the edge map image to calculate a quadratic energy function. Figure 7. Perspective projection: background image at far plane and cups centres at far plane. 3.. Perspective Projection We use perspective projection (Figure 7) to project cup models on the background image. We draw the background image on the far plane and set the size of far plane equals to the size of the background image. Then we put cups centres at far/ plane so that cups will not be occluded by anything. The position of far plane is determined by the height of the background image H and fovy (i.e. field of view y): Z far = tan(90 o fovy ) H For the near plane, we just simply set Z near := 1. Together with parameter (x 0, y 0 ), the 3-D coordinate of the centre of cup in World Coordinate System is (x 0, y 0, Z far ) Estimation Cup model is controlled by seven parameters. We yield these values by two steps, estimation and optimization. Good estimated initial values will release the burden of optimization no matter what optimization method is used. Recall that for each cup, the output of the detection part includes a bounding box and a template ID. For parameter (x 0, y 0 ) which determine the position of the model, we simply initialize it by the center of the bounding box: x 0 = y 0 = x b + w b W y b + h b H where (x b, y b ) is the coordinate of the lower-left corner of the bounding box, (w b, h b ) is the width and height of the bounding box, and (W, H) is the size of background image. The estimated initial values should all be divided by two, since cup models are drawn at far/ plane. For other parameters r 0,r 1,< x, y, z > which determine the shape and the orientation of the model, we initial them by pre-set values according to the template ID. We believe this is a naive yet effective way to estimate the shape and the orientation, since the object is classified to certain template for reasons and the number of classification template is limit. However, different objects with same template ID will share the same estimated shape and orientation (Figure 8). It is optimization s task that vary different objects Optimization The target of optimization is to make the projected cup fit the original object well. Specifically, we want to select parameters which minimize this quadratic energy function E = (x i, y i ) (x ei, y ei ), where (x i, y i ) are coordinates of projected representative points on the cup model (we select representative points, see Figure 9 for more details) and (x ei, y ei ) is the nearest edge point (Figure 9) of projected point (x i, y i ). And we discuss two algorithm used to minimize this quadratic energy function in the following respectively Naive algorithm We only implement a naive algorithm so far, which is just to enumerate each parameters. Specifically, each parameter has seven offset choices, from 3δ to +3δ. For instance, we try parameter r 0 by r 0 3δ r0, r 0 δ r0, r 0 δ r0, r 0, r 0 + δ r0, r 0 + δ r0, and r 0 + 3δ r0, where r 0 is the estimated initial value and δ r0 is determined by the size of bounding box. For each set of parameters θ = (x 0, y 0, r 0, r 1, x, y, z ), we can calculate E(θ) by E = (x i, y i ) (x ei, y ei ) = dis[x i ][y i ] where dis[x i ][y i ] is the distance between projected point (x i, y i ) to its nearest edge point(x ei, y ei ), which can be precalculated by the detection part. Therefore, the time com- 4

5 Estimated Original (c) Optimized Figure 8. Comparison between original image, result after estimation and result after optimization. plexity of this algorithm is O(W H) for distance map calculation and O(7 7 ( )) for enumeration. The bottleneck is the number of choices for each parameter during enumeration. Comparing with the estimated one (Figure 8), the result of this optimization algorithm (Figure 8(c)) fit the original (Figure 8) better: the left one get bigger, the middle one get smaller, and the right one get skewed a little bit. Figure 10. Image editing results 3.4. EM algorithm We can use Expectation-Maximization algorithm instead of the former naive one to optimize parameters. Gradient Descent is not applicable, since the nearest edge point for each projected point will be updated. However, we can use EM algorithm which applies gradient descent in each iteration: Initialize θ by estimated value (x 0, y 0, r 0, r 1, x, y, z) while(e(θ) > threshold) { (x i, y i ) := (f xi (θ), f yi (θ)), for all i; Find each point s nearest edge point; Formulize E θ ; Apply Gradient Descent; } Although we do not implement it in our system yet, we believe this guided optimization algorithm will lead to a better results. It is one of our future works. 4. Image Editing Part After reconstruction, we can do image editing on the user interface. Image editing in our system simply means changing cup models parameters. User can translate the cup by right click the target cup and dragging it, rotate the cup by left click the target cup, and scaling a target cup by left and right click together. Some sample results are shown in Figure 10. References [1] T. Chen, Z. Zhu, A. Shamir, S.-M. Hu, and D. Cohen-Or. 3- sweep: Extracting editable objects from a single photo. ACM Transactions on Graphics (Proceedings of SIGGRAPH Asia 013), 3(6):Article 195, 013. [] N. Dalal and B. Triggs. Histograms of oriented gradients for human detection. In Computer Vision and Pattern Recognition, 005. CVPR 005. IEEE Computer Society Conference on, volume 1, pages IEEE, 005. [3] P. F. Felzenszwalb, R. B. Girshick, D. McAllester, and D. Ramanan. Object detection with discriminatively trained partbased models. Pattern Analysis and Machine Intelligence, IEEE Transactions on, 3(9): , 010. [4] A. Krizhevsky, I. Sutskever, and G. E. Hinton. Imagenet classification with deep convolutional neural networks. In NIPS, volume 1, page 4, 01. [5] D. G. Lowe. Object recognition from local scale-invariant features. In Computer vision, The proceedings of the seventh IEEE international conference on, volume, pages Ieee, [6] L. Olsen and F. F. Samavati. Image-assisted modeling from sketches. In Proceedings of Graphics Interface 010, GI 10, pages 5 3, Toronto, Ont., Canada, Canada, 010. Canadian Information Processing Society. [7] A. Shtof, A. Agathos, Y. Gingold, A. Shamir, and D. Cohen- Or. Geosemantic snapping for sketch-based modeling. Computer Graphics Forum, 3():45 53, 013. Proceedings of Eurographics

Deformable Part Models with CNN Features

Deformable Part Models with CNN Features Deformable Part Models with CNN Features Pierre-André Savalle 1, Stavros Tsogkas 1,2, George Papandreou 3, Iasonas Kokkinos 1,2 1 Ecole Centrale Paris, 2 INRIA, 3 TTI-Chicago Abstract. In this work we

More information

Image Classification for Dogs and Cats

Image Classification for Dogs and Cats Image Classification for Dogs and Cats Bang Liu, Yan Liu Department of Electrical and Computer Engineering {bang3,yan10}@ualberta.ca Kai Zhou Department of Computing Science kzhou3@ualberta.ca Abstract

More information

Recognizing Cats and Dogs with Shape and Appearance based Models. Group Member: Chu Wang, Landu Jiang

Recognizing Cats and Dogs with Shape and Appearance based Models. Group Member: Chu Wang, Landu Jiang Recognizing Cats and Dogs with Shape and Appearance based Models Group Member: Chu Wang, Landu Jiang Abstract Recognizing cats and dogs from images is a challenging competition raised by Kaggle platform

More information

Lecture 2: The SVM classifier

Lecture 2: The SVM classifier Lecture 2: The SVM classifier C19 Machine Learning Hilary 2015 A. Zisserman Review of linear classifiers Linear separability Perceptron Support Vector Machine (SVM) classifier Wide margin Cost function

More information

Lecture 6: Classification & Localization. boris. ginzburg@intel.com

Lecture 6: Classification & Localization. boris. ginzburg@intel.com Lecture 6: Classification & Localization boris. ginzburg@intel.com 1 Agenda ILSVRC 2014 Overfeat: integrated classification, localization, and detection Classification with Localization Detection. 2 ILSVRC-2014

More information

Convolutional Feature Maps

Convolutional Feature Maps Convolutional Feature Maps Elements of efficient (and accurate) CNN-based object detection Kaiming He Microsoft Research Asia (MSRA) ICCV 2015 Tutorial on Tools for Efficient Object Detection Overview

More information

Local features and matching. Image classification & object localization

Local features and matching. Image classification & object localization Overview Instance level search Local features and matching Efficient visual recognition Image classification & object localization Category recognition Image classification: assigning a class label to

More information

Assessment. Presenter: Yupu Zhang, Guoliang Jin, Tuo Wang Computer Vision 2008 Fall

Assessment. Presenter: Yupu Zhang, Guoliang Jin, Tuo Wang Computer Vision 2008 Fall Automatic Photo Quality Assessment Presenter: Yupu Zhang, Guoliang Jin, Tuo Wang Computer Vision 2008 Fall Estimating i the photorealism of images: Distinguishing i i paintings from photographs h Florin

More information

Drawing a histogram using Excel

Drawing a histogram using Excel Drawing a histogram using Excel STEP 1: Examine the data to decide how many class intervals you need and what the class boundaries should be. (In an assignment you may be told what class boundaries to

More information

MVA ENS Cachan. Lecture 2: Logistic regression & intro to MIL Iasonas Kokkinos Iasonas.kokkinos@ecp.fr

MVA ENS Cachan. Lecture 2: Logistic regression & intro to MIL Iasonas Kokkinos Iasonas.kokkinos@ecp.fr Machine Learning for Computer Vision 1 MVA ENS Cachan Lecture 2: Logistic regression & intro to MIL Iasonas Kokkinos Iasonas.kokkinos@ecp.fr Department of Applied Mathematics Ecole Centrale Paris Galen

More information

siftservice.com - Turning a Computer Vision algorithm into a World Wide Web Service

siftservice.com - Turning a Computer Vision algorithm into a World Wide Web Service siftservice.com - Turning a Computer Vision algorithm into a World Wide Web Service Ahmad Pahlavan Tafti 1, Hamid Hassannia 2, and Zeyun Yu 1 1 Department of Computer Science, University of Wisconsin -Milwaukee,

More information

The Delicate Art of Flower Classification

The Delicate Art of Flower Classification The Delicate Art of Flower Classification Paul Vicol Simon Fraser University University Burnaby, BC pvicol@sfu.ca Note: The following is my contribution to a group project for a graduate machine learning

More information

The Role of Size Normalization on the Recognition Rate of Handwritten Numerals

The Role of Size Normalization on the Recognition Rate of Handwritten Numerals The Role of Size Normalization on the Recognition Rate of Handwritten Numerals Chun Lei He, Ping Zhang, Jianxiong Dong, Ching Y. Suen, Tien D. Bui Centre for Pattern Recognition and Machine Intelligence,

More information

The Visual Internet of Things System Based on Depth Camera

The Visual Internet of Things System Based on Depth Camera The Visual Internet of Things System Based on Depth Camera Xucong Zhang 1, Xiaoyun Wang and Yingmin Jia Abstract The Visual Internet of Things is an important part of information technology. It is proposed

More information

Lecture 6: CNNs for Detection, Tracking, and Segmentation Object Detection

Lecture 6: CNNs for Detection, Tracking, and Segmentation Object Detection CSED703R: Deep Learning for Visual Recognition (206S) Lecture 6: CNNs for Detection, Tracking, and Segmentation Object Detection Bohyung Han Computer Vision Lab. bhhan@postech.ac.kr 2 3 Object detection

More information

3D Distance from a Point to a Triangle

3D Distance from a Point to a Triangle 3D Distance from a Point to a Triangle Mark W. Jones Technical Report CSR-5-95 Department of Computer Science, University of Wales Swansea February 1995 Abstract In this technical report, two different

More information

Part-Based Recognition

Part-Based Recognition Part-Based Recognition Benedict Brown CS597D, Fall 2003 Princeton University CS 597D, Part-Based Recognition p. 1/32 Introduction Many objects are made up of parts It s presumably easier to identify simple

More information

3D Model based Object Class Detection in An Arbitrary View

3D Model based Object Class Detection in An Arbitrary View 3D Model based Object Class Detection in An Arbitrary View Pingkun Yan, Saad M. Khan, Mubarak Shah School of Electrical Engineering and Computer Science University of Central Florida http://www.eecs.ucf.edu/

More information

Pedestrian Detection with RCNN

Pedestrian Detection with RCNN Pedestrian Detection with RCNN Matthew Chen Department of Computer Science Stanford University mcc17@stanford.edu Abstract In this paper we evaluate the effectiveness of using a Region-based Convolutional

More information

Simultaneous Gamma Correction and Registration in the Frequency Domain

Simultaneous Gamma Correction and Registration in the Frequency Domain Simultaneous Gamma Correction and Registration in the Frequency Domain Alexander Wong a28wong@uwaterloo.ca William Bishop wdbishop@uwaterloo.ca Department of Electrical and Computer Engineering University

More information

Canny Edge Detection

Canny Edge Detection Canny Edge Detection 09gr820 March 23, 2009 1 Introduction The purpose of edge detection in general is to significantly reduce the amount of data in an image, while preserving the structural properties

More information

AN IMPROVED DOUBLE CODING LOCAL BINARY PATTERN ALGORITHM FOR FACE RECOGNITION

AN IMPROVED DOUBLE CODING LOCAL BINARY PATTERN ALGORITHM FOR FACE RECOGNITION AN IMPROVED DOUBLE CODING LOCAL BINARY PATTERN ALGORITHM FOR FACE RECOGNITION Saurabh Asija 1, Rakesh Singh 2 1 Research Scholar (Computer Engineering Department), Punjabi University, Patiala. 2 Asst.

More information

Face Recognition in Low-resolution Images by Using Local Zernike Moments

Face Recognition in Low-resolution Images by Using Local Zernike Moments Proceedings of the International Conference on Machine Vision and Machine Learning Prague, Czech Republic, August14-15, 014 Paper No. 15 Face Recognition in Low-resolution Images by Using Local Zernie

More information

Pedestrian Detection using R-CNN

Pedestrian Detection using R-CNN Pedestrian Detection using R-CNN CS676A: Computer Vision Project Report Advisor: Prof. Vinay P. Namboodiri Deepak Kumar Mohit Singh Solanki (12228) (12419) Group-17 April 15, 2016 Abstract Pedestrian detection

More information

CS 1699: Intro to Computer Vision. Deep Learning. Prof. Adriana Kovashka University of Pittsburgh December 1, 2015

CS 1699: Intro to Computer Vision. Deep Learning. Prof. Adriana Kovashka University of Pittsburgh December 1, 2015 CS 1699: Intro to Computer Vision Deep Learning Prof. Adriana Kovashka University of Pittsburgh December 1, 2015 Today: Deep neural networks Background Architectures and basic operations Applications Visualizing

More information

A Learning Based Method for Super-Resolution of Low Resolution Images

A Learning Based Method for Super-Resolution of Low Resolution Images A Learning Based Method for Super-Resolution of Low Resolution Images Emre Ugur June 1, 2004 emre.ugur@ceng.metu.edu.tr Abstract The main objective of this project is the study of a learning based method

More information

Taking Inverse Graphics Seriously

Taking Inverse Graphics Seriously CSC2535: 2013 Advanced Machine Learning Taking Inverse Graphics Seriously Geoffrey Hinton Department of Computer Science University of Toronto The representation used by the neural nets that work best

More information

Scalable Object Detection by Filter Compression with Regularized Sparse Coding

Scalable Object Detection by Filter Compression with Regularized Sparse Coding Scalable Object Detection by Filter Compression with Regularized Sparse Coding Ting-Hsuan Chao, Yen-Liang Lin, Yin-Hsi Kuo, and Winston H Hsu National Taiwan University, Taipei, Taiwan Abstract For practical

More information

Topological Data Analysis Applications to Computer Vision

Topological Data Analysis Applications to Computer Vision Topological Data Analysis Applications to Computer Vision Vitaliy Kurlin, http://kurlin.org Microsoft Research Cambridge and Durham University, UK Topological Data Analysis quantifies topological structures

More information

Classification of Fingerprints. Sarat C. Dass Department of Statistics & Probability

Classification of Fingerprints. Sarat C. Dass Department of Statistics & Probability Classification of Fingerprints Sarat C. Dass Department of Statistics & Probability Fingerprint Classification Fingerprint classification is a coarse level partitioning of a fingerprint database into smaller

More information

Administrivia. Traditional Recognition Approach. Overview. CMPSCI 370: Intro. to Computer Vision Deep learning

Administrivia. Traditional Recognition Approach. Overview. CMPSCI 370: Intro. to Computer Vision Deep learning : Intro. to Computer Vision Deep learning University of Massachusetts, Amherst April 19/21, 2016 Instructor: Subhransu Maji Finals (everyone) Thursday, May 5, 1-3pm, Hasbrouck 113 Final exam Tuesday, May

More information

OBJECT TRACKING USING LOG-POLAR TRANSFORMATION

OBJECT TRACKING USING LOG-POLAR TRANSFORMATION OBJECT TRACKING USING LOG-POLAR TRANSFORMATION A Thesis Submitted to the Gradual Faculty of the Louisiana State University and Agricultural and Mechanical College in partial fulfillment of the requirements

More information

Recognition. Sanja Fidler CSC420: Intro to Image Understanding 1 / 28

Recognition. Sanja Fidler CSC420: Intro to Image Understanding 1 / 28 Recognition Topics that we will try to cover: Indexing for fast retrieval (we still owe this one) History of recognition techniques Object classification Bag-of-words Spatial pyramids Neural Networks Object

More information

Segmentation & Clustering

Segmentation & Clustering EECS 442 Computer vision Segmentation & Clustering Segmentation in human vision K-mean clustering Mean-shift Graph-cut Reading: Chapters 14 [FP] Some slides of this lectures are courtesy of prof F. Li,

More information

Subspace Analysis and Optimization for AAM Based Face Alignment

Subspace Analysis and Optimization for AAM Based Face Alignment Subspace Analysis and Optimization for AAM Based Face Alignment Ming Zhao Chun Chen College of Computer Science Zhejiang University Hangzhou, 310027, P.R.China zhaoming1999@zju.edu.cn Stan Z. Li Microsoft

More information

Linear Threshold Units

Linear Threshold Units Linear Threshold Units w x hx (... w n x n w We assume that each feature x j and each weight w j is a real number (we will relax this later) We will study three different algorithms for learning linear

More information

Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches

Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches PhD Thesis by Payam Birjandi Director: Prof. Mihai Datcu Problematic

More information

A Study on SURF Algorithm and Real-Time Tracking Objects Using Optical Flow

A Study on SURF Algorithm and Real-Time Tracking Objects Using Optical Flow , pp.233-237 http://dx.doi.org/10.14257/astl.2014.51.53 A Study on SURF Algorithm and Real-Time Tracking Objects Using Optical Flow Giwoo Kim 1, Hye-Youn Lim 1 and Dae-Seong Kang 1, 1 Department of electronices

More information

Semantic Recognition: Object Detection and Scene Segmentation

Semantic Recognition: Object Detection and Scene Segmentation Semantic Recognition: Object Detection and Scene Segmentation Xuming He xuming.he@nicta.com.au Computer Vision Research Group NICTA Robotic Vision Summer School 2015 Acknowledgement: Slides from Fei-Fei

More information

Lecture 8 February 4

Lecture 8 February 4 ICS273A: Machine Learning Winter 2008 Lecture 8 February 4 Scribe: Carlos Agell (Student) Lecturer: Deva Ramanan 8.1 Neural Nets 8.1.1 Logistic Regression Recall the logistic function: g(x) = 1 1 + e θt

More information

Adobe Illustrator CS5 Part 1: Introduction to Illustrator

Adobe Illustrator CS5 Part 1: Introduction to Illustrator CALIFORNIA STATE UNIVERSITY, LOS ANGELES INFORMATION TECHNOLOGY SERVICES Adobe Illustrator CS5 Part 1: Introduction to Illustrator Summer 2011, Version 1.0 Table of Contents Introduction...2 Downloading

More information

Latest Advances in Deep Learning. Yao Chou

Latest Advances in Deep Learning. Yao Chou Latest Advances in Deep Learning Yao Chou Outline Introduction Images Classification Object Detection R-CNN Traditional Feature Descriptor Selective Search Implementation Latest Application Deep Learning

More information

Determining optimal window size for texture feature extraction methods

Determining optimal window size for texture feature extraction methods IX Spanish Symposium on Pattern Recognition and Image Analysis, Castellon, Spain, May 2001, vol.2, 237-242, ISBN: 84-8021-351-5. Determining optimal window size for texture feature extraction methods Domènec

More information

Environmental Remote Sensing GEOG 2021

Environmental Remote Sensing GEOG 2021 Environmental Remote Sensing GEOG 2021 Lecture 4 Image classification 2 Purpose categorising data data abstraction / simplification data interpretation mapping for land cover mapping use land cover class

More information

Application of Face Recognition to Person Matching in Trains

Application of Face Recognition to Person Matching in Trains Application of Face Recognition to Person Matching in Trains May 2008 Objective Matching of person Context : in trains Using face recognition and face detection algorithms With a video-surveillance camera

More information

Neural Network based Vehicle Classification for Intelligent Traffic Control

Neural Network based Vehicle Classification for Intelligent Traffic Control Neural Network based Vehicle Classification for Intelligent Traffic Control Saeid Fazli 1, Shahram Mohammadi 2, Morteza Rahmani 3 1,2,3 Electrical Engineering Department, Zanjan University, Zanjan, IRAN

More information

PowerPoint: Graphics and SmartArt

PowerPoint: Graphics and SmartArt PowerPoint: Graphics and SmartArt Contents Inserting Objects... 2 Picture from File... 2 Clip Art... 2 Shapes... 3 SmartArt... 3 WordArt... 3 Formatting Objects... 4 Move a picture, shape, text box, or

More information

Creating an invitation

Creating an invitation Creating an invitation Michaela Maginot About the author Michaela Maginot lives in Unterhaching, Germany, not too far from Munich. She graduated from the Deutschen Meisterschule für Mode (German Master

More information

APPLYING COMPUTER VISION TECHNIQUES TO TOPOGRAPHIC OBJECTS

APPLYING COMPUTER VISION TECHNIQUES TO TOPOGRAPHIC OBJECTS APPLYING COMPUTER VISION TECHNIQUES TO TOPOGRAPHIC OBJECTS Laura Keyes, Adam Winstanley Department of Computer Science National University of Ireland Maynooth Co. Kildare, Ireland lkeyes@cs.may.ie, Adam.Winstanley@may.ie

More information

UNIVERSITY OF OSLO. Faculty of Mathematics and Natural Sciences

UNIVERSITY OF OSLO. Faculty of Mathematics and Natural Sciences UNIVERSITY OF OSLO Faculty of Mathematics and Natural Sciences Exam: INF 4300 Digital image analysis Date: Friday December 11, 2009 Exam hours: 14.30-17.30 Number of pages: 7 pages plus 1 page enclosure

More information

What s New V 11. Preferences: Parameters: Layout/ Modifications: Reverse mouse scroll wheel zoom direction

What s New V 11. Preferences: Parameters: Layout/ Modifications: Reverse mouse scroll wheel zoom direction What s New V 11 Preferences: Reverse mouse scroll wheel zoom direction Assign mouse scroll wheel Middle Button as Fine tune Pricing Method (Manufacturing/Design) Display- Display Long Name Parameters:

More information

Are we ready for Autonomous Driving? The KITTI Vision Benchmark Suite

Are we ready for Autonomous Driving? The KITTI Vision Benchmark Suite Are we ready for Autonomous Driving? The KITTI Vision Benchmark Suite Philip Lenz 1 Andreas Geiger 2 Christoph Stiller 1 Raquel Urtasun 3 1 KARLSRUHE INSTITUTE OF TECHNOLOGY 2 MAX-PLANCK-INSTITUTE IS 3

More information

Segmentation as Selective Search for Object Recognition

Segmentation as Selective Search for Object Recognition Segmentation as Selective Search for Object Recognition Koen E. A. van de Sande Jasper R. R. Uijlings Theo Gevers Arnold W. M. Smeulders University of Amsterdam University of Trento Amsterdam, The Netherlands

More information

Visualization and Feature Extraction, FLOW Spring School 2016 Prof. Dr. Tino Weinkauf. Flow Visualization. Image-Based Methods (integration-based)

Visualization and Feature Extraction, FLOW Spring School 2016 Prof. Dr. Tino Weinkauf. Flow Visualization. Image-Based Methods (integration-based) Visualization and Feature Extraction, FLOW Spring School 2016 Prof. Dr. Tino Weinkauf Flow Visualization Image-Based Methods (integration-based) Spot Noise (Jarke van Wijk, Siggraph 1991) Flow Visualization:

More information

3 hours One paper 70 Marks. Areas of Learning Theory

3 hours One paper 70 Marks. Areas of Learning Theory GRAPHIC DESIGN CODE NO. 071 Class XII DESIGN OF THE QUESTION PAPER 3 hours One paper 70 Marks Section-wise Weightage of the Theory Areas of Learning Theory Section A (Reader) Section B Application of Design

More information

Android Ros Application

Android Ros Application Android Ros Application Advanced Practical course : Sensor-enabled Intelligent Environments 2011/2012 Presentation by: Rim Zahir Supervisor: Dejan Pangercic SIFT Matching Objects Android Camera Topic :

More information

Finding people in repeated shots of the same scene

Finding people in repeated shots of the same scene Finding people in repeated shots of the same scene Josef Sivic 1 C. Lawrence Zitnick Richard Szeliski 1 University of Oxford Microsoft Research Abstract The goal of this work is to find all occurrences

More information

Design document Goal Technology Description

Design document Goal Technology Description Design document Goal OpenOrienteering Mapper is a program to draw orienteering maps. It helps both in the surveying and the following final drawing task. Support for course setting is not a priority because

More information

Pro/ENGINEER Wildfire 4.0 Basic Design

Pro/ENGINEER Wildfire 4.0 Basic Design Introduction Datum features are non-solid features used during the construction of other features. The most common datum features include planes, axes, coordinate systems, and curves. Datum features do

More information

Edge Boxes: Locating Object Proposals from Edges

Edge Boxes: Locating Object Proposals from Edges Edge Boxes: Locating Object Proposals from Edges C. Lawrence Zitnick and Piotr Dollár Microsoft Research Abstract. The use of object proposals is an effective recent approach for increasing the computational

More information

In mathematics, there are four attainment targets: using and applying mathematics; number and algebra; shape, space and measures, and handling data.

In mathematics, there are four attainment targets: using and applying mathematics; number and algebra; shape, space and measures, and handling data. MATHEMATICS: THE LEVEL DESCRIPTIONS In mathematics, there are four attainment targets: using and applying mathematics; number and algebra; shape, space and measures, and handling data. Attainment target

More information

Mean-Shift Tracking with Random Sampling

Mean-Shift Tracking with Random Sampling 1 Mean-Shift Tracking with Random Sampling Alex Po Leung, Shaogang Gong Department of Computer Science Queen Mary, University of London, London, E1 4NS Abstract In this work, boosting the efficiency of

More information

Online Learning for Offroad Robots: Using Spatial Label Propagation to Learn Long Range Traversability

Online Learning for Offroad Robots: Using Spatial Label Propagation to Learn Long Range Traversability Online Learning for Offroad Robots: Using Spatial Label Propagation to Learn Long Range Traversability Raia Hadsell1, Pierre Sermanet1,2, Jan Ben2, Ayse Naz Erkan1, Jefferson Han1, Beat Flepp2, Urs Muller2,

More information

A Simple Feature Extraction Technique of a Pattern By Hopfield Network

A Simple Feature Extraction Technique of a Pattern By Hopfield Network A Simple Feature Extraction Technique of a Pattern By Hopfield Network A.Nag!, S. Biswas *, D. Sarkar *, P.P. Sarkar *, B. Gupta **! Academy of Technology, Hoogly - 722 *USIC, University of Kalyani, Kalyani

More information

Image Segmentation and Registration

Image Segmentation and Registration Image Segmentation and Registration Dr. Christine Tanner (tanner@vision.ee.ethz.ch) Computer Vision Laboratory, ETH Zürich Dr. Verena Kaynig, Machine Learning Laboratory, ETH Zürich Outline Segmentation

More information

Statistical Machine Learning

Statistical Machine Learning Statistical Machine Learning UoC Stats 37700, Winter quarter Lecture 4: classical linear and quadratic discriminants. 1 / 25 Linear separation For two classes in R d : simple idea: separate the classes

More information

Analecta Vol. 8, No. 2 ISSN 2064-7964

Analecta Vol. 8, No. 2 ISSN 2064-7964 EXPERIMENTAL APPLICATIONS OF ARTIFICIAL NEURAL NETWORKS IN ENGINEERING PROCESSING SYSTEM S. Dadvandipour Institute of Information Engineering, University of Miskolc, Egyetemváros, 3515, Miskolc, Hungary,

More information

GIS Tutorial 1. Lecture 2 Map design

GIS Tutorial 1. Lecture 2 Map design GIS Tutorial 1 Lecture 2 Map design Outline Choropleth maps Colors Vector GIS display GIS queries Map layers and scale thresholds Hyperlinks and map tips 2 Lecture 2 CHOROPLETH MAPS Choropleth maps Color-coded

More information

Vehicle Tracking by Simultaneous Detection and Viewpoint Estimation

Vehicle Tracking by Simultaneous Detection and Viewpoint Estimation Vehicle Tracking by Simultaneous Detection and Viewpoint Estimation Ricardo Guerrero-Gómez-Olmedo, Roberto López-Sastre, Saturnino Maldonado-Bascón, and Antonio Fernández-Caballero 2 GRAM, Department of

More information

SHORTCUTS PHOTO PERSONA

SHORTCUTS PHOTO PERSONA SHORTCUTS PHOTO PERSONA esc ± SHOWHIDE TABS F1 F2 F3 F4 F5 F6 F7 F8 F9 F10 F11 F12 1 2 3 4 5 6 7 8 9 0 100% 200% 400% 800% ACTUAL PIXEL QUIT CLOSE MERGE RULERS CHAR OPEN Q W E DOWN R T ACTER Y U I O P

More information

Artificial Neural Networks and Support Vector Machines. CS 486/686: Introduction to Artificial Intelligence

Artificial Neural Networks and Support Vector Machines. CS 486/686: Introduction to Artificial Intelligence Artificial Neural Networks and Support Vector Machines CS 486/686: Introduction to Artificial Intelligence 1 Outline What is a Neural Network? - Perceptron learners - Multi-layer networks What is a Support

More information

Alphacam Art combines Vectric s Aspire artistic design software with the market leading Alphacam manufacturing software.

Alphacam Art combines Vectric s Aspire artistic design software with the market leading Alphacam manufacturing software. Alphacam Art Alphacam Art - CNC Routing For Artists & Ideal Jewellery Cad Cam Software Alphacam Art combines Vectric s Aspire artistic design software with the market leading Alphacam manufacturing software.

More information

Creating a Poster Presentation using PowerPoint

Creating a Poster Presentation using PowerPoint Creating a Poster Presentation using PowerPoint Course Description: This course is designed to assist you in creating eye-catching effective posters for presentation of research findings at scientific

More information

RUN-LENGTH ENCODING FOR VOLUMETRIC TEXTURE

RUN-LENGTH ENCODING FOR VOLUMETRIC TEXTURE RUN-LENGTH ENCODING FOR VOLUMETRIC TEXTURE Dong-Hui Xu, Arati S. Kurani, Jacob D. Furst, Daniela S. Raicu Intelligent Multimedia Processing Laboratory, School of Computer Science, Telecommunications, and

More information

Support Vector Machines with Clustering for Training with Very Large Datasets

Support Vector Machines with Clustering for Training with Very Large Datasets Support Vector Machines with Clustering for Training with Very Large Datasets Theodoros Evgeniou Technology Management INSEAD Bd de Constance, Fontainebleau 77300, France theodoros.evgeniou@insead.fr Massimiliano

More information

Probabilistic Latent Semantic Analysis (plsa)

Probabilistic Latent Semantic Analysis (plsa) Probabilistic Latent Semantic Analysis (plsa) SS 2008 Bayesian Networks Multimedia Computing, Universität Augsburg Rainer.Lienhart@informatik.uni-augsburg.de www.multimedia-computing.{de,org} References

More information

Vision based Vehicle Tracking using a high angle camera

Vision based Vehicle Tracking using a high angle camera Vision based Vehicle Tracking using a high angle camera Raúl Ignacio Ramos García Dule Shu gramos@clemson.edu dshu@clemson.edu Abstract A vehicle tracking and grouping algorithm is presented in this work

More information

Level Set Framework, Signed Distance Function, and Various Tools

Level Set Framework, Signed Distance Function, and Various Tools Level Set Framework Geometry and Calculus Tools Level Set Framework,, and Various Tools Spencer Department of Mathematics Brigham Young University Image Processing Seminar (Week 3), 2010 Level Set Framework

More information

Binary Image Scanning Algorithm for Cane Segmentation

Binary Image Scanning Algorithm for Cane Segmentation Binary Image Scanning Algorithm for Cane Segmentation Ricardo D. C. Marin Department of Computer Science University Of Canterbury Canterbury, Christchurch ricardo.castanedamarin@pg.canterbury.ac.nz Tom

More information

Introduction to Pattern Recognition

Introduction to Pattern Recognition Introduction to Pattern Recognition Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr CS 551, Spring 2009 CS 551, Spring 2009 c 2009, Selim Aksoy (Bilkent University)

More information

Big Data: Image & Video Analytics

Big Data: Image & Video Analytics Big Data: Image & Video Analytics How it could support Archiving & Indexing & Searching Dieter Haas, IBM Deutschland GmbH The Big Data Wave 60% of internet traffic is multimedia content (images and videos)

More information

Robert Collins CSE598G. More on Mean-shift. R.Collins, CSE, PSU CSE598G Spring 2006

Robert Collins CSE598G. More on Mean-shift. R.Collins, CSE, PSU CSE598G Spring 2006 More on Mean-shift R.Collins, CSE, PSU Spring 2006 Recall: Kernel Density Estimation Given a set of data samples x i ; i=1...n Convolve with a kernel function H to generate a smooth function f(x) Equivalent

More information

ENHANCED WEB IMAGE RE-RANKING USING SEMANTIC SIGNATURES

ENHANCED WEB IMAGE RE-RANKING USING SEMANTIC SIGNATURES International Journal of Computer Engineering & Technology (IJCET) Volume 7, Issue 2, March-April 2016, pp. 24 29, Article ID: IJCET_07_02_003 Available online at http://www.iaeme.com/ijcet/issues.asp?jtype=ijcet&vtype=7&itype=2

More information

3 An Illustrative Example

3 An Illustrative Example Objectives An Illustrative Example Objectives - Theory and Examples -2 Problem Statement -2 Perceptron - Two-Input Case -4 Pattern Recognition Example -5 Hamming Network -8 Feedforward Layer -8 Recurrent

More information

What Makes a Great Picture?

What Makes a Great Picture? What Makes a Great Picture? Robert Doisneau, 1955 With many slides from Yan Ke, as annotated by Tamara Berg 15-463: Computational Photography Alexei Efros, CMU, Fall 2011 Photography 101 Composition Framing

More information

Detecting falls and poses in image silhouettes

Detecting falls and poses in image silhouettes Detecting falls and poses in image silhouettes Master s thesis in Complex Adaptive Systems NIKLAS SCHRÄDER Department of Applied Mechanics Division of Vehicle Engineering and Autonomous Systems CHALMERS

More information

Object Recognition. Selim Aksoy. Bilkent University saksoy@cs.bilkent.edu.tr

Object Recognition. Selim Aksoy. Bilkent University saksoy@cs.bilkent.edu.tr Image Classification and Object Recognition Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr Image classification Image (scene) classification is a fundamental

More information

Using Microsoft Picture Manager

Using Microsoft Picture Manager Using Microsoft Picture Manager Storing Your Photos It is suggested that a county store all photos for use in the County CMS program in the same folder for easy access. For the County CMS Web Project it

More information

Segmentation of building models from dense 3D point-clouds

Segmentation of building models from dense 3D point-clouds Segmentation of building models from dense 3D point-clouds Joachim Bauer, Konrad Karner, Konrad Schindler, Andreas Klaus, Christopher Zach VRVis Research Center for Virtual Reality and Visualization, Institute

More information

G E N E R A L A P P R O A CH: LO O K I N G F O R D O M I N A N T O R I E N T A T I O N I N I M A G E P A T C H E S

G E N E R A L A P P R O A CH: LO O K I N G F O R D O M I N A N T O R I E N T A T I O N I N I M A G E P A T C H E S G E N E R A L A P P R O A CH: LO O K I N G F O R D O M I N A N T O R I E N T A T I O N I N I M A G E P A T C H E S In object categorization applications one of the main problems is that objects can appear

More information

An Energy-Based Vehicle Tracking System using Principal Component Analysis and Unsupervised ART Network

An Energy-Based Vehicle Tracking System using Principal Component Analysis and Unsupervised ART Network Proceedings of the 8th WSEAS Int. Conf. on ARTIFICIAL INTELLIGENCE, KNOWLEDGE ENGINEERING & DATA BASES (AIKED '9) ISSN: 179-519 435 ISBN: 978-96-474-51-2 An Energy-Based Vehicle Tracking System using Principal

More information

Enhanced LIC Pencil Filter

Enhanced LIC Pencil Filter Enhanced LIC Pencil Filter Shigefumi Yamamoto, Xiaoyang Mao, Kenji Tanii, Atsumi Imamiya University of Yamanashi {daisy@media.yamanashi.ac.jp, mao@media.yamanashi.ac.jp, imamiya@media.yamanashi.ac.jp}

More information

Fast Matching of Binary Features

Fast Matching of Binary Features Fast Matching of Binary Features Marius Muja and David G. Lowe Laboratory for Computational Intelligence University of British Columbia, Vancouver, Canada {mariusm,lowe}@cs.ubc.ca Abstract There has been

More information

Footwear Print Retrieval System for Real Crime Scene Marks

Footwear Print Retrieval System for Real Crime Scene Marks Footwear Print Retrieval System for Real Crime Scene Marks Yi Tang, Sargur N. Srihari, Harish Kasiviswanathan and Jason J. Corso Center of Excellence for Document Analysis and Recognition (CEDAR) University

More information

The Scientific Data Mining Process

The Scientific Data Mining Process Chapter 4 The Scientific Data Mining Process When I use a word, Humpty Dumpty said, in rather a scornful tone, it means just what I choose it to mean neither more nor less. Lewis Carroll [87, p. 214] In

More information

An Experimental Comparison of Online Object Tracking Algorithms

An Experimental Comparison of Online Object Tracking Algorithms An Experimental Comparison of Online Object Tracking Algorithms Qing Wang a, Feng Chen a, Wenli Xu a, and Ming-Hsuan Yang b a Tsinghua University, Beijing, China b University of California at Merced, Calfironia,

More information

Template-based Eye and Mouth Detection for 3D Video Conferencing

Template-based Eye and Mouth Detection for 3D Video Conferencing Template-based Eye and Mouth Detection for 3D Video Conferencing Jürgen Rurainsky and Peter Eisert Fraunhofer Institute for Telecommunications - Heinrich-Hertz-Institute, Image Processing Department, Einsteinufer

More information

Circle Object Recognition Based on Monocular Vision for Home Security Robot

Circle Object Recognition Based on Monocular Vision for Home Security Robot Journal of Applied Science and Engineering, Vol. 16, No. 3, pp. 261 268 (2013) DOI: 10.6180/jase.2013.16.3.05 Circle Object Recognition Based on Monocular Vision for Home Security Robot Shih-An Li, Ching-Chang

More information

ROBOTRACKER A SYSTEM FOR TRACKING MULTIPLE ROBOTS IN REAL TIME. by Alex Sirota, alex@elbrus.com

ROBOTRACKER A SYSTEM FOR TRACKING MULTIPLE ROBOTS IN REAL TIME. by Alex Sirota, alex@elbrus.com ROBOTRACKER A SYSTEM FOR TRACKING MULTIPLE ROBOTS IN REAL TIME by Alex Sirota, alex@elbrus.com Project in intelligent systems Computer Science Department Technion Israel Institute of Technology Under the

More information

Human Pose Estimation from RGB Input Using Synthetic Training Data

Human Pose Estimation from RGB Input Using Synthetic Training Data Human Pose Estimation from RGB Input Using Synthetic Training Data Oscar Danielsson and Omid Aghazadeh School of Computer Science and Communication KTH, Stockholm, Sweden {osda02, omida}@kth.se arxiv:1405.1213v2

More information