Convolutional Feature Maps
|
|
|
- Asher Fitzgerald
- 9 years ago
- Views:
Transcription
1 Convolutional Feature Maps Elements of efficient (and accurate) CNN-based object detection Kaiming He Microsoft Research Asia (MSRA) ICCV 2015 Tutorial on Tools for Efficient Object Detection
2 Overview of this section Quick introduction to convolutional feature maps Intuitions: into the black boxes How object detection networks & region proposal networks are designed Bridging the gap between hand-engineered and deep learning systems Focusing on forward propagation (inference) Backward propagation (training) covered by Ross s section
3 Object Detection = What, and Where Localization Where? person : Recognition What? car : horse : dog : person : We need a building block that tells us what and where
4 Object Detection = What, and Where Convolutional: sliding-window operations Feature: encoding what (and implicitly encoding where ) Map: explicitly encoding where
5 Convolutional Layers Convolutional layers are locally connected a filter/kernel/window slides on the image or the previous map the position of the filter explicitly provides information for localizing local spatial information w.r.t. the window is encoded in the channels
6 Convolutional Layers Convolutional layers share weights spatially: translation-invariant Translation-invariant: a translated region will produce the same response at the correspondingly translated position A local pattern s convolutional response can be re-used by different candidate regions
7 Convolutional Layers Convolutional layers can be applied to images of any sizes, yielding proportionally-sized outputs W H W H W 2 H 2 W 4 H 4
8 HOG by Convolutional Layers Steps of computing HOG: Computing image gradients Binning gradients into 18 directions Computing cell histograms Normalizing cell histograms Convolutional perspectives: Horizontal/vertical edge filters Directional filters + gating (non-linearity) Sum/average pooling Local response normalization (LRN) see [Mahendran & Vedaldi, CVPR 2015] HOG, dense SIFT, and many other hand-engineered features are convolutional feature maps. Aravindh Mahendran & Andrea Vedaldi. Understanding Deep Image Representations by Inverting Them. CVPR 2015
9 Feature Maps = features and their locations Convolutional: sliding-window operations Feature: encoding what (and implicitly encoding where ) Map: explicitly encoding where
10 Feature Maps = features and their locations ImageNet images with strongest responses of this channel one feature map of conv 5 (#55 in 256 channels of a model trained on ImageNet) Intuition of this response: There is a circle-shaped object (likely a tire) at this position. What Where Kaiming He, Xiangyu Zhang, Shaoqing Ren, & Jian Sun. Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition. ECCV 2014.
11 Feature Maps = features and their locations ImageNet images with strongest responses of this channel one feature map of conv 5 (#66 in 256 channels of a model trained on ImageNet) Intuition of this response: There is a λ-shaped object (likely an underarm) at this position. What Where Kaiming He, Xiangyu Zhang, Shaoqing Ren, & Jian Sun. Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition. ECCV 2014.
12 Feature Maps = features and their locations Visualizing one response (by Zeiler and Fergus)? image a feature map keep one response (e.g., the strongest) Matthew D. Zeiler & Rob Fergus. Visualizing and Understanding Convolutional Networks. ECCV 2014.
13 Feature Maps = features and their locations Visualizing one response conv3 image credit: Zeiler & Fergus Matthew D. Zeiler & Rob Fergus. Visualizing and Understanding Convolutional Networks. ECCV 2014.
14 Feature Maps = features and their locations Visualizing one response Intuition of this visualization: There is a dog-head shape at this position. Location of a feature: explicitly represents where it is. Responses of a feature: encode what it is, and implicitly encode finer position information finer position information is encoded in the channel dimensions (e.g., bbox regression from responses at one pixel as in RPN) conv5 image credit: Zeiler & Fergus Matthew D. Zeiler & Rob Fergus. Visualizing and Understanding Convolutional Networks. ECCV 2014.
15 Receptive Field Receptive field of the first layer is the filter size Receptive field (w.r.t. input image) of a deeper layer depends on all previous layers filter size and strides Correspondence between a feature map pixel and an image pixel is not unique Map a feature map pixel to the center of the receptive field on the image in the SPP-net paper Kaiming He, Xiangyu Zhang, Shaoqing Ren, & Jian Sun. Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition. ECCV 2014.
16 Receptive Field How to compute the center of the receptive field A simple solution For each layer, pad F/2 pixels for a filter size F (e.g., pad 1 pixel for a filter size of 3) On each feature map, the response at (0, 0) has a receptive field centered at (0, 0) on the image On each feature map, the response at (x, y) has a receptive field centered at (Sx, Sy) on the image (stride S) A general solution See [Karel Lenc & Andrea Vedaldi] R-CNN minus R. BMVC Kaiming He, Xiangyu Zhang, Shaoqing Ren, & Jian Sun. Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition. ECCV 2014.
17 Region-based CNN Features warped region CNN aeroplane? no... person? yes... tvmonitor? no. figure credit: R. Girshick et al. input image region proposals ~2,000 1 CNN for each region classify regions R-CNN pipeline R. Girshick, J. Donahue, T. Darrell, & J. Malik. Rich feature hierarchies for accurate object detection and semantic segmentation. CVPR 2014.
18 Region-based CNN Features Given proposal regions, what we need is a feature for each region R-CNN: cropping an image region + CNN on region, requires 2000 CNN computations What about cropping feature map regions?
19 Regions on Feature Maps image region feature map region Compute convolutional feature maps on the entire image only once. Project an image region to a feature map region (using correspondence of the receptive field center) Extract a region-based feature from the feature map region Kaiming He, Xiangyu Zhang, Shaoqing Ren, & Jian Sun. Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition. ECCV 2014.
20 Regions on Feature Maps? warp image region feature map region Fixed-length features are required by fully-connected layers or SVM But how to produce a fixed-length feature from a feature map region? Solutions in traditional compute vision: Bag-of-words, SPM Kaiming He, Xiangyu Zhang, Shaoqing Ren, & Jian Sun. Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition. ECCV 2014.
21 Bag-of-words & Spatial Pyramid Matching SIFT/HOG-based feature maps + level level level pooling pooling pooling Bag-of-words [J. Sivic & A. Zisserman, ICCV 2003] Spatial Pyramid Matching (SPM) [K. Grauman & T. Darrell, ICCV 2005] [S. Lazebnik et al, CVPR 2006] figure credit: S. Lazebnik et al. Kaiming He, Xiangyu Zhang, Shaoqing Ren, & Jian Sun. Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition. ECCV 2014.
22 Spatial Pyramid Pooling (SPP) Layer fix the number of bins (instead of filter sizes) adaptively-sized bins a finer level maintains explicit spatial information pooling concatenate, fc layers a coarser level removes explicit spatial information (bag-of-features) Kaiming He, Xiangyu Zhang, Shaoqing Ren, & Jian Sun. Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition. ECCV 2014.
23 Spatial Pyramid Pooling (SPP) Layer Pre-trained nets often have a single-resolution pooling layer (7x7 for VGG nets) To adapt to a pre-trained net, a single-level pyramid is useable Region-of-Interest (RoI) pooling [R. Girshick, ICCV 2015] pooling concatenate, fc layers Kaiming He, Xiangyu Zhang, Shaoqing Ren, & Jian Sun. Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition. ECCV 2014.
24 Single-scale and Multi-scale Feature Maps Feature Pyramid Resize the input image to multiple scales Compute feature maps for each scale Used for HOG/SIFT features and convolutional features (OverFeat [Sermanet et al. 2013]) feature pyramid image pyramid Kaiming He, Xiangyu Zhang, Shaoqing Ren, & Jian Sun. Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition. ECCV 2014.
25 Single-scale and Multi-scale Feature Maps But deep convolutional feature maps perform well at a single scale SPP-net 1-scale SPP-net 5-scale pool fc fine-tuned fc fine-tuned fc fine-tuned fc 7 bbox reg conv time 0.053s 0.293s fc time 0.089s 0.089s total time 0.142s 0.382s detection map on PASCAL VOC 2007, with ZF-net pre-trained on ImageNet this table is from [K. He, et al. 2014] Also observed in Fast R-CNN and VGG nets Good speed-vs-accuracy tradeoff Learn to be scale-invariant from pretraining data (ImageNet) (note: but if good accuracy is desired, feature pyramids are still needed) Kaiming He, Xiangyu Zhang, Shaoqing Ren, & Jian Sun. Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition. ECCV 2014.
26 R-CNN vs. Fast R-CNN (forward pipeline) feature feature feature feature feature feature feature SPP/RoI pooling CNN CNN CNN CNN CNN image image R-CNN Extract image regions 1 CNN per region (2000 CNNs) Classify region-based features SPP-net & Fast R-CNN (the same forward pipeline) 1 CNN on the entire image Extract features from feature map regions Classify region-based features Ross Girshick. Fast R-CNN. ICCV Kaiming He, Xiangyu Zhang, Shaoqing Ren, & Jian Sun. Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition. ECCV 2014.
27 R-CNN vs. Fast R-CNN (forward pipeline) feature feature feature feature feature feature feature SPP/RoI pooling CNN CNN CNN CNN CNN image image R-CNN Complexity: ~ SPP-net & Fast R-CNN (the same forward pipeline) Complexity: ~ ~160x faster than R-CNN Ross Girshick. Fast R-CNN. ICCV Kaiming He, Xiangyu Zhang, Shaoqing Ren, & Jian Sun. Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition. ECCV 2014.
28 Region Proposal from Feature Maps Object detection networks are fast (0.2s) but what about region proposal? Selective Search [Uijlings et al. ICCV 2011]: 2s per image EdgeBoxes [Zitnick & Dollar. ECCV 2014]: 0.2s per image Can we do region proposal on the same set of feature maps? Shaoqing Ren, Kaiming He, Ross Girshick, & Jian Sun. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. NIPS 2015.
29 Feature Maps = features and their locations Convolutional: sliding-window operations Feature: encoding what (and implicitly encoding where ) Map: explicitly encoding where
30 Region Proposal from Feature Maps By decoding one response at a single pixel, we can still roughly see the object outline * Revisiting visualizations from Zeiler & Fergus Finer localization information has been encoded in the channels of a convolutional feature response Extract this information for better localization * Zeiler & Fergus s method traces unpooling information so the visualization involves more than a single response. But other visualization methods reveal similar patterns. Shaoqing Ren, Kaiming He, Ross Girshick, & Jian Sun. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. NIPS 2015.
31 Region Proposal from Feature Maps a feature vector (e.g., 256-d) The channels of this feature vector encodes finer localization information image feature map The spatial position of this feature provides coarse locations Shaoqing Ren, Kaiming He, Ross Girshick, & Jian Sun. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. NIPS 2015.
32 Region Proposal Network Slide a small window on the feature map Build a small network for: classifying object or not-object, and regressing bbox locations classify obj./not-obj. regress box locations Position of the sliding window provides localization information with reference to the image Box regression provides finer localization information with reference to this sliding window sliding window convolutional feature map Shaoqing Ren, Kaiming He, Ross Girshick, & Jian Sun. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. NIPS 2015.
33 Anchors as references Anchors: pre-defined reference boxes Box regression is with reference to anchors: regressing an anchor box to a ground-truth box n scores 4n coordinates n anchors Object probability is with reference to anchors, e.g.: anchors as positive samples: if IoU > 0.7 or IoU is max anchors as negative samples: if IoU < d Shaoqing Ren, Kaiming He, Ross Girshick, & Jian Sun. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. NIPS 2015.
34 Anchors as references Anchors: pre-defined reference boxes n scores 4n coordinates n anchors Translation-invariant anchors: the same set of anchors are used at each sliding position the same prediction functions (with reference to the sliding window) are used a translated object will have a translated prediction 256-d Shaoqing Ren, Kaiming He, Ross Girshick, & Jian Sun. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. NIPS 2015.
35 Anchors as references Anchors: pre-defined reference boxes n scores 4n coordinates n anchors Multi-scale/size anchors: multiple anchors are used at each position: e.g., 3 scales (128 2, 256 2, ) and 3 aspect ratios (2:1, 1:1, 1:2) yield 9 anchors each anchor has its own prediction function single-scale features, multi-scale predictions 256-d Shaoqing Ren, Kaiming He, Ross Girshick, & Jian Sun. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. NIPS 2015.
36 Anchors as references Comparisons of multi-scale strategies feature image Image/Feature Pyramid Filter Pyramid Anchor Pyramid Shaoqing Ren, Kaiming He, Ross Girshick, & Jian Sun. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. NIPS 2015.
37 Region Proposal Network RPN is fully convolutional [Long et al. 2015] n scores 4n coordinates n anchors RPN is trained end-to-end 256-d RPN shares convolutional feature maps with the detection network (covered in Ross s section) Shaoqing Ren, Kaiming He, Ross Girshick, & Jian Sun. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. NIPS 2015.
38 detector Faster R-CNN RoI pooling proposals RPN feature map system time 07 data data R-CNN ~50s Fast R-CNN ~2s Faster R-CNN 198ms detection map on PASCAL VOC 2007, with VGG-16 pre-trained on ImageNet CNN image Shaoqing Ren, Kaiming He, Ross Girshick, & Jian Sun. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. NIPS 2015.
39 Example detection results of Faster R-CNN car : person : horse : person : dog : cat : dog : bus: boat : person : person : person : person : person : Shaoqing Ren, Kaiming He, Ross Girshick, & Jian Sun. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. NIPS 2015.
40 Keys to efficient CNN-based object detection Feature sharing R-CNN => SPP-net & Fast R-CNN: sharing features among proposal regions Fast R-CNN => Faster R-CNN: sharing features between proposal and detection All are done by shared convolutional feature maps Efficient multi-scale solutions Single-scale convolutional feature maps are good trade-offs Multi-scale anchors are fast and flexible
41 Conclusion of this section Quick introduction to convolutional feature maps Intuitions: into the black boxes How object detection networks & region proposal networks are designed Bridging the gap between hand-engineered and deep learning systems Focusing on forward propagation (inference) Backward propagation (training) covered by Ross s section
Lecture 6: CNNs for Detection, Tracking, and Segmentation Object Detection
CSED703R: Deep Learning for Visual Recognition (206S) Lecture 6: CNNs for Detection, Tracking, and Segmentation Object Detection Bohyung Han Computer Vision Lab. [email protected] 2 3 Object detection
Module 5. Deep Convnets for Local Recognition Joost van de Weijer 4 April 2016
Module 5 Deep Convnets for Local Recognition Joost van de Weijer 4 April 2016 Previously, end-to-end.. Dog Slide credit: Jose M 2 Previously, end-to-end.. Dog Learned Representation Slide credit: Jose
Fast R-CNN. Author: Ross Girshick Speaker: Charlie Liu Date: Oct, 13 th. Girshick, R. (2015). Fast R-CNN. arxiv preprint arxiv:1504.08083.
Fast R-CNN Author: Ross Girshick Speaker: Charlie Liu Date: Oct, 13 th Girshick, R. (2015). Fast R-CNN. arxiv preprint arxiv:1504.08083. ECS 289G 001 Paper Presentation, Prof. Lee Result 1 67% Accuracy
Fast R-CNN Object detection with Caffe
Fast R-CNN Object detection with Caffe Ross Girshick Microsoft Research arxiv code Latest roasts Goals for this section Super quick intro to object detection Show one way to tackle obj. det. with ConvNets
Lecture 6: Classification & Localization. boris. [email protected]
Lecture 6: Classification & Localization boris. [email protected] 1 Agenda ILSVRC 2014 Overfeat: integrated classification, localization, and detection Classification with Localization Detection. 2 ILSVRC-2014
Bert Huang Department of Computer Science Virginia Tech
This paper was submitted as a final project report for CS6424/ECE6424 Probabilistic Graphical Models and Structured Prediction in the spring semester of 2016. The work presented here is done by students
Image and Video Understanding
Image and Video Understanding 2VO 710.095 WS Christoph Feichtenhofer, Axel Pinz Slide credits: Many thanks to all the great computer vision researchers on which this presentation relies on. Most material
Deformable Part Models with CNN Features
Deformable Part Models with CNN Features Pierre-André Savalle 1, Stavros Tsogkas 1,2, George Papandreou 3, Iasonas Kokkinos 1,2 1 Ecole Centrale Paris, 2 INRIA, 3 TTI-Chicago Abstract. In this work we
Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks
1 Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks Shaoqing Ren, Kaiming He, Ross Girshick, and Jian Sun arxiv:1506.01497v3 [cs.cv] 6 Jan 2016 Abstract State-of-the-art object
CS 1699: Intro to Computer Vision. Deep Learning. Prof. Adriana Kovashka University of Pittsburgh December 1, 2015
CS 1699: Intro to Computer Vision Deep Learning Prof. Adriana Kovashka University of Pittsburgh December 1, 2015 Today: Deep neural networks Background Architectures and basic operations Applications Visualizing
Pedestrian Detection with RCNN
Pedestrian Detection with RCNN Matthew Chen Department of Computer Science Stanford University [email protected] Abstract In this paper we evaluate the effectiveness of using a Region-based Convolutional
Compacting ConvNets for end to end Learning
Compacting ConvNets for end to end Learning Jose M. Alvarez Joint work with Lars Pertersson, Hao Zhou, Fatih Porikli. Success of CNN Image Classification Alex Krizhevsky, Ilya Sutskever, Geoffrey E. Hinton,
R-CNN minus R. 1 Introduction. Karel Lenc http://www.robots.ox.ac.uk/~karel. Department of Engineering Science, University of Oxford, Oxford, UK.
LENC, VEDALDI: R-CNN MINUS R 1 R-CNN minus R Karel Lenc http://www.robots.ox.ac.uk/~karel Andrea Vedaldi http://www.robots.ox.ac.uk/~vedaldi Department of Engineering Science, University of Oxford, Oxford,
CAP 6412 Advanced Computer Vision
CAP 6412 Advanced Computer Vision http://www.cs.ucf.edu/~bgong/cap6412.html Boqing Gong Jan 26, 2016 Today Administrivia A bigger picture and some common questions Object detection proposals, by Samer
Pedestrian Detection using R-CNN
Pedestrian Detection using R-CNN CS676A: Computer Vision Project Report Advisor: Prof. Vinay P. Namboodiri Deepak Kumar Mohit Singh Solanki (12228) (12419) Group-17 April 15, 2016 Abstract Pedestrian detection
arxiv:1604.08893v1 [cs.cv] 29 Apr 2016
Faster R-CNN Features for Instance Search Amaia Salvador, Xavier Giró-i-Nieto, Ferran Marqués Universitat Politècnica de Catalunya (UPC) Barcelona, Spain {amaia.salvador,xavier.giro}@upc.edu Shin ichi
Deep Residual Networks
Deep Residual Networks Deep Learning Gets Way Deeper 8:30-10:30am, June 19 ICML 2016 tutorial Kaiming He Facebook AI Research* *as of July 2016. Formerly affiliated with Microsoft Research Asia 7x7 conv,
Image Classification for Dogs and Cats
Image Classification for Dogs and Cats Bang Liu, Yan Liu Department of Electrical and Computer Engineering {bang3,yan10}@ualberta.ca Kai Zhou Department of Computing Science [email protected] Abstract
Learning and transferring mid-level image representions using convolutional neural networks
Willow project-team Learning and transferring mid-level image representions using convolutional neural networks Maxime Oquab, Léon Bottou, Ivan Laptev, Josef Sivic 1 Image classification (easy) Is there
Semantic Recognition: Object Detection and Scene Segmentation
Semantic Recognition: Object Detection and Scene Segmentation Xuming He [email protected] Computer Vision Research Group NICTA Robotic Vision Summer School 2015 Acknowledgement: Slides from Fei-Fei
Latest Advances in Deep Learning. Yao Chou
Latest Advances in Deep Learning Yao Chou Outline Introduction Images Classification Object Detection R-CNN Traditional Feature Descriptor Selective Search Implementation Latest Application Deep Learning
Administrivia. Traditional Recognition Approach. Overview. CMPSCI 370: Intro. to Computer Vision Deep learning
: Intro. to Computer Vision Deep learning University of Massachusetts, Amherst April 19/21, 2016 Instructor: Subhransu Maji Finals (everyone) Thursday, May 5, 1-3pm, Hasbrouck 113 Final exam Tuesday, May
Automatic 3D Reconstruction via Object Detection and 3D Transformable Model Matching CS 269 Class Project Report
Automatic 3D Reconstruction via Object Detection and 3D Transformable Model Matching CS 69 Class Project Report Junhua Mao and Lunbo Xu University of California, Los Angeles [email protected] and lunbo
MulticoreWare. Global Company, 250+ employees HQ = Sunnyvale, CA Other locations: US, China, India, Taiwan
1 MulticoreWare Global Company, 250+ employees HQ = Sunnyvale, CA Other locations: US, China, India, Taiwan Focused on Heterogeneous Computing Multiple verticals spawned from core competency Machine Learning
arxiv:1504.08083v2 [cs.cv] 27 Sep 2015
Fast R-CNN Ross Girshick Microsoft Research [email protected] arxiv:1504.08083v2 [cs.cv] 27 Sep 2015 Abstract This paper proposes a Fast Region-based Convolutional Network method (Fast R-CNN) for object
Object Detection in Video using Faster R-CNN
Object Detection in Video using Faster R-CNN Prajit Ramachandran University of Illinois at Urbana-Champaign [email protected] Abstract Convolutional neural networks (CNN) currently dominate the computer
Local features and matching. Image classification & object localization
Overview Instance level search Local features and matching Efficient visual recognition Image classification & object localization Category recognition Image classification: assigning a class label to
Recognition. Sanja Fidler CSC420: Intro to Image Understanding 1 / 28
Recognition Topics that we will try to cover: Indexing for fast retrieval (we still owe this one) History of recognition techniques Object classification Bag-of-words Spatial pyramids Neural Networks Object
Do Convnets Learn Correspondence?
Do Convnets Learn Correspondence? Jonathan Long Ning Zhang Trevor Darrell University of California Berkeley {jonlong, nzhang, trevor}@cs.berkeley.edu Abstract Convolutional neural nets (convnets) trained
arxiv:1409.1556v6 [cs.cv] 10 Apr 2015
VERY DEEP CONVOLUTIONAL NETWORKS FOR LARGE-SCALE IMAGE RECOGNITION Karen Simonyan & Andrew Zisserman + Visual Geometry Group, Department of Engineering Science, University of Oxford {karen,az}@robots.ox.ac.uk
Segmentation as Selective Search for Object Recognition
Segmentation as Selective Search for Object Recognition Koen E. A. van de Sande Jasper R. R. Uijlings Theo Gevers Arnold W. M. Smeulders University of Amsterdam University of Trento Amsterdam, The Netherlands
SSD: Single Shot MultiBox Detector
SSD: Single Shot MultiBox Detector Wei Liu 1, Dragomir Anguelov 2, Dumitru Erhan 3, Christian Szegedy 3, Scott Reed 4, Cheng-Yang Fu 1, Alexander C. Berg 1 1 UNC Chapel Hill 2 Zoox Inc. 3 Google Inc. 4
Steven C.H. Hoi School of Information Systems Singapore Management University Email: [email protected]
Steven C.H. Hoi School of Information Systems Singapore Management University Email: [email protected] Introduction http://stevenhoi.org/ Finance Recommender Systems Cyber Security Machine Learning Visual
Multi-view Face Detection Using Deep Convolutional Neural Networks
Multi-view Face Detection Using Deep Convolutional Neural Networks Sachin Sudhakar Farfade Yahoo [email protected] Mohammad Saberian Yahoo [email protected] Li-Jia Li Yahoo [email protected]
arxiv:1502.01852v1 [cs.cv] 6 Feb 2015
Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification Kaiming He Xiangyu Zhang Shaoqing Ren Jian Sun arxiv:1502.01852v1 [cs.cv] 6 Feb 2015 Abstract Rectified activation
Exploit All the Layers: Fast and Accurate CNN Object Detector with Scale Dependent Pooling and Cascaded Rejection Classifiers
Exploit All the Layers: Fast and Accurate CNN Object Detector with Scale Dependent Pooling and Cascaded Rejection Classifiers Fan Yang 1,2, Wongun Choi 2, and Yuanqing Lin 2 1 Department of Computer Science,
Recognizing Cats and Dogs with Shape and Appearance based Models. Group Member: Chu Wang, Landu Jiang
Recognizing Cats and Dogs with Shape and Appearance based Models Group Member: Chu Wang, Landu Jiang Abstract Recognizing cats and dogs from images is a challenging competition raised by Kaggle platform
Return of the Devil in the Details: Delving Deep into Convolutional Nets
CHATFIELD ET AL.: RETURN OF THE DEVIL 1 Return of the Devil in the Details: Delving Deep into Convolutional Nets Ken Chatfield [email protected] Karen Simonyan [email protected] Andrea Vedaldi [email protected]
Task-driven Progressive Part Localization for Fine-grained Recognition
Task-driven Progressive Part Localization for Fine-grained Recognition Chen Huang Zhihai He [email protected] University of Missouri [email protected] Abstract In this paper we propose a task-driven
arxiv:1511.02300v2 [cs.cv] 9 Mar 2016
Deep Sliding Shapes for Amodal 3D Object Detection in RGB-D Images Shuran Song Jianxiong Xiao Princeton University http://dss.cs.princeton.edu arxiv:1511.02300v2 [cs.cv] 9 Mar 2016 Abstract We focus on
Edge Boxes: Locating Object Proposals from Edges
Edge Boxes: Locating Object Proposals from Edges C. Lawrence Zitnick and Piotr Dollár Microsoft Research Abstract. The use of object proposals is an effective recent approach for increasing the computational
RECOGNIZING objects and localizing them in images is
1 Region-based Convolutional Networks for Accurate Object Detection and Segmentation Ross Girshick, Jeff Donahue, Student Member, IEEE, Trevor Darrell, Member, IEEE, and Jitendra Malik, Fellow, IEEE Abstract
HE Shuncheng [email protected]. March 20, 2016
Department of Automation Association of Science and Technology of Automation March 20, 2016 Contents Binary Figure 1: a cat? Figure 2: a dog? Binary : Given input data x (e.g. a picture), the output of
Semantic Image Segmentation and Web-Supervised Visual Learning
Semantic Image Segmentation and Web-Supervised Visual Learning Florian Schroff Andrew Zisserman University of Oxford, UK Antonio Criminisi Microsoft Research Ltd, Cambridge, UK Outline Part I: Semantic
Large Scale Semi-supervised Object Detection using Visual and Semantic Knowledge Transfer
Large Scale Semi-supervised Object Detection using Visual and Semantic Knowledge Transfer Yuxing Tang 1 Josiah Wang 2 Boyang Gao 1,3 Emmanuel Dellandréa 1 Robert Gaizauskas 2 Liming Chen 1 1 Ecole Centrale
Understanding Deep Image Representations by Inverting Them
Understanding Deep Image Representations by Inverting Them Aravindh Mahendran University of Oxford [email protected] Andrea Vedaldi University of Oxford [email protected] Abstract Image representations,
SEMANTIC CONTEXT AND DEPTH-AWARE OBJECT PROPOSAL GENERATION
SEMANTIC TEXT AND DEPTH-AWARE OBJECT PROPOSAL GENERATION Haoyang Zhang,, Xuming He,, Fatih Porikli,, Laurent Kneip NICTA, Canberra; Australian National University, Canberra ABSTRACT This paper presents
The Delicate Art of Flower Classification
The Delicate Art of Flower Classification Paul Vicol Simon Fraser University University Burnaby, BC [email protected] Note: The following is my contribution to a group project for a graduate machine learning
Scalable Object Detection by Filter Compression with Regularized Sparse Coding
Scalable Object Detection by Filter Compression with Regularized Sparse Coding Ting-Hsuan Chao, Yen-Liang Lin, Yin-Hsi Kuo, and Winston H Hsu National Taiwan University, Taipei, Taiwan Abstract For practical
Going Deeper with Convolutional Neural Network for Intelligent Transportation
Going Deeper with Convolutional Neural Network for Intelligent Transportation by Tairui Chen A Thesis Submitted to the Faculty of the WORCESTER POLYTECHNIC INSTITUTE In partial fulfillment of the requirements
Conditional Random Fields as Recurrent Neural Networks
Conditional Random Fields as Recurrent Neural Networks Shuai Zheng 1, Sadeep Jayasumana *1, Bernardino Romera-Paredes 1, Vibhav Vineet 1,2, Zhizhong Su 3, Dalong Du 3, Chang Huang 3, and Philip H. S. Torr
SIGNAL INTERPRETATION
SIGNAL INTERPRETATION Lecture 6: ConvNets February 11, 2016 Heikki Huttunen [email protected] Department of Signal Processing Tampere University of Technology CONVNETS Continued from previous slideset
Weakly Supervised Fine-Grained Categorization with Part-Based Image Representation
ACCEPTED BY IEEE TIP 1 Weakly Supervised Fine-Grained Categorization with Part-Based Image Representation Yu Zhang, Xiu-Shen Wei, Jianxin Wu, Member, IEEE, Jianfei Cai, Senior Member, IEEE, Jiangbo Lu,
The Visual Internet of Things System Based on Depth Camera
The Visual Internet of Things System Based on Depth Camera Xucong Zhang 1, Xiaoyun Wang and Yingmin Jia Abstract The Visual Internet of Things is an important part of information technology. It is proposed
Getting Started with Caffe Julien Demouth, Senior Engineer
Getting Started with Caffe Julien Demouth, Senior Engineer What is Caffe? Open Source Framework for Deep Learning http://github.com/bvlc/caffe Developed by the Berkeley Vision and Learning Center (BVLC)
Convolutional Neural Networks with Intra-layer Recurrent Connections for Scene Labeling
Convolutional Neural Networks with Intra-layer Recurrent Connections for Scene Labeling Ming Liang Xiaolin Hu Bo Zhang Tsinghua National Laboratory for Information Science and Technology (TNList) Department
Fast Accurate Fish Detection and Recognition of Underwater Images with Fast R-CNN
Fast Accurate Fish Detection and Recognition of Underwater Images with Fast R-CNN Xiu Li 1, 2, Min Shang 1, 2, Hongwei Qin 1, 2, Liansheng Chen 1, 2 1. Department of Automation, Tsinghua University, Beijing
3D Model based Object Class Detection in An Arbitrary View
3D Model based Object Class Detection in An Arbitrary View Pingkun Yan, Saad M. Khan, Mubarak Shah School of Electrical Engineering and Computer Science University of Central Florida http://www.eecs.ucf.edu/
Applications of Deep Learning to the GEOINT mission. June 2015
Applications of Deep Learning to the GEOINT mission June 2015 Overview Motivation Deep Learning Recap GEOINT applications: Imagery exploitation OSINT exploitation Geospatial and activity based analytics
CNN Based Object Detection in Large Video Images. WangTao, [email protected] IQIYI ltd. 2016.4
CNN Based Object Detection in Large Video Images WangTao, [email protected] IQIYI ltd. 2016.4 Outline Introduction Background Challenge Our approach System framework Object detection Scene recognition Body
Object Categorization using Co-Occurrence, Location and Appearance
Object Categorization using Co-Occurrence, Location and Appearance Carolina Galleguillos Andrew Rabinovich Serge Belongie Department of Computer Science and Engineering University of California, San Diego
Online Tracking by Learning Discriminative Saliency Map with Convolutional Neural Network
Online Tracking by Learning Discriminative Saliency Map with Convolutional Neural Network Seunghoon Hong 1 Tackgeun You 1 Suha Kwak 2 Bohyung Han 1 1 Dept. of Computer Science and Engineering, POSTECH,
Group Sparse Coding. Fernando Pereira Google Mountain View, CA [email protected]. Dennis Strelow Google Mountain View, CA strelow@google.
Group Sparse Coding Samy Bengio Google Mountain View, CA [email protected] Fernando Pereira Google Mountain View, CA [email protected] Yoram Singer Google Mountain View, CA [email protected] Dennis Strelow
Object Detection from Video Tubelets with Convolutional Neural Networks
Object Detection from Video Tubelets with Convolutional Neural Networks Kai Kang Wanli Ouyang Hongsheng Li Xiaogang Wang Department of Electronic Engineering, The Chinese University of Hong Kong {kkang,wlouyang,hsli,xgwang}@ee.cuhk.edu.hk
siftservice.com - Turning a Computer Vision algorithm into a World Wide Web Service
siftservice.com - Turning a Computer Vision algorithm into a World Wide Web Service Ahmad Pahlavan Tafti 1, Hamid Hassannia 2, and Zeyun Yu 1 1 Department of Computer Science, University of Wisconsin -Milwaukee,
Taking Inverse Graphics Seriously
CSC2535: 2013 Advanced Machine Learning Taking Inverse Graphics Seriously Geoffrey Hinton Department of Computer Science University of Toronto The representation used by the neural nets that work best
arxiv:submit/1533655 [cs.cv] 13 Apr 2016
Bags of Local Convolutional Features for Scalable Instance Search Eva Mohedano, Kevin McGuinness and Noel E. O Connor Amaia Salvador, Ferran Marqués, and Xavier Giró-i-Nieto Insight Center for Data Analytics
Tattoo Detection for Soft Biometric De-Identification Based on Convolutional NeuralNetworks
1 Tattoo Detection for Soft Biometric De-Identification Based on Convolutional NeuralNetworks Tomislav Hrkać, Karla Brkić, Zoran Kalafatić Faculty of Electrical Engineering and Computing University of
High Level Describable Attributes for Predicting Aesthetics and Interestingness
High Level Describable Attributes for Predicting Aesthetics and Interestingness Sagnik Dhar Vicente Ordonez Tamara L Berg Stony Brook University Stony Brook, NY 11794, USA [email protected] Abstract
Cees Snoek. Machine. Humans. Multimedia Archives. Euvision Technologies The Netherlands. University of Amsterdam The Netherlands. Tree.
Visual search: what's next? Cees Snoek University of Amsterdam The Netherlands Euvision Technologies The Netherlands Problem statement US flag Tree Aircraft Humans Dog Smoking Building Basketball Table
Applying Deep Learning to Car Data Logging (CDL) and Driver Assessor (DA) October 22-Oct-15
Applying Deep Learning to Car Data Logging (CDL) and Driver Assessor (DA) October 22-Oct-15 GENIVI is a registered trademark of the GENIVI Alliance in the USA and other countries Copyright GENIVI Alliance
Marr Revisited: 2D-3D Alignment via Surface Normal Prediction
Marr Revisited: 2D-3D Alignment via Surface Normal Prediction Aayush Bansal Bryan Russell2 Abhinav Gupta Carnegie Mellon University 2 Adobe Research http://www.cs.cmu.edu/ aayushb/marrrevisited/ Input&Image&
Keypoint Density-based Region Proposal for Fine-Grained Object Detection and Classification using Regions with Convolutional Neural Network Features
Keypoint Density-based Region Proposal for Fine-Grained Object Detection and Classification using Regions with Convolutional Neural Network Features JT Turner 1, Kalyan Gupta 1, Brendan Morris 2, & David
Object class recognition using unsupervised scale-invariant learning
Object class recognition using unsupervised scale-invariant learning Rob Fergus Pietro Perona Andrew Zisserman Oxford University California Institute of Technology Goal Recognition of object categories
Lecture 2: The SVM classifier
Lecture 2: The SVM classifier C19 Machine Learning Hilary 2015 A. Zisserman Review of linear classifiers Linear separability Perceptron Support Vector Machine (SVM) classifier Wide margin Cost function
Classification using Intersection Kernel Support Vector Machines is Efficient
To Appear in IEEE Computer Vision and Pattern Recognition 28, Anchorage Classification using Intersection Kernel Support Vector Machines is Efficient Subhransu Maji EECS Department U.C. Berkeley [email protected]
PANDA: Pose Aligned Networks for Deep Attribute Modeling
PANDA: Pose Aligned Networks for Deep Attribute Modeling Ning Zhang 1,2, Manohar Paluri 1, Marc Aurelio Ranzato 1, Trevor Darrell 2, Lubomir Bourdev 1 1 Facebook AI Research 2 EECS, UC Berkeley {nzhang,
Stochastic Pooling for Regularization of Deep Convolutional Neural Networks
Stochastic Pooling for Regularization of Deep Convolutional Neural Networks Matthew D. Zeiler Department of Computer Science Courant Institute, New York University [email protected] Rob Fergus Department
arxiv:1505.04597v1 [cs.cv] 18 May 2015
U-Net: Convolutional Networks for Biomedical Image Segmentation Olaf Ronneberger, Philipp Fischer, and Thomas Brox arxiv:1505.04597v1 [cs.cv] 18 May 2015 Computer Science Department and BIOSS Centre for
TouchPaper - An Augmented Reality Application with Cloud-Based Image Recognition Service
TouchPaper - An Augmented Reality Application with Cloud-Based Image Recognition Service Feng Tang, Daniel R. Tretter, Qian Lin HP Laboratories HPL-2012-131R1 Keyword(s): image recognition; cloud service;
Image Super-Resolution Using Deep Convolutional Networks
1 Image Super-Resolution Using Deep Convolutional Networks Chao Dong, Chen Change Loy, Member, IEEE, Kaiming He, Member, IEEE, and Xiaoou Tang, Fellow, IEEE arxiv:1501.00092v3 [cs.cv] 31 Jul 2015 Abstract
Are we ready for Autonomous Driving? The KITTI Vision Benchmark Suite
Are we ready for Autonomous Driving? The KITTI Vision Benchmark Suite Philip Lenz 1 Andreas Geiger 2 Christoph Stiller 1 Raquel Urtasun 3 1 KARLSRUHE INSTITUTE OF TECHNOLOGY 2 MAX-PLANCK-INSTITUTE IS 3
Big Data: Image & Video Analytics
Big Data: Image & Video Analytics How it could support Archiving & Indexing & Searching Dieter Haas, IBM Deutschland GmbH The Big Data Wave 60% of internet traffic is multimedia content (images and videos)
Learning to Process Natural Language in Big Data Environment
CCF ADL 2015 Nanchang Oct 11, 2015 Learning to Process Natural Language in Big Data Environment Hang Li Noah s Ark Lab Huawei Technologies Part 1: Deep Learning - Present and Future Talk Outline Overview
CIKM 2015 Melbourne Australia Oct. 22, 2015 Building a Better Connected World with Data Mining and Artificial Intelligence Technologies
CIKM 2015 Melbourne Australia Oct. 22, 2015 Building a Better Connected World with Data Mining and Artificial Intelligence Technologies Hang Li Noah s Ark Lab Huawei Technologies We want to build Intelligent
Locality-constrained Linear Coding for Image Classification
Locality-constrained Linear Coding for Image Classification Jinjun Wang, Jianchao Yang,KaiYu, Fengjun Lv, Thomas Huang, and Yihong Gong Akiira Media System, Palo Alto, California Beckman Institute, University
An Analysis of Single-Layer Networks in Unsupervised Feature Learning
An Analysis of Single-Layer Networks in Unsupervised Feature Learning Adam Coates 1, Honglak Lee 2, Andrew Y. Ng 1 1 Computer Science Department, Stanford University {acoates,ang}@cs.stanford.edu 2 Computer
MVA ENS Cachan. Lecture 2: Logistic regression & intro to MIL Iasonas Kokkinos [email protected]
Machine Learning for Computer Vision 1 MVA ENS Cachan Lecture 2: Logistic regression & intro to MIL Iasonas Kokkinos [email protected] Department of Applied Mathematics Ecole Centrale Paris Galen
EdVidParse: Detecting People and Content in Educational Videos
EdVidParse: Detecting People and Content in Educational Videos by Michele Pratusevich S.B., Massachusetts Institute of Technology (2013) Submitted to the Department of Electrical Engineering and Computer
Image Captioning A survey of recent deep-learning approaches
Image Captioning A survey of recent deep-learning approaches Kiran Vodrahalli February 23, 2015 The task We want to automatically describe images with words Why? 1) It's cool 2) Useful for tech companies
Practical Tour of Visual tracking. David Fleet and Allan Jepson January, 2006
Practical Tour of Visual tracking David Fleet and Allan Jepson January, 2006 Designing a Visual Tracker: What is the state? pose and motion (position, velocity, acceleration, ) shape (size, deformation,
Learning Motion Categories using both Semantic and Structural Information
Learning Motion Categories using both Semantic and Structural Information Shu-Fai Wong, Tae-Kyun Kim and Roberto Cipolla Department of Engineering, University of Cambridge, Cambridge, CB2 1PZ, UK {sfw26,
arxiv:1312.6034v2 [cs.cv] 19 Apr 2014
Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps arxiv:1312.6034v2 [cs.cv] 19 Apr 2014 Karen Simonyan Andrea Vedaldi Andrew Zisserman Visual Geometry Group,
Two-Stream Convolutional Networks for Action Recognition in Videos
Two-Stream Convolutional Networks for Action Recognition in Videos Karen Simonyan Andrew Zisserman Visual Geometry Group, University of Oxford {karen,az}@robots.ox.ac.uk Abstract We investigate architectures
