Announcements. Image formation. Recap: Grouping & fitting. Recap: Features and filters. Multiple views. Plan 3/21/2011. CS 376 Lecture 15 1

Similar documents
Lecture 12: Cameras and Geometry. CAP 5415 Fall 2010

Feature Tracking and Optical Flow

Camera geometry and image alignment

Image Formation. 7-year old s question. Reference. Lecture Overview. It receives light from all directions. Pinhole

Probabilistic Latent Semantic Analysis (plsa)

The Geometry of Perspective Projection

A Study on SURF Algorithm and Real-Time Tracking Objects Using Optical Flow

Optical Tracking Using Projective Invariant Marker Pattern Properties

Geometric Camera Parameters

Local features and matching. Image classification & object localization

Lenses and Telescopes

Epipolar Geometry. Readings: See Sections 10.1 and 15.6 of Forsyth and Ponce. Right Image. Left Image. e(p ) Epipolar Lines. e(q ) q R.

Distinctive Image Features from Scale-Invariant Keypoints

Thin Lenses Drawing Ray Diagrams

3D Scanner using Line Laser. 1. Introduction. 2. Theory

INTRODUCTION TO RENDERING TECHNIQUES

Geometric Optics Converging Lenses and Mirrors Physics Lab IV

Randomized Trees for Real-Time Keypoint Recognition

LIST OF CONTENTS CHAPTER CONTENT PAGE DECLARATION DEDICATION ACKNOWLEDGEMENTS ABSTRACT ABSTRAK

PHOTOGRAMMETRIC TECHNIQUES FOR MEASUREMENTS IN WOODWORKING INDUSTRY

Build Panoramas on Android Phones

Solution Guide III-C. 3D Vision. Building Vision for Business. MVTec Software GmbH

CS 534: Computer Vision 3D Model-based recognition

Edge detection. (Trucco, Chapt 4 AND Jain et al., Chapt 5) -Edges are significant local changes of intensity in an image.

WHITE PAPER. Are More Pixels Better? Resolution Does it Really Matter?

EFFICIENT VEHICLE TRACKING AND CLASSIFICATION FOR AN AUTOMATED TRAFFIC SURVEILLANCE SYSTEM

Segmentation & Clustering

Polarization of Light

Tracking in flussi video 3D. Ing. Samuele Salti

Practical Tour of Visual tracking. David Fleet and Allan Jepson January, 2006

Revision problem. Chapter 18 problem 37 page 612. Suppose you point a pinhole camera at a 15m tall tree that is 75m away.

Recognition. Sanja Fidler CSC420: Intro to Image Understanding 1 / 28

Endoscope Optics. Chapter Introduction

Projective Geometry. Projective Geometry

Robot Perception Continued

Monash University Clayton s School of Information Technology CSE3313 Computer Graphics Sample Exam Questions 2007

Canny Edge Detection

Convolution. 1D Formula: 2D Formula: Example on the web:

VECTORAL IMAGING THE NEW DIRECTION IN AUTOMATED OPTICAL INSPECTION

Relating Vanishing Points to Catadioptric Camera Calibration

EXPERIMENT 6 OPTICS: FOCAL LENGTH OF A LENS

Chapter 6 Telescopes: Portals of Discovery. How does your eye form an image? Refraction. Example: Refraction at Sunset.

Introduction. Selim Aksoy. Bilkent University

Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches

Understanding Line Scan Camera Applications

Digital Photography Composition. Kent Messamore 9/8/2013

Integrated sensors for robotic laser welding

Optical Flow. Shenlong Wang CSC2541 Course Presentation Feb 2, 2016

PDF Created with deskpdf PDF Writer - Trial ::

9/16 Optics 1 /11 GEOMETRIC OPTICS

3D OBJECT MODELING AND RECOGNITION IN PHOTOGRAPHS AND VIDEO

Point Matching as a Classification Problem for Fast and Robust Object Pose Estimation

3D Model based Object Class Detection in An Arbitrary View

Object tracking & Motion detection in video sequences

Automatic 3D Reconstruction via Object Detection and 3D Transformable Model Matching CS 269 Class Project Report

2) A convex lens is known as a diverging lens and a concave lens is known as a converging lens. Answer: FALSE Diff: 1 Var: 1 Page Ref: Sec.

MVA ENS Cachan. Lecture 2: Logistic regression & intro to MIL Iasonas Kokkinos Iasonas.kokkinos@ecp.fr

Bildverarbeitung und Mustererkennung Image Processing and Pattern Recognition

A PHOTOGRAMMETRIC APPRAOCH FOR AUTOMATIC TRAFFIC ASSESSMENT USING CONVENTIONAL CCTV CAMERA

Lecture 17. Image formation Ray tracing Calculation. Lenses Convex Concave. Mirrors Convex Concave. Optical instruments

The use of computer vision technologies to augment human monitoring of secure computing facilities

Part-Based Recognition

Android Ros Application

DESIGN & DEVELOPMENT OF AUTONOMOUS SYSTEM TO BUILD 3D MODEL FOR UNDERWATER OBJECTS USING STEREO VISION TECHNIQUE

Understanding astigmatism Spring 2003

Imaging Systems Laboratory II. Laboratory 4: Basic Lens Design in OSLO April 2 & 4, 2002

Experiment 3 Lenses and Images

Spatial location in 360 of reference points over an object by using stereo vision

Introduction to Computer Graphics

Projective Geometry: A Short Introduction. Lecture Notes Edmond Boyer

CS 4204 Computer Graphics

Solution Derivations for Capa #14

Beginners Guide to Digital Camera Settings

4BA6 - Topic 4 Dr. Steven Collins. Chap. 5 3D Viewing and Projections

Feature Point Selection using Structural Graph Matching for MLS based Image Registration

Lesson 29: Lenses. Double Concave. Double Convex. Planoconcave. Planoconvex. Convex meniscus. Concave meniscus

Wii Remote Calibration Using the Sensor Bar

Digital Imaging and Multimedia. Filters. Ahmed Elgammal Dept. of Computer Science Rutgers University

What Makes a Great Picture?

3D Human Face Recognition Using Point Signature

A Comparative Study between SIFT- Particle and SURF-Particle Video Tracking Algorithms

4. CAMERA ADJUSTMENTS

Mean-Shift Tracking with Random Sampling

Create a Camera Obscura

Computer Vision. Image acquisition. 25 August Copyright by NHL Hogeschool and Van de Loosdrecht Machine Vision BV All rights reserved

Segmentation of building models from dense 3D point-clouds

How an electronic shutter works in a CMOS camera. First, let s review how shutters work in film cameras.

Static Environment Recognition Using Omni-camera from a Moving Vehicle

Light and its effects

Convex Mirrors. Ray Diagram for Convex Mirror

Introduction Epipolar Geometry Calibration Methods Further Readings. Stereo Camera Calibration

OBJECT TRACKING USING LOG-POLAR TRANSFORMATION

Edge-based Template Matching and Tracking for Perspectively Distorted Planar Objects

Automated Location Matching in Movies

MetropoGIS: A City Modeling System DI Dr. Konrad KARNER, DI Andreas KLAUS, DI Joachim BAUER, DI Christopher ZACH

Space Perception and Binocular Vision

siftservice.com - Turning a Computer Vision algorithm into a World Wide Web Service

Online Learning of Patch Perspective Rectification for Efficient Object Detection

A Prototype For Eye-Gaze Corrected

Template-based Eye and Mouth Detection for 3D Video Conferencing

Transcription:

Announcements Image formation Reminder: Pset 3 due March 30 Midterms: pick up today Monday March 21 Prof. UT Austin Recap: Features and filters Recap: Grouping & fitting Transforming and describing images; textures, colors, edges Clustering, segmentation, fitting; what parts belong together? [fig from Shi et al] Multiple views Multi-view geometry, matching, invariant features, stereo vision Plan Today: Local feature matching btwn views (wrap-up) Image formation, geometry of a single view Hartley and Zisserman Lowe Wednesday: Multiple views and epipolar geometry Monday: Approaches for stereo correspondence CS 376 Lecture 15 1

Previously Local invariant features Detection of interest points Harris corner detection Scale invariant blob detection: LoG Description of local patches SIFT : Histograms of oriented gradients Matching descriptors Local features: main components 1) Detection: Identify the interest points 2) Description:Extract vector feature descriptor surrounding each interest point. 3) Matching: Determine correspondence between descriptors in two views (1) x [ x,, x (1) 1 1 d ] (2) x [ x,, x (2) 2 1 d ] Recall: Corners as distinctive interest points Harris Detector: Steps edge : 1 >> 2 2 >> 1 corner : 1 and 2 are large, 1 ~ 2 ; flat region 1 and 2 are small; One way to score the cornerness: Harris Detector: Steps Compute corner response f Harris Detector: Steps CS 376 Lecture 15 2

Blob detection in 2D: scale selection 2 2 2 g g Laplacian-of-Gaussian = blob detector g 2 2 x y Blob detection in 2D We define the characteristic scale as the scale that produces peak of Laplacian response filter scales characteristic scale img1 img2 img3 Slide credit: Lana Lazebnik Example Original image at ¾ the size Scale invariant interest points Interest points are local maxima in both position and scale. L ( ) L ( ) xx yy scale Squared filter response maps List of (x, y, σ) Scale-space blob detector: Example Local features: main components 1) Detection: Identify the interest points 2) Description:Extract vector feature descriptor surrounding each interest point. (1) x [ x,, x (1) 1 1 d ] T. Lindeberg. Feature detection with automatic scale selection. IJCV 1998. 3) Matching: Determine correspondence between descriptors in two views (2) x [ x,, x (2) 2 1 d ] CS 376 Lecture 15 3

SIFT descriptor [Lowe 2004] Making descriptor rotation invariant Use histograms to bin pixels within sub-patches according to their orientation. 0 2 Why subpatches? Why does SIFT have some illumination invariance? CSE 576: Computer Vision Rotate patch according to its dominant gradient orientation This puts the patches into a canonical orientation. Image from Matthew Brown SIFT descriptor [Lowe 2004] SIFT properties Extraordinarily robust matching technique Can handle changes in viewpoint Up to about 60 degree out of plane rotation Can handle significant changes in illumination Sometimes even day vs. night (below) Fast and efficient can run in real time Lots of code available http://people.csail.mit.edu/albert/ladypack/wiki/index.php/known_implementations_of_sift Invariant to Scale Rotation Partially invariant to Illumination changes Camera viewpoint Occlusion, clutter Steve Seitz Local features: main components Matching local features 1) Detection: Identify the interest points 2) Description:Extract vector feature descriptor surrounding each interest point. 3) Matching: Determine correspondence between descriptors in two views CS 376 Lecture 15 4

Matching local features Ambiguous matches????? Image 1 Image 2 To generate candidate matches, find patches that have the most similar appearance (e.g., lowest SSD) Simplest approach: compare them all, take the closest (or closest k, or within a thresholded distance) Image 1 Image 2 At what SSD value do we have a good match? To add robustness to matching, can consider ratio : distance to best match / distance to second best match If low, first match looks good. If high, could be ambiguous match. Matching SIFT Descriptors Nearest neighbor (Euclidean distance) Threshold ratio of nearest to 2 nd nearest descriptor Recap: robust feature-based alignment Derek Hoiem Lowe IJCV 2004 Source: L. Lazebnik Recap: robust feature-based alignment Recap: robust feature-based alignment Extract features Extract features Compute putative matches Source: L. Lazebnik Source: L. Lazebnik CS 376 Lecture 15 5

Recap: robust feature-based alignment Recap: robust feature-based alignment Extract features Compute putative matches Loop: Hypothesize transformation T (small group of putative matches that are related by T) Source: L. Lazebnik Extract features Compute putative matches Loop: Hypothesize transformation T (small group of putative matches that are related by T) Verify transformation (search for other matches consistent with T) Source: L. Lazebnik Recap: robust feature-based alignment Extract features Compute putative matches Loop: Hypothesize transformation T (small group of putative matches that are related by T) Verify transformation (search for other matches consistent with T) Applications of local invariant features Wide baseline stereo Motion tracking Panoramas Mobile robot navigation 3D reconstruction Recognition Source: L. Lazebnik Automatic mosaicing Wide baseline stereo http://www.cs.ubc.ca/~mbrown/autostitch/autostitch.html [Image from T. Tuytelaars ECCV 2006 tutorial] CS 376 Lecture 15 6

Recognition of specific objects, scenes Schmid and Mohr 1997 Sivic and Zisserman, 2003 Plan Today: Local feature matching btwn views (wrap-up) Image formation, geometry of a single view Wednesday: Multiple views and epipolar geometry Monday: Approaches for stereo correspondence Rothganger et al. 2003 Lowe 2002 Image formation How are objects in the world captured in an image? Physical parameters of image formation Geometric Type of projection Camera pose Optical Sensor s lens type focal length, field of view, aperture Photometric Type, direction, intensity of light reaching sensor Surfaces reflectance properties Image formation Pinhole camera Let s design a camera Idea 1: put a piece of film in front of an object Do we get a reasonable image? Add a barrier to block off most of the rays This reduces blurring The opening is known as the aperture How does this transform the image? Slide by Steve Seitz Slide by Steve Seitz CS 376 Lecture 15 7

Pinhole camera Camera obscura Pinhole camera is a simple model to approximate imaging process, perspective projection. Image plane In Latin, means dark room If we treat pinhole as a point, only one ray from any given point can enter the camera. Fig from Forsyth and Ponce Virtual image pinhole "Reinerus Gemma-Frisius, observed an eclipse of the sun at Louvain on January 24, 1544, and later he used this illustration of the event in his book De Radio Astronomica et Geometrica, 1545. It is thought to be the first published illustration of a camera obscura..." Hammond, John H., The Camera Obscura, A Chronicle http://www.acmi.net.au/aic/camera_obscura.html Camera obscura Camera obscura at home Jetty at Margate England, 1898. An attraction in the late 19 th century Around 1870s http://brightbytes.com/cosite/collection2.html Adapted from R. Duraiswami Sketch from http://www.funsci.com/fun3_en/sky/sky.htm http://blog.makezine.com/archive/2006/02/how_to_room_ sized_camera_obscu.html Perspective effects Perspective effects Far away objects appear smaller Forsyth and Ponce CS 376 Lecture 15 8

Perspective effects Perspective effects Parallel lines in the scene intersect in the image Converge in image on horizon line Image plane (virtual) pinhole Scene Projection properties Many-to-one: any points along same ray map to same point in image Points points Lines lines (collinearity preserved) Distances and angles are not preserved Perspective and art Use of correct perspective projection indicated in 1 st century B.C. frescoes Skill resurfaces in Renaissance: artists develop systematic methods to determine perspective projection (around 1480-1515) Degenerate cases: Line through focal point projects to a point. Plane through focal point projects to line Plane perpendicular to image plane projects to part of the image. Raphael Durer, 1525 Perspective projection equations 3d world mapped to 2d projection in image plane Image plane Focal length Homogeneous coordinates Is this a linear transformation? no division by z is nonlinear Trick: add one more coordinate: Camera frame Optical axis Scene / world points homogeneous image coordinates homogeneous scene coordinates Converting from homogeneous coordinates Scene point Image coordinates Forsyth and Ponce Slide by Steve Seitz CS 376 Lecture 15 9

Perspective Projection Matrix Projection is a matrix multiplication using homogeneous coordinates: 1 0 0 0 1 0 0 0 1/ f ' x 0 x y 0 y z 0 z / f ' 1 x y ( f ', f ' ) z z divide by the third coordinate to convert back to non-homogeneous coordinates Weak perspective Approximation: treat magnification as constant Assumes scene depth << average distance to camera Image plane World points: Slide by Steve Seitz Orthographic projection Given camera at constant distance from scene World points projected along rays parallel to optical access Physical parameters of image formation Geometric Type of projection Camera pose Optical Sensor s lens type focal length, field of view, aperture Photometric Type, direction, intensity of light reaching sensor Surfaces reflectance properties Pinhole size / aperture How does the size of the aperture affect the image we d get? Larger Adding a lens focal point f Smaller A lens focuses light onto the film Rays passing through the center are not deviated All parallel rays converge to one point on a plane located at the focal length f Slide by Steve Seitz CS 376 Lecture 15 10

Pinhole vs. lens Cameras with lenses F optical center (Center Of Projection) focal point A lens focuses parallel rays onto a single focal point Gather more light, while keeping focus; make pinhole perspective projection practical Thin lens equation Thin lens equation u v u v How to relate distance of object from optical center (u) to the distance at which it will be in focus (v), given focal length f? How to relate distance of object from optical center (u) to the distance at which it will be in focus (v), given focal length f? Thin lens equation Thin lens equation u v How to relate distance of object from optical center (u) to the distance at which it will be in focus (v), given focal length f? u v 1 1 1 f u v Any object point satisfying this equation is in focus CS 376 Lecture 15 11

Focus and depth of field Focus and depth of field Depth of field: distance between image planes where blur is tolerable Thin lens: scene points at distinct depths come in focus at different image planes. (Real camera lens systems have greater depth of field.) circles of confusion Image credit: cambridgeincolour.com Shapiro and Stockman Focus and depth of field How does the aperture affect the depth of field? Angular measure of portion of 3d space seen by the camera Field of view A smaller aperture increases the range in which the object is approximately in focus Flower images from Wikipedia http://en.wikipedia.org/wiki/depth_of_field Slide from S. Seitz Images from http://en.wikipedia.org/wiki/angle_of_view Field of view depends on focal length Field of view depends on focal length As f gets smaller, image becomes more wide angle more world points project onto the finite image plane As f gets larger, image becomes more telescopic smaller part of the world projects onto the finite image plane from R. Duraiswami Smaller FOV = larger Focal Length Slide by A. Efros CS 376 Lecture 15 12

Digital cameras Film sensor array Often an array of charge coupled devices Each CCD is light sensitive diode that converts photons (light energy) to electrons CCD array camera optics frame grabber computer Summary Image formation affected by geometry, photometry, and optics. Projection equations express how world points mapped to 2d image. Homogenous coordinates allow linear system for projection equations. Lenses make pinhole model practical. Parameters (focal length, aperture, lens diameter, ) affect image obtained. Next time Geometry of two views scene point optical center image plane CS 376 Lecture 15 13