Gaze Estimation Using Image Segmentation, Feature Extraction, & Neural Networks. by Katie Morgan
|
|
|
- Alexina Edwards
- 9 years ago
- Views:
Transcription
1 Gaze Estimation Using Image Segmentation, Feature Extraction, & Neural Networks by Katie Morgan
2 Introduction Human gaze estimation is a rapidly growing field of research with many useful applications. It is especially useful for human-computer interaction (HCI) applications, such as video conferencing and eye-typing Many eye and gaze tracking methods exist, but not all are desirable for the average computer user, either because they are intrusive, expensive, or require frequent recalibration
3 Previous Works One method [10] uses an infrared camera and an infrared light to obtain bright and dark pupil images which highlight the pupil and cause an infrared glint on the eye. By fitting an ellipse to the pupil and using the glint coordinates, user gaze is accurately approximated Requires the use of expensive and specialized equipment
4 Previous Works Another method [3] uses a simple camera to capture grayscale images of the user. A small eye image is segmented from this large image, and is then input to a neural network, which returns the gaze position coordinates Although straightforward, it was thought that the neural networks in this method would be very large, since it would need to represent images containing 600 pixel values.
5 Proposed Feature Extraction Method The method researched combines the approaches of the infrared method and the image neural networking method 1. A small eye image is segmented from the larger image of the user 2. Pupil is identified and an ellipse is fit around it 3. The area of the sclera (white of the eye) to the left and right of the pupil is calculated 4. Parameters from the fitted ellipse and the ratio of the left and right sclera areas are saved in feature vectors used to train 2 neural networks to approximate screen gaze coordinates By using feature vectors instead of images to train the neural networks, it was hypothesized that the networks would be smaller and more efficient
6 Equipment and Setup The system setup was designed to run in an open office environment Users sat about one meter away from the camera in front of the left monitor, after which the camera was manually adjusted once to put the subject s right eye near the center of the resulting images
7 Feature Extraction Eye segmentation Using thresholding and known information about average pupil features, extract the small image of the user s right eye from the larger image
8 Feature Extraction Pupil Ellipse Fitting Identify the pupil in the eye image and fit an ellipse around it
9 Feature Extraction Pupil Ellipse Fitting Histogram normalization is performed on the eye image to enhance the brightest and darkest values This image is thresholded so that only those points with the lowest 3% of pixel values remain The resulting clusters of points are analyzed based on size, and the cluster with the largest area is assumed to be the pupil The image indices corresponding to the pupil cluster points are input to an ellipse fitting method to determine the parameters of the ellipse which best fits the pupil points Angle of ellipse orientation, x & y coordinates of the ellipse center, and the ratio of it s major to minor axis are the 4 ellipse parameters used to train the networks.
10 Feature Extraction Sclera Detection & Area Calculation Detect the sclera to the left and right of the pupil and calculate the ratio of their areas
11 Feature Extraction Sclera Detection & Area Calculation Starting from the center of the pupil and moving left, compare the t difference in value between neighboring pixels If two pixels are found whose difference is greater than the difference ference threshold, then the edge between the iris and sclera has been found If no edge is detected, the difference threshold decreases and the t process is repeated until an edge is found The maximum and minimum values in a neighborhood around the first sclera pixel determine the range of pixel values of the sclera Using these values, a recursive algorithm searches a set window for neighboring pixels within value range, and increases the sclera area count until all appropriate pixels are found The same procedure repeated to find the right sclera area Once both counts are obtained, their ratio is calculated and used d as the 5 th element in the feature vector
12 Training Data Collection Four black, square points were produced in the center of the four screen quadrants One by one, in random order, each of these points was made much larger than the remaining three As a point was enlarged, the program allowed a one second delay time for the user to focus on the point The interface captured three different images to avoid having an image of the subject blinking The images and the position of the corresponding gaze point were appended to the end of a data file.
13 Training Data Collection Segmentation was then used to attempt to find the subject s right eye If an eye candidate was not found, then the program segmented the second image, and then the third If no eye candidates were found, the test data was rejected Once an eye candidate was found, the segmented image was saved to the same file as the data and full image This procedure was repeated for all 4 gaze points. The operator then manually checked the segmented images of the saved data Any segmented image not of an eye, then this image, along with the full image and data corresponding to it, were deleted By doing this, the neural networks would only be trained with relevant data.
14 Timing Results Avg. Time (s) Image Segmentation (fps) Feature Extraction (fps) Combined Method (fps) Image segmentation IR Feature Extraction Image NN Combined Method Timing results from the average runtimes of 77 images (combined) and 50 successful images (segmentation and feature extraction) Improving segmentation time will improve the combined speed
15 Feature Extraction Results The method was tested on twelve test subjects of varying age, gender, and race (Asian and Caucasian) The method performed fairly well for many test subjects
16 Feature Extraction Results Once the image was correctly segmented, features were accurately extracted in most cases. Glare, squints, and eyelashes were the usual causes of incorrect extraction. A correlation between gaze position and ellipse orientation can be seen although there is some difference for each test subject, the main ellipse shape and orientation remains similar for each of the four gaze points
17 Feature Extraction Results The feature extraction procedure also performs well for some users wearing glasses, although the results are dependent on the type of glasses worn, which affects the position of the glare on the lens.
18 Incorrect Results Hair Detected Eye Corner Detected One Sided Sclera Eyebrow Fit with Ellipse
19 Future Work Training the Neural Networks At this time, there is not sufficient training data for the appropriate training of the neural networks there was simply not enough time before this presentation to collect the hundreds of necessary feature vectors required The 50 feature vectors obtained in the data collection were from tests of many different people In order to effectively train the neural networks for this many subjects, many more feature vectors per person would be required Until the feature extraction and segmentation procedure can be improved and sped up, the neural network data would also be inconsistent and incorrect
20 Future Work Training the Neural Networks Eventually, there will be two neural networks, each trained with the same large set of input feature vectors and target gaze coordinates One network will approximate the vertical coordinate, and the other the horizontal, since there is a larger difference between horizontal movement than vertical movement Many other methods use General Regression Neural Networks because of their faster training times However, they often use varying numbers of layers, a capability not available in Matlab,, so the use of Radial Basis Neural Networks will also be examined during future study, since the two networks are very similar
21 Future Work There is much future work to be done to explore and improve this method Testing the method with different head angles and experimenting with the eye segmentation portion of the method Using edge enhancement to improve sclera detection Improving the accuracy of the system to divide the screen into 8 regions Implementing the method with a web-camera instead of the Sony DVR camera would also be a desirable future study in order to decrease the cost of this method even further
22 Works Cited [1] A. García,, A. Pérez,, F. Sánchez,, J.L. Pedraza,, M.L. Córdoba,, M.L. Muñoz,, and R.Méndez,, A Precise Eye-Gaze Detection and Tracking System, in Proceedings of the 11th International Conference in Central Europe on Computer Graphics, Visualization and Computer Vision,, [2] A.S. Johansen, D.W. Hansen, J.P. Hansen, and M. Nielsen, Eye Typing using Markov and Active Appearance Models, Sixth IEEE Workshop on Applications of Computer Vision,, pp , 136, [3] D. Machin,, L.-Q. Xu,, and P. Sheppard, A Novel Approach to Real-time Non-intrusive Gaze Finding, in British Machine Vision Conference,, pp , 437, [4] J. Nusairat,, N.O. Nawari,, and R. Liang,, Artificial Intelligence Techniques for the Design and Analysis s of Deep Foundations, The Electronic Journal of Geotechnical Engineering,, [5] J. Yang, and J. Zhu, Subpixel Eye Gaze Tracking, Proceedings of the Fifth IEEE International Conference on Automatic Face and Gesture Recognition,, p.131, [6] K. Fujimura, Q. Ji,, and Z. Zhu, Combining Kalman Filtering and Mean Shift for Real Time Eye Tracking Under Active IR Illumination, in International Conference on Pattern Recognition,, pp: IV , 321, [7] K. Fujimura, Q. Ji,, and Z. Zhu, Real-Time Eye Detection and Tracking Under Various Light Conditions, Eye Tracking Research & Application, ACM Press, New York, NY, USA, pp , 144, [8] N. Mukawa,, and T. Ohno,, A free-head, Simple Calibration, Gaze Tracking System That Enables Gaze-Based Interaction, Proceedings of the Eye tracking research & applications symposium m on Eye tracking research & applications, ACM Press, New York, NY, USA, pp , 122, [9] Q. Ji,, and X. Yang, Real Time 3D Face Pose Discrimination Based On Active A IR Illumination, Proceedings, 16th International Conference on Pattern Recognition, pp , 313, [10] Q. Ji,, and Z. Zhu, Eye and gaze Tracking for Interactive Graphic Display, ACM International Conference Proceeding Series; Vol. 24,, ACM Press, New York, NY, USA, pp , [11] Y. Ebisawa,, Improved Video-Based Eye-Gaze Detection Method, IEEE Trans. Instr and Meas., vol. 47, no.4, pp , 955, 1998.
23
24
25
26
27
28
29
30
31
32
33
Eye Tracking Instructions
Eye Tracking Instructions [1] Check to make sure that the eye tracker is properly connected and plugged in. Plug in the eye tracker power adaptor (the green light should be on. Make sure that the yellow
Analecta Vol. 8, No. 2 ISSN 2064-7964
EXPERIMENTAL APPLICATIONS OF ARTIFICIAL NEURAL NETWORKS IN ENGINEERING PROCESSING SYSTEM S. Dadvandipour Institute of Information Engineering, University of Miskolc, Egyetemváros, 3515, Miskolc, Hungary,
Face detection is a process of localizing and extracting the face region from the
Chapter 4 FACE NORMALIZATION 4.1 INTRODUCTION Face detection is a process of localizing and extracting the face region from the background. The detected face varies in rotation, brightness, size, etc.
Locating and Decoding EAN-13 Barcodes from Images Captured by Digital Cameras
Locating and Decoding EAN-13 Barcodes from Images Captured by Digital Cameras W3A.5 Douglas Chai and Florian Hock Visual Information Processing Research Group School of Engineering and Mathematics Edith
Starburst: A hybrid algorithm for video-based eye tracking combining feature-based and model-based approaches
Starburst: A hybrid algorithm for video-based eye tracking combining feature-based and model-based approaches Dongheng Li, David Winfield, Derrick J. Parkhurst Human Computer Interaction Program Iowa State
A concise guide to Safety Glasses, the different standards and the effects of light on the eye. Contents. Links. Year of publication: 2010
A concise guide to Safety Glasses, the different standards and the effects of light on the eye Year of publication: 2010 Produced by the leading supplier of Safety Glasses in the UK. All Rights Reserved.
Tracking and Recognition in Sports Videos
Tracking and Recognition in Sports Videos Mustafa Teke a, Masoud Sattari b a Graduate School of Informatics, Middle East Technical University, Ankara, Turkey [email protected] b Department of Computer
HANDS-FREE PC CONTROL CONTROLLING OF MOUSE CURSOR USING EYE MOVEMENT
International Journal of Scientific and Research Publications, Volume 2, Issue 4, April 2012 1 HANDS-FREE PC CONTROL CONTROLLING OF MOUSE CURSOR USING EYE MOVEMENT Akhil Gupta, Akash Rathi, Dr. Y. Radhika
PRODUCT SHEET. [email protected] [email protected] www.biopac.com
EYE TRACKING SYSTEMS BIOPAC offers an array of monocular and binocular eye tracking systems that are easily integrated with stimulus presentations, VR environments and other media. Systems Monocular Part
Face Recognition For Remote Database Backup System
Face Recognition For Remote Database Backup System Aniza Mohamed Din, Faudziah Ahmad, Mohamad Farhan Mohamad Mohsin, Ku Ruhana Ku-Mahamud, Mustafa Mufawak Theab 2 Graduate Department of Computer Science,UUM
Poker Vision: Playing Cards and Chips Identification based on Image Processing
Poker Vision: Playing Cards and Chips Identification based on Image Processing Paulo Martins 1, Luís Paulo Reis 2, and Luís Teófilo 2 1 DEEC Electrical Engineering Department 2 LIACC Artificial Intelligence
How To Use Eye Tracking With A Dual Eye Tracking System In A Collaborative Collaborative Eye Tracking (Duet)
Framework for colocated synchronous dual eye tracking Craig Hennessey Department of Electrical and Computer Engineering University of British Columbia Mirametrix Research [email protected] Abstract Dual
Tobii Technology AB. Accuracy and precision Test report. X2-60 fw 1.0.5. Date: 2013-08-16 Methodology/Software version: 2.1.7
Tobii Technology AB Accuracy and precision Test report X2-60 fw 1.0.5 Date: 2013-08-16 Methodology/Software version: 2.1.7 1. Introduction This document provides an overview of tests in terms of accuracy
Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches
Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches PhD Thesis by Payam Birjandi Director: Prof. Mihai Datcu Problematic
LIST OF CONTENTS CHAPTER CONTENT PAGE DECLARATION DEDICATION ACKNOWLEDGEMENTS ABSTRACT ABSTRAK
vii LIST OF CONTENTS CHAPTER CONTENT PAGE DECLARATION DEDICATION ACKNOWLEDGEMENTS ABSTRACT ABSTRAK LIST OF CONTENTS LIST OF TABLES LIST OF FIGURES LIST OF NOTATIONS LIST OF ABBREVIATIONS LIST OF APPENDICES
An Energy-Based Vehicle Tracking System using Principal Component Analysis and Unsupervised ART Network
Proceedings of the 8th WSEAS Int. Conf. on ARTIFICIAL INTELLIGENCE, KNOWLEDGE ENGINEERING & DATA BASES (AIKED '9) ISSN: 179-519 435 ISBN: 978-96-474-51-2 An Energy-Based Vehicle Tracking System using Principal
Circle Object Recognition Based on Monocular Vision for Home Security Robot
Journal of Applied Science and Engineering, Vol. 16, No. 3, pp. 261 268 (2013) DOI: 10.6180/jase.2013.16.3.05 Circle Object Recognition Based on Monocular Vision for Home Security Robot Shih-An Li, Ching-Chang
Real-Time Eye Gaze Tracking With an Unmodified Commodity Webcam Employing a Neural Network
To appear in Proceedings of ACM Conference on Human Factors in Computing Systems (CHI), Atlanta, GA, 2010 Real-Time Eye Gaze Tracking With an Unmodified Commodity Webcam Employing a Neural Network Weston
Mean-Shift Tracking with Random Sampling
1 Mean-Shift Tracking with Random Sampling Alex Po Leung, Shaogang Gong Department of Computer Science Queen Mary, University of London, London, E1 4NS Abstract In this work, boosting the efficiency of
An Approach for Utility Pole Recognition in Real Conditions
6th Pacific-Rim Symposium on Image and Video Technology 1st PSIVT Workshop on Quality Assessment and Control by Image and Video Analysis An Approach for Utility Pole Recognition in Real Conditions Barranco
Virtual Fitting by Single-shot Body Shape Estimation
Virtual Fitting by Single-shot Body Shape Estimation Masahiro Sekine* 1 Kaoru Sugita 1 Frank Perbet 2 Björn Stenger 2 Masashi Nishiyama 1 1 Corporate Research & Development Center, Toshiba Corporation,
International Journal of Advanced Information in Arts, Science & Management Vol.2, No.2, December 2014
Efficient Attendance Management System Using Face Detection and Recognition Arun.A.V, Bhatath.S, Chethan.N, Manmohan.C.M, Hamsaveni M Department of Computer Science and Engineering, Vidya Vardhaka College
Template-based Eye and Mouth Detection for 3D Video Conferencing
Template-based Eye and Mouth Detection for 3D Video Conferencing Jürgen Rurainsky and Peter Eisert Fraunhofer Institute for Telecommunications - Heinrich-Hertz-Institute, Image Processing Department, Einsteinufer
Tobii AB. Accuracy and precision Test report. Tobii Pro X3-120 fw 1.7.1. Date: 2015-09-14 Methodology/Software version: 2.1.7*
Tobii AB Accuracy and precision Test report Tobii Pro X3-120 fw 1.7.1 Date: 2015-09-14 Methodology/Software version: 2.1.7* 1. Introduction This document provides an overview of tests in terms of accuracy
A Dynamic Approach to Extract Texts and Captions from Videos
Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology IJCSMC, Vol. 3, Issue. 4, April 2014,
Appearance-based Eye Gaze Estimation
Appearance-based Eye Gaze Estimation Kar-Han Tan Department of CS and Beckman Institute University of Illinois at Urbana-Champaign David J Kriegman Department of CS University of California at San Diego
A Learning Based Method for Super-Resolution of Low Resolution Images
A Learning Based Method for Super-Resolution of Low Resolution Images Emre Ugur June 1, 2004 [email protected] Abstract The main objective of this project is the study of a learning based method
The Role of Size Normalization on the Recognition Rate of Handwritten Numerals
The Role of Size Normalization on the Recognition Rate of Handwritten Numerals Chun Lei He, Ping Zhang, Jianxiong Dong, Ching Y. Suen, Tien D. Bui Centre for Pattern Recognition and Machine Intelligence,
Automatic Calibration of an In-vehicle Gaze Tracking System Using Driver s Typical Gaze Behavior
Automatic Calibration of an In-vehicle Gaze Tracking System Using Driver s Typical Gaze Behavior Kenji Yamashiro, Daisuke Deguchi, Tomokazu Takahashi,2, Ichiro Ide, Hiroshi Murase, Kazunori Higuchi 3,
Colour Image Segmentation Technique for Screen Printing
60 R.U. Hewage and D.U.J. Sonnadara Department of Physics, University of Colombo, Sri Lanka ABSTRACT Screen-printing is an industry with a large number of applications ranging from printing mobile phone
Novelty Detection in image recognition using IRF Neural Networks properties
Novelty Detection in image recognition using IRF Neural Networks properties Philippe Smagghe, Jean-Luc Buessler, Jean-Philippe Urban Université de Haute-Alsace MIPS 4, rue des Frères Lumière, 68093 Mulhouse,
Eye-Tracking with Webcam-Based Setups: Implementation of a Real-Time System and an Analysis of Factors Affecting Performance
Universitat Autònoma de Barcelona Master in Computer Vision and Artificial Intelligence Report of the Master Project Option: Computer Vision Eye-Tracking with Webcam-Based Setups: Implementation of a Real-Time
Segmentation of building models from dense 3D point-clouds
Segmentation of building models from dense 3D point-clouds Joachim Bauer, Konrad Karner, Konrad Schindler, Andreas Klaus, Christopher Zach VRVis Research Center for Virtual Reality and Visualization, Institute
RESEARCH ON SPOKEN LANGUAGE PROCESSING Progress Report No. 29 (2008) Indiana University
RESEARCH ON SPOKEN LANGUAGE PROCESSING Progress Report No. 29 (2008) Indiana University A Software-Based System for Synchronizing and Preprocessing Eye Movement Data in Preparation for Analysis 1 Mohammad
Mouse Control using a Web Camera based on Colour Detection
Mouse Control using a Web Camera based on Colour Detection Abhik Banerjee 1, Abhirup Ghosh 2, Koustuvmoni Bharadwaj 3, Hemanta Saikia 4 1, 2, 3, 4 Department of Electronics & Communication Engineering,
Static Environment Recognition Using Omni-camera from a Moving Vehicle
Static Environment Recognition Using Omni-camera from a Moving Vehicle Teruko Yata, Chuck Thorpe Frank Dellaert The Robotics Institute Carnegie Mellon University Pittsburgh, PA 15213 USA College of Computing
How To Filter Spam Image From A Picture By Color Or Color
Image Content-Based Email Spam Image Filtering Jianyi Wang and Kazuki Katagishi Abstract With the population of Internet around the world, email has become one of the main methods of communication among
Simultaneous Gamma Correction and Registration in the Frequency Domain
Simultaneous Gamma Correction and Registration in the Frequency Domain Alexander Wong [email protected] William Bishop [email protected] Department of Electrical and Computer Engineering University
Component Ordering in Independent Component Analysis Based on Data Power
Component Ordering in Independent Component Analysis Based on Data Power Anne Hendrikse Raymond Veldhuis University of Twente University of Twente Fac. EEMCS, Signals and Systems Group Fac. EEMCS, Signals
Neural Network based Vehicle Classification for Intelligent Traffic Control
Neural Network based Vehicle Classification for Intelligent Traffic Control Saeid Fazli 1, Shahram Mohammadi 2, Morteza Rahmani 3 1,2,3 Electrical Engineering Department, Zanjan University, Zanjan, IRAN
A Simple Feature Extraction Technique of a Pattern By Hopfield Network
A Simple Feature Extraction Technique of a Pattern By Hopfield Network A.Nag!, S. Biswas *, D. Sarkar *, P.P. Sarkar *, B. Gupta **! Academy of Technology, Hoogly - 722 *USIC, University of Kalyani, Kalyani
Automatic 3D Reconstruction via Object Detection and 3D Transformable Model Matching CS 269 Class Project Report
Automatic 3D Reconstruction via Object Detection and 3D Transformable Model Matching CS 69 Class Project Report Junhua Mao and Lunbo Xu University of California, Los Angeles [email protected] and lunbo
Assessment. Presenter: Yupu Zhang, Guoliang Jin, Tuo Wang Computer Vision 2008 Fall
Automatic Photo Quality Assessment Presenter: Yupu Zhang, Guoliang Jin, Tuo Wang Computer Vision 2008 Fall Estimating i the photorealism of images: Distinguishing i i paintings from photographs h Florin
Projection Center Calibration for a Co-located Projector Camera System
Projection Center Calibration for a Co-located Camera System Toshiyuki Amano Department of Computer and Communication Science Faculty of Systems Engineering, Wakayama University Sakaedani 930, Wakayama,
Digital Image Requirements for New Online US Visa Application
Digital Image Requirements for New Online US Visa Application As part of the electronic submission of your DS-160 application, you will be asked to provide an electronic copy of your photo. The photo must
Do-It-Yourself Eye Tracker: Impact of the Viewing Angle on the Eye Tracking Accuracy
Do-It-Yourself Eye Tracker: Impact of the Viewing Angle on the Eye Tracking Accuracy Michal Kowalik Supervised by: Radoslaw Mantiuk Faculty of Computer Science West Pomeranian University of Technology
Crater detection with segmentation-based image processing algorithm
Template reference : 100181708K-EN Crater detection with segmentation-based image processing algorithm M. Spigai, S. Clerc (Thales Alenia Space-France) V. Simard-Bilodeau (U. Sherbrooke and NGC Aerospace,
INTRODUCTION TO MACHINE LEARNING 3RD EDITION
ETHEM ALPAYDIN The MIT Press, 2014 Lecture Slides for INTRODUCTION TO MACHINE LEARNING 3RD EDITION [email protected] http://www.cmpe.boun.edu.tr/~ethem/i2ml3e CHAPTER 1: INTRODUCTION Big Data 3 Widespread
Limitations of Human Vision. What is computer vision? What is computer vision (cont d)?
What is computer vision? Limitations of Human Vision Slide 1 Computer vision (image understanding) is a discipline that studies how to reconstruct, interpret and understand a 3D scene from its 2D images
REAL TIME TRAFFIC LIGHT CONTROL USING IMAGE PROCESSING
REAL TIME TRAFFIC LIGHT CONTROL USING IMAGE PROCESSING Ms.PALLAVI CHOUDEKAR Ajay Kumar Garg Engineering College, Department of electrical and electronics Ms.SAYANTI BANERJEE Ajay Kumar Garg Engineering
Designing eye tracking experiments to measure human behavior
Designing eye tracking experiments to measure human behavior Eindhoven, The Netherlands August, 2010 Ricardo Matos Tobii Technology Steps involved in measuring behaviour 1. Formulate and initial question
ANALYZING A CONDUCTORS GESTURES WITH THE WIIMOTE
ANALYZING A CONDUCTORS GESTURES WITH THE WIIMOTE ICSRiM - University of Leeds, School of Computing & School of Music, Leeds LS2 9JT, UK [email protected] www.icsrim.org.uk Abstract This paper presents
Technical Considerations Detecting Transparent Materials in Particle Analysis. Michael Horgan
Technical Considerations Detecting Transparent Materials in Particle Analysis Michael Horgan Who We Are Fluid Imaging Technology Manufacturers of the FlowCam series of particle analyzers FlowCam HQ location
LOCAL SURFACE PATCH BASED TIME ATTENDANCE SYSTEM USING FACE. [email protected]
LOCAL SURFACE PATCH BASED TIME ATTENDANCE SYSTEM USING FACE 1 S.Manikandan, 2 S.Abirami, 2 R.Indumathi, 2 R.Nandhini, 2 T.Nanthini 1 Assistant Professor, VSA group of institution, Salem. 2 BE(ECE), VSA
Fast Subsequent Color Iris Matching in large Database
Fast Subsequent Color Iris Matching in large Database Adnan Alam Khan 1, Safeeullah Soomro 2 and Irfan Hyder 3 1 PAF-KIET Department of Telecommunications, Employer of Institute of Business Management
PHOTOGRAMMETRIC TECHNIQUES FOR MEASUREMENTS IN WOODWORKING INDUSTRY
PHOTOGRAMMETRIC TECHNIQUES FOR MEASUREMENTS IN WOODWORKING INDUSTRY V. Knyaz a, *, Yu. Visilter, S. Zheltov a State Research Institute for Aviation System (GosNIIAS), 7, Victorenko str., Moscow, Russia
ECE 533 Project Report Ashish Dhawan Aditi R. Ganesan
Handwritten Signature Verification ECE 533 Project Report by Ashish Dhawan Aditi R. Ganesan Contents 1. Abstract 3. 2. Introduction 4. 3. Approach 6. 4. Pre-processing 8. 5. Feature Extraction 9. 6. Verification
HAND GESTURE BASEDOPERATINGSYSTEM CONTROL
HAND GESTURE BASEDOPERATINGSYSTEM CONTROL Garkal Bramhraj 1, palve Atul 2, Ghule Supriya 3, Misal sonali 4 1 Garkal Bramhraj mahadeo, 2 Palve Atule Vasant, 3 Ghule Supriya Shivram, 4 Misal Sonali Babasaheb,
Multimodal Biometric Recognition Security System
Multimodal Biometric Recognition Security System Anju.M.I, G.Sheeba, G.Sivakami, Monica.J, Savithri.M Department of ECE, New Prince Shri Bhavani College of Engg. & Tech., Chennai, India ABSTRACT: Security
International Journal of Computer Science Trends and Technology (IJCST) Volume 3 Issue 3, May-June 2015
RESEARCH ARTICLE OPEN ACCESS Data Mining Technology for Efficient Network Security Management Ankit Naik [1], S.W. Ahmad [2] Student [1], Assistant Professor [2] Department of Computer Science and Engineering
FACE RECOGNITION BASED ATTENDANCE MARKING SYSTEM
Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology IJCSMC, Vol. 3, Issue. 2, February 2014,
COLOR-BASED PRINTED CIRCUIT BOARD SOLDER SEGMENTATION
COLOR-BASED PRINTED CIRCUIT BOARD SOLDER SEGMENTATION Tz-Sheng Peng ( 彭 志 昇 ), Chiou-Shann Fuh ( 傅 楸 善 ) Dept. of Computer Science and Information Engineering, National Taiwan University E-mail: [email protected]
Camera Technology Guide. Factors to consider when selecting your video surveillance cameras
Camera Technology Guide Factors to consider when selecting your video surveillance cameras Introduction Investing in a video surveillance system is a smart move. You have many assets to protect so you
BCC Multi Stripe Wipe
BCC Multi Stripe Wipe The BCC Multi Stripe Wipe is a similar to a Horizontal or Vertical Blind wipe. It offers extensive controls to randomize the stripes parameters. The following example shows a Multi
A Method of Caption Detection in News Video
3rd International Conference on Multimedia Technology(ICMT 3) A Method of Caption Detection in News Video He HUANG, Ping SHI Abstract. News video is one of the most important media for people to get information.
A Method for Controlling Mouse Movement using a Real- Time Camera
A Method for Controlling Mouse Movement using a Real- Time Camera Hojoon Park Department of Computer Science Brown University, Providence, RI, USA [email protected] Abstract This paper presents a new
COMPARISON OF OBJECT BASED AND PIXEL BASED CLASSIFICATION OF HIGH RESOLUTION SATELLITE IMAGES USING ARTIFICIAL NEURAL NETWORKS
COMPARISON OF OBJECT BASED AND PIXEL BASED CLASSIFICATION OF HIGH RESOLUTION SATELLITE IMAGES USING ARTIFICIAL NEURAL NETWORKS B.K. Mohan and S. N. Ladha Centre for Studies in Resources Engineering IIT
ROBUST VEHICLE TRACKING IN VIDEO IMAGES BEING TAKEN FROM A HELICOPTER
ROBUST VEHICLE TRACKING IN VIDEO IMAGES BEING TAKEN FROM A HELICOPTER Fatemeh Karimi Nejadasl, Ben G.H. Gorte, and Serge P. Hoogendoorn Institute of Earth Observation and Space System, Delft University
A MACHINE LEARNING APPROACH TO FILTER UNWANTED MESSAGES FROM ONLINE SOCIAL NETWORKS
A MACHINE LEARNING APPROACH TO FILTER UNWANTED MESSAGES FROM ONLINE SOCIAL NETWORKS Charanma.P 1, P. Ganesh Kumar 2, 1 PG Scholar, 2 Assistant Professor,Department of Information Technology, Anna University
AN IMPROVED DOUBLE CODING LOCAL BINARY PATTERN ALGORITHM FOR FACE RECOGNITION
AN IMPROVED DOUBLE CODING LOCAL BINARY PATTERN ALGORITHM FOR FACE RECOGNITION Saurabh Asija 1, Rakesh Singh 2 1 Research Scholar (Computer Engineering Department), Punjabi University, Patiala. 2 Asst.
Protocol for Microscope Calibration
Protocol for Microscope Calibration A properly calibrated system is essential for successful and efficient software use. The following are step by step instructions on how to calibrate the hardware using
3D Scanner using Line Laser. 1. Introduction. 2. Theory
. Introduction 3D Scanner using Line Laser Di Lu Electrical, Computer, and Systems Engineering Rensselaer Polytechnic Institute The goal of 3D reconstruction is to recover the 3D properties of a geometric
Face Recognition in Low-resolution Images by Using Local Zernike Moments
Proceedings of the International Conference on Machine Vision and Machine Learning Prague, Czech Republic, August14-15, 014 Paper No. 15 Face Recognition in Low-resolution Images by Using Local Zernie
OptimizedIR in Axis cameras
Whitepaper OptimizedIR in Axis cameras Leveraging new, smart and long-life LED technology Table of contents Introduction 3 1. Why do you need ir illumination? 3 2. Built-in or stand-alone ir illumination?
Image Processing Based Automatic Visual Inspection System for PCBs
IOSR Journal of Engineering (IOSRJEN) ISSN: 2250-3021 Volume 2, Issue 6 (June 2012), PP 1451-1455 www.iosrjen.org Image Processing Based Automatic Visual Inspection System for PCBs Sanveer Singh 1, Manu
Galaxy Morphological Classification
Galaxy Morphological Classification Jordan Duprey and James Kolano Abstract To solve the issue of galaxy morphological classification according to a classification scheme modelled off of the Hubble Sequence,
Taking Inverse Graphics Seriously
CSC2535: 2013 Advanced Machine Learning Taking Inverse Graphics Seriously Geoffrey Hinton Department of Computer Science University of Toronto The representation used by the neural nets that work best
To Install EdiView IP camera utility on iphone, follow the following instructions:
To Install EdiView IP camera utility on iphone, follow the following instructions: Please download Ediview application from Apple s Appstore. You can launch EdiView by clicking EdiView icon. You ll see
VECTORAL IMAGING THE NEW DIRECTION IN AUTOMATED OPTICAL INSPECTION
VECTORAL IMAGING THE NEW DIRECTION IN AUTOMATED OPTICAL INSPECTION Mark J. Norris Vision Inspection Technology, LLC Haverhill, MA [email protected] ABSTRACT Traditional methods of identifying and
PASSENGER/PEDESTRIAN ANALYSIS BY NEUROMORPHIC VISUAL INFORMATION PROCESSING
PASSENGER/PEDESTRIAN ANALYSIS BY NEUROMORPHIC VISUAL INFORMATION PROCESSING Woo Joon Han Il Song Han Korea Advanced Science and Technology Republic of Korea Paper Number 13-0407 ABSTRACT The physiological
Performance Comparison of Visual and Thermal Signatures for Face Recognition
Performance Comparison of Visual and Thermal Signatures for Face Recognition Besma Abidi The University of Tennessee The Biometric Consortium Conference 2003 September 22-24 OUTLINE Background Recognition
A Real Time Hand Tracking System for Interactive Applications
A Real Time Hand Tracking System for Interactive Applications Siddharth Swarup Rautaray Indian Institute of Information Technology Allahabad ABSTRACT In vision based hand tracking systems color plays an
VISUAL RECOGNITION OF HAND POSTURES FOR INTERACTING WITH VIRTUAL ENVIRONMENTS
VISUAL RECOGNITION OF HAND POSTURES FOR INTERACTING WITH VIRTUAL ENVIRONMENTS Radu Daniel VATAVU Ştefan-Gheorghe PENTIUC "Stefan cel Mare" University of Suceava str.universitatii nr.13, RO-720229 Suceava
MACHINE VISION MNEMONICS, INC. 102 Gaither Drive, Suite 4 Mount Laurel, NJ 08054 USA 856-234-0970 www.mnemonicsinc.com
MACHINE VISION by MNEMONICS, INC. 102 Gaither Drive, Suite 4 Mount Laurel, NJ 08054 USA 856-234-0970 www.mnemonicsinc.com Overview A visual information processing company with over 25 years experience
The Implementation of Face Security for Authentication Implemented on Mobile Phone
The Implementation of Face Security for Authentication Implemented on Mobile Phone Emir Kremić *, Abdulhamit Subaşi * * Faculty of Engineering and Information Technology, International Burch University,
Vision based approach to human fall detection
Vision based approach to human fall detection Pooja Shukla, Arti Tiwari CSVTU University Chhattisgarh, [email protected] 9754102116 Abstract Day by the count of elderly people living alone at home
