A Fully Automatic Approach for Human Recognition from Profile Images Using 2D and 3D Ear Data



Similar documents
LOCAL SURFACE PATCH BASED TIME ATTENDANCE SYSTEM USING FACE.

International Journal of Advanced Information in Arts, Science & Management Vol.2, No.2, December 2014

Normalisation of 3D Face Data

Face Model Fitting on Low Resolution Images

Canny Edge Detection

Biometric Authentication using Online Signatures

Framework for Biometric Enabled Unified Core Banking

Performance Comparison of Visual and Thermal Signatures for Face Recognition

A Various Biometric application for authentication and identification

FACE RECOGNITION BASED ATTENDANCE MARKING SYSTEM

Multimodal Biometric Recognition Security System

Expression Invariant 3D Face Recognition with a Morphable Model

Face Recognition in Low-resolution Images by Using Local Zernike Moments

Signature Region of Interest using Auto cropping

Efficient Attendance Management: A Face Recognition Approach

Keywords Input Images, Damage face images, Face recognition system, etc...

A PHOTOGRAMMETRIC APPRAOCH FOR AUTOMATIC TRAFFIC ASSESSMENT USING CONVENTIONAL CCTV CAMERA

A Study of Automatic License Plate Recognition Algorithms and Techniques

High Resolution Fingerprint Matching Using Level 3 Features

AN IMPROVED DOUBLE CODING LOCAL BINARY PATTERN ALGORITHM FOR FACE RECOGNITION

Face Recognition For Remote Database Backup System

High-accuracy ultrasound target localization for hand-eye calibration between optical tracking systems and three-dimensional ultrasound

Practical Tour of Visual tracking. David Fleet and Allan Jepson January, 2006

Classroom Monitoring System by Wired Webcams and Attendance Management System

Index Terms: Face Recognition, Face Detection, Monitoring, Attendance System, and System Access Control.


HANDS-FREE PC CONTROL CONTROLLING OF MOUSE CURSOR USING EYE MOVEMENT

Advances in Face Recognition Research Second End-User Group Meeting - Feb 21, 2008 Dr. Stefan Gehlen, L-1 Identity Solutions AG, Bochum, Germany

Palmprint Recognition. By Sree Rama Murthy kora Praveen Verma Yashwant Kashyap

Interactive person re-identification in TV series

ECE 533 Project Report Ashish Dhawan Aditi R. Ganesan

3D Human Face Recognition Using Point Signature

Using Real Time Computer Vision Algorithms in Automatic Attendance Management Systems

Poker Vision: Playing Cards and Chips Identification based on Image Processing

Segmentation of building models from dense 3D point-clouds

Face detection is a process of localizing and extracting the face region from the

HAND GESTURE BASEDOPERATINGSYSTEM CONTROL

A Comparison of Photometric Normalisation Algorithms for Face Verification

Face Recognition: Some Challenges in Forensics. Anil K. Jain, Brendan Klare, and Unsang Park

Circle Object Recognition Based on Monocular Vision for Home Security Robot

CS 534: Computer Vision 3D Model-based recognition

Template-based Eye and Mouth Detection for 3D Video Conferencing

DESIGN OF DIGITAL SIGNATURE VERIFICATION ALGORITHM USING RELATIVE SLOPE METHOD

EFFICIENT VEHICLE TRACKING AND CLASSIFICATION FOR AN AUTOMATED TRAFFIC SURVEILLANCE SYSTEM

May For other information please contact:

A secure face tracking system

SIGNATURE VERIFICATION

MACHINE VISION MNEMONICS, INC. 102 Gaither Drive, Suite 4 Mount Laurel, NJ USA

Palmprint as a Biometric Identifier

Introduction to Pattern Recognition

Palmprint Classification

Mean-Shift Tracking with Random Sampling

Low-resolution Image Processing based on FPGA

Active Learning with Boosting for Spam Detection

FRACTAL RECOGNITION AND PATTERN CLASSIFIER BASED SPAM FILTERING IN SERVICE

Open-Set Face Recognition-based Visitor Interface System

An Active Head Tracking System for Distance Education and Videoconferencing Applications

Automatic Extraction of Signatures from Bank Cheques and other Documents

Extracting a Good Quality Frontal Face Images from Low Resolution Video Sequences

Image Normalization for Illumination Compensation in Facial Images

The Visual Internet of Things System Based on Depth Camera

Robust Real-Time Face Detection

Document Image Retrieval using Signatures as Queries

Optical Digitizing by ATOS for Press Parts and Tools

Environmental Remote Sensing GEOG 2021

Digital Identity & Authentication Directions Biometric Applications Who is doing what? Academia, Industry, Government

Visual-based ID Verification by Signature Tracking

Using Data Mining for Mobile Communication Clustering and Characterization

An Approach for Utility Pole Recognition in Real Conditions

Analecta Vol. 8, No. 2 ISSN

Simultaneous Gamma Correction and Registration in the Frequency Domain

Tracking performance evaluation on PETS 2015 Challenge datasets

Support Vector Machine-Based Human Behavior Classification in Crowd through Projection and Star Skeletonization

ACCURACY ASSESSMENT OF BUILDING POINT CLOUDS AUTOMATICALLY GENERATED FROM IPHONE IMAGES

Understanding The Face Image Format Standards

Naive-Deep Face Recognition: Touching the Limit of LFW Benchmark or Not?

MetropoGIS: A City Modeling System DI Dr. Konrad KARNER, DI Andreas KLAUS, DI Joachim BAUER, DI Christopher ZACH

MULTIMODAL BIOMETRICS IN IDENTITY MANAGEMENT

Very Low Frame-Rate Video Streaming For Face-to-Face Teleconference

Efficient Storage, Compression and Transmission

The Role of Size Normalization on the Recognition Rate of Handwritten Numerals

Development of Attendance Management System using Biometrics.

Extend Table Lens for High-Dimensional Data Visualization and Classification Mining

Illumination, Expression and Occlusion Invariant Pose-Adaptive Face Recognition System for Real- Time Applications

Interactive Data Mining and Visualization

What is the Right Illumination Normalization for Face Recognition?

How To Create A Surface From Points On A Computer With A Marching Cube

Jiří Matas. Hough Transform

The Scientific Data Mining Process

High Quality Image Magnification using Cross-Scale Self-Similarity

Fingerprint s Core Point Detection using Gradient Field Mask

SECURE WEB MARKETING USING EAR BIOMETRICS

Low-resolution Character Recognition by Video-based Super-resolution

Transcription:

A Fully Automatic Approach for Human Recognition from Profile Images Using 2D and 3D Ear Data S. M. S. Islam, M. Bennamoun, A. S. Mian and R. Davies School of Computer Science and Software Engineering, The University of Western Australia 35, Stirling Hwy, Crawley, WA 6009 {shams, bennamou, ajmal, rowan}@csse.uwa.edu.au Abstract The use of ear shape as a biometric trait for recognizing people in different applications is one of the most recent trends in the research communities. In this work, a fully automatic and fast technique based on the AdaBoost algorithm is used to detect a subject s ear from his/her 2D and corresponding 3D profile images. A modified version of the Iterative Closest Point (ICP) algorithm is then used for the matching of this extracted probe ear to the previously stored ear data in a gallery database. A coarse-tofine hierarchical technique is used where the ICP algorithm is first applied on low and then on high resolution meshes of 3D ear data. We obtain a rank one recognition rate of 93% while testing with the University of Notre Dame Biometrics Database. The proposed recognition approach does not require any manual intervention or sharp extraction of ear contour from the detected ear region. No segmentation of the extracted ear is required and more importantly, the system performance does not rely on the presence of a particular feature of the ear. 1. Introduction Due to instances of fraud with the traditional ID based systems, biometric recognition systems are gaining popularity day by day. In such a system, one or more physiological (e.g. face, fingerprint, palmprint, iris and DNA) or behavioral (e.g. handwriting, gait and voice) traits of a subject are taken into consideration for automatic recognition. Although the ear as a biometric trait is not as accurate as iris or DNA, it is non-intrusive and easy to be collected. The face is also non-intrusive but its appearance is affected by changes in facial expressions, use of cosmetics or eye glasses. The ear is also smaller in size but rich in features and its shape does not change with aging between 8 years and 70 years [7]. It can be used separately or in a multimodal approach with the face for effective human recognition in many applications including some national IDs, security, surveillance and law enforcement applications. However, in any of these applications, accurate recognition of the ear is an important step. A biometric recognition system may operate in one or both of two modes: authentication and identification. In authentication, one-to-one matching is performed to compare a user s biometric to the template of the claimed identity. In identification, one-to-many matching is done to associate an identity with the user by matching it against every identity in the database. For both modes, an accurate detection is an essential pre-requisite. However, ear detection from arbitrary profile images is a challenging problem due to the fact that the ear is sometimes occluded by hair and earrings and ear images can vary in appearance under different viewing and illumination conditions. Consequently, most of the existing ear recognition algorithms assume that the ear has been accurately detected [9]. One of the earliest ear detection methods uses Canny edge maps to detect the ear contour [3]. Hurley et al. [6] proposed the force field transformation for ear detection. Alvarez et al. [1] used a modified active contour algorithm and Ovoid model for detecting the ear. Yan and Bowyer [14] proposed taking a predefined sector from the nose tip to locate the ear region. The non-ear portion from that sector is cropped out by skin detection and the ear pit was detected using Gaussian smoothing and curvature estimation. Then, they applied an active contour algorithm to extract the ear contour. The system is automatic but fails if the ear pit is not visible. Li Yuan and Mu [15] used a modified CAMSHIF T algorithm to roughly track the profile image as the region of interest (ROI). Then, contour fitting is operated on ROI for further accurate localization using the contour information of the ear. Most recently, Islam et al. [8] proposed an ear detection approach based on the AdaBoost algorithm [12]. The system was trained with rectangular Haar-like features and using a dataset of varied races, sexes, appearances, orientations and illuminations. The data was collected by cropping

and synthesizing from the University of Notre Dame (UND) biometrics database [13, 14], the NIST Mugshot Identification Database (MID), the XM2VTSDB [10], the USTB, the MIT-CBL and the UMIST database. The approach is fully automatic, provides 100% detection while tested with 203 non-occluded images of the UND profile face database and also works well with some occluded and degraded images. As summarized in the survey of Pun et al. [11] and Islam et al. [9], most of the proposed ear recognition approaches use either PCA (Principal Component Analysis) or the ICP algorithm for matching. Choras [4] proposed a different automated geometrical method. Testing with 240 images (20 different views) of 12 subjects, 100% recognition rate is reported. Genetic local search and the force field transformation based approaches have also been proposed by Yuizono et al. [16] and Hurley et al. [6] respectively. The first ever ear recognition system tested with a larger database of 415 subjects is proposed by Yan and Bowyer [14]. Using a modified version of the ICP, they achieved an accuracy of 95.7% with occlusion and 97.8 % without occlusion (with an Equal-error rate (EER) of 1.2%). The system does not work well if the ear pit is not visible. In this work, we have adopted the work of Islam et al. [8] for detecting the ear from 2D profile images and then, extended it for cropping the corresponding 3D profile face data. After ear detection, we apply a variant of the Iterative Closest Point (ICP) algorithm for recognition of the ear at different mesh resolutions of the extracted 3D ear data. Using two different resolutions hierarchically, we obtain a rank 1 recognition rate of 93%. The proposed system is fully automatic and does not rely on the presence of a particular feature of the ear ( e.g. ear pit). It also does not require an accurate extraction of the ear contour and hence reduces the computational cost. Besides, the ear recognition results can be combined with other biometric modalities such as 2D and 3D faces to obtain a more robust and accurate human recognition system. The paper is organized as follows. The proposed system for 3D ear detection and that for 3D ear recognition are described in Section 2 and 3 respectively. Results obtained are reported and discussed in Section 4. Section 5 concludes our findings. 2. 3D Ear Detection and Normalization The proposed 3D ear detection approach utilizes the AdaBoost based 2D ear detection technique developed by Islam et al. [8]. Since the 3D profile data are co-registered with corresponding the 2D images, the 2D ear detector is first scanned through the whole profile image to localize the ear. A rectangle is placed covering the ear. The 3D data corresponding to this rectangular region is cropped for use in 3D ear recognition. The complete ear detection process from 2D and 3D face profile image/data is shown in Figure 1. A sample of a profile image and corresponding 2D and 3D ear data detected by our system is also shown in the same figure. Data collection (2D and 3D profile faces) Pre-processing the 2D data 2D 3D 2D Validation and update of the training with false positives Training of the classifier using Cascaded AdaBoost Creation of rectangular features Off-line On-line Crop 2D ear region using the trained classifier Extract the corresponding 3D Ear data Figure 1. Block diagram of the 3D ear detection approach. Once the 3D ear is detected, we remove all the spikes and holes by filtering the data. We perform triangulation on the data points, remove edges longer than a threshold of 0.6 and finally, remove disconnected points. The data is then normalized by shifting to its mean and then, uniformly sampled using a grid of size 82 by 56 pixels. 3. 3D Ear Matching and Recognition The extracted 3D ear data of a subject (also called probe data) is matched with one (for authentication) or all (for identification) 3D ear data (also called gallery data) stored in the gallery database built off-line. Matching can be performed based on the error of registering between the two data sets, more specifically, two clouds of points. The ICP algorithm [2] is considered as one of the most accurate algorithms for this purpose. However, it is computationally expensive and may converge to a local minimum if the two data sets are not nearly registered. To minimize these limitations, we adopted a coarse-to-fine hierarchical technique. ICP is first applied on low and then, on high resolution meshes of the 3D ear data. The mesh reduction was performed using the surface simplification algorithm of Garland and Heckbert [5] as it preserves features on the surfaces of a mesh (unlike the sampling reduction which removes points without any regard to the features). The degree of mesh reduction can be controlled by fixing the number of triangles to be remained in the reduced mesh or by fixing the surface error approximations (using quadric matrices). We achieved better result using the later option (see Section 4.3). Initially, reduced meshes (see Figure 2) of the probe (created on-line) and the gallery (created off-line) ear data are used for coarse registration with the ICP. The rotation and translation resulting from this coarse registration are applied to the original data set and then, the ICP algorithm is applied to them to get a finer match. The matching approach which

(a) Full mesh with 55006 triangles (b) Reduced mesh with 2000 triangles (c) Reduced mesh with 400 triangles Figure 2. Sample of the full and reduced meshes of the extracted 3D ear data. we term as the two-step ICP is shown in the flowchart of figure 3. Subscripts p and q in the flowchart are used for the probe and the gallery respectively. Start Take dp=3d data of the probe, N=total number of gallery images, i=1 Create probe mesh Mp for dp Reduce mesh Mp to Mp_r with data points dp_r 4.1. Dataset Used The recognition performance of the proposed system was evaluated against two datasets collected from the UND Biometrics Database [13, 14]. Dataset A consists of arbitrarily selected 200 profile images of 100 different subjects each with the size of 640 by 480 pixels. Among the chosen images, 100 images collected in the year 2003 are used in the gallery database and another 100 images of the same subjects collected in the year 2004 are used for the probe database. No image of this dataset was used in the training of the detection classifiers. In Dataset B, the whole UND database is used excluding the images of two subjects (due to data error). 300 images taken in the year of 2003 are used as gallery. One image out of multiple images of the same subjects is arbitrarily chosen as probe. 4.2. Recognition Rate of the Single-step ICP Using Dataset A, we obtained recognition rate of 93% and 94% rank one and rank ten respectively with the singlestep ICP. Testing with Dataset B provided recognition rate of 93.98%, 94.31%, 95.31% and 96.32% rank one, two, three and ten respectively. Figure 4 shows a sample of the correct recognition. It is noticed that the system works even in the presence of partial occlusions due to hair and earrings. i>=n? Sort E in ascending order Create gallery mesh Mg_i for dg_i Reduce mesh Mg_i to Mg_i_r with data points dg_i_r Find rotation R and translation t needed for registering dp_r with dg_i_r using the ICP algorithm Calculate dp=dp R + t Find error e for registering dp with dg_i using the ICP algorithm Assign E(i)=e and i=i+1 Find dg_i corresponding to E(1) as the rank-1, that to E(2) as rank-2 and that to E(3) as rank-3 match Stop Figure 3. Flowchart of the matching algorithm using coarse-to-fine hierarchical technique with ICP. 4. Results and Discussion In this section, results of our recognition system using both two and three levels of mesh resolution are reported. The reasons for misclassifications are also discussed. (a) (b) Figure 4. Example of recognition: (a) 2D and range image of the gallery ear (b) 2D and range image of the probe ear with a small ear-ring and hair (This figure is best seen in color). 4.3. Improvement Obtained with the Two-step ICP The recognition rate with Dataset A improved to 93%, 94% and 95% for rank 1, rank 2 and rank 7 respectively when we applied the two-step ICP as mentioned in Section 3. In our experiment, we choose quadric error factor of 10 for the mesh reduction. The improvement is shown by the plots in Figure 5. It is worth mentioning that the recognition result becomes worse (84%, 88% and 88% respectively) when we fix the number of triangles to be 400 for mesh reduction. This is due to the fact that the size of the detected window is not same for all the ear images. The two-step ICP provides better results in the cases where the single-step ICP converged to a local minimum and also where the probe and the gallery images differ

96 95 Recognition Rate 94 93 (a) (b) 92 91 90 Single-step ICP Two-step ICP 1 2 3 4 5 6 7 8 9 10 Rank Figure 5. Recognition rates with Single-step and Two-step ICP. slightly in rotation and translation. An example of this improvement is shown in figure 6. (a) Figure 6. Examples of improvement with the two-step ICP which was not correctly recognized by the single-step ICP [left image is the gallery and right one is the probe]. 4.4. Analysis of the Misclassification Both single-step and two-step ICP failed to recognize some of the probe images. Visual inspection of these probe images and the corresponding gallery images indicates that the ICP algorithm cannot work properly if there is large variation in pose (rotation and translation) as shown in Figure 7-(a, b, c), occlusion by hair or ear-rings as shown in Figure 7-(a, d) or missing of data as in Figure 7-(e) is present. The pose variation might occur during the capture of the probe profile image as mentioned in section 4.1 or by the detection program (see Figure 7-(d)). 5. Conclusion Our method for ear detection and recognition from profile images is fully automatic and robust to some degrees of occlusion due to hair and ear-rings. One of its major (b) (c) (d) (e) Figure 7. Examples of misclassification:(a, b, c) 2D image of a gallery (left) and the corresponding probe (right) with pose variations, (d) 2D image of a probe with occlusions (e) Range image of a probe having missing data. strengths is that it does not assume accurate prior detection of the ear. It also does not rely on the presence of particular features like the ear pit. Thus, it is suitable for any non-intrusive biometric application. In this paper, it is also shown that the recognition performance of the ICP algorithm improves with the proposed hierarchical approach. ICP performance may be further improved by providing better initial registration. This can be done by invariant feature matching. The recognition system can be made more robust to occlusion by representing the extracted ear data with some occlusion-invariant representations. These include our future research tasks. Acknowledgements Authors acknowledge the use of the University of Notre Dame Biometrics Database for 3D ear detection and recognition. This research is sponsored by ARC grants DP0664228 and DP0881813. References [1] L. Alvarez, E. Gonzalez, and L. Mazorra. Fitting ear contour using an ovoid model. In Proc. of Int l Carnahan Conf. on Security Technology, 2005., pages 145 148, Oct. 2005. [2] P. J. Besl and N. D. McKay. A Method for Registration of 3-D Shapes. IEEE Trans. on PAMI, 14(2):239 256, 1992. [3] M. Burge and W. Burger. Ear biometrics in computer vision. In Proc. of the ICPR 00, pages 822 826, Sept 2000. [4] M. Choras. Ear biometrics based on geometrical feature extraction. Electronic Letters on Computer Vision and Image Analysis, Vol. 5:84 95, 2005. [5] M. Garland and C. Heckbert. Surface Simplification Using Quadric Error Metrics. In SIGGRAPH, 1997.

[6] D. J. Hurley, M. S. Nixon, and J. N. Carter. Force field feature extraction for ear biometrics. Computer Vision and Image Understanding, 98(3):491 512, June 2005. [7] A. Iannarelli. Ear Identification. Forensic Identification Series. Paramount Publishing Company, Fremont, California, 1989. [8] S. Islam, M. Bennamoun, and R. Davies. Fast and Fully Automatic Ear Detection Using Cascaded AdaBoost. Proc. of IEEE Workshop on Application of Computer Vision WACV 2008, Jan. 2008. [9] S. Islam, M. Bennamoun, R. Owens, and R. Davies. Biometric Approaches of 2D-3D Ear and Face: A Survey. Proc. of Int l Conf. on Systems, Computing Sciences and Software Engineering, SCSS 2007, Dec. 2007. [10] K. Messer, J. Matas, J. Kittler, J. Luettin, and G. Maitre. Xm2vtsbd: The extended m2vts database. In Proc. of the 2nd Conf. on Audio and Video-base Biometric Personal Verification, Springer Verlag, New York, pages 1 6, May 1999. [11] K. H. Pun and Y. S. Moon. Recent advances in ear biometrics. In Proc. of the Sixth IEEE Int l Conf. on Automatic Face and Gesture Recognition, pages 164 169, May 2004. [12] R. Schapire and Y. Singer. Improved boosting algorithms using confidence-rated predictions. Mach. Learn., 37(3):297 336, 1999. [13] University of Notre Dame Biometrics Database. http://www.nd.edu/ cvrl/undbiometricsdatabase.html. [14] P. Yan and K. W. Bowyer. Biometric recognition using 3d ear shape. IEEE Trans. on PAMI, 29(8):1297 1308, Aug. 2007. [15] L. Yuan and Z.-C. Mu. Ear detection based on skin-color and contour information. In Proc. of the Int l Conf. on Machine Learning and Cybernetics, Vol. 4:2213 2217, Aug. 2007. [16] T. Yuizono, Y. Wang, K. Satoh, and S. Nakayama. Study on individual recognition for ear images by using genetic local search. In Proc. of Congress on Evolutionary Computation, pages 237 242, 2002.