Face detection is a process of localizing and extracting the face region from the



Similar documents
ECE 533 Project Report Ashish Dhawan Aditi R. Ganesan

Drawing a histogram using Excel

Solving Simultaneous Equations and Matrices

Assessment. Presenter: Yupu Zhang, Guoliang Jin, Tuo Wang Computer Vision 2008 Fall

International Journal of Advanced Information in Arts, Science & Management Vol.2, No.2, December 2014


Template-based Eye and Mouth Detection for 3D Video Conferencing

HANDS-FREE PC CONTROL CONTROLLING OF MOUSE CURSOR USING EYE MOVEMENT

A Method of Caption Detection in News Video

Efficient Attendance Management: A Face Recognition Approach

The Role of Size Normalization on the Recognition Rate of Handwritten Numerals

Geometry of Vectors. 1 Cartesian Coordinates. Carlo Tomasi

Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches

3D Scanner using Line Laser. 1. Introduction. 2. Theory

Geometric Camera Parameters

Detection and Restoration of Vertical Non-linear Scratches in Digitized Film Sequences

Statistics Revision Sheet Question 6 of Paper 2

Largest Fixed-Aspect, Axis-Aligned Rectangle

Locating and Decoding EAN-13 Barcodes from Images Captured by Digital Cameras

Signature Region of Interest using Auto cropping

Optical Illusions Essay Angela Wall EMAT 6690

FACE RECOGNITION BASED ATTENDANCE MARKING SYSTEM

Arrangements And Duality

GeoGebra. 10 lessons. Gerrit Stols

Bildverarbeitung und Mustererkennung Image Processing and Pattern Recognition

Enhanced Bar Code Engine

A Study of Automatic License Plate Recognition Algorithms and Techniques

Section 1.1. Introduction to R n

Diagrams and Graphs of Statistical Data

Component Ordering in Independent Component Analysis Based on Data Power

MetaMorph Software Basic Analysis Guide The use of measurements and journals

Using Microsoft Picture Manager

Research on Chinese financial invoice recognition technology

MACHINE VISION MNEMONICS, INC. 102 Gaither Drive, Suite 4 Mount Laurel, NJ USA

1.3. DOT PRODUCT If θ is the angle (between 0 and π) between two non-zero vectors u and v,

Visual Structure Analysis of Flow Charts in Patent Images

Glencoe. correlated to SOUTH CAROLINA MATH CURRICULUM STANDARDS GRADE 6 3-3, , , 4-9

Super-resolution method based on edge feature for high resolution imaging

Reflection and Refraction

Skew Detection of Scanned Document Images

IBM SPSS Direct Marketing 23

Galaxy Morphological Classification

Multimodal Biometric Recognition Security System

Vision based Vehicle Tracking using a high angle camera

IBM SPSS Direct Marketing 22

Prentice Hall Mathematics: Course Correlated to: Arizona Academic Standards for Mathematics (Grades 6)

Big Data: Rethinking Text Visualization

Open Access A Facial Expression Recognition Algorithm Based on Local Binary Pattern and Empirical Mode Decomposition

Investigating Area Under a Curve

To determine vertical angular frequency, we need to express vertical viewing angle in terms of and. 2tan. (degree). (1 pt)

Introduction. Chapter 1

3D Viewer. user's manual _2

Copyright 2011 Casa Software Ltd. Centre of Mass

A Short Introduction to Computer Graphics

STATS8: Introduction to Biostatistics. Data Exploration. Babak Shahbaba Department of Statistics, UCI

Instructions for Creating a Poster for Arts and Humanities Research Day Using PowerPoint

Determining optimal window size for texture feature extraction methods

Installation and Operating Instructions for the FluidX IntelliCode software Rev 1, release date: July 2014

AN IMPROVED DOUBLE CODING LOCAL BINARY PATTERN ALGORITHM FOR FACE RECOGNITION

Understanding The Face Image Format Standards

Utilizing Microsoft Access Forms and Reports

A vector is a directed line segment used to represent a vector quantity.

IntelliSpace PACS 4.4. Image Enabled EMR Workflow and Enterprise Overview. Learning Objectives

Common Core Unit Summary Grades 6 to 8

Analecta Vol. 8, No. 2 ISSN

Classification of Fingerprints. Sarat C. Dass Department of Statistics & Probability

Palmprint Recognition. By Sree Rama Murthy kora Praveen Verma Yashwant Kashyap

Excel Unit 4. Data files needed to complete these exercises will be found on the S: drive>410>student>computer Technology>Excel>Unit 4

Tutorial 2: Using Excel in Data Analysis

Self-Portrait Steps Images taken from Andrew Loomis Drawing Head & Hands (there are many sites to download this out of print book for free)

Normalisation of 3D Face Data

Enhanced LIC Pencil Filter

Understand the Sketcher workbench of CATIA V5.

Line and Polygon Clipping. Foley & Van Dam, Chapter 3

SIGNATURE VERIFICATION

A PHOTOGRAMMETRIC APPRAOCH FOR AUTOMATIC TRAFFIC ASSESSMENT USING CONVENTIONAL CCTV CAMERA

Math 215 HW #6 Solutions

Automatic Labeling of Lane Markings for Autonomous Vehicles

House Design Tutorial

MEP Y9 Practice Book A

How to build text and objects in the Titler

MEMORY WORK - MATH FACTS 1

Image Interpolation by Pixel Level Data-Dependent Triangulation

How To Filter Spam Image From A Picture By Color Or Color

LINES AND PLANES CHRIS JOHNSON

AP Physics - Chapter 8 Practice Test

Determination of Acceleration due to Gravity

HISTOGRAMS, CUMULATIVE FREQUENCY AND BOX PLOTS

OPERATING P L AYA B I L I T Y P R E C I S I O N D E S I G N

Quick use reference book

SECTION 2-1: OVERVIEW SECTION 2-2: FREQUENCY DISTRIBUTIONS

Visualization of 2D Domains

Static Environment Recognition Using Omni-camera from a Moving Vehicle

Automatic Extraction of Signatures from Bank Cheques and other Documents

Algebra Geometry Glossary. 90 angle

Transcription:

Chapter 4 FACE NORMALIZATION 4.1 INTRODUCTION Face detection is a process of localizing and extracting the face region from the background. The detected face varies in rotation, brightness, size, etc. in different images even for the same person. Those features are independent of face features and will affect the recognition rate significantly. One method to solve the problem is to normalize the detected faces. The objective of the face normalization is to reduce the effect of useless, interferential and redundant information such as background, hair, cloth etc. so as to enhance the recognition process. In order to normalize the face, the basic feature points are commonly selected by using the center points of the eyes. Many eye location algorithms have been presented before [79] [80] [81] [82]. However they are either computationally expensive or unable to precisely localize the double eyeballs. As shown in Figure 4.1, we proposed an efficient algorithm for face normalization, normalizing geometry of faces so as to improve the recognition rate and real time efficiency. Algorithm assumes that the face is already detected before detecting the features such as eyes and mouth using ROI. Then the face normalization is performed. Geometric normalization is carried out by a rotation based on the eyeball s location and an affine translation defined by locations of the eyeballs. Then, the brightness is normalized by the grey information. Since the algorithm is mainly based on the histogram analysis, it can normalize the face efficiently. The experiments carried out showed that the proposed algorithm can effectively and efficiently 89

Start Feature Detection Using ROI Left Eye ball Center Mouth Center Right Eye ball Centre Orientation Normalization Scale Normalization Figure 4.1: Proposed face normalization algorithm normalize the face images. The experiments showed that after the normalization, the face recognition rate was improved by about 4 to5 percent for various methods on the ORL face image database. 4.2 FEATURE DETECTION The system of feature detection has detected three features such as left eyeball center, right eyeball center & mouth center. Each feature is detected using Region of Interests (ROI). Following are the steps required for feature detection module. 4.2.1 Eyeball Localization 90

The eyeball s position is located mainly based on the brightness histogram using pattern matching. In order to reduce the computational cost, we used the brightness histogram and bright feature to detect the eyeball s position. In this algorithm the gray image is processed directly. For simplicity, we defined the left- top point of the image as the origin point of the image coordinate frame (Figure 4.2) [83] [84]. 0 X, Selected Window h W y Figure 4.2 : The window selection for eyeball The eye s region can be roughly estimated by selecting a candidate window as region of interest. In order to cover all the eye regions the rectangle window should be big enough. Denote height and width of the face by and, respectively. The parameters of the window have been determined after extensive experimentation. Coordinates selected for left top window are (30, 12, 50, 30) i.e. height = 20 and width = 18. It can be established that in the selected window, that brightness of the eyeballs and the eyebrow is darker than other regions. The selected window as shown in Figure 4.4 (b) is displaced left to right and top to bottom taking initial displacement as zero and increase it in the step of 3 till it reaches 12. All pixels in the window are then sorted by the grey value. First value is taken as threshold and all such non- 91

repetitive thresholds are then determined. Maximum threshold is the last value. Eyeball candidate region is chosen as the 15% parts with smallest gray value., 0.15 (4.1) 0 255 (4.2) The segment threshold will be obtained by equation (4.1), where, is grey value of pixel with sorted index number as, and its coordinate is,. is the total number of the pixels in the window. Eyeballs and eyebrow from the face can be segmented. The image can be translated to binary image by setting accepted pixels with grey value 0 and setting refused pixels with value 255 as equation (4.2). Thus threshold image is obtained by setting all pixels in ROI with intensity greater than or equal to threshold and others as black as shown in Figure 4.4 (c). Complement the binary image and project the binary image window to horizontal and vertical histogram, (4.3), (4.4) We can establish the histogram as equation (4.3) and (4.4). Commonly the horizontal histogram has two minimum extremes. The lower one corresponds to the eye and the upper one corresponds to the brow. We then sort horizontal histogram in descending order. If distance between two indexes of horizontal histogram is between 4 to 10 then horizontal left index is 2 otherwise it is assigned as 1. If there are at least 4 white pixels to the left or right of this pixel then it will be a probable left eye coordinate candidate pixel. Else repeat above steps. Similarly find out left eye coordinate 92

candidate pixel. Taking median of these will give final point of left eye. Horizontal and vertical position of the right eyeballs can also be obtained similar to the method of left eyeball center. If coordinates of the center point are between the ranges given then only we will select it, otherwise finding center point procedure will be carried out again. 4.2.2 Mouth Center Localization Mouth window position is established from left and right eye center points. It s coordinates are determined as follows; Horizontal position of left eye + (Horizontal position of right eye)/2 + 25; Horizontal position of left eye + (Horizontal position of right eye) /2+60; Vertical position of left eye + (Vertical position of right eye)/2-15; Vertical position of left eye + (Vertical position of right eye)/2 + 15; Taking percentage threshold as 40, horizontal and vertical position of the mouth center is determined using same procedure as that of right eye ball center. 4.3 FACE NORMALIZATION After detection of features these faces are now normalized by geometric normalization. Then normalized database is created.these normalized faces are recognized applying PCA, LDA, EBGM, PCA and LDA taken together. This section presents major steps in the face normalization process which is to normalize face images of the same person to those in similar orientation. If vertical distance between right and left eye coordinate is between 20 and 30, then only the face is normalized. A line segment is drawn joining two center points of the detected eyeballs as shown in Figure 4.3. Slope between left and right eye ball center is 93

Figure 4.3: The rotation correction procedure when eye locations are given calculated. Angle between this line and the horizontal line is then determined. If this angle is less than 6 then normalized image is obtained by rotating image by angle in degrees. Otherwise given image is considered as normalized image. 4.4 BRIGHTNESS NORMALIZATION ALGORITHM The face images captured at different times or positions often have different brightness. These commonly affect the recognition significantly. In order to reduce the affection of the brightness, we use a brightness normalization method to let the image have the same mean and the same variance value. 1. Mean center the pixel values in the source image. 2. Linearly smooth the edges of the source image across a 30 pixel border. 3. Normalize the contrast of the new image with histogram equalization. The edge of the image is smoothed to reduce its effect in frequency space. Pixel values are weighted by 30 where 30 is the distance from the edge. Pixels greater 94

than 30 pixels from the edge are not modified. This smoothes the image edge from actual pixel values in the center of the image to zero at the edge of the image. The algorithm then performs contrast and brightness normalizations. 4.5 EXPERIMENTAL RESULTS Figure 4.4 shows some of the experimental result of left eyeball, right eyeball and mouth center detection and normalization of face image. Figure 4.5 shows result of brightness normalization. First column of Figure 4.6 shows the original images from ORL face database, and second column shows left and right eyeball center detection and mouth center localization. And the third column shows normalized face image. It is easily found that the normalization results are obviously more standard and more like each other than original groups, which can reduce the difficulty of recognition. The proposed algorithm has been implemented on ORL database consisting of 400 images to verify the results. [24]. Figure 4.4 shows the results, obtained at the major steps of the geometric normalization (G Norm) algorithm, of face images. The input images were properly selected so that the eyeballs are not easy to detect. For some images given in Figure 4.6 it was hard to see the eyeballs of the person, and hence their localization was difficult. However, the results show that the algorithm can detect their positions with an acceptable accuracy. As shown in Figure 4.6 the algorithm can correctly localize the eyeball even with the glasses. The face images were trained and recognized based on PCA, FLD, and Gabor Wavelet face recognition method. Algorithm used 1 to 10 images of a person to train and used the other remaining as well as all images for testing. The correct recognition 95

rate is given in table 4.1. The results shown in table 4.2 prove that this method can improve the recognition rate of faces. (a) (b) (c) Figure 4.4:(a:c) The result of eyeball s and mouth center localization and face normalization Figure 4.5: Result of brightness normalization. 96

Figure 4.6: Experimental results on ORL database Table 4.1: Recognition rate for various face recognition methods Training Images 1 2 3 4 5 6 7 8 9 10 Train + Test Images PCA 80 89.2 89.2 90 92 95.2 95.5 95.7 96.7 97 PCA & G Norm 80.5 88.7 89.2 90.2 93.2 95.2 95.7 95.7 96 96 FLD -- 97.75 98.25 99.5 99 95.5 92.75 92.25 93 94.75 FLD & G Norm Gabor Wavelet Gabor Wavelet & B Norm -- 90.75 89 94 93.5 95.5 94.75 97 98.75 99.5 54.5 67.75 76.25 82.25 86.75 89.25 92.25 95 97.25 98.5 52.75 67.25 79.75 86.25 89.25 92.25 93.75 95.25 96.25 97.25 97

4.6 CONCLUSION In this work effective and efficient feature detection method based on eyes location has been proposed. We presented results of performance test of face recognition algorithms on the ORL database. The well known PCA algorithm is first tested as a reference performance. Then FLD and Gabor Wavelet algorithm were tested. All the three algorithms were tested with and without normalization. Highest recognition rate of 99.5% was found for FLD along with G Norm. It has been found that as the number of training images is increased recognition rate increases. Lowest recognition rate of 52.75% is observed for Gabor Wavelet along with B Norm algorithm when there is only one training image sample. With only one training sample this result is 80.5 % in case of PCA. This may be because of the fact that for PCA whole image is taken for face recognition whereas in case of Gabor wavelet only 5 feature points are considered. Results shown in table 4.1 demonstrate that normalization method can effectively detect the feature such as eyeball center with and without spectacles and normalize the face images by geometric as well as brightness normalization and improve recognition rate. 98