Comparison And Combination Of Ear And Face Images In .

3y ago
19 Views
3 Downloads
1.29 MB
6 Pages
Last View : Today
Last Download : 3m ago
Upload by : Cannon Runnels
Transcription

1160IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE,Short PapersKyong Chang, Kevin W. Bowyer, Fellow, IEEE,Sudeep Sarkar, Member, IEEE, andBarnabas VictorAbstract—Researchers have suggested that the ear may have advantages overthe face for biometric recognition. Our previous experiments with ear and facerecognition, using the standard principal component analysis approach, showedlower recognition performance using ear images. We report results of similarexperiments on larger data sets that are more rigorously controlled for relativequality of face and ear images. We find that recognition performance is notsignificantly different between the face and the ear, for example, 70.5 percentversus 71.6 percent, respectively, in one experiment. We also find that multimodalrecognition using both the ear and face results in statistically significantimprovement over either individual biometric, for example, 90.9 percent in theanalogous experiment.Index Terms—Biometrics, multimodal biometrics, face recognition, earrecognition, appearance-based recognition, principal component analysis.æINTRODUCTIONWHILE good face recognition performance has been reported undercertain conditions, there is still a great need for better performance inbiometrics appropriate for use in video surveillance. Possibleavenues for improved performance include the use of a differentsource of biometric information, and/or the combination ofinformation from multiple sources. One other possible biometricsource is the ear. Iannarelli performed important early research on amanual approach to using the ear for human identification [1].Recent works that explore computer vision techniques for earbiometrics include those of Burge and Burger [2] and Hurley et al.[3]. In particular, Burge and Burger assert that the ear offers thepromise of similar performance to the face:Facial biometrics fail due to the changes in features caused byexpressions, cosmetics, hair styles, and the growth of facial hair aswell as the difficulty of reliably extracting them in an unconstrainedenvironment exhibiting imaging problems such as lighting andshadowing.Therefore, we propose a new class of biometrics forpassive identification based upon ears which have both reliable androbust features which are extractable from a distance.identificationby ear biometrics is promising because it is passive like facerecognition, but instead of the difficult to extract face biometrics,robust and simply extracted biometrics like those in fingerprints canbe used. ([2], p. 275)In the context of Iannarelli’s earlier work and the current popularityof face recognition research, this assertion that the ear could offerimproved biometric performance relative to the face deserves carefulevaluation. The experiments reported in this paper are aimed at1) testing the hypothesis that images of the ear provide betterbiometric performance than images of the face and 2) exploringwhether a combination of ear and face images may provide betterperformance than either one individually. The results reported here. K. Chang and K.W. Bowyer are with the Department of Computer Scienceand Engineering, University of Notre Dame, Notre Dame, IN 46556.E-mail: {kchang, kwb}@cse.nd.edu. S. Sarkar and B. Victor are with the Department of Computer Science andEngineering, University of South Florida, Tampa, FL 33620.E-mail: {sarkar, bvictor}@csee.usf.edu.Manuscript received 20 June 2002; revised 13 Dec. 2002; accepted 16 Feb. 2003.Recommended for acceptance by M. Pietikainen.For information on obtaining reprints of this article, please send e-mail to:tpami@computer.org, and reference IEEECS Log Number 116812.0162-8828/03/ 17.00 ß 2003 IEEENO. 9,SEPTEMBER 2003Comparison and Combination of Ear and FaceImages in Appearance-Based Biometrics1VOL. 25,Published by the IEEE Computer Societyfollow up on those reported in an earlier study [4]. Using larger datasets and more rigorous assurance of similar relative quality in the earand face images, we obtain somewhat different results than in theearlier study. In the experiments reported here, recognitionperformance is essentially identical using ear images or face imagesand combining the two for multimodal recognition results in astatistically significant performance improvement. For example, inone experiment the rank-one recognition rates for face and ear were70.5 percent and 71.6 percent, respectively, whereas the corresponding multimodal recognition rate was 90.9 percent. To our knowledge,ours is the only work to present any experimental results ofcomputer algorithms for biometric recognition based on the ear.2“EIGEN-FACES” AND “EIGEN-EARS”Extensive work has been done on face recognition algorithmsbased on principal component analysis (PCA), popularly known as“eigenfaces” [5]. The FERET evaluation protocol [6] is the de factostandard in evaluation of face recognition algorithms, andcurrently uses PCA-based recognition performance as a baseline.A standard implementation of the PCA-based algorithm [7] is usedin the experiments reported here. This implementation requires thelocation of two landmark points for image registration. For the faceimages, the landmark points are the centers of the eyes. Manuallyidentified eye center coordinates are supplied with the face imagesin the Human ID database. For the ear images, the manuallyidentified coordinates of the triangular fossa and the antitragus [1]are used. See Fig. 1 for an illustration of the landmark points.The PCA-based approach begins with using a set of trainingimages to create a “face space” or “ear space.” First, the landmarkpoints are identified and used to crop the image to a standard sizelocated around the landmark points. In our experiments, originalface images are cropped to 768 1; 024 and original ear images to400 500. In these images, one pixel covers essentially the same sizearea on the face or the ear. Next, the cropped images are normalizedto the 130 150 size used by the PCA software. At this point, onepixel in an ear image represents a finer-grain metric area than in aface image. The normalized images are masked to “gray out” thebackground and leave only the face or ear, respectively. The faceimages use the mask that comes with the standard implementation[7]. For the ear images, we experimented with several differentlevels of masking in order to tune this algorithm parameter for goodperformance. Last, the image is histogram equalized. The eigenvalues and eigenvectors are computed for the set of training images,and a “face space” or “ear space” is selected based on theeigenvectors associated with the largest eigenvalues. Followingthe FERET approach, we use the eigenvectors corresponding to thefirst 60 percent of the large eigenvalues and drop the firsteigenvector as it typically represents illumination variation [6]. Thisapproach uses the same dimension of face space and ear space, 117,in this case (Table 1). Another approach is to use whatever numberof eigenvectors accounts for some fixed percent of the total variation,resulting in a different dimension of face space and ear space. Whichof these approaches is used does not substantially affect ourconclusions, as is shown later in the paper.The set of training images consists of data for 197 subjects, each ofwhom had both a face image and an ear image taken under the sameconditions at the same image acquisition session. These images wereacquired at the University of South Florida (USF) between August2000 and November 2001. A subject’s images were dropped from ourstudy if either the face or ear was substantially obscured by hair, if thesubject wore an earring or analogous face jewelry or if either imagehad technical problems. Some of the gallery and probe images for thefirst experiment were acquired at USF during the same time frame.Additional gallery and probe images for the first experiment, and allgallery and probe images for the second and third experiments, wereacquired at the University of Notre Dame in November 2002.There is a separate (gallery, probe) data set for each of threeexperiments. The gallery images represent the “watch list,” that is,

IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE,VOL. 25, NO. 9,SEPTEMBER 20031161TABLE 1A Number of Eigenvectors Used to Create the EigenspaceFig. 1. Illustration of points used for geometric normalization of face and earimages. The triangular fossa is the upper point on the ear image and the antitragusis the lower point.the people who are enrolled in the system to be recognized. A probeimage is an image given to the system to be matched against thegallery. Each of the three experiments represents a single factorbeing varied in a consistent way between the gallery and probe. Forthe day variation experiment, 88 subjects had both an ear and a faceimage taken under the same conditions in one acquisition sessionand then another ear and face image taken under the sameconditions on a different day. The face images are the standardFERET “FA ” (“normal expression”) images [6]. The ear images are ofthe right ear. For each subject, the earlier image is used as the galleryimage and the later image is used as the probe image. Thisexperiment looks at the recognition rate when gallery and probeimages of a subject are obtained on different days, but under similarconditions of pose and lighting.For the lighting variation experiment, 111 subjects had an earand a face image taken under the same conditions in one sessionand then another face and ear image taken in the same session, butunder a different lighting condition. The standard lighting usestwo side spotlights and one above-center spotlight and the alteredlighting uses just the above-center spotlight. The images takenunder the standard lighting are gallery images and the imagestaken under altered lighting are probe images. This experimentlooks at the recognition rate when gallery and probe images of asubject are obtained in the same session and with similar pose, butunder distinctly different lighting.Fig. 2. An example of the gallery and probe face and ear images used in this study.For the pose variation experiment, 101 subjects had both an earand a face image taken under the same conditions in oneacquisition session and then another face and ear image taken at22.5 degree rotation in the same acquisition session. The imagestaken from a straight-on view are the gallery set, and the imagestaken at a 22.5 degree rotation are the probe set. This experimentlooks at the recognition rate when gallery and probe images of asubject are obtained in the same session and with the samelighting, but with a different pose. An example of the gallery anddifferent probe conditions for one subject appear in Fig. 2.Not all subjects attended all acquisition sessions and somesubjects were dropped from some experiments after image qualitycontrol checks and, so, the three experiments have different numbersof subjects. The same standard face and ear images of some subjectsmay appear in the gallery set for each of the three experiments.However, since the probe sets are the changed conditions, there areno images in common across the three probe sets.3EXPERIMENTAL RESULTS: FACE VERSUS EARThe null hypothesis for these experiments is that there is nosignificant difference in performance between using the face or theear as a biometric, given 1) use of the same PCA-based algorithmimplementation, 2) the same subject pool represented in both thegallery and probe sets, and 3) controlled variation in one parameterof image acquisition between the gallery and probe images. Therecognition experiment is to compute the cumulative matchcharacteristic (CMC) curve for the gallery and probe set and toconsider the statistical significance of the difference in rank-onerecognition rates.The baseline is the day variation experiment. This experimentlooks at the recognition performance for gallery and probe imagestaken under the same conditions but on different days. The

1162IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE,VOL. 25,NO. 9,SEPTEMBER 2003Fig. 3. Recognition performance comparison between face and ear. (a) Face and ear recognition performance in the day variation experiment. (b) Face and earrecognition performance in the lighting variation experiment. (c) Face and ear recognition performance in the pose variation experiment.CMC curves for face and ear recognition are shown in Fig. 3. TheCMC curves are computed in two ways. One uses the 197-imagetraining set that has no subjects in common with the gallery andprobe sets. The other uses the gallery set as the training set. There isno substantial difference in the results between the two trainingmethods. Numbers reported for statistical significance tests aretaken from the results using the 197-image training set. Note that theear and face performance represented in the CMC curves is quitesimilar, with the curves actually crossing at some point. The rankone recognition rates of 70.5 percent for face and 71.6 percent for ear

IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE,VOL. 25, NO. 9,SEPTEMBER 20031163Fig. 4. Recognition performance of face, ear, and combined face-ear. (a) Face combined with ear recognition performance in the day variation experiment. (b) Facecombined with ear recognition performance in the lighting variation experiment. (c) Face combined with ear recognition performance in the pose variation experiment.are not statistically significantly different at the 0:05 level using aMcNemar test [8].Relative to the baseline experiment, the lighting variationexperiment looks at how a lighting change between the galleryimage and the probe image affects the recognition rate. Performancefor either the face or the ear is slightly lower than in the baselineexperiment. Similar to the baseline experiment, there is relativelylittle difference between the CMC curves for the face and the ear,especially at lower ranks. The rank-one recognition rates of

1164IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE,VOL. 25,NO. 9,SEPTEMBER 2003Fig. 5. Performance based on different selection of eigenvectors in face and ear spaces.64.9 percent for face and 68.5 percent for ear are not statisticallysignificantly different at the 0:05 level using a McNemar test.Relative to the baseline experiment, the pose variation experiment looks at how a 22.5 degree rotation to the left between thegallery and the probe images affects the recognition rate.Performance, in this case, is much lower than for either the baselineor the lighting change experiment. There also appears to be a largergap between face and ear performance than in the other twoexperiments, but still the difference is not statistically significant. Inany case, performance at this low of a level is not likely to bepractically meaningful.Overall, the results of our experiments do not provide anysignificant evidence for rejecting the null hypothesis that the faceand the ear have equal potential as the source for appearancebased biometric recognition. Of course, there may still be somebiometric algorithm, other than PCA, for which one of the face orthe ear offers significantly better recognition performance than theother. Also, there may be particular application scenarios in whichit is not practical to acquire ear and face images that meet similarquality control conditions. For example, in an outdoor sportscontext many people may wear sunglasses or in a formal indoorevent many people may wear earrings.4EXPERIMENTAL RESULTS: FACE PLUS EARMULTIMODAL BIOMETRICAnother experiment was performed to investigate the value of amultimodal biometric using the face and ear images. A very simplecombination technique is used. The normalized, masked ear and faceimages of a subject are concatenated to form a combined face-plusear image. This was done with the data from each of the threeexperiments and Fig. 4 shows the resulting CMC curves. TheCMC curves for the day variation and lighting variation experimentssuggest that the multimodal biometric offers substantial performance gain. The difference in the rank-one recognition rates for theday variation experiment using the 197-image training sets is90.9 percent for the multimodal biometric versus 71.6 percent forthe ear and 70.5 percent for the face. A McNemar’s test forsignificance of the difference in accuracy in the rank-one matchbetween the multimodal biometric and either the ear or the face aloneshows that multimodal performance is significantly greater at the0:05 level. Of the 88 probes, the multimodal and the ear are correct on62, both incorrect on 6, multimodal only is correct on 18, and ear onlyis correct on 2. The difference between the multimodal biometric andeither the face or the ear alone is again statistically significant in thelighting change experiment, 87.4 percent rank-one recognition rateversus 64.9 percent or 68.5 percent, for the face or ear, respectively.However, because the overall performance is so low, the difference inthe pose change experiment is not statistically significant. Theseresults suggest that it is worthwhile to explore the combination ofmultiple biometric sources that could be acquired in a surveillancescenario.5DISCUSSIONOverall, our experimental results suggest that the ear and the facemay have similar value for biometric recognition. Our results do notsupport a conclusion that an ear-based or face-based biometricshould necessarily offer better performance than the other. Ofcourse, this is not the same as proving that there is no usefulbiometric algorithm for which one would offer better performance.Research into new algorithms that take advantage of specificfeatures of the ear or the face may produce improved performanceusing one or the other.Our results do support the conclusion that a multimodalbiometric using both the ear and the face can out-perform abiometric using either one alone. There is substantial related workin multimodal biometrics. For example, Hong and Jain [9] usedface and fingerprint in multimodal biometric identification, andVerlinde et al. [10] used face and voice. However, use of the faceand ear in combination seems more relevant to surveillanceapplications. We are aware of just one other work specifically onmultimodal biometrics appropriate to surveillance, this one usingface and gait [11]. This would seem to be an especially rich andpromising area of research. It might be expanded to include otherbiometric sources, such as face, ear, and gait. It might also beexpanded to investigate more sophisticated methods of combiningevidence from the different biometrics.The results presented so far are based on using the same fixednumber of eigenvectors for both the face and ear space. It is alsopossible to create the spaces based on the same percent of energy,allowing the number of eigenvectors to vary as appropriate.CMC curves computed using both spaces for the day variationexperiment appear in Fig. 5. Performance is essentially the samewhether the spaces are created based on a fixed number ofeigenvectors in this case, or a floating number of eigenvectorscorresponding to a fixed percent of total energy. (The authors wouldlike to thank the anonymous reviewer who suggested inclusion ofthis comparison.)The PCA-based face recognition approach has been informallytuned through use over time and, inevitably, an accumulation ofexpertise is embedded in the standard implementation [7]. Severaloptions were explored in an attempt to ensure that the use of thePCA approach was appropriately tuned for use with ear images.For example, five different levels of masking for the ear imageswere tried. Also, a total of four landmark points were marked oneach ear image and experiments were run with a different pair oflandmark points. The results reported here are for the best level ofmasking and pair of landmark points.

IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE,VOL

the face for biometric recognition. Our previous experiments with ear and face recognition, using the standard principal component analysis approach, showed lower recognition performance using ear images. We report results of similar experiments on larger data sets that are more rigorously controlled for relative quality of face and ear images.

Related Documents:

Furuncle of external ear. H60.00 Abscess of external ear, unspecified ear. H60.01 Abscess of right external ear. H60.02 Abscess of left external ear. H60.03 Abscess of external ear, bilateral. H60.1 Cellulitis of external ear. Cellulitis of auricle Cellulitis of external auditory canal. H60.10 Cellulitis of external ear, unspecified ear

Care of Ear Cleaning Equipment 7. Ear Cleaning Equipment Single-use [(e.g. Tips (Elephant Ear Washer tips , Otoclear tips)], Syringes) Multi-patient reusable Patient owned or dedicated to a single patient (e.g. Elephant Ear Washer or Rhino Ear Washer ) (e.g. Elephant Ear Washer , Rhi

H53.10 Unspecified subjective visual disturbances H53.1 1 Day blindness Ear Diseases H61 .20 Impacfed cerumen, unspecified ear H61 .2t Impacted cerumen, right ear H6t 22 Impacted cerumen left ear H61 23 Impacted cerumen bilatera H60.00 Abscess of external ear, unspecified ear H60.01 Abscess of right external ear

whole middle ear space is the ear drum and the tympanic membrane the drum skin. 2.2.2. The Middle Ear The middle ear is an air filled space connected to the back of the nose by a long, thin tube called the Eustachian tube. The middle ear space houses three little bones, the hammer, anvil and

1. The ear is the sense organ that is sensitive to sound stimuli produced by vibrating object. 2. A human ear has three main part. i. the outer ear, filled with air. ii. The middle ear, filled with air. iii. The inner ear, filled with liquid Draw figure 1.15, page 14 3. Function of the different parts of the human ear. Salin table 1.4, page 15

Massachusetts Eye and Ear (Mass. Eye and Ear) is a specialty hospital dedicated to excellence in the care of disorders that affect the eye, ear, nose, throat, and adjacent regions of the head and neck. Mass. Eye and Ear also provides primary care and serves as a referral center for outpatient and inpatient medical and surgical care.

problems in the middle ear. It’s possible for infection to spread from the nose and throat area through the Eustachian tube to the middle ear, which is one of the causes of middle ear infections. The third part of the auditory system is the inner ear. The inner ear has two sections: one that is responsible for balance and the other for hearing.

Albert Woodfox were properly convicted for the 1972 murder of prison guard Brent Miller. Supporters of Wallace and Woodfox, who make up two-thirds of a group known to supporters as the "Angola Three," say that the convictions were at least partly because of the men's involvement with the Black Panther Party. "Under this new governor's office, this new day, we are making sure we right the .