![人脸识别文献翻译中英文[共15页]_第1页](http://file2.renrendoc.com/fileroot_temp3/2021-6/7/c8ef08ed-6bb3-4c6b-8c8d-c8b8e072b331/c8ef08ed-6bb3-4c6b-8c8d-c8b8e072b3311.gif)
![人脸识别文献翻译中英文[共15页]_第2页](http://file2.renrendoc.com/fileroot_temp3/2021-6/7/c8ef08ed-6bb3-4c6b-8c8d-c8b8e072b331/c8ef08ed-6bb3-4c6b-8c8d-c8b8e072b3312.gif)
![人脸识别文献翻译中英文[共15页]_第3页](http://file2.renrendoc.com/fileroot_temp3/2021-6/7/c8ef08ed-6bb3-4c6b-8c8d-c8b8e072b331/c8ef08ed-6bb3-4c6b-8c8d-c8b8e072b3313.gif)
![人脸识别文献翻译中英文[共15页]_第4页](http://file2.renrendoc.com/fileroot_temp3/2021-6/7/c8ef08ed-6bb3-4c6b-8c8d-c8b8e072b331/c8ef08ed-6bb3-4c6b-8c8d-c8b8e072b3314.gif)
![人脸识别文献翻译中英文[共15页]_第5页](http://file2.renrendoc.com/fileroot_temp3/2021-6/7/c8ef08ed-6bb3-4c6b-8c8d-c8b8e072b331/c8ef08ed-6bb3-4c6b-8c8d-c8b8e072b3315.gif)
版权说明:本文档由用户提供并上传,收益归属内容提供方,若内容存在侵权,请进行举报或认领
文档简介
1、附录( 原文及译文 )翻译原文来自Thomas David Heseltine BSc. Hons. The University of YorkDepartment of Computer ScienceFor the Qualification of PhD. - September 2005 -Face Recognition: Two-Dimensional and Three-Dimensional Techniques4 Two-dimensional Face Recognition4.1 Feature LocalizationBefore discussing the met
2、hods of comparing two facial images we now take a brief look atsome at the preliminary processes of facial feature alignment. This process typically consists oftwo stages: face detection and eye localisation. Depending on the application, if the position ofthe face within the image is known beforeha
3、nd (for a cooperative subject in a door access systemfor example) then the face detection stage can often be skipped, as the region of interest isalready known. Therefore, we discuss eye localisation here, with a brief discussion of facedetection in the literature review(section 3.1.1).The eye local
4、isation method is used to align the 2D face images of the various test sets usedthroughout this section. However, to ensure that all results presented arerepresentative of the face recognition accuracy and not a product of the performance of the eyelocalisation routine, all image alignments are manu
5、ally checked and any errors corrected, prior totesting and evaluation.We detect the position of the eyes within an image using a simple template basedmethod. A training set of manually pre-aligned images of faces is taken, and eachimage cropped to an area around both eyes. The average image is calcu
6、lated and usedas a template.Figure 4-1 - The average eyes. Used as a template for eye detection.Both eyes are included in a single template, rather than individually searching for each eyein turn, as the characteristic symmetry of the eyes either side of the nose, provides a usefulfeature that helps
7、 distinguish between the eyes and other false positives that may be picked up inthe background. Although this method is highly susceptible to scale(i.e. subject distance from thecamera) and also introduces the assumption that eyes in the image appear near horizontal. Somepreliminary experimentation
8、also reveals that it is advantageous to include the area of skin justbeneath the eyes. The reason being that in some cases the eyebrows can closely match thetemplate, particularly if there are shadows in the eye-sockets, but the area of skin below the eyeshelps to distinguish the eyes from eyebrows
9、(the area just below the eyebrows contain eyes,whereas the area below the eyes contains only plain skin).A window is passed over the test images and the absolute difference taken to that of the averageeye image shown above. The area of the image with the lowest difference is taken as the regionof in
10、terest containing the eyes. Applying the same procedure using a smaller template of theindividual left and right eyes then refines each eye position.This basic template-based method of eye localisation, although providing fairlypreciselocalisations, often fails to locate the eyes completely. However
11、, we are able toimprove performance by including a weighting scheme.Eye localisation is performed on the set of training images, which is then separated into twosets: those in which eye detection was successful; and those in which eye detection failed.Taking the set of successful localisations we co
12、mpute the average distance from the eye template(Figure 4-2 top). Note that the image is quite dark, indicating that the detected eyes correlateclosely to the eye template, as we would expect. However, bright points do occur near the whitesof the eye, suggesting that this area is often inconsistent,
13、 varying greatly from the average eyetemplate.Figure 4-2 Distance to the eye template for successful detections (top) indicating variancedue tonoise and failed detections (bottom) showing credible variance due to miss-detectedfeatures.In the lower image (Figure 4-2 bottom), we have taken the set of
14、failed localisations(imagesof the forehead, nose, cheeks, background etc. falsely detected by the localisation routine) andonce again computed the average distance from the eye template. The bright pupils surroundedby darker areas indicate that a failed match is often due to the high correlation of
15、the nose andcheekbone regions overwhelming the poorly correlated pupils. Wanting to emphasise thedifference of the pupil regions for these failed matches and minimise the variance of the whitesof the eyes for successful matches, we divide the lower image values by the upper image toproduce a weights
16、 vector as shown in Figure 4-3. When applied to the difference image beforesumming a total error, this weighting scheme provides a much improved detection rate.Figure 4-3 - Eye template weights used to give higher priority to those pixels that bestrepresent the eyes.4.2 The Direct Correlation Approa
17、chWe begin our investigation into face recognition with perhaps the simplest approach,knownas the direct correlation method (also referred to as template matching by Brunelli and Poggio 29 ) involving the direct comparison of pixel intensity values taken from facial images. We usethe term Direct Cor
18、relation to encompass all techniques in which face images are compareddirectly, without any form of image space analysis, weighting schemes or feature extraction,regardless of the dsi tance metric used. Therefore, we do not infer that Pearson s correlatiapplied as the similarity function (although s
19、uch an approach would obviously come under ourdefinition of direct correlation). We typically use the Euclidean distance as our metric in theseinvestigations (inversely related to Pearson s correlation and can be considered as a scaletranslation sensitive form of image correlation), as this persists
20、 with the contrast made betweenimage space and subspace approaches in later sections.Firstly, all facial images must be aligned such that the eye centres are located at twospecified pixel coordinates and the image cropped to remove any backgroundinformation. These images are stored as greyscale bitm
21、aps of 65 by 82 pixels and prior torecognition converted into a vector of 5330 elements (each element containing the correspondingpixel intensity value). Each corresponding vector can be thought of as describing a point within a5330 dimensional image space. This simple principle can easily be extend
22、ed to much largerimages: a 256 by 256 pixel image occupies a single point in 65,536-dimensional image spaceand again, similar images occupy close points within that space. Likewise, similar faces arelocated close together within the image space, while dissimilar faces are spaced far apart.Calculatin
23、g the Euclidean distanced, between two facial image vectors (often referred to as thequery image q, and gallery imageg), we get an indication of similarity. A threshold is thenapplied to make the final verification decision.d q g (d threshold ? accept) (d threshold ? reject ) . Equ. 4-14.2.1 Verific
24、ation TestsThe primary concern in any face recognition system is its ability to correctly verify aclaimed identity or determine a persons most likely identity from a set of potential matches in adatabase. In order to assess a given system s ability to perform these tasks, a variety ofevaluation meth
25、odologies have arisen. Some of these analysis methods simulate a specific modeof operation (i.e. secure site access or surveillance), while others provide a more mathematicaldescription of data distribution in someclassification space. In addition, the results generated from each analysis method may
26、be presented in a variety of formats. Throughout the experimentations in this thesis, we primarilyuse the verification test as our method of analysis and comparison, although we also use FisherLinear Discriminant to analyse individual subspace components in section 7 and theidentification test for t
27、he final evaluations described in section 8. The verification test measures asystem s ability to correctly accept or reject the proposed identity of an individual. At afunctional level, this reduces to two images being presented for comparison, for which thesystem must return either an acceptance (t
28、he two images are of the same person) or rejection (thetwo images are of different people). The test is designed to simulate the application area ofsecure site access. In this scenario, a subject will present some form of identification at a point ofentry, perhaps as a swipe card, proximity chip or
29、PIN number. This number is then used toretrieve a stored image from a database of known subjects (often referred to as the target orgallery image) and compared with a live image captured at the point of entry (the query image).Access is then granted depending on the acceptance/rejection decision.The
30、 results of the test are calculated according to how many times the accept/reject decisionis made correctly. In order to execute this test we must first define our test set of face images.Although the number of images in the test set does not affect the results produced (as the errorrates are specif
31、ied as percentages of image comparisons), it is important to ensure that the test setis sufficiently large such that statistical anomalies become insignificant (for example, a couple ofbadly aligned images matching well). Also, the type of images (high variation in lighting, partialocclusions etc.)
32、will significantly alter the results of the test. Therefore, in order to comparemultiple face recognition systems, they must be applied to the same test set.However, it should also be noted that if the results are to be representative of systemperformance in a real world situation, then the test dat
33、a should be captured under precisely thesame circumstances as in the application environment.On the other hand, if the purpose of theexperimentation is to evaluate and improve a method of face recognition, which may be appliedto a range of application environments, then the test data should present
34、the range of difficultiesthat are to be overcome. This may mean including a greater percentage of difficultwould be expected in the perceived operating conditions and hence higher error rates in theresults produced. Below we provide the algorithm for executing the verification test. Thealgorithm is
35、applied to a single test set of face images, using a single function call to the facerecognition algorithm: CompareFaces(FaceA, FaceB). This call is used to compare two facialimages, returning a distance score indicating how dissimilar the two face images are: the lowerthe score the more similar the
36、 two face images. Ideally, images of the same face should producelow scores, while images of different faces should produce high scores.Every image is compared with every other image, no image is compared with itself and nopair is compared more than once (we assume that the relationship is symmetric
37、al). Once twoimages have been compared, producing a similarity score, the ground-truth is used to determineif the images are of the same person or different people. In practical tests this information isoften encapsulated as part of the image (by means of a unique person identifier). Scores arethen
38、stored in one of two lists: a list containing scores produced by comparing images ofdifferent people and a list containing scores produced by comparing images of the same person.The final acceptance/rejection decision is made by application of a threshold. Any incorrectdecision is recorded as either
39、 a false acceptance or false rejection. The false rejection rate (FRR)is calculated as the percentage of scores from the same people that were classified as rejections.The false acceptance rate (FAR) is calculated as the percentage of scores from different peoplethat were classified as acceptances.F
40、or IndexA = 0 to length(TestSet)For IndexB = IndexA+1 to length(TestSet)Score = CompareFaces(TestSetIndexA, TestSetIndexB)If IndexA and IndexB are the same personAppend Score to AcceptScoresListElseAppend Score to RejectScoresListFor Threshold = Minimum Score to Maximum Score:FalseAcceptCount, False
41、RejectCount = 0For each Score in RejectScoresListIf Score ThresholdIncrease FalseRejectCountFalseAcceptRate = FalseAcceptCount / Length(AcceptScoresList)FalseRejectRate = FalseRejectCount / length(RejectScoresList)Add plot to error curve at (FalseRejectRate, FalseAcceptRate)These two error rates exp
42、ress the inadequacies of the system when operating at aspecific threshold value. Ideally, both these figures should be zero, but in reality reducing eitherthe FAR or FRR (by altering the threshold value) will inevitably resultin increasing the other. Therefore, in order to describe the full operatin
43、g range of aparticular system, we vary the threshold value through the entire range of scoresproduced. The application of each threshold value produces an additional FAR, FRRpair, which when plotted on a graph produces the error rate curve shown below.Figure 4-5 - Example Error Rate Curve produced b
44、y the verification test.The equal error rate (EER) can be seen as the point at which FAR is equal to FRR. This EERvalue is often used as a single figure representing the general recognitionperformance of a biometric system and allows for easy visual comparison of multiplemethods. However, it is impo
45、rtant to note that the EER does not indicate the level oferror that would be expected in a real world application. It is unlikely that any realsystem would use a threshold value such that the percentage of false acceptances wereequal to the percentage of false rejections. Secure site access systems
46、would typicallyset the threshold such that false acceptances were significantly lower than false rejections:unwilling to tolerate intruders at the cost of inconvenient access denials.Surveillance systems on the other hand would require low false rejection rates tosuccessfully identify people in a le
47、ss controlled environment. Therefore we should bear in mindthat a system with a lower EER might not necessarily be the better performer towards theextremes of its operating capability.There is a strong connection between the above graph and the receiver operatingcharacteristic (ROC) curves, also use
48、d in such experiments. Both graphs are simply twovisualisations of the same results, in that the ROC format uses the True Acceptance Rate(TAR),where TAR = 1.0 FRR in place of the FRR, effectively flipping the graph vertically. Anothervisualisation of the verification test results is to display both
49、the FRR and FAR as functions ofthe threshold value. This presentation format provides a reference to determine the thresholdvalue necessary to achieve a specific FRR and FAR. The EER can be seen as the point where thetwo curves intersect.Figure 4-6 - Example error rate curve as a function of the sco
50、re thresholdThe fluctuation of these error curves due to noise and other errors is dependant on thenumber of face image comparisons made to generate the data. A small dataset that only allowsfor a small number of comparisons will results in a jagged curve, in which large steps correspondto the influ
51、ence of a single image on a high proportion of thecomparisons made. A typical dataset of 720 images (as used in section 4.2.2) provides258,840 verification operations, hence a drop of 1% EER represents an additional 2588 correctdecisions, whereas the quality of a single image could cause the EER tof
52、luctuate by up to 0.28.4.2.2 ResultsAs a simple experiment to test the direct correlation method, we apply the techniquedescribed above to a test set of 720 images of 60 different people, taken from the AR FaceDatabase 39 . Every image is compared with every other image in the test set to produce al
53、ikeness score, providing 258,840 verification operations from which to calculate falseacceptance rates and false rejection rates. The error curve produced is shown in Figure 4-7.Figure 4-7 - Error rate curve produced by the direct correlation method using no imagepreprocessing.We see that an EER of
54、25.1% is produced, meaning that at the EER thresholdapproximately one quarter of all verification operations carried out resulted in anincorrect classification. There are a number of well-known reasons for this poor levelof accuracy. Tiny changes in lighting, expression or head orientation cause the
55、 location in imagespace to change dramatically. Images in face space are moved far apart due to these imagecapture conditions, despite being of the same person s face. The distance between imagesdifferent people becomes smaller than the area of face space covered by images of the sameperson and henc
56、e false acceptances and false rejections occur frequently. Other disadvantagesinclude the large amount of storage necessary for holding many face images and the intensiveprocessing required for each comparison, making this method unsuitable for applications appliedto a large database. In section 4.3
57、 we explore the eigenface method, which attempts to addresssome of these issues.4 二维人脸识别4.1 功能定位在讨论比较两个人脸图像,我们现在就简要介绍的方法一些在人脸特征的初步调整过程。这一过程通常两个阶段组成:人脸检测和眼睛定位。根据不同的申请时,如果在面部图像的立场是众所周知事先(对于合作的主题,例如在门禁系统) ,那么人脸检测阶段通常可以跳过,作为地区的利益是已知的。因此,我们讨论眼睛定位在这里,有一个人脸检测的文献简短讨论 (第 3.1.1 )。眼睛定位方法用于对齐的各种测试二维人脸图像集通篇使用这一节。但是,为了确保所有的结果都呈现代表面部识别准确率,而不是对产品的性能例行的眼睛定位,所有图像路线是手动检查,若有错误更正前的测试和评价。我们发现在一个使用图像的眼睛一个简单的基于模板的位置方法。训练集的前脸手动对齐图像是采取和各图片进行裁剪,以两只眼睛周围的地区。平均计算,用形象作为一个模板。图 4-1 - 平均眼睛,用作模板的眼睛检测两个眼睛都包括在一个模板,而不是单独为每个搜索,因为眼睛的任一鼻子两边对称的特点,提供了一个有用的功能,可以帮助区分眼睛和其他可能误报被拾起的背景。虽然这种方法在介绍了假设眼近水平的形象出现后很容易受到
温馨提示
- 1. 本站所有资源如无特殊说明,都需要本地电脑安装OFFICE2007和PDF阅读器。图纸软件为CAD,CAXA,PROE,UG,SolidWorks等.压缩文件请下载最新的WinRAR软件解压。
- 2. 本站的文档不包含任何第三方提供的附件图纸等,如果需要附件,请联系上传者。文件的所有权益归上传用户所有。
- 3. 本站RAR压缩包中若带图纸,网页内容里面会有图纸预览,若没有图纸预览就没有图纸。
- 4. 未经权益所有人同意不得将文件中的内容挪作商业或盈利用途。
- 5. 人人文库网仅提供信息存储空间,仅对用户上传内容的表现方式做保护处理,对用户上传分享的文档内容本身不做任何修改或编辑,并不能对任何下载内容负责。
- 6. 下载文件中如有侵权或不适当内容,请与我们联系,我们立即纠正。
- 7. 本站不保证下载资源的准确性、安全性和完整性, 同时也不承担用户因使用这些下载资源对自己和他人造成任何形式的伤害或损失。
最新文档
- 《餐饮服务与管理》课件-教学课件:闽浙湘徽
- 2025年合作联盟合同范本
- 肾小管性酸中毒的临床护理
- 五年级数学下《分数的基本性质》公开课教学设计
- 三尖瓣畸形的临床护理
- 2025劳动合同试用期限是多少个月
- 2025劳动合同范本【简单】
- 《2025年光纤传输网络建设合同》
- 2025年上海市各区高三二模语文试题汇编《文言文一》含答案
- 初中历史金与南宋对峙课件-2024-2025学年统编版七年级历史下册
- 2025年装维智企工程师(三级)复习模拟100题及答案
- 国家管网集团西南管道昆明输油气分公司突发环境事件综合应急预案
- 施工现场临时用电安全
- 停送电培训课件
- 医院培训课件:《核心制度-护理值班和交接班制度》
- 解题秘籍05 圆的综合问题(9种题型汇-总+专题训练)(解析版)-2025年中考数学重难点突破
- 无线网络施工方案
- 电商平台居间合同
- 阮乐器美术课件
- 中国大唐集团有限公司陆上风电工程标杆造价指标(2023年)
- 2025年上半年赤峰新正电工技术服务限公司社会招聘66人(内蒙古)易考易错模拟试题(共500题)试卷后附参考答案
评论
0/150
提交评论