Keypoint Detection and Local Feature Matching for Textured 3D Face Recognition

Holistic face recognition algorithms are sensitive to expressions, illumination, pose, occlusions and makeup. On the other hand, feature-based algorithms are robust to such variations. In this paper, we present a feature-based algorithm for the recognition of textured 3D faces. A novel keypoint dete...

Full description

Saved in:
Bibliographic Details
Published inInternational journal of computer vision Vol. 79; no. 1; pp. 1 - 12
Main Authors Mian, Ajmal S., Bennamoun, Mohammed, Owens, Robyn
Format Journal Article
LanguageEnglish
Published Boston Springer US 01.08.2008
Springer
Springer Nature B.V
Subjects
Online AccessGet full text
ISSN0920-5691
1573-1405
DOI10.1007/s11263-007-0085-5

Cover

More Information
Summary:Holistic face recognition algorithms are sensitive to expressions, illumination, pose, occlusions and makeup. On the other hand, feature-based algorithms are robust to such variations. In this paper, we present a feature-based algorithm for the recognition of textured 3D faces. A novel keypoint detection technique is proposed which can repeatably identify keypoints at locations where shape variation is high in 3D faces. Moreover, a unique 3D coordinate basis can be defined locally at each keypoint facilitating the extraction of highly descriptive pose invariant features. A 3D feature is extracted by fitting a surface to the neighborhood of a keypoint and sampling it on a uniform grid. Features from a probe and gallery face are projected to the PCA subspace and matched. The set of matching features are used to construct two graphs. The similarity between two faces is measured as the similarity between their graphs. In the 2D domain, we employed the SIFT features and performed fusion of the 2D and 3D features at the feature and score-level. The proposed algorithm achieved 96.1% identification rate and 98.6% verification rate on the complete FRGC v2 data set.
Bibliography:SourceType-Scholarly Journals-1
ObjectType-Feature-1
content type line 14
ObjectType-Article-2
content type line 23
ISSN:0920-5691
1573-1405
DOI:10.1007/s11263-007-0085-5