Beruflich Dokumente
Kultur Dokumente
Abstract—This paper presents an effective method for 3-D face due to factors as diverse as pose and facial expression etc. are
recognition using a novel geometric facial representation along usually greater than interclass ones. The past decades have wit-
with a local feature hybrid matching scheme. The proposed facial
nessed tremendous efforts firstly focused on 2-D face images [2]
surface description is based on a set of facial depth maps extracted
by multiscale extended Local Binary Patterns (eLBP) and enables and more recently on 3-D face models or scans [3]. Despite great
an efficient and accurate description of local shape changes; it thus progress achieved so far within the field [2], face recognition
enhances the distinctiveness of smooth and similar facial range (FR) using 2-D facial texture images is still not reliable enough
images generated by preprocessing steps. The following matching [4], especially in the presence of pose and lighting variations [5].
strategy is SIFT-based and performs in a hybrid way that com-
bines local and holistic analysis, robustly associating the keypoints With the rapid development in 3-D imaging systems, 2.5-D and
between two facial representations of the same subject. As a 3-D facial scans have emerged as a major alternative in dealing
result, the proposed approach proves robust to facial expression with the unsolved issues in 2-D face recognition, i.e. changes
variations, partial occlusions, and moderate pose changes, and the of illumination and pose [3], [6]. Meanwhile, even though 3-D
last property makes our system registration-free for nearly frontal
face models. The proposed method was experimented on three
facial scans capture exact shape information of facial surfaces,
public datasets, i.e. FRGC v2.0, Gavab, and Bosphorus. It displays and are thereby theoretically reputed to be robust to variations
a rank-one recognition rate of 97.6% and a verification rate of in illumination, they are likely to be more sensitive to expres-
98.4% at a 0.001 FAR on the FRGC v2.0 database without any sion changes. Furthermore, they generally require an accurate
face alignment. Additional experiments on the Bosphorus dataset
registration step before 3-D shape-based face matching.
further highlight the advantages of the proposed method with
regard to expression changes and external partial occlusions. The
last experiment carried out on the Gavab database demonstrates A. Related Work
that the entire system can also deal with faces under large pose
variations and even partially occluded ones, when only aided by a Generally, how to describe facial surface is a core topic in 3-D
coarse alignment process. face recognition. “Good” features of facial surfaces should have
Index Terms—Extended LBP, geometric facial description, hy- the following properties [7]: first, they can tolerate within-class
brid matching, SIFT, 3-D face recognition and verification. variations while discriminating different classes well; second,
they can easily be extracted from raw facial data to allow fast
I. INTRODUCTION processing; finally, they should lie in a space with moderate di-
mensionality to avoid high computational cost in matching. As a
3-D face recognition techniques can also be categorized ac- of a set of MultiScale extended Local Binary Pattern Depth
cording to the nature of their matching strategies, even though it Faces (MS-eLBP-DFs). This method improves the discrim-
is highly dependent on the adopted features. Zhao et al. [2] have inative power of LBP for 3-D facial surface description by
roughly classified 2-D face recognition approaches into three two solutions, i.e. encoding exact gray value differences be-
main streams: holistic, e.g. PCA [27] and LDA [28]; feature- tween the central pixel and the neighboring ones as well as
based such as Elastic Bunch Graph Matching (EBGM) [29]; embedding a MultiScale (MS) scheme.
hybrid like Component Eigenfaces [30]. This taxonomy can 2) In order to extract sufficient repeatable local features on
be extended to 3-D face recognition. The holistic stream con- smooth facial surfaces, we propose to apply SIFT [46] to
tains ICP-based matching [12], annotated deformable model these MultiScale extended LBP-based facial representa-
[31], and isometry invariant description [32] etc. This matching tions, i.e. MS-eLBP-DFs, interpreted as simple gray level
scheme is based on holistic facial features and hence generally images, for the detection of keypoints and the characteri-
requires an accurate normalization step with respect to pose and zation of local geometric properties.
scale changes. Furthermore, it has proved sensitive to expres- 3) A hybrid matching approach is designed to measure sim-
sion variations and partial occlusions. The feature-based one ilarities between gallery and probe facial scans once they
utilizes local features of 3-D facial scans and has been explored are represented in terms of the geometric facial descrip-
in several works in the literature, such as point signature [24] tion, i.e. MS-eLBP-DFs. This matching method inherits the
and more recently keypoint detection and local feature matching principles of local feature matching along with graph based
for textured 3-D face recognition by Mian et al. [33]. In this last matching as in [33], and it also extends the latter by incor-
work, an original keypoint detection method on a 3-D facial porating a facial component constraint.
surface was introduced, and a graph-based matching scheme 4) Thanks to the proposed local feature-based approach and
along with a dynamic fusion strategy performed at score level. the hybrid matching scheme, the proposed 3-D face recog-
As also highlighted in [33], feature-based matching has the po- nition algorithm is robust to facial expression changes and
tential advantage of being robust to facial expression, pose and partial occlusions whilst remaining tolerant of moderate
lighting changes and even to partial occlusions. The downside pose variations. The last property makes our method a reg-
of this scheme is the difficulty in extracting sufficient informa- istration-free technique for recognizing nearly frontal 3-D
tive feature points from similar or smooth 3-D facial surfaces. facial scans as can be the case of most user cooperative
Some studies also exist which present hybrid matching com- biometric applications, e.g. access control. This is clearly
bining global facial features with local ones: Region-ICP [34], in contrast to the overwhelming majority of state-of-the-art
multiple region-based matching [35] and the component and 3-D face recognition algorithms requiring the costly 3-D
morphable model-based approach [36]. As this scheme tends face alignment step before face matching.
to combine facial configuration information with local proper- As a local feature-based 3-D face recognition algorithm,
ties of faces, it is theoretically the most powerful [2]. However, the proposed approach and that of Mian et al. in [33] share
it also risks inheriting both types of shortcomings: sensitivity some similarities: particularly the overall local-feature ori-
to pose variations, difficulty in generating sufficient stable de- ented framework (though the features used in each of the two
scriptive features, etc. approaches are very different), the use of the SIFT method
(though SIFT is applied to MS-eLBP-DFs, instead of texture
B. Motivation and Approach Overview maps), the matching strategy combining local feature matching
In this paper, our basic assumption, as the one behind fea- and graph based matching that is further extended to include
ture-based face recognition algorithms, is that, when a variation a facial component constraint-based matching, and finally the
such as an expression or an occlusion occurs on a probe 3-D fa- weighted fusion scheme at score level with dynamic weight
cial scan, there still exist some small local areas, e.g. the nose calculation.
region during a facial expression, that change slightly or remain The proposed method was evaluated on three public datasets,
invariant as compared to the corresponding 3-D facial scan in namely FRGC v2.0 [48], Bosphorus [68] and Gavab DB [39].
the gallery set. Once located and characterized accurately, these As experimented on the FRGC v2.0 database for both the tasks
local regions can be utilized to identify 3-D faces, thereby pro- of 3-D face recognition and verification, our approach achieves
viding robustness to facial expression changes and partial oc- a rank-one recognition rate of 97.6% and a 98.4% verification
clusions through a proper matching process. rate with a FAR of 0.1% respectively. Since the 3-D facial scans
Motivated by this intuition, this paper proposes a novel ap- in FRGC are nearly frontal, the costly step of 3-D face alignment
proach to 3-D face recognition, making use of a geometric fa- was not required. The experiment carried out on the Bosphorus
cial representation along with a local feature hybrid matching dataset further highlights the ability of the proposed method
strategy. Our main contributions can be summarized as follows: to identify nearly frontal 3-D face models owning expression
1) Because after basic preprocessing, e.g. spike removal and changes and external occlusions. When it handles large pose
hole filling, 3-D facial surfaces to be identified are generally variations including left and right profiles which probably lead
smooth and similar, to achieve accurate representations of to self-occlusions, a coarse alignment based on a few landmarks
facial surfaces and enhance their distinctiveness, we propose is sufficient as a preprocessing step of this approach. This is
a 3-D shape based geometric facial description, consisting demonstrated by the experiments on the Gavab dataset.
HUANG et al.: 3-D FACE RECOGNITION USING eLBP-BASED FACIAL DESCRIPTION 1553
Fig. 4. Confusion case of LBP when it encodes similar but different local geo- eLBP code of its corresponding layer, e.g., is composed of
metric shapes.
and its decimal value is 64; is composed of
and its decimal value is 54; finally is composed
similar shapes with different Shape Index (SI) values [42] while of and its decimal value is 234. As a result, when
indeed sharing the same LBP code: shape (A) is a spherical cap; describing two similar local shapes, although the first layer LBP
shape (B) is a dome. This lack of descriptive power is problem- is not discriminative enough (both marked by the decimal value
atic when one needs to derive a facial description to enhance of 211), the information encoded in their additional layers can be
distinctiveness for the task of face recognition. used to distinguish them so long as the values of the two shapes
To address such a problem, we propose to adopt two comple- in all corresponding layers are not exactly the same.
mentary solutions. The first solution aims to improve the dis- Theoretically, in one image, the maximum value of s is
criminative ability of LBP with the eLBP coding method, and 255 (between 0 and 255), which means that 8 additional bi-
the other one focuses on providing a more comprehensive geo- nary units are required to encode s , and
metrical representation of a given neighborhood by applying a thus 7 additional layers should be produced. Nevertheless, we
multiscale strategy. Each solution is discussed in the following do not need so many layers in eLBP. Preprocessed range faces
two subsections respectively. are indeed very smooth; the s in a local surface generally do
not vary dramatically. Some preliminary statistical work reveals
B. Extended Local Binary Patterns that more than 80% s are smaller than 7 between points
The fact that LBP is not competent to distinguish similar local within eight pixels. Therefore, the number of additional binary
shapes is due to its operation mode. It only encodes relative dif- units, , is determined by . Meanwhile can also be ex-
ferences between a central pixel and its neighboring ones. In ploited to control the trade-off between the description expres-
this section, we introduce eLBP in order to better describe local siveness of local shapes and the computational simplicity of
surface properties. Instead of the original LBP, eLBP not only eLBP. All the which are larger than can be as-
extracts the relative gray value difference between the central signed to to decrease computational cost. In this study,
pixel and its neighbors provided by LBP, but also focuses on three additional layers are extracted and analyzed to illustrate
their absolute differences which are also critical to describe local their contributions to the final accuracy.
shapes. ELBP is a generalized version the 3DLBP [38] origi-
C. Multiscale Strategy
nally proposed for histogram-based 3-D face recognition.
Specifically, the eLBP code consists of several LBP codes The original LBP operator was extended later with different
in multiple layers that encode the exact gray value differences sizes of local neighborhood to deal with various scales [40].
between the central pixel and its neighbors. The first layer The local neighborhood of the LBP operator is defined as a set
of eLBP is actually the original LBP code encoding the sign. of sampling points evenly spaced on a circle centered on the
The following layers of eLBP then encode the absolute value pixel to be labeled. These sampling points which do not fall
of . Basically, each absolute value is firstly encoded exactly on the pixels are expressed using bilinear interpolation,
in its binary representation and then all the binary values at a thus allowing any radius value and any number of points in the
given layer result in an additional local binary pattern. The ex- neighborhood. Fig. 6 shows different LBP neighborhoods. The
ample of Fig. 1 can be expressed by eLBP as shown in Fig. 5. notation denotes the neighborhood of sampling points
The first layer of eLBP code is simply the original LBP code on a circle of radius . By adopting the same protocol, the eLBP
that encodes the sign of , yielding a decimal number of 211 operator can also handle different sampling points and scales.
from its binary form . The absolute values of , Some LBP histogram-based tasks change the neighborhood
i.e. 1, 5, 3, 2, 1, 2, 3, 0, are first encoded in their binary num- of the LBP operator for improved performance. By varying the
bers: , , , , , etc. Using the same value of radius , the LBP of different resolutions is thus ob-
eighting scheme of LBP on all the binary bits, we generate the tained. The multiscale strategy was originally used for texture
HUANG et al.: 3-D FACE RECOGNITION USING eLBP-BASED FACIAL DESCRIPTION 1555
Fig. 6. LBP operator examples: circular (8, 1), (16, 2), and (8, 2).
classification [40], and it was also introduced to 2-D face recog- Fig. 7. MS-eLBP-DFs of a facial range image with different radii from 1 to 8
(from left to right).
nition [43], [44]. In [45], Shan and Gritti studied MS-LBP for
facial expression recognition by firstly extracting MS-LBP his-
togram-based facial features and then using AdaBoost to learn
the most discriminative bins. They reported that the boosted
classifiers of MS-LBP consistently outperform those based on
single-scale LBP, and the selected LBP bins distribute at all
scales. MS-LBP can hence be regarded as an efficient method
for facial representation. When considering it in 3-D face anal-
ysis, this multiscale technique can be also applied to enhance
the descriptive power of LBP.
Fig. 10. SIFT-based keypoints of an original facial range image and its four IV. THE HYBRID MATCHING PROCESS
associated eLBP-DFs.
Once local features have been extracted from MS-eLBP-DFs,
a hybrid matching process is carried out, which combines a local
matching step using the SIFT-based features with a global one
keypoint are given more emphasis than the ones far from it (see
under the facial component and configuration constraints.
[46] for SIFT parameter settings). The orientation histograms
of 4 4 sampling regions are calculated, each with eight ori- A. Local Feature-Based Matching
entation bins. Hence a feature vector with a dimension of 128
is produced. Given local facial features extracted from each
The SIFT operator works on each MS-eLBP-DF separately. MS-eLBP-DFs pair of the gallery and the probe face scan
Because each MS-eLBP-DFs highlights the local shape changes respectively, two facial keypoint sets can be matched. Matching
of an original smooth facial range image by encoding local bi- one keypoint to another is accepted only if the matching
nary patterns at different scales and thus providing various de- distance is less than a predefined threshold times the distance
tails, many more SIFT-based keypoints can be detected for the to the second closest match. Here, denotes the
following matching step than in the original smoothed range number of matched keypoints in the layer of an eLBP-DFs
face. Some statistical work was done using all the 3-D facial pair, generated by eLBP from range face images with a
scans in the FRGC v2.0 database. The average number of de- parameter setting of .
scriptors extracted from each of MS-eLBP-DFs is 553, while
that of each original facial range image is limited to 41 and the B. Holistic Facial Matching
detected keypoints are often located on the edge of the face. As Unlike the samples used in the domain of object detection,
a result, the number of detected keypoints is increased by 10 all human faces have the same physical components and share
times more on MS-eLBP-DFs as compared to the original face a similar global configuration. Holistic matching is thus carried
range image; and the number of keypoints only covers 0.3% of out to constrain the matched local features with respect to the
pixels in each MS-eLBP-DF that has an average resolution of facial components and configuration.
400 400. Fig. 10 shows the SIFT-based keypoints extracted 1) Facial Component Constraint: we propose to divide the
from one facial range image and its four associated eLBP-DFs entire facial range image into nonoverlapped subregions,
respectively. each of which contains roughly one component of nearly
Meanwhile, we also studied the repeatability of these key- frontal faces. Different from the division scheme used for
points detected on MS-eLBP-DFs across different scans of the histogram statistics, the one in our method is to restrict the
same subject based on three manually labeled facial landmarks, matched keypoints of gallery and probe face scans only
i.e. the nose tip and two inner corners of eyes. 25 subjects that to those with similar physical meaning. That means the
have more than 4 facial scans were randomly selected in FRGC matched keypoints from the same facial region should be
v2.0 for this experiment. To each keypoint detected on an more important. Instead of the costly clustering step [47] to
MS-eLBP-DF, we can compute a vector of three distances, automatically construct subregions based on keypoint lo-
, each of which is the one between the keypoint and cations from training samples, we simply use facial com-
one of these three landmarks. These three distances are then ponent position, and divide the face area into 3 3 rec-
used as the coordinates of . Now a keypoint detected by tangle blocks of the same size. The similarity measure-
SIFT on an MS-eLBP-DF with the coordinates is ment of the facial component constraint is defined from this
considered as in correspondence or matched with the keypoint facial composition scheme. An MS-eLBP-DF, , is repre-
detected from another MS-eLBP-DF possessing the coordi- sented as ; is 9 in our case and is
nates , as long as the difference between each pair the number of detected SIFT keypoints that fall within the
HUANG et al.: 3-D FACE RECOGNITION USING eLBP-BASED FACIAL DESCRIPTION 1557
All facial range images are normalized to a certain size where corresponds to the three similarities: , , and ,
to build a public 2-D coordinate system. For each 3-D and operators and produce the first and
face scan, the MS-eLBP-DFs are extracted from the range second maximum value of vector . The gallery face which
image. Therefore, all the keypoints of the proposed facial has the maximum value in vector is declared as the identity
surface representations share the same XY-plane with the of the probe face image when the decision is to be made on each
range face image, and the pixel values of the corresponding MS-eLBP-DF independently.
facial range image can be regarded as the Z-axis values
of these keypoints. Hence, each keypoint has its position
in 3-D space. After local feature-based matching, a 3-D V. EXPERIMENTAL RESULTS
graph is formed for each MS-eLBP-DF of a probe , by
To demonstrate the effectiveness of the proposed method, we
simply linking every two keypoints which have a matching
utilized three public datasets for experiments, i.e. FRGC v2.0
relationship with keypoints detected on the corresponding
[48], Bosphorus [68] and Gavab DB [39]. The first is for eval-
MS-eLBP-DF of a gallery face . The matched keypoints
uating performance with large number of subjects that have
of also construct a corresponding graph of . Since all
slight pose variations and various facial expression changes; the
the facial range images are of the same scale, intuitively,
second is to further observe its robustness to facial expression
if faces and are from the same subject, their corre-
variations as well as external occlusions; while the last one is
sponding graphs should have similar shapes in 3-D space.
to analyze its accuracy on 3-D face samples with extreme pose
changes (left and right profiles).
The similarity measure between the two graphs can be de-
fined as:
A. Experiments on the FRGC v2.0 Dataset
Fig. 11. Examples of preprocessed 3-D facial scans of the same subject from the FRGC dataset.
TABLE IV
COMPARISON OF VERIFICATION RATES AT 0.001 FAR USING THE EXPRESSION
PROTOCOL ON THE FRGC V2.0 DATASET
Fig. 13. Some example 3-D facial scans of the same subject in the Bosphorus
dataset.
Fig. 14. Examples of all the 3-D facial scans of the same subject from the Gavab dataset [58].
The experimental protocol in [67] has some differences from that in this
study and [69].
Fig. 16. Face matching demonstration. The upper maps are generated by ; while the bottom ones with : (a) matching with
nearly frontal probe face; (b) matching with probe rotated by 5 ; (c) matching with probe rotated by 10 ; (d) matching with probe rotated by 15 ; (e) matching
with probe rotated by 20 .
the others. Moreover, when evaluating the robustness to severe face is rotated by 15 while this number over the whole Gavab
pose variations (Table VII (B)), we achieved an overall accu- dataset excluding the left and right profiles is close to 10 in the
racy of 91.4% on these four subsets while that reported by [58] latter case. However, when probe facial scans are rotated by 20 ,
is 88.9%. it becomes difficult to discriminate intraclass variations from
To sum up, the experimental results achieved on the Gavab interclass ones in terms of the matched keypoint pairs. We con-
dataset clearly prove that only aided by a coarse alignment, our ducted the experiment 10 times, the phenomena are similar.
method can deal with large pose changes and even partial oc-
clusions. D. Experiment Summary
3) Analysis of the Impact by Pose Changes on Face Recogni- In the Sections V-A, V-B and V-C, eight experiments (Exper-
tion Accuracy (Experiment H): The previous experiments tend iment A-H) are carried out on three public datasets, i.e. FRGC
to suggest that our method can cope with moderate 3-D face v2.0, Bosphours and Gavab. The results achieved in Experiment
pose changes, in particular for nearly frontal face models as this B, D and E demonstrate that our method, without the registra-
can be the case for user cooperative biometric applications (e.g., tion step for nearly frontal faces displays comparable to the best
access control), without any prior registration. When aided by ones so far presented in the literature. Moreover, Experiments C
a coarse alignment step using very few landmarks, our method and F show that the proposed approach is robust to facial expres-
can even deal with 3-D face models with large pose variations sion variations whereas experiment F further highlights its tol-
such as the ones in the Gavab dataset. However, the question erance to partial external occlusions. Additionally, Experiment
remains of determining to what extent the proposed approach is G indicates that supported by a coarse registration step applied
still robust in pose changes. To give some insight into this ques- on only a few landmarks, our approach can deal with the faces
tion, we designed the following experiment. Two nearly frontal with large pose variations and even self-occlusions.
facial scans of the same subject were randomly chosen from the At the same time, we observe the computational expense of
Gavab database: one used as the gallery face and the other as the the proposed approach along with the experiments conducted
probe. We rotate the probe one around the yaw axis by 5 , 10 , on the FRGC v2.0 database. Currently, an unoptimized imple-
15 and 20 , respectively, thus leading to range images (shown mentation of the proposed method with MATLAB (R2010a) can
in Fig. 15) subsequently employed for the analysis on the im- perform one match between one pair of eLBP-DFs of the gallery
pact of rotated faces to the matching result. and probe faces in about 0.32 s using a machine where Intel(R)
We generated the eLBP based facial representations for a Core(TM) i5 CPU (2.60 GHz) and 4 GB RAM are equipped.
gallery face (a) as well as each of the probe faces (b–f). We Since the similarity scores based on different eLBP-DFs can
then computed the number of keypoints detected on each probe be calculated independently, if implemented in a parallel com-
face (c)–(f) and found that the numbers almost remain stable puting device, our method is quite promising in providing deci-
when compared with the one of the original probe face (b). In the sions in real time.
matching phase, when the probe face Fig. 15(b) is rotated suc-
cessively from Figs. 15(c) to (f) each time by 5 , the number of VI. CONCLUSION
matched keypoints with the gallery face (a) keeps decreasing, as We have presented an effective approach to 3-D face recog-
illustrated in Fig. 16, from roughly 300 pairs in Fig. 16(a) down nition using a novel geometric facial representation and local
to only around 10 pairs in Fig. 16(e) when the probe face is ro- feature hybrid matching. The proposed facial representation is
tated by 20 from the frontal one. Meanwhile, the main quality based on MS-eLBP and allows for accurate and fast description
of a similarity measurement for the purpose of face recogni- of local shape variations, thus enhancing the distinctiveness of
tion is its ability to distinguish intraclass changes from interclass range faces. SIFT-based hybrid matching that combines local
ones. From this perspective, the number of matched keypoints and holistic analysis further robustly associates keypoints be-
on each eLBP-DF pair used as face similarity measurement be- tween two faces of the same subject. The proposed method was
tween the gallery face and the probe rotated by 15 can still evaluated in 3-D face recognition and verification, achieving
clearly discriminate the facial scans of the same subject from a recognition rate of 97.6% and a 98.4% verification rate with
the ones of different subjects. Indeed, the average number of a 0.001 FAR respectively on the FRGC v2.0 database which
matched keypoints in the former case is above 20 even as the consists of nearly frontal 3-D facial scans with rich facial
HUANG et al.: 3-D FACE RECOGNITION USING eLBP-BASED FACIAL DESCRIPTION 1563
expression changes. Additional experiments on the Bosphorus [19] A. B. Moreno, A. Sanchez, J. F. Velez, and F. J. Diaz, “Face recognition
database further confirm the advantages of the proposed using 3D surface-extracted descriptors,” in Proc. Irish Machine Vision
and Image Processing Conf., Coleraine, Ireland, 2003.
method with regard to facial expression variations and external [20] S. Gupta, J. K. Aggatwal, M. K. Markey, and A. C. Bovik, “3D face
partial occlusions. The results achieved on the Gavab database recognition founded on the structural diversity of human faces,” in
containing severe pose changes clearly illustrate that the entire Proc. IEEE Conf. Computer Vision and Pattern Recognition, Min-
neapolis, MN, 2007.
system also provides a promising solution to recognizing par- [21] T. Nagamine, T. Uemura, and I. Masuda, “3D facial image analysis for
tially occluded faces. Moreover, generally costly registration human identification,” in Proc. Int. Conf. Pattern Recognition, 1992,
was not needed thanks to the relative tolerance of the proposed pp. 324–327.
[22] C. Beumier and M. Acheroy, “Automatic 3D face authentication,”
hybrid matching strategy to nearly frontal faces like the ones Image Vis. Comput., vol. 18, no. 4, pp. 315–321, 2000.
in the FRGC v2.0 and the subset of Bosphorus. When dealing [23] C. Samir, A. Srivastava, and M. Daoudi, “Three-dimensional face
with extreme poses, e.g. left or right profiles, a coarse alignment recognition using shapes of facial curves,” IEEE Trans. Pattern Anal.
Mach. Intell., vol. 28, no. 11, pp. 1858–1863, Nov. 2006.
step based on a few manually landmarked points was sufficient [24] C. Chua, F. Han, and Y. Ho, “3D human face recognition using point
in preprocessing as indicated by the experiments on the Gavab signature,” in Proc. Int. Conf. Automatic Face and Gesture Recogni-
database. tion, 2000, pp. 233–238.
[25] H. T. Tanaka, M. Ikeda, and H. Chiaki, “Curvature-based face sur-
face recognition using spherical correlation—Principal directions for
REFERENCES curved object recognition,” in Proc. Int. Conf. Automatic Face and
Gesture Recognition, 1998, pp. 372–377.
[1] A. K. Jain, A. Ross, and S. Prabhakar, “An introduction to biometric [26] Y. Wang, J. Liu, and X. Tang, “Robust 3D face recognition by local
recognition,” IEEE Trans. Circuits Syst. Video Technol., vol. 14, no. 1, shape difference boosting,” IEEE Trans. Pattern Anal. Mach. Intell.,
pp. 4–20, Jan. 2004. vol. 32, no. 10, pp. 1858–1870, Oct. 2010.
[2] W. Zhao, R. Chellappa, P. J. Phillips, and A. Rosenfeld, “Face recog- [27] M. Turk and A. Pentland, “Eigenfaces for recognition,” J. Cognitive
nition: A literature survey,” ACM Comput. Survey, vol. 35, no. 4, pp. Neuro-Sci., vol. 3, no. 1, pp. 71–86, 1991.
399–458, 2003. [28] P. Belhumeur, J. Hespanha, and D. Kriegman, “Eigenfaces vs. fisher-
[3] K. W. Bowyer, K. Chang, and P. J. Flynn, “A survey of approaches and faces: Recognition using class specific linear projection,” IEEE Trans.
challenges in 3D and multi-modal 3D+2D face recognition,” Comput. Pattern Anal. Mach. Intell., vol. 19, no. 7, pp. 711–720, Jul. 1997.
Vis. Image Understand., vol. 101, no. 1, pp. 1–15, 2006. [29] L. Wiskott, J. M. Fellous, N. Kruger, and C. V. D. Malsburg, “Face
[4] A. F. Abate, M. Nappi, D. Riccio, and G. Sabatino, “2D and 3D face recognition by elastic bunch graph matching,” IEEE Trans. Pattern
recognition: A survey,” Pattern Recognit. Lett., vol. 28, no. 14, pp. Anal. Mach. Intell., vol. 19, no. 7, pp. 775–779, Jul. 1997.
1885–1906, 2007. [30] A. Pentland, B. Moghaddam, and T. Starner, “View-based and mod-
[5] P. J. Phillips, H. Moon, P. J. Rauss, and S. Rizvi, “The FERET evalu- ular eigenspaces for face recognition,” in Proc. IEEE Conf. Computer
ation methodology for face recognition algorithms,” IEEE Trans. Pat- Vision and Pattern Recognition, 1994, pp. 84–91.
tern Anal. Mach. Intell., vol. 22, no. 10, pp. 1090–1104, Oct. 2000. [31] I. A. Kakadiaris, G. Passalis, G. Toderici, M. N. Murtuza, Y. Lu, N.
[6] A. Scheenstra, A. Ruifrok, and R. C. Veltkamp, “A survey of Karampatziakis, and T. Theoharis, “Three-dimensional face recogni-
3D face recognition methods,” in Proc. Int. Conf. Audio- and tion in the presence of facial expressions: An annotated deformable
Video-Based Biometric Person Authentication, 2005, pp. 891–899. model approach,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 29, no.
[7] A. Hadid, M. Pietikäinen, and T. Ahonen, “A discriminative feature 4, pp. 640–649, Apr. 2007.
space for detecting and recognizing faces,” in Proc. IEEE Conf. Com- [32] A. M. Bronstein, M. M. Bronstein, and R. Kimmel, “Three dimensional
puter Vision and Pattern Recognition, 2004, pp. 797–804. face recognition,” Int. J. Comput. Vis., vol. 64, no. 1, pp. 5–30, 2005.
[8] B. Achermann, X. Jiang, and H. Bunke, “Face recognition using range [33] A. S. Mian, M. Bennamoun, and R. Owens, “Keypoint detection
images,” in Proc. Int. Conf. Virtual Systems and MultiMedia, 1997, pp. and local feature matching for textured 3D face recognition,” Int. J.
129–136. Comput. Vis., vol. 79, no. 1, pp. 1–12, 2008.
[9] C. Hesher, A. Srivastava, and G. Erlebacher, “A novel technique for [34] K. Ouji, B. B. Amor, M. Ardabilian, L. Chen, and F. Ghorbel, “3D face
face recognition using range imaging,” in Proc. Int. Symp. Signal Pro- recognition using R-ICP and geodesic coupled approach,” in Proc. Int.
cessing and its Applications, 2003, pp. 201–204. Conf. Multimedia Modeling, 2009, pp. 390–400.
[10] A. M. Bronstein, M. M. Bronstein, and R. Kimmel, “Expression in- [35] A. S. Mian, M. Bennamoun, and R. Owens, “Region-based matching
variant 3D face recognition,” in Proc. Int. Conf. Audio- and Video- for robust 3D face recognition,” in Proc. British Machine Vision Conf.,
Based Biometric Person Authentication, 2003, pp. 62–70. Oxford, U.K., 2005.
[11] P. J. Besl and N. D. McKay, “A method for registration of 3-D shapes,” [36] J. Huang, B. Heisele, and V. Blanz, “Component-based face recog-
IEEE Trans. Pattern Anal. Mach. Intell., vol. 14, no. 2, pp. 239–256, nition with 3D morphable models,” in Proc. Int. Conf. Audio- and
Feb. 1992. Video-Based Biometric Person Authentication, 2003, pp. 27–34.
[12] X. Lu, A. K. Jain, and D. Colbry, “Matching 2.5D face scans to 3D [37] S. Z. Li, C. Zhao, M. Ao, and Z. Lei, “Learning to fuse
models,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 28, no. 1, pp. 3D+2D based face recognition at both feature and decision
31–43, Jan. 2006. levels,” in Proc. Workshop on Analysis and Modeling of Faces
[13] X. Lu and A. K. Jain, “Deformation modeling for robust 3D face and Gestures, 2005, pp. 44–54.
matching,” in Proc. IEEE Conf. Computer Vision and Pattern Recog- [38] Y. Huang, Y. Wang, and T. Tan, “Combining statistics of geometrical
nition, 2006, pp. 1377–1383. and correlative features for 3D face recognition,” in Proc. British Ma-
[14] K. Chang, K. W. Bowyer, and P. Flynn, “Effects on facial expression chine Vision Conf., Edinburgh, U.K., 2006.
in 3D face recognition,” in Proc. SPIE Conf. Biometric Technology for [39] A. B. Moreno and A. Sanchez, “GavabDB: A 3D face database,” in
Human Identification, 2005, vol. 5779, pp. 132–143. Proc. COST Workshop on Biometrics on the Internet: Fundamentals,
[15] T. Maurer, D. Guigonis, I. Maslov, B. Pesenti, A. Tsaregorodtsev, D. Advances and Applications, 2004, pp. 77–82.
West, and G. Medioni, “Performance of geometrix active ID 3D face [40] T. Ojala, M. Pietikäinen, and T. Maenpaa, “Multiresolution gray-scale
recognition engine on the FRGC data,” in Proc. IEEE Workshop on and rotation invariant texture classification with local binary patterns,”
FRGC Experiments, San Diego, CA, 2005. IEEE Trans. Pattern Anal. Mach. Intell., vol. 24, no. 7, pp. 971–987,
[16] Y. Lee and J. Shim, “Curvature-based human face recognition using Jul. 2002.
depth-weighted Hausdorff distance,” in Proc. Int. Conf. Image Pro- [41] T. Ahonen, A. Hadid, and M. Pietikäinen, “Face recognition with
cessing, 2004, pp. 1429–1432. local binary patterns,” in Proc. Eur. Conf. Computer Vision, 2004, pp.
[17] T. D. Russ, M. W. Koch, and C. Q. Little, “A 2D range Hausdorff 469–481.
approach for 3D face recognition,” in Proc. IEEE Workshop on FRGC [42] J. J. Koenderink and A. J. Doorn, “Surface shape and curvature scales,”
Experiments, San Diego, CA, 2005. Image Vis. Comput., vol. 10, no. 8, pp. 557–565, 1992.
[18] G. G. Gordon, “Face recognition based on depth and curvature fea- [43] S. Yan, H. Wang, X. Tang, and T. S. Huang, “Exploring feature de-
tures,” in Proc. IEEE Conf. Computer Vision and Pattern Recognition, scriptors for face recognition,” in Proc. Int. Conf. Acoustics, Speech,
1992, pp. 808–810. and Signal Processing, 2007, pp. 629–632.
1564 IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, VOL. 7, NO. 5, OCTOBER 2012
[44] C. Chan, J. Kittler, and K. Messer, “Multi-scale local binary pattern [66] C. C. Queirolo, L. Silva, O. R. P. Bellon, and M. P. Segundo, “3D face
histograms for face recognition,” in Proc. Int. Conf. Biometrics, 2007, recognition using simulated annealing and the surface interpenetration
pp. 809–818. measure,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 32, no. 2, pp.
[45] C. Shan and T. Gritti, “Learning discriminative LBP-histogram bins for 206–219, Feb. 2010.
facial expression recognition,” in Proc. British Machine Vision Conf., [67] N. Alyuz, B. Gokberk, and L. Akarun, “Regional registration for ex-
Leeds, U.K., 2008. pression resistant 3-D face recognition,” IEEE Trans. Inf. Forensics
[46] D. G. Lowe, “Distinctive image features from scale-invariant key- Security, vol. 5, no. 3, pp. 425–440, Sep. 2010.
points,” Int. J. Comput. Vis., vol. 60, no. 4, pp. 91–110, 2004. [68] A. Savran, N. Alyüz, H. Dibeklioğlu, O. Çeliktutan, B. Gökberk, B.
[47] J. Luo, Y. Ma, E. Takikawa, S. Lao, M. Kawade, and B. Lu, “Person- Sankur, and L. Akarun, “Bosphorus database for 3D face analysis,” in
specific SIFT features for face recognition,” in Proc. IEEE Int. Conf. The First COST 2101 Workshop on Biometrics and Identity Manage-
Acoustics, Speech and Signal Processing, 2007, pp. 593–596. ment, 2008, pp. 47–56.
[48] P. J. Phillips, P. J. Flynn, T. Scruggs, K. W. Bowyer, K. I. Chang, K. [69] C. Maes, T. Fabry, J. Keustermans, D. Smeets, P. Suetens, and D. Van-
Hoffman, J. Marques, J. Min, and W. Worek, “Overview of the face dermeulen, “Feature detection on 3D face surfaces for pose normali-
recognition grand challenge,” in Proc. IEEE Conf. Computer Vision sation and recognition,” in Proc. IEEE Int. Conf. Biometrics: Theory,
and Pattern Recognition, 2005, vol. I, pp. 947–954. Applications and Systems, Washington, DC, 2010.
[49] A. S. Mian, M. Bennamoun, and R. Owens, “An efficient multimodal
2D-3D hybrid approach to automatic face recognition,” IEEE Trans. Di Huang (S’10–M’11) received the B.S. and M.S.
Pattern Anal. Mach. Intell., vol. 29, no. 11, pp. 1927–1943, Nov. 2007. degrees in computer science from Beihang Univer-
[50] D. Huang, G. Zhang, M. Ardabilian, Y. Wang, and L. Chen, “3D face sity, Beijing, China, and the Ph.D. degree in computer
recognition using distinctiveness enhanced facial representations and science from Ecole Centrale de Lyon, Lyon, France,
local feature hybrid matching,” in Proc. IEEE Int. Conf. Biometrics: in 2005, 2008, and 2011, respectively.
Theory, Applications and Systems, Washington, DC, 2010. In December 2011, he joined the Laboratory of In-
telligent Recognition and Image Processing, Beijing
[51] Y. Wang, G. Pan, and Z. Wu, “3D face recognition in the presence
Key Laboratory of Digital Media, School of Com-
of expression: Guidance-based constraint deformation approach,”
puter Science and Engineering, Beihang University,
in Proc. IEEE Conf. Computer Vision and Pattern Recognition,
as a Faculty Member. His current research interests
Minneapolis, MN, 2007.
include biometrics, especially on 2-D/3-D face anal-
[52] K. I. Chang, K. W. Bowyer, and P. J. Flynn, “Adaptive rigid multi-re- ysis, image/video processing, and pattern recognition.
gion selection for handling expression variation in 3D face recogni-
tion,” in Proc. IEEE Workshop on FRGC Experiments, San Diego, CA,
2005.
[53] G. Passalis, I. Kakadiaris, T. Theoharis, G. Tederici, and N. Murtaza,
“Evaluation of 3D face recognition in the presence of facial expres- Mohsen Ardabilian received the M.S. and the Ph.D.
degrees in computer science from Université Tech-
sions: An annotated deformable model approach,” in Proc. IEEE Work-
nologie de Compiègne, France, in 1996 and 2001, re-
shop on FRGC Experiments, San Diego, CA, 2005.
spectively.
[54] M. Husken, M. Brauckmann, S. Gehlen, and C. V. D. Malsburg,
In 2001, he founded Avisias Company, special-
“Strategies and benefits of fusion of 2D and 3D face recognition,”
izing in media asset management with several other
in Proc. IEEE Workshop on FRGC Experiments, San Diego,
confirmed industrial players of Thomson, Canal+
CA, 2005.
Technologies, and Philips, where he served as Sci-
[55] J. Cook, V. Chandran, and C. Fookes, “3D face recognition using Log-
entific Expert from 2001 to 2003. In 2003, he joined
Gabor templates,” in Proc. British Machine Vision Conf., Edinburgh,
Ecole Centrale de Lyon as an associate professor.
U.K., 2006.
His current research interests include computer
[56] T. Faltemier, K. W. Bowyer, and P. Flynn, “A region ensemble for 3D
vision and multimedia analysis, particularly 3-D acquisition and modeling, 3-D
face recognition,” IEEE Trans. Inf. Forensics Security, vol. 3, no. 1,
face analysis and recognition.
pp. 62–73, Mar. 2008.
[57] P. Szeptycki, M. Ardabilian, and L. Chen, “A coarse-to-fine curvature
analysis-based rotation invariant 3D face landmarking,” in Proc. IEEE
Int. Conf. Biometrics: Theory, Applications and Systems, Washington,
DC, 2009. Yunhong Wang (M’98) received the B.S. degree in
[58] H. Drira, B. B. Amor, M. Daoudi, and A. Srivastava, “Pose and ex- electronic engineering from Northwestern Polytech-
pression-invariant 3D face recognition using elastic radial curves,” in nical University, Xi’an, China, in 1989, and the M.S.
Proc. British Machine Vision Conf., Aberystwyth, U.K., 2010. and Ph.D. degrees in electronic engineering from the
[59] X. Li, T. Jia, and H. Zhang, “Expression-insensitive 3D face recogni- Nanjing University of Science and Technology, Nan-
tion using sparse representation,” in Proc. IEEE Conf. Computer Vision jing, China, in 1995 and 1998, respectively.
and Pattern Recognition, Miami, FL, 2009. From 1998 to 2008, she worked at the National
[60] A. B. Moreno, A. Sanchez, J. F. Velez, and F. J. Díaz, “Face recognition Laboratory of Pattern Recognition, Institute of
using 3D local geometrical features: Pca vs.svm,” in Proc. Int. Symp. Automation, Chinese Academy of Sciences, Beijing.
Image and Signal Processing and Analysis, 2005, pp. 185–190. Since 2004, she has been a Professor with the School
[61] M. H. Mahoor and M. Abdel-Mottaleb, “Face recognition based on 3D of Computer Science and Engineering, Beihang
ridge images obtained from range data,” Pattern Recognit., vol. 42, no. University, Beijing, China, where she is also the Director of Laboratory of
3, pp. 445–451, 2009. Intelligent Recognition and Image Processing, Beijing Key Laboratory of
[62] S. Berretti, A. Del Bimbo, and P. Pala, “3D face recognition by Digital Media. Her research interests include biometrics, pattern recognition,
modeling the arrangement of concave and convex regions,” in computer vision, data fusion, and image processing.
Proc. Int. Workshop on Adaptive Multimedia Retrieval: Adaptive
Multimedia Retrieval: User, Context, and Feedback, 2006, pp.
108–118.
[63] M. H. Mousavi, K. Faez, and A. Asgharim, “Three dimensional face Liming Chen (M’05) received the B.Sc. degree in
recognition using svm classifier,” in Proc. Int. Conf. Computer and mathematics and computer science from Université
Information Science, 2008, pp. 208–213. de Nantes, Nantes, France, in 1984, and the M.S. and
[64] D. Huang, M. Ardabilian, Y. Wang, and L. Chen, “A novel geometric Ph.D. degrees, both in computer science, from the
facial representation based on multi-scale extended local binary pat- University of Paris 6, Paris, France, in 1986 and 1989,
terns,” in Proc. IEEE Int. Conf. Automatic Face and Gesture Recogni- respectively.
tion, 2011, pp. 1–7. From 1990 to 1998, he worked as an Associate
[65] D. Huang, C. Shan, M. Ardabilian, Y. Wang, and L. Chen, “Local bi- Professor with the Université de Technologies de
nary patterns and its application to facial image analysis: A survey,” Compiègne, Compiègne, France, and then he joined
IEEE Trans. Syst., Man, Cybern. C, Applicat. Rev., vol. 41, no. 6, pp. Ecole Centrale de Lyon, Lyon, France, in 1998 as a
765–781, Nov. 2011. Professor, where he has been leading an advanced
HUANG et al.: 3-D FACE RECOGNITION USING eLBP-BASED FACIAL DESCRIPTION 1565
research team on multimedia computing and pattern recognition. From 2001 Dr. Chen has been the Chairman and a Program Committee Member for var-
to 2003, he was the Chief Scientific Officer with the Paris-based company ious international conferences and journals. He has been a Reviewer for many
Avivias, where he specializes in media asset management. During 2005, he was conferences and journals, e.g., IEEE SIGNAL PROCESSING LETTERS, Computer
a Scientific Expert in multimedia at France Telecom R&D, China. Since 2007, Vision and Image Understanding, IEEE TRANSACTIONS ON SYSTEMS, MAN,
he has been Chairman with the Department of Mathematics and Computer AND CYBERNETICS, the Proceedings of the International Conference on Image
Science, Laboratoire d’InfoRmatique en Image et Systèmes d’information Processing, IEEE TRANSACTIONS ON IMAGE PROCESSING, and Pattern Recog-
(LIRIS), Ecole Centrale de Lyon. Since 1995, he has been the author of nition Letters. He was a Guest Editor for the special issue on Automatic Audio
three patents and more than 100 publications in international journals and Classification of the European Association for Signal Processing (EURASIP)
conferences. He has directed more than 15 Ph.D. dissertations. His research Journal on Audio, Speech, and Music Processing.
interests include face analysis and recognition in 3-D and 2-D, image and video
categorization, and affective computing on multimedia data.