Beruflich Dokumente
Kultur Dokumente
Information Fusion
journal homepage: www.elsevier.com/locate/inffus
a r t i c l e
i n f o
Article history:
Received 31 December 2013
Received in revised form 28 May 2014
Accepted 2 June 2014
Available online 21 June 2014
Keywords:
Face recognition
Wavelet transform
Pixel-level fusion
Feature-level fusion
Principle component analysis
a b s t r a c t
The traditional wavelet-based approaches directly use the low frequency sub-band of wavelet transform
to extract facial features. However, the high frequency sub-bands also contain some important
information corresponding to the edge and contour of face, reecting the details of face, especially the
top-levels high frequency sub-bands. In this paper, we propose a novel technique which is a joint of
pixel-level and feature-level fusion at the top-levels wavelet sub-bands for face recognition. We convert
the problem of nding the best pixel-level fusion coefcients of high frequency wavelet sub-bands to two
optimization problems with the help of principal component analysis and linear discriminant analysis,
respectively; and propose two alternating direction methods to solve the corresponding optimization
problems for nding transformation matrices of dimension reduction and optimal fusion coefcients
of the high frequency wavelet sub-bands. The proposed methods make full use of four top-levels wavelet
sub-bands rather than the low frequency sub-band only. Experiments are carried out on the FERET, ORL
and AR face databases, which indicate that our methods are effective and robust.
2014 Elsevier B.V. All rights reserved.
1. Introduction
Face recognition, as a biological feature recognition, has been
one of the most active research areas in computer vision, pattern
recognition and biometrics [1]. Face recognition has several advantages over other biometric modalities such as ngerprint and iris.
Besides being natural and nonintrusive, the most important advantage of face recognition is that it can be captured at a distance and
in a friendly manner. Various face recognition algorithms have
been devised in the literature. However, up to now, face recognition is still faced with a number of challenges such as varying illumination, facial expression and poses [24].
Feature extraction and classication are two key steps of a face
recognition system. Feature extraction provides an effective representation of face images to decrease the computational complexity
of the classier, which can greatly enhance the performance of a
face recognition system; while classication is to distinguish those
features with a good classier. Therefore, in order to improve the
recognition rate of a face recognition system, it is crucial to nd
a good feature extractor and an effective classier [5]. In this paper,
Corresponding author at: Department of Mathematics, School of Science,
Tianjin University, Tianjin 300072, PR China. Tel./fax: +86 22 27403615.
E-mail address: zt69069882@tju.edu.cn (T. Zhang).
http://dx.doi.org/10.1016/j.inffus.2014.06.001
1566-2535/ 2014 Elsevier B.V. All rights reserved.
we focus on feature extraction methods by using wavelet transform with the help of the classical principal component analysis
(PCA) [6] and linear discriminant analysis (LDA) [7].
Wavelet transform, which is an increasingly popular tool in
image processing and computer vision, has been investigated
in many applications, such as compression, detection, recognition, image retrieval, among others, due to its great advantages
with the nice features of space-frequency localization and multiresolution. Researchers have developed face recognition algorithms by combining discrete wavelet transform (DWT) with
other methods (see, for example, [813] and references therein).
Through a two-dimensional DWT (2D-DWT), an image of face is
transformed into two parts: a low frequency sub-band and three
high frequency sub-bands. On one hand, the low frequency subband plays a dominant role in four sub-bands for approximation
of the original image; and on the other hand, when the expression changes or the image is affected by occlusion, the low frequency sub-band has no obvious change but the high frequency
sub-bands change obviously, so the low frequency sub-band is
usually single-handed used for identication in image recognition [10]. The authors in [14] directly used the low frequency
sub-band of wavelet transform to extract facial features.
However, although the high frequency sub-bands do not contain
96
2. Model of TWSBF
Let L2 R denote the square integrable space. The continuous
wavelet transform of a one-dimensional function f 2 L2 R is
dened as
W f a; b
f twa;b tdt;
1
tb
;
wa;b t jaj2 w
a
in which wt is called mother wavelet and the parameters a and b
stand for the scale and position, respectively. Eq. (1) can be discretized by imposing restrictions on a and b with a 2n and b 2 Z.
The DWT of a one-dimensional signal is processed by transforming it into two parts with a low-pass lter and a high-pass lter. The low frequency part is split again into two parts of high and
low frequencies [21]. In image processing, similar to the onedimensional DWT, the DWT for a two-dimensional image can be
constructed by applying the one-dimensional DWT at horizontal
and vertical directions, respectively. The process of 2D-DWT is
shown in Fig. 1, in which an image is rst ltered in the horizontal
direction with low-pass and high-pass lters. Then the ltered outputs are downsampled by a factor of 2. Moreover, the same process
is applied in the vertical direction. Thus, an image is decomposed
into 4 sub-bands denoted by LL; HL; LH; HH. Each sub-band can be
thought of a smaller version of the image representing different
properties. The sub-band LL is a coarser approximation to the original image; the sub-bands HL and LH record the changes of the
image along horizontal and vertical directions, respectively; and
the sub-band HH shows the changes of the image along diagonal
direction. The sub-bands HL; LH and HH are all the high frequency
components of the image. Further decomposition can be conducted
on the LL sub-band. The 1-level 2D-DWT means that an original
image is decomposed into a low frequency sub-band and three
high frequency sub-bands. Generally, the t-level 2D-DWT (t > 1)
means that the low frequency sub-band obtained in the t 1level 2D-DWT is further decomposed into a low frequency
sub-band and three high frequency sub-bands. Fig. 2(a) shows
the owchart of a 2-level 2D-DWT and Fig. 2(b) shows the face
image with a 1-level 2D-DWT.
Generally, sub-band LL is the low frequency component of the
image which contains most information of the original image;
and sub-bands HL; LH and HH stand for the high-frequency components of the image which reect the details of images. Since the
low frequency sub-band plays a dominant role in four sub-bands
for approximation of the original image, if a t-level wavelet-based
method is used, then the low frequency sub-band obtained by the
t-th-level wavelet transform (i.e., the top-levels low frequency
sub-band) is usually single-handed used for identication in image
recognition in the literature. However, although the high frequency
sub-bands do not contain as much information as the low frequency sub-band, they also contain much important information.
Thus, it is more reasonable to use the information of all top-levels
sub-bands for face recognition. Based on these analysis, we now
propose a new wavelet-based technique (TWSBF) for feature
extraction, which makes full use of all top-levels wavelet subbands. Such a technique is described as follows.
Let Ix; y be a face image, by using 1-level or multi-level
wavelet transform to Ix; y, we get the top-levels low frequency
sub-band matrix LL 2 Rmn and high frequency sub-bands matrices
HL; LH; HH 2 Rmn . Without loss of generality, let L; H; V; D 2 Rl represent the vectorization of LL; HL; LH; HH, where l m n. Suppose
97
to represent the face Ix; y. Fig. 3 vividly illustrates this fusion ow.
The appropriate function fH is focused in this article.
3. TWSBF plus PCA
3.1. Classical PCA
DL fL L 2 Rd1 ;
S u1 H u2 V u3 D Au 2 R ;
T
where u u1 ; u2 ; u3 2 R is a coefcient vector of the high frequency wavelet sub-band fusion, which gives a model of pixel-level
fusion of three top-levels high frequency wavelet sub-bands.
In order to make such a fusion technique of wavelet sub-bands
more effective, it is necessary to nd the optimal coefcient vector
u in (3) so that S is the best representation of A for face recognition.
Suppose that fH is the function of this process, and S is expressed as
DS fH S fH Au 2 Rd2 ;
DL
fL L
2 Rd1 d2
DS
fH S
The feature extraction is to transform data from a high dimensional space to a lower dimensional space, which can be linear or
nonlinear. PCA [6] is a linear dimension reduction technique which
extracts the desired principal components of the multidimensional
data.
Suppose that the set fx1 ; x2 ; . . . ; xm g with xi 2 Rl comes from m
training samples, then PCA is trying to nd a transformation matrix
P
W 2 Rlc such that yi W T xi x, where yi 2 Rc and x m1 m
i1 xi is
the mean of the samples. To make yi be on behalf of xi in the projection subspace, the covariance of yi must be as large as possible.
The covariance matrix of yi in projection subspace is dened by
m
m h
ih
iT
X
X
yi yTi
W T xi x W T xi x
i1
"
m
X
T
i1
xi xxi xT W W T SPCA W;
i1
P
T
where SPCA m
the covariance matrix of
i1 xi xxi x is
training samples. So, the classical PCA needs to solve the following
optimization problem to obtain the transformation matrix W:
SPCA w kw:
Suppose that w1 ; w2 ; . . . ; wc are eigenvectors of (6) which correspond to the c largest eigenvalues k1 ; k2 ; . . . ; kc satisfying
k1 P k2 P P kc , then the solution of (5) is W w1 w2 wc .
3.2. Model of TWSBF plus PCA
In this subsection, we combine TWSBF with PCA to propose a
new model for feature extraction. Let m be the number of training
98
Comparing the classical PCA model (5) with the new model (9),
it is easy to see that the new model (9) is more difcult to solve
than the classical PCA model (5), since, in the new model (9), we
need to nd the optimal coefcient vector u besides the transformation matrix W. Fortunately, the new model (9) is computationally tractable, which is investigated in the following subsection.
Si u1 Hi u2 V i u3 Di Ai u 2 Rl ;
In order to design the effective method to solve the optimization problem (9), the following property is a key.
SW; u
"
#
m
X
T
Ai u AuAi u Au W
i1
"
#
m
X
T
Ai AuuT Ai A W:
i1
m
X
Ai AuuT Ai A 2 Rll ;
i1
!
m
X
T
tr
uT Ai A WW T Ai Au
i1
T
i1
tr u
m
X
! !
T
Ai A WW Ai A u
i1
truT Lc Wu;
SW; u W T Sc uW:
Therefore, we need to solve the following optimization problem
to obtain the optimal transformation matrix W and the optimal
coefcient vector u:
m
X
T
trW Sc uW tr
W T Ai AuuT Ai A W
T
W;u
10
Then, we have
i1
i1
Sc u
m
X
T
Ai A WW T Ai A 2 R33 :
#
m
X
T
Si SSi S W
WT
Lc W
"
WT
W
i1
m
m h
ih
iT
X
X
yi yTi
W T Si S W T Si S
i1
99
k1
; uk1 trSW k ; uk < e;
trSW
then
Update u uk1 ; k k 1: Go to 4:
end if
7: Compute DSi through u and W . Then implement featurelevel fusion on DLi and DSi and obtain F i .
4. TWSBF plus LDA
4.1. Classical LDA
From Section 3, we know that PCA is a dimension reduction
method without discrimination because there is no use of the sample label information during the discussion of PCA. However, the
dimension reduction method with discrimination is more advantageous for classication problems in general. LDA is a widely used
criterion of dimension reduction method whose main idea is to
make the samples distribution in the same class in the low dimensional space compact as much as possible (minimize within-class
scatter); and make the distance of samples in different classes in
the low dimensional space separate as much as possible (maximize
between-class scatter) [23,24].
Given c training classes of faces. Suppose that each class i has mi
training face images and m is the total number of face images satP
isfying ci1 mi m. Let xij be the vector from j-th training image in
e
Sb
i1
yi yyi yT
i1
"
c
X
mi
T
i1
e
Sw
c
X
mi
i1
WT
W T xi W T xW T xi W T x
"
"
i
T
1 X
W T xij W T xi W T xij W T xi
m mi 1 j1
(
"
c
X
mi
i1
xi xxi xT W W T Sb W;
i
1 X
y yi yij yi T
m mi 1 j1 ij
c
X
mi
i1
c
X
mi
i
1 X
xij xi xij xi T
m mi 1 j1
#)
W W T Sw W:
max
W
trW T Sb W
trW T Sw W
11
Sb wi kSw wi :
Suppose that fwi ji 1; 2; . . . ; cg is the set of generalized eigenvectors of matrices Sb and Sw corresponding to the c largest generalized eigenvalues fki ji 1; 2; . . . ; cg. Then the solution of (11) is
W w1 wc .
4.2. Model of TWSBF plus LDA
Given a set of training face images with class labels, we can generate a pixel-level combined feature-level fusion of the wavelet
sub-bands for each image. In order to achieve the better recognition performance, we combine TWSBF with the classical LDA to
propose a new model for feature extraction.
Let c be the number of training face classes, and Lij ; Hij ; V ij ; Dij be
the top-levels wavelet sub-bands of j-th face image in class i,
where i 1; 2; . . . ; c; j 1; 2; . . . ; mi , and mi denotes the number
of training samples in class i. Firstly, we implement dimension
reduction in the low frequency sub-band Lij through LDA and
obtain DLij . Then we implement pixel-level fusion on the high frequency sub-bands as follows.
Let Aij Hij V ij Dij be the top-levels high frequency wavelet
sub-band matrix of j-th face image in class i, The mean high frequency wavelet image of training images in class i is
Pmi
Ai m1 j1
Aij Hi V i Di and the mean high frequency wavelet
i
P Pmi
Aij H V D, where
image of all training images is A m1 ci1 j1
P
m is the total number of training images satisfying m ci1 mi .
Suppose that the pixel-level fusion of three high frequency wavelet
sub-bands of Aij Hij V ij Dij is Sij , i.e.,
Sb
c
X
mi
i1
Sw
#
mi
1 X
T
xij xi xij xi ;
m mi 1 j1
c
X
mi
i1
xi xxi xT ;
"
where u u1 ; u2 ; u3 T 2 R3 is a coefcient vector of the high frequency wavelet sub-band fusion. Denote the mean vector in class
i by Si Ai u and the total mean vector by S Au. Then, the
between-class scatter matrix Sb u and the within-class scatter
matrix Sw u are dened, respectively, by
100
Sb u
c
c
X
X
T
T
Pi Si SSi S
Pi Ai AuAi Au
i1
i1
c
X
T
Pi Ai AuuT Ai A ;
12
i1
Sw u
c
X
Pi
"
i1
c
X
Pi
i1
c
X
Pi
i1
i
T
1 X
Sij Si Sij Si
mi 1 j1
1
mi 1
mi
X
Aij Ai uAij Ai u
tr W T Sw uW uT Lw Wu c;
T
i
T
1 X
Aij Ai uuT Aij Ai ;
mi 1 j1
trW T Sb uW
trW T Sw uW
13
W;u
15
16
i1
c
X
Pi
Lw W
i1
1
mi 1
mi
X
T
Aij Ai WW T Aij Ai 2 R33 :
Lb W
17
1: Let Lij ; Hij ; V ij ; Dij (i 1; 2; . . . ; c; j 1; 2; . . . ; mi ) be four wavelet sub-bands obtained by the t-level 2D-DWT.
2: Implement LDA dimension reduction on Lij and obtain DLij .
3: Denote Aij Hij V ij Dij . Initialize u u0 . Set Pi mi =m
(i 1; 2; . . . ; c) and k 0.
4: Compute Sb u and Sw u by (12) and (13), respectively.
5: Compute ST u by (18); and its eigenvectors v 1 ; . . . ; v d corresponding to the d largest eigenvalues of ST u.
6: Let V v 1 ; . . . ; v d . Compute Sb u and Sw u according to
Sb u V T Sb uV; Sw u V T Sw uV, and their generalized
eigenvectors y1 ; . . . ; yc corresponding to the c largest eigenvalues of Sb u and Sw u. Let w1 Vy1 ; . . . ; wc Vyc ;
W k1 w1 w2 wc .
7: Compute Lb W k1 and Lw W k1 by (16) and (17), respectively; and their generalized eigenvector uk1 corresponding
to the largest eigenvalue of Lb W k1 and Lw W k1 .
8: If k > 0 and
k1
; uk1 JW k ; uk < e;
JW
then
j1
Then, we have
else
ST u
c X
i
T
1X
Sij SSij S
m i1 j1
m
c X
i
T
1X
Aij AuuT Aij A :
m i1 j1
uT 1=cLw Wu 1:
14
max tr W T Sb uW s:t: W T Sw uW I;
i:e:;
j1
where P i is the prior probability of class i which is generally calculated with P i mi =m. From the analysis of Subsection 4.1, we need
to maximize the following function:
JW; u
18
Update u uk1 ; k k 1: Go to 4:
end if
9: Compute DSij through u and W . Then implement featurelevel fusion on DLij and DSij and obtain F ij .
5. Numerical experiments
Based on the observation although the high frequency subbands do not contain as much information as the low frequency
sub-band, they also contain important information corresponding
to the edge and contour of face, reecting the details of face, we
have proposed two methods for face recognition in Sections 3
and 4, respectively. Compared with traditional methods which
only single-handed use the low frequency information, the proposed methods not only use the low frequency information, but
also make full use of the top-levels high frequency information.
101
102
Table 1
Average recognition rates on the FERET database.
Method
34.09
38.37
34.96
39.65
40.56
45.51
42.75
47.17
54.00
59.55
41.78
50.73
43.40
54.06
45.93
57.26
46.60
59.45
58.90
73.00
Table 2
Average recognition rates on the FERET database.
Table 4
Average recognition rates on the ORL database.
Method
Method
GMTR (%)
PCRC (%)
CRC (%)
SRC (%)
PNN (%)
TWSBF PCA (%)
TWSBF LDA (%)
6.21
22.85
28.75
28.57
39.92
38.37
50.73
10.70
27.65
35.57
32.23
44.81
39.65
54.06
13.88
29.00
40.11
35.93
53.96
45.51
57.26
14.72
23.37
39.30
33.07
56.87
47.17
59.45
20.25
24.60
48.55
40.25
74.30
59.55
73.00
GMTR (%)
PCRC (%)
CRC (%)
SRC (%)
PNN (%)
TWSBF PCA (%)
TWSBF LDA (%)
76.53
67.64
86.46
87.28
88.25
88.10
92.07
83.45
72.79
89.20
89.91
92.66
92.58
95.04
87.35
76.45
92.25
92.30
94.95
94.25
97.15
89.62
77.93
92.87
93.00
96.12
95.50
97.87
91.91
79.58
93.00
93.25
96.66
97.25
98.66
91.37
79.62
93.37
93.75
97.25
97.50
99.00
Table 3
Average recognition rates on the ORL database.
Table 5
Average recognition rates on the AR database.
Method
Method
10
11
12
13
14
86.64
88.10
90.54
92.58
92.60
94.25
94.06
95.50
95.83
97.25
96.00
97.50
83.56
84.80
87.73
88.55
91.08
91.33
92.04
92.28
93.36
93.69
93.37
93.40
91.00
92.07
94.12
95.04
97.10
97.15
97.62
97.87
97.66
98.66
98.25
99.00
90.47
94.23
93.35
95.76
95.35
97.20
95.87
97.36
96.32
97.51
96.35
97.56
103
10
11
12
13
14
GMTR (%)
PCRC (%)
CRC (%)
SRC (%)
PNN (%)
TWSBF PCA (%)
TWSBF LDA (%)
34.08
85.13
89.79
91.07
92.24
84.80
94.23
36.23
86.94
90.26
91.40
93.48
88.55
95.76
37.88
87.45
90.31
91.56
94.38
91.33
97.20
39.10
87.91
89.39
90.48
95.22
92.28
97.36
40.66
88.09
87.06
88.44
95.82
93.69
97.51
42.06
88.29
83.88
85.67
95.99
93.40
97.56
Table 7
Average recognition rates on the AR database using the db4 wavelet.
Method
10
11
12
13
14
86.41
89.26
89.10
91.73
91.59
93.93
92.42
94.38
93.26
95.08
93.54
95.42
91.45
94.06
94.31
95.22
96.15
96.88
96.27
97.07
96.74
97.27
96.58
97.09
Secondly, no matter which kind of wavelet to be chosen, our methods achieve the higher recognition rates on all experiments than
DWT + PCA and DWT + LDA.
6. Conclusions
In this paper, we proposed two effective methods of feature
extraction for face recognition. There are four important points in
our methods: Firstly, the data fusion technique proposed in this
paper ensures four top-levels wavelet sub-bands are all used,
which avoids the top-levels information to be lost. Secondly, in
order to make the proposed high frequency sub-bands pixel-level
fusion technique more effective, by combining it with PCA and
LDA respectively, we converted the problem of nding the best
fusion coefcients to two optimization problems and designed
two alternative direction methods to solve the corresponding
problems. Thirdly, the optimal fusion coefcients of wavelet subbands and the optimal transformation matrices of dimension
reduction were obtained simultaneously. Fourthly, we combined
feature-level fusion with pixel-level fusion effectively.
From the numerical results given in Section 5, it is easy to see
that the recognition rates of our methods are always higher than
those of the classical wavelet-based methods, which demonstrates
that the high frequency sub-bands indeed contain some important
information and our data fusion technique by making full use of all
top-levels wavelet sub-bands indeed has some advantage than the
traditional methods which only single-handed use the low frequency information; it is also easy to see that the recognition rates
of our methods are always higher than those of popular methods
recently given in the literature, which demonstrates that the methods proposed in this paper are effective for face recognition.
Some further issues are worth studying in the future. Firstly, our
methods were proposed for face recognition. We believe the methods proposed in this paper can be further used in some other elds
about data fusion of images. Secondly, our methods were proposed
by combining the sub-band fusion technique with PCA and LDA,
respectively. In fact, many other approaches for dimension reduction have been proposed in the literature, such as independent
component analysis, kernel principal component analysis, kernel
linear discriminant analysis, and so on. It is possible that some
more effective methods can be proposed by combining the subband fusion technique proposed in this paper with other
approaches for dimension reduction. Thirdly, in our methods, only
four top-levels wavelet sub-bands were effectively fused. It is possible that the proposed fusion technique of the top-levels subbands can be further developed to fuse the information of more
wavelet sub-bands including those located in other wavelet levels
instead of the top-level only. Fourthly, the recognition rates in the
FERET database are low generally. The FERET database contains
images under different poses, expression and illumination. In our
opinion, the pose inuences the rates greatly. It is our further work
to revise our methods in pose estimation.
Acknowledgments
This work was partially supported by the National Natural Science Foundations of China (Grant No. 11171252).
104
References
[1] W. Zhao, R. Chellappa, P.J. Phillips, A. Rosenfeld, Face recognition: a literature
survey, ACM Comput. Surv. 35 (4) (2003) 399458.
[2] Y. Cheng, Y.K. Hou, C.X. Zhao, Z.Y. Li, Y. Hu, C.L. Wang, Robust face recognition
based on illumination invariant in nonsubsampled contourlet transform
domain, Neurocomputing 73 (1012) (2010) 22172224.
[3] X.Z. Zhang, Y.S. Gao, Face recognition across pose: a review, Pattern Recogn. 42
(11) (2009) 28762896.
[4] A.M. Martinez, Recognizing imprecisely localized, partially occluded, and
expression variant faces from a single sample per class, IEEE Trans. Pattern
Anal. Machine Intell. 24 (6) (2002) 748763.
[5] J.W. Zhao, Z.H. Zhou, F.L. Cao, Human face recognition based on ensemble of
polyharmonic extreme learning machine, Neural Comput. Appl. 24 (6) (2014)
13171326.
[6] I.T. Jolliffe, Principal Component Analysis, rst ed., Springer, New York, USA,
1986.
[7] P.N. Belhumeur, J.P. Hespanha, D. Kriegman, Eigenfaces vs. Fisherfaces:
recognition using class specic linear projection, IEEE Trans. Pattern Anal.
Machine Intell. 19 (7) (1997) 711720.
[8] K.C. Kwak, W. Pedrycz, Face recognition using fuzzy integral and wavelet
decomposition method, IEEE Trans. Syst. Man Cybernet. Part B: Cybernet. 34
(4) (2004) 16661675.
[9] D.Q. Li, W. Pedrycz, N.J. Pizzi, Fuzzy wavelet packet based feature extraction
method and its application to biomedical signal classication, IEEE Trans.
Biomed. Eng. 52 (6) (2005) 11321139.
[10] C.L. Fan, X.T. Chen, N.D. Jin, Research of face recognition based on wavelet
transform and principal component analysis, in: Proceedings of Eighth
International Conference on Natural Computation, 2012, pp. 575578.
[11] J.T. Chien, C.C. Wu, Discriminant waveletfaces and nearest feature classiers
for face recognition, IEEE Trans. Pattern Anal. Machine Intell. 24 (12) (2002)
16441649.
[12] V. Vidya, N. Farheen, K. Manikantan, S. Ramachandran, Face recognition using
threshold based DWT feature extraction and selective illumination
enhancement technique, Proc. Technol. 6 (2012) 334343.
[13] D.V. Jadhav, R.S. Holambe, Feature extraction using Radon and wavelet
transforms with application to face recognition, Neurocomputing 72 (79)
(2009) 19511959.
[14] B. Li, Y.H. Liu, When eigenfaces are combined with wavelets, Knowl.-Based
Syst. 15 (56) (2002) 343347.
[15] H.F. Hu, Variable lighting face recognition using discrete wavelet transform,
Pattern Recogn. Lett. 32 (13) (2011) 15261534.
[16] J. Yang, J.Y. Yang, D. Zhang, J.F. Lu, Feature fusion: parallel strategy vs. serial
strategy, Pattern Recogn. 36 (6) (2003) 13691381.