covga a novel descriptor based on symmetry of regions for...

12
CovGa: A novel descriptor based on symmetry of regions for head pose estimation Bingpeng Ma a,n , Annan Li b , Xiujuan Chai c , Shiguang Shan c a School of Computer and Control Engineering, University of Chinese Academy of Sciences, Beijing, China b Institute for Infocomm Research, Agency for Science, Technology and Research, Singapore c Key Lab of Intelligent Information Processing of Chinese Academy of Sciences (CAS), Institute of Computing Technology, CAS, Beijing 100190, China article info Article history: Received 14 January 2014 Received in revised form 8 April 2014 Accepted 6 June 2014 Communicated by Qingshan Liu Available online 14 June 2014 Keywords: Head pose estimation Covariance descriptors Gabor lters Symmetry abstract This paper proposes a novel method to estimate the head yaw rotation using the symmetry of regions. We argue that the symmetry of 2D regions located in the same horizontal row is more intrinsically relevant to the yaw rotation of head than the symmetry of 1D signals, while at the same time insensitive to the identity of the face. Specically, the proposed method relies on the effective combination of Gabor lters and covariance descriptors. We rst extract the multi-scale and multi-orientation Gabor representations of the input face image, and then use covariance descriptors to compute the symmetry between two regions in terms of Gabor representations under the same scale and orientation. Since the covariance matrix can alleviate the inuence caused by rotations and illumination, the proposed method is robust to such variations. In addition, the proposed method is further improved by combining it with a metric learning method named aa KISS MEtric learning (KISSME). Experiments on four challenging databases demonstrated that the proposed method outperformed the state of the art. & 2014 Elsevier B.V. All rights reserved. 1. Introduction During the last decades, there has been a signicant progress in the face recognition research. However, robust and accurate face recognition is still a challenging problem especially because of pose variations. To achieve robustness to pose variations, one might have to process face images differently according to their poses. Therefore, head pose estimation has been an active research topic for many years. In this paper, we focus on the challenging problem of estimating the head yaw pose from the input face images. Head pose estimation from images is a challenging problem due to large variations because of pose, illumination, facial expressions, and subject variability. A generic (i.e. person-inde- pendent) algorithm for head pose estimation has to be robust to such intrinsic variations and extrinsic variations such as occlusion, noise, illumination and perspective distortion. The different meth- ods about head pose estimation are introduced in Section 2. The survey on the topic of head pose estimation can be found in [1]. Recently, Ma et al. have proposed the GaFour method which estimates the head yaw pose by using the symmetry of the facial appearance [2] and shows great improvement in head pose estimation. They argue that the symmetry of the intensities in each row of the face image is closely relevant to the yaw rotation of head. Their motivation comes from the observation that with the pose varying from the front to the half-prole, the symmetry of the face image decreases gradually. Since this decrease of the symmetry is insensitive to the identity of the input face, it can be explored to estimate the head pose. Specically, in GaFour, taking the intensities of each row of the face image as a 1D signal, 1D Gabor lters are rst convolved with the row signals to reduce noise and extract the local information, and then Fourier analysis is applied to compute the symmetry features of the head, i.e., to represent the pose. Although the symmetry based on the 1D signal is related to the head pose, it is easily affected by other factors, such as the misalignment. For example, when there is a rotation in the plane for a frontal face image, the pixels in the same row are no longer symmetrical. In fact, the problem of misalignment happens inevitably since the input images in the head pose estimation problem often come from the output of automatic face detectors. Generally speaking, many automatic face detectors only care about whether there are faces in the image but not their orientations. In this scenario, the robustness to misalignment is one of the draw- backs of the methods based on the 1D signals. In this paper, we propose a novel method to improve the accuracy of head pose estimation using the symmetry of 2D regions of the input face image. In the proposed method, 2D Gabor lters are rst convolved with the input face image to Contents lists available at ScienceDirect journal homepage: www.elsevier.com/locate/neucom Neurocomputing http://dx.doi.org/10.1016/j.neucom.2014.06.014 0925-2312/& 2014 Elsevier B.V. All rights reserved. n Corresponding author. E-mail addresses: [email protected] (B. Ma), [email protected] (A. Li), [email protected] (X. Chai), [email protected] (S. Shan). Neurocomputing 143 (2014) 97108

Upload: others

Post on 31-Dec-2019

1 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: CovGa A novel descriptor based on symmetry of regions for ...vipl.ict.ac.cn/uploadfile/upload/2015116181238205.pdf · the face recognition research. However, robust and accurate face

CovGa: A novel descriptor based on symmetry of regions for headpose estimation

Bingpeng Ma a,n, Annan Li b, Xiujuan Chai c, Shiguang Shan c

a School of Computer and Control Engineering, University of Chinese Academy of Sciences, Beijing, Chinab Institute for Infocomm Research, Agency for Science, Technology and Research, Singaporec Key Lab of Intelligent Information Processing of Chinese Academy of Sciences (CAS), Institute of Computing Technology, CAS, Beijing 100190, China

a r t i c l e i n f o

Article history:Received 14 January 2014Received in revised form8 April 2014Accepted 6 June 2014Communicated by Qingshan LiuAvailable online 14 June 2014

Keywords:Head pose estimationCovariance descriptorsGabor filtersSymmetry

a b s t r a c t

This paper proposes a novel method to estimate the head yaw rotation using the symmetry of regions.We argue that the symmetry of 2D regions located in the same horizontal row is more intrinsicallyrelevant to the yaw rotation of head than the symmetry of 1D signals, while at the same time insensitiveto the identity of the face. Specifically, the proposed method relies on the effective combination of Gaborfilters and covariance descriptors. We first extract the multi-scale and multi-orientation Gaborrepresentations of the input face image, and then use covariance descriptors to compute the symmetrybetween two regions in terms of Gabor representations under the same scale and orientation. Since thecovariance matrix can alleviate the influence caused by rotations and illumination, the proposed methodis robust to such variations. In addition, the proposed method is further improved by combining it witha metric learning method named aa KISS MEtric learning (KISSME). Experiments on four challengingdatabases demonstrated that the proposed method outperformed the state of the art.

& 2014 Elsevier B.V. All rights reserved.

1. Introduction

During the last decades, there has been a significant progress inthe face recognition research. However, robust and accurate facerecognition is still a challenging problem especially because ofpose variations. To achieve robustness to pose variations, onemight have to process face images differently according to theirposes. Therefore, head pose estimation has been an active researchtopic for many years. In this paper, we focus on the challengingproblem of estimating the head yaw pose from the input faceimages.

Head pose estimation from images is a challenging problemdue to large variations because of pose, illumination, facialexpressions, and subject variability. A generic (i.e. person-inde-pendent) algorithm for head pose estimation has to be robust tosuch intrinsic variations and extrinsic variations such as occlusion,noise, illumination and perspective distortion. The different meth-ods about head pose estimation are introduced in Section 2. Thesurvey on the topic of head pose estimation can be found in [1].

Recently, Ma et al. have proposed the GaFour method whichestimates the head yaw pose by using the symmetry of the facialappearance [2] and shows great improvement in head pose

estimation. They argue that the symmetry of the intensities ineach row of the face image is closely relevant to the yaw rotationof head. Their motivation comes from the observation that withthe pose varying from the front to the half-profile, the symmetryof the face image decreases gradually. Since this decrease of thesymmetry is insensitive to the identity of the input face, it can beexplored to estimate the head pose. Specifically, in GaFour, takingthe intensities of each row of the face image as a 1D signal, 1DGabor filters are first convolved with the row signals to reducenoise and extract the local information, and then Fourier analysisis applied to compute the symmetry features of the head, i.e., torepresent the pose.

Although the symmetry based on the 1D signal is related to thehead pose, it is easily affected by other factors, such as themisalignment. For example, when there is a rotation in the planefor a frontal face image, the pixels in the same row are no longersymmetrical. In fact, the problem of misalignment happensinevitably since the input images in the head pose estimationproblem often come from the output of automatic face detectors.Generally speaking, many automatic face detectors only care aboutwhether there are faces in the image but not their orientations. Inthis scenario, the robustness to misalignment is one of the draw-backs of the methods based on the 1D signals.

In this paper, we propose a novel method to improve theaccuracy of head pose estimation using the symmetry of 2Dregions of the input face image. In the proposed method, 2DGabor filters are first convolved with the input face image to

Contents lists available at ScienceDirect

journal homepage: www.elsevier.com/locate/neucom

Neurocomputing

http://dx.doi.org/10.1016/j.neucom.2014.06.0140925-2312/& 2014 Elsevier B.V. All rights reserved.

n Corresponding author.E-mail addresses: [email protected] (B. Ma), [email protected] (A. Li),

[email protected] (X. Chai), [email protected] (S. Shan).

Neurocomputing 143 (2014) 97–108

Page 2: CovGa A novel descriptor based on symmetry of regions for ...vipl.ict.ac.cn/uploadfile/upload/2015116181238205.pdf · the face recognition research. However, robust and accurate face

alleviate the influence of noise and illumination and extract thelocal information. The Gabor feature at each scale and orientationis then divided into many regions with the same size. Covariancedescriptor is used to extract the similarity of the symmetricalregions in the same horizontal row. The similarities of all thescales and orientations are concatenated to form the symmetryrepresentation of the face image. To further enhance the discrimi-native ability and reduce the dimension of the proposed repre-sentation, metric learning methods, specifically, KISS MEtriclearning (KISSME) [3], are applied to extract features. To gain thefinal pose, the Nearest Centroid (NC) classifier is exploited finally.

Compared with the symmetry of the 1D signals in GaFour, thesymmetry of the 2D regions in the proposed representation ismore relevant to pose variations. On the one hand, when the in-plane rotation happens, even if a point and its symmetric point inthe face are not in the same row of the image, they are mostprobably still in the symmetrical regions. By using the symmetryof regions, the influence of misalignment can be reduced greatly.On the other hand, the symmetry features of GaFour are extractedby Fourier transform, which means that even though the featuresinclude the information of symmetry, they are not a directsymmetry measure. On the contrary, in the proposed method,the similarity of 2D regions is computed by the distance betweencovariance descriptors. The similarity is taken as the symmetrymeasure of regions. So, we conjecture that the proposed descriptorcan outperform GaFour.

The remaining part of this paper is organized as follows: inSection 2 we introduce the related methods for head poseestimation; in Section 3, we show the advantage of the symmetryof 2D regions compared with 1D signals and the motivation ofusing the symmetry of 2D regions to improve the performance ofhead pose estimation. In Section 4, the proposed method isintroduced in detail. Experiments on four challenging databasesare shown in Section 5 to demonstrate the effectiveness of theproposed method. Conclusions are drawn in Section 6 with somediscussions on the future work.

2. Related work

During the last decades, more and more researches put theirattentions on the task of head pose estimation. There exists a largeamount of literature on the topic of head pose estimation, see [1]for a review. Broadly speaking, most previous work can becategorized into three main groups: algorithms based on facialfeatures [4–6], model-based algorithms [7,8], and appearance-based algorithms [9,10].

In the algorithms based on facial features, the 3D face structureis exploited along with a priori anthropometric information inorder to define the head pose. Representative geometric featuresused to estimate the head pose include the elliptic shape of theface, the mouth–nose region geometry, the line connecting the eyecenters, the line connecting the mouth corners and the facesymmetry. This category of algorithms has a major disadvantage:they are sensitive to the misalignment of the facial feature points,while the accurate and robust localization of facial landmarksremains as an open problem, especially for the non-frontal faces.

Using the 3D structure of human head, the model-basedalgorithms build a priori known 3D models for human faces andattempt to match the facial features such as the face contour andthe facial components of the 3D face model with their 2Dprojections. Once the correspondences from 3D to 2D are foundbetween the input data and the face model, conventional poseestimation techniques are exploited to provide the head pose. Themain problem for these algorithms is that it is difficult to precisely

build the head model for different persons and to define the bestmapping of the 3D model to the 2D face images.

The appearance-based algorithms typically assume that thereexists a certain relationship between the 3D face pose and someproperties of the 2D facial image and infer the relationship byusing a large number of training images and statistical learningtechniques. Intuitively, these appearance-based algorithms cannaturally avoid the drawbacks of the model-based methods.Therefore, they have attracted more and more attentions. In thesealgorithms, instead of using facial landmarks or face models, thewhole image of the face is used for pose classification.

Specially, Gong et al. studied the trajectories of multi-viewfaces in linear Principal component analysis (PCA) feature space[11,12]. They use two Sobel operators (horizontal and vertical) tofilter the training images. PCA is then performed to reduce thedimensionality of the training examples. Finally, SVM regression isutilized to construct two pose estimators, for the tilt and yawangles. Darrell et al. compute a separate eigen-space for each faceunder each possible pose [13]. The head pose is determined byprojecting the input image onto each eigen-space and selectingthe one with the lowest residual error. In some sense, this methodcan be formulated as an MAP estimation problem. Li et al. exploitIndependent Component Analysis (ICA) and its variants, subspaceanalysis and topographic ICA for pose estimation [14]. ICA takesinto account higher order statistics required to characterize theview of objects and suitable for the learning of view subspaces.Wei et al. propose that the optimal orientation of the Gabor filterscan be selected for each pose to enhance pose information andeliminate other distractive information like variable facial appear-ance or changing environmental illumination [10]. In theirmethod, a distribution-based pose model is used to model eachpose cluster in Gabor eigen-space.

Since the set of the facial images with various poses intrinsicallyforms a 3D manifold in image space, manifold learning [15–17] forhead pose estimation is getting popular recently [18–23]. In [18], bythinking globally and fitting locally, Fu and Huang propose to usethe graph embedded analysis method for head pose estimation.They first construct the neighborhood weighted graph in the senseof supervised locally linear embedding [15]. The unified projectionis calculated in a closed-form solution based on the graph embed-ding linearization, and then project new data into the embeddedlow-dimensional subspace with the identical projection. To over-come the disadvantage that most embedding based methods areunsupervised in nature and do not extract features that incorporateclass information, in [24], Huang et. al. present the methodSupervised Local Subspace Learning (SL2), which learns a locallinear model from a sparse and non-uniformly sampled trainingset. The authors argue that SL2 is robust to under-sampled regions,over-fitting and image noise.

After extracting the representation of the face image, neuralnetworks, Bayesian approaches and Boosting are also applied to getthe final angle of the input image. In [25], a neural network-basedapproach is presented, in which a multilayer perception is trained foreach pose angle (pan and tilt) by feeding it with preprocessed facialimages captured by a panoramic camera. In [26,27], based on aBayesian formulation, Ba et al. propose an algorithm that coupleshead tracking and pose estimation in a mixed state particle filterframework. In [28], the authors use boosting regression and simpleHaar-type features to estimate the head pose.

With the recent development and availability of 3D sensingtechnologies are becoming ever more afford able and reliable,more and more researchers use the additional depth informationto over come some problems inherent of methods based on 2Ddata [29,30].

In our case, the proposed method belongs to the appearance-based methods. The features used by all the above methods are

B. Ma et al. / Neurocomputing 143 (2014) 97–10898

Page 3: CovGa A novel descriptor based on symmetry of regions for ...vipl.ict.ac.cn/uploadfile/upload/2015116181238205.pdf · the face recognition research. However, robust and accurate face

extracted from the entire face. On the one hand, the features aregenerally vectorized as 1D vectors which lose the face structure insome sense; on the other hand, these features contain not onlypose information, but also information about identity, lighting,expression, etc. Different with the above methods, the proposedmethod is based on a commonly accepted assumption that humanhead is bilaterally symmetrical and the symmetry in the 2D imageis decreased gradually when the face rotates from the front to theprofile. Motivated by this assumption, our feature is based on thesimilarity of face regions. In other words, our feature is indeedclosely related to the pose variations and at the same timeindependent of other facial variations, especially the identity.

3. Relationship between pose variations and symmetryof regions

In [2], the authors illustrate their motivation of using thesymmetry of 1D row signals to estimate the head pose. Theyintroduce the relationship between the center line of the imagesand the symmetry plane of heads. In Fig. 1, the dash line denotesthe center line of the 2D images while the solid line denotes thesymmetry line of 3D faces. In the front-view image, the two linesare overlapping. With the head pose varying from the front to thehalf-profile, the deviation between two lines increases gradually.This deviation is related to pose variations and the symmetry ofthe images at the same time. Finally, the authors conclude that thesymmetry is closely relevant to pose variations.

Though the symmetry is relevant to pose variations, thesymmetry defined in [2] is based on the 1D row signals, whichis easily affected by other factors. For example, when there is arotation in the plane for a frontal face image, the pixels in the samerow are not symmetrical. In this section, we argue that comparedwith the symmetry of 1D row signals, the symmetry of 2D regionsis more relevant to the pose variations. To demonstrate this, weconduct the following two experiments using the images fromCAS-PEAL face database [31].

First, we define the symmetry measure of 1D signals and 2Dregions and show their measures on the CAS-PEAL database. Wedefine the symmetry measure E1D of 1D signals for a face image asfollows:

E1D ¼ 1wh

∑h

i ¼ 1∑w

j ¼ 1ðIij� ITijÞ2 ð1Þ

where Iij is the image intensity at position ði; jÞ, w and h are thewidth and the height of the input image, respectively. The image Iis flipped horizontally and we can gain a new face image IT. InFig. 2, we show the image I and its correlative IT. In fact, image Ihas the same information with image IT because the pixel (x, y) of Iis the same with the pixel ðwþ1�x; yÞ of IT. By using IT, the regionswith the same position of two images are much easier to beaccepted by human brain than the different positions of oneimage. Obviously, the lower the value of E1D, the greater theamount of symmetry and vice versa. For all the images in the CAS-PEAL database, we compute their symmetry measure E1D andshow the means and the standard deviations of E1D under differentposes in Fig. 3(a). In the figure, the horizontal axis represents thepose while the vertical axis shows the measure of symmetry. From

the figure, we can know that E1D is increased with pose varyingfrom the front to the half-profile, which means that the symmetryis decreased.

We also define the symmetry measure E2D of 2D regions. First,we divide a face image into many regions. We name a region usingthe position of the left-upper pixel. Then the symmetry measureE2D of the input face image is defined by

E2D ¼ 1n∑ijsymðRij;R

TijÞ ð2Þ

where symð�Þ is the similarity function of two regions, Rij is theregion in the original input image I and Rij

T is the region in image IT.In the experiment, we take the covariance descriptor as the

Fig. 1. The relationship between the symmetry plane of the head and the center lines of the images.

Fig. 2. A face image and its symmetric image.

−45 −30 −15 0 15 30 45−2

−1

0

1

2

−45 −30 −15 0 15 30 45−2

−1

0

1

2

Fig. 3. The symmetry measures of the different poses on the CAS-PEAL database.The horizontal axes represent the poses and the vertical axes represent themeasures. (a) The symmetry measure of the 1D signals. (b) The symmetry measureof 2D regions. This figure shows that the symmetry of 2D regions is more related tothe pose variations than that of 1D signals. The values in the figures are normalized.

B. Ma et al. / Neurocomputing 143 (2014) 97–108 99

Page 4: CovGa A novel descriptor based on symmetry of regions for ...vipl.ict.ac.cn/uploadfile/upload/2015116181238205.pdf · the face recognition research. However, robust and accurate face

function of symð�Þ, which is used in the proposed method and isintroduced in the following section. Obviously, the lower the valueof E2D, the greater the amount of symmetry and vice versa. Werepeat the experiment in Fig. 3(a) and show the means and thestandard deviations of the symmetry measure E2D of differentposes in Fig. 3(b). From Fig. 3(b), we can also clearly see that thesymmetry measures decrease when the pose varies from the frontto the half-profile, which means that the symmetry of 2D regionsis also related to the pose variations and can be applied in headpose estimation. Compared with the means of E1D in Fig. 3(a), themeans of E2D are more close to a straight line while the standarddeviations of E2D are much smaller than those of E1D when posevaries from the front to the half-profile. From the comparison, we

can conclude that the symmetry of 2D regions is more relevant tothe pose variations than the symmetry of 1D row signals.

We conduct the second experiment to show that the symmetryof 2D regions is more robust to the rotation in the plane than thatof 1D signals. For all the images in the CAS-PEAL database, we firstrotate them in the plane about 251 and then compute thesymmetry measures of 1D signals and 2D regions again. Themeans and the standard deviations of E1D and E2D under differentposes are shown in Fig. 4(a) and (b), respectively. From Fig. 4, wecan see that when the image is rotated in the plane, the symmetryof 1D signals is destroyed. But the symmetry measures of 2Dregions are still near a straight line when the poses varies from thefront to the right half profile and from �151 to the left half profile.From this scene, we can conclude that compared with thesymmetry of 1D signals, the symmetry of 2D region is more robustto the rotation in the plane.

On the whole, compared with the symmetry of 1D row signals,the symmetry of 2D regions is more related to the pose variantsand robust to the rotation in the plane. Considering the advantageof the symmetry of 2D regions, in this paper, we try to proposea novel method which use the symmetry of 2D regions to improvethe accuracy of head pose estimation.

4. Covariance descriptor of Gabor filters

In this section, we first introduce the proposed feature descrip-tor named Covariance Descriptor of Gabor filters (CovGa). Then,we introduce how to improve the discriminative ability of CovGaby using the metric leaning methods, e.g., KISSME. Fig. 5 shows theflowchart of the proposed CovGa. CovGa is a representation withtwo stages: Gabor features are first extracted and then thesymmetrical regions of Gabor filters are encoded by covariancedescriptors. In the following, we introduce each stage in detail.

4.1. Gabor features in CovGa

Though we have shown the advantage of the symmetry of 2Dregions for head pose estimation, in the proposed method, we stillcannot use the symmetry measure directly to estimate the headpose. On the one hand, it is difficult to distinguish the right pose

−45 −30 −15 0 15 30 45−2

−1

0

1

2

−45 −30 −15 0 15 30 45−2

−1

0

1

2

Fig. 4. The symmetry measures of the different poses on the CAS-PEAL database.The images are rotated in the plane about 251. (a) The symmetry of 1D signals.(b) The symmetry of 2D regions. This figure shows that compared with thesymmetry of 1D signals, the symmetry of 2D regions is more robust to the in-plane rotation. The values in the figures are normalized.

Ori.image

Gaborfeatures

MAXoperation

Covariancedescriptors

Symmetryunder

Covariancedescriptors

CovGafeatures

concatenateBand 1(Scale 1&2)

Band 2

Band 7

Band 8

... ... ... ...

G11

G21

G71

G81

C11,1

C21,1

C71,1

C81,1

C11,1’

C21,1’

C71,1’

C81,1’

d11,1d11,2

d21,1

d71,1

d71,1

d21,2

d71,2

d81,2

D={d11,1,..,dM,K,R}

Fig. 5. The flowchart of the proposed CovGa.

B. Ma et al. / Neurocomputing 143 (2014) 97–108100

Page 5: CovGa A novel descriptor based on symmetry of regions for ...vipl.ict.ac.cn/uploadfile/upload/2015116181238205.pdf · the face recognition research. However, robust and accurate face

and the left pose only from the measure. On the other hand, thestandard deviations are too large to estimate yaw accurately.However, considering the advantages of Gabor filters in facerecognition and the related areas, we exploit the multi-resolution and multi-orientation Gabor filters to de-compositethe input face images for sequential feature extraction. Then, wecompute the symmetry of the regions on each Gaborrepresentation.

The Gabor filters are inspired by the human visual system andtheir kernels are very similar to the 2D receptive field profiles ofthe mammalian cortical simple cells. For an image Iðx; yÞ, wecompute its convolution with Gabor filters according to thefollowing equation [32]:

Gðμ;νÞ ¼ Iðx; yÞnψμ;νðzÞ ð3Þ

where

ψμ;νðzÞ ¼‖kμ;ν‖2

σ2 eð�‖kμ;ν‖2‖z‖2=2σ2Þ½eikμ;νz�e�σ2=2� ð4Þ

kμ;ν ¼ kνeiϕμ ; kν ¼ 2�ðνþ2Þ=2π; ϕμ ¼ μπ8

ð5Þ

where μ and ν are the scale and orientation parameters, respec-tively. In our work, the number of scales is fixed to 16 while ν isquantized into 8 orientations.

In our method, two neighborhood scales (within the sameorientation) are grouped into one band (we therefore have8 different bands). We apply MAX pooling over two consecutivescales:

Gij ¼maxfGð2i�1; jÞ;Gð2i; jÞg; i; j¼ 1;…;8 ð6Þ

where i and j represent scales and orientations, respectively. TheMAX pooling operation increases the tolerance to small scalechanges which often occurs in face images since face images aremisaligned or only roughly aligned.

4.2. Covariance descriptor in CovGa

The critical component of the proposed CovGa is how tomeasure the symmetry of 2D regions. In this paper, we employthe covariance descriptor as the metric of the symmetry of 2Dregions. Covariance descriptor was firstly proposed by Tuzel et al.for object detection [33], and then widely used in other fields suchas pedestrian detection [34] and object tracking. Covariancedescriptor is able to capture shape, location and color information.It is shown that the performance of the covariance features issuperior to other methods as rotation and illumination changesare absorbed by the covariance matrix.

In the second step of CovGa, first, for each pixel of the Gaborrepresentation Gij, a 7-dimensional feature vector f ijðx; yÞ is com-puted to capture the spatial, intensity, texture and shape statistics:

f ijðx; yÞ ¼ ½x; y;Gijðx; yÞ;Gijx ðx; yÞ;Gijy ðx; yÞ;Gijxx ðx; yÞ;Gijyy ðx; yÞ� ð7Þ

where x and y are the pixel coordinate, Gijðx; yÞ is the Gabor featureat position (x, y), Gijx ðx; yÞ and Gijy ðx; yÞ are the gradient of image Gij

at position (x, y) in direction x and y , respectively. Gijxx ðx; yÞ andGijyy ðx; yÞ are the second-order gradient of image Gij at position(x, y) in direction x and y, respectively, which can be computed bythe convolution between ½�1 2 �1� and image Gij.

Then, Gabor features are divided into small overlapping rec-tangular regions. The covariance descriptor of region r is computedas

Cij;r ¼1

n�1∑

ðx;yÞA rðf ijðx; yÞ� f ij Þðf ijðx; yÞ� f ij ÞT ð8Þ

where f ij is the mean of all the f ijðx; yÞ in region r and n is thenumber of the pixels in the region.

The flipped image of Gabor feature image Gij is GijT and region rT

is the flipped version of region r. For regions r and rT, we take theirsimilarity under covariance descriptors as the symmetry ofregions. Since the covariance matrices are not in Euclidean space,we use the distance definition proposed by [33] to compute thesimilarity between regions r and rT:

dij;r ¼ dðCij;r ;Cij;rT Þ ¼ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi∑P

p ¼ 1ln2 λpðCij;r ;Cij;rT Þ

sð9Þ

where λpðCij;r ;Cij;rT Þ is the p-th generalized eigenvalues of Cij;r andCij;rT .

4.3. CovGa

Besides the symmetry of regions, considering the success ofGabor filters in many area, we also include the Gabor features ofregions in the representation of CovGa. Generally speaking, Gaborfeatures can be seen as the appearance-based features. As thesimilarity metric of Gabor features, the symmetry of regions canbe seen as the features' feature. To a certain extent, Gabor featuresand the symmetry of region are in the different level of featurerepresentations. In this scene, Gabor features can be the supple-ment of the proposed symmetry features.

Since the central point of CovGa is the symmetry of 2D regionsand we just select Gabor features as the supplement of thesymmetry feature, we do not use all the Gabor features directly.Instead, when we compute the similarity of two regions, we alsocompute the means of Gabor features of these two regions. Notethat Gij;r and Gij;rT are the mean of Gabor features of regions r andrT, respectively. We take the mean gij;r of Gij;r and Gij;rT as the Gaborfeatures of these two regions:

gij;r ¼Gij;rþGij;rT

2ð10Þ

This way, we need little extra computation while the Gaborfeatures is complemented in CovGa.

Since the values of dij;r and gij;r are not in the same range, beforecombining them together, dij;r and gij;r are normalized, respec-tively, as follows:

dij;r ¼dij;rffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi

∑Mi ¼ 0∑

Kj ¼ 0∑

Rr ¼ 0d

2ij;r

q ð11Þ

g ij;r ¼gij;rffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi

∑Mi ¼ 0∑

Kj ¼ 0∑

Rr ¼ 0g

2ij;r

q ð12Þ

where R is the number of regions of image Gij, M and K are thenumber of Gabor bands and orientations, respectively.

Finally, in CovGa, for region r, its representation Dij;r is

Dij;r ¼23dij;r

13g ij;r

� �ð13Þ

For simplicity reason and because it is not the central part of thepaper, we have empirically set the above mixing weights, givingmore emphasis to the proposed descriptor. Learning the weightsfrom the data should improve the results further.

Finally, the symmetry metrics of different regions underdifferent bands and orientations are concatenated to form theimage representation:

D¼ ðD11;1;…;D11;R;…;D1K ;R;…;DMK;RÞ ð14Þ

In CovGa, the similarity between two face images Ii and Ij isobtained by computing the Euclidian distance between their

B. Ma et al. / Neurocomputing 143 (2014) 97–108 101

Page 6: CovGa A novel descriptor based on symmetry of regions for ...vipl.ict.ac.cn/uploadfile/upload/2015116181238205.pdf · the face recognition research. However, robust and accurate face

representations Di and Dj:

dðIi; IjÞ ¼ JDi�Dj J ð15Þ

4.4. Analysis of CovGa

We argue that the advantage of CovGa resides in the followingaspects. First, since Gabor filters and covariance descriptors areboth known to be tolerant to illumination variations, the CovGarepresentation is also robust to illumination variations. In fact,illumination variations are still an open difficult problem whichaffects the accuracy of head pose estimation and face recognition.

Second, CovGa is also robust to the small misalignment. InCovGa, when computing the covariance matrix of the region, thestructure of region is destroyed. In other words, the covariancematrix of the region is based on the pixels not their position. Oncea small misalignment happens, the covariance matrix of the regionis mostly the same because the pixels in this region are still in thisregion though their positions have changed.

Finally, our descriptor makes a very different use of thecovariance descriptor. Generally, the similarity under covariancedescriptors is computed on two different images. On the contrary,CovGa computes the similarity of covariance descriptors withinthe same image. We take the similarities as the image signature,and the difference of probe and gallery images is obtained bycomputing the l2 distance between their signatures. By doing so,we avoid the difficult of computing the similarity between oneprobe image and each gallery image under covariance descriptors,which could be extremely time-consuming when the gallery setis large.

4.5. Enhancement of CovGa using metric learning

The dimensionality of the original CovGa representation is veryhigh because of the multi-scales and multi-orientations of Gaborfilters. Reducing the dimensionality makes the proposed methodmore efficient. Here we show that the simple methods, such asPrincipal Component Analysis (PCA) [35], can work well for theproposed representation. PCA is a traditional linear transformationtechnique, which can greatly reduce the dimensionality of fea-tures. The projection matrix Wpca is composed of the orthogonaleigenvectors of the covariance matrix of all the training samples.After using PCA, we can get the low dimension representation DP

of CovGa by DP ¼Wpca � D.In addition, head pose estimation, as a classification problem,

evidently needs the features to be discriminative besides theirgood representation ability. Therefore, we need to combine dis-criminant analysis methods or metric learning methods with theCovGa representation in order to improve the recognition perfor-mance. Both the discriminant analysis methods or metric learningmethods improve the discriminative ability by using the informa-tion of training sample's label and can be seen as the supervisedmethod. Specially, metric learning methods are based on the classof Mahalanobis distance functions, which has gained considerableinterest in the computer vision area.

In this paper, considering its great success in face recognitionand person re-identification [3], we use KISSME as our choice ofmetric learning methods to improve the discriminative ability ofCovGa. We propose the method Kiss Covariance Descriptor ofGabor filters (kCovGa), which is the supervised version of CovGa.As shown in the experiments, the performance of the supervisedmethod is much better than that of the unsupervised method.

KISSME is motivated by a statistical inference perspective basedon a likelihood-ratio test. It considers two independent generationprocesses for observed commonalities of similar and dissimilarpairs. The dissimilarity is defined by the plausibility of belonging

either to one or the other. Compared with other metric learningmethod, KISSME do not rely on a tedious iterative optimizationprocedure. Therefore, it is scalable to large datasets, as it justinvolves computation of two small sized covariance matrices. InKISSME, semi-definite matrix M is computed by the followingequation:

M¼Σ�1yij ¼ 1�Σ�1

yij ¼ 0 ð16Þ

where

Σyij ¼ 1 ¼ ∑yij ¼ 1

ðxi�xjÞðxi�xjÞT ð17Þ

Σyij ¼ 0 ¼ ∑yij ¼ 0

ðxi�xjÞðxi�xjÞT ð18Þ

yi is the label of sample xi. yij ¼ 1 means similar pairs, i.e., if thesamples share the same class label (yi¼yj) and yij ¼ 0 otherwise.More details about KISSME can be found in [3].

In kCovGa, we revise KISSME to combine it with CovGa morenaturally. Considering that the number of the negative pairs is farmore than that of the positive pairs, we introduce a weight factor λwhich is set to the ratio of the number of positive pairs andnegative pairs, and Eq. (16) can be rewritten as

M¼Σ�1yij ¼ 1�λΣ�1

yij ¼ 0 ð19Þ

where

λ¼ Number of positive pairsNumber of negative pairs

ð20Þ

By this way, we can alleviate the influence caused by theunbalance data. In general, the Mahalanobis distance metricmeasures the squared distance between two data points DP

i andDP

j as follows:

d2MðDPi ;D

Pj Þ ¼ ðDP

i �DPj ÞTMðDP

i �DPj Þ ð21Þ

But in head pose estimation, a projection matrix is more con-venient than the semi-definite matrix M in computing the projec-tion of a new sample. We then use Cholesky factorization toproduces an upper triangular matrix Wkiss which is fitted to

M¼WTkiss �Wkiss ð22Þ

We take Wkiss as the projection matrix of the new samples. It mustbe pointed out that the dimension of Wkiss is the same as that ofWpca. In other words, KISSME do not reduce the dimension offeatures.

Finally, for the input representation D of CovGa, the finalrepresentation DK of kCovGa can be gained by the followingequation:

DK ¼Wkiss �Wpca � D ð23Þ

4.6. CovGa for head pose estimation

Since the extraction of CovGa/kCovGa can be regarded as thepreprocessing step for yaw estimation, it should be combined withclassifiers to get the yaw pose of the input image. In this paper, NCis selected as the classifier to evaluate the performance of theproposed features. In the NC classifier, for the training sampleswith the same pose, the k-means method is applied to find the kcentroids. Then we compute the distance between the inputfeature and each class centroid, and take the label of the classwith the smallest Euclidean distance as the output label. Com-pared with the Nearest Neighbor (NN) classifier, the NC classifiercan eliminate the error caused by the identify since the imagedifference between the different poses of the same person may be

B. Ma et al. / Neurocomputing 143 (2014) 97–108102

Page 7: CovGa A novel descriptor based on symmetry of regions for ...vipl.ict.ac.cn/uploadfile/upload/2015116181238205.pdf · the face recognition research. However, robust and accurate face

less than the image difference between different persons with thesame pose.

5. Experiments

In this section, we perform experiments on four different headpose databases to show the effectiveness of CovGa an kCovGa.

5.1. Pointing'04 database

We first evaluate the performance of our descriptors on thePointing'04 Head Pose Image Database [36] which consists of 2790color face images for 15 subjects. The head pose of each personranges from �90 to 901 both in the horizontal and verticaldirections. The image of each subject has 93 different poses,including 13 yaw angles and 7 pitch angles, plus two extremecases with pitch angle 901 and �901. The bounding box contain-ing the face for each image is provided. All face images are down-scaled to 32�32 pixels in the gray-scale space. We estimate thehead pose of unknown faces by using a Leave One Person Out(LOPO) strategy. In LOPO, all images are used for training, exceptthe images of one subject, which will be used for testing. Theperson to be tested is then changed at each step. By this way, noimages of the same person are both in the training and testingparts. The final results are the average of all the tests. Somesamples of Pointing'04 database are shown in Fig. 6.

A comparison between our methods and other state-of-the-artmethods is shown in Table 1. In our experiment, yaw estimationand pitch estimation are handled separately. In Table 1, we showthe Mean Absolute Error (MAE) of CovGa and kCovGa when thenumber of center is 10. MAE is mean absolute error between the

continuous predicted pose and the ground truth pose. From thetable, we can see that our methods are the best on both yawestimation and pitch estimation. Compared with the advantage ofCovGa in pitch estimation, the advantage in yaw estimation ismore obviously. We attribute the advantage in yaw estimation tothe motivation of CovGa that CovGa is based on the symmetry ofthe regions, which is more related to yaw variations.

Besides the results from the references, some methods byourself also come true and we show their results under the NC

Fig. 6. The face images in the Pointing'04 database.

Table 1Comparison of the proposed techniques with leading methods on the Pointing'04 database.

Method Reference Yaw error Pitch error Notes

Neural network Stiefelhagen (2004) [37] 9.5 9.7 {13, 9}Human performance Gourier et al. [38] 11.8 9.4 {13, 9}Associative memories Gourier et al. [38] 10.1 15.9 {13, 9}High-order SVD Tu et al. [19] 12.9 17.97 {13, 9}PCA Tu et al. [19] 14.11 14.98 {13, 9}Locally embedded analysis Tu (2007) [19] 15.88 17.44 {13, 9}Random forest regression Li et al. [39] 9.6 13.9 {13, 9}Convex regularized sparse regression Ji et al. [40] 8.6 12.1 {13, 9}CovGa – 7.27 8.69 {13, 9}kCovGa – 6.24 7.14 {13, 9}

1 2 3 4 5 6 7 8 9 106

6.5

7

7.5

8

8.5

9

9.5

10

10.5

11

The number of center

The

mea

n ab

solu

te e

rror

PCALDAGaborGFCGaFourGFFFHOGsHOGSIFTsSIFTCovGakCovGa

Fig. 7. The MAEs of head yaw estimation on the Pointing'04 database. The x-axisrepresents the center number of each class and the y-axis represents the MAE.

B. Ma et al. / Neurocomputing 143 (2014) 97–108 103

Page 8: CovGa A novel descriptor based on symmetry of regions for ...vipl.ict.ac.cn/uploadfile/upload/2015116181238205.pdf · the face recognition research. However, robust and accurate face

classifier. We compare the performance of CovGa with thefollowing unsupervised methods: PCA, GaFour, HOG, Gabor andSIFT. We also compare the performance of kCovGa with thesupervised method of LDA, GFFF, sHOG, GFC and sSIFT. GFFF(sHOG, GFC, sSIFT) is the supervised version of GaFour (HOG,GFC and SIFT) by using LDA. As one of the baseline methods in facerecognition, PCA [35] is also the baseline method in appearance-based pose estimation. We compare our descriptors with Gabor,HOG and SIFT since these descriptors are the general texturedescriptor and have been applied in many areas. For all themethods, PCA is used after feature extraction to reduce thedimension of features and 97% of total energy of eigenvalues iskept. For the supervised methods, we apply PCA first for dimen-sionality reduction and then LDA for discriminant analysis. In theexperiment, the region of CovGa is set to 4�4 with overlapping2�2. To improve the performance of CovGa, we also compute thesymmetry of regions in the same column.

The accuracies of yaw estimation and pitch estimation with thecenter number ranging from one to ten are shown in Figs. 7 and 8,respectively. The x-axis represents the center number of eachposes and the y-axis represents MAE.

As the figure shows, the MAEs of kCovGa are the lowest underall the center number k for yaw estimation and pitch estimation,which show the good performance of the proposed descriptor. Wealso can find that the performance of GFC is much better than the

other two texture descriptors sHOG and sSIFT, which shows theadvantage of Gabor features in head pose estimation. These resultsvalidate that extracting the symmetry of regions on the Gaborfeatures but not the intensities is reasonable.

5.2. MultiPIE database

The second database is the CMU Multi-PIE face database [41].There are four sessions in Multi-PIE database. In our experiment,we only use the images in the first session. This session contains3735 images from different subjects. The head angles varies from�901 to 901 with an interval of 151.

For all the images, a face detection method [42] is first appliedto locate the face region from the input images, and all the faceregions are then normalized to the same size of 32�32. Finally,histogram equalization is used to reduce the influence of lightingvariations. In Fig. 9, we show some face images in the MultiPIEdatabase. From the figure, we can see that the faces in the MultiPIEdatabase are misaligned. Therefore, from the results of thisexperiment, we can investigate the robustness of different meth-ods to misalignment.

In this experiment, 3-fold cross-validation is used to avoidover-training. Specifically, we rank all the images by subjects anddivide them into three subsets. Two subsets are taken as thetraining set and the other is taken as the testing set. In this way,the persons for training and testing are totally different, thusavoiding the over-fitting in identity. Testing is repeated threetimes, by taking each subset as the testing set. The reportedresults are the average of all the tests.

Different from MAE on the Pointing'04 database, we show theaccuracies of different methods on the MultiPIE database in Fig. 10.From the above figure, we can see that on the MultiPIE database,the results of CovGa are obviously better than those of otherunsupervised methods. After using the metric learning method,the results of kCovGa are the best of all methods. For theunsupervised methods, it can also be seen that the accuracyincreases with the increase of the number of centers k when k issmall. However, for the supervised methods, such as kCovGa, theaccuracies are nearly equal for different k, which actually impliesthe excellent compactness of each class in the feature spaceobtained by metric learning. Especially, for different numbers ofcenters, the accuracies of kCovGa are constantly 98.5% while theresults of GFFF are 96.8%. The robustness of kCovGa to the numberof centers shows the good discriminative-ability of kCovGa. In thereal system, we can just use 5 or 6 centers for each pose, which candecrease the computing cost of matching the gallery samples andthe probe sample.

1 2 3 4 5 6 7 8 9 107

8

9

10

11

12

13

14

15

The number of center

The

mea

n ab

solu

te e

rror

PCALDAGaborGFCGaFourGFFFHOGsHOGSIFTsSIFTCovGakCovGa

Fig. 8. The MAEs of head pitch estimation on the Pointing'04 database. The x-axisrepresents the center number of each class and the y-axis represents the MAE.

Fig. 9. The face images in the MultiPIE database.

B. Ma et al. / Neurocomputing 143 (2014) 97–108104

Page 9: CovGa A novel descriptor based on symmetry of regions for ...vipl.ict.ac.cn/uploadfile/upload/2015116181238205.pdf · the face recognition research. However, robust and accurate face

5.3. Experiments on the CAS-PEAL database

We also evaluate the performances of different methods on thepublic CAS-PEAL database [31]. The CAS-PEAL database contains21 poses combining seven yaw poses [�451, 451] with intervals of151 and three pitch poses (301, 01 and �301). We use a subsetcontaining totally 4200 images of 200 subjects whose IDs rangefrom 401 through 600. Considering the cross-validation, there aretotally about 400ð ¼ 600=3� 2Þ samples for each pose in thetraining set. The experimental setting is the same as that ofMultiPIE database. Some images in the CAS-PEAL database areshown in Fig. 11.

In Fig. 12, we show the accuracies of different methods on theCAS-PEAL database. From the figure, we can see that for differentcenter numbers, the accuracies of kCovGa are about 94.2% whilethe accuracies of GFFF are about 90.7%. This proves again that thesymmetry of 2D regions is much better than the symmetry of 1Dsignals. It must be pointed out that the accuracies of Gaborfeatures are better than those of CovGa. But after combining withthe metric learning method, the accuracies of kCovGa are justa little higher than those of GFC, which means that the CovGafeatures contain more information of pose and are suitable to becombined with KISSME.

5.4. Experiments on the Multi-Pose database

The fourth experiment is performed on the private Multi-Posesdatabase, which consists of 3030 images of 102 subjects taken undernormal indoor lighting conditions and fixed background. The yawposes and the pitch poses range within [�501, þ501] with intervalsof 11. The sample number is 30 for each class (i.e. yaw pose). Someimages of the results of face detection are shown in Fig. 13.Considering that the sample number is about 40ð ¼ 60=3� 2Þ foreach class (pose) in the training set, the maximal centroid numberfor each pose is limited to 7, which is different from that in theexperiment on the other databases. The MAEs when the centernumber varies from 1 to 7 are shown in Fig. 14.

From the above figure, we can see that on the MultiPosesdatabase the advantage of the proposed method is more obvious.The results of CovGa are the best of all the unsupervised methodsand the results of kCovGa are the best of all the methods.Especially, for different numbers of centers, the MAEs of kCovGaare close to 3.81. The MAEs of CovGa are close to 4.51 which areeven better than those of the supervised methods. The goodperformance of the proposed method again shows that theproposed method can improve the performance of head poseestimation by using the symmetry of regions.

1 2 3 4 5 6 7 8 9 10

82

84

86

88

90

92

94

96

98

100

The number of center

The

accu

racy

PCALDAGaborGFCGaFourGFFFHOGsHOGSIFTsSIFTCovGakCovGa

Fig. 10. The accuracies on the MultiPIE database. The x-axis represents the centernumber of each class and the y-axis represents the accuracy.

Fig. 11. The face images of one subject in the CAS-PEAL database.

1 2 3 4 5 6 7 8 9 1076

78

80

82

84

86

88

90

92

94

The number of center

The

accu

racy

PCALDAGaborGFCGaFourGFFFHOGsHOGSIFTsSIFTCovGakCovGa

Fig. 12. The accuracies on the CAS-PEAL database. The x-axis represents the centernumber of each class and the y-axis represents the accuracy.

B. Ma et al. / Neurocomputing 143 (2014) 97–108 105

Page 10: CovGa A novel descriptor based on symmetry of regions for ...vipl.ict.ac.cn/uploadfile/upload/2015116181238205.pdf · the face recognition research. However, robust and accurate face

5.5. Experimental results with different region sizes on the CAS-PEALdatabase

In CovGa, one important parameter is the size of the region.To show the impact of the region size, we also repeat theexperiments with the different region sizes. In Fig. 15, we showthe accuracies of kCovGa on the CAS-PEAL database while theregion sizes are 4�4, 8�8, 12�12 and 16�16. The overlapping isset to a half of the region size. To show the effectiveness ofsymmetry, we only use the symmetry feature dij;r and discard theGabor feature g ij;r in Eq. (13).

From Fig. 15, it is easy to see that on the one hand, the largerthe region sizes, the higher the accuracies of head pose estimation.

On the other hand, the accuracies under larger region sizes aremuch more robust to the center numbers of the NC classifier. Bothscenes show that the discriminative ability of CovGa is improvedwhen the region size decreases. We attribute the improvement ofthe discriminate ability to that using smaller region sizes keepingthe detail of the image better. In Table 2, we also show thedimension of CovGa (before using PCA) under different regionsizes. From Table 2, we can know that the dimension of CovGa isalso increased when we use the smaller region sizes, which meansthat it increases the computation complexity and storage require-ment. Fortunately, compared with the accuracies of GFFF in Fig. 12,we can know that the accuracies of CovGa when the region size is16 are close to the accuracies of GFFF and its dimension is only 576.

Fig. 13. The face images of one subject in the Multi-Pose database.

1 2 3 4 5 6 7

4

4.5

5

5.5

6

The number of center

The

mea

n ab

solu

te e

rror

PCALDAGaborGFCGaFourGFFFHOGsHOGSIFTsSIFTCovGakCovGa

Fig. 14. The MAEs on the Multi-Poses database. The x-axis represents the centernumber of each class and the y-axis represents the MAE.

2 4 6 8 10 12 14 16 18 2090.5

91

91.5

92

92.5

93

93.5

94

94.5

The number of center

The

accu

racy

4x48x812x1216x16

Fig. 15. The accuracies of kCovGa under different region sizes on the CAS-PEALdatabase. The x-axis represents the center number of each class and the y-axisrepresents the accuracy.

B. Ma et al. / Neurocomputing 143 (2014) 97–108106

Page 11: CovGa A novel descriptor based on symmetry of regions for ...vipl.ict.ac.cn/uploadfile/upload/2015116181238205.pdf · the face recognition research. However, robust and accurate face

This encourages us to use relatively bigger region sizes in thescenario where the speed of system is of more concern.

5.6. Experimental results with different dimensions on the CAS-PEALdatabase

Fig. 16 shows the performance of the proposed method withrespect to a varying subspace dimension. Since the number of thetraining samples is 2800¼ ð4200=3n2Þ and there is the limitationthat the dimension is smaller than the number of training samples,the max dimension is set to 2500. From the figure, we can see thatwith the increase of dimension, the accuracy is also increasedwhen the dimension is smaller than 1200. But the accuraciesconverge to about 94.2% when the dimension is larger than 1200.

In this scene, we use PCA to reduce the dimension of CovGa whilekeeping the 97% data variance.

5.7. Experimental results of different poses on the Pointing'04database

In Fig. 17, we also show the MAEs and standard deviations ofkCovGa under different poses on the Pointing'04 database. Fromthe figure, we can know that the performance of kCovGa isdecreased when the head varies from the frontal to the profile.We attribute this to the higher variations in the image when thepose is near to the profile.

6. Conclusions

Based on the relationship between the symmetry of the faceimage and the head pose, we propose in this paper a novel facerepresentation method for head yaw estimation. Compared withthe symmetry of 1D signal from one row of the image, thesymmetry of 2D regions is much more intrinsically related to thepose variations of head. Specially, to extract the symmetry ofregions which are located in the same horizontal position, we usecovariance descriptor on the Gabor representations. The results onseveral databases show the effectiveness of the proposed method.

There are also several aspects to be studied in the future. First,considering the necessity of the real-time system, the covariancedescriptor should be replaced by its faster variants. Second, thedimension of CovGa is still very high. Therefore, more effortsshould be made to further reduce the dimension. For example, it isworth studying how to set the optimal parameters of Gabor filtersso that only a few Gabor filters can be selected which will reducethe dimension of CovGa greatly. Finally, different regions are notequivalently important to the performance. We should probablyemphasize some regions more in the symmetry encoding process.

Acknowledgments

This paper is partially supported by National Natural ScienceFoundation of China under Contract nos. 61003103, 61173065 and61332016 and the President Fund of UCAS.

References

[1] E. Murphy-Chutorian, M.M. Trivedi, Head pose estimation in computer vision:a survey, IEEE Trans. Pattern Anal. Mach. Intell. 31 (4) (2009) 607–626.

[2] B. Ma, S. Shan, X. Chen, W. Gao, Head yaw estimation from asymmetry of facialappearance, IEEE Trans. Syst. Man Cybern. Part B: Cybern. 38 (6) (2008)1501–1512.

[3] M. Köstinger, M. Hirzer, P. Wohlhart, P.M. Roth, H. Bischof, Large scale metriclearning from equivalence constraints, in: Proceedings of the IEEE Conferenceon Computer Vision and Pattern Recognition, 2012.

[4] A. Nikolaidis, I. Pitas, Facial feature extraction and determination of pose, in:Proceedings of the NOBLESSE Workshop on Nonlinear Model Based ImageAnalysis, 1998, pp. 257–262.

[5] F. Fleuret, D. Geman, Fast face detection with precise pose estimation, in:Proceedings of the IEEE International Conference on Pattern Recognition, vol. 1,2002, pp. 235–238.

[6] R. Stiefelhagen, J. Yang, A. Waibel, A model-based gaze tracking system, in:IEEE International Joint Symposia on intelligence and Systems, 1996, pp. 304–310.

[7] J. Xiao, T. Moriyama, T. Kanade, J.F. Cohn, Robust full-motion recovery of headby dynamic templates and re-registration techniques, Int. J. Imag. Syst.Technol. 13 (2003) 85–94.

[8] Q. Ji, R. Hu, 3D face pose estimation and tracking from a monocular camera,Image Vis. Comput. 20 (7) (2002) 499–511.

[9] L. Chen, L. Zhang, Y. Hu, M. Li, H. Zhang, Head pose estimation using fishermanifold learning, in: Proceeding of the IEEE International Workshop onAnalysis and Modeling of Faces and Gestures, 2003, pp. 203–207.

Table 2The dimensions of kCovGa under different region sizes.

Region size 4�4 8�8 12�12 16�16

Dim 14 440 3136 1024 576

500 1000 1500 2000 250091.5

92

92.5

93

93.5

94

94.5

Dim

The

accu

racy

Fig. 16. Accuracy vs. dimension. The x-axis represents the dimension and the y-axisrepresents the accuracy.

−105 −75 −45 −15 15 45 75 1053

4

5

6

7

8

9

10

11

12

13

The pose

The

mea

n ab

solu

te e

rror

meanstd

Fig. 17. The MAE under different poses on the Pointing'04 database. The x-axisrepresents the head poses and the y-axis represents the MAE.

B. Ma et al. / Neurocomputing 143 (2014) 97–108 107

Page 12: CovGa A novel descriptor based on symmetry of regions for ...vipl.ict.ac.cn/uploadfile/upload/2015116181238205.pdf · the face recognition research. However, robust and accurate face

[10] Y. Wei, L. Fradet, T. Tan, Head pose estimation using Gabor eigenspacemodeling, in: Proceedings of the IEEE International Conference on ImageProcessing, 2002, pp. 281–284.

[11] Y. Li, S. Gong, H. Liddel, Support vector regression and classification basedmulti-view face detection and recognition, in: Proceedings of the IEEEInternational Conference on Automatic Face and Gesture Recognition, 2000,pp. 300–305.

[12] J.N.S. Kwong, S. Gong, Learning support vector machines for a multi-view facemodel, in: Proceedings of the British Machine Vision Conference, 1999,pp. 300–305.

[13] T. Darrell, B. Moghaddam, A.P. Pentland, Active face tracking and poseestimation in an interactive room, in: Proceedings of the IEEE Conference onComputer Vision and Pattern Recognition, 1996.

[14] Z. Stan, X. Li, X. Lu, X. Hou, X. Peng, Q. Cheng, Learning multiview facesubspaces and facial pose estimation using independent component analysis,IEEE Trans. Image Process. 14 (6) (2005) 705–712.

[15] S.T. Roweis, L.K. Saul, Nonlinear dimensionality reduction by locally linearembedding, Science 290 (2000) 2323–2326.

[16] J.B. Tenenbaum, V. de Silva, J.C. Langford, A global geometric framework fornonlinear dimensionality reduction, Science 290 (2000) 2319–2323.

[17] M. Belkin, P. Niyogi, Laplacian eigenmaps and spectral techniques for embed-ding and clustering, in: Advances in Neural Information Processing Systems,vol. 15, 2001, pp. 585–591.

[18] Y. Fu, T.S. Huang, Graph embedded analysis for head pose estimation, in:Proceeding of the IEEE International Conference on Automatic Face andGesture Recognition, 2006, pp. 3–8.

[19] J. Tu, Y. Fu, Y. Hu, T.S. Huang, Evaluation of head pose estimation for studiodata, in: First International Evaluation Workshop on Classification of Events,Activities and Relationships, 2007, pp. 281–290.

[20] N. Hu, W. Huang, S. Ranganath, Head pose estimation by non-linear embed-ding and mapping, in: Proceedings of the IEEE International Conference onImage Processing, vol. 2, 2005, pp. 342–345.

[21] B. Raytchev, I. Yoda, K. Sakaue, Head pose estimation by nonlinear manifoldlearning, in: Proceedings of the IEEE International Conference on PatternRecognition, vol. 4, pp. 462–466, 2004.

[22] V.N. Balasubramanian, J. Ye, S. Panchanathan, Biased manifold embedding: aframework for person-independent head pose estimation, in: IEEE Conferenceon Computer Vision and Pattern Recognition, June 2007, pp. 1–7.

[23] S. Yan, X. Xu, D. Xu, Q. Liu, Learning the object location, view and scale forimage categorization with adapted classifiers, Inf. Sci. (2014), http://dx.doi.org/10.1016/j.ins.2014.03.041, in press.

[24] D. Huang, M. Storer, F. Torre, H. Bischof, Head pose estimation by non-linearembedding and mapping, in: Proceedings of the IEEE Conference on ComputerVision and Pattern Recognition, 2011.

[25] R. Stiefelhagen, J. Yang, A. Waibel, Modeling focus of attention for meetingindexing based on multiple cues, IEEE Trans. Neural Netw. 13 (4) (2002)928–938.

[26] S.O. Ba, J.M. Odobez, A probabilistic framework for joint head tracking andpose estimation, in: Proceedings of the IEEE International Conference onPattern Recognition, 2004.

[27] S.O. Ba, J.M. Odobez, Evaluation of multiple cue head pose estimationalgorithms in natural environements, in: Proceedings of the IEEE InternationalConference on Multimedia and Expo, 2005, pp. 1330–1333.

[28] Z. Yang, H. Ai, B. Wu, S. Lao, L. Cai, Face pose estimation and its application invideo shot selection, in: Proceedings of the IEEE International Conference onPattern Recognition, 2004.

[29] G. Fanelli, J. Gall, L. Van Gool, Real time head pose estimation with randomregression forests, in: Proceedings of the IEEE Conference on Computer Visionand Pattern Recognition, 2011.

[30] G. Fanelli, J. Gall, L. Van Gool, Random forests for real time 3D face analysis,Int. J. Comput. Vis. (2012).

[31] W. Gao, B. Cao, S. Shan, X. Chen, D. Zhou, X. Zhang, D. Zhao, The CAS-PEALlarge-scale chinese face database and baseline evaluations, IEEE Trans. Syst.Man Cybern. Part A 38 (1) (2008) 149–161.

[32] L. Wiskott, J.M. Fellous, N. Krüger, C.V.D. Malsburg, Face recognition by elasticbunch graph matching, IEEE Trans. Pattern Anal. Mach. Intell. 19 (7) (1997)775–779.

[33] O. Tuzel, F. Porikli, P. Meer, Region covariance: a fast descriptor for detectionand classification, in: Proceedings of the European Conference on ComputerVision, 2006, pp. 589–600.

[34] O. Tuzel, F. Porikli, P. Meer, Pedestrian detection via classification onRiemannian manifolds, IEEE Trans. Pattern Anal. Mach. Intell. 30 (10) (2008)1713–1727.

[35] M.A. Turk, A.P. Pentland, Eigenfaces for recognition, J. Cogn. Neurosci. 3 (1)(1991) 71–86.

[36] N. Gourier, D. Hall, J.L. Crowley, Estimating face orientation from robustdetection of salient facial features, in: Proceeding of the IEEE InternationalConference on Pattern Recognition International Workshop on Visual Obser-vation of Deictic Gestures, 2004, pp. 183–191.

[37] R. Stiefelhagen, Estimating head pose with neural networks results on thePointing04 ICPR workshop evaluation data, in: Proceeding of the IEEEInternational Conference on Pattern Recognition Workshop, 2004.

[38] N. Gourier, J. Maisonnasse, D. Hall, J.L. Crowley, Head pose estimation on lowresolution images, First International Evaluation Workshop on Classification ofEvents, Activities and Relationships, 2006, pp. 270–280.

[39] Y. Li, S. Wang, X. Ding, Person-Independent Head Pose Estimation Based onRandom Forest Regression, September 2010, pp. 1521–1524.

[40] H. Ji, R. Liu, F. Su, Z. Su, Y. Tian, Robust head pose estimation via convexregularized sparse regression, in: Proceeding of the IEEE International Con-ference on Image Processing, 2011, pp. 3617–3620.

[41] R. Gross, I. Matthews, J. Cohn, T. Kanade, S. Baker, Guide to the CMU Multi-PIEdatabase, Technical Report, Carnegie Mellon University, 2007.

[42] S. Yan, S. Shan, X. Chen, W. Gao, J. Chen, Matrix-structural learning (MSL) ofcascaded classifier from enormous training set, in: Proceeding of the IEEEConference on Computer Vision and Pattern Recognition, 2007.

Bingpeng Ma received the B.S. degree in mechanics, in1998, and the M.S. degree in mathematics, in 2003,from Huazhong University of Science and Technology.He received the Ph.D. degree in computer science at theInstitute of Computing Technology, Chinese Academy ofSciences, P.R. China, in 2009. He was a post-doctorialresearcher in University of Caen, France, from 2011 to2012. He joined the School of Computer and ControlEngineering, University of Chinese Academy ofSciences, Beijing, in March 2013 and now he is anassistant professor. His research interests cover imageanalysis, pattern recognition, and computer vision.

Annan Li received the B.S. and M.S. degrees in compu-ter science from the Harbin Institute of Technology,Harbin, China, in 2003 and 2006, respectively and thePh.D. degree in computer science from Institute ofComputing Technology, Chinese Academy of Sciences,Beijing, China, in 2011. From 2011 to 2013, he was apostdoctoral research fellow at National University ofSingapore. He is currently a research scientist in Insti-tute for Infocomm Research, Agency for Science, Tech-nology and Research, Singapore. His research interestsinclude computer vision, pattern recognition, and sta-tistical learning.

Xiujuan Chai received the B.S., M.S., and Ph.D. degreesin computer science from the Harbin Institute ofTechnology, Harbin, China, in 2000, 2002, and 2007,respectively. She was a Post-doctorial researcher inNokia Research Center (Beijing), from 2007 to 2009.She joined the Institute of Computing Technology,Chinese Academy Sciences, Beijing, in July 2009 andnow she is an associate professor. Her research inter-ests include computer vision, pattern recognition, andmultimodal human computer interaction.

Shiguang Shan is now a full professor (since 2010)with the Institute of Computing Technology (ICT),Chinese Academy of Sciences (CAS), where he receivedthe Ph.D. degree in computer science, in 2004. He isnow the executive director of the Key Lab of IntelligentInformation Processing of CAS. He published more than150 papers in refereed journals and proceedings in theareas of Computer Vision and Pattern Recognition. Heespecially focuses on face recognition relatedresearches. He is the winner of the Best Student PosterAward Runner-up in CVPR'08, the China's State Scien-tific and Technological Progress Awards in 2005 for hiswork on face recognition technologies. He has served as

Area Chair for a number of international conferences including ICCV'11, ICPR'12,ACCV'12, FG'13, ICPR'14, and ICASSP'14. He is Associate Editor of the IEEETransactions on Image Processing, the Neurocomputing, and the EURASIP Journalof Image and Video Processing.

B. Ma et al. / Neurocomputing 143 (2014) 97–108108