incremental tensor biased discriminant analysis: a new color-based visual tracking method

13
Incremental tensor biased discriminant analysis: A new color-based visual tracking method Jing Wen a , Xinbo Gao a, , Yuan Yuan b , Dacheng Tao c , Jie Li a a School of Electronic Engineering, Xidian University, No. 2, South Taibai Road, Xi’an 710071, China b School of Engineering and Applied Science, Aston University, Birmingham B4 7ET, United Kingdom c School of Computer Engineering, Nanyang Technological University, Singapore article info Article history: Received 2 May 2009 Received in revised form 13 August 2009 Accepted 6 October 2009 Communicated by X. Li Available online 20 November 2009 Keywords: Visual tracking Biased discriminant analysis Tensor biased discriminant analysis Incremental tensor biased discriminant analysis abstract Most existing color-based tracking algorithms utilize the statistical color information of the object as the tracking clues, without maintaining the spatial structure within a single chromatic image. Recently, the researches on the multilinear algebra provide the possibility to hold the spatial structural relationship in a representation of the image ensembles. In this paper, a third-order color tensor is constructed to represent the object to be tracked. Considering the influence of the environment changing on the tracking, the biased discriminant analysis (BDA) is extended to the tensor biased discriminant analysis (TBDA) for distinguishing the object from the background. At the same time, an incremental scheme for the TBDA is developed for the tensor biased discriminant subspace online learning, which can be used to adapt to the appearance variant of both the object and background. The experimental results show that the proposed method can track objects precisely undergoing large pose, scale and lighting changes, as well as partial occlusion. & 2009 Elsevier B.V. All rights reserved. 1. Introduction Visual tracking is an important and essential component of visual perception, and has been an active research topic in computer vision community for decades. Influenced by the environment change and object motion, the appearance of the object takes on variety and variability, which is a challenge for describing the object effectively. Many works have been developed for visual tracking to extract various low-level features (e.g., color [1,3,7], shape [2,6], texture and contour [4,5]), and build object appearance model (e.g., spatial histogram [1,2,7], AAMs [10], and subspace method [11–17]). Birchfield [1] and Wang et al. [2] presented the facial model with integration of shape and color. Hayashi and Fujiyoshi [3] developed color tracking method based on meanshift in luminance change. Isard and Blake [4] proposed a conditional density propagation of a parametric spine curve. Cootes et al. [6] employed the combination of shape with appearance representa- tions for tracking. Perez et al. [7] presented multi-part color modeling to capture a rough spatial layout ignored by global histograms, without taking the background color into account. Comaniciu et al. [8] proposed a new object tracking based on kernel method and meanshift algorithm. Avidan [9] developed an ensemble tracking by using AdaBoost to distinguish the object from the background. These methods above usually utilize the low-level features of the image, but ignore the high-level semantic knowledge. Moreover, these methods often assume that the object takes on consistency and similarity in respect of the texture, gradient and so on, and obtain the segmentation-like tracking results. However, the corresponding locations/pixels within these segmentation-like tracked image patches usually have not coherent sense in the context. For the purpose of getting more accurate results, a series of appearance-based methods were developed recently. Gross et al. [10] employed AAMs to track face efficiently in videos containing occlusion, but a complicated training process is inevitable before the tracking. The subspace-based methods were developed recently and applied to many research areas widely [11–35]. Based on the success of EigenTracking and incremental extension of the subspace learning, Ross et al. [11] presented an adaptive probabilistic visual tracking by updating subspace incrementally. Lin et al. [12] proposed a discriminative generative model for visual tracking. On the basis of Lin’s work, Shen et al. [13] developed a kernelized version for tracking. Yuan et al. [32] employed the incremental principal components analysis to scene segmentation for visual surveillance. Due to the limitation of this ARTICLE IN PRESS Contents lists available at ScienceDirect journal homepage: www.elsevier.com/locate/neucom Neurocomputing 0925-2312/$ - see front matter & 2009 Elsevier B.V. All rights reserved. doi:10.1016/j.neucom.2009.10.013 Corresponding author. Tel.: + 86 2988201838; fax: + 86 2988201620. E-mail addresses: [email protected] (J. Wen), [email protected] (X. Gao), [email protected] (Y. Yuan), [email protected] (D. Tao). [email protected] (J. Li). Neurocomputing 73 (2010) 827–839

Upload: jing-wen

Post on 10-Sep-2016

215 views

Category:

Documents


1 download

TRANSCRIPT

Page 1: Incremental tensor biased discriminant analysis: A new color-based visual tracking method

ARTICLE IN PRESS

Neurocomputing 73 (2010) 827–839

Contents lists available at ScienceDirect

Neurocomputing

0925-23

doi:10.1

� Corr

E-m

(X. Gao

leejie@m

journal homepage: www.elsevier.com/locate/neucom

Incremental tensor biased discriminant analysis: A new color-based visualtracking method

Jing Wen a, Xinbo Gao a,�, Yuan Yuan b, Dacheng Tao c, Jie Li a

a School of Electronic Engineering, Xidian University, No. 2, South Taibai Road, Xi’an 710071, Chinab School of Engineering and Applied Science, Aston University, Birmingham B4 7ET, United Kingdomc School of Computer Engineering, Nanyang Technological University, Singapore

a r t i c l e i n f o

Article history:

Received 2 May 2009

Received in revised form

13 August 2009

Accepted 6 October 2009

Communicated by X. Liconstructed to represent the object to be tracked. Considering the influence of the environment

Available online 20 November 2009

Keywords:

Visual tracking

Biased discriminant analysis

Tensor biased discriminant analysis

Incremental tensor biased discriminant

analysis

12/$ - see front matter & 2009 Elsevier B.V. A

016/j.neucom.2009.10.013

esponding author. Tel.: +86 2988201838; fax

ail addresses: [email protected] (J. Wen

), [email protected] (Y. Yuan), [email protected]

ail.xidian.edu.cn (J. Li).

a b s t r a c t

Most existing color-based tracking algorithms utilize the statistical color information of the object as

the tracking clues, without maintaining the spatial structure within a single chromatic image. Recently,

the researches on the multilinear algebra provide the possibility to hold the spatial structural

relationship in a representation of the image ensembles. In this paper, a third-order color tensor is

changing on the tracking, the biased discriminant analysis (BDA) is extended to the tensor biased

discriminant analysis (TBDA) for distinguishing the object from the background. At the same time, an

incremental scheme for the TBDA is developed for the tensor biased discriminant subspace online

learning, which can be used to adapt to the appearance variant of both the object and background. The

experimental results show that the proposed method can track objects precisely undergoing large pose,

scale and lighting changes, as well as partial occlusion.

& 2009 Elsevier B.V. All rights reserved.

1. Introduction

Visual tracking is an important and essential component ofvisual perception, and has been an active research topic incomputer vision community for decades. Influenced by theenvironment change and object motion, the appearance of theobject takes on variety and variability, which is a challenge fordescribing the object effectively.

Many works have been developed for visual tracking toextract various low-level features (e.g., color [1,3,7], shape [2,6],texture and contour [4,5]), and build object appearance model(e.g., spatial histogram [1,2,7], AAMs [10], and subspace method[11–17]). Birchfield [1] and Wang et al. [2] presented the facialmodel with integration of shape and color. Hayashi and Fujiyoshi[3] developed color tracking method based on meanshift inluminance change. Isard and Blake [4] proposed a conditionaldensity propagation of a parametric spine curve. Cootes et al. [6]employed the combination of shape with appearance representa-tions for tracking. Perez et al. [7] presented multi-part colormodeling to capture a rough spatial layout ignored by global

ll rights reserved.

: +86 2988201620.

), [email protected]

du.sg (D. Tao).

histograms, without taking the background color into account.Comaniciu et al. [8] proposed a new object tracking based onkernel method and meanshift algorithm. Avidan [9] developed anensemble tracking by using AdaBoost to distinguish the objectfrom the background. These methods above usually utilize thelow-level features of the image, but ignore the high-levelsemantic knowledge. Moreover, these methods often assume thatthe object takes on consistency and similarity in respect of thetexture, gradient and so on, and obtain the segmentation-liketracking results. However, the corresponding locations/pixelswithin these segmentation-like tracked image patches usuallyhave not coherent sense in the context.

For the purpose of getting more accurate results, a series ofappearance-based methods were developed recently. Gross et al.[10] employed AAMs to track face efficiently in videos containingocclusion, but a complicated training process is inevitable beforethe tracking. The subspace-based methods were developedrecently and applied to many research areas widely [11–35].Based on the success of EigenTracking and incremental extensionof the subspace learning, Ross et al. [11] presented an adaptiveprobabilistic visual tracking by updating subspace incrementally.Lin et al. [12] proposed a discriminative generative model forvisual tracking. On the basis of Lin’s work, Shen et al. [13]developed a kernelized version for tracking. Yuan et al. [32]employed the incremental principal components analysis to scenesegmentation for visual surveillance. Due to the limitation of this

Page 2: Incremental tensor biased discriminant analysis: A new color-based visual tracking method

ARTICLE IN PRESS

J. Wen et al. / Neurocomputing 73 (2010) 827–839828

image-as-vector way to build subspace, a new representationbased on high order tensor has drawn many researchers’ interest,and been introduced into the tracking and recognition. Tao et al.[18,33] introduced the multilinear representation for gait recog-nition. Moreover, Sun et al. [24] proposed incremental tensoranalysis to deal with the learning of dynamic/online data. Later,series of discriminant methods [21,27,28,31] for tensor analysisand their applications in retrieval [21,31], video semantic [30],gait recognition [28,33] were developed. Tao et al. [23,25]proposed a Bayesian tensor analysis method and applied it to3-D face modeling, as well as the kernelization [22] andprobabilistic [26] version. Li and Lee [14] presented a motionsaliency-based visual tracking. Shao et al. [15] developed anappearance-based method using the three-dimensional trilineartensor. Li et al. [16] employed a three-dimensional temporaltensor subspace learning for visual tracking. It should be noticedthat those appearance-based tracking algorithms above have thefollowing characteristics:

They only make use of the intensity information [5,11–13,15–17], which will lose chromatic knowledge in color videosequences. �

Sw

Sb

Sx

Sy

They usually regard the visual tracking as a two-class fisherdiscriminant problem [12,13], whereas the classificationbetween the object and the background should be a (1+x)-class formulation.

In visual tracking, the object belongs to the positive sample set,while the background belongs to the negative sample set relativeto the object of interest. Though the appearances of both theobject and background vary with time, the variant between theobject and background is so different. Since there are only pose,scale and illumination changes for the object, the changingappearances of the object are similar in some degree during thetracking, which can be regarded to a class as the blue symbolsshown in Fig. 1. However, with the object moving the backgroundchanges drastically, which is shown in Fig. 1 with the orangesymbols. Therefore, it is unreasonable to assign the background toone class. With this consideration, biased discriminant analysis(BDA) was developed by Zhou and Huang [19,20], which alsoleads to the small sample size (SSS) problem. Tao et al. [21]proposed a direct kernel biased discriminate analysis to deal withthe SSS problem, and to provide a relevance feedback scheme forcontent-based image retrieval.

However, these discriminant analysis methods still hold themanner of the image-as-vector representation, and lose thespatial structure of the two-dimensional image. The SSS problemcould be coped with by introducing the tensor representation[22]. In this paper, we propose a tensor biased discriminantanalysis (TBDA) which could solve the SSS problem existed in theBDA, develop an online learning scheme for the TBDA named asincremental tensor biased discriminant analysis (ITBDA), andpresent a third-order color tensor-based visual tracking byemploying the ITBDA to distinguish the objects from background.The contributions of this paper can be summarized as follows: (1)provide a new appearance construction method by integrating thespatial color information into a third-order tensor representation,which is more distinguishable for the appearance model; (2)

t1 t2 t3 ti ti+K

Fig. 1. Data stream for classification.

propose a tensor biased discriminant analysis (TBDA), which isthe generalization of the BDA for tensor representation, and ableto deal with the SSS caused by the vectorization of the BDA; (3)present an incremental tensor biased discriminant analysis(ITBDA) suitable for online distinguishing the objects from theobject-like background.

The reminder of this paper is organized as follows: the relatedprevious work, say LDA and BDA, is described in Section 2. TheTBDA and ITBDA are then proposed in Section 3. In Section 4, acolor-based visual tracking algorithm is presented. Section 5conducts several experiments to validate the effectiveness of theproposed tracking method undergoing large pose, scale andlighting changes for the single-object tracking, as well as thepartial occlusion for the multiple-object tracking. The final sectiondraws conclusions and future works.

2. Previous works

In this section, previous works are introduced including thelinear discriminant analysis (LDA) and biased discriminantanalysis (BDA).

2.1. Linear discriminant analysis (LDA)

Linear discriminant analysis is a supervised learning method,which tries to seek directions that are effective for discrimination.The objective function of LDA is to obtain a set of vectors W,which maximizing the ratio between Sw and Sb, the within-class

scatter matrix and the between-class scatter matrix,

W¼ argmaxW

JWT SbWJ

JWT SwWJð1Þ

Suppose we have a set of N samples x1, x2, y, xN, belonging tok classes and each class has Ni samples. Then Sb and Sw are definedas

Sb ¼1

N

Xk

i ¼ 1

Niðmi�mÞðmi�mÞT

Sw ¼1

N

Xk

i ¼ 1

XNi

j ¼ 1

ðxij�miÞðx

ij�miÞ

T ; xijAki

8>>>>>><>>>>>>:

ð2Þ

where N¼Pk

i ¼ 1 Ni. m¼ ð1=NÞPN

j ¼ 1 xj is the mean vector of the

total training set, and mi ¼ ð1=NiÞPNi

j ¼ 1 xij is the mean vector for

the individual class ki. xij is the jth samples belonging to class ki.

Therefore, W can be computed from the eigenvectors of S�1w Sb.

Given two classes, i.e., k=2, LDA is reduced to Fisher discriminantanalysis (FDA) as shown in Fig. 2; otherwise, multiplediscriminant analysis (MDA).

Fig. 2. FDA versus BDA, the left is FDA and the right is BDA.

Page 3: Incremental tensor biased discriminant analysis: A new color-based visual tracking method

ARTICLE IN PRESS

J. Wen et al. / Neurocomputing 73 (2010) 827–839 829

2.2. Biased discriminant analysis (BDA)

Zhou and Huang [19,20] proposed the biased discriminantanalysis, which defines the (1+x)-class classification problem. Thismethod is applicable for the cases that there are an unknownnumber of classes (the symbols in orange color in Fig. 2) but theuser only concern with one class (the symbols in blue color inFig. 2), i.e., the user is biased toward one class (in blue).

The biased discriminant analysis tries to seek the subspace todiscriminate the positive samples (the only class of concern to theuser) from the negative samples (unknown number of classes).The objective aims to find a set of vector W maximizing the ratiobetween Sx and Sy, the positive-class scatter matrix and thenegative-class scatter matrix,

W¼ argW

maxJWT SyWJ

JWT SxWJð3Þ

where the scatter matrix can be obtained by

Sx ¼XNx

i ¼ 1

ðxi�mxÞðxi�mxÞT

Sy ¼XNy

i ¼ 1

ðyi�mxÞðyi�mxÞT

8>>>>><>>>>>:

ð4Þ

where xi and yi denote the positive and negative samples,respec-tively.Nx and Ny are the numbers of the positive and negativesamples, and mx ¼ ð1=NxÞ

PNx

i ¼ 1 xi is the mean vector of thepositive samples. W can be computed from the eigenvectors ofS�1

x Sy. BDA minimizes the variance of the positive samples andmaximizes the scatter distance of all negative samples from thecenter of the positive samples. However, BDA still has the SSSproblem, since the valid dimensionality is determined by theminimum of the positive and negative sample, while for the FDA,it is only one [20].

However, it is pointed out that when the number of thetraining measurements is limited, the vectorization operationalways leads to the SSS problem [22]. For this reason, thesevectorized discriminant methods often use the regularizationmethod, which has been reported to be not a good choice bymany papers on face recognition [34,35]. Therefore, wepropose a new method based on tensor representation whichcould cope with the problem in some extent in the nextsection.

3. Incremental tensor biased discriminant analysis

In this section, a new supervised subspace method—tensorbiased discriminant analysis (TBDA), the generalization of thebiased discriminant analysis (BDA) [19] by introducing tensorrepresentation, is developed to distinguish the positive andnegative classes and mainly focus on the class of interest. Dueto tensor representation makes full use of the structureinformation of the object, which is a reasonable constraint[22] to reduce the number of the unknown parameters used torepresent a learning model, the proposed TBDA could cope withthe SSS problem [22] in some extent by introducing tensorrepresentation. In addition, an incremental tensor biaseddiscriminant analysis (ITBDA) algorithm for distinguishing thepositive from negative class adaptively when the data areavailable incrementally/online. Moreover, the definition of thetensor algebra operation involved in this section could bereferred to [18,22].

3.1. Tensor biased discriminant analysis

Before presenting the TBDA, we introduce the tensor LDA inadvance. A tensor can be regarded as a multidimensional matrix.An M-order tensor X is an M-dimensional matrix. Here, we denotea tensor sample as XARL1�����LM . The traditional LDA aims to find aprojection which is optimal for separating different classes in alow-dimensional space. Similarly, tensor LDA is to obtain theoptimal discriminative projections for the tensor data in eachorder,

Wd ¼ argWd

maxJWT SðdÞb WJ

JWT SðdÞw WJ; d¼ 1; . . . ;M ð5Þ

where

SðdÞb ¼1

N

Xk

i ¼ 1

NiðXðdÞ

i �XðdÞÞðXðdÞ

i �XðdÞÞT

SðdÞw ¼1

N

Xk

i ¼ 1

XNi

j ¼ 1

ðXðdÞi;j �XðdÞ

i ÞðXðdÞi;j �X

ðdÞ

i ÞT ; Xi;jAki

8>>>>>><>>>>>>:

ð6Þ

in which the symbol �ðdÞ means the mode-d matricizing unfolding[18]. X

ðdÞand X

ðdÞ

i are the mean of the total samples, the mean forthe individual class ki, respectively. XðdÞi;j is the jth samplebelonging to class ki. Wd, which is the mode-d discriminantprojection, can be computed from the eigenvectors of ðSðdÞw Þ

�1SðdÞb .Similarly, for the case there are unknown numbers of classes

while we only care for one class, tensor biased discriminantanalysis (TBDA) is proposed for necessity. Accordingly, thepositive and negative variance matrices can be defined as follows:

SðdÞX ¼XNx

i ¼ 1

ðXðdÞi �XðdÞÞðXðdÞi �X

ðdÞÞT

SðdÞY ¼XNY

i ¼ 1

ðYðdÞi �XðdÞÞðYðdÞi �X

ðdÞÞT

8>>>>><>>>>>:

ð7Þ

where NX and NY are the numbers of the positive and the negativesamples, respectively. We can compute the biased discriminantprojection in each mode by the following objective function:

Wd ¼ argmaxWd

JWT SðdÞY WJ

JWT SðdÞX WJ; d¼ 1; . . . ;M ð8Þ

The eigenvectors are computed from ðSðdÞX Þ�1SðdÞY .

3.2. Incremental tensor biased discriminant analysis

We extend TBDA to an incremental learning fashion, named asincremental tensor biased discriminant analysis (ITBDA), so as toadapt to the situation that the data cannot be obtained at onetime, especially in the context of visual tracking, the appearancemodel of the object should be updated to adapt to the change ofboth the object and the background with time drift. For visualtracking, the object measurement is the positive class of which isconcerned only, while the background belongs to the negativeclasses with unknown numbers of classes as illustrated in Fig. 1.Due to the appearances of both the object and background varywith time, the two variance matrices SðdÞX and SðdÞY should beupdated as the arrival of new data (video frames).

Notice the difference of the computation of SðdÞY from SðdÞX , SðdÞY isthe deviation of the negative samples from the mean of thepositive class, while SðdÞX is the variance matrix of the positivesamples. We describe the online updating on SðdÞX at first, which issimilar to [17].

Suppose there are a set of old tensor data fXi; i¼ 1; . . . ;nXg, andthe new arrival tensor data fXj; j¼ 1; . . . ;mXg, where nX and mX are

Page 4: Incremental tensor biased discriminant analysis: A new color-based visual tracking method

ARTICLE IN PRESS

Table 1The pseudo code of the ITBDA algorithm.

Input:New data from positive and negative observations:

XjjmX

j ¼ 1 ARL1�����LM , YjjmY

j ¼ 1 ARL1�����LM

Old means of the positive data: XðdÞ

oldjMd ¼ 1

Old projection and energy matrix of the positive data: UðdÞX jMd ¼ 1DðdÞX j

Md ¼ 1

Old variance matrices of the negative data: SðdÞYoldjMd ¼ 1

The numbers of the old positive and negative tensor data: nX, nY

The numbers of the new arrival positive and negative tensor data: mX , mY

For d=1 to M

S1: Update the positive mean XðdÞ¼ðanX X

ðdÞ

old þmX XðdÞ

new Þ

anX þmX, with X

ðdÞ

new computed by

Eq. (12)

S2: Update the positive variance

SðdÞX ¼ a2UðdÞX DðdÞX UðdÞ

T

X þSðdÞXnewþ

anX mXanX þmX

ðXðdÞ

old�XðdÞ

newÞðXðdÞ

old�XðdÞ

newÞT

with SðdÞXnewcomputed by Eq. (12)

S3: Update the negative variance

SðdÞY ¼ a2SðdÞYold

þSðdÞYnew

with SðdÞYnewcomputed by Eq. (16)

S4: Update the numbers of the positive and negative tensor data

nX ¼ anXþmX ;nY ¼ anYþmY

S5: Compute the discriminant projection

W ðdÞ ¼ SVDððSðdÞ�1

X ÞSðdÞY Þ

S6: Compute the new projection and energy matrix for positive variance:

UðdÞX DðdÞX UðdÞT

X ’SVD

SðdÞX

End For

Output:New updated variance matrices of the positive and negative data:

SðdÞX jMd ¼ 1, SðdÞY j

Md ¼ 1

New means of the positive data: XðdÞjMd ¼ 1

New projection and energy matrix of the positive data: UðdÞX jMd ¼ 1DðdÞX j

Md ¼ 1

The updated numbers of the old positive and negative tensor data: nX , nY

The discriminative project WðdÞjMd ¼ 1

J. Wen et al. / Neurocomputing 73 (2010) 827–839830

the old and new arrival data numbers. The total variance matrixfor the mode-d tensor can be defined as follows:

SðdÞX ¼XnX þmX

i ¼ 1

ðXðdÞi �XðdÞÞðXðdÞi �X

ðdÞÞT

ð9Þ

where

XðdÞ¼ðnXX

ðdÞ

oldþmXXðdÞ

newÞ

nXþmXð10Þ

and

XðdÞ

old ¼1

nX

XnX

i ¼ 1XðdÞi

XðdÞ

new ¼1

mX

XmX

j ¼ 1XðdÞj

8>>><>>>:

ð11Þ

where XðdÞ

is the total mean of the old and new tensor data, and

XðdÞ

old and XðdÞ

new are the old and new mean,respectively. By defining

the old and new variance matrices,

SðdÞXold¼XnX

i ¼ 1

ðXðdÞi �XðdÞÞðXðdÞi �X

ðdÞÞT

SðdÞXnew¼XmX

j ¼ 1

ðXðdÞj �XðdÞÞðXðdÞj �X

ðdÞÞT

8>>>>><>>>>>:

ð12Þ

we have Eq. (9) as

SðdÞX ¼XnX

i ¼ 1

ðXðdÞi �XðdÞÞðXðdÞi �X

ðdÞÞTþXmX

j ¼ 1

ðXðdÞj �XðdÞÞðXðdÞj �X

ðdÞÞT

¼XnX

i ¼ 1

ðXðdÞi �XðdÞ

oldþXðdÞ

old�XðdÞÞðXðdÞi �X

ðdÞ

oldþXðdÞ

old�XðdÞÞT

þXmX

j ¼ 1

ðXðdÞj �XðdÞ

newþXðdÞ

new�XðdÞÞðXðdÞj �X

ðdÞ

newþXðdÞ

new�XðdÞÞT

¼ SðdÞXoldþSðdÞXnew

þnXðXðdÞ

old�XðdÞÞðXðdÞ

old�XðdÞÞTþmXðX

ðdÞ

new�XðdÞÞðXðdÞ

new�XðdÞÞT

ð13Þ

subscribe Eq. (10) into Eq. (13), and we get

SðdÞX ¼ SðdÞXoldþSðdÞXnew

þnXmX

nXþmXðXðdÞ

old�XðdÞ

newÞðXðdÞ

old�XðdÞ

newÞT

ð14Þ

By analyzing Eq. (14), we can come out to the conclusion thatwhen the new data are received, the updated variance matrix is

not the simple sum of the variance matrix of the old data SðdÞXoldand

the new arrival data SðdÞXnew, there also exist a new term

ðnXmX=ðnXþmXÞÞðXðdÞ

old�XðdÞ

newÞðXðdÞ

old�XðdÞ

newÞT . It should be mentioned

that SðdÞXoldis stored as UðdÞX DðdÞX UðdÞ

T

X ’SVD

SðdÞXoldfor the sake of reducing

the memory requirement and facilitating the computation of the

likelihood in Section 5, where UðdÞX is an unitary matrix composed

of basis vectors and DðdÞX is a diagonal matrix with nonnegative real

numbers on the diagonal corresponding to the singular-value

decomposition (SVD) of SðdÞXold.

Notice that the computation of the matrix SðdÞY is different from

that of SðdÞX . We only need sum up the two matrix SðdÞYoldand SðdÞYnew

.

The reason for this is that the only interested class for the trackingtask is the object, while the background observation also changesat the same time. We do not need to update the negative(background observation) mean, and should keep the biasedcenter at the updated positive (object observation) mean only,with the following equation:

SðdÞY ¼ SðdÞYoldþSðdÞYnew

ð15Þ

where

SðdÞYold¼XnY

i ¼ 1

ðYðdÞi �XðdÞ

oldÞðYðdÞi �X

ðdÞ

oldÞT

SðdÞYnew¼XmY

j ¼ 1

ðYðdÞj �XðdÞ

newÞðYðdÞj �X

ðdÞ

newÞT

8>>>>><>>>>>:

ð16Þ

where nY and mY are the old and new data numbers for thebackground observation.

For the sake of the learning speed, we use the forgetting factora to control the proportion of the old data in the total data, withthe pseudo-code of ITBDA shown in Table 1.

Complexity: The space consumption for incremental tensorbiased discriminant analysis algorithm is

OYM

d ¼ 1

Ld

!þO

XMd ¼ 1

L2d

!:

For a high mode tensor data, the main factor is from the first

term OðQM

d ¼ 1 LdÞ, which comes from the storage of both the new

arrival positive and negative data, while for a low mode tensor

data, the main factor is OðPM

d ¼ 1 L2dÞ because of the storage of the

variance matrix. The computation cost is:

OXMi ¼ 1

Li

YMj ¼ 1

Lj

0@

1A

from updating the variance matrix, OðPM

i ¼ 1 L3i Þ from SVD decom-

position. Note that for a low or medium mode tensor (say, orderMr5), the latter term is the main cost.

Page 5: Incremental tensor biased discriminant analysis: A new color-based visual tracking method

ARTICLE IN PRESS

iti

Fig. 3. Sample an image Ii by parameterized by git , and construct a third-order

tensor in RGB space.

J. Wen et al. / Neurocomputing 73 (2010) 827–839 831

4. ITBDA-based visual tracking

Visual tracking could be formulated as a classification problembetween the object and background observations, which areencoded with the locations or motion parameters of the objectsthrough the unobservable states, and the task is to infer theunobservable states from the observed images over time.

4.1. Bayesian inference

We regard the visual tracking as a state estimation problem ina way similar to [17]. Denote zt as an image region observed attime t and Zt ¼ ðz1; . . . ; ztÞ is a set of image regions observed fromthe beginning to time t. A visual tracking is a process to infer stategt from the observation Zt. Due to the motion of the object fromone frame to the next can be modeled by a one-order Markovmodel, this inference problem is defined as a recursive equation:

pðgtjZtÞ ¼ kpðztjgtÞ

Zpðgtjgt�1Þpðgt�1jZt�1Þdgt�1 ð17Þ

where k is a constant, pðztjgtÞ and pðgtjgt�1Þ correspond to theobservation model and dynamic model, respectively. pðgt�1jZt�1Þ isthe state estimation given all the prior observation up to timet�1, and pðztjgtÞ is the likelihood of the observed image zt at gt.Then the posterior estimation pðgtjZtÞ at time t could be computedefficiently by Eq. (17). Therefore, the key steps for the visualtracking based on Bayesian inference are the computation of thedynamical model and observation model.

For visual tracking, an ideal distribution pðgtjZtÞ should beconvex at zt, i.e., gt matching the observed object’s location zt.While the integral in Eq. (17) predicts the regions where objects islikely to appear given all the prior observation, the observationmodel pðztjgtÞ could determine the most likely state that matchesthe observation at time t.

Dynamic model: The location of an object in an image can bedescribed by an affine image warp with six parameters. In thispaper, the state at time t consists of the six parameters of an affinetransformation gt ¼ fat ; bt ;yt ; st ;at ;jtg, where at ; bt ; yt ; st ;at ;jt ,denote the vertical, horizontal transition, rotation angle, scale,aspect ratio, and skew direction at time t,

pðgtjgt�1Þ ¼Nðgt; gt�1;CÞ ð18Þ

where C is a diagonal covariance matrix whose elements arethe corresponding variances of affine parameters, i.e.,s2

a ;s2b ;s

2y ;s

2s ;s2

a;s2j. Each parameter in gt is modeled indepen-

dently by a Gaussian distribution around its counterpart in gt�1.And the warp of the observations patches between the successiveframes is an affine transformation. In order to get the most likelystate, a sampling scheme is adopted to collect a set of statesfgi

t�1; i¼ 1; . . . ; kg as many as possible. Then the optimal state canbe estimated by maximizing the likelihood pðzt jgtÞ among thesestates.

Observation model: Given a set samples I¼ fI1; . . . ; Ikgwhere Ii isthe appearance tensor data (as shown in Fig. 3) collected in zt�1

based on fgit�1; i¼ 1; . . . ; kg. That is, we draw a set samples

parameterized by fgit�1; i¼ 1; . . . ; kg in zt�1 that have large

pðzt�1jgit�1Þ, but the low posterior pðgi

t�1jZt�1Þ. pðzitjgtÞ measures

the probability of zijki ¼ 1 as samples being generated by the objectclass,

pðzitjgtÞpexp

��:ðIi

t�XðdÞÞ�ðIi

t�XðdÞÞYM

d ¼ 1

�dðUdXUdT

X Þ:2

F

�; i¼ 1; . . . ; k

ð19Þ

where J � J2F is Frobenius norm, and � d is the mode product in

tensor algebra [18].

However, there exists both the object X and background Y

observation in samples I. We want to find a projectionWðdÞjMd ¼ 1,

which could discriminate the object samples from the backgroundsamples. Then we can have a set of binary value bi

tjki ¼ 1 ¼ f0;1g

with the object samples are 1 and the background samples are 0by applying the discriminant analysis to all the samples at time k.So we have the likelihood in Eq. (19) modified as

pðzitjgtÞpbi

texp��:ðIi

t�XðdÞÞ�ðIi

t�XðdÞÞYM

d ¼ 1

�dðUdXUdT

X Þ:2

F

�; i¼ 1; . . . ; k

ð20Þ

where b is the label by applying the ITBDA.Feature selection for tracking: We keep the best estimated

sample at every frame parameterized by regarding the optimalstate as the positive sample. For the negative class, thesamples with medium likelihood usually confuse the estimation,due to their appearances are close to the object class andcontaining both the observation of the object and background.Therefore, we will keep both the background samples and object-like samples, which have low values in Eq. (20) as negativesamples.

4.2. Proposed tracking algorithm

In this paper, we adopt the RGB space as the appearancevalue, since color images can be thought as real valuedthird-order tensors as shown in Fig. 3. Both the object andbackground observations are the warped images, obtained byusing affine transform with the six state parameters. The warpedimages are kept with their original chromatic structures, as third-order tensors with the first two orders as the location and thethird order as the RGB color components as shown in the right ofFig. 3. The proposed tracking algorithm is based on maximumlikelihood estimate given all the observations up to that timeinstance,

g�t ¼ argmaxgt

pðgtjZtÞ ð21Þ

The proposed tracking method can be initialized by manuallylabeling the object of interest or automatically detecting theobject to be tracked through the available detection method. Inthis paper, we initialized the tracking by manually labeling theobject in the first frame. The proposed tracking method is apropagation process for the conditional density of the objectobservation based on the affine state parameters. Firstly, accord-ing to the dynamical model in Eq. (18), the tracker draws k

samples which represent the possible locations of the object attime t�1 based on fgi

t�1; i¼ 1; . . . ; kg, and obtains k ‘‘particle’’observations by warping the color image observationsI¼ fI1; . . . ; Ikg through the affine transformation based on thesestate parameters. Then, discriminate the object and backgroundsamples by the ITBDA among all these k image observations, and

Page 6: Incremental tensor biased discriminant analysis: A new color-based visual tracking method

ARTICLE IN PRESS

ITBDA

video sequence

dynamic model

t-1 t

negativetensor data

observation model

positivetensor data

positive subspace

negative variance discriminant projection

optimalestimation

Fig. 4. The proposed tracking flowchart.

Fig. 5. The results for tracking a human face mainly undergoing large pose variant with the frame numbers of the results #7, #23, #45, #66, #78, #100, #115, #130, # 159,

#181, #216, #230, # 252, #299, #376, #381, #420, # 457, #473, and #494, in order.

J. Wen et al. / Neurocomputing 73 (2010) 827–839832

compute the likelihood in Eq. (20) for the positive samples asthe weights for propagation of the conditional density by theobservation model. Thirdly, the optimal object sample and thebackground samples are assigned to positive and negativesamples set, respectively, for updating the discriminative projec-tion by ITBDA. The detailed tracking flow is given in Fig. 4.

1 From: http://homepages.inf.ed.ac.uk/rbf/CAVIAR/.

5. Experimental results

In order to evaluate the performance of the proposed trackingalgorithm, we collected eight videos with human face andpedestrians as the tracking objects, where the first three videosequences are captured indoor undergoing large pose variant and

drastic illumination, the last five video sequences1 are recorded inshopping center in Portugal.

We test five and three video sequences for the single- andmultiple-objects tracking, respectively, to verify the trackingperformance of the proposed method. For the single objecttracking in our experimental results, the red box shows themaximum likelihood of the observation. Moreover, the mean andoriginal tracked images are shown in the bottom row of eachpanel, as well as the error image and reconstructive image basedon the positive subspace.

Page 7: Incremental tensor biased discriminant analysis: A new color-based visual tracking method

ARTICLE IN PRESS

J. Wen et al. / Neurocomputing 73 (2010) 827–839 833

5.1. Tracking a human face

In this subsection, we desire two experiments to evaluate thetracking performance mainly undergoing pose and drastic lightvariant, in Examples 1 and 2, respectively.

Example 1. In this video sequence, the object, i.e., the humanface, walks fast in the moving camera, undergoing large pose(#7, #45, #78, #130, #230, and #299), scale (#7, #78, #181, #252, #376, and #494) variant as shown in Fig. 5. Our method isable to efficiently learn a compact color representation while thecolor histogram method usually gives an inaccurate location forthe object. All the eigenbases are constructed automatically fromthe RGB-space images and constantly updated to model theappearance of the object, as well as the background, both of which

Fig. 6. A person undergoing drastic light change (from the light to the dark), slight pose

#9, #48, #90, # 130, #168, #205, #248, #287, #326, and #384, in order.

Fig. 7. A person undergoing drastic light change (from the dark to the light), and slight

#149, #170, #196, and #213, in order.

Fig. 8. A pedestrian moves far away from the camera with another one, with the other

#126, #190, #202, #269, #325, and #361, in order.

will be combined to compute the discriminant projection for thedistinguishment at next frame.

Example 2. In this experiment, we aim to test the performanceof the proposed tracking algorithm under different lightconditions. In Figs. 6 and 7, the human faces are both underthe environment of illumination change, as well as slightpose changing. In Fig. 6, the human face comes close to thecamera, with a slow change of the light condition from thebright to the dark (#9, #168, and #384), with slight pose (#9, #48,and #248), expression (#9, #48, and #326) and scale variant(#9 and #287) during the light darkening. In Fig. 7, thehuman face undergoes a drastic illumination, walking froma very dark environment to a bright place (#6, #129, and #213),with slight pose (#27, #50, #109, and #170) during thelight darkening. Our method makes full use of the color

and expression, as well as the scale variant. The frame numbers of the results are

pose variant. The frame numbers of the results are #6, #27, #50, #70, #109, #129,

pedestrian occlusion. The frame numbers of the results are #50, #89, #106, #116,

Page 8: Incremental tensor biased discriminant analysis: A new color-based visual tracking method

ARTICLE IN PRESS

Fig. 9. A pedestrian moves far away from the camera, with another pedestrian occlusion. The frame numbers of the results are #171, #185, #203, #221, #233, #281, #378,

#448, #516, and #558, in order.

Fig. 10. A comparison of our tracking method (the bottom row) with the SMOG tracker [2] (the top row), the ITSL tracker (the second row) and the ICTSL tracker (the third

row) on the video sequence of a person moving fast in the camera, with pose variant, cast shadow and scale changes.

J. Wen et al. / Neurocomputing 73 (2010) 827–839834

information, which is combined with spatial structure in thethird-order tensor data, could get good results under the variouslight conditions.

5.2. Tracking a pedestrian

In this subsection, the object is a pedestrian as shown in Figs. 8and 9. The appearances of the pedestrians do not keep constantvariance in spatial structure as the human face any more. Thoughthe changing of the pedestrian observation seems periodic, theappearance among a period is so different from each other andusually accompanies with the out-of-plane rotation, which resultsin the nonperiodicity of the appearance changing.

The object in Fig. 8 experiences twice short-time occlusions inabout #106, #116 frame and #190 frame, the object in red clothesis different in color from the two different occlusions, could betracked well after the occlusions removed.

The object in Fig. 9 has been occluded two times by the sameocclusion pedestrian in about frame #106 and #378. Though thecolors for the two pedestrians are very similar, the spatialconstraint for the color information could keep the trackerperforms well, while the methods based on histogram orstatistical model often lose the track by following the occlusionpedestrian away.

It should be pointed out that, in the bottom row of each panelin Figs. 5–9, the tracked image regions have the same structure forthe appearance, even for the pedestrian tracking. This character ofthe proposed tracking method could provide reasonable trackedblob to the recognition task.

5.3. Qualitative comparison

In this subsection, two tracking methods are run, as aqualitative benchmark, on three video sequences: In [2], a spatialcolor mixture of Gaussians (SMOG) appearance model for particlefilters were proposed, by considering both the colors in a region

Page 9: Incremental tensor biased discriminant analysis: A new color-based visual tracking method

ARTICLE IN PRESS

J. Wen et al. / Neurocomputing 73 (2010) 827–839 835

and the spatial layout of the colors. However, the spatialinformation in [2] used the color histogram combined spatialinformation partly, which is irreversible feature extraction, that is,the original image structure cannot be reconstructed by thefeatures. In order to maintain the true object appearance,incremental tensor subspace learning (ITSL) method [16] is alsoimplemented in the experiments, which utilized only grayinformation, by integrating the temporal information to a third-order tensor incrementally. Then, our proposed method withoutdiscriminant analysis is also run for comparative objective byusing the likelihood measurement in Eq. (19), named as incre-mental chromatic tensor subspace learning (ICTSL) here.

As shown in Figs. 10–12, our method provides comparableperformance to the three methods above. The SMOG tracker(listed in the top row) is much more robust to the case that theappearance variant is slow as in Fig. 12 at about frame #151 and#181. However, it is not steady in Fig. 10 (#41, #151, and #271),Fig. 11 (#181 and #213) and Fig. 12 (#211 and #241), and losesthe tracking object at the earlier frame number than the otherthree methods, due to the fast motion, view and scale changes,cast shadow and drastic illumination.

The ITSL tracker (the second row) gets better trackingperformance than the SMOG tracker when there are the smallpose changes in Fig. 11. However, it is sensitive to the appearancevariant resulting from the fast motion and occlusion as shown inFigs. 10 and 12, and loses the tracking object due to the drasticlight condition in Fig. 11.

The tracking performance of our proposed method withoutdiscriminant analysis (the third row) is much better than theabove two methods, especially in Fig. 12, because it does not onlyhold the structure information of the appearance in the tensorrepresentation, but also combine the color information with thetensor. However, without the discriminant analysis between theobject and the background, the tracking results are not satisfac-tory in Fig. 10 (#369 and #384) and Fig. 11 (#433 and #451), dueto the non-rigid motion and large light variant.

Fig. 11. A comparison of our tracking method (the bottom row) with the SMOG tracker

row) on the video sequence of a person moving close to the camera, from the bright p

The proposed tracking method based on the ITBDA (the bottomrow) could track the object effectively undergoing the variousappearance changes caused by both the instinct and extrinsicfactors such as pose (Fig. 10), light (Fig. 11) and scale variant(Figs. 10 and 12). The reason for that is our method could not onlykeep the color information in the structure representation, butalso distinguish the object from the background by using the colorstructural information. Moreover, the particles number can bereduced in some degree, when the biased discriminant analysis isused to ‘‘filter’’ the particles. Thus, it is the true object observation,but the arbitrary image patches (maybe containing the back-ground patches), whose conditional densities are propagatedframe by frame.

5.4. Quantitative comparison

To evaluate the tracking precision quantitatively, we con-ducted a face tracking experiment on in the first and second videosequences to compare our tracking error results with the twoavailable methods, i.e., SMOG and ITSL. We computed the locationerror between the estimated facial locations and the manuallylabeled ‘‘ground truth’’.

The tracking errors are plotted for each frame in Figs. 13(a)and 14(a) for the first and second videos, respectively. For mostframes our tracking results match the ground truth well in thefirst video as shown in Fig. 10, the largest errors occurring duringlarge scale and fast pose changes. In Fig. 13, the SMOG trackerperforms poorly, experiencing significant drift off the targetobject. This can be attributed to the appearance model of theSMOG tracker, based on the spatial histograms of the pixelintensities, which is not stable to the variance of the scale andpose, and does not adapt over time. The rest three trackingmethods could get similar performance before about #330. Thelarge errors occur during large scale and fast pose changessimultaneously. After about frame #330, the ITSL tracker loses the

[2] (the top row), the ITSL tracker (the second row) and the ICTSL tracker (the third

lace to the dark one.

Page 10: Incremental tensor biased discriminant analysis: A new color-based visual tracking method

ARTICLE IN PRESS

J. Wen et al. / Neurocomputing 73 (2010) 827–839836

object, and could not recover to the right position, because of theless of the color and discriminative information. After about frame#369, the ICTSL tracker loses the track and drift off. However, theproposed method could keep steady tracking performance.

Due to the scale and motion change gently in the second video(Fig. 11) compared to the first video (Fig. 10), the tracking errors of thefour methods are also smaller in Fig. 14(a) than those inFig. 13(a). The same conclusion could be drawn by comparingthe location errors in Figs. 13(b) and 14(b). The location error inFig. 10 is more than that in Fig. 11. Moreover, since the location andscale variant is small for the example in Fig. 11, the mean (symbol ‘‘*’’)and derivation of the location error in Fig. 14(b) would be smaller

0 50 100 150 200 250 300 350 4000

20

40

60

80

100

120

140

160

Frame number

Trac

king

err

or

SMOGITSLICTSLITBDA

Fig. 13. The comparison for the tracking errors by the four methods shown in Fig. 10:

location and (b) the comparison of the location precision and derivation by the four m

Fig. 12. A comparison of our tracking method (the bottom row) with the SMOG tracker

row) on the video sequence of a person moving close to the camera, from the bright t

than that in Fig. 13(b), compared to the example in Fig. 10, which isreasonable theoretically. However, in the second video sequence, theobject appearance is characterized by the drastic illumination, whichwould result in large variant of the pixel intensity value and lost trackas seen in Fig. 14(a). The SMOG tracker still takes on instable trackingresult respect to the ground truth. The tracking error of the ITSLtracker could get steady performance compared to the SMOG tracker,but the error is still a little larger than the rest two trackers, due toonly use the illumination intensity and lost the chromatic spatialinformation. The ICTSL could get more accurate tracking performancecompared with the two methods mentioned before, but the trackerstill lost due to it does not take the background into account. The

0.5 1 1.5 2 2.5 3 3.5 4 4.5

-10

0

10

20

30

40

SMOG ICTSL ITBDA

Loca

tion

erro

r

ITSL

(a) the tracking error between the estimated locations and the manually labeled

ethods.

[2] (the top row), the ITSL tracker (the second row) and the ICTSL tracker (the third

o the dark place.

Page 11: Incremental tensor biased discriminant analysis: A new color-based visual tracking method

ARTICLE IN PRESS

50 100 150 200 250 300 350 400 4500

10

20

30

40

50

60

Frame number

Trac

king

err

or

SMOGITSLICTSLITBDA

0.5 1 1.5 2 2.5 3 3.5 4 4.5-2

0

2

4

6

8

10

12

14

16

SMOG ICTSL ITBDA

Loca

tion

erro

r

ITSL

Fig. 14. The comparison for the tracking errors by the four methods shown in Fig. 11: (a) the tracking error between the estimated locations and the manually labeled

location and (b) the comparison of the location precision and derivation by the four methods.

Fig. 15. Two pedestrians walk far away from the camera and toward the shop entry. The frame number is #9, #56, #102, #145, #174, #183, #191, and # 213 in order.

Fig. 16. Three pedestrians walk far away from the camera and toward the shop entry. The frame number is #25, #65, #123, #178, #243, #279, #310, and # 332 in order.

J. Wen et al. / Neurocomputing 73 (2010) 827–839 837

proposed tracking method is able to get rather satisfied results asshown in Fig. 14(a).

In addition, our simulation environment is Matlab 7.7 on a PC(2.0 GHz Intel Core Duo, 1 GB RAM). The computational time ofthe tracking algorithm is about 0.245 fps, with 300 particlenumbers, for the single-object tracking. It will cost as the objectsnumber times as the single object, due to the objects areindependent to each other for the multiple-objects tracking.

5.5. Tracking multiple objects

All the experimental results above are the single-objecttracking examples. Here, we use the rest three video sequencesto verify the multiple objects tracking ability of the proposedmethod in Figs. 15–17, respectively.

In Fig. 15, there are two objects (pedestrians) of interest,labeled by white and cyan rectangle, respectively. The two

Page 12: Incremental tensor biased discriminant analysis: A new color-based visual tracking method

ARTICLE IN PRESS

Fig. 17. Three pedestrians walk far away from the camera and toward the shop entry. The frame number is #4, #19, #35, #48, #64, #74, #78, and # 88 in order.

Fig. 18. The failure of the proposed tracking method to track three pedestrians in the video sequence in Fig. 17. The frame number is #108, #122, #135, and #178 in order.

J. Wen et al. / Neurocomputing 73 (2010) 827–839838

object walk from the bottom right scene into the shop, experien-cing the scale and shape change, as well as the partial occlusionbefore they go into the shop. The object labeled by cyan rectanglewalks in front of the other one, which results in that theappearance of the black dressed pedestrian is occluded partiallyby the red dressed one (#174, #183, and #191). After that, the reddressed pedestrian is occluded by another pedestrian (#213), whois walking towards the camera. The proposed tracking methodcan be able to track the two pedestrians well for this case.

In Fig. 16, there are three pedestrians walking close to thecamera. In this video sequence, the three pedestrians walktowards the camera, so the scales of the objects get biggergradually, accompanying various postures obviously during theirbeing closer. The tensor representation in the proposed trackingmethod, could maintain the chromatic appearance spatial-structure, even if the appearance of the object is warpedirregularly. In Fig. 17, there are three pedestrians walking farfrom the camera, accompanying two of them exchanging theirposition. The proposed tracking method could track the assignedobjects even though there is occlusion during their exchanges.

Like Fig. 17, there are also two pedestrians who changed theirposition in Fig. 18. The difference is that these two pedestrianslabeled by red and cyan rectangles have similar color. It is the similarcolor that results in the tracking failure of the proposed method. So,the proposed tracking method would lose the object of interestwhen there is the full-body occlusion, especially the occluding objecthas the same or similar color as the occluded object.

6. Conclusion and future work

This paper proposes a new color-clue-based visual trackingmethod, which can incrementally learn the object color structureinformation and discriminate the object from the background inonline way. For this application purpose, we extend the biaseddiscriminant analysis from the vector-based method to tensor-based way in order to keep the structure information wellcombine with the color information, present batch tensor biaseddiscriminant analysis (TBDA) and its incremental version, ITBDA,for distinguish the object and background. Moreover, the TBDAcould be able to deal with the SSS problems, as well as give a good

talk on the extension of vector-vise method to the high order one.The experimental results show the robustness of the proposedalgorithm to tracking object undergoes large pose variant,illumination change and occlusions.

Although our method could track object well for most cases,we have to point out the proposed method is a little sensitive tothe very drastic illumination, and not adaptive to the full-bodyocclusion whose color is same as the object’s. Moreover, theproposed tracking algorithm cannot be implemented in real time,which will be one of the future works.

Acknowledgements

We want to thank the helpful comments and suggestions fromthe anonymous reviewers. This research was supported by theNational Natural Science Foundation of China (60771068,60702061, 60832005), the Open Project Program of the NationalLaboratory of Pattern Recognition (NLPR) in China and theNational Laboratory of Automatic Target Recognition, ShenzhenUniversity, China.

References

[1] S. Birchfield, Elliptical head tracking using intensity gradient and colorhistograms, in: Proceedings of the 1998 IEEE International Conference onComputer Vision and Pattern Recognition, 1998, pp. 232–237.

[2] H.Z. Wang, D. Suter, K. Schindler, C.H. Shen, Adaptive object tracking based onan effective appearance filter, IEEE Transactions on Pattern Analysis andMachine Intelligence 29 (9) (2007) 1661–1667.

[3] Y. Hayashi, H. Fujiyoshi, Mean-shift-based color tracking in illuminancechange, Lecture Notes in Artificial Intelligence 5001 (2008) 302–311.

[4] M. Isard, A. Blake, Contour tracking by stochastic propagation of conditionaldensity, in: The Fourth European Conference on Computer Vision, LNCX 1064,1996, pp. 343–356.

[5] M.S. Allili, D. Ziou, Object tracking in videos using adaptive mixture modelsand active contours, Neurocomputing 71 (10–12) (2008) 2001–2011.

[6] T. Cootes, G. Edwards, C. Taylor, Robust real-time periodic motion detectionanalysis and applications, IEEE Transactions on Pattern Analysis and MachineIntelligence 23 (6) (2001) 681–685.

[7] P. Perez, C. Hue, J. Vermaak, M. Gangnet, Color-based probabilistic tracking,in: The Seventh European Conference on Computer Vision, Lecture Notes inComputer Science, vol. 2350, 2002, pp. 661–675.

[8] D. Comaniciu, V. Ramesh, P. Meer, Kernel-based object tracking, IEEE Transac-tions on Pattern Analysis and Machine Intelligence 25 (5) (2003) 564–577.

[9] S. Avidan, Ensemble tracking, in: Proceedings of the 2005 IEEE InternationalConference on Computer Vision and Pattern Recognition, 2005, pp. 494–501.

Page 13: Incremental tensor biased discriminant analysis: A new color-based visual tracking method

ARTICLE IN PRESS

J. Wen et al. / Neurocomputing 73 (2010) 827–839 839

[10] R. Gross, I. Matthews, S. Baker, Active appearance models with occlusion,Image and Vision Computing 24 (2006) 593–604.

[11] D. Ross, J. Lim, M.H. Yang, Adaptive probabilistic visual tracking with incrementalsubspace update, in: The Eighth European Conference on Computer Vision,Lecture Notes in Computer Science, vol. 2350, 2004, pp. 707–720.

[12] R.S. Lin, D. Ross, R.S. Lin, M.H. Yang, Adaptive discriminative generativemodel and its applications, Advances in Neural Information ProcessingSystems (2004) 801–808.

[13] C.H. Shen, A.V.D. Hengel, M.J. Brooks, Visual tracking via efficient kerneldiscriminant subspace learning, in: Proceedings of the 2005 IEEE Interna-tional Conference on Image Processing, 2005, pp. 590–593.

[14] S. Li, M.C. Lee, Fast visual tracking using motion saliency in video, in:Proceedings of the 2007 IEEE International Conference on Acoustics, Speech,and Signal Processing, vol. 1, 2007, pp. 1073–1076.

[15] J. Shao, S.H.K. Zhou, R. Chellappa, Appearance-based tracking and recogni-tion using the 3d trilinear tensor, in: Proceedings of the 2004 IEEE InternationalConference on Acoustics, Speech, and Signal Processing, vol. 3, 2004, pp. 613–616.

[16] X. Li, W.M. Hu, Z.F. Zhang, et al., Robust visual tracking based on incrementaltensor subspace learning, in: Proceedings of the 2007 IEEE InternationalConference on Computer Vision, October 2007, pp. 1–8.

[17] J. Wen, X. Gao, X. Li, D. Tao. Incremental learning of weighted tensor subspace forvisual tracking, in: Proceedings of the 2009 IEEE International Conference onSystems, Man, and Cybernetics, San Antonio, USA, October 2009, pp. 3788–3793.

[18] D. Tao, X. Li, X. Wu, S.J. Maybank, General tensor discriminant analysis andgabor features for gait recognition, IEEE Transactions on Pattern Analysis andMachine Intelligence 29 (10) (2007) 1700–1715.

[19] X. Zhou, T. Huang, Small sample learning during multimedia retrieval usingbiasmap, in: Proceedings of the 2001 IEEE International Conference onComputer Vision and Pattern Recognition, vol. 1, 2001, pp. 11–17.

[20] X. Zhou, T. Huang, Comparing discriminate transformations and SVM forlearning during multimedia retrieval, in: Proceedings of the ACM Conferenceon Multimedia, 2001, pp. 137–146.

[21] D. Tao, X. Tang, X. Li, Y. Rui, Kernel direct biased discriminant analysis: a newcontent-based image retrieval relevance feedback algorithm, IEEE Transac-tions on Multimedia 8 (4) (2006) 716–725.

[22] D. Tao, X. Li, W. Hu, et al., Supervised tensor analysis, in: Proceedings of the5th IEEE International Conference on Data Mining, 2005, pp. 450–457.

[23] D. Tao, M. Song, X. Li, J. Shen, et al., Bayesian tensor approach for 3-D facemodelling, IEEE Transactions on Circuits and Systems for Video Technology18 (10) (2008) 1397–1410.

[24] J. Sun, D. Tao, S. Papadimitriou, et al., Incremental tensor analysis: theory andapplications, ACM Transactions on Knowledge Discovery in Data 2 (3) (2008).

[25] D. Tao, J. Sun, J. Shen, et al., Bayesian tensor analysis, in: Proceedings of the IEEEInternational Joint Conference on Neural Networks, 2008, pp. 1403–1410.

[26] D. Tao, J. Sun, J. Shen, et al., Probabilistic tensor analysis with Akaike andBayesian information criteria, in: Proceedings of the 14th InternationalConference on Neural Information Processing 2007, Part I, Lecture Notes inComputer Science, vol. 4984, 2008, pp. 791–801.

[27] D. Tao, X. Li, X. Wu, S. Maybank, Tensor rank one discriminant analysis—

a convergent method for discriminative multilinear subspace selection,Neurocomputing 71 (10–12) (2008) 1866–1882.

[28] D. Tao, X. Li, X. Wu, S. Maybank, Elapsed time in human gait recognition: anew approach, in: Proceedings of the 2006 IEEE International Conference onAcoustics, Speech, and Signal Processing, vol. 2, 2006, pp. 177–180.

[29] D. Tao, X. Li, X. Wu, S. Maybank, Geometry mean for subspace selection, IEEETransactions on Pattern Analysis and Machine Intelligence 31 (2) (2009) 260–274.

[30] X. Gao, Y. Yang, D. Tao, X. Li, Discriminative optical flow tensor for video semanticanalysis, Computer Vision and Image Understanding 113 (3) (2009) 372–383.

[31] X. Gao, X. Li, J. Feng, D. Tao, Shot-based video retrieval with optical flow tensor andHMM, Pattern Recognition Letters 30 (2) (2009) 140–147.

[32] Y. Yuan, Y. Pang, J. Pan, X. Li, Scene segmentation based on IPCA for visualsurveillance, Neurocomputing 72 (10–12) (2009) 2450–2454.

[33] D. Xu, S. Yan, D. Tao, et al., Human gait recognition with matrixrepresentation, IEEE Transactions on Circuits and Systems for VideoTechnology 16 (7) (2006) 896–903.

[34] H. Yu, J. Yang, A direct LDA algorithm for high-dimensional data withapplication to face recognition, Pattern Recognition 34 (2001) 2067–2070.

[35] J. Lu, K. Plataniotis, A. Venetsanopoulos, Face recognition using kernel directdiscriminant analysis algorithms, IEEE Transactions on Neural Network 14 (1)(2003) 117–126.

Jing Wen received the B.Sc. degree in Electronic Informa-tion Science and Technology from Shanxi University,Taiyuan, China, in 2003, and the M.Eng. degree in Signaland Information Processing from Xidian University, Xi’an,China, in 2006. Since August 2006, she has been pursuingher Ph.D. degree in Pattern Recognition and IntelligentSystem at Xidian University. Her research interestsinclude pattern recognition and computer vision.

Xinbo Gao received the B.Sc., M.Sc. and Ph.D. degrees insignal and information processing from Xidian University,China, in 1994, 1997 and 1999, respectively. From 1997 to1998, he was a research fellow in the Department ofComputer Science at Shizuoka University, Japan. From2000 to 2001, he was a postdoctoral research fellow in theDepartment of Information Engineering at the ChineseUniversity of Hong Kong. Since 2001, he joined the Schoolof Electronic Engineering at Xidian University. Currently,he is a Professor of Pattern Recognition and IntelligentSystem, and Director of the VIPS Lab, Xidian University.His research interests are computational intelligence,

machine learning, computer vision, pattern recognition

and artificial intelligence. In these areas, he has published 4 books and around 100technical articles in refereed journals and proceedings including IEEE TIP, TCSVT, TNN,TSMC, etc. He is on the editorial boards of journals including EURASIP Signal Processing(Elsevier) and Neurocomputing (Elsevier). He served as general chair/co-chair orprogram committee chair/co-chair or PC member for around 30 major internationalconferences.

Yuan Yuan is currently a Lecturer with the School ofEngineering and Applied Science, Aston University, UnitedKingdom. She received her B.Eng. degree from theUniversity of Science and Technology of China, China,and the Ph.D. degree from the University of Bath, UnitedKingdom. She has over 60 scientific publications injournals and conferences on visual information processing,compression, retrieval, etc. She is an associate editor ofInternational Journal of Image and Graphics (WorldScientific), an editorial board member of Journal ofMultimedia (Academy Publisher), a guest editor of SignalProcessing (Elsevier), and a guest editor of Recent Patents

on Electrical Engineering. She was a chair of some

conference sessions, and a member of program committees of many conferences. Sheis a reviewer for several IEEE transactions, other international journals and conferences.

Dacheng Tao received the B.Eng. degree from theUniversity of Science and Technology of China (USTC),the M.Phil. degree from the Chinese University of HongKong (CUHK), and the Ph.D. degree from the Universityof London (Lon). Currently, he is a Nanyang AssistantProfessor with the School of Computer Engineering inthe Nanyang Technological University and holds avisiting post in Lon. He is a Visiting Professor in the XiDian University and a Guest Professor in the Wu HanUniversity. His research is mainly on applying statis-tics and mathematics for data analysis problems incomputer vision, multimedia, machine learning, data

mining, and video surveillance. He has published more

than 100 scientific papers including IEEE TPAMI, TIP, TKDE, CVPR, ECCV, NIPS,ICDM; ACM TKDD, Multimedia, KDD, etc., with best paper runner up awards andfinalists. One of his TPAMI papers received an interview with ScienceWatch.com(Thomson Scientific). His H-Index in google scholar is 14 and his Erdos number is3. He holds the K.C. WONG Education Foundation Award.

Jie Li received the B.Sc., M.Sc. and Ph.D. degrees in Circuitand System from Xidian University, China, in 1995, 1998and 2005, respectively. Since 1998, she joined the Schoolof Electronic Engineering at Xidian University. Currently,she is an Associate Professor of Xidian University. Herresearch interests include computational intelligence,machine learning, and image processing. In these areas,she has published over 30 technical articles in refereedjournals and proceedings including IEEE TCSVT, IJFS, etc.