projection under pairwise distance controls

35
HAL Id: hal-01420662 https://hal.archives-ouvertes.fr/hal-01420662v2 Preprint submitted on 3 Dec 2017 (v2), last revised 23 Dec 2020 (v5) HAL is a multi-disciplinary open access archive for the deposit and dissemination of sci- entific research documents, whether they are pub- lished or not. The documents may come from teaching and research institutions in France or abroad, or from public or private research centers. L’archive ouverte pluridisciplinaire HAL, est destinée au dépôt et à la diffusion de documents scientifiques de niveau recherche, publiés ou non, émanant des établissements d’enseignement et de recherche français ou étrangers, des laboratoires publics ou privés. Projection under pairwise distance controls Alawieh Hiba, Nicolas Wicker, Christophe Biernacki To cite this version: Alawieh Hiba, Nicolas Wicker, Christophe Biernacki. Projection under pairwise distance controls. 2017. hal-01420662v2

Upload: others

Post on 02-Feb-2022

1 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Projection under pairwise distance controls

HAL Id: hal-01420662https://hal.archives-ouvertes.fr/hal-01420662v2Preprint submitted on 3 Dec 2017 (v2), last revised 23 Dec 2020 (v5)

HAL is a multi-disciplinary open accessarchive for the deposit and dissemination of sci-entific research documents, whether they are pub-lished or not. The documents may come fromteaching and research institutions in France orabroad, or from public or private research centers.

L’archive ouverte pluridisciplinaire HAL, estdestinée au dépôt et à la diffusion de documentsscientifiques de niveau recherche, publiés ou non,émanant des établissements d’enseignement et derecherche français ou étrangers, des laboratoirespublics ou privés.

Projection under pairwise distance controlsAlawieh Hiba, Nicolas Wicker, Christophe Biernacki

To cite this version:Alawieh Hiba, Nicolas Wicker, Christophe Biernacki. Projection under pairwise distance controls.2017. �hal-01420662v2�

Page 2: Projection under pairwise distance controls

Noname manuscript No.(will be inserted by the editor)

Projection under pairwise distance control

Alawieh Hiba · Nicolas Wicker · ChristopheBiernacki

Received: date / Accepted: date

Abstract Visualization of high-dimensional and possibly complex (non con-tinuous for instance) data onto a low-dimensional space may be di�cult. Sev-eral projection methods have been already proposed for displaying such high-dimensional structures on a lower-dimensional space, but the information lostis not always easy to use. Here, a new projection paradigm is presented todescribe a non-linear projection method that takes into account the projec-tion quality of each projected point in the reduced space, this quality beingdirectly available in the same scale as this reduced space. More speci�cally,this novel method allows a straightforward visualization data in R2 with asimple reading of the approximation quality, and provides then a novel variantof dimensionality reduction.

Keywords Data visualization · dimension reduction · multidimensionalscaling · principal component analysis.

1 Introduction

Several domains in science use data with large number of variables in theirstudies such as in biological [12,18], chemical [31], geographical [33], �nancial[22] studies and many others studies. These data can be viewed as a largematrix and extracting results from this type of matrix is often hard and com-plicate. In such cases, it is desirable to reduce the number of dimensions of databy conserving as much information as possible from the given initial matrix.Di�erent types of multivariate data analysis methods were developed to studythese data as dimensionality reduction, variable selection, cluster analysis andothers. These methods are directly related to the main goal of the researcher

H. Alawieh, N. Wicker, C. BiernackiUniversité Lille 1, - UFR de Mathématiques, cité scienti�que, 59655 Villeneuve d'Ascq,FranceE-mail: [email protected]

Page 3: Projection under pairwise distance controls

2 Alawieh Hiba, Nicolas Wicker, Christophe Biernacki

such as dimensionality reduction to summarize the data, variable selection tochoose the pertinent variables from the set of candidate variables and clus-ter analysis to group the objects or variables. In our study, we are focus ondimension reduction.

Dimensionality reduction techniques can be used in di�erent ways like datadimensionality reduction that projects the data form a high-dimensional spaceto a low-dimensional space or data visualization that provides a simple inter-pretation of the data in R2

or R3

.

Many data dimensionality reduction and data visualization methods have beenproposed to drop the di�culties associated to the high dimensional data [27,17,13,25,9]. To quote a few, principal component analysis (PCA) [21], multi-dimensional scaling (MDS) [32], scatter plot matrix [14], parallel coordinates[20] and Sammon's mapping [28] are some of the known used methods.

Scatter plot matrix, parallel coordinates and Sammon's mapping methods arewidely used to visualize multidimensional data sets. The �rst two methodshave as inconvenience that when the number of dimensions grows, importantdimensional relationships might not be visualized. Concerning Sammon's map-ping method, the inconvenience is similar to that founded in PCA and MDSfrom the point of view of projection quality. Indeed, the quality of projec-tion assessed by the percentage of variance that is conserved or by the stressfactor is a global quality measure and takes only into account what happensglobally but in some projection methods like PCA, a local measure is de�nedto indicate the projection quality of each projected point taken individually.This local measure is evaluated by the squared cosine of angle between theprincipal space and the vector of the point. A good representation in the pro-jected space is hinted by high squared cosine values. This measure is usefulin cases of linear projection as happens in PCA but cannot be applied tothe case of non-linear projection. Moreover, PCA will fail to give usa "good" representation in case of non-linear con�gurations there-fore, Kernel PCA has been developed to extract non-linear principalcomponents. Many disadvantages of Kernel PCA can be founded es-pecially by comparing with other method. The storage of the data asdot products in the Gram matrix is too expensive since the size ofkernel matrix increases quadratically with the number of data [5]. Additionally, in kernel PCA we can �nd the same local qualityprojection measure problem designed in PCA.

In this paper, we propose a new non-linear projection method that projectsthe points in a reduced space by using the pairwise distance between pairs ofpoints and by taking into account the projection quality of each point takenindividually. This projection leads to a representation of the points as circleswith a di�erent radius associated to each point. Henceforth, this method will becalled "Projection under pairwise distance control". The main contributions ofthis study are to give a simple data visualization in R2 with a straightforwardinterpretation and provide a new variant of dimensionality reduction. First,the new projection method is presented in Section 2. Then, in Section 3, thealgorithms used in the resolution of optimization problems related to this

Page 4: Projection under pairwise distance controls

Projection under pairwise distance control 3

method are illustrated. Next, Section 4 shows the application of this methodto various real data sets. Finally, Section 5 concludes this work.

2 Projection under pairwise distance control

Let us consider n points given by their pairwise distance noted dij for i, j ∈{1, . . . , n}. The task here is to project these points using distances into areduced space Rm by introducing additional variables, called hereafter radii,that indicate to which extent the projection of each point is accurate. Thelocal quality is then given by the values of the radii. A good quality projectionof point i is indicated by a small radius value noted ri. It will be important tonote that both units of dij 's and ri's are identical, allowing direct comparison.Before developing our method, an overview of principal component analysis(PCA) is presented to highlight the interest of our method.

2.1 Principal Component Analysis (PCA)

PCA method is the most used method in the data visualization and dimension-ality reduction. This method is a linear projection technique applied when thedata is linearly separable. PCA problem can be stated as an optimization prob-lem involving the squared Euclidean distances [27]. This optimization problemis the following:

PPCA :

min

A∈Mp×q

∑1≤i<j≤n

|d2ij − ‖Ayi −Ayj‖2|

s.t. rank(A) = mAAT = Ip

where yi ∈ Rp is the original coordinates vector of point i, d2ij is the squared

distance for couple (i, j) given by ‖yi− yj‖2 and A is the projection matrix ofdimension p× q with q is the reduced space dimension.By construction, PCA cannot take into account non-linear struc-tures, since it describes the data in terms of a linear subspace.Therefore, a new concept of PCA has been developed to lead withnon-linear structures that is called Kernel PCA

2.2 Kernel PCA (KPCA)

The KPCA idea is to perform PCA in a feature space noted Fproduced by a non-linear mapping of data from its space into thefeature space F, where the low-dimensional latent structure is, hope-fully, easier to discover. The mapping function noted Φ is consideredas:

Page 5: Projection under pairwise distance controls

4 Alawieh Hiba, Nicolas Wicker, Christophe Biernacki

Φ : Rp → FX → Φ(X)

The original data yi is then represented in the feature space asΦ(yi), where F and Φ are not known explicitly but obtained thanksto the "kernel trick". So, from the Gram matrix K given by K = kijwith kij = K(yi, yj) = 〈Φ(yi), Φ(yj)〉, KPCA is based to �nd the �rst meigenvectors corresponding to the largest eigenvalues λi of the Grammatrix K. Let Vv for v = 1, . . . ,m are the eigenvectors in the featurespace and PΦ(yi) is the projection of Φ(yi) onto the subspace V1, . . . , Vm.KPCA problem can be represented as a minimization problem of thefollowing error:

EKPCA : ‖Φ(y)− PΦ(y)‖22

with PΦ(y) =

m∑v=1

〈Φ(y);Vv〉Vv

Furthermore, the only measures used to evaluate the projection qualityof points are the squared cosines values which can only be used in the caseof linear projection. Thus, the individual control of projection is no moreguaranteed using non-linear projection method.

2.3 Multidimensional Scaling (MDS)

Likewise PCA, Multidimensional scaling (MDS) consists in �nding a new datacon�guration in a reduced space. The main di�erence between these two meth-ods is that the input data in MDS are typically comprehensive measures ofsimilarity or dissimilarity between objects, they are called "proximity". Thekey idea of MDS is to perform dimensionality reduction in a way to approx-imate high-dimensional distances noted dij by the low-dimensional distancesδij where δij is equal to the distance between xi and xj the coordinates ofi and j in the reduced space. In the classical and simplest case of MDS, theleast-squares loss function noted Stress is given as follows:

Stress =

√ ∑1≤i<j≤n

(|dij − ‖xi − xj‖)2·

By minimizing the stress function, we �nd the best con�guration of (x1, . . . , xn)such that the distances �t to the initial distances dij .

Now if we consider n variables r1, . . . , rn ∈ R+, the sum of which bounds thestress function, the optimization problem PMDS can be equivalently rewrittenas:

Page 6: Projection under pairwise distance controls

Projection under pairwise distance control 5

PMDS :

min

r1,...,rn∈R+

n∑i=1

ri

s.t.

n∑i=1

ri ≥1

n− 1

√ ∑1≤i<j≤n

(|dij − ‖xi − xj‖)2

Then, we can observe that the constraint on∑ni=1 ri can be modi�ed to have a

stronger control on each dij in the following way: |dij−‖xi−xj‖| ≤ ri+rj wherexi and xj are projection coordinates of points i and j. Here the projectioncoordinates are not obtained necessarily by linear projection anymore.

Our new non-linear projection method that controls individually the pro-jection of points is developed hereafter.

2.4 Our proposed method

Let x1, . . . , xn be the coordinates of the projected points in Rm. Radiiare an important element of the paper introduced to assess howmuch the distance between two projected points (i, j) given by ‖xi − xj‖is far from given distance dij. Indeed, radii (ri, rj) for couple (i, j)are small when ‖xi − xj‖ is close to dij. Figure 1 depicts this idea:for each point Pi ∈ {1, . . . , n}, the projected point of point Pi belongsto a sphere with center xPi

and radius rPisuch that for each couple

(i, j) ∈ {1, . . . , n} we have ‖xi − xj‖ − (ri + rj) ≤ dij ≤ ‖xi − xj‖+ ri + rj.This idea can be expressed by �nding the value of radii that satisfy these two

xi xj

ri rj

dij

• •

Fig. 1: Examples of radii for bounding the original distance dij

constraints:

•n∑i=1

ri is minimum.

• dij ∈ {‖xi − xj‖ − ri − rj , ‖xi − xj‖+ ri + rj}, for 1 ≤ i < j ≤ n.

Page 7: Projection under pairwise distance controls

6 Alawieh Hiba, Nicolas Wicker, Christophe Biernacki

The projection under pairwise distance control problem can be written as thefollowing optimization problem:

Pr,x :

minr1,...,rn∈R+,x1,...,xn∈Rk

n∑i=1

ri

s.t |dij − ‖xi − xj‖| ≤ ri + rj , for 1 ≤ i < j ≤ n

Of course, by �xing the coordinates vectors xi for all i ∈ {1, . . . , n}, usingprincipal component analysis or any other projection method, the problemcan easily be solved in (r1, . . . , rn) using linear programming. This problemcan be written as follows:

Pr :

minr1,...,rn∈R+

n∑i=1

ri

s.t |dij − ‖xi − xj‖| ≤ ri + rj , for 1 ≤ i < j ≤ n

We can remark that a solution of Pr always exists. Indeed, to satisfy the con-straints it is enough to increase all ri. Besides, solving Pr with �xed coordinates(x1, . . . , xn) does not lead in general to the optimum of problem Pr,x.

2.5 Visualization example

Let us apply our projection method to a simple example by taking a tetrahe-dron with all pairwise distance equal to 1. For problem Pr, the coordinatesxi for i = 1, . . . , 4 are obtained using multidimensional scaling. Using linearand non-linear optimization packages in Matlab respectively for problems Prand Pr,x give a value of

∑ni=1 ri equal to 0.7935 for problem Pr and 0.4226

for Pr,x. Figure 2a corresponds to the �rst solution and Figure 2b correspondsto the second one. In Figures 2a and 2b, we depict circles with di�erent radii.The circle color is related to the radius values, the shades of gray lie betweenwhite and black in the descending direction of the radius values; the smallerthe radius, the darker circle. The points that have circles with small radii areconsidered as well projected points. Note that the points that are representedas points and not circles are very well projected, having radii almost equal tozero. In Figure 2a, half of the points is well projected whereas the other halfhave large radii indicating that they are not well projected. In Figure 2b justone circle appears marking that the projection quality using problem Pr,x isbetter than Pr.

2.6 Link with other methods

Multidimensional �tting (MDF) [4] is a method that modi�es the coordinatesof a set of points in order to make the distances calculated on the modi�edcoordinates similar to given distances on the same set of points. We call �target

Page 8: Projection under pairwise distance controls

Projection under pairwise distance control 7

-1 -0.8 -0.6 -0.4 -0.2 0 0.2 0.4 0.6 0.8 1

-1

-0.8

-0.6

-0.4

-0.2

0

0.2

0.4

0.6

0.8

1

(a)

-1 -0.8 -0.6 -0.4 -0.2 0 0.2 0.4 0.6 0.8 1

-1

-0.8

-0.6

-0.4

-0.2

0

0.2

0.4

0.6

0.8

1

(b)

Fig. 2: Projected points after solving Pr and Pr,x. (a) shows the projectionobtained from the solution of Pr using MDS and (b) shows that obtained fromthe solution of Pr,x.

matrix� the matrix that contains the points coordinates and �reference matrix�the matrix that contains the given distances.

Let us note X = {x1| · · · |xn} the target matrix and D = {dij} the referencematrix. The objective function of MDF problem is given by:

∑1≤i<j≤n

|dij − ‖xi − xj‖|.

Property 1 Problem Pr,x is bounded below by1

n− 1

∑1≤i<j≤n

|dij − ‖xi − xj‖|

where x1, . . . , xn is the optimum of the associated MDF problem.

Proof By summing all the constraints of problem Pr,x we obtain:

∑1≤i<j≤n

|dij − ‖xi − xj‖| ≤∑

1≤i<j≤n

ri + rj = (n− 1)

n∑i=1

ri

So,

n∑i=1

ri ≥1

n− 1

∑1≤<i<j≤n

|dij − ‖xi − xj‖|, which concludes the proof.

Page 9: Projection under pairwise distance controls

8 Alawieh Hiba, Nicolas Wicker, Christophe Biernacki

3 Optimization tools

3.1 Lower Bound of the objective function of problem Pr,x

Minimization problem Pr,x is too hard to be solved analytically. A way toassess how good a solution is, is to provide a lower bound on the objectivefunction. Then, if the bound is close to the best found solution, we can concludethat this solution is �xed. Thus, in this section we present such a lower boundof problem Pr,x.Let x1, · · · , xn; r1, · · · , rn a feasible solution of Pr,x and M ∈ R such thatM = max

(i,j)‖ xi − xj ‖. We consider three functions noted f , g, h depending on

M as follows:

� f(M) =

√(1− n

3

)M2 +

1

n− 1

∑i<j

d2ij −M .

� g(M) = |M − dmax|.� h(M) = min

i<jmax

k<l;k,l 6=i,jmin

{L1ijkl;L

2ijkl

where:

� dmax = max1≤i<j≤n

{dij},

� L1ijkl = max

{djk − dkl

2;djl − dkl

2; |dij −M |

},

� L2ijkl = max

{dik − dkl

2;dil − dkl

2; |dij −M |

}.

Using results presented in appendix B, we can write:

n∑i=1

ri ≥ f(M)

n∑i=1

ri ≥ g(M)

n∑i=1

ri ≥ h(M)·

Consequently,

n∑i=1

ri ≥ max{f(M); g(M);h(M)}· (1)

The inequality (1) is true for all solutions of Pr,x particularly for the optimalsolution. Thus:

n∑i=1

ropti ≥ max{f(M); g(M);h(M)}· (2)

Page 10: Projection under pairwise distance controls

Projection under pairwise distance control 9

Hence, the lower bound is given by:

n∑i=1

ropti ≥ minM

max{f(M); g(M);h(M)} for all feasible solutions·

Given M , a lower bound of problem Pr,x is derived. Afterwards, a bound freeof M is given by minimizing the bounds depending on M .By applying this bound to the tetrahedron example, the three functions areplotted. The result is shown in Figure 3. The lower bound is equal to 0.1276for M = 1.1276 and as we have seen the minimum obtained by solving Pr,x isequal to 0.42.

M

0 0.5 1 1.5 2

∑r i

0

0.5

1

1.5

f

g

h

Fig. 3: The curves of the three functions f, g and h. Functions g and h are equaldue to the fact that all distances are equal to 1 . The minimal intersection isgiven by the black circle for M = 1.1276 and

∑ni=1 ri > 0.1276.

3.2 Initialization point of problem Pr,x

Di�erent resolutions of problem Pr,x can be obtained using di�erent initialvalues of matrix X. Three possible initial values can be used. The �rst ofthem is the matrix obtained by PCA or another projection method. In whatfollows, we present two other possibilities.

Initial point using squared distances The optimization problem Pr,x can bechanged by taking the squared distances between points instead of the dis-tances. Rewriting r2i as Ri, the problem is changed into

PR,x :

minR1,...,Rn∈R+,x1,...,xn∈Rk

n∑i=1

Ri

s.t. |d2ij − ‖xi − xj‖2| ≤ Ri +Rj , for 1 ≤ i < j ≤ n.

Page 11: Projection under pairwise distance controls

10 Alawieh Hiba, Nicolas Wicker, Christophe Biernacki

The transformation is interesting as if the constraints of problem PR,x aresatis�ed, the constraints of problem Pr,x will also be satis�ed. Indeed:

|d2ij − ‖xi − xj‖2| ≤ Ri +Rj = r2i + r2j

⇒ (dij − ‖xi − xj‖) (dij + ‖xi − xj‖) ≤ r2i + r2j ≤ (ri + rj)2

⇒ |dij − ‖xi − xj‖|2 ≤ (ri + rj)2

⇒ |dij − ‖xi − xj‖| ≤ (ri + rj)·

That way problem PR,x can serve as an initial step for solving problem Pr,x.

Initial point using improved solution of problem Pr First, we give two proper-ties which provide a way to improve the optimization results of problem Pr,x.

Property 2 Let us consider a point xi such that for an index j, the followinginequality is satis�ed :

|dij − ‖xi − xj‖| ≤ ri + rj ,

and the other inequalities involving i are not satis�ed. Then, the correspondingsolution can be improved by moving xi along the line xj − x∗i with x∗i is thevector of coordinates for point j in order decrease ri and |dij − ‖xi − xj‖|.

Proof The above condition means that xi is rewritten xi + a(xj − xi) witha ∈ R and we look for a such that |dij − ‖xi + a(xj − xi) − xj‖| < ri + rj .In particular a ≤ 0 if dij − ‖xi − xj‖ ≥ 0 and a > 0 otherwise. Let us nowconsider the other inequalities corresponding to index pairs (i, k) with k 6= j.For each of them, either ∃a ∈ [a

k, a′′

k ] with a′

k < 0 and a′′

k > 0 such that

|dij − ‖xi + a(xj − xi)− xj‖| ≤ ri + rj ,

as these constraints are unsaturated. Finally, if we take a di�erent from 0 in[a′, a′′] with a

′= maxk a

k and a′′= mink a

′′

k , all constraints involving i getunsaturated so that ri can be decreased, decreasing so the objective function.Depending on whether amust be negative or positive, we take a = a

′or a = a

′′

respectively.

Another manner to improve the resolution of problem Pr,x is to e�ectuate ascale change by multiplying the coordinates xi, for i = 1, . . . , n, by a constanta ∈ R. Thus, the new optimization problem is given by:

Pr,a :

minr1,...,rn,a∈R+

n∑i=1

ri

s.t. |dij − a‖xi − xj‖| ≤ ri + rj

Page 12: Projection under pairwise distance controls

Projection under pairwise distance control 11

Property 3 Let r1, . . . , rn;x1, . . . , xn be a feasible solution of Pr,x, if ∃a such

that η(a) <

n∑i=1

ri with η(a) =∑

1≤i<j≤n

|dij − a‖xi − xj‖|, then ∃ r̃1, . . . , r̃n a

solution of Pr,a such that

n∑i=1

r̃i <

n∑i=1

ri.

Proof Let us consider r1, . . . , rn;x1, . . . , xn a feasible solution of problem Pr,xand a, r̃1, r̃2, . . . , r̃n the optimal solution of Pr,a. For the solution of Pr,a, foreach point i, we have a certain saturated constraint associated to point k notedCik(i), otherwise it would not be an optimum. So, we have:

|di1 − a‖xi − x1‖| ≤ r̃i + r̃1...

|dik(i) − a‖xi − xk(i)‖| = r̃i + r̃k(i)...

|dij − a‖xi − xj‖| ≤ r̃i + r̃j...

|din − a‖xi − xn‖| ≤ r̃i + r̃n.

Then, |dik(i) − a‖xi − xk(i)‖| = r̃i + r̃k(i) ≥ r̃i. By summing all points i, fori = 1, . . . , n, we obtain:

n∑i=1

|dik(i) − a‖xi − xk(i)‖| ≥n∑i=1

r̃i.

Thus

∑1≤i<j≤n

|dij − a‖xi − xj‖| ≥n∑i=1

|dik(i) − a‖xi − xk(i)‖| ≥n∑i=1

r̃i.

Note η(a) =∑

1≤i<j≤n

|dij − a‖xi−xj‖|, then if η(a) <

n∑i=1

ri there is a solution

of Pr,a such that

n∑i=1

r̃i <

n∑i=1

ri.

The new initial point is then given by using these two properties as follows:

� �rstly, improve the solution of Pr using property 2 by solving Pr,a.� secondly, improve the solution of Pr,a using property 1.

Page 13: Projection under pairwise distance controls

12 Alawieh Hiba, Nicolas Wicker, Christophe Biernacki

3.3 Algorithm 1

Using the di�erent initial values of matrix X presented above, we solve nowproblem Pr,x. For this task, we introduce a new algorithm denoted algorithm1 which gives the best solution that can be obtained using the di�erent initialvalues cited above. This algorithm is consisted of two steps: initialization stepand optimization step and it is presented as follows:

Algorithm 1

Input: D: distance matrix, N : number of iterations.Initialization step

Project the points using PCA or MDS.Solve Pr using an interior-point method. Obtained solution: (XPr

, rPr).

Solve PR,x using an active-set method and starting from the solution of Pr obtained atthe previous step. Obtained solution: (XPR,x

, RPR,x).

X0 ← XPR,X.

for t = 1 to N do

Solve Pr,a starting from X0 using an interior-point method.Improve the solution of Pr,a using property 1. Obtained solution: (XI

Pr,a,rIPr,a

).

X0 ← XIPr,a

.

end for

Optimization step

Optimize Pr,x using an active-set method and starting from X0, XPrand XPR,x

.

Choose the minimal solution obtained by these three di�erent starting points.

3.4 Algorithm 2

Problem Pr,x is a hard problem, so it is natural to resort to stochastic op-timization methods. In the present case, Metropolis-Hastings algorithm [23]allows us to build a Markov chain with a desired stationary distribution. Theonly delicate part is the choice of the proposal distribution and the necessityto solve a Pr problem at each iteration. In details, this Metropolis-Hastingsalgorithm requires:

1- A target distribution:The target distribution is related with the objective function of problemPr,x and it is given by:

π(s) ∝ exp

(−E(x)

T

),

with E an application given by:

E : Rn 7−→ Rx = (x1, . . . , xn) 7−→ E(x) = Solution of problem Pr with x �x.

The variable T is the temperature parameter, to be �xed according to thevalue range of E.

Page 14: Projection under pairwise distance controls

Projection under pairwise distance control 13

2- A proposal distribution:The choice of the proposal distribution is very important to obtain interest-ing results. It should be chosen in such a way that the proposal distributiongets close to the target distribution. The proposal distribution q(X → .)has been constructed as follows, giving priority to the selection of pointsinvolved in saturated constraints:� For each point i, choose a point j(i) with probability equal to:

Pj(i) =λ exp

(−λ(ri + rj(i) − |dij(i) − ‖xi − xj(i)‖|)

)n∑

k=1,k 6=i

λ exp (−λ(ri + rk − |dik − ‖xi − xk‖|))·

� Choose a constant cij(i) using Gaussian distribution Nk(0, σ).� Generate a matrix X∗ by moving each vector xi of matrix Xt−1 asfollows:• If dij(i) − ‖xi − xj(i)‖ > 0 then x∗i = xi + |cij(i) |Lij(i) .• else x∗i = xi − |cij(i) |Lij(i) .

with Lij(i) =xi − xj‖xi − xj‖

.

3- A linear optimization problem:For the matrix X generated in each iteration, we solve the linear optimiza-tion problem Pr.

4 Numerical application

The presented projection method has been applied to di�erent types of realdata sets so as to illustrate its generality.

4.1 The data

Four real data sets are used and divided into three categories:

� Quantitative data: Iris and car data sets.� Categorical data: Soybean data set.� Functional data: Co�ee data set.

The Iris data set [1] is a famous data set and is presented to show that theprojection is as expected. This data set contains 3 classes of 50 instances each,where each class refers to a type of iris plant. The four variables studied in thisdata set are: sepal length, sepal width, petal length and petal width (in cm).Car data set [29] is a data set studied in the book of Saporta (Table 17.1, page428). This data set describes 18 cars according to various variables (cylinders,power, length, width, weight, speed).

Page 15: Projection under pairwise distance controls

14 Alawieh Hiba, Nicolas Wicker, Christophe Biernacki

The soybean data set [30] from UCI Machine Learning Repository character-izes 47 soybean disease case histories de�ned over 35 attributes. Each obser-vation is identi�ed by one of the 4 diseases: Diaporthe Stem Canker (D1),charcoal Rot (D2), Rhizoctonia Root Rot (D3) and Phytophthora Rot (D3).

The co�ee data set is a time series data set used in chemometrics to classifyfood types. This kind of time series is seen in many applications in food safetyand quality insurance. This data set is taken from UCR time Series Classi�ca-tion and Clustering website [10]. Co�ea Arabica and Co�ea Canephora variantRobusta are the two species of co�ee bean which have acquired a worldwideeconomic importance and many methods have been developed to discriminatebetween these two species by chemical analysis [8].

4.2 Experimental setup

In practice, we have tested our method on the di�erent data sets by solving theoptimization problem Pr,x using algorithm 1 and also the proposed Metropolis-Hastings algorithm (algorithm 2). Each time, a distance matrix is required.For the quantitative data, we compute the Euclidean distance between points

yi, for i = 1, . . . , n, by the known formula dij =

√√√√ p∑k=1

(yik − yjk)2. For cate-

gorical data, the distance between two soybean diseases (i, j) is given throughEskin dissimilarity (or proximity) measure [7] computed by the formula pij =

Q∑t=1

wtptij where p

tij =

1n2k

n2k + 2

if it = jt

else, ptij is the per-attribute Eskin dis-

similarity between two values for the categorical attribute indexed by t, wt isthe weight assigned to the attribute t, Q is the number of attributes and nt isthe number of values taken by each attribute. Then, using the formula whichtransforms the dissimilarity into similarity: pij = 1 − sij , the distances canbe given by the standard transformation formula from similarity to distance:dij =

√sii − 2sij + sjj . On top of that, to compute the distances between

the curves of functional data, we have chosen a measure of proximity similarto that studied in [19]. In this article, the authors develop a proper classi�-cation designed to distinguish the grouping structure by using a functionalk-means clustering procedure with three sorts of distances. So, in our workwe choose one of these three proximity measures forasmuch their results aresimilar. Thus, the proximity measure chosen between two curves Fi and Fj is

the following: d0(Fi, Fj) =

√∫T(F 0i (t)− F 0

j (t))2dt. This measure is calculated

using the function metric.lp() of the fda.usc package for the R software.

To solve the di�erent optimization problems, we have used the optimizationtoolbox in MATLAB. For problems Pr and Pr,a, we apply �rstly PCA � forquantitative data � or MDS � for categorical and functional data � and then a

Page 16: Projection under pairwise distance controls

Projection under pairwise distance control 15

linear programming package is used to solve the optimization problems usingan interior-point algorithm. Problems Pr,x and PR,x are non-linear optimiza-tion problems, therefore we use a non-linear programming package to solve itselecting the active-set algorithm to obtain the best values of (x1, . . . , xn) and(r1, . . . , rn). This iterative algorithm is composed of two phases. In the �rstphase (the feasibility phase), the objective function is ignored while a feasiblepoint is found for the constraints, in the second phase (the optimality phase),the objective function is minimized while feasibility is maintained [34].

Our proposed Metropolis-Hastings algorithm can provide a good solution ifparameters λ, σ and T are chosen adequately. For instance, λ should be suchthat the points belonging to unsaturated constraints are chosen with smallprobabilities. Therefore, we take it equal to 100. For the other parameters σand T , we take their values respectively in a range from 0.01 and 100.

As we have mentioned in the section of visualization, the visualization ofthe projection of each point i in R2 is presented as a circle having xi as centerand ri as radius so as the projected point belongs to this circle and this is thespeci�city of our method. For each data set, we show the circles obtained foreach point after resolution of optimization problem Pr,x. To compare theprojection quality of our representation with that obtained by PCAand KPCA, we use the squared cosine values as projection quality.Moreover, the comparison of the projection quality of our represen-tation with MDS is done using Stress-per-point (SPP) proposed byBorn and Groenen in [11]. The SPP of the point i is given by:

SPPi =

∑nj=1,j 6=i(dij − ‖xi − xj‖)2∑n

j=1,j 6=i d2ij

Stress

with Stress =

∑n1≤i<j≤n(dij − ‖xi − xj‖)2∑n

1≤i<j≤n d2ij

Furthermore, the lower bound de�ned in section 3.1 is each time computed.

4.3 Results

4.3.1 Visualization data in R2

The optimization results for these four data sets are given in Table 1. For eachdata, we give the algorithm 1 and Metropolis-Hastings results with whichinitial starting point is used in algorithm 1. The lower bound value for eachdata set is also given in this table, we observe that in one case (cars), thislower bound indicates that the found solution is not far from the optimumbut in the other cases, it seems that the lower bound while providing a goodstarting point can be improved.

Figures 5 and 7 depict the results of projection under pairwisedistance control for quantitative data. This projection is compared

Page 17: Projection under pairwise distance controls

16 Alawieh Hiba, Nicolas Wicker, Christophe Biernacki

Table 1: Optimization solution of problem Pr,x for di�erent data sets.

∑r Algo 1i

∑rMHi Lower Bound

Iris 16.19 17.2 1.07Cars 3.27 3.35 1.21

Soybean 3.98 3.93 0.29Co�ee 21.68 21.97 0.89

with the projection given by PCA, KPCA and MDS. For PCA andKPCA, we have plotted the projection of the points indexed by theirsquared cosine values. For MDS, we have used the smacof packagein R to compute the stress-per-point and to plot the bubble plotrepresented the stress-per-point.In the projection of Iris data set showed in Figure 5, it is interesting to remarkthat appealingly two areas are well separated. This corresponds to the well-known fact that Iris versicolor and virginica are close whereas the species Irissetosa are more distant.Concerning car data set, the projection of points using projection under pair-wise distance control is given in Figure 7. The expensive cars as "Audi 100","Alfetta-1.66", "Dastun-200L", "Renault 30" are well-separated from the low-standard cars as "Lada-1300", "Toyota Corolla", "Citroen GS Club", "Simca1300". We remark that the expensive cars are located on the right and low-standard one are located on the left.For these two data sets, we want to compare the projection qualityfor each method presented in Figures 4 and 6 with the projectionquality obtained using projection under pairwise distance contral inFigures 5 and 7. For PCA, we can say that our method projected thepoints without giving any importance to any group. Indeed, Figure4a depicts a group with small values of quality measure and a groupwith high values of quality measure whereas the radii obtained byprojection under pairwise distance control method are distributed inan equivalent way. Additionally, from Figure 7, we can assert thatthe projected points obtained using projection under pairwise dis-tance control method are well separated as there is any intersectionbetween the circle. For KPCA,we have plotted the squared cosineas a circle to make the representation more clear especially for irisdata set as the Iris setosa specie are projected next to each other.From Figure 4b we can conclude that in each category, the pointswhich have close quality values are located side by side. The sameconclusion is appeared for cars data set in Figure 6b, we can see thatthe point with navy circle are located almost in the same Y axis co-ordinates similarly for the red circles. So the local quality in KPCAdependent on the place on the points. Additionally, by comparingthe projection of our method with those obtained in MDS, we can�nd the same results appear in PCA for Iris data set, the points

Page 18: Projection under pairwise distance controls

Projection under pairwise distance control 17

2 3 4 5 6 7 8 9 10

4

4.5

5

5.5

6

6.5

7

0.997

0.807

0.866

0.9871

0.998

1

1

0.981

0.998

0.997

0.988

0.988

0.995

0.99

0.996

0.971

1

0.995

0.998

0.976

0.997

0.9570.994

0.9860.992

0.982

0.979

0.997

0.9920.984

0.9970.998

0.965

0.995

0.999

0.988

0.985

0.945

0.988

0.994

0.9940.998

0.986

0.985

0.969

0.985

0.996

0.997

0.996

0.995

0.995

0.991

0.862

0.999

0.876

0.871

0.976

0.982

0.708

0.782

1

0.815

0.856

0.945

0.787

0.795

0.360.832

0.9370.982

0.6460.829

0.947

0.782

0.713

0.779

0.749 0.964

0.982

0.915

0.9920.977

0.967

0.971

0.2

0.81

0.897

0.86

0.314

0.995 0.942

0.946

0.966

0.997

0.964

0.354

0.821

0.828

0.971

0.972

0.806

0.991

0.962

0.959

0.959

0.719

0.919

0.966

0.981

0.962

0.999

0.994

0.893

0.684

0.871

0.988

0.993

0.956

0.984

0.983

0.649

0.919

0.99

0.978

0.956

0.978

0.902

0.977

0.872

0.918

0.976

0.962

0.951

0.888

0.957

0.803

0.962

0.83

0.99

0.949

0.946

0.866

0.9770.92

0.944

0.985

0.9810.771

0.75

Se

VeCo

ViCa

(a)

-0.6 -0.4 -0.2 0 0.2 0.4 0.6 0.8 1-0.8

-0.6

-0.4

-0.2

0

0.2

0.4

0.6

0.8

VeCo

ViCa

VeCo

ViCa ViCa

ViCaViCa

VeCo

Se

VeCo

VeCo

ViCa

VeCo

Se

ViCa

Se

VeCo

ViCa

Se

ViCa

Se

ViCa

SeSe

VeCo

ViCa

Se

VeCo

ViCa

VeCo

ViCa

Se

VeCoVeCo

SeSeSe

ViCa

Se

ViCaViCa

VeCo

VeCo

Se Se

VeCo

Se

VeCo

ViCa

VeCo

VeCo

ViCa

SeSe

ViCa

Se

ViCa

VeCo

SeSeSeSeSe

ViCaViCa

VeCo

VeCo

ViCa

SeSe

ViCa

ViCa

VeCo

SeSeSe

ViCa

Se

VeCo

VeCo

VeCo

Se

VeCo

ViCa

Se

ViCa

VeCo

VeCoVeCo

SeSe

ViCa

Se

VeCo

VeCo

SeViCa

ViCa

VeCo

VeCoVeCo

Se

VeCo

Se

ViCa

VeCo

VeCo

VeCo

VeCo

ViCaViCa

VeCo

ViCa

VeCo

ViCa

ViCaVeCo

ViCa

ViCaVeCo

VeCo

ViCaSe

ViCa

VeCoVeCo

Se SeSeSe

ViCa

SeSe

ViCa

Se

ViCa

SeSe

ViCa

Se

VeCoVeCo VeCoVeCo

Se

ViCa

ViCa

ViCaViCaViCa

Kernel PCA

(b)

(c)

Fig. 4: Projection of Iris data set using PCA, KPCA and MDS respectively.

Page 19: Projection under pairwise distance controls

18 Alawieh Hiba, Nicolas Wicker, Christophe Biernacki

Fig. 5: Projection of Iris data set using projection under pairwise distancecontrol methods respectively. Two well separated groups can be observed.

are projected by giving an importance to Iris setosa group. Indeed,almost all the red circles (indicating the very well projection) aregiven to the Iris setosa specie.

Moreover, the pairwise distances are signi�cant in our method and give aninterpretation on the position between points whereas the distances betweenthe projected points using PCA, KPCA and MDS are not interpretable as thecosine values and the Stress-per-point can not be interpreted as distances inPCA and KPCA and MDS. This is the particular strength of our method.Hence, projection under pairwise distance control suggests an absolute inter-pretation whereas the other methods give a relative one. From this, we canconclude from Figure 7 that there is a big di�erence between the two cars"Toyota" and "Renault 3" as the distances between this two cars is very im-portant. Conversely, the distance between "Lada1300" and "Citroen" cars aresmall indicating then the closeness of these two cars. Note here that these twocars are very well projected leading to a very good interpretation.

For the qualitative and functional data sets, it is necessary to verify thatthe matrix B obtained by MDS method is semi-de�nite positive to use thesquared cosine as quality measure because the starting point of optimizationis obtained from MDS. After that, in case of positiveness of matrix B, wecan calculate the quality measure. In the projection of the soybean data set,four classes have been shown in Figure 8 and each class contains the diseasesnumber of the class. But basically, the whole set of points can be divided in

Page 20: Projection under pairwise distance controls

Projection under pairwise distance control 19

-4 -3 -2 -1 0 1 2 3 4 5-4

-3

-2

-1

0

1

2

3

4

5

ALFASUD-T1-1350(0.944)

AUDI(0.714)

SIMCS(0.791)

CITROEN(0.979)

FIAT(0.571)

LANCA-BETA(0.116)

PEUGOT(0.884)RENAULT-16-TL(0.845)

RENAULT-30-TS(0.944)

TOYOTA(0.978)

ALFETTA(0.864)

PRINCESS(0.894)

DATSUN(0.806)

TAUNUS(0.801)

RANCHO(0.653)

MAZDA(0.402)

OPEL-REKORD(0.864)

LADA-1300(0.929)

(a)

−0.6 −0.4 −0.2 0 0.2 0.4 0.6 0.8−0.8

−0.6

−0.4

−0.2

0

0.2

0.4

0.6

ALFETTA

RENAULT−30−TSALFASUD−T1−1350

DATSUNAUDI

FIAT

OPEL−REKORD

PEUGOT

PRINCESS

TAUNUS

MAZDA

RENAULT−16−TL

LADA−1300

TOYOTA

CITROEN

LANCA−BETA

SIMCS

RANCHO

Projection using Kernel PCA

(b)

−1 −0.8 −0.6 −0.4 −0.2 0 0.2 0.4 0.6 0.8 1−1

−0.5

0

0.5

1

1.5

RENAULT−16−TL

RENAULT−30−TS

OPEL−REKORD

AUDI

CITROEN

TAUNUS

PRINCESSALFASUD−T1−1350

MAZDA

PEUGOT

FIAT

LANCA−BETATOYOTA

ALFETTA

DATSUN

RANCHO

SIMCS

Projection using MDS

(c)

Fig. 6: Projection of car data set using PCA, KPCA and MDS. For PCA, thevalues of the quality are given between parentheses near each car.

Page 21: Projection under pairwise distance controls

20 Alawieh Hiba, Nicolas Wicker, Christophe Biernacki

−4 −3 −2 −1 0 1 2 3 4 5−2.5

−2

−1.5

−1

−0.5

0

0.5

1

1.5

2

2.5

DATSUN

AUDI

LANCA−BETA

OPEL−REKORD

RENAULT−16−TL

RANCHO

TAUNUSMAZDA

FIAT

SIMCS

LADA−1300

ALFASUD−T1−1350

PEUGOT

ALFETTA

RENAULT−30−TS

CITROENTOYOTA

PRINCESS

Projection using projection under pairwise distance control method

Fig. 7: Projection of car data set using projection under pairwise distancecontrol.

two large classes. Indeed, It is clear that class 2 is well separated from theothers classes as there is no intersection between the circles of class 2 and thecircles of others classes. Moreover, class 1 can be considered as well separatedclass from classes 3 and 4 if we do not take into account the point D∗3 . Classes3 and 4 are not at all well separated as we can exhibit that there are di�erentintersections between the circles of these two classes. This result is �gured in[30] which lists the value "normal" for the �rst two classes and "irrelevant"for the later two classes. The comparison of projection under pairwise distancecontrol result with PCA and KPCA is not possible for this data set becausethe matrix B is not semi-de�nite positive.

By comparing our projection with the projection obtained byMDS, we can see that the 4 groups of soybean are well separatedusing MDS than our method and the bubble are small than our cir-cles. So we can conclude that our method detect more the misplacedpoints in the groups and the points which are not well projected.

The co�ee data set has been studied in several articles ([8,3]) and di�erentclassi�cation methods have shown the di�erent groups contained in this dataset using our method and the other methods. We can see clearly in Figures 10and 11 the grouping structure that is obtained.

In Figure 11, we show that we have succeeded in di�erentiating the Arabicafrom Robusta co�ee. These two classes are clearly presented, the �rst class in-dexed by 1 corresponding to Arabica co�ee and the second one indexed by 2

Page 22: Projection under pairwise distance controls

Projection under pairwise distance control 21

−0.5 −0.4 −0.3 −0.2 −0.1 0 0.1 0.2 0.3 0.4 0.5−0.5

−0.4

−0.3

−0.2

−0.1

0

0.1

0.2

0.3

0.4

D3

D4

D2

D3

D3

D4

D3

D1

D4

D4

D4

D1

D2

D1D1

D3

D3

D1

D1

D2

D4

D2

D1D1

D2

D3

D2

D4

D4

D4

D2

D2

D3

D4

D1

D3

D1

D3

D4

D2

D4

D2

D4

D4

D4D4D4

Projection using projection under pairwise distance control method

Fig. 8: Projection under pairwise distance control for soybean dat set. Fourgroups are presented, indexed by D1, D2, D3 and D4.

corresponding to Robusta co�ee. These classes are not well separated by com-paring with the results of quantitative data, since there are many intersections.Therefore, the representation of the points as circles and not as coordinatespoints gives more information about the real class of points and shows thepoints who have the possibility to be misplaced in a class.

Figure 10a and 10b show the projection quality using PCA and MDSrespectively. As all the eigenvalues of matrix B are positive, so wecan compute the quality measure given by PCA. Comparing the pro-jection quality of PCA and projection under pairwise distance con-trol given respectively by Figures 10a and 11, we can observe thatthe quality of projection of the set of points is pretty steady.

The comparison with KPCA projection is not allowed because allthe local quality projection obtained using KPCA is equal to 1.

Additionally, Metropolis-Hastings has been applied to these data sets. Thetrace plots of the optimization problem Pr,x are shown in Figure 12 after

Page 23: Projection under pairwise distance controls

22 Alawieh Hiba, Nicolas Wicker, Christophe Biernacki

−0.8 −0.6 −0.4 −0.2 0 0.2 0.4 0.6 0.8 1−0.8

−0.6

−0.4

−0.2

0

0.2

0.4

0.6

0.8

D2

D1

D3

D4

D4

D3

D3

D3

D3

D2

D4

D1

D2

D2

D2

D3

D1

D4

D1D3

D3

D1

D2

D1

D4

D3

D4

D4

D1

D4

D2D4

D1

D1

D3

D2

D4

D2

D2

D4

D4

D4

D4

D4

D4

D1

Projection using MDS

Fig. 9: MDS for soybean dat set. Four groups are presented, indexed by D1,D2, D3 and D4.

5000 iterations. Returning to Table 1, we can exhibit that Metropolis-Hastingsalgorithm solutions are very close to those obtained using the optimizationpackage of Matlab and reciprocally. Thus, the obtained radii should be closeto the optimum.Finally, we present the lower bound computed from the three functions de-scribed in Section 3. The lower bound is taken by the minimal intersection ofthese functions. Returning to Table 1, it is clear that the value of the boundis small compared to the value of the solution obtained by algorithm 1 andMetropolis-Hastings. Thus, this bound while providing a good starting point,should be improved Note that this bound for tetrahedron example gives alsogood results as algorithm 1 provides a solution three times smaller than thebound.

4.3.2 Dimensionality reduction results

One of high-dimensional data studies objectives is to choose from a large num-ber of variables those which are important for understanding the underlyingphenomena of study. So, the aim will be to reduce the dimension rather thanto visualize data in R2. So, our method can also serve to reduce the numberof variables by taking into account the minimal value of

∑ni=1 ri.

Here, we have solved the problem Pr,x using the di�erent possible dimensionvalues. We have plotted in Figure 13 the values of

∑ni=1 ri as a guide for

choosing the reduced number of variables. This �gure shows the values of∑ni=1 ri for the di�erent data sets using di�erent dimensions. It is clear to see

that the value of∑ni=1 ri decreases when the dimension increases.

Page 24: Projection under pairwise distance controls

Projection under pairwise distance control 23

-4 -3 -2 -1 0 1 2 3

-4

-3

-2

-1

0

1

2

3

0.38391

0.74248

0.94061

0.80253

0.88018

0.78104

0.43144

0.83067

0.99892

0.57037

0.99071

0.18942

0.113880.74653

0.813770.99906

0.242070.80966

0.26657

0.62758

0.90738

0.90807

0.625580.99823

0.44215

0.5948

0.85415

0.90669

0.59223

0.52788

0.59057

0.97906

0.14908

0.68023

0.66059

0.96702

0.87681

0.95121

0.98311

0.49414

0.31723

0.40596

0.97095

0.55609

0.941170.97962

0.985150.78354

0.99982

0.75495

0.393450.93179

0.96395

0.81388

0.99172

0.74377

Class 1

Class 2

(a)

−1 −0.5 0 0.5 1 1.5−1.5

−1

−0.5

0

0.5

1

2

1

2

1

1

2

2

1

1

1

1

2

2

1

1

2

2

1

1

11

1

2

1

1

2

1

2

1

1

1

1

1

2

1

212

2

1

2

222

1

1

2

2

1

2

2

22

2

1

Projection using MDS

(b)

Fig. 10: Projection of co�ee data set using PCA and MDS.

Page 25: Projection under pairwise distance controls

24 Alawieh Hiba, Nicolas Wicker, Christophe Biernacki

−5 −4 −3 −2 −1 0 1 2 3 4−2.5

−2

−1.5

−1

−0.5

0

0.5

1

1.5

2

2.5

2

2

2

1

1 1

12

2

1

1

2

1

1

1

1

1

1

1

1

1

1

2

2

1

2

1

2 2

1

1

2

1

1

2

2

2

2

2

2

2 1

2

2

1

2

2

11

2

1

2

2

2

1

1

Projection using projection under pairwise distance control method

Fig. 11: Projection of co�ee data set using projection under pairwise distancecontrol. Two clusters indexed 1 and 2 indicate respectively Arabica and Ro-busta classes.

The main problem which is widely posed in dimension reduction methods isthe determination of the number of components that are needed to be retained.Many methods have been discussed in the literature [24,6] to determine thedimension of reduced space relying on di�erent strategies related to the goodexplanation or the good prediction. So, with our method the choice of thereduced space dimension is related to the locally projection quality of pointsand how much the user is interested by the projection quality of points.

Concerning the quantitative data sets (Iris and car), if the main objective ofthe user is to obtain a very good projection quality then a choice of threecomponents against 4 for iris and 6 for cars can be a good choice as thevalue of

∑ni=1 ri is small and there is not a big di�erence between this value

for this dimension and the values of the higher dimensions. For co�ee dataset, a dimensionality reduction from 56 sample time series down to 6 simpleextracted features is considered as a good choice. The same idea can be seenfor soybean data set, a reduced space dimension equal to 4 can be consideredas e�cient reduced space.

Moreover, a comparison of our results with the existent results shows a coher-ence between them. For Iris data set, [15] and [26] conclude that the number ofvariables can be reduced to 2 as the petal length and petal width variables arethe most important variables among all variables. Similarly, this result can beseen for car data set, Saporta in his book [29] (Table 7.4.1 page 178) notices

Page 26: Projection under pairwise distance controls

Projection under pairwise distance control 25

0 1000 2000 3000 4000 500017

18

19

20

21

22

23

24

25

26

(a) Iris data set

0 1000 2000 3000 4000 50003

3.5

4

4.5

5

5.5

6

6.5

(b) car data set

0 1000 2000 3000 4000 500021

22

23

24

25

26

27

28

29

30

31

(c) Co�ee data set

0 1000 2000 3000 4000 5000

3.5

4

4.5

5

5.5

6

6.5

(d) Soybean data set

Fig. 12: Trace plots of Metropolis Hastings for di�erent data sets. The x-axiscorresponds to the iteration number and y-axis to the value of

∑ni=1 ri.

that the conservation of two dimensions leads to the explanation of 88% of in-ertia where the term inertia re�ects the importance of a component.So, these results seem very similar to our results, the important decrease islocated between dimensions 1 and 2. The other decreases are negligible forthese two data sets. Selection variables is studied on time series co�ee dataset in [2]. Using several analysis methods, the number of selected variablesranges between 2 and 13. This result is also seen using our method, a numberof reduced variables taken between 2 and 9 gives a good quality projection ofthe points. Concerning soybean data set, Dela Cruz shows in his paper [16]that the 35 attributes can be reduced to 15 and here with our method, wehave succeeded to reduce the attributes to 6 by having a very good projection.Hence, the presented results con�rm that we can reduce the dimension non-linearly and still keep a way of assessing as reasonable number of dimensions.and that is e�cient as a dimensionality reduction method.

4.4 Advantages of projection under pairwise distance control method

As we have seen, our presented method has several advantages. To summarize:

Page 27: Projection under pairwise distance controls

26 Alawieh Hiba, Nicolas Wicker, Christophe Biernacki

Dimensions

1 2 3 4

Sum

of ra

dii

0

5

10

15

20

25

30

35

40

45

47.73

16.57

3.987

1.0069e-13

Iris data set

Dimensions

1 2 3 4 5 6

Su

m o

f ra

dii

0

2

4

6

8

10

12 11.5371

3.2886

1.4494

0.534810.1783 3.6175e-15

Voitures data set

Dimensions

1 2 3 4 5 6 7 8 9 10

Sum

of ra

dii

0

5

10

15

20

25

30

35

40

45

46.2756

21.6818

13.219

8.3141

5.55713.8738

2.9723 2.3825 1.9985

0

Coffee data set

Dimensions

1 2 3 4 5 6 7 8 9 10

Sum

of ra

dii

0

1

2

3

4

5

6

7

6.3655

3.9825

2.6753

1.8578

1.3762

0.95801

0

Soybean data set

Fig. 13: The scree plot of∑ni=1 ri for di�erent dimensions for the four data

sets.

�rstly, it is a non-linear projection method which takes into account the pro-jection quality of each point individually. Secondly, the distances between pro-jected points are related to the initial distances between points o�ering a wayto interpret easily the distances observed in the projection plane. Thirdly, thequality distribution between the points seems to be evenly distributed.

5 Conclusion

The purpose of this article was to outline a new non-linear projection methodbased on a new local measure of projection quality. Of course, in some pro-jection methods a local measure is given but this measure cannot be appliedunless in cases of linear projections, and even then it is not suitable for graph-ical representation.

The quality of projection is given here by additional variables called radii,which enable to give a bound on the original distances. We have shown thatthe idea can be written as an optimization problem in order to minimize thesum of the radii under some constraints. As the solution of this problem cannotbe obtained exactly, we have developed di�erent algorithms and proposed alower bound for the objective function. As such, the method described here

Page 28: Projection under pairwise distance controls

Projection under pairwise distance control 27

needs further research to improve the lower bound in order to assess how closethe algorithms are from the minimum.

References

1. Anderson E, The Irises of the Gaspé Peninsula, Bull. Am. Iris Soc., 59, 2�5 (1935)2. Andrews, J. L. and McNicholas, P.D., Variable Selection for Clustering and Classi�cation,Journal of Classi�cation, 31, 136-153 (2014)

3. Bagnall, A., Davis, L., Hills, J., and Lines, J., Transformation Based Ensembles for TimeSeries Classi�cation, Proceedings of the 12th SIAM International Conference on DataMining, 307�319 (2012)

4. Berge, C., Frolo�, N., Kalathur, RK., Maumy, M., Poch, O., Ra�elsberger, W. andWicker, N., Multidimensional �tting for multivariate data analysis. Journal of Com-putational Biology, 17, 723�732 (2010)

5. Xiaofang, L., Chun, Y., Greedy kernel PCA for training data reduction and nonlinearfeature extraction in classi�cation. Proceedings of SPIE - The International Society forOptical Engineering, 2009.

6. Besse, P., PCA stability and choice of dimensionality, Statistics & Probability Letters,13, 405-410, (1992)

7. Boriah, S., Chandola, V., and Kumar, V., Similarity Measures for Categorical Data:A Comparative Evaluation, Proceedings of the SIAM International Conference on DataMining (2008)

8. Briandet, R., Kemsley, E. K., and Wilson, R. H., Discrimination of arabica and robustain instant co�ee by fourier transform infrared spectroscopy and chemometrics, J. Agric.Food Chem., 44 (1), 170�174 (1996)

9. Chan, W. W-Y., A survey on multivariate data visualization in Science and technology,Department of Computer Science and Engineering Hong Kong, University of Science andTechnology, 8(6), 1�29 (2006)

10. Chen, Y., Keogh, E., Hu, B., Begum, N., Bagnall, A., Mueen, A. and Batista, G., TheUCR Time Series Classi�cation Archive, www.cs.ucr.edu/~eamonn/time_series_data/

(2015).11. Borg, I., Groenen, P. Modern Multidimensional Scaling: Theory and Applications (2nded.). New York: Springer-Verlag (2005).

12. Cheung, L.W., Classi�cation approaches for microarray gene expression data analysis,Methods Mol Biol., 802, 73�85 (2012)

13. Chinchilli, V.M., Sen, P.K., Multivariate Data Analysis: Its Methods, Chemometricsand Intelligent Laboratory Systems, 2, 29�36 (1987)

14. Cleveland, W.S., McGill, M.E., Dynamic Graphics for Statistics, Wadsworth andBrooks/Cole, Paci�c Grove, CA, (1988)

15. Chiu, S. L., Method and Software for Extracting Fuzzy Classi�cation Rules by Subtrac-tive Clustering, In Proceedings of North American Fuzzy Information Processing SocietyConference (1996)

16. Dela Cruz, G. B., Comparative Study of Data Mining Classi�cation Techniques overSoybean Disease by Implementing PCA-GA, International Journal of Engineering Re-search and General Science, 3(5), 6�11 (2015)

17. Dempster, A.P., An overview of multivariate data analysis, Journal of MultivariateAnalysis, 1(3), 316�346, (1971)

18. Golub, T. R., Slonim, D. K., Tamayo, P., Huard, C., Gaasenbeek, M., Mesirov, J. P.,Coller, H., Loh, M. L., Downing, J. R., Caligiuri, M. A., Bloom�eld, C. D. and Lan-der, E. S., Molecular classi�cation of cancer: class discovery and class prediction by geneexpression monitoring, Science, 286, 531�537 (1999)

19. Ieva, F., Paganoni, A.M., Pigoli, D., and Vitelli., V., Multivariate functional cluster-ing for the analysis of ECG curves morphology, Journal of the Royal Statistical Society,Applied Statistics, series C., 62(3), 401�418 (2012)

20. Inselberg, A., The Plane with Parallel Coordinates, Special Issue on ComputationalGeometry, The Visual Computer, 1, 69�91 (1985)

Page 29: Projection under pairwise distance controls

28 Alawieh Hiba, Nicolas Wicker, Christophe Biernacki

21. Jackson, J., A Users Guide to Principal Components, John Wiley & Sons, New York(1991)

22. Jagannathan, R. and Ma, T., Risk reduction in large portfolios: why imposing the wrongconstraints helps, J. Finan., 58, 1651�1683 (2003)

23. Johansen, A. M. and Evers, L. Monte Carlo Methods, Department of Mathematics,University of Bristol (2007)

24. Jollife, I.T., Principal Component Analysis, Springer, New York (1986)25. Keim, D. A. and Kriegel, H. P., Visualization Techniques for Mining Large Databases:A Comparison, IEEE Transactions on Knowledge and Data Engineering, 8(6), 923-938(1996)

26. Liu, H., Setiono, R., Chi2: feature selection and discretization of numeric attributes,In Proceedings., Seventh International Conference on Tools with Arti�cial Intelligence(TAT'95) (1995)

27. Mardia, K.V., Kent, J.T. and Bibby, J.M., Multivariate analysis, Academic Press, Lon-don (1979)

28. Sammon, J., A non-linear mapping for data structure analysis, IEEE Transactions onComputers, 18(5), 401�409 (1969)

29. Saporta, G., Probabilités, analyse des données et statistique, Technip (2006)30. Stepp, R., Conjunctive conceptual clustering, Doctoral dissertation, department of com-puter science, university of Illinois, Urbana-Champaign, IL (1984)

31. Svante W., C. Albano, W. J. DunnIII, U. Edlund, K. Esbensen, P. Geladi, S. Hellberg,E. Johansson, W. Lindberg , M. Sjostrom., Multivariate Data Analysis in Chemistry,Chemometrics, 138, 17�95 (1984)

32. Togerson, W.S., Theory and methods of scaling, New York: Wiley (1958)33. Van der Hilst, R., de Hoop, M., Wang, P., Shim, S.-H., Ma, P. and Tenorio, L., Seismo-stratigraphy and thermal structure of earth's core-mantle boundary region, Science, 315,1813�1817 (2007)

34. Wong, E., Active-Set Methods for Quadratic Programming, Ph.D. thesis, university ofCalifornia, San Diego (2011)

Page 30: Projection under pairwise distance controls

29

Appendices

Appendix A: Two lemmas

Let (C) be a circle with center O and radius r. Let consider n points with coordinatesx1, · · · , xn such that for all i = 1, . . . , n, ‖xi − O‖ ≤ r and having g as center of gravity.This hypothesis is used in Lemma 1 and Corollary 1.

Lemma 1 For all points x1, . . . , xn, we have:

‖xi −O‖ = r when

n∑i=1

‖xi − g‖2 is maximum .

Proof Let A and B two points among the n points such that A is inside the circle (C) andB belonging (C). The point A and B have as coordinates (a1, a2) and (r, 0) respectively.We want to show that by moving B by small movements along the circle, we can approachthe point A to the circle border increasing thus the inertia

∑ni=1 ‖xi − g‖2.

We note B′, A′ the new positions after movements of B and A respectively. Let θ be theangle between (OB) and (OB′), uθ the displacement of B and (a′1, a

′2) the coordinate of

point A′.Approaching A to the circle border requires the opposite movements of A and B with equallength. This constraint is necessary to keep the center in the same position. We distinguishtwo cases:

1- A having a2 < 0.2- A having a2 > 0.

The two cases are illustrated in Figure 14.

O B

B′

θu+θ

A

A′

O B

B′

θ

u−θ

A

A′

case 1 case 2

Fig. 14: Representation of movements of points A and B in cases 1 and 2.

For case 1, A in the lower half of the circle requires that B moves positively with angle

θ ∈ [0; π2]. In this case, the vector uθ is given by: u+θ =

(cos θ − 1sin θ

)and the inner product

〈u+θ , AA′〉 is given by:

〈u+θ , AA′〉 = (a′1 − a1)(cos θ − 1) + (a′2 − a2) sin θ· (3)

Here, we have a′1 ≥ a1 and a′2 ≤ a2 that imply a′1 − a1 ≥ 0 and a′2 − a2 ≤ 0. Moreover, we

have 0 ≤ cos θ ≤ 1 and sin θ ≥ 0 that give 〈u+θ , AA′〉 < 0.

Page 31: Projection under pairwise distance controls

30

For case 2, A in the upper half of the circle requires that B moves negatively with

angle θ ∈ [3π2, 2π]. In this case, the vector uθ is given by: u−θ =

(cos θ − 1− sin θ

)and the inner

product is given by:

〈u−θ , AA′〉 = (a′1 − a1)(cos θ − 1)− (a′2 − a2) sin θ (4)

Here, we have a′1 ≥ a1 and a′2 ≤ a2 that imply a′1 − a1 ≥ 0 and a′2 − a2 ≤ 0. Moreover, we

have 0 ≤ cos θ ≤ 1 and sin θ ≤ 0 that give 〈u−θ , AA′〉 < 0.

Corollary 1 The center of gravity g of x1, . . . , xn is the center of circle (C) i.e. O = gwhen

∑ni=1 ‖xi − g‖2 is maximum.

Proof We have:

n∑i=1

‖xi − g‖2 =

n∑i=1

‖xi −O +O − g‖2

=

n∑i=1

‖xi −O‖2 +

n∑i=1

‖O − g‖2 + 2

n∑i=1

(xi −O)′(O − g)

=

n∑i=1

‖xi −O‖2 +

n∑i=1

‖O − g‖2 + 2n(g −O)′(O − g)

=n∑i=1

‖xi −O‖2 + n‖O − g‖2 − 2n‖O − g‖2

=n∑i=1

‖xi −O‖2 − n‖O − g‖2

All the points belong to the circle (C) as a result of Lemma 1. So, ‖xi − O‖2 is �x and

equal to r2. Thus, maximizing inertian∑i=1

‖xi − g‖2 amounts to minimize ‖O − g‖2. Then,

the minimum of ‖O − g‖2 is zero so that O = g.

Lemma 2 If∑ni=1 ‖ xi−g ‖2 is maximum for points x1, . . . , xn under constraints, for all

couple (i, j), ‖xi − xj‖ ≤M , then an upper bound of∑ni=1 ‖ xi − g ‖2 is given by:

nM2

Proof Let (C) be the smallest circle containing the n points with coordinates x1, . . . , xn. Weconsider three points noted A,B and C among the n points and having xA, xB and xC ascoordinates . We suppose that A, B and C belong to the circle (C) and the distance betweenB and C is equal to M i .e. ‖xB − xC‖ =M . By hypothesis, we have ‖xA − xB‖ ≤M . Wenote θ the angle between (AB) and (BC). Figure 15 illustrates the situation.

If θ >π

3then ‖xA − xB‖ ≤ M and ‖xA − xC‖ > M , by reversing the role of xB and xC

we get θ =π

3. So, A = A′ (A′ is at the position indicated in �gure 15) and then r =

M√3.

Now, if we consider the n points y1, . . . , yn and using Lemma 1 and Corollary 1, the maxi-mum of the inertia

∑ni=1 ‖ yi − g ‖2, under the constraint that y1, . . . , yn are inside (C) is

equal to nr2 = nM2

3. Thus, the maximum of

n∑i=1

‖ xi − g ‖2 is upper bounded by nM2

3.

Page 32: Projection under pairwise distance controls

31

g

BC

AA′

θr

M2

M2 ••

Fig. 15: Representation of the points on the circle.

Appendix B: Three functions used to compute the lower bound

of∑n

i=1 ri

Recall optimization problem Pr,x:

minr1,...,rn∈R+,x1,...,xn∈Rk

n∑i=1

ri

s.t dij− ‖ xi − xj ‖≤ ri + rj

‖ xi − xj ‖ −dij ≤ ri + rj

Appendix B.1: Function f(M)

Using the �rst constraint, we have:

dij ≤ ‖ xi − xj ‖ +ri + rj

d2ij ≤ (‖ xi − xj ‖ +ri + rj)2∑

i<j

d2ij ≤∑i<j

‖ xi − xj ‖2 +∑i<j

(ri + rj)2 + 2

∑i<j

(‖ xi − xj ‖)(ri + rj)

∑i<j

d2ij ≤∑i<j

‖ xi − xj ‖2 +∑i<j

(ri + rj)2 + 2M

∑i<j

(ri + rj) as ‖ xi − xj ‖≤M · (5)

Let g be the center of gravity of the projected points x1, . . . , xn, so:

‖ xi − xj ‖ = ‖ xi − g − xj + g ‖‖ xi − xj ‖2 = ‖ xi − g ‖2 + ‖ xj − g ‖2 +2(xi − g)′(xj − g)∑

i<j

‖ xi − xj ‖2 =∑i<j

(‖ xi − g ‖2 + ‖ xj − g ‖2

)+ 2

∑i<j

(xi − g)′(xj − g)·

Page 33: Projection under pairwise distance controls

32

As∑i<j

(xi − g)′(xj − g) = 0 then:

∑i<j

‖ xi − xj ‖2 = (n− 1)

n∑i=1

‖ xi − g ‖2 · (6)

Replacing equation (6) in (5) gives :

(n− 1)n∑i=1

‖ xi − g ‖2 +∑i<j

(ri + rj)2 + 2M

∑i<j

(ri + rj)−∑i<j

d2ij ≥ 0· (7)

The quantity∑ni=1 ‖ xi − g ‖2 is the inertia of the projected points x1, . . . , xn. As long as

we want to conserve the initial information, the inertia must be maximal under constraints‖xi − xj‖ ≤M for all 1 ≤ i < j ≤ n.Recalling equation (7) and by using Lemma 2, we obtain:

n(n− 1)

3M2 +

∑i<j

(ri + rj)2 + 2M

∑i<j

(ri + rj)−∑i<j

d2ij ≥ 0

n(n− 1)

3M2 + (n− 1)

(n∑i=1

ri

)2

+ 2(n− 1)M

n∑i=1

ri −∑i<j

d2ij ≥ 0

(n∑i=1

ri

)2

+ 2M

(n∑i=1

ri

)+n

3M2 −

1

n− 1

∑i<j

d2ij ≥ 0· (8)

The discriminant of equation (8) is given by: ∆ = 4(1−

n

3

)M2 +

4

n− 1

∑i<j

d2ij and as

ri ≥ 0, ∀i = 1, . . . , n we get:

n∑i=1

ri ≥

√√√√(1− n

3

)M2 +

1

n− 1

∑i<j

d2ij −M ·

We note f(M) =

√(1−

n

3

)M2 +

1

n− 1

∑i<j

d2ij −M ·

Appendix B.2: Function g(M)

Two situations are possible:

1. ∃(i′, j′) such that ‖ xi′ − xj′ ‖=M , that gives:

ri′ + rj′ ≥‖ xi′ − xj′ ‖ −di′j′ ≥M − di′j′ ≥M − dmax·

Asn∑i=1

ri ≥ ri′ + rj′ , we obtain:

n∑i=1

ri ≥M − dmax

2. ∃(i∗, j∗) such that di∗j∗ = dmax, that gives:

ri∗ + rj∗ ≥ di∗j∗− ‖ xi∗ − xj∗ ‖≥ dmax −M ·

Then, we obtain:n∑i=1

ri ≥ dmax −M ·

Page 34: Projection under pairwise distance controls

33

Hence:

n∑i=1

ri ≥ |M − dmax|· (9)

We note g(M) = |M − dmax|.

Appendix B.3: Function h(M)

Let us consider four distinct points i, j, k and l. We suppose that there is a couple (i, j) suchthat ‖xi − xj‖ = M and one of their coordinates is equal to zero (xi = 0 or xj = 0). Wedistinguish two cases:

1. xi = 0.2. xj = 0.

Case 1: For xi = 0, we take xj = αxk + βxl with α, β ∈ [0, 1]. The constraints related tothese four points are the following:

|‖xj‖ − dij | ≤ ri + rj (C1)

|‖xk‖ − dik| ≤ ri + rk (C2)

|‖xl‖ − dil| ≤ ri + rl (C3)

|‖xj − xk‖ − djk| ≤ rj + rk (C4)

|‖xj − xl‖ − djl| ≤ rj + rl (C5)

|‖xk − xl‖ − dkl| ≤ rk + rl (C6)

Firstly, using constraints (C4) and (C6) we obtain:

2

n∑t=1

rt ≥ djk − dkl + ‖xk − xl‖ − ‖xj − xk‖·

Additionally, as xj = αxk + βxl with α, β ∈ [0, 1], then

‖xk − xj‖ = ‖xk − αxk − βxl‖ = ‖(1− α)xk − βxl‖ ≤ ‖xk − xl‖,

which gives

n∑t=1

rt ≥djk − dkl

2· (10)

Secondly, using constraints (C5) and (C6) we obtain:

2n∑t=1

rt ≥ djl − dkl + ‖xk − xl‖ − ‖xj − xl‖·

As ‖xl − xj‖ ≤ ‖xk − xl‖ we obtain:n∑t=1

rt ≥djl − dkl

2· (11)

Thirdly, constraint (C1) and initial hypothesis ‖xi − xj‖ =M lead to:

n∑t=1

rt ≥ |dij −M |· (12)

Consequently, equations (10), (11) and (12) involve:

n∑t=1

rt ≥ max

{djk − dkl

2;djl − dkl

2; |dij −M |

}denoted L1

ijkl

Page 35: Projection under pairwise distance controls

34

Case 2: For xj = 0, we take xi = αxk + βxl with α, β ∈ [0, 1]. By analogy with case 1, weobtain:

n∑t=1

rt ≥ max

{dik − dkl

2;dil − dkl

2; |dij −M |

}denoted L2

ijkl·

Due to the choice of one case among cases 1 and 2, we take the minimum of L1ijkl and L

2ijkl.

Thus:n∑t=1

rt ≥ min{L1ijkl;L

2ijkl

Moreover, for a given i, j, this inequality is veri�ed. So that:

n∑t=1

rt ≥ mini<j

maxk<l;k,l6=i,j

min{L1ijkl;L

2ijkl

We note h(M) = mini<j

maxk<l;k,l6=i,j

min{L1ijkl;L

2ijkl