functional correspondence by matrix completion · functional maps let x and y denote two manifolds...

1
Functional correspondence by matrix completion A. Kovnatsky 1 , M. M. Bronstein 1 , Xavier Bresson 2 , and Pierre Vandergheynst 2 1 Faculty of Informatics, Università della Svizzera italiana (USI). 2 Signal Processing Lab, École Polytechnique Fédérale de Lausanne (EPFL). We deal with the problem of finding dense intrinsic correspondence between manifolds in the functional formulation ([4], Fig. 1). We propose treating functional correspondence as geometric matrix completion. We show that our method compares favorably to state-of-the-art methods for non-rigid shape correspondence on the challenging Princeton benchmark [2]. The advantage of our method is especially pronounced in the scarce data set- tings. The proposed method is completely generic and can be applied to any manifolds and high-dimensional geometric data. Functional maps Let X and Y denote two manifolds sampled at n and m points, respectively. Ovsjanikov et al. [4] propose to model the functional correspondence between the spaces L 2 (X ) and L 2 ( Y ) as the m × n matrix T 0 , which maps a function f L 2 (X ) into T 0 f = g L 2 ( Y ). Traditional point-wise correspondence is a particular case of this model wherein T 0 maps delta functions into delta functions. n m t i 00 t i 0 t i x i x i 0 x i 00 y j y j 0 y j 00 x i 00 Figure 1: Illustration of functional correspondence as matrix completion. Column t i = Tδ i of matrix T is the functional map of a delta at x i . Geo- metric structure is imposed on the columns to ensure smoothness, i.e. t i , t i 0 corresponding to spatially close points x i , x i 0 on X are similar. Functional maps as matrix completion Kalofolias et al. [1] studied the problem of matrix completion on graphs, where the rows and columns of the matrix representing the correspondence have underlying geometric struc- ture. They show that adding geometric structure to the standard matrix com- pletion problem improves recovery results. We use the same philosophy to formulate the problem of finding a func- tional map as matrix completion, whose rows and columns are interpreted as functions on the respective manifolds X and Y . For this purpose, we consider the matrix T as a collection of columns T =(t 1 ,..., t n ) or rows T =(t 1> ,..., t m> ) > , where t i and t j denote the ith column and jth row of T, respectively. The column t i = Tδ i is the function on Y corresponding to a delta located at point x i on X . Similarly, the row t j = δ > j T is the function on X corresponding to a delta located at point y j on Y . As in the classical matrix completion problem, we aim at recovering the unknown correspondence matrix T 0 from a few observations of the form T 0 F = G, where F =(f 1 ,..., f q ) and G =(g 1 ,..., g q ) are q corresponding functions on X , Y , respectively. Furthermore, matrix T T 0 should explain the data in a “simple” way, in the sense discussed in the following. Our problem comprises the following terms: Data term The correspondence should respect the data, which is achieved by minimizing kTF - Gk F . Smoothness The correspondence must be smooth, in the sense that if x i , x i 0 are two close points on X , then the respective corresponding functions are similar, i.e., t i t i 0 (see Figure 1). Similarly, for close points y j , y j 0 on Y , the rows t j t j 0 . Smoothness is achieved by minimizing the row and column Dirichlet energy tr(T > L Y T)+ tr(TL X T > ). Localization The correspondence is localized using the L 1 -penalty kTk 1 , which, in combination with smoothness, results in a few non-zero elements that are close in space. This is an extended abstract. The full paper is available at the Computer Vision Foundation webpage. “Simplicity” By simplicity, we mean here that the correspondence matrix is ‘explained’ using a small number of degrees of freedom. The follow- ing models are commonly used in matrix completion and recommendation systems literature. Low rank. A popular model is to find the matrix with smallest rank(T). However, this minimization is known to be NP-hard, and the nuclear norm kTk * is typically used as a convex proxy, leading to min T kTF - Gk 2 F + μ 1 tr(TL X T > )+ μ 2 tr(T > L Y T)+ μ 3 kTk 1 + μ 4 kTk * . (1) where μ 1 , μ 2 , μ 3 > 0 are parameters determining the tradeoff between smooth- ness and localization. Low norm. Srebro et al. [6] rewrite problem (1) using the decom- position T = UV > with U and V of size m × k and n × k, respectively. Note that k can be arbitrarily large. The nuclear norm is written as kTk * = 1 2 (kUk 2 F + kVk 2 F ). Unlike (1), this problem is non-convex w.r.t. both U and V, however behaves well for large k [6]. Subspace parametrization Let L X and L Y be the Laplacians of X and Y , and let Φ k =(φ 1 ,... φ k ) and Ψ k =(ψ 1 ,... ψ k ) be the respective truncated Laplacian eigenbases. The number of variables in the problem (1) depends on the number of samples m, n, which may result in scalability issues for large (m, n 10 6 ) manifolds. To overcome this issue, we approximate the n × k and m × k factors U, V in the truncated Laplacian eigenbases of X and Y using k 0 k first expansion terms, U Ψ k 0 A and V Φ k 0 B, where matrices A, B of the expansion coefficients are of size k 0 × k. This leads to a subspace version of our problem, min A,B kAB > Φ > k 0 F - Ψ > k 0 Gk 2 F + μ 1 tr(AB > Λ X ,k 0 BA > )+ (2) μ 2 tr(BA > Λ Y,k 0 AB > )+ μ 3 kΨ k 0 AB > Φ > k 0 k 1 + μ 4 2 (kΨ k 0 Ak 2 F + kΦ k 0 Bk 2 F ) where we used the invariance of the Frobenius norm to orthogonal transfor- mations and the fact that Φ > k 0 L X Φ k 0 = Λ X ,k 0 , Ψ > k 0 L Y Ψ k 0 = Λ Y,k 0 to simplify the expressions (Λ ·,k denotes a diagonal matrix of k respective eigenvalues). Note that now the number of variables 2kk 0 is independent of the number of samples. We emphasize that k 0 , k can be arbitrarily large and are dictated only by complexity considerations and not by the amount of data. This is one of the major advantages of our approach compared to [3, 4, 5]. Figure 2: Examples of correspondence between non-isometric shapes. Leftmost shape is used as reference. Similar colors encode corresponding points. [1] Vassilis Kalofolias, Xavier Bresson, Michael Bronstein, and Pierre Van- dergheynst. Matrix completion on graphs. arXiv:1408.1717, 2014. [2] Vladimir G Kim, Yaron Lipman, and Thomas Funkhouser. Blended intrinsic maps. TOG, 30(4):79, 2011. [3] Artiom Kovnatsky, Michael M Bronstein, Alexander M Bronstein, Klaus Glashoff, and Ron Kimmel. Coupled quasi-harmonic bases. Computer Graphics Forum, 32:439–448, 2013. [4] M. Ovsjanikov, M. Ben-Chen, J. Solomon, A. Butscher, and L. Guibas. Func- tional maps: A flexible representation of maps between shapes. TOG, 31(4), 2012. [5] Jonathan Pokrass et al. Sparse modeling of intrinsic correspondences. Computer Graphics Forum, 32:459–468, 2013. [6] Nathan Srebro, Jason Rennie, and Tommi S Jaakkola. Maximum-margin matrix factorization. In Proc. NIPS, 2004.

Upload: others

Post on 17-Jul-2020

3 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Functional correspondence by matrix completion · Functional maps Let X and Y denote two manifolds sampled at n and m points, respectively. Ovsjanikov et al. [4] propose to model

Functional correspondence by matrix completion

A. Kovnatsky1, M. M. Bronstein1, Xavier Bresson2, and Pierre Vandergheynst21Faculty of Informatics, Università della Svizzera italiana (USI). 2Signal Processing Lab, École Polytechnique Fédérale de Lausanne (EPFL).

We deal with the problem of finding dense intrinsic correspondence betweenmanifolds in the functional formulation ([4], Fig. 1). We propose treatingfunctional correspondence as geometric matrix completion. We show thatour method compares favorably to state-of-the-art methods for non-rigidshape correspondence on the challenging Princeton benchmark [2]. Theadvantage of our method is especially pronounced in the scarce data set-tings. The proposed method is completely generic and can be applied to anymanifolds and high-dimensional geometric data.

Functional maps Let X and Y denote two manifolds sampled at n and mpoints, respectively. Ovsjanikov et al. [4] propose to model the functionalcorrespondence between the spaces L2(X) and L2(Y ) as the m× n matrixT0, which maps a function f ∈ L2(X) into T0f = g ∈ L2(Y ). Traditionalpoint-wise correspondence is a particular case of this model wherein T0maps delta functions into delta functions.

n

m

ti′′ ti′ ti

xi

xi′

xi′′

y jy j′

y j′′xi′′

Figure 1: Illustration of functional correspondence as matrix completion.Column ti = Tδ i of matrix T is the functional map of a delta at xi. Geo-metric structure is imposed on the columns to ensure smoothness, i.e. ti, ti′

corresponding to spatially close points xi, xi′ on X are similar.

Functional maps as matrix completion Kalofolias et al. [1] studied theproblem of matrix completion on graphs, where the rows and columns of thematrix representing the correspondence have underlying geometric struc-ture. They show that adding geometric structure to the standard matrix com-pletion problem improves recovery results.

We use the same philosophy to formulate the problem of finding a func-tional map as matrix completion, whose rows and columns are interpretedas functions on the respective manifolds X and Y . For this purpose, weconsider the matrix T as a collection of columns T = (t1, . . . , tn) or rowsT = (t1>, . . . , tm>)>, where ti and t j denote the ith column and jth row ofT, respectively. The column ti = Tδ i is the function on Y corresponding toa delta located at point xi on X . Similarly, the row t j = δ

>j T is the function

on X corresponding to a delta located at point y j on Y .As in the classical matrix completion problem, we aim at recovering the

unknown correspondence matrix T0 from a few observations of the formT0F = G, where F = (f1, . . . , fq) and G = (g1, . . . ,gq) are q correspondingfunctions on X , Y , respectively. Furthermore, matrix T≈ T0 should explainthe data in a “simple” way, in the sense discussed in the following. Ourproblem comprises the following terms:

Data term The correspondence should respect the data, which is achievedby minimizing ‖TF−G‖F.

Smoothness The correspondence must be smooth, in the sense that if xi,xi′

are two close points on X , then the respective corresponding functions aresimilar, i.e., ti ≈ ti′ (see Figure 1). Similarly, for close points y j,y j′ on Y , therows t j ≈ t j′ . Smoothness is achieved by minimizing the row and columnDirichlet energy tr(T>LYT)+ tr(TLXT>).

Localization The correspondence is localized using the L1-penalty ‖T‖1,which, in combination with smoothness, results in a few non-zero elementsthat are close in space.

This is an extended abstract. The full paper is available at the Computer Vision Foundationwebpage.

“Simplicity” By simplicity, we mean here that the correspondence matrixis ‘explained’ using a small number of degrees of freedom. The follow-ing models are commonly used in matrix completion and recommendationsystems literature.

Low rank. A popular model is to find the matrix with smallest rank(T).However, this minimization is known to be NP-hard, and the nuclear norm‖T‖∗ is typically used as a convex proxy, leading to

minT‖TF−G‖2

F +µ1tr(TLX T>)+µ2tr(T>LY T)+µ3‖T‖1 +µ4‖T‖∗. (1)

where µ1,µ2,µ3 > 0 are parameters determining the tradeoff between smooth-ness and localization.

Low norm. Srebro et al. [6] rewrite problem (1) using the decom-position T = UV> with U and V of size m× k and n× k, respectively.Note that k can be arbitrarily large. The nuclear norm is written as ‖T‖∗ =12 (‖U‖

2F +‖V‖2

F). Unlike (1), this problem is non-convex w.r.t. both U andV, however behaves well for large k [6].

Subspace parametrization Let LX and LY be the Laplacians of X and Y ,and let Φk = (φ 1, . . .φ k) and Ψk = (ψ1, . . .ψk) be the respective truncatedLaplacian eigenbases. The number of variables in the problem (1) dependson the number of samples m,n, which may result in scalability issues forlarge (m,n ∼ 106) manifolds. To overcome this issue, we approximate then× k and m× k factors U,V in the truncated Laplacian eigenbases of Xand Y using k′ ≥ k first expansion terms, U ≈ Ψk′A and V ≈ Φk′B, wherematrices A,B of the expansion coefficients are of size k′×k. This leads to asubspace version of our problem,

minA,B

‖AB>Φ>k′F−Ψ

>k′G‖

2F +µ1tr(AB>ΛX ,k′BA>)+ (2)

µ2tr(BA>ΛY,k′AB>)+µ3‖Ψk′AB>Φ>k′‖1 +

µ4

2(‖Ψk′A‖2

F +‖Φk′B‖2F)

where we used the invariance of the Frobenius norm to orthogonal transfor-mations and the fact that Φ>k′LX Φk′ = ΛX ,k′ , Ψ>k′LY Ψk′ = ΛY,k′ to simplifythe expressions (Λ·,k denotes a diagonal matrix of k respective eigenvalues).

Note that now the number of variables 2kk′ is independent of the numberof samples. We emphasize that k′,k can be arbitrarily large and are dictatedonly by complexity considerations and not by the amount of data. This isone of the major advantages of our approach compared to [3, 4, 5].

Figure 2: Examples of correspondence between non-isometric shapes.Leftmost shape is used as reference. Similar colors encode correspondingpoints.

[1] Vassilis Kalofolias, Xavier Bresson, Michael Bronstein, and Pierre Van-dergheynst. Matrix completion on graphs. arXiv:1408.1717, 2014.

[2] Vladimir G Kim, Yaron Lipman, and Thomas Funkhouser. Blended intrinsicmaps. TOG, 30(4):79, 2011.

[3] Artiom Kovnatsky, Michael M Bronstein, Alexander M Bronstein, KlausGlashoff, and Ron Kimmel. Coupled quasi-harmonic bases. Computer GraphicsForum, 32:439–448, 2013.

[4] M. Ovsjanikov, M. Ben-Chen, J. Solomon, A. Butscher, and L. Guibas. Func-tional maps: A flexible representation of maps between shapes. TOG, 31(4),2012.

[5] Jonathan Pokrass et al. Sparse modeling of intrinsic correspondences. ComputerGraphics Forum, 32:459–468, 2013.

[6] Nathan Srebro, Jason Rennie, and Tommi S Jaakkola. Maximum-margin matrixfactorization. In Proc. NIPS, 2004.