dimensionality reduction for exponential family data · questions i how to characterize common...

27
Dimensionality Reduction for Exponential Family Data Yoonkyung Lee* Department of Statistics The Ohio State University *joint work with Andrew Landgraf July 2-6, 2018 Computational Strategies for Large-Scale Statistical Data Analysis Workshop ICMS, Edinburgh, UK

Upload: others

Post on 31-Jan-2020

0 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Dimensionality Reduction for Exponential Family Data · Questions I How to characterize common factors underlying a set of binary variables? I Can we apply PCA to binary data? Any

Dimensionality Reductionfor Exponential Family Data

Yoonkyung Lee*Department of Statistics

The Ohio State University*joint work with Andrew Landgraf

July 2-6, 2018Computational Strategies

for Large-Scale Statistical Data Analysis WorkshopICMS, Edinburgh, UK

Page 2: Dimensionality Reduction for Exponential Family Data · Questions I How to characterize common factors underlying a set of binary variables? I Can we apply PCA to binary data? Any

Patient-Diagnosis Matrix

Data source: ICU patients at OSU Medical Center (2007-2010)

Page 3: Dimensionality Reduction for Exponential Family Data · Questions I How to characterize common factors underlying a set of binary variables? I Can we apply PCA to binary data? Any

Questions

I How to characterize common factors underlying a set ofbinary variables?

I Can we apply PCA to binary data?

Any implicit link between PCA and Gaussian distributions?

I How to extend PCA to exponential family data?

I Should we define those factors differently if prediction of aresponse is concerned?

How to make use of the response?

Page 4: Dimensionality Reduction for Exponential Family Data · Questions I How to characterize common factors underlying a set of binary variables? I Can we apply PCA to binary data? Any

Outline

I Dimensionality reduction for non-Gaussian data

{exponential family PCA, generalized PCA}

I Supervised dimensionality reduction for exponential familydata

{supervised generalized PCA, supervised matrix factorization}

Page 5: Dimensionality Reduction for Exponential Family Data · Questions I How to characterize common factors underlying a set of binary variables? I Can we apply PCA to binary data? Any

Generalization of PCACollins et al. (2001), A generalization of principal componentsanalysis to the exponential family

I Draws on the ideas from the exponential family andgeneralized linear models

I For Gaussian data, assume that xi ∼ Np(θi , Ip) and θi ∈ Rp

lies in a k dimensional subspace:

for a basis {b`}k`=1, θi =k∑

`=1

ai`b` = Bp×kai

I To find Θ = [θij ], maximize the log likelihood or equivalentlyminimize the negative log likelihood (or deviance):

n∑i=1

‖xi − θi‖2 = ‖X −Θ‖2F = ‖X − AB>‖2F

Page 6: Dimensionality Reduction for Exponential Family Data · Questions I How to characterize common factors underlying a set of binary variables? I Can we apply PCA to binary data? Any

Generalization of PCAI According to Eckart-Young theorem, the best rank-k

approximation of X (= Un×pDp×pV>p×p) is given by therank-k truncated singular value decomposition UkDk︸ ︷︷ ︸

A

V>k︸︷︷︸B>

I For exponential family data, factorize the matrix of naturalparameter values Θ as AB> with rank-k matrices An×k andBp×k (of orthogonal columns) by maximizing the loglikelihood

I For binary data X = [xij ] with P = [pij ], “logistic PCA” looks

for a factorization of Θ =[log pij

1−pij

]= AB> that maximizes

`(X ; Θ) =∑i,j

{xij(a>i bj∗)− log(1 + exp(a>i bj∗))

}

subject to B>B = Ik

Page 7: Dimensionality Reduction for Exponential Family Data · Questions I How to characterize common factors underlying a set of binary variables? I Can we apply PCA to binary data? Any

Drawbacks of the Matrix Factorization Formulation

I Involves estimation of both case-specific (or row-specific)scores A and variable-specific (or column-specific) factorsB: more of extension of SVD than PCA

I The number of parameters increases with the number ofobservations

I The scores of generalized PC for new data involveadditional optimization while PC scores for standard PCAare simple linear combinations of the data

Page 8: Dimensionality Reduction for Exponential Family Data · Questions I How to characterize common factors underlying a set of binary variables? I Can we apply PCA to binary data? Any

Alternative Interpretation of Standard PCA

I Assuming that data are centered, minimize

n∑i=1

‖xi − VV>xi‖2 = ‖X − XVV>‖2F

subject to V>V = Ik

I XVV> can be viewed as a rank-k projection of the matrixof natural parameters (“means” in this case) of thesaturated model Θ̃ (best possible fit) for Gaussian data

I Standard PCA finds the best rank-k projection of Θ̃ byminimizing the deviance under Gaussian distribution

Page 9: Dimensionality Reduction for Exponential Family Data · Questions I How to characterize common factors underlying a set of binary variables? I Can we apply PCA to binary data? Any

Natural Parameters of the Saturated Model

I For an exponential family distribution with naturalparameter θ and pdf

f (x |θ) = exp (θx − b(θ) + c(x)) ,

E(X ) = b′(θ) and the canonical link function is the inverseof b′.

θ b(θ) canonical linkN(µ,1) µ θ2/2 identityBernoulli(p) logit(p) log(1 + exp(θ)) logitPoisson(λ) log(λ) exp(θ) log

I Take Θ̃ = [canonical link(xij)]

Page 10: Dimensionality Reduction for Exponential Family Data · Questions I How to characterize common factors underlying a set of binary variables? I Can we apply PCA to binary data? Any

New Formulation of Logistic PCA

Landgraf and Lee (2015), Dimensionality Reduction for BinaryData through the Projection of Natural Parameters

I Given xij ∼ Bernoulli(pij), the natural parameter (logit pij )of the saturated model is

θ̃ij = logit(xij) =∞× (2xij − 1)

We will approximate θ̃ij ≈ m × (2xij − 1) for large m > 0

I Project Θ̃ to a k -dimensional subspace by using thedeviance D(X ; Θ) = −2{`(X ; Θ)− `(X ; Θ̃)} as a loss:

minV∈Rp×k

D(X ; Θ̃VV>︸ ︷︷ ︸Θ̂

) = −2∑i,j

{xij θ̂ij − log(1 + exp(θ̂ij))

}

subject to V>V = Ik

Page 11: Dimensionality Reduction for Exponential Family Data · Questions I How to characterize common factors underlying a set of binary variables? I Can we apply PCA to binary data? Any

Logistic PCA vs Logistic SVDI The previous logistic SVD (matrix factorization) gives an

approximation of logit P:

Θ̂LSVD = AB>

I Alternatively, our logistic PCA gives

Θ̂LPCA = Θ̃V︸︷︷︸A

V>,

which has much fewer parameters

I Computation of PC scores on new data only requiresmatrix multiplication for logistic PCA while logistic SVDrequires fitting k -dimensional logistic regression for eachnew observation

I Logistic SVD with additional A is prone to overfit

Page 12: Dimensionality Reduction for Exponential Family Data · Questions I How to characterize common factors underlying a set of binary variables? I Can we apply PCA to binary data? Any

Geometry of Logistic PCA

● ●

−5

0

5

−5 0 5θ1

θ 2

● ●

●●

0

1

0 1X1

X2

Probability

●●●

0.1

0.2

0.3

0.4

●●

●●

PCA

LPCA

Figure: Logistic PCA projection in the natural parameter space withm = 5 (left) and in the probability space (right) compared to the PCAprojection

Page 13: Dimensionality Reduction for Exponential Family Data · Questions I How to characterize common factors underlying a set of binary variables? I Can we apply PCA to binary data? Any

New Formulation of Generalized PCA

Landgraf and Lee (2015), Generalized PCA: Projection ofSaturated Model Parameters

I The idea can be applied to any exponential familydistribution

I Project the matrix of natural parameters from the saturatedmodel Θ̃ to a k -dimensional subspace by using thedeviance D(X ; Θ) = −2{`(X ; Θ)− `(X ; Θ̃)} as a loss:

minV∈Rp×k

D(X ; Θ̃VV>︸ ︷︷ ︸Θ̂

)

subject to V>V = Ik

I If desired, main effects µ can be added to theapproximation of Θ:

Θ̂ = 1µ> + (Θ̃− 1µ>)VV>

Page 14: Dimensionality Reduction for Exponential Family Data · Questions I How to characterize common factors underlying a set of binary variables? I Can we apply PCA to binary data? Any

MM Algorithm for Generalized PCA

I Majorize the objective function with a simpler objective ateach iterate, and minimize the majorizing function.(Hunter and Lange, 2004)

I From the quadratic approximation of the Bernoulli devianceat Θ(t), step t solution, and the fact that p(1− p) ≤ 1/4,

D(X ; 1µ> + (Θ̃− 1µ>)VV>)

≤ 14‖1µ> + (Θ̃− 1µ>)VV> − Z (t+1)‖2F + C,

where Z (t+1) = Θ(t) + 4(X − P̂(t))

I Update Θ at step (t + 1):averaging for µ(t+1) given V (t) andeigen-analysis of a p × p matrix for V (t+1) given µ(t+1)

Page 15: Dimensionality Reduction for Exponential Family Data · Questions I How to characterize common factors underlying a set of binary variables? I Can we apply PCA to binary data? Any

Medical Diagnosis Data

I Part of electronic health record data on 12,000 adultpatients admitted to the intensive care units (ICU) in OhioState University Medical Center from 2007 to 2010

I Patients are classified as having one or more diseases ofover 800 disease categories from the InternationalClassification of Diseases (ICD-9).

I Interested in characterizing the co-morbidity as latentfactors, which can be used to define patient profiles forprediction of other clinical outcomes (e.g. pressure ulcer)

I Analysis is based on a sample of 1,000 patients, whichreduced the number of disease categories to about 600

Page 16: Dimensionality Reduction for Exponential Family Data · Questions I How to characterize common factors underlying a set of binary variables? I Can we apply PCA to binary data? Any

Deviance Explained by Components

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●●●●●●●●● ●

●●●

●●●●●●

●●●●●●●

●●●

●●

●●

●●●●●

Cumulative Marginal

0%

25%

50%

75%

0.0%

2.5%

5.0%

7.5%

0 10 20 30 0 10 20 30Number of Principal Components

% o

f Dev

ianc

e E

xpla

ined

LPCA

LSVD

PCA

Figure: Cumulative and marginal percent of deviance explained byprincipal components of LPCA, LSVD, and PCA

Page 17: Dimensionality Reduction for Exponential Family Data · Questions I How to characterize common factors underlying a set of binary variables? I Can we apply PCA to binary data? Any

Deviance Explained by Parameters

0%

25%

50%

75%

0k 10k 20k 30k 40kNumber of free parameters

% o

f dev

ianc

e ex

plai

ned

LPCA

LSVD

PCA

Figure: Cumulative percent of deviance explained by principalcomponents of LPCA, LSVD, and PCA versus the number of freeparameters

Page 18: Dimensionality Reduction for Exponential Family Data · Questions I How to characterize common factors underlying a set of binary variables? I Can we apply PCA to binary data? Any

Predictive Deviance

●●●●●●●●●●●●●●●●●●●●●●●●●●●●●●

●●

●●●

●●

●●

●●●

●●

●●

●●●

Cumulative Marginal

0%

10%

20%

30%

40%

0%

2%

4%

6%

0 10 20 30 0 10 20 30Number of principal components

% o

f pre

dict

ive

devi

ance

LPCA

PCA

Figure: Cumulative and marginal percent of predictive deviance overtest data (1,000 patients) by the principal components of LPCA andPCA

Page 19: Dimensionality Reduction for Exponential Family Data · Questions I How to characterize common factors underlying a set of binary variables? I Can we apply PCA to binary data? Any

Interpretation of Loadings

●●

●● ●

●● ●

●●

●●

●● ● ●● ●

● ●●

● ● ●

● ●

●●

●●

●●

● ●

●●

● ● ●

●●

●●

●●

●●

●●

●●

● ●

● ●●

●●

●●

●●

●●

●●●

●● ●

● ●

●●

●●

●●

●●

● ●● ●

●●

● ●

● ●●● ●

●●

●●

●●

●●●

● ●

● ●● ●

● ●

●●

●●

●●

● ●

●●●

●●

●●

●●

●●

●●

●●

● ●●

●● ●

●●

●●

●●●●

● ●●

●● ● ●

●●

●● ●●

●●

● ●●● ● ● ●● ●●

●●●

●●●

● ●

● ●●●●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●●●

●●

●● ●●

●●●

●●● ●

● ●● ●●●●

●●● ●

●●●●

● ● ●●●

●●

● ●●

● ●●● ●●●●

●●

● ●●●

●●● ●●

●●

●●

●●●

●●

●●●

● ●●●

●●

● ●● ●

●●

● ●●

● ●

●●

●●

●●

● ● ●

●●

●●

●●

●●

●●

●●

●●

●●● ●

● ●●

●●

●●

●●

●● ●

● ●●

●●

●●

● ●

●●●

● ●●

● ●

● ●●●

●●

● ●

●●

●●

●●

●● ●●

● ●

●●

●● ●

●●

●●

● ●●●

●●

● ●

● ●

● ●

●●

●●

● ●

●●

●●

●● ●

●●● ● ●

●●

●●

●●

●●

●●● ●

●●

●●

●●

●●

●●

●●

●●

●● ●

●● ●

● ●● ●

●●

● ●

● ●●

●●

●●

●●

●●

●●

●●●●

●●● ●●● ●● ● ●● ●● ● ●

●●

●●●

●●●

●●

●●

●●●

●●●●

●●

●●

●●

● ● ●

●●

● ●

● ●

●●

●●

●●

●●

●●

●●

● ●● ●

●●

●● ●●● ●● ●

●●

●●●● ●

●●

●●

●● ●

●●●

●●

●●●

●●●

●●● ●

●●●●

● ●●

● ●

●●

● ●●

●●

●● ●

● ●●●

●● ●

● ●

01:Gram−neg septicemia NEC03:Hyperpotassemia

08:Acute respiratry failure

10:Acute kidney failure NOS16:Speech Disturbance NEC

17:Asphyxiation/strangulat

03:DMII wo cmp nt st uncntr

03:Hyperlipidemia NEC/NOS

04:Leukocytosis NOS

07:Mal hy kid w cr kid I−IV

07:Old myocardial infarct

07:Crnry athrscl natve vssl

07:Systolic hrt failure NOS

09:Oral soft tissue dis NEC

10:Chr kidney dis stage IIIV:Gastrostomy status

−0.2

0.0

0.2

Component 1 Component 2

Figure: The first component is characterized by common seriousconditions that bring patients to ICU, and the second component isdominated by diseases of the circulatory system (07’s).

Page 20: Dimensionality Reduction for Exponential Family Data · Questions I How to characterize common factors underlying a set of binary variables? I Can we apply PCA to binary data? Any

Supervised Generalized PCA

I Extend generalized PCA to the supervised setting with aresponse Y

I Represent predictors X by latent factor scores Θ̃X V andpredict Y with the scores

I Combine deviance for dimensionality reduction andprediction and minimize:

D(Y ; Θ̃X Vβ)︸ ︷︷ ︸prediction

+ α D(X ; Θ̃X VV>)︸ ︷︷ ︸dim reduction

I Dimensionality reduction is a form of regularization

Page 21: Dimensionality Reduction for Exponential Family Data · Questions I How to characterize common factors underlying a set of binary variables? I Can we apply PCA to binary data? Any

Matrix Factorization Approach

Rish et al. (2008), Closed-form supervised dimensionalityreduction with generalized linear models

I Extending Collins et al.’s matrix factorization of exponentialfamily data, consider a latent representation A of X through

ΘX = AB>

and relate A to Y

I Minimize a combination of dimensionality reduction andprediction criteria

D(Y ; Aβ) + α D(X ; AB>)

Page 22: Dimensionality Reduction for Exponential Family Data · Questions I How to characterize common factors underlying a set of binary variables? I Can we apply PCA to binary data? Any

Comparison of Two Approaches

I Representation of latent factor scoresI Previous method (GenSupMF): An×k

The number of parameters increases with the number ofobservations

I Our method (GenSupPCA): Θ̃X Vp×kThe latent factor scores are interpretable as linearcombinations

I As α ↓ 0,I GenSupMF: min D(Y ; Aβ)

Does not use covariates and fits Y perfectly

I GenSupPCA: min D(Y ; Θ̃X Vβ)Reduces to GLM with Θ̃X V as covariates

Page 23: Dimensionality Reduction for Exponential Family Data · Questions I How to characterize common factors underlying a set of binary variables? I Can we apply PCA to binary data? Any

Predicting on New Data

I GenSupMF requires solving for Anew with new data XnewI Given fixed B,

Anew = arg minA

D(Xnew ; AB>)

I When Xnew = Xold for training, prediction will be differentfrom the original fit as the latter involves

minA,B,β

D(Yold ; Aβ) + α D(Xold ; AB>)

I GenSupPCA only requires a linear combination of Θ̃Xnew

and predictions can be made online

Page 24: Dimensionality Reduction for Exponential Family Data · Questions I How to characterize common factors underlying a set of binary variables? I Can we apply PCA to binary data? Any

Computation

I Minimize

D(Y ; Θ̃X Vβ) + α D(X ; Θ̃X VV>)

under the orthonormality constraint:

V>V = Ik

I Algorithm1. With V fixed, find β via GLM fitting

2. With β fixed, minimize V over the Stiefel manifold Vk (Rp)

(Used a gradient based method in Wen and Yin (2013) fororthonormal V )

3. Repeat until convergence

Page 25: Dimensionality Reduction for Exponential Family Data · Questions I How to characterize common factors underlying a set of binary variables? I Can we apply PCA to binary data? Any

Concluding Remarks

I Generalized PCA via projections of the natural parametersof the saturated model using GLM framework

I Proposed a supervised dimensionality reduction methodfor exponential family data by combining generalized PCAfor covariates and a generalized linear model for aresponse

I Impose other constraints on the loadings than rank fordesirable properties (e.g. sparsity)

I R package, logisticPCA is available at CRAN andgeneralizedPCA and genSupPCA are available at GitHub

Page 26: Dimensionality Reduction for Exponential Family Data · Questions I How to characterize common factors underlying a set of binary variables? I Can we apply PCA to binary data? Any

Acknowledgments

Andrew Landgraf@ Battelle Memorial Institute

Sookyung Hyun and Cheryl Newton@ College of Nursing, OSU

DMS-15-13566

Page 27: Dimensionality Reduction for Exponential Family Data · Questions I How to characterize common factors underlying a set of binary variables? I Can we apply PCA to binary data? Any

References

Collins, M., S. Dasgupta, and R. E. Schapire (2001).A generalization of principal components analysis to the exponential family.In T. G. Dietterich, S. Becker, and Z. Ghahramani (Eds.), Advances in Neural Information ProcessingSystems 14, pp. 617–624.

Landgraf, A. J. and Y. Lee (2015a).Dimensionality reduction for binary data through the projection of natural parameters.Technical Report 890, Department of Statistics, The Ohio State University.Also available at arXiv:1510.06112.

Landgraf, A. J. and Y. Lee (2015b).Generalized principal component analysis: Projection of saturated model parameters.Technical Report 892, Department of Statistics, The Ohio State University.

Rish, I., G. Grabarnik, G. Cecchi, F. Pereira, and G. J. Gordon (2008).Closed-form supervised dimensionality reduction with generalized linear models.In Proceedings of the 25th International Conference on Machine Learning, pp. 832–839. ACM.

Wen, Z. and W. Yin (2013).A feasible method for optimization with orthogonality constraints.Mathematical Programming 142(1-2), 397–434.