challenging common assumptions in the unsupervised ... · first theoretically show that the...

37
Challenging Common Assumptions in the Unsupervised Learning of Disentangled Representations Francesco Locatello 12 Stefan Bauer 2 Mario Lucic 3 Gunnar Rätsch 1 Sylvain Gelly 3 Bernhard Schölkopf 2 Olivier Bachem 3 Abstract The key idea behind the unsupervised learning of disentangled representations is that real-world data is generated by a few explanatory factors of variation which can be recovered by unsupervised learning algorithms. In this paper, we provide a sober look at recent progress in the field and challenge some common assumptions. We first theoretically show that the unsupervised learning of disentangled representations is fundamentally impossible without inductive biases on both the models and the data. Then, we train more than 12 000 models covering most prominent methods and evaluation metrics in a reproducible large-scale experimental study on seven different data sets. We observe that while the different methods successfully enforce prop- erties “encouraged” by the corresponding losses, well-disentangled models seemingly cannot be identified without supervision. Furthermore, increased disentanglement does not seem to lead to a decreased sample complexity of learning for downstream tasks. Our results suggest that future work on disentanglement learning should be explicit about the role of inductive biases and (implicit) supervision, investigate concrete ben- efits of enforcing disentanglement of the learned representations, and consider a reproducible experimental setup covering several data sets. 1. Introduction In representation learning it is often assumed that real-world observations x (e.g., images or videos) are generated by 1 ETH Zurich, Department for Computer Science 2 Max- Planck Institute for Intelligent Systems 3 Google Research, Brain Team. Correspondence to: Francesco Locatello <[email protected]>, Olivier Bachem <[email protected]>. Proceedings of the 36 th International Conference on Machine Learning, Long Beach, California, PMLR 97, 2019. Copyright 2019 by the author(s). a two-step generative process. First, a multivariate latent random variable z is sampled from a distribution P (z). In- tuitively, z corresponds to semantically meaningful factors of variation of the observations (e.g., content + position of objects in an image). Then, in a second step, the observation x is sampled from the conditional distribution P (x|z). The key idea behind this model is that the high-dimensional data x can be explained by the substantially lower dimensional and semantically meaningful latent variable z which is mapped to the higher-dimensional space of observations x. Informally, the goal of representation learning is to find useful transformations r(x) of x that “make it easier to extract useful information when building classifiers or other predictors” (Bengio et al., 2013). A recent line of work has argued that representations that are disentangled are an important step towards a better representation learning (Bengio et al., 2013; Peters et al., 2017; LeCun et al., 2015; Bengio et al., 2007; Schmidhuber, 1992; Lake et al., 2017; Tschannen et al., 2018). They should contain all the information present in x in a compact and interpretable structure (Bengio et al., 2013; Kulkarni et al., 2015; Chen et al., 2016) while being independent from the task at hand (Goodfellow et al., 2009; Lenc & Vedaldi, 2015). They should be useful for (semi-)supervised learning of downstream tasks, transfer and few shot learning (Bengio et al., 2013; Schölkopf et al., 2012; Peters et al., 2017). They should enable to integrate out nuisance factors (Kumar et al., 2017), to perform interventions, and to answer counterfactual questions (Pearl, 2009; Spirtes et al., 1993; Peters et al., 2017). While there is no single formalized notion of disentangle- ment (yet) which is widely accepted, the key intuition is that a disentangled representation should separate the distinct, informative factors of variations in the data (Bengio et al., 2013). A change in a single underlying factor of variation z i should lead to a change in a single factor in the learned representation r(x). This assumption can be extended to groups of factors as, for instance, in Bouchacourt et al. (2018) or Suter et al. (2018). Based on this idea, a variety of disentanglement evaluation protocols have been proposed leveraging the statistical relations between the learned arXiv:1811.12359v4 [cs.LG] 18 Jun 2019

Upload: others

Post on 14-May-2020

3 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Challenging Common Assumptions in the Unsupervised ... · first theoretically show that the unsupervised learning of disentangled representations is fundamentally impossible without

Challenging Common Assumptions in the Unsupervised Learning ofDisentangled Representations

Francesco Locatello 1 2 Stefan Bauer 2 Mario Lucic 3 Gunnar Rätsch 1 Sylvain Gelly 3 Bernhard Schölkopf 2

Olivier Bachem 3

AbstractThe key idea behind the unsupervised learningof disentangled representations is that real-worlddata is generated by a few explanatory factors ofvariation which can be recovered by unsupervisedlearning algorithms. In this paper, we providea sober look at recent progress in the field andchallenge some common assumptions. Wefirst theoretically show that the unsupervisedlearning of disentangled representations isfundamentally impossible without inductivebiases on both the models and the data. Then,we train more than 12 000 models covering mostprominent methods and evaluation metrics in areproducible large-scale experimental study onseven different data sets. We observe that whilethe different methods successfully enforce prop-erties “encouraged” by the corresponding losses,well-disentangled models seemingly cannot beidentified without supervision. Furthermore,increased disentanglement does not seem to leadto a decreased sample complexity of learningfor downstream tasks. Our results suggest thatfuture work on disentanglement learning shouldbe explicit about the role of inductive biases and(implicit) supervision, investigate concrete ben-efits of enforcing disentanglement of the learnedrepresentations, and consider a reproducibleexperimental setup covering several data sets.

1. IntroductionIn representation learning it is often assumed that real-worldobservations x (e.g., images or videos) are generated by

1ETH Zurich, Department for Computer Science 2Max-Planck Institute for Intelligent Systems 3Google Research,Brain Team. Correspondence to: Francesco Locatello<[email protected]>, Olivier Bachem<[email protected]>.

Proceedings of the 36 th International Conference on MachineLearning, Long Beach, California, PMLR 97, 2019. Copyright2019 by the author(s).

a two-step generative process. First, a multivariate latentrandom variable z is sampled from a distribution P (z). In-tuitively, z corresponds to semantically meaningful factorsof variation of the observations (e.g., content + position ofobjects in an image). Then, in a second step, the observationx is sampled from the conditional distribution P (x|z). Thekey idea behind this model is that the high-dimensional datax can be explained by the substantially lower dimensionaland semantically meaningful latent variable z which ismapped to the higher-dimensional space of observationsx. Informally, the goal of representation learning is to finduseful transformations r(x) of x that “make it easier toextract useful information when building classifiers or otherpredictors” (Bengio et al., 2013).

A recent line of work has argued that representations thatare disentangled are an important step towards a betterrepresentation learning (Bengio et al., 2013; Peters et al.,2017; LeCun et al., 2015; Bengio et al., 2007; Schmidhuber,1992; Lake et al., 2017; Tschannen et al., 2018). Theyshould contain all the information present in x in a compactand interpretable structure (Bengio et al., 2013; Kulkarniet al., 2015; Chen et al., 2016) while being independentfrom the task at hand (Goodfellow et al., 2009; Lenc &Vedaldi, 2015). They should be useful for (semi-)supervisedlearning of downstream tasks, transfer and few shotlearning (Bengio et al., 2013; Schölkopf et al., 2012; Peterset al., 2017). They should enable to integrate out nuisancefactors (Kumar et al., 2017), to perform interventions, andto answer counterfactual questions (Pearl, 2009; Spirteset al., 1993; Peters et al., 2017).

While there is no single formalized notion of disentangle-ment (yet) which is widely accepted, the key intuition is thata disentangled representation should separate the distinct,informative factors of variations in the data (Bengio et al.,2013). A change in a single underlying factor of variationzi should lead to a change in a single factor in the learnedrepresentation r(x). This assumption can be extended togroups of factors as, for instance, in Bouchacourt et al.(2018) or Suter et al. (2018). Based on this idea, a variety ofdisentanglement evaluation protocols have been proposedleveraging the statistical relations between the learned

arX

iv:1

811.

1235

9v4

[cs

.LG

] 1

8 Ju

n 20

19

Page 2: Challenging Common Assumptions in the Unsupervised ... · first theoretically show that the unsupervised learning of disentangled representations is fundamentally impossible without

Challenging Common Assumptions in the Unsupervised Learning of Disentangled Representations

representation and the ground-truth factor of variations.Disentanglement is then measured as a particular structuralproperty of these relations (Higgins et al., 2017a; Kim &Mnih, 2018; Eastwood & Williams, 2018; Kumar et al.,2017; Chen et al., 2018; Ridgeway & Mozer, 2018).

State-of-the-art approaches for unsupervised disentangle-ment learning are largely based on Variational Autoencoders(VAEs) (Kingma & Welling, 2014): One assumes a specificprior P (z) on the latent space and then uses a deep neuralnetwork to parameterize the conditional probability P (x|z).Similarly, the distribution P (z|x) is approximated using avariational distribution Q(z|x), again parametrized using adeep neural network. The model is then trained by minimiz-ing a suitable approximation to the negative log-likelihood.The representation for r(x) is usually taken to be the meanof the approximate posterior distribution Q(z|x). Severalvariations of VAEs were proposed with the motivation thatthey lead to better disentanglement (Higgins et al., 2017a;Burgess et al., 2017; Kim & Mnih, 2018; Chen et al., 2018;Kumar et al., 2017; Rubenstein et al., 2018). The commontheme behind all these approaches is that they try to enforcea factorized aggregated posterior

∫xQ(z|x)P (x)dx, which

should encourage disentanglement.

Our contributions. In this paper, we challenge commonlyheld assumptions in this field in both theory and practice.Our key contributions can be summarized as follows:

• We theoretically prove that (perhaps unsurprisingly) theunsupervised learning of disentangled representations isfundamentally impossible without inductive biases bothon the considered learning approaches and the data sets.

• We investigate current approaches and their inductivebiases in a reproducible large-scale experimental study1

with a sound experimental protocol for unsupervised dis-entanglement learning. We implement six recent unsu-pervised disentanglement learning methods as well as sixdisentanglement measures from scratch and train morethan 12 000 models on seven data sets.

• We release disentanglement_lib2, a new libraryto train and evaluate disentangled representations. As re-producing our results requires substantial computationaleffort, we also release more than 10 000 trained modelswhich can be used as baselines for future research.

• We analyze our experimental results and challenge com-mon beliefs in unsupervised disentanglement learning: (i)While all considered methods prove effective at ensuringthat the individual dimensions of the aggregated posterior(which is sampled) are not correlated, we observe that the

1Reproducing these experiments requires approximately 2.52GPU years (NVIDIA P100).

2https://github.com/google-research/disentanglement_lib

dimensions of the representation (which is taken to be themean) are correlated. (ii) We do not find any evidencethat the considered models can be used to reliably learndisentangled representations in an unsupervised manneras random seeds and hyperparameters seem to mattermore than the model choice. Furthermore, good trainedmodels seemingly cannot be identified without accessto ground-truth labels even if we are allowed to transfergood hyperparameter values across data sets. (iii) Forthe considered models and data sets, we cannot validatethe assumption that disentanglement is useful for down-stream tasks, for example through a decreased samplecomplexity of learning.

• Based on these empirical evidence, we suggest three crit-ical areas of further research: (i) The role of inductive bi-ases and implicit and explicit supervision should be madeexplicit: unsupervised model selection persists as a keyquestion. (ii) The concrete practical benefits of enforcinga specific notion of disentanglement of the learned rep-resentations should be demonstrated. (iii) Experimentsshould be conducted in a reproducible experimental setupon data sets of varying degrees of difficulty.

2. Other related workIn a similar spirit to disentanglement, (non-)linear indepen-dent component analysis (Comon, 1994; Bach & Jordan,2002; Jutten & Karhunen, 2003; Hyvarinen & Morioka,2016) studies the problem of recovering independent com-ponents of a signal. The underlying assumption is that thereis a generative model for the signal composed of the com-bination of statistically independent non-Gaussian compo-nents. While the identifiability result for linear ICA (Comon,1994) proved to be a milestone for the classical theory offactor analysis, similar results are in general not obtainablefor the nonlinear case and the underlying sources gener-ating the data cannot be identified (Hyvarinen & Pajunen,1999). The lack of almost any identifiability result in non-linear ICA has been a main bottleneck for the utility of theapproach (Hyvarinen et al., 2018) and partially motivatedalternative machine learning approaches (Desjardins et al.,2012; Schmidhuber, 1992; Cohen & Welling, 2015). Giventhat unsupervised algorithms did not initially perform wellon realistic settings most of the other works have consid-ered some more or less explicit form of supervision (Reedet al., 2014; Zhu et al., 2014; Yang et al., 2015; Kulka-rni et al., 2015; Cheung et al., 2015; Mathieu et al., 2016;Narayanaswamy et al., 2017; Suter et al., 2018). (Hintonet al., 2011; Cohen & Welling, 2014) assume some knowl-edge of the effect of the factors of variations even thoughthey are not observed. One can also exploit known relationsbetween factors in different samples (Karaletsos et al., 2015;Goroshin et al., 2015; Whitney et al., 2016; Fraccaro et al.,

Page 3: Challenging Common Assumptions in the Unsupervised ... · first theoretically show that the unsupervised learning of disentangled representations is fundamentally impossible without

Challenging Common Assumptions in the Unsupervised Learning of Disentangled Representations

2017; Denton & Birodkar, 2017; Hsu et al., 2017; Yingzhen& Mandt, 2018) or explicit inductive biases (Locatello et al.,2018). This is not a limiting assumption especially in se-quential data, i.e., for videos. We focus our study on thesetting where factors of variations are not observable at all,i.e. we only observe samples from P (x).

3. Impossibility resultThe first question that we investigate is whether unsuper-vised disentanglement learning is even possible for arbitrarygenerative models. Theorem 1 essentially shows that with-out inductive biases both on models and data sets the taskis fundamentally impossible. The proof is provided in Ap-pendix A.

Theorem 1. For d > 1, let z ∼ P denote any distributionwhich admits a density p(z) =

∏di=1 p(zi). Then, there

exists an infinite family of bijective functions f : supp(z)→supp(z) such that ∂fi(u)

∂uj6= 0 almost everywhere for all

i and j (i.e., z and f(z) are completely entangled) andP (z ≤ u) = P (f(z) ≤ u) for all u ∈ supp(z) (i.e., theyhave the same marginal distribution).

Consider the commonly used “intuitive” notion of disentan-glement which advocates that a change in a single ground-truth factor should lead to a single change in the repre-sentation. In that setting, Theorem 1 implies that unsu-pervised disentanglement learning is impossible for arbi-trary generative models with a factorized prior3 in the fol-lowing sense: Assume we have p(z) and some P (x|z)defining a generative model. Consider any unsuperviseddisentanglement method and assume that it finds a repre-sentation r(x) that is perfectly disentangled with respectto z in the generative model. Then, Theorem 1 impliesthat there is an equivalent generative model with the la-tent variable z = f(z) where z is completely entangledwith respect to z and thus also r(x): as all the entriesin the Jacobian of f are non-zero, a change in a singledimension of z implies that all dimensions of z change.Furthermore, since f is deterministic and p(z) = p(z) al-most everywhere, both generative models have the samemarginal distribution of the observations x by construction,i.e., P (x) =

∫p(x|z)p(z)dz =

∫p(x|z)p(z)dz. Since the

(unsupervised) disentanglement method only has access toobservations x, it hence cannot distinguish between the twoequivalent generative models and thus has to be entangledto at least one of them.

This may not be surprising to readers familiar with thecausality and ICA literature as it is consistent with thefollowing argument: After observing x, we can construct

3Theorem 1 only applies to factorized priors; however, weexpect that a similar result can be extended to non-factorizingpriors.

infinitely many generative models which have the samemarginal distribution of x. Any one of these models couldbe the true causal generative model for the data, and theright model cannot be identified given only the distributionof x (Peters et al., 2017). Similar results have been obtainedin the context of non-linear ICA (Hyvarinen & Pajunen,1999). The main novelty of Theorem 1 is that it allows theexplicit construction of latent spaces z and z that are com-pletely entangled with each other in the sense of (Bengioet al., 2013). We note that while this result is very intuitivefor multivariate Gaussians it also holds for distributionswhich are not invariant to rotation, for example multivariateuniform distributions.

While Theorem 1 shows that unsupervised disentanglementlearning is fundamentally impossible for arbitrary genera-tive models, this does not necessarily mean it is an impossi-ble endeavour in practice. After all, real world generativemodels may have a certain structure that could be exploitedthrough suitably chosen inductive biases. However, Theo-rem 1 clearly shows that inductive biases are required bothfor the models (so that we find a specific set of solutions)and for the data sets (such that these solutions match the truegenerative model). We hence argue that the role of inductivebiases should be made explicit and investigated further asdone in the following experimental study.

4. Experimental designConsidered methods. All the considered methods augmentthe VAE loss with a regularizer: The β-VAE (Higgins et al.,2017a), introduces a hyperparameter in front of the KL reg-ularizer of vanilla VAEs to constrain the capacity of theVAE bottleneck. The AnnealedVAE (Burgess et al., 2017)progressively increase the bottleneck capacity so that theencoder can focus on learning one factor of variation at thetime (the one that most contribute to a small reconstruc-tion error). The FactorVAE (Kim & Mnih, 2018) and theβ-TCVAE (Chen et al., 2018) penalize the total correla-tion (Watanabe, 1960) with adversarial training (Nguyenet al., 2010; Sugiyama et al., 2012) or with a tractable butbiased Monte-Carlo estimator respectively. The DIP-VAE-Iand the DIP-VAE-II (Kumar et al., 2017) both penalize themismatch between the aggregated posterior and a factorizedprior. Implementation details and further discussion on themethods can be found in Appendix B and G.

Considered metrics. The BetaVAE metric (Higgins et al.,2017a) measures disentanglement as the accuracy of a linearclassifier that predicts the index of a fixed factor of variation.Kim & Mnih (2018) address several issues with this metricin their FactorVAE metric by using a majority vote classifieron a different feature vector which accounts for a cornercase in the BetaVAE metric. The Mutual Information Gap(MIG) (Chen et al., 2018) measures for each factor of vari-

Page 4: Challenging Common Assumptions in the Unsupervised ... · first theoretically show that the unsupervised learning of disentangled representations is fundamentally impossible without

Challenging Common Assumptions in the Unsupervised Learning of Disentangled Representations

ation the normalized gap in mutual information between thehighest and second highest coordinate in r(x). Instead, theModularity (Ridgeway & Mozer, 2018) measures if eachdimension of r(x) depends on at most a factor of variationusing their mutual information. The Disentanglement metricof Eastwood & Williams (2018) (which we call DCI Disen-tanglement for clarity) computes the entropy of the distribu-tion obtained by normalizing the importance of each dimen-sion of the learned representation for predicting the value ofa factor of variation. The SAP score (Kumar et al., 2017) isthe average difference of the prediction error of the two mostpredictive latent dimensions for each factor. Implementationdetails and further descriptions can be found in Appendix C.

Data sets. We consider four data sets in which x is ob-tained as a deterministic function of z: dSprites (Higginset al., 2017a), Cars3D (Reed et al., 2015), SmallNORB (Le-Cun et al., 2004), Shapes3D (Kim & Mnih, 2018). Wealso introduce three data sets where the observations x arestochastic given the factor of variations z: Color-dSprites,Noisy-dSprites and Scream-dSprites. In Color-dSprites, theshapes are colored with a random color. In Noisy-dSprites,we consider white-colored shapes on a noisy background.Finally, in Scream-dSprites the background is replaced witha random patch in a random color shade extracted from thefamous The Scream painting (Munch, 1893). The dSpritesshape is embedded into the image by inverting the color ofits pixels. Further details on the preprocessing of the datacan be found in Appendix H.

Inductive biases. To fairly evaluate the different ap-proaches, we separate the effect of regularization (in theform of model choice and regularization strength) from theother inductive biases (e.g., the choice of the neural architec-ture). Each method uses the same convolutional architecture,optimizer, hyperparameters of the optimizer and batch size.All methods use a Gaussian encoder where the mean and thelog variance of each latent factor is parametrized by the deepneural network, a Bernoulli decoder and latent dimensionfixed to 10. We note that these are all standard choices inprior work (Higgins et al., 2017a; Kim & Mnih, 2018).

We choose six different regularization strengths, i.e., hy-perparameter values, for each of the considered methods.The key idea was to take a wide enough set to ensure thatthere are useful hyperparameters for different settings foreach method and not to focus on specific values known towork for specific data sets. However, the values are par-tially based on the ranges that are prescribed in the literature(including the hyperparameters suggested by the authors).

We fix our experimental setup in advance and we run all theconsidered methods on each data set for 50 different randomseeds and evaluate them on the considered metrics. Thefull details on the experimental setup are provided in theAppendix G. Our experimental setup, the limitations of this

study, and the differences with previous implementationsare extensively discussed in Appendices D-F.

5. Key experimental resultsIn this section, we highlight our key findings with plotsspecifically picked to be representative of our main results.In Appendix I, we provide the full experimental results witha complete set of plots for different methods, data sets anddisentanglement metrics.

5.1. Can current methods enforce a uncorrelatedaggregated posterior and representation?

While many of the considered methods aim to enforce afactorizing and thus uncorrelated aggregated posterior (e.g.,regularizing the total correlation of the sampled representa-tion), they use the mean vector of the Gaussian encoder asthe representation and not a sample from the Gaussian en-coder. This may seem like a minor, irrelevant modification;however, it is not clear whether a factorizing aggregatedposterior also ensures that the dimensions of the mean rep-resentation are uncorrelated. To test the impact of this, wecompute the total correlation of both the mean and the sam-pled representation based on fitting Gaussian distributionsfor each data set, model and hyperparameter value (seeAppendix C and I.2 for details).

Figure 1 (left) shows the total correlation based on a fittedGaussian of the sampled representation plotted against theregularization strength for each method except Annealed-VAE on Color-dSprites. We observe that the total correlationof the sampled representation generally decreases with theregularization strength. One the other hand, Figure 1 (right)shows the total correlation of the mean representation plot-ted against the regularization strength. It is evident that thetotal correlation of the mean representation generally in-creases with the regularization strength. The only exceptionis DIP-VAE-I for which we observe that the total correlation

0.0 0.2 0.4 0.6 0.8 1.0Regularization strength

0.00

0.02

0.04

0.06

0.08

0.10

0.12

Valu

e

Metric = TC (sampled)

VAE β-VAE FactorVAE β-TCVAE DIP-VAE-I DIP-VAE-II

0.0 0.2 0.4 0.6 0.8 1.0Regularization strength

0.0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8Metric = TC (mean)

Figure 1. Total correlation based on a fitted Gaussian of the sam-pled (left) and the mean representation (right) plotted against reg-ularization strength for Color-dSprites and approaches (exceptAnnealedVAE). The total correlation of the sampled representationdecreases while the total correlation of the mean representationincreases as the regularization strength is increased.

Page 5: Challenging Common Assumptions in the Unsupervised ... · first theoretically show that the unsupervised learning of disentangled representations is fundamentally impossible without

Challenging Common Assumptions in the Unsupervised Learning of Disentangled Representations

of the mean representation is consistently low. This is notsurprising as the DIP-VAE-I objective directly optimizes thecovariance matrix of the mean representation to be diagonalwhich implies that the corresponding total correlation (aswe measure it) is low. These findings are confirmed by ourdetailed experimental results in Appendix I.2 (in particularFigures 8-9) which considers all different data sets. Further-more, we observe largely the same pattern if we consider theaverage mutual information between different dimensionof the representation instead of the total correlation (seeFigures 27-28 in Appendix J).

Implications. Overall, these results lead us to concludewith minor exceptions that the considered methods are effec-tive at enforcing an aggregated posterior whose individualdimensions are not correlated but that this does not seemto imply that the dimensions of the mean representation(usually used for representation) are uncorrelated.

(A) (B) (C) (D) (E) (F)

BetaVAE Score (A)

FactorVAE Score (B)

MIG (C)

DCI Disentanglement (D)

Modularity (E)

SAP (F)

100 80 44 41 46 37

80 100 49 52 25 38

44 49 100 76 6 42

41 52 76 100 -8 38

46 25 6 -8 100 13

37 38 42 38 13 100

Dataset = Noisy-dSprites

Figure 2. Rank correlation of different metrics on Noisy-dSprites.Overall, we observe that all metrics except Modularity seem mildlycorrelated with the pairs BetaVAE and FactorVAE, and MIG andDCI Disentanglement strongly correlated with each other.

5.2. How much do the disentanglement metrics agree?

As there exists no single, common definition of disentangle-ment, an interesting question is to see how much differentproposed metrics agree. Figure 2 shows the Spearman rankcorrelation between different disentanglement metrics onNoisy-dSprites whereas Figure 12 in Appendix I.3 shows thecorrelation for all the different data sets. We observe that allmetrics except Modularity seem to be correlated strongly onthe data sets dSprites, Color-dSprites and Scream-dSpritesand mildly on the other data sets. There appear to be twopairs among these metrics that capture particularly similarnotions: the BetaVAE and the FactorVAE score as well asthe MIG and DCI Disentanglement.

Implication. All disentanglement metrics except Modular-ity appear to be correlated. However, the level of correlationchanges between different data sets.

5.3. How important are different models andhyperparameters for disentanglement?

The primary motivation behind the considered methods isthat they should lead to improved disentanglement. This

0 1 2 3 4 5Model

0.60

0.65

0.70

0.75

0.80

0.85

0.90

0.95

1.00

Valu

e

Data

set =

Cars3

D

Metric = FactorVAE Score

0.0 0.2 0.4 0.6 0.8 1.0Regularization strength

0.60

0.65

0.70

0.75

0.80

0.85

0.90

0.95

1.00

Valu

e

Model =

Facto

rVA

EMetric = FactorVAE Score

Figure 3. (left) FactorVAE score for each method on Cars3D.Models are abbreviated (0=β-VAE, 1=FactorVAE, 2=β-TCVAE,3=DIP-VAE-I, 4=DIP-VAE-II, 5=AnnealedVAE). The variance isdue to different hyperparameters and random seeds. The scores areheavily overlapping. (right) Distribution of FactorVAE scores forFactorVAE model for different regularization strengths on Cars3D.In this case, the variance is only due to the different random seeds.We observe that randomness (in the form of different randomseeds) has a substantial impact on the attained result and that agood run with a bad hyperparameter can beat a bad run with agood hyperparameter.

raises the question how disentanglement is affected by themodel choice, the hyperparameter selection and randomness(in the form of different random seeds). To investigate this,we compute all the considered disentanglement metrics foreach of our trained models.

In Figure 3 (left), we show the range of attainable Factor-VAE scores for each method on Cars3D. We observe thatthese ranges are heavily overlapping for different modelsleading us to (qualitatively) conclude that the choice of hy-perparameters and the random seed seems to be substantiallymore important than the choice of objective function. Theseresults are confirmed by the full experimental results on allthe data sets presented in Figure 13 of Appendix I.4: Whilecertain models seem to attain better maximum scores onspecific data sets and disentanglement metrics, we do notobserve any consistent pattern that one model is consistentlybetter than the other. At this point, we note that in our studywe have fixed the range of hyperparameters a priori to sixdifferent values for each model and did not explore addi-tional hyperparameters based on the results (as that wouldbias our study). However, this also means that specific mod-els may have performed better than in Figure 13 (left) if wehad chosen a different set of hyperparameters.

In Figure 3 (right), we further show the impact of random-ness in the form of random seeds on the disentanglementscores. Each violin plot shows the distribution of the Fac-torVAE metric across all 50 trained FactorVAE models foreach hyperparameter setting on Cars3D. We clearly see thatrandomness (in the form of different random seeds) has asubstantial impact on the attained result and that a goodrun with a bad hyperparameter can beat a bad run with agood hyperparameter in many cases. Again, these findings

Page 6: Challenging Common Assumptions in the Unsupervised ... · first theoretically show that the unsupervised learning of disentangled representations is fundamentally impossible without

Challenging Common Assumptions in the Unsupervised Learning of Disentangled Representations

0.0 0.2 0.4 0.6 0.8 1.0Regularization strength

0.70

0.75

0.80

0.85

0.90

0.95

Valu

e

Dataset = Cars3D | Metric = FactorVAE Score

β-VAE

FactorVAE

β-TCVAE

DIP-VAE-I

DIP-VAE-II

AnnealedVAE

(A) (B) (C) (D) (E) (F)

Reconstruction

TC (sampled)

KL

ELBO

-30 -4 59 22 -21 27

1 5 -11 -8 -11 -2

-14 -1 -38 -31 -11 -29

-38 -9 48 9 -25 15

Dataset = Shapes3D

(I) (II) (III) (IV) (V) (VI) VII

dSprites (I)

Color-dSprites (II)

Noisy-dSprites (III)

Scream-dSprites (IV)

SmallNORB (V)

Cars3D (VI)

Shapes3D (VII)

100 95 65 65 34 64 46

95 100 61 60 21 63 47

65 61 100 68 17 64 59

65 60 68 100 36 93 69

34 21 17 36 100 21 -9

64 63 64 93 21 100 85

46 47 59 69 -9 85 100

Metric = DCI Disentanglement

Figure 4. (left) FactorVAE score vs hyperparameters for each score on Cars3d. There seems to be no model dominating all the othersand for each model there does not seem to be a consistent strategy in choosing the regularization strength. (center) Unsupervised scoresvs disentanglement metrics on Shapes3D. Metrics are abbreviated ((A)=BetaVAE Score, (B)=FactorVAE Score, (C)=MIG , (D)=DCIDisentanglement, (E)=Modularity, (F)=SAP). The unsupervised scores we consider do not seem to be useful for model selection. (right)Rank-correlation of DCI disentanglement metric across different data sets. Good hyperparameters only seem to transfer between dSpritesand Color-dSprites but not in between the other data sets.

are consistent with the complete set of plots provided inFigure 14 of Appendix I.4.

Finally, we perform a variance analysis by trying to predictthe different disentanglement scores using ordinary leastsquares for each data set: If we allow the score to dependonly on the objective function (treated as a categorical vari-able), we are only able to explain 37% of the variance of thescores on average (see Table 5 in Appendix I.4 for furtherdetails). Similarly, if the score depends on the Cartesianproduct of objective function and regularization strength(again categorical), we are able to explain 59% of the vari-ance while the rest is due to the random seed.

Implication. The disentanglement scores of unsupervisedmodels are heavily influenced by randomness (in the formof the random seed) and the choice of the hyperparameter(in the form of the regularization strength). The objectivefunction appears to have less impact.

5.4. Are there reliable recipes for model selection?

In this section, we investigate how good hyperparameterscan be chosen and how we can distinguish between goodand bad training runs. In this paper, we advocate that thatmodel selection should not depend on the considered dis-entanglement score for the following reasons: The point ofunsupervised learning of disentangled representation is thatthere is no access to the labels as otherwise we could incor-porate them and would have to compare to semi-supervisedand fully supervised methods. All the disentanglement met-rics considered in this paper require a substantial amountof ground-truth labels or the full generative model (for ex-ample for the BetaVAE and the FactorVAE metric). Hence,one may substantially bias the results of a study by tun-ing hyperparameters based on (supervised) disentanglementmetrics. Furthermore, we argue that it is not sufficient to fixa set of hyperparameters a priori and then show that one ofthose hyperparameters and a specific random seed achievesa good disentanglement score as it amounts to showing the

existence of a good model, but does not guide the practi-tioner in finding it. Finally, in many practical settings, wemight not even have access to adequate labels as it maybe hard to identify the true underlying factor of variations,in particular, if we consider data modalities that are lesssuitable to human interpretation than images.

In the remainder of this section, we hence investigate andassess different ways how hyperparameters and good modelruns could be chosen. In this study, we focus on choosingthe learning model and the regularization strength corre-sponding to that loss function. However, we note that inpractice this problem is likely even harder as a practitionermight also want to tune other modeling choices such archi-tecture or optimizer.

General recipes for hyperparameter selection. We firstinvestigate whether we may find generally applicable “rulesof thumb” for choosing the hyperparameters. For this, weplot in Figure 4 (left) the FactorVAE score against differentregularization strengths for each model on the Cars3D dataset whereas Figure 16 in Appendix I.5 shows the same plotfor all data sets and disentanglement metrics. The valuescorrespond to the median obtained values across 50 randomseeds for each model, hyperparameter and data set. Overall,there seems to be no model consistently dominating allthe others and for each model there does not seem to be aconsistent strategy in choosing the regularization strength tomaximize disentanglement scores. Furthermore, even if wecould identify a good objective function and correspondinghyperparameter value, we still could not distinguish betweena good and a bad training run.

Model selection based on unsupervised scores. Anotherapproach could be to select hyperparameters based on un-supervised scores such as the reconstruction error, the KLdivergence between the prior and the approximate posterior,the Evidence Lower BOund or the estimated total corre-lation of the sampled representation (mean representationgives similar results). This would have the advantage that

Page 7: Challenging Common Assumptions in the Unsupervised ... · first theoretically show that the unsupervised learning of disentangled representations is fundamentally impossible without

Challenging Common Assumptions in the Unsupervised Learning of Disentangled Representations

Table 1. Probability of outperforming random model selection on adifferent random seed. A random disentanglement metric and dataset is sampled and used for model selection. That model is thencompared to a randomly selected model: (i) on the same metricand data set, (ii) on the same metric and a random different dataset, (iii) on a random different metric and the same data set, and(iv) on a random different metric and a random different data set.The results are averaged across 10 000 random draws.

Random data set Same data set

Random metric 54.9% 62.6%Same metric 59.3% 80.7%

we could select specific trained models and not just goodhyperparameter settings whose median trained model wouldperform well. To test whether such an approach is fruitful,we compute the rank correlation between these unsuper-vised metrics and the disentanglement metrics and presentit in Figure 4 (center) for Shapes3D and in Figure 16 ofAppendix I.5 for all the different data sets. While we doobserve some correlations, no clear pattern emerges whichleads us to conclude that this approach is unlikely to besuccessful in practice.

Hyperparameter selection based on transfer. The finalstrategy for hyperparameter selection that we consider isbased on transferring good settings across data sets. Thekey idea is that good hyperparameter settings may be in-ferred on data sets where we have labels available (such asdSprites) and then applied to novel data sets. Figure 4 (right)shows the rank correlations obtained between different datasets for the DCI disentanglement (whereas Figure 17 in Ap-pendix I.5 shows it for all data sets). We find a strong andconsistent correlation between dSprites and Color-dSprites.While these results suggest that some transfer of hyper-parameters is possible, it does not allow us to distinguishbetween good and bad random seeds on the target data set.

To illustrate this, we compare such a transfer based approachto hyperparameter selection to random model selection asfollows: First, we sample one of our 50 random seeds, arandom disentanglement metric and a data set and use themto select the hyperparameter setting with the highest attainedscore. Then, we compare that selected hyperparameter set-ting to a randomly selected model on either the same ora random different data set, based on either the same or arandom different metric and for a randomly sampled seed.Finally, we report the percentage of trials in which thistransfer strategy outperforms or performs equally well asrandom model selection across 10 000 trials in Table 1. Ifwe choose the same metric and the same data set (but a dif-ferent random seed), we obtain a score of 80.7%. If we aimto transfer for the same metric across data sets, we achievearound 59.3%. Finally, if we transfer both across metricsand data sets, our performance drops to 54.9%.

LR10

LR100

LR1000

LR10000

GB

T10

GB

T100

GB

T1000

GB

T10000

Eff

icie

ncy

(LR

)

Eff

icie

ncy

(G

BT)

BetaVAE Score

FactorVAE Score

MIG

DCI Disentanglement

Modularity

SAP

18 65 28 28 67 78 75 76 50 50

13 49 13 12 58 73 71 71 43 46

18 63 20 -1 71 86 86 87 62 47

19 65 18 4 75 94 94 94 62 54

-3 -9 15 18 -6 -17 -19 -13 -19 -14

12 64 20 12 71 77 74 75 56 49

Dataset = dSprites

Figure 5. Rank correlations between disentanglement metrics anddownstream performance (accuracy and efficiency) on dSprites.

Implications. Unsupervised model selection remains an un-solved problem. Transfer of good hyperparameters betweenmetrics and data sets does not seem to work as there appearsto be no unsupervised way to distinguish between good andbad random seeds on the target task.

5.5. Are these disentangled representations useful fordownstream tasks in terms of the samplecomplexity of learning?

One of the key motivations behind disentangled rep-resentations is that they are assumed to be useful forlater downstream tasks. In particular, it is argued thatdisentanglement should lead to a better sample complexityof learning (Bengio et al., 2013; Schölkopf et al., 2012;Peters et al., 2017). In this section, we consider the simplestdownstream classification task where the goal is to recoverthe true factors of variations from the learned representationusing either multi-class logistic regression (LR) or gradientboosted trees (GBT).

Figure 5 shows the rank correlations between the disen-tanglement metrics and the downstream performance ondSprites. We observe that all metrics except Modularityseem to be correlated with increased downstream perfor-mance on the different variations of dSprites and to somedegree on Shapes3D but not on the other data sets. However,it is not clear whether this is due to the fact that disentangledrepresentations perform better or whether some of thesescores actually also (partially) capture the informativenessof the evaluated representation. Furthermore, the full resultsin Figure 19 of Appendix I.6 indicate that the correlation isweaker or inexistent on other data sets (e.g. Cars3D).

To assess the sample complexity argument we computefor each trained model a statistical efficiency score whichwe define as the average accuracy based on 100 samplesdivided by the average accuracy based on 10 000 samples.Figure 6 show the sample efficiency of learning (based onGBT) versus the FactorVAE Score on dSprites. We do notobserve that higher disentanglement scores reliably lead toa higher sample efficiency. This finding which appears to beconsistent with the results in Figures 20-23 of Appendix I.6.

Page 8: Challenging Common Assumptions in the Unsupervised ... · first theoretically show that the unsupervised learning of disentangled representations is fundamentally impossible without

Challenging Common Assumptions in the Unsupervised Learning of Disentangled Representations

0.2 0.4 0.6 0.8 1.0Value

0.40

0.48

0.56

0.64

Eff

icie

ncy

(G

BT)

Data

set =

dSprite

s

Metric = FactorVAE Score

β-VAE

FactorVAE

β-TCVAE

DIP-VAE-I

DIP-VAE-II

AnnealedVAE

Figure 6. Statistical efficiency of the FactorVAE Score for learninga GBT downstream task on dSprites.

Implications. While the empirical results in this sectionare negative, they should also be interpreted with care.After all, we have seen in previous sections that themodels considered in this study fail to reliably producedisentangled representations. Hence, the results in thissection might change if one were to consider a different setof models, for example semi-supervised or fully supervisedone. Furthermore, there are many more potential notionsof usefulness such as interpretability and fairness thatwe have not considered in our experimental evaluation.Nevertheless, we argue that the lack of concrete examples ofuseful disentangled representations necessitates that futurework on disentanglement methods should make this pointmore explicit. While prior work (Steenbrugge et al., 2018;Laversanne-Finot et al., 2018; Nair et al., 2018; Higginset al., 2017b; 2018) successfully applied disentanglementmethods such as β-VAE on a variety of downstream tasks,it is not clear to us that these approaches and trained modelsperformed well because of disentanglement.

6. ConclusionsIn this work we first theoretically show that the unsupervisedlearning of disentangled representations is fundamentallyimpossible without inductive biases. We then performeda large-scale empirical study with six state-of-the-artdisentanglement methods, six disentanglement metrics onseven data sets and conclude the following: (i) A factorizingaggregated posterior (which is sampled) does not seem tonecessarily imply that the dimensions in the representation(which is taken to be the mean) are uncorrelated. (ii)Random seeds and hyperparameters seem to matter morethan the model but tuning seem to require supervision. (iii)We did not observe that increased disentanglement impliesa decreased sample complexity of learning downstreamtasks. Based on these findings, we suggest three maindirections for future research:

Inductive biases and implicit and explicit supervision.Our theoretical impossibility result in Section 3 highlightsthe need of inductive biases while our experimentalresults indicate that the role of supervision is crucial. Ascurrently there does not seem to exist a reliable strategyto choose hyperparameters in the unsupervised learningof disentangled representations, we argue that future workshould make the role of inductive biases and implicit andexplicit supervision more explicit. We would encourageand motivate future work on disentangled representationlearning that deviates from the static, purely unsupervisedsetting considered in this work. Promising settings (thathave been explored to some degree) seem to be for example(i) disentanglement learning with interactions (Thomaset al., 2017), (ii) when weak forms of supervision e.g.through grouping information are available (Bouchacourtet al., 2018), or (iii) when temporal structure is availablefor the learning problem. The last setting seems to beparticularly interesting given recent identifiability resultsin non-linear ICA (Hyvarinen & Morioka, 2016).

Concrete practical benefits of disentangled representa-tions. In our experiments we investigated whether higherdisentanglement scores lead to increased sample efficiencyfor downstream tasks and did not find evidence that thisis the case. While these results only apply to the settingand downstream task used in our study, we are also notaware of other prior work that compellingly shows theusefulness of disentangled representations. Hence, weargue that future work should aim to show concrete benefitsof disentangled representations. Interpretability and fairnessas well as interactive settings seem to be particularlypromising candidates to evaluate usefulness. One potentialapproach to include inductive biases, offer interpretability,and generalization is the concept of independent causalmechanisms and the framework of causal inference (Pearl,2009; Peters et al., 2017).

Experimental setup and diversity of data sets. Ourstudy also highlights the need for a sound, robust, andreproducible experimental setup on a diverse set of data setsin order to draw valid conclusions. We have observed thatit is easy to draw spurious conclusions from experimentalresults if one only considers a subset of methods, metricsand data sets. Hence, we argue that it is crucial for futurework to perform experiments on a wide variety of datasets to see whether conclusions and insights are generallyapplicable. This is particularly important in the settingof disentanglement learning as experiments are largelyperformed on toy-like data sets. For this reason, we releaseddisentanglement_lib, the library we created to trainand evaluate the different disentanglement methods onmultiple data sets. We also released more than 10 000trained models to provide a solid baseline for futuremethods and metrics.

Page 9: Challenging Common Assumptions in the Unsupervised ... · first theoretically show that the unsupervised learning of disentangled representations is fundamentally impossible without

Challenging Common Assumptions in the Unsupervised Learning of Disentangled Representations

AcknowledgementsThe authors thank Ilya Tolstikhin, Paul Rubenstein and JosipDjolonga for helpful discussions and comments. This re-search was partially supported by the Max Planck ETHCenter for Learning Systems and by an ETH core grant(to Gunnar Rätsch). This work was partially done whileFrancesco Locatello was at Google Research Zurich.

ReferencesArcones, M. A. and Gine, E. On the bootstrap of u and v

statistics. The Annals of Statistics, pp. 655–674, 1992.

Bach, F. R. and Jordan, M. I. Kernel independent componentanalysis. Journal of machine learning research, 3(Jul):1–48, 2002.

Bengio, Y., LeCun, Y., et al. Scaling learning algorithmstowards ai. Large-scale kernel machines, 34(5):1–41,2007.

Bengio, Y., Courville, A., and Vincent, P. Representationlearning: A review and new perspectives. IEEE transac-tions on pattern analysis and machine intelligence, 35(8):1798–1828, 2013.

Bouchacourt, D., Tomioka, R., and Nowozin, S. Multi-levelvariational autoencoder: Learning disentangled represen-tations from grouped observations. In AAAI, 2018.

Burgess, C. P., Higgins, I., Pal, A., Matthey, L., Watters, N.,Desjardins, G., and Lerchner, A. Understanding disentan-gling in beta-vae. In Workshop on Learning DisentangledRepresentations at the 31st Conference on Neural Infor-mation Processing Systems, 2017.

Chen, T. Q., Li, X., Grosse, R. B., and Duvenaud, D. K.Isolating sources of disentanglement in variational autoen-coders. In Advances in Neural Information ProcessingSystems, pp. 2615–2625, 2018.

Chen, X., Duan, Y., Houthooft, R., Schulman, J., Sutskever,I., and Abbeel, P. Infogan: Interpretable representationlearning by information maximizing generative adversar-ial nets. In Advances in Neural Information ProcessingSystems, pp. 2172–2180, 2016.

Cheung, B., Livezey, J. A., Bansal, A. K., and Olshausen,B. A. Discovering hidden factors of variation in deepnetworks. In Workshop at International Conference onLearning Representations, 2015.

Cohen, T. and Welling, M. Learning the irreducible repre-sentations of commutative lie groups. In InternationalConference on Machine Learning, pp. 1755–1763, 2014.

Cohen, T. S. and Welling, M. Transformation properties oflearned visual representations. In International Confer-ence on Learning Representations, 2015.

Comon, P. Independent component analysis, a new concept?Signal processing, 36(3):287–314, 1994.

Denton, E. L. and Birodkar, v. Unsupervised learning ofdisentangled representations from video. In Advances inNeural Information Processing Systems, pp. 4414–4423,2017.

Desjardins, G., Courville, A., and Bengio, Y. Disentan-gling factors of variation via generative entangling. arXivpreprint arXiv:1210.5474, 2012.

Eastwood, C. and Williams, C. K. I. A framework for thequantitative evaluation of disentangled representations.In International Conference on Learning Representations,2018.

Fraccaro, M., Kamronn, S., Paquet, U., and Winther, O. Adisentangled recognition and nonlinear dynamics modelfor unsupervised learning. In Advances in Neural Infor-mation Processing Systems, pp. 3601–3610, 2017.

Goodfellow, I., Lee, H., Le, Q. V., Saxe, A., and Ng, A. Y.Measuring invariances in deep networks. In Advancesin neural information processing systems, pp. 646–654,2009.

Goroshin, R., Mathieu, M. F., and LeCun, Y. Learningto linearize under uncertainty. In Advances in NeuralInformation Processing Systems, pp. 1234–1242, 2015.

Higgins, I., Matthey, L., Pal, A., Burgess, C., Glorot, X.,Botvinick, M., Mohamed, S., and Lerchner, A. beta-vae: Learning basic visual concepts with a constrainedvariational framework. In International Conference onLearning Representations, 2017a.

Higgins, I., Pal, A., Rusu, A., Matthey, L., Burgess, C.,Pritzel, A., Botvinick, M., Blundell, C., and Lerchner,A. Darla: Improving zero-shot transfer in reinforcementlearning. In International Conference on Machine Learn-ing, pp. 1480–1490, 2017b.

Higgins, I., Sonnerat, N., Matthey, L., Pal, A., Burgess, C. P.,Bošnjak, M., Shanahan, M., Botvinick, M., Hassabis, D.,and Lerchner, A. Scan: Learning hierarchical composi-tional visual concepts. In International Conference onLearning Representations, 2018.

Hinton, G. E., Krizhevsky, A., and Wang, S. D. Trans-forming auto-encoders. In International Conference onArtificial Neural Networks, pp. 44–51. Springer, 2011.

Page 10: Challenging Common Assumptions in the Unsupervised ... · first theoretically show that the unsupervised learning of disentangled representations is fundamentally impossible without

Challenging Common Assumptions in the Unsupervised Learning of Disentangled Representations

Hsu, W.-N., Zhang, Y., and Glass, J. Unsupervised learningof disentangled and interpretable representations fromsequential data. In Advances in Neural Information Pro-cessing Systems, pp. 1878–1889, 2017.

Hyvarinen, A. and Morioka, H. Unsupervised feature ex-traction by time-contrastive learning and nonlinear ica.In Advances in Neural Information Processing Systems,pp. 3765–3773, 2016.

Hyvarinen, A. and Pajunen, P. Nonlinear independent com-ponent analysis: Existence and uniqueness results. NeuralNetworks, 12(3):429–439, 1999.

Hyvarinen, A., Sasaki, H., and Turner, R. E. Nonlinearica using auxiliary variables and generalized contrastivelearning. arXiv preprint arXiv:1805.08651, 2018.

Jutten, C. and Karhunen, J. Advances in nonlinear blindsource separation. In Proc. of the 4th Int. Symp. on Inde-pendent Component Analysis and Blind Signal Separation(ICA2003), pp. 245–256, 2003.

Karaletsos, T., Belongie, S., and Rätsch, G. Bayesian repre-sentation learning with oracle constraints. arXiv preprintarXiv:1506.05011, 2015.

Kim, H. and Mnih, A. Disentangling by factorising. InProceedings of the 35th International Conference on Ma-chine Learning, pp. 2649–2658, 2018.

Kingma, D. P. and Welling, M. Auto-encoding variationalbayes. In International Conference on Learning Repre-sentations, 2014.

Kulkarni, T. D., Whitney, W. F., Kohli, P., and Tenenbaum,J. Deep convolutional inverse graphics network. InAdvances in neural information processing systems, pp.2539–2547, 2015.

Kumar, A., Sattigeri, P., and Balakrishnan, A. Variationalinference of disentangled latent concepts from unlabeledobservations. In International Conference on LearningRepresentations, 2017.

Lake, B. M., Ullman, T. D., Tenenbaum, J. B., and Gersh-man, S. J. Building machines that learn and think likepeople. Behavioral and Brain Sciences, 40, 2017.

Laversanne-Finot, A., Pere, A., and Oudeyer, P.-Y. Curiositydriven exploration of learned disentangled goal spaces.In Conference on Robot Learning, pp. 487–504, 2018.

LeCun, Y., Huang, F. J., and Bottou, L. Learning methodsfor generic object recognition with invariance to pose andlighting. In Computer Vision and Pattern Recognition,2004. CVPR 2004. Proceedings of the 2004 IEEE Com-puter Society Conference on, volume 2, pp. II–104. IEEE,2004.

LeCun, Y., Bengio, Y., and Hinton, G. Deep learning. nature,521(7553):436, 2015.

Lenc, K. and Vedaldi, A. Understanding image representa-tions by measuring their equivariance and equivalence. InIEEE conference on computer vision and pattern recogni-tion, pp. 991–999, 2015.

Locatello, F., Vincent, D., Tolstikhin, I., Rätsch, G., Gelly,S., and Schölkopf, B. Competitive training of mixturesof independent deep generative models. arXiv preprintarXiv:1804.11130, 2018.

Mathieu, M. F., Zhao, J. J., Zhao, J., Ramesh, A., Sprech-mann, P., and LeCun, Y. Disentangling factors of varia-tion in deep representation using adversarial training. InAdvances in Neural Information Processing Systems, pp.5040–5048, 2016.

Munch, E. The scream, 1893.

Nair, A. V., Pong, V., Dalal, M., Bahl, S., Lin, S., andLevine, S. Visual reinforcement learning with imaginedgoals. In Advances in Neural Information ProcessingSystems, pp. 9209–9220, 2018.

Narayanaswamy, S., Paige, T. B., Van de Meent, J.-W.,Desmaison, A., Goodman, N., Kohli, P., Wood, F., andTorr, P. Learning disentangled representations with semi-supervised deep generative models. In Advances inNeural Information Processing Systems, pp. 5925–5935,2017.

Nguyen, X., Wainwright, M. J., and Jordan, M. I. Estimatingdivergence functionals and the likelihood ratio by convexrisk minimization. IEEE Transactions on InformationTheory, 56(11):5847–5861, 2010.

Pearl, J. Causality. Cambridge university press, 2009.

Pedregosa, F., Varoquaux, G., Gramfort, A., Michel, V.,Thirion, B., Grisel, O., Blondel, M., Prettenhofer, P.,Weiss, R., Dubourg, V., Vanderplas, J., Passos, A., Cour-napeau, D., Brucher, M., Perrot, M., and Duchesnay, E.Scikit-learn: Machine learning in Python. Journal ofMachine Learning Research, 12:2825–2830, 2011.

Peters, J., Janzing, D., and Schölkopf, B. Elements of causalinference: foundations and learning algorithms. MITpress, 2017.

Reed, S., Sohn, K., Zhang, Y., and Lee, H. Learning todisentangle factors of variation with manifold interaction.In International Conference on Machine Learning, pp.1431–1439, 2014.

Reed, S. E., Zhang, Y., Zhang, Y., and Lee, H. Deep visualanalogy-making. In Advances in Neural InformationProcessing Systems, pp. 1252–1260, 2015.

Page 11: Challenging Common Assumptions in the Unsupervised ... · first theoretically show that the unsupervised learning of disentangled representations is fundamentally impossible without

Challenging Common Assumptions in the Unsupervised Learning of Disentangled Representations

Ridgeway, K. and Mozer, M. C. Learning deep disentan-gled embeddings with the f-statistic loss. In Advancesin Neural Information Processing Systems, pp. 185–194,2018.

Rubenstein, P. K., Schoelkopf, B., and Tolstikhin, I. Learn-ing disentangled representations with wasserstein auto-encoders. In Workshop at International Conference onLearning Representations, 2018.

Schmidhuber, J. Learning factorial codes by predictabilityminimization. Neural Computation, 4(6):863–879, 1992.

Schölkopf, B., Janzing, D., Peters, J., Sgouritsa, E., Zhang,K., and Mooij, J. On causal and anticausal learning.In International Conference on Machine Learning, pp.1255–1262, 2012.

Spirtes, P., Glymour, C., and Scheines, R. Causation, pre-diction, and search. Springer-Verlag. (2nd edition MITPress 2000), 1993.

Steenbrugge, X., Leroux, S., Verbelen, T., and Dhoedt, B.Improving generalization for abstract reasoning tasks us-ing disentangled feature representations. In Workshopon Relational Representation Learning at Conference onNeural Information Processing Systems, 2018.

Sugiyama, M., Suzuki, T., and Kanamori, T. Density-ratiomatching under the bregman divergence: a unified frame-work of density-ratio estimation. Annals of the Instituteof Statistical Mathematics, 64(5):1009–1044, 2012.

Suter, R., Miladinovic, Ð., Bauer, S., and Schölkopf, B.Interventional robustness of deep latent variable models.arXiv preprint arXiv:1811.00007, 2018.

Thomas, V., Bengio, E., Fedus, W., Pondard, J., Beaudoin,P., Larochelle, H., Pineau, J., Precup, D., and Bengio,Y. Disentangling the independently controllable factorsof variation by interacting with the world. In Workshopon Learning Disentangled Representations at the 31stConference on Neural Information Processing Systems,2017.

Tschannen, M., Bachem, O., and Lucic, M. Recent advancesin autoencoder-based representation learning. arXivpreprint arXiv:1812.05069, 2018.

Watanabe, S. Information theoretical analysis of multivari-ate correlation. IBM Journal of research and development,4(1):66–82, 1960.

Whitney, W. F., Chang, M., Kulkarni, T., and Tenenbaum,J. B. Understanding visual concepts with continuationlearning. In Workshop at International Conference onLearning Representations, 2016.

Yang, J., Reed, S. E., Yang, M.-H., and Lee, H. Weakly-supervised disentangling with recurrent transformationsfor 3d view synthesis. In Advances in Neural InformationProcessing Systems, pp. 1099–1107, 2015.

Yingzhen, L. and Mandt, S. Disentangled sequential autoen-coder. In International Conference on Machine Learning,pp. 5656–5665, 2018.

Zhu, Z., Luo, P., Wang, X., and Tang, X. Multi-view per-ceptron: a deep model for learning face identity and viewrepresentations. In Advances in Neural Information Pro-cessing Systems, pp. 217–225, 2014.

Page 12: Challenging Common Assumptions in the Unsupervised ... · first theoretically show that the unsupervised learning of disentangled representations is fundamentally impossible without

Challenging Common Assumptions in the Unsupervised Learning of Disentangled Representations

A. Proof of Theorem 1Proof. To show the claim, we explicitly construct a family of functions f using a sequence of bijective functions. Let d > 1be the dimensionality of the latent variable z and consider the function g : supp(z)→ [0, 1]d defined by

gi(v) = P (zi ≤ vi) ∀i = 1, 2, . . . , d.

Since P admits a density p(z) =∏i p(zi), the function g is bijective and, for almost every v ∈ supp(z), it holds that

∂gi(v)∂vi

6= 0 for all i and ∂gi(v)∂vj

= 0 for all i 6= j. Furthermore, it is easy to see that, by construction, g(z) is a independentd-dimensional uniform distribution. Similarly, consider the function h : (0, 1]d → Rd defined by

hi(v) = ψ−1(vi) ∀i = 1, 2, . . . , d,

where ψ(·) denotes the cumulative density function of a standard normal distribution. Again, by definition, h is bijectivewith ∂hi(v)

∂vi6= 0 for all i and ∂hi(v)

∂vj= 0 for all i 6= j. Furthermore, the random variable h(g(z)) is a d-dimensional standard

normal distribution.

Let A ∈ Rd×d be an arbitrary orthogonal matrix with Aij 6= 0 for all i = 1, 2, . . . , d and j = 1, 2, . . . , d. An infinite familyof such matrices can be constructed using a Householder transformation: Choose an arbitrary α ∈ (0, 0.5) and consider

the vector v with v1 =√α and vi =

√1−αd−1 for i = 2, 3, . . . , d. By construction, we have vTv = 1 and both vi 6= 0 and

vi 6=√

12 for all i = 1, 2, . . . , d. Define the matrix A = Id − 2vvT and note that Aii = 1− 2v2i 6= 0 for all 1, 2, . . . , d as

well as Aij = −vivj 6= 0 for all i 6= j. Furthermore, A is orthogonal since

ATA =(Id − 2vvT

)T (Id − 2vvT

)= Id − 4vvT + 4v(vTv)vT = Id.

Since A is orthogonal, it is invertible and thus defines a bijective linear operator. The random variable Ah(g(z)) ∈ Rd ishence an independent, multivariate standard normal distribution since the covariance matrix ATA is equal to Id.

Since h is bijective, it follows that h−1(Ah(g(z))) is an independent d-dimensional uniform distribution. Define thefunction f : supp(z)→ supp(z)

f(u) = g−1(h−1(Ah(g(u))))

and note that by definition f(z) has the same marginal distribution as z under P , i.e., P (z ≤ u) = P (f(z) ≤ u) for all u.Finally, for almost every u ∈ supp(z), it holds that

∂fi(u)

∂uj=

Aij · ∂hj(g(u))∂vj

· ∂gj(u)∂uj

∂hi(h−1i (Ah(g(u)))

∂vi· ∂gi(g

−1(h−1(Ah(g(u)))))∂vi

6= 0,

as claimed. Since the choice of A was arbitrary, there exists an infinite family of such functions f .

B. Unsupervised learning of disentangled representations with VAEsVariants of variational autoencoders (Kingma & Welling, 2014) are considered the state-of-the-art for unsupervised

disentanglement learning. One assumes a specific prior P (z) on the latent space and then parameterizes the conditionalprobability P (x|z) with a deep neural network. Similarly, the distribution P (z|x) is approximated using a variationaldistribution Q(z|x), again parametrized using a deep neural network. One can then derive the following approximation tothe maximum likelihood objective,

maxφ,θ

Ep(x)[Eqφ(z|x)[log pθ(x|z)]−DKL(qφ(z|x)‖p(z))] (1)

which is also know as the evidence lower bound (ELBO). By carefully considering the KL term, one can encourage variousproperties of the resulting presentation. We will briefly review the main approaches.

Page 13: Challenging Common Assumptions in the Unsupervised ... · first theoretically show that the unsupervised learning of disentangled representations is fundamentally impossible without

Challenging Common Assumptions in the Unsupervised Learning of Disentangled Representations

Bottleneck capacity. Higgins et al. (2017a) propose the β-VAE, introducing a hyperparameter in front of the KLregularizer of vanilla VAEs. They maximize the following expression:

Ep(x)[Eqφ(z|x)[log pθ(x|z)]− βDKL(qφ(z|x)‖p(z))]

By setting β > 1, the encoder distribution will be forced to better match the factorized unit Gaussian prior. This procedureintroduces additional constraints on the capacity of the latent bottleneck, encouraging the encoder to learn a disentangledrepresentation for the data. Burgess et al. (2017) argue that when the bottleneck has limited capacity, the network will beforced to specialize on the factor of variation that most contributes to a small reconstruction error. Therefore, they propose toprogressively increase the bottleneck capacity, so that the encoder can focus on learning one factor of variation at the time:

Ep(x)[Eqφ(z|x)[log pθ(x|z)]− γ|DKL(qφ(z|x)‖p(z))− C|]

where C is annealed from zero to some value which is large enough to produce good reconstruction. In the following, werefer to this model as AnnealedVAE.

Penalizing the total correlation. Let I(x; z) denote the mutual information between x and z and note that the secondterm in equation 1 can be rewritten as

Ep(x)[DKL(qφ(z|x)‖p(z))] = I(x; z) +DKL(q(z)‖p(z)).

Therefore, when β > 1, β-VAE penalizes the mutual information between the latent representation and the data, thusconstraining the capacity of the latent space. Furthermore, it pushes q(z), the so called aggregated posterior, to matchthe prior and therefore to factorize, given a factorized prior. Kim & Mnih (2018) argues that penalizing I(x; z) is neithernecessary nor desirable for disentanglement. The FactorVAE (Kim & Mnih, 2018) and the β-TCVAE (Chen et al., 2018)augment the VAE objective with an additional regularizer that specifically penalizes dependencies between the dimensionsof the representation:

Ep(x)[Eqφ(z|x)[log pθ(x|z)]−DKL(qφ(z|x)‖p(z))]− γDKL(q(z)‖d∏j=1

q(zj)).

This last term is also known as total correlation (Watanabe, 1960). The total correlation is intractable and vanilla Monte Carloapproximations require marginalization over the training set. (Kim & Mnih, 2018) propose an estimate using the densityratio trick (Nguyen et al., 2010; Sugiyama et al., 2012) (FactorVAE). Samples from

∏dj=1 q(zj) can be obtained shuffling

samples from q(z) (Arcones & Gine, 1992). Concurrently, Chen et al. (2018) propose a tractable biased Monte-Carloestimate for the total correlation (β-TCVAE).

Disentangled priors. Kumar et al. (2017) argue that a disentangled generative model requires a disentangled prior. Thisapproach is related to the total correlation penalty, but now the aggregated posterior is pushed to match a factorized prior.Therefore

Ep(x)[Eqφ(z|x)[log pθ(x|z)]−DKL(qφ(z|x)‖p(z))]− λD(q(z)‖p(z)),

where D is some (arbitrary) divergence. Since this term is intractable when D is the KL divergence, they propose to matchthe moments of these distribution. In particular, they regularize the deviation of either Covp(x)[µφ(x)] or Covqφ [z] from theidentity matrix in the two variants of the DIP-VAE. This results in maximizing either the DIP-VAE-I objective

Ep(x)[Eqφ(z|x)[log pθ(x|z)]−DKL(qφ(z|x)‖p(z))]− λod∑i 6=j

[Covp(x)[µφ(x)]

]2ij− λd

∑i

([Covp(x)[µφ(x)]

]ii− 1)2

or the DIP-VAE-II objective

Ep(x)[Eqφ(z|x)[log pθ(x|z)]−DKL(qφ(z|x)‖p(z))]− λod∑i6=j

[Covqφ [z]

]2ij− λd

∑i

([Covqφ [z]

]ii− 1)2.

Page 14: Challenging Common Assumptions in the Unsupervised ... · first theoretically show that the unsupervised learning of disentangled representations is fundamentally impossible without

Challenging Common Assumptions in the Unsupervised Learning of Disentangled Representations

C. Implementation of metricsAll our metrics consider the expected representation of training samples (except total correlation for which we also considerthe sampled representation as described in Section 5).

BetaVAE metric. Higgins et al. (2017a) suggest to fix a random factor of variation in the underlying generative model andto sample two mini batches of observations x. Disentanglement is then measured as the accuracy of a linear classifier thatpredicts the index of the fixed factor based on the coordinate-wise sum of absolute differences between the representationvectors in the two mini batches. We sample two batches of 64 points with a random factor fixed to a randomly sampled valueacross the two batches and the others varying randomly. We compute the mean representations for these points and take theabsolute difference between pairs from the two batches. We then average these 64 values to form the features of a training (ortesting) point. We train a Scikit-learn logistic regression with default parameters on 10 000 points. We test on 5000 points.

FactorVAE metric Kim & Mnih (2018) address several issues with this metric by using a majority vote classifier thatpredicts the index of the fixed ground-truth factor based on the index of the representation vector with the least variance.First, we estimate the variance of each latent dimension by embedding 10 000 random samples from the data set and weexclude collapsed dimensions with variance smaller than 0.05. Second, we generate the votes for the majority vote classifierby sampling a batch of 64 points, all with a factor fixed to the same random value. Third, we compute the variance of eachdimension of their latent representation and divide by the variance of that dimension we computed on the data withoutinterventions. The training point for the majority vote classifier consists of the index of the dimension with the smallestnormalized variance. We train on 10 000 points and evaluate on 5000 points.

Mutual Information Gap. Chen et al. (2018) argue that the BetaVAE metric and the FactorVAE metric are neither generalnor unbiased as they depend on some hyperparameters. They compute the mutual information between each ground truthfactor and each dimension in the computed representation r(x). For each ground-truth factor zk, they then consider the twodimensions in r(x) that have the highest and second highest mutual information with zk. The Mutual Information Gap (MIG)is then defined as the average, normalized difference between the highest and second highest mutual information of eachfactor with the dimensions of the representation. The original metric was proposed evaluating the sampled representation.Instead, we consider the mean representation, in order to be consistent with the other metrics. We estimate the discretemutual information by binning each dimension of the representations obtained from 10 000 points into 20 bins. Then, thescore is computed as follows:

1

K

K∑k=1

1

Hzk

(I(vjk , zk)−max

j 6=jkI(vj , zk)

)Where zk is a factor of variation, vj is a dimension of the latent representation and jk = arg maxj I(vj , zk).

Modularity. Ridgeway & Mozer (2018) argue that two different properties of representations should be considered, i.e.,Modularity and Explicitness. In a modular representation each dimension of r(x) depends on at most a single factor ofvariation. In an explicit representation, the value of a factor of variation is easily predictable (i.e. with a linear model) fromr(x). They propose to measure the Modularity as the average normalized squared difference of the mutual information ofthe factor of variations with the highest and second-highest mutual information with a dimension of r(x). They measureExplicitness as the ROC-AUC of a one-versus-rest logistic regression classifier trained to predict the factors of variation. Inthis study, we focus on Modularity as it is the property that corresponds to disentanglement. For the modularity score, wesample 10 000 points for which we obtain the latent representations. We discretize these points into 20 bins and compute themutual information between representations and the values of the factors of variation. These values are stored in a matrix m.For each dimension of the representation i, we compute a vector ti as:

ti,f =

{θi if f = arg maxgmi,g

0 otherwise

where θi = maxgmig . The modularity score is the average over the dimensions of the representation of 1− δi where:

δi =

∑f (mif − tif )2

θ2i (N − 1)

and N is the number of factors.

Page 15: Challenging Common Assumptions in the Unsupervised ... · first theoretically show that the unsupervised learning of disentangled representations is fundamentally impossible without

Challenging Common Assumptions in the Unsupervised Learning of Disentangled Representations

DCI Disentanglement. Eastwood & Williams (2018) consider three properties of representations, i.e., Disentanglement,Completeness and Informativeness. First, Eastwood & Williams (2018) compute the importance of each dimension ofthe learned representation for predicting a factor of variation. The predictive importance of the dimensions of r(x) canbe computed with a Lasso or a Random Forest classifier. Disentanglement is the average of the difference from one ofthe entropy of the probability that a dimension of the learned representation is useful for predicting a factor weighted bythe relative importance of each dimension. Completeness, is the average of the difference from one of the entropy of theprobability that a factor of variation is captured by a dimension of the learned representation. Finally, the Informativenesscan be computed as the prediction error of predicting the factors of variations. We sample 10 000 and 5000 training andtest points respectively. For each factor, we fit gradient boosted trees from Scikit-learn with the default setting. Fromthis model, we extract the importance weights for the feature dimensions. We take the absolute value of these weightsand use them to form the importance matrix R, whose rows correspond to factors and columns to the representation.To compute the disentanglement score, we first subtract from 1 the entropy of each column of this matrix (we treat thecolumns as a distribution by normalizing them). This gives a vector of length equal to the dimensionality of the latent space.Then, we compute the relative importance of each dimension by ρi =

∑j Rij/

∑ij Rij and the disentanglement score as∑

i ρi(1−H(Ri)).

SAP score. Kumar et al. (2017) propose to compute the R2 score of the linear regression predicting the factor values fromeach dimension of the learned representation. For discrete factors, they propose to train a classifier. The Separated AttributePredictability (SAP) score is the average difference of the prediction error of the two most predictive latent dimensionsfor each factor. We sample 10 000 points for training and 5000 for testing. We then compute a score matrix containingthe prediction error on the test set for a linear SVM with C = 0.01 predicting the value of a factor from a single latentdimension. The SAP score is computed as the average across factors of the difference between the top two most predictivelatent dimensions.

Downstream task. We sample training sets of different sizes: 10, 100, 1000 and 10 000 points. We always evaluate on5000 samples. We consider as a downstream task the prediction of the values of each factor from r(x). For each factorwe fit a different model and report then report the average test accuracy across factors. We consider two different models.First, we train a cross validated logistic regression from Scikit-learn with 10 different values for the regularization strength(Cs = 10) and 5 folds. Finally, we train a gradient boosting classifier from Scikit-learn with default parameters.

Total correlation based on fitted Gaussian. We sample 10 000 points and obtain their latent representation r(x) byeither sampling from the encoder distribution of by taking its mean. We then compute the mean µr(x) and covariance matrixΣr(x) of these points and compute the total correlation of a Gaussian with mean µr(x) and covariance matrix Σr(x), i.e.,

DKL

N (µr(x),Σr(x))∥∥∥∏

j

N (µr(x)j ,Σr(x)jj )

where j indexes the dimensions in the latent space. We choose this approach for the following reasons: In this study, wecompute statistics of r(x) which can be either sampled from the probabilistic encoder or taken to be its mean. We argue thatestimating the total correlation as in (Kim & Mnih, 2018) is not suitable for this comparison as it consistently underestimatethe true value (see Figure 7 in (Kim & Mnih, 2018)) and depends on a non-convex optimization procedure (for fittingthe discriminator). The estimate of (Chen et al., 2018) is also not suitable as the mean representation is a deterministicfunction for the data, therefore we cannot use the encoder distribution for the estimate. Furthermore, we argue that the totalcorrelation based on the fitted Gaussian provides a simple and robust way to detect if a representation is not factorizingbased on the first two moments. In particular, if it is high, it is a strong signal that the representation is not factorizing (whilea low score may not imply the opposite). We note that this procedure is similar to the penalty of DIP-VAE-I. Therefore, it isnot surprising that DIP-VAE-I achieves a low score for the mean representation.

Page 16: Challenging Common Assumptions in the Unsupervised ... · first theoretically show that the unsupervised learning of disentangled representations is fundamentally impossible without

Challenging Common Assumptions in the Unsupervised Learning of Disentangled Representations

D. Experimental conditions and guiding principles.In our study, we seek controlled, fair and reproducible experimental conditions. We consider the case in which we can

sample from a well defined and known ground-truth generative model by first sampling the factors of variations from adistribution P (z) and then sampling an observation from P (x|z). Our experimental protocol works as follows: Duringtraining, we only observe the samples of x obtained by marginalizing P (x|z) over P (z). After training, we obtain arepresentation r(x) by either taking a sample from the probabilistic encoder Q(z|x) or by taking its mean. Typically,disentanglement metrics consider the latter as the representation r(x). During the evaluation, we assume to have access tothe whole generative model, i.e. we can draw samples from both P (z) and P (x|z). In this way, we can perform interventionson the latent factors as required by certain evaluation metrics. We explicitly note that we effectively consider the statisticallearning problem where we optimize the loss and the metrics on the known data generating distribution. As a result, we donot use separate train and test sets but always take i.i.d. samples from the known ground-truth distribution. This is justifiedas the statistical problem is well defined and it allows us to remove the additional complexity of dealing with overfitting andempirical risk minimization.

E. Limitations of our study.While we aim to provide a useful and fair experimental study, there are clear limitations to the conclusions that can be drawnfrom it due to design choices that we have taken. In all these choices, we have aimed to capture what is considered thestate-of-the-art inductive bias in the community.

On the data set side, we only consider images with a heavy focus on synthetic images. We do not explore other modalitiesand we only consider the toy scenario in which we have access to a data generative process with uniformly distributedfactors of variations. Furthermore, all our data sets have a small number of independent discrete factors of variations withoutany confounding variables.

For the methods, we only consider the inductive bias of convolutional architectures. We do not test fully connectedarchitectures or additional techniques such as skip connections. Furthermore, we do not explore different activationfunctions, reconstruction losses or different number of layers. We also do not vary any other hyperparameters other than theregularization weight. In particular, we do not evaluate the role of different latent space sizes, optimizers and batch sizes.We do not test the sample efficiency of the metrics but simply set the size of the train and test set to large values.

Implementing the different disentanglement methods and metrics has proven to be a difficult endeavour. Few “official” opensource implementations are available and there are many small details to consider. We take a best-effort approach to theseimplementations and implemented all the methods and metrics from scratch as any sound machine learning practitionermight do based on the original papers. When taking different implementation choices than the original papers, we explicitlystate and motivate them.

F. Differences with previous implementations.As described above, we use a single choice of architecture, batch size and optimizer for all the methods which might deviatefrom the settings considered in the original papers. However, we argue that unification of these choices is the only way toguarantee a fair comparison among the different methods such that valid conclusions may be drawn in between methods.The largest change is that for DIP-VAE and for β-TCVAE we used a batch size of 64 instead of 400 and 2048 respectively.However, Chen et al. (2018) shows in Section H.2 of the Appendix that the bias in the mini-batch estimation of the totalcorrelation does not significantly affect the performances of their model even with small batch sizes. For DIP-VAE-II, wedid not implement the additional regularizer on the third order central moments since no implementation details are providedand since this regularizer is only used on specific data sets.

Our implementations of the disentanglement metrics deviate from the implementations in the original papers as follows:First, we strictly enforce that all factors of variations are treated as discrete variables as this corresponds to the assumedground-truth model in all our data sets. Hence, we used classification instead of regression for the SAP score and thedisentanglement score of (Eastwood & Williams, 2018). This is important as it does not make sense to use regression ontrue factors of variations that are discrete (e.g., shape on dSprites). Second, wherever possible, we resorted to using thedefault, well-tested Scikit-learn (Pedregosa et al., 2011) implementations instead of using custom implementations withpotentially hard to set hyperparameters. Third, for the Mutual Information Gap (Chen et al., 2018), we estimate the discrete

Page 17: Challenging Common Assumptions in the Unsupervised ... · first theoretically show that the unsupervised learning of disentangled representations is fundamentally impossible without

Challenging Common Assumptions in the Unsupervised Learning of Disentangled Representations

Table 2. Encoder and Decoder architecture for the main experiment.

Encoder Decoder

Input: 64× 64× number of channels Input: R10

4× 4 conv, 32 ReLU, stride 2 FC, 256 ReLU4× 4 conv, 32 ReLU, stride 2 FC, 4× 4× 64 ReLU4× 4 conv, 64 ReLU, stride 2 4× 4 upconv, 64 ReLU, stride 24× 4 conv, 64 ReLU, stride 2 4× 4 upconv, 32 ReLU, stride 2FC 256, F2 2× 10 4× 4 upconv, 32 ReLU, stride 2

4× 4 upconv, number of channels, stride 2

Table 3. Model’s hyperparameters. We allow a sweep over a single hyperparameter for each model.

Model Parameter Values

β-VAE β [1, 2, 4, 6, 8, 16]AnnealedVAE cmax [5, 10, 25, 50, 75, 100]

iteration threshold 100000γ 1000

FactorVAE γ [10, 20, 30, 40, 50, 100]DIP-VAE-I λod [1, 2, 5, 10, 20, 50]

λd 10λodDIP-VAE-II λod [1, 2, 5, 10, 20, 50]

λd λodβ-TCVAE β [1, 2, 4, 6, 8, 10]

mutual information (as opposed to continuous) on the mean representation (as opposed to sampled) on a subset of thesamples (as opposed to the whole data set). We argue that this is the correct choice as the mean is usually taken to be therepresentation. Hence, it would be wrong to consider the full Gaussian encoder or samples thereof as that would correspondto a different representation. Finally, we fix the number of sampled train and test points across all metrics to a large value toensure robustness.

G. Main experiment hyperparametersIn our study, we fix all hyperparameters except one per each model. Model specific hyperparameters can be found in Table 3.The common architecture is depicted in Table 2 along with the other fixed hyperparameters in Table 4a. For the discriminatorin FactorVAE we use the architecture in Table 4b with hyperparameters in Table 4c. All the hyperparameters for which wereport single values were not varied and are selected based on the literature.

H. Data sets and preprocessingAll the data sets contains images with pixels between 0 and 1. Color-dSprites: Every time we sample a point, we alsosample a random scaling for each channel uniformly between 0.5 and 1. Noisy-dSprites: Every time we sample a point, wefill the background with uniform noise. Scream-dSprites: Every time we sample a point, we sample a random 64 × 64patch of The Scream painting. We then change the color distribution by adding a random uniform number to each channeland divide the result by two. Then, we embed the dSprites shape by inverting the colors of each of its pixels.

Page 18: Challenging Common Assumptions in the Unsupervised ... · first theoretically show that the unsupervised learning of disentangled representations is fundamentally impossible without

Challenging Common Assumptions in the Unsupervised Learning of Disentangled Representations

Table 4. Other fixed hyperparameters.

(a) Hyperparameters common to each ofthe considered methods.

Parameter Values

Batch size 64Latent space dimension 10Optimizer AdamAdam: beta1 0.9Adam: beta2 0.999Adam: epsilon 1e-8Adam: learning rate 0.0001Decoder type BernoulliTraining steps 300000

(b) Architecture for the discriminator inFactorVAE.

Discriminator

FC, 1000 leaky ReLUFC, 1000 leaky ReLUFC, 1000 leaky ReLUFC, 1000 leaky ReLUFC, 1000 leaky ReLUFC, 1000 leaky ReLUFC, 2

(c) Parameters for the discriminator inFactorVAE.

Parameter Values

Batch size 64Optimizer AdamAdam: beta1 0.5Adam: beta2 0.9Adam: epsilon 1e-8Adam: learning rate 0.0001

I. Detailed experimental resultsGiven the breadth of the experimental study, we summarized our key findings in Section 5 and presented figures that wepicked to be representative of our results. This section contains a self-contained presentation of all our experimental results.In particular, we present a complete set of plots for the different methods, data sets and disentanglement metrics.

I.1. Can one achieve a good reconstruction error across data sets and models?

First, we check for each data set that we manage to train models that achieve reasonable reconstructions. Therefore, for eachdata set we sample a random model and show real samples next to their reconstructions. The results are depicted in Figure 7.As expected, the additional variants of dSprites with continuous noise variables are harder than the original data set. OnNoisy-dSprites and Color-dSprites the models produce reasonable reconstructions with the noise on Noisy-dSprites beingignored. Scream-dSprites is even harder and we observe that the shape information is lost. On the other data sets, we observethat reconstructions are blurry but objects are distinguishable. SmallNORB seems to be the most challenging data set.

I.2. Can current methods enforce a uncorrelated aggregated posterior and representation?

We investigate whether the considered unsupervised disentanglement approaches are effective at enforcing a factorizingand thus uncorrelated aggregated posterior. For each trained model, we sample 10 000 images and compute a sample fromthe corresponding approximate posterior. We then fit a multivariate Gaussian distribution over these 10 000 samples bycomputing the empirical mean and covariance matrix. Finally, we compute the total correlation of the fitted Gaussian andreport the median value for each data set, method and hyperparameter value.

Figure 8 shows the total correlation of the sampled representation plotted against the regularization strength for eachdata set and method except AnnealedVAE. On all data sets except SmallNORB, we observe that plain vanilla variationalautoencoders (i.e. the β-VAE model with β = 1) exhibit the highest total correlation. For β-VAE and β-TCVAE, it can beclearly seen that the total correlation of the sampled representation decreases on all data sets as the regularization strength (inthe form of β) is increased. The two variants of DIP-VAE exhibit low total correlation across the data sets except DIP-VAE-Iwhich incurs a slightly higher total correlation on SmallNORB compared to a vanilla VAE. Increased regularization in theDIP-VAE objective also seems to lead a reduced total correlation, even if the effect is not as pronounced as for β-VAE andβ-TCVAE. While FactorVAE achieves a low total correlation on all data sets except on SmallNORB, we observe that thetotal correlation does not seem to decrease with increasing regularization strength. We further observe that AnnealedVAE(shown in Figure 25) is much more sensitive to the regularization strength. However, on all data sets except Scream-dSprites(on which AnnealedVAE performs poorly), the total correlation seems to decrease with increased regularization strength.

While many of the considered methods aim to enforce a factorizing aggregated posterior, they use the mean vector of theGaussian encoder as the representation and not a sample from the Gaussian encoder. This may seem like a minor, irrelevantmodification; however, it is not clear whether a factorizing aggregated posterior also ensures that the dimensions of themean representation are uncorrelated. To test whether this is true, we compute the mean of the Gaussian encoder for thesame 10 000 samples, fit a multivariate Gaussian and compute the total correlation of that fitted Gaussian. Figure 9 shows

Page 19: Challenging Common Assumptions in the Unsupervised ... · first theoretically show that the unsupervised learning of disentangled representations is fundamentally impossible without

Challenging Common Assumptions in the Unsupervised Learning of Disentangled Representations

(a) DIP-VAE-I trained on dSprites. (b) β-VAE trained on Noisy-dSprites.

(c) FactorVAE trained on Color-dSprites. (d) FactorVAE trained on Scream-dSprites.

(e) AnneaeledVAE trained on Shapes3D. (f) β-TCVAE trained on SmallNORB.

(g) Reconstructions for a DIP-VAE-II trained on Cars3D.

Figure 7. Reconstructions for different data sets and methods. Odd columns show real samples and even columns their reconstruction. Asexpected, the additional variants of dSprites with continuous noise variables are harder than the original data set. On Noisy-dSprites andColor-dSprites the models produce reasonable reconstructions with the noise on Noisy-dSprites being ignored. Scream-dSprites is evenharder and we observe that the shape information is lost. On the other data sets, we observe that reconstructions are blurry but objects aredistinguishable.

Page 20: Challenging Common Assumptions in the Unsupervised ... · first theoretically show that the unsupervised learning of disentangled representations is fundamentally impossible without

Challenging Common Assumptions in the Unsupervised Learning of Disentangled Representations

0.0 0.2 0.4 0.6 0.8 1.00.00

0.01

0.02

0.03

0.04

0.05

TC

(sa

mple

d)

Dataset = dSprites

Model VAE β-VAE FactorVAE β-TCVAE DIP-VAE-I DIP-VAE-II

0.0 0.2 0.4 0.6 0.8 1.00.00

0.02

0.04

0.06

0.08

0.10

0.12Dataset = Color-dSprites

0.0 0.2 0.4 0.6 0.8 1.00.000

0.005

0.010

0.015

0.020

0.025

0.030

0.035

0.040Dataset = Noisy-dSprites

0.0 0.2 0.4 0.6 0.8 1.0Regularization strength

0.00

0.01

0.02

0.03

0.04

0.05

0.06Dataset = Scream-dSprites

0.0 0.2 0.4 0.6 0.8 1.0Regularization strength

0.002

0.004

0.006

0.008

0.010

0.012

0.014

TC

(sa

mple

d)

Dataset = SmallNORB

0.0 0.2 0.4 0.6 0.8 1.0Regularization strength

0.00

0.05

0.10

0.15

0.20

0.25

0.30Dataset = Cars3D

0.0 0.2 0.4 0.6 0.8 1.0Regularization strength

0.00

0.05

0.10

0.15

0.20

0.25Dataset = Shapes3D

Figure 8. Total correlation of sampled representation plotted against regularization strength for different data sets and approaches (exceptAnnealedVAE). The total correlation of the sampled representation decreases as the regularization strength is increased.

0.0 0.2 0.4 0.6 0.8 1.00.0

0.2

0.4

0.6

0.8

1.0

1.2

1.4

TC

(m

ean)

Dataset = dSprites

Model VAE β-VAE FactorVAE β-TCVAE DIP-VAE-I DIP-VAE-II

0.0 0.2 0.4 0.6 0.8 1.00.0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8Dataset = Color-dSprites

0.0 0.2 0.4 0.6 0.8 1.00.0

0.2

0.4

0.6

0.8

1.0

1.2

1.4

1.6Dataset = Noisy-dSprites

0.0 0.2 0.4 0.6 0.8 1.0Regularization strength

0.0

0.1

0.2

0.3

0.4

0.5

0.6Dataset = Scream-dSprites

0.0 0.2 0.4 0.6 0.8 1.0Regularization strength

0.0

0.5

1.0

1.5

2.0

2.5

3.0

3.5

4.0

TC

(m

ean)

Dataset = SmallNORB

0.0 0.2 0.4 0.6 0.8 1.0Regularization strength

0.0

0.2

0.4

0.6

0.8

1.0

1.2Dataset = Cars3D

0.0 0.2 0.4 0.6 0.8 1.0Regularization strength

0.0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8Dataset = Shapes3D

Figure 9. Total correlation of mean representation plotted against regularization strength for different data sets and approaches (exceptAnnealedVAE). The total correlation of the mean representation does not necessarily decrease as the regularization strength is increased.

Page 21: Challenging Common Assumptions in the Unsupervised ... · first theoretically show that the unsupervised learning of disentangled representations is fundamentally impossible without

Challenging Common Assumptions in the Unsupervised Learning of Disentangled Representations

7 6 5 4 3 2 18

6

4

2

0

2

4

log T

C (

mean)

Dataset = dSprites

Model β-VAE FactorVAE β-TCVAE DIP-VAE-I DIP-VAE-II AnnealedVAE

7 6 5 4 3 2 17

6

5

4

3

2

1

0

1

2Dataset = Color-dSprites

8 6 4 2 0 2 48

6

4

2

0

2

4Dataset = Noisy-dSprites

7 6 5 4 3 2 1log TC (sampled)

6

4

2

0

2

Dataset = Scream-dSprites

8 6 4 2 0 2 4log TC (sampled)

6

4

2

0

2

4

6

log T

C (

mean)

Dataset = SmallNORB

8 6 4 2 0 2 4log TC (sampled)

6

4

2

0

2

4Dataset = Cars3D

7 6 5 4 3 2 1 0 1log TC (sampled)

8

6

4

2

0

2

4Dataset = Shapes3D

Figure 10. Log total correlation of mean vs sampled representations. For a large number of models, the total correlation of the meanrepresentation is higher than that of the sampled representation.

the total correlation of the mean representation plotted against the regularization strength for each data set and methodexcept AnnealedVAE. We observe that, for β-VAE and β-TCVAE, increased regularization leads to a substantially increasedtotal correlation of the mean representations. This effect can also be observed for for FactorVAE, albeit in a less extremefashion. For DIP-VAE-I, we observe that the total correlation of the mean representation is consistently low. This is notsurprising as the DIP-VAE-I objective directly optimizes the covariance matrix of the mean representation to be diagonalwhich implies that the corresponding total correlation (as we compute it) is low. The DIP-VAE-II objective which enforcesthe covariance matrix of the sampled representation to be diagonal seems to lead to a factorized mean representation onsome data sets (for example Shapes3D and Cars3d), but also seems to fail on others (dSprites). For AnnealedVAE (shownin Figure 26), we overall observe mean representations with a very high total correlation. In Figure 10, we further plotthe log total correlations of the sampled representations versus the mean representations for each of the trained models.It can be clearly seen that for a large number of models, the total correlation of the mean representations is much higherthan that of the sampled representations. The same trend can be seen computing the average discrete mutual informationof the representation. In this case, the DIP-VAE-I exhibit increasing mutual information in both the mean and sampledrepresentation. This is to be expected as DIP-VAE-I enforces a variance of one for the mean representation. We remark thatas the regularization terms and hyperparameter values are different for different losses, one should not draw conclusionsfrom comparing different models at nominally the same regularization strength. From these plots one can only compare theeffect of increasing the regularization in the different models.

Implications. Overall, these results lead us to conclude with minor exceptions that the considered methods are effective atenforcing an aggregated posterior whose individual dimensions are not correlated but that this does not seem to imply thatthe dimensions of the mean representation (usually used for representation) are uncorrelated.

I.3. How much do existing disentanglement metrics agree?

As there exists no single, common definition of disentanglement, an interesting question is to see how much differentproposed metrics agree. Figure 11 shows pairwise scatter plots of the different considered metrics on dSprites whereeach point corresponds to a trained model, while Figure 12 shows the Spearman rank correlation between differentdisentanglement metrics on different data sets. Overall, we observe that all metrics except Modularity seem to be correlatedstrongly on the data sets dSprites, Color-dSprites and Scream-dSprites and mildly on the other data sets. There appear to betwo pairs among these metrics that capture particularly similar notions: the BetaVAE and the FactorVAE score as well as the

Page 22: Challenging Common Assumptions in the Unsupervised ... · first theoretically show that the unsupervised learning of disentangled representations is fundamentally impossible without

Challenging Common Assumptions in the Unsupervised Learning of Disentangled Representations

0.4

0.8

Beta

VA

E S

core

0.5

1.0

Fact

orV

AE S

core

0.0

MIG

0.0

0.6

DC

I D

isenta

ngle

ment

0.6

0.8

1.0

Modula

rity

0.5 1.0BetaVAE Score

0.0

SA

P

0.6FactorVAE Score

0.0 0.6MIG

0.0DCI Disentanglement

0.8Modularity

0.0 0.1SAP

Model

β-VAE

FactorVAEβ-TCVAE

DIP-VAE-I

DIP-VAE-II

AnnealedVAE

Figure 11. Pairwise scatter plots of different disentanglement metrics on dSprites. All the metrics except Modularity appear to becorrelated. The strongest correlation seems to be between MIG and DCI Disentanglement.

Page 23: Challenging Common Assumptions in the Unsupervised ... · first theoretically show that the unsupervised learning of disentangled representations is fundamentally impossible without

Challenging Common Assumptions in the Unsupervised Learning of Disentangled Representations

Beta

VA

E S

core

Fact

orV

AE S

core

MIG

DC

I D

isenta

ngle

ment

Modula

rity

SA

P

BetaVAE Score

FactorVAE Score

MIG

DCI Disentanglement

Modularity

SAP

100 82 81 84 1 81

82 100 72 77 -5 67

81 72 100 93 -5 83

84 77 93 100 -14 84

1 -5 -5 -14 100 -10

81 67 83 84 -10 100

Dataset = dSprites

Beta

VA

E S

core

Fact

orV

AE S

core

MIG

DC

I D

isenta

ngle

ment

Modula

rity

SA

P

BetaVAE Score

FactorVAE Score

MIG

DCI Disentanglement

Modularity

SAP

100 76 80 87 9 79

76 100 68 76 2 62

80 68 100 91 1 82

87 76 91 100 -7 82

9 2 1 -7 100 -7

79 62 82 82 -7 100

Dataset = Color-dSprites

Beta

VA

E S

core

Fact

orV

AE S

core

MIG

DC

I D

isenta

ngle

ment

Modula

rity

SA

P

BetaVAE Score

FactorVAE Score

MIG

DCI Disentanglement

Modularity

SAP

100 80 44 41 46 37

80 100 49 52 25 38

44 49 100 76 6 42

41 52 76 100 -8 38

46 25 6 -8 100 13

37 38 42 38 13 100

Dataset = Noisy-dSprites

Beta

VA

E S

core

Fact

orV

AE S

core

MIG

DC

I D

isenta

ngle

ment

Modula

rity

SA

P

BetaVAE Score

FactorVAE Score

MIG

DCI Disentanglement

Modularity

SAP

100 95 94 92 47 95

95 100 95 89 49 90

94 95 100 95 50 87

92 89 95 100 44 87

47 49 50 44 100 41

95 90 87 87 41 100

Dataset = Scream-dSprites

Beta

VA

E S

core

Fact

orV

AE S

core

MIG

DC

I D

isenta

ngle

ment

Modula

rity

SA

P

BetaVAE Score

FactorVAE Score

MIG

DCI Disentanglement

Modularity

SAP

100 63 56 45 -60 39

63 100 34 39 -69 2

56 34 100 79 -50 76

45 39 79 100 -42 58

-60 -69 -50 -42 100 -28

39 2 76 58 -28 100

Dataset = SmallNORB

Beta

VA

E S

core

Fact

orV

AE S

core

MIG

DC

I D

isenta

ngle

ment

Modula

rity

SA

P

BetaVAE Score

FactorVAE Score

MIG

DCI Disentanglement

Modularity

SAP

100 37 22 38 29 13

37 100 66 60 39 40

22 66 100 67 52 43

38 60 67 100 57 27

29 39 52 57 100 18

13 40 43 27 18 100

Dataset = Cars3D

Beta

VA

E S

core

Fact

orV

AE S

core

MIG

DC

I D

isenta

ngle

ment

Modula

rity

SA

P

BetaVAE Score

FactorVAE Score

MIG

DCI Disentanglement

Modularity

SAP

100 76 24 54 46 57

76 100 44 49 23 66

24 44 100 67 9 70

54 49 67 100 52 58

46 23 9 52 100 17

57 66 70 58 17 100

Dataset = Shapes3D

Figure 12. Rank correlation of different metrics on different data sets. Overall, we observe that all metrics except Modularity seem to bestrongly correlated on the data sets dSprites, Color-dSprites and Scream-dSprites and mildly on the other data sets. There appear to betwo pairs among these metrics that capture particularly similar notions: the BetaVAE and the FactorVAE score as well as the MutualInformation Gap and DCI Disentanglement.

Mutual Information Gap and DCI Disentanglement.

Implication. All disentanglement metrics except Modularity appear to be correlated. However, the level of correlationchanges between different data sets.

I.4. How important are different models and hyperparameters for disentanglement?

The primary motivation behind the considered methods is that they should lead to improved disentanglement scores. Thisraises the question how disentanglement is affected by the model choice, the hyperparameter selection and randomness (inthe form of different random seeds). To investigate this, we compute all the considered disentanglement metrics for each ofour trained models. In Figure 13, we show the range of attainable disentanglement scores for each method on each dataset. We observe that these ranges are heavily overlapping for different models leading us to (qualitatively) conclude thatthe choice of hyperparameters and the random seed seems to be substantially more important than the choice of objectivefunction. While certain models seem to attain better maximum scores on specific data sets and disentanglement metrics,we do not observe any consistent pattern that one model is consistently better than the other. Furthermore, we note that inour study we have fixed the range of hyperparameters a priori to six different values for each model and did not exploreadditional hyperparameters based on the results (as that would bias our study). However, this also means that specificmodels may have performed better than in Figure 13 if we had chosen a different set of hyperparameters. In Figure 14, wefurther show the impact of randomness in the form of random seeds on the disentanglement scores. Each violin plot showsthe distribution of the disentanglement metric across all 50 trained models for each model and hyperparameter setting onCars3D. We clearly see that randomness (in the form of different random seeds) has a substantial impact on the attainedresult and that a good run with a bad hyperparameter can beat a bad run with a good hyperparameter in many cases.

Finally, we perform a variance analysis by trying to predict the different disentanglement scores using ordinary least squaresfor each data set: If we allow the score to depend only on the objective function (categorical variable), we are only able toexplain 37% of the variance of the scores on average. Similarly, if the score depends on the Cartesian product of objectivefunction and regularization strength (again categorical), we are able to explain 59% of the variance while the rest is dueto the random seed. In Table 5, we report the percentage of variance explained for the different metrics in each data set

Page 24: Challenging Common Assumptions in the Unsupervised ... · first theoretically show that the unsupervised learning of disentangled representations is fundamentally impossible without

Challenging Common Assumptions in the Unsupervised Learning of Disentangled Representations

0.4

0.5

0.6

0.7

0.8

0.9

1.0

Valu

e

Metric = BetaVAE Score

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1.0 Metric = FactorVAE Score

0.1

0.0

0.1

0.2

0.3

0.4

0.5 Metric = MIG

0.1

0.0

0.1

0.2

0.3

0.4

0.5

0.6 Metric = DCI Disentanglement

0.60

0.65

0.70

0.75

0.80

0.85

0.90

0.95

1.00 Metric = Modularity

0.02

0.00

0.02

0.04

0.06

0.08

0.10

Data

set =

dSprite

s

Metric = SAP

0.5

0.6

0.7

0.8

0.9

Valu

e

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1.0

0.05

0.00

0.05

0.10

0.15

0.20

0.25

0.30

0.35

0.40

0.1

0.0

0.1

0.2

0.3

0.4

0.5

0.6

0.70

0.75

0.80

0.85

0.90

0.95

1.00

0.02

0.00

0.02

0.04

0.06

0.08

0.10

Data

set =

Colo

r-dSprite

s

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1.0

Valu

e

0.1

0.0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.05

0.00

0.05

0.10

0.15

0.20

0.25

0.30

0.05

0.00

0.05

0.10

0.15

0.20

0.25

0.30

0.35

0.5

0.6

0.7

0.8

0.9

1.0

0.01

0.00

0.01

0.02

0.03

0.04

0.05

0.06

0.07

Data

set =

Noisy

-dSprite

s

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

Valu

e

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.05

0.00

0.05

0.10

0.15

0.20

0.25

0.30

0.05

0.00

0.05

0.10

0.15

0.20

0.25

0.30

0.35

0.55

0.60

0.65

0.70

0.75

0.80

0.85

0.90

0.95

0.01

0.00

0.01

0.02

0.03

0.04

0.05

0.06

0.07

0.08

Data

set =

Scre

am

-dSprite

s

0.4

0.5

0.6

0.7

0.8

0.9

1.0

Valu

e

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

0.05

0.00

0.05

0.10

0.15

0.20

0.25

0.30

0.35

0.00

0.05

0.10

0.15

0.20

0.25

0.30

0.35

0.40

0.65

0.70

0.75

0.80

0.85

0.90

0.95

1.00

1.05

0.05

0.00

0.05

0.10

0.15

0.20

Data

set =

Sm

allN

OR

B

0.965

0.970

0.975

0.980

0.985

0.990

0.995

1.000

1.005

Valu

e

0.60

0.65

0.70

0.75

0.80

0.85

0.90

0.95

1.00

0.05

0.00

0.05

0.10

0.15

0.20

0.25

0.30

0.0

0.1

0.2

0.3

0.4

0.5

0.6

0.70

0.75

0.80

0.85

0.90

0.95

1.00

0.01

0.00

0.01

0.02

0.03

0.04

0.05

0.06

Data

set =

Cars3

D

0 1 2 3 4 5Model

0.4

0.5

0.6

0.7

0.8

0.9

1.0

1.1

Valu

e

0 1 2 3 4 5Model

0.2

0.4

0.6

0.8

1.0

0 1 2 3 4 5Model

0.2

0.0

0.2

0.4

0.6

0.8

1.0

0 1 2 3 4 5Model

0.0

0.2

0.4

0.6

0.8

1.0

1.2

0 1 2 3 4 5Model

0.65

0.70

0.75

0.80

0.85

0.90

0.95

1.00

1.05

0 1 2 3 4 5Model

0.05

0.00

0.05

0.10

0.15

0.20

0.25

0.30

Data

set =

Shapes3

D

Figure 13. Score for each method for each score (column) and data set (row). Models are abbreviated (0=β-VAE, 1=FactorVAE, 2=β-TCVAE, 3=DIP-VAE-I, 4=DIP-VAE-II, 5=AnnealedVAE). The scores are heavily overlapping and we do not observe a consistent pattern.We conclude that hyperparameters matter more than the model choice.

Page 25: Challenging Common Assumptions in the Unsupervised ... · first theoretically show that the unsupervised learning of disentangled representations is fundamentally impossible without

Challenging Common Assumptions in the Unsupervised Learning of Disentangled Representations

0.985

0.990

0.995

1.000

Valu

e

Metric = BetaVAE Score

0.60

0.65

0.70

0.75

0.80

0.85

0.90

0.95

1.00 Metric = FactorVAE Score

0.05

0.00

0.05

0.10

0.15

0.20 Metric = MIG

0.05

0.10

0.15

0.20

0.25

0.30

0.35

0.40

0.45

0.50 Metric = DCI Disentanglement

0.75

0.80

0.85

0.90

0.95

1.00 Metric = Modularity

0.01

0.00

0.01

0.02

0.03

0.04

0.05

Model =

β-VA

E

Metric = SAP

0.965

0.970

0.975

0.980

0.985

0.990

0.995

1.000

1.005

Valu

e

0.60

0.65

0.70

0.75

0.80

0.85

0.90

0.95

1.00

0.00

0.02

0.04

0.06

0.08

0.10

0.12

0.14

0.16

0.10

0.15

0.20

0.25

0.30

0.35

0.40

0.45

0.50

0.80

0.82

0.84

0.86

0.88

0.90

0.92

0.94

0.96

0.98

0.01

0.00

0.01

0.02

0.03

0.04

0.05

0.06

Model =

Facto

rVA

E

0.9980

0.9985

0.9990

0.9995

1.0000

1.0005

Valu

e

0.65

0.70

0.75

0.80

0.85

0.90

0.95

1.00

0.05

0.00

0.05

0.10

0.15

0.20

0.25

0.30

0.0

0.1

0.2

0.3

0.4

0.5

0.6

0.75

0.80

0.85

0.90

0.95

1.00

0.005

0.000

0.005

0.010

0.015

0.020

0.025

0.030

0.035

Model =

β-TC

VA

E

0.0000

0.0001

0.0002

0.0003

0.0004

0.0005

0.0006

Valu

e

+9.995e 1

0.65

0.70

0.75

0.80

0.85

0.90

0.95

1.00

0.00

0.02

0.04

0.06

0.08

0.10

0.12

0.14

0.16

0.05

0.10

0.15

0.20

0.25

0.30

0.35

0.40

0.45

0.70

0.75

0.80

0.85

0.90

0.95

0.01

0.00

0.01

0.02

0.03

0.04

0.05

0.06

Model =

DIP

-VA

E-I

0.9988

0.9990

0.9992

0.9994

0.9996

0.9998

1.0000

1.0002

Valu

e

0.60

0.65

0.70

0.75

0.80

0.85

0.90

0.95

1.00

0.02

0.00

0.02

0.04

0.06

0.08

0.10

0.12

0.14

0.16

0.00

0.05

0.10

0.15

0.20

0.25

0.30

0.35

0.40

0.70

0.75

0.80

0.85

0.90

0.95

1.00

0.01

0.00

0.01

0.02

0.03

0.04

0.05

Model =

DIP

-VA

E-II

0.0 0.2 0.4 0.6 0.8 1.0Regularization strength

0.970

0.975

0.980

0.985

0.990

0.995

1.000

1.005

Valu

e

0.0 0.2 0.4 0.6 0.8 1.0Regularization strength

0.65

0.70

0.75

0.80

0.85

0.90

0.95

1.00

0.0 0.2 0.4 0.6 0.8 1.0Regularization strength

0.02

0.00

0.02

0.04

0.06

0.08

0.10

0.12

0.14

0.0 0.2 0.4 0.6 0.8 1.0Regularization strength

0.00

0.05

0.10

0.15

0.20

0.25

0.30

0.35

0.0 0.2 0.4 0.6 0.8 1.0Regularization strength

0.70

0.75

0.80

0.85

0.90

0.95

1.00

0.0 0.2 0.4 0.6 0.8 1.0Regularization strength

0.005

0.000

0.005

0.010

0.015

0.020

0.025

0.030

0.035

Model =

Anneale

dV

AE

Figure 14. Distribution of scores for different models, hyperparameters and regularization strengths on Cars3D. We clearly see thatrandomness (in the form of different random seeds) has a substantial impact on the attained result and that a good run with a badhyperparameter can beat a bad run with a good hyperparameter in many cases.

Page 26: Challenging Common Assumptions in the Unsupervised ... · first theoretically show that the unsupervised learning of disentangled representations is fundamentally impossible without

Challenging Common Assumptions in the Unsupervised Learning of Disentangled Representations

Table 5. Variance of the disentanglement scores explained by the objective function or its cartesian product with the hyperparameters. Thevariance explained is computed regressing using ordinary least squares.

(a) Percentage of variance explained regressing the disentanglement scores on the different data sets from the objective function only.

BetaVAE Score DCI Disentanglement FactorVAE Score MIG Modularity SAP

Cars3D 1% 36% 26% 34% 37% 13%Color-dSprites 30% 39% 52% 26% 23% 29%Noisy-dSprites 17% 21% 17% 11% 41% 6%Scream-dSprites 89% 50% 76% 45% 60% 56%Shapes3D 31% 21% 14% 20% 26% 10%SmallNORB 68% 71% 58% 71% 62% 62%dSprites 29% 41% 47% 26% 29% 31%

(b) Percentage of variance explained regressing the disentanglement scores on the different data sets from the Cartesian product ofobjective function and regularization strength.

BetaVAE Score DCI Disentanglement FactorVAE Score MIG Modularity SAP

Cars3D 4% 69% 42% 59% 51% 17%Color-dSprites 69% 80% 61% 76% 40% 56%Noisy-dSprites 26% 42% 25% 29% 50% 20%Scream-dSprites 93% 74% 83% 66% 68% 75%Shapes3D 61% 78% 53% 59% 49% 35%SmallNORB 87% 89% 81% 88% 72% 82%dSprites 59% 77% 54% 72% 39% 56%

considering the regularization strength or not.

Implication. The disentanglement scores of unsupervised models are heavily influenced by randomness (in the form ofthe random seed) and the choice of the hyperparameter (in the form of the regularization strength). The objective functionappears to have less impact.

I.5. Are there reliable recipes for model selection?

In this section, we investigate how good hyperparameters can be chosen and how we can distinguish between good andbad training runs. In this paper, we advocate that model selection should not depend on the considered disentanglementscore for the following reasons: The point of unsupervised learning of disentangled representation is that there is no accessto the labels as otherwise we could incorporate them and would have to compare to semi-supervised and fully supervisedmethods. All the disentanglement metrics considered in this paper require a substantial amount of ground-truth labels or thefull generative model (for example for the BetaVAE and the FactorVAE metric). Hence, one may substantially bias theresults of a study by tuning hyperparameters based on (supervised) disentanglement metrics. Furthermore, we argue thatit is not sufficient to fix a set of hyperparameters a priori and then show that one of those hyperparameters and a specificrandom seed achieves a good disentanglement score as it amounts to showing the existence of a good model, but does notguide the practitioner in finding it. Finally, in many practical settings, we might not even have access to adequate labels as itmay be hard to identify the true underlying factor of variations, in particular, if we consider data modalities that are lesssuitable to human interpretation than images.

In the remainder of this section, we hence investigate and assess different ways how hyperparameters and good model runscould be chosen. In this study, we focus on choosing the learning model and the regularization strength corresponding tothat loss function. However, we note that in practice this problem is likely even harder as a practitioner might also want totune other modeling choices such architecture or optimizer.

General recipes for hyperparameter selection. We first investigate whether we may find generally applicable “rules ofthumb” for choosing the hyperparameters. For this, we plot in Figure 15 different disentanglement metrics against differentregularization strengths for each model and each data set. The values correspond to the median obtained values across

Page 27: Challenging Common Assumptions in the Unsupervised ... · first theoretically show that the unsupervised learning of disentangled representations is fundamentally impossible without

Challenging Common Assumptions in the Unsupervised Learning of Disentangled Representations

Table 6. Probability of outperforming random model selection on a different random seed. A random disentanglement metric and data setis sampled and used for model selection. That model is then compared to a randomly selected model: (i) on the same metric and data set,(ii) on the same metric and a random different data set, (iii) on a random different metric and the same data set, and (iv) on a randomdifferent metric and a random different data set. The results are averaged across 10 000 random draws.

Random different data set Same data set

Random different metric 54.9% 62.6%Same metric 59.3% 80.7%

50 random seeds for each model, hyperparameter and data set. There seems to be no model dominating all the othersand for each model there does not seem to be a consistent strategy in choosing the regularization strength to maximizedisentanglement scores. Furthermore, even if we could identify a good objective function and corresponding hyperparametervalue, we still could not distinguish between a good and a bad training run.

Model selection based on unsupervised scores. Another approach could be to select hyperparameters based on unsu-pervised scores such as the reconstruction error, the KL divergence between the prior and the approximate posterior, theEvidence Lower Bound or the estimated total correlation of the sampled representation. This would have the advantagethat we could select specific trained models and not just good hyperparameter settings whose median trained model wouldperform well. To test whether such an approach is fruitful, we compute the rank correlation between these unsupervisedmetrics and the disentanglement metrics and present it in Figure 16. While we do observe some correlations, no clear patternemerges which leads us to conclude that this approach is unlikely to be successful in practice.

Hyperparameter selection based on transfer. The final strategy for hyperparameter selection that we consider is basedon transferring good settings across data sets. The key idea is that good hyperparameter settings may be inferred on data setswhere we have labels available (such as dSprites) and then applied to novel data sets. To test this idea, we plot in Figure 18the different disentanglement scores obtained on dSprites against the scores obtained on other data sets. To ensure robustnessof the results, we again consider the median across all 50 runs for each model, regularization strength, and data set. Weobserve that the scores on Color-dSprites seem to be strongly correlated with the scores obtained on the regular version ofdSprites. Figure 17 further shows the rank correlations obtained between different data sets for each disentanglement scores.This confirms the strong and consistent correlation between dSprites and Color-dSprites. While these result suggest thatsome transfer of hyperparameters is possible, it does not allow us to distinguish between good and bad random seeds on thetarget data set.

To illustrate this, we compare such a transfer based approach to hyperparameter selection to random model selection asfollows: We first randomly sample one of our 50 random seeds and consider the set of trained models with that random seed.First, we sample one of our 50 random seeds, a random disentanglement metric and a data set and use them to select thehyperparameter setting with the highest attained score. Then, we compare that selected hyperparameter setting to a randomlyselected model on either the same or a random different data set, based on either the same or a random different metricand for a randomly sampled seed. Finally, we report the percentage of trials in which this transfer strategy outperforms orperforms equally well as random model selection across 10 000 trials in Table 6. If we choose the same metric and the samedata set (but a different random seed), we obtain a score of 80.7%. If we aim to transfer for the same metric across data sets,we achieve around 59.3%. Finally, if we transfer both across metrics and data sets, our performance drops to 54.9%.

Implications. Unsupervised model selection remains an unsolved problem. Transfer of good hyperparameters betweenmetrics and data sets does not seem to work as there appears to be no unsupervised way to distinguish between good andbad random seeds on the target task.

I.6. Are these disentangled representations useful for downstream tasks in terms of the sample complexity oflearning?

One of the key motivations behind disentangled representations is that they are assumed to be useful for later downstreamtasks. In particular, it is argued that disentanglement should lead to a better sample complexity of learning (Bengio et al.,2013; Schölkopf et al., 2012; Peters et al., 2017). In this section, we consider the simplest downstream classification task

Page 28: Challenging Common Assumptions in the Unsupervised ... · first theoretically show that the unsupervised learning of disentangled representations is fundamentally impossible without

Challenging Common Assumptions in the Unsupervised Learning of Disentangled Representations

0.0 0.2 0.4 0.6 0.8 1.00.65

0.70

0.75

0.80

0.85

Valu

e

Metric = BetaVAE Score

Model β-VAE FactorVAE β-TCVAE DIP-VAE-I DIP-VAE-II AnnealedVAE

0.0 0.2 0.4 0.6 0.8 1.00.45

0.50

0.55

0.60

0.65

0.70

0.75

0.80 Metric = FactorVAE Score

0.0 0.2 0.4 0.6 0.8 1.00.00

0.05

0.10

0.15

0.20

0.25

0.30

0.35

0.40 Metric = MIG

0.0 0.2 0.4 0.6 0.8 1.00.0

0.1

0.2

0.3

0.4

0.5 Metric = DCI Disentanglement

0.0 0.2 0.4 0.6 0.8 1.00.80

0.82

0.84

0.86

0.88

0.90 Metric = Modularity

0.0 0.2 0.4 0.6 0.8 1.00.00

0.01

0.02

0.03

0.04

0.05

0.06

0.07

0.08

Data

set =

dSprite

s

Metric = SAP

0.0 0.2 0.4 0.6 0.8 1.00.60

0.65

0.70

0.75

0.80

0.85

0.90

Valu

e

0.0 0.2 0.4 0.6 0.8 1.00.50

0.55

0.60

0.65

0.70

0.75

0.80

0.0 0.2 0.4 0.6 0.8 1.00.00

0.05

0.10

0.15

0.20

0.25

0.30

0.35

0.40

0.0 0.2 0.4 0.6 0.8 1.00.00

0.05

0.10

0.15

0.20

0.25

0.30

0.35

0.40

0.45

0.0 0.2 0.4 0.6 0.8 1.00.80

0.82

0.84

0.86

0.88

0.90

0.92

0.0 0.2 0.4 0.6 0.8 1.00.00

0.01

0.02

0.03

0.04

0.05

0.06

0.07

0.08

Data

set =

Colo

r-dSprite

s

0.0 0.2 0.4 0.6 0.8 1.00.60

0.65

0.70

0.75

Valu

e

0.0 0.2 0.4 0.6 0.8 1.00.35

0.40

0.45

0.50

0.55

0.60

0.0 0.2 0.4 0.6 0.8 1.00.01

0.02

0.03

0.04

0.05

0.06

0.07

0.08

0.0 0.2 0.4 0.6 0.8 1.00.02

0.04

0.06

0.08

0.10

0.12

0.14

0.16

0.0 0.2 0.4 0.6 0.8 1.0

0.75

0.80

0.85

0.0 0.2 0.4 0.6 0.8 1.00.004

0.006

0.008

0.010

0.012

0.014

0.016

0.018

0.020

Data

set =

Noisy

-dSprite

s

0.0 0.2 0.4 0.6 0.8 1.00.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

Valu

e

0.0 0.2 0.4 0.6 0.8 1.00.2

0.3

0.4

0.5

0.6

0.7

0.8

0.0 0.2 0.4 0.6 0.8 1.00.00

0.05

0.10

0.15

0.20

0.25

0.0 0.2 0.4 0.6 0.8 1.00.00

0.05

0.10

0.15

0.20

0.25

0.0 0.2 0.4 0.6 0.8 1.0

0.65

0.70

0.75

0.80

0.0 0.2 0.4 0.6 0.8 1.00.00

0.01

0.02

0.03

0.04

0.05

0.06

Data

set =

Scre

am

-dSprite

s

0.0 0.2 0.4 0.6 0.8 1.00.50

0.55

0.60

0.65

0.70

0.75

0.80

0.85

0.90

Valu

e

0.0 0.2 0.4 0.6 0.8 1.00.30

0.35

0.40

0.45

0.50

0.55

0.60

0.65

0.70

0.0 0.2 0.4 0.6 0.8 1.00.00

0.05

0.10

0.15

0.20

0.25

0.30

0.0 0.2 0.4 0.6 0.8 1.00.10

0.15

0.20

0.25

0.30

0.35

0.40

0.0 0.2 0.4 0.6 0.8 1.00.75

0.80

0.85

0.90

0.95

1.00

0.0 0.2 0.4 0.6 0.8 1.00.02

0.04

0.06

0.08

0.10

0.12

0.14

0.16

Data

set =

Sm

allN

OR

B

0.0 0.2 0.4 0.6 0.8 1.00.00000

0.00005

0.00010

0.00015

0.00020

0.00025

Valu

e

+9.998e 1

0.0 0.2 0.4 0.6 0.8 1.00.70

0.75

0.80

0.85

0.90

0.95

0.0 0.2 0.4 0.6 0.8 1.00.02

0.04

0.06

0.08

0.10

0.12

0.14

0.16

0.0 0.2 0.4 0.6 0.8 1.00.10

0.15

0.20

0.25

0.30

0.35

0.40

0.45

0.0 0.2 0.4 0.6 0.8 1.00.82

0.84

0.86

0.88

0.90

0.92

0.94

0.96

0.0 0.2 0.4 0.6 0.8 1.00.006

0.008

0.010

0.012

0.014

0.016

0.018

0.020

Data

set =

Cars3

D

0.0 0.2 0.4 0.6 0.8 1.0Regularization strength

0.65

0.70

0.75

0.80

0.85

0.90

0.95

1.00

Valu

e

0.0 0.2 0.4 0.6 0.8 1.0Regularization strength

0.4

0.5

0.6

0.7

0.8

0.9

1.0

0.0 0.2 0.4 0.6 0.8 1.0Regularization strength

0.0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.0 0.2 0.4 0.6 0.8 1.0Regularization strength

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

0.0 0.2 0.4 0.6 0.8 1.0Regularization strength

0.82

0.84

0.86

0.88

0.90

0.92

0.94

0.96

0.98

0.0 0.2 0.4 0.6 0.8 1.0Regularization strength

0.03

0.04

0.05

0.06

0.07

0.08

0.09

0.10

0.11

Data

set =

Shapes3

D

Figure 15. Score vs hyperparameters for each score (column) and data set (row). There seems to be no model dominating all the othersand for each model there does not seem to be a consistent strategy in choosing the regularization strength.

Page 29: Challenging Common Assumptions in the Unsupervised ... · first theoretically show that the unsupervised learning of disentangled representations is fundamentally impossible without

Challenging Common Assumptions in the Unsupervised Learning of Disentangled Representations

Beta

VA

E S

core

Fact

orV

AE S

core

MIG

DC

I D

isenta

ngle

ment

Modula

rity

SA

P

Reconstruction

TC (sampled)

KL

ELBO

2 3 37 28 -5 11

-1 22 5 11 -12 10

-57 -51 -73 -78 31 -56

-8 -5 28 17 2 4

Dataset = dSprites

Beta

VA

E S

core

Fact

orV

AE S

core

MIG

DC

I D

isenta

ngle

ment

Modula

rity

SA

P

Reconstruction

TC (sampled)

KL

ELBO

12 18 46 34 -6 20

11 35 7 18 -8 13

-68 -53 -73 -80 21 -61

5 13 40 26 -3 15

Dataset = Color-dSprites

Beta

VA

E S

core

Fact

orV

AE S

core

MIG

DC

I D

isenta

ngle

ment

Modula

rity

SA

P

Reconstruction

TC (sampled)

KL

ELBO

-4 -12 37 27 4 6

11 22 -2 6 -11 11

28 22 -43 -42 32 -6

13 -0 31 20 19 9

Dataset = Noisy-dSprites

Beta

VA

E S

core

Fact

orV

AE S

core

MIG

DC

I D

isenta

ngle

ment

Modula

rity

SA

P

Reconstruction

TC (sampled)

KL

ELBO

-9 -18 -15 -10 -35 -7

-21 -25 -25 -25 -17 -20

-53 -46 -47 -48 -9 -55

-14 -24 -20 -14 -34 -13

Dataset = Scream-dSprites

Beta

VA

E S

core

Fact

orV

AE S

core

MIG

DC

I D

isenta

ngle

ment

Modula

rity

SA

P

Reconstruction

TC (sampled)

KL

ELBO

-83 -73 -59 -54 69 -38

-12 41 -21 -2 -12 -46

20 49 -19 -18 -22 -39

-79 -57 -72 -65 61 -55

Dataset = SmallNORB

Beta

VA

E S

core

Fact

orV

AE S

core

MIG

DC

I D

isenta

ngle

ment

Modula

rity

SA

P

Reconstruction

TC (sampled)

KL

ELBO

-19 9 42 20 38 12

-18 -6 -3 -14 2 0

-14 -10 -46 -45 -40 -5

-24 5 35 12 29 11

Dataset = Cars3D

Beta

VA

E S

core

Fact

orV

AE S

core

MIG

DC

I D

isenta

ngle

ment

Modula

rity

SA

P

Reconstruction

TC (sampled)

KL

ELBO

-30 -4 59 22 -21 27

1 5 -11 -8 -11 -2

-14 -1 -38 -31 -11 -29

-38 -9 48 9 -25 15

Dataset = Shapes3D

Figure 16. Rank correlation between unsupervised scores and supervised disentanglement metrics. The unsupervised scores we considerdo not seem to be useful for model selection.

dSpri

tes

Colo

r-dSpri

tes

Nois

y-d

Spri

tes

Scr

eam

-dSpri

tes

Sm

allN

OR

B

Cars

3D

Shapes3

D

dSprites

Color-dSprites

Noisy-dSprites

Scream-dSprites

SmallNORB

Cars3D

Shapes3D

100 83 -22 54 -11 34 21

83 100 -24 69 -28 39 38

-22 -24 100 -24 -14 16 -16

54 69 -24 100 -6 63 72

-11 -28 -14 -6 100 -3 13

34 39 16 63 -3 100 65

21 38 -16 72 13 65 100

Metric = BetaVAE Score

dSpri

tes

Colo

r-dSpri

tes

Nois

y-d

Spri

tes

Scr

eam

-dSpri

tes

Sm

allN

OR

B

Cars

3D

Shapes3

DdSprites

Color-dSprites

Noisy-dSprites

Scream-dSprites

SmallNORB

Cars3D

Shapes3D

100 91 4 61 4 61 13

91 100 13 61 3 68 12

4 13 100 -20 62 34 -22

61 61 -20 100 -16 51 50

4 3 62 -16 100 30 -38

61 68 34 51 30 100 34

13 12 -22 50 -38 34 100

Metric = FactorVAE Score

dSpri

tes

Colo

r-dSpri

tes

Nois

y-d

Spri

tes

Scr

eam

-dSpri

tes

Sm

allN

OR

B

Cars

3D

Shapes3

D

dSprites

Color-dSprites

Noisy-dSprites

Scream-dSprites

SmallNORB

Cars3D

Shapes3D

100 94 75 59 3 80 58

94 100 75 54 -10 85 65

75 75 100 55 -23 74 68

59 54 55 100 28 64 18

3 -10 -23 28 100 -23 -48

80 85 74 64 -23 100 68

58 65 68 18 -48 68 100

Metric = MIG

dSpri

tes

Colo

r-dSpri

tes

Nois

y-d

Spri

tes

Scr

eam

-dSpri

tes

Sm

allN

OR

B

Cars

3D

Shapes3

D

dSprites

Color-dSprites

Noisy-dSprites

Scream-dSprites

SmallNORB

Cars3D

Shapes3D

100 95 65 65 34 64 46

95 100 61 60 21 63 47

65 61 100 68 17 64 59

65 60 68 100 36 93 69

34 21 17 36 100 21 -9

64 63 64 93 21 100 85

46 47 59 69 -9 85 100

Metric = DCI Disentanglement

dSpri

tes

Colo

r-dSpri

tes

Nois

y-d

Spri

tes

Scr

eam

-dSpri

tes

Sm

allN

OR

B

Cars

3D

Shapes3

D

dSprites

Color-dSprites

Noisy-dSprites

Scream-dSprites

SmallNORB

Cars3D

Shapes3D

100 86 78 -17 -6 -55 2

86 100 71 -11 -4 -32 30

78 71 100 3 5 -57 10

-17 -11 3 100 -45 -6 25

-6 -4 5 -45 100 21 -25

-55 -32 -57 -6 21 100 31

2 30 10 25 -25 31 100

Metric = Modularity

dSpri

tes

Colo

r-dSpri

tes

Nois

y-d

Spri

tes

Scr

eam

-dSpri

tes

Sm

allN

OR

B

Cars

3D

Shapes3

DdSprites

Color-dSprites

Noisy-dSprites

Scream-dSprites

SmallNORB

Cars3D

Shapes3D

100 86 27 63 17 51 35

86 100 39 65 10 43 43

27 39 100 -3 -34 21 2

63 65 -3 100 38 24 68

17 10 -34 38 100 -16 -2

51 43 21 24 -16 100 -4

35 43 2 68 -2 -4 100

Metric = SAP

Figure 17. Rank-correlation of different disentanglement metrics across different data sets. Good hyperparameters only seem to transferbetween dSprites and Color-dSprites but not in between the other data sets.

Page 30: Challenging Common Assumptions in the Unsupervised ... · first theoretically show that the unsupervised learning of disentangled representations is fundamentally impossible without

Challenging Common Assumptions in the Unsupervised Learning of Disentangled Representations

0.6 0.7 0.8 0.90.6

0.7

0.8

0.9

Metr

ic v

alu

e

Metric = BetaVAE Score

Model β-VAE FactorVAE β-TCVAE DIP-VAE-I DIP-VAE-II AnnealedVAE

0.45 0.60 0.75

0.45

0.60

0.75

Metric = FactorVAE Score

0.0 0.1 0.2 0.3 0.40.0

0.1

0.2

0.3

0.4 Metric = MIG

0.00 0.15 0.30 0.450.00

0.15

0.30

0.45

Metric = DCI Disentanglement

0.80 0.84 0.88 0.92

0.80

0.84

0.88

0.92 Metric = Modularity

0.000 0.025 0.050 0.0750.000

0.025

0.050

0.075 Data

set =

dSprite

s

Metric = SAP

0.6 0.7 0.8 0.90.6

0.7

0.8

0.9

Metr

ic v

alu

e

0.45 0.60 0.750.45

0.60

0.75

0.0 0.1 0.2 0.3 0.4

0.00

0.15

0.30

0.00 0.15 0.30 0.450.00

0.15

0.30

0.45

0.80 0.84 0.88 0.92

0.80

0.84

0.88

0.92

0.000 0.025 0.050 0.0750.000

0.025

0.050

0.075

Data

set =

Colo

r-dSprite

s

0.6 0.7 0.8 0.9

0.56

0.64

0.72

0.80

Metr

ic v

alu

e

0.45 0.60 0.75

0.40

0.48

0.56

0.0 0.1 0.2 0.3 0.40.000

0.025

0.050

0.075

0.00 0.15 0.30 0.45

0.04

0.08

0.12

0.16

0.80 0.84 0.88 0.92

0.72

0.80

0.88

0.000 0.025 0.050 0.0750.000

0.008

0.016

0.024

Data

set =

Noisy

-dSprite

s

0.6 0.7 0.8 0.9

0.25

0.50

0.75

Metr

ic v

alu

e

0.45 0.60 0.75

0.2

0.4

0.6

0.8

0.0 0.1 0.2 0.3 0.4

0.0

0.1

0.2

0.00 0.15 0.30 0.45

0.0

0.1

0.2

0.3

0.80 0.84 0.88 0.920.60

0.65

0.70

0.75

0.80

0.000 0.025 0.050 0.075

0.000

0.025

0.050

Data

set =

Scre

am

-dSprite

s

0.6 0.7 0.8 0.9

0.60

0.75

Metr

ic v

alu

e

0.45 0.60 0.750.30

0.45

0.60

0.0 0.1 0.2 0.3 0.4

0.0

0.1

0.2

0.3

0.00 0.15 0.30 0.45

0.1

0.2

0.3

0.4

0.80 0.84 0.88 0.92

0.80

0.88

0.96

0.000 0.025 0.050 0.0750.00

0.04

0.08

0.12

0.16

Data

set =

Sm

allN

OR

B

0.6 0.7 0.8 0.9

0.002

0.004

0.006

Metr

ic v

alu

e

+9.96e 1

0.45 0.60 0.75

0.72

0.80

0.88

0.0 0.1 0.2 0.3 0.4

0.04

0.08

0.12

0.16

0.00 0.15 0.30 0.450.1

0.2

0.3

0.4

0.5

0.80 0.84 0.88 0.92

0.84

0.88

0.92

0.96

0.000 0.025 0.050 0.0750.000

0.008

0.016

0.024

Data

set =

Cars3

D

0.6 0.7 0.8 0.9dSprites

0.75

0.90

Metr

ic v

alu

e

0.45 0.60 0.75dSprites

0.4

0.6

0.8

1.0

0.0 0.1 0.2 0.3 0.4dSprites

0.0

0.2

0.4

0.6

0.00 0.15 0.30 0.45dSprites

0.25

0.50

0.75

0.80 0.84 0.88 0.92dSprites

0.80

0.85

0.90

0.95

1.00

0.000 0.025 0.050 0.075dSprites

0.04

0.08

0.12

Data

set =

Shapes3

D

Figure 18. Disentanglement scores on dSprites vs other data sets. Good hyperparameters only seem to transfer consistently from dSpritesto Color-dSprites.

Page 31: Challenging Common Assumptions in the Unsupervised ... · first theoretically show that the unsupervised learning of disentangled representations is fundamentally impossible without

Challenging Common Assumptions in the Unsupervised Learning of Disentangled Representations

Beta

VA

E S

core

Fact

orV

AE S

core

MIG

DC

I D

isenta

ngle

ment

Modula

rity

SA

P

LR10

LR100

LR1000

LR10000

GBT10

GBT100

GBT1000

GBT10000

Efficiency (LR)

Efficiency (GBT)

18 13 18 19 -3 12

65 49 63 65 -9 64

28 13 20 18 15 20

28 12 -1 4 18 12

67 58 71 75 -6 71

78 73 86 94 -17 77

75 71 86 94 -19 74

76 71 87 94 -13 75

50 43 62 62 -19 56

50 46 47 54 -14 49

Dataset = dSprites

Beta

VA

E S

core

Fact

orV

AE S

core

MIG

DC

I D

isenta

ngle

ment

Modula

rity

SA

P

LR10

LR100

LR1000

LR10000

GBT10

GBT100

GBT1000

GBT10000

Efficiency (LR)

Efficiency (GBT)

1 -11 -1 -5 15 -7

63 43 63 65 2 55

32 12 15 18 19 16

24 1 -1 5 23 8

68 57 71 75 -4 71

81 68 87 93 -5 77

76 66 84 89 -5 69

72 61 82 84 5 66

38 36 59 54 -15 43

36 34 32 41 -21 41

Dataset = Color-dSprites

Beta

VA

E S

core

Fact

orV

AE S

core

MIG

DC

I D

isenta

ngle

ment

Modula

rity

SA

P

LR10

LR100

LR1000

LR10000

GBT10

GBT100

GBT1000

GBT10000

Efficiency (LR)

Efficiency (GBT)

28 19 18 12 27 9

53 33 -0 -11 55 15

60 40 2 -5 58 18

57 40 -8 -16 58 17

19 14 26 25 19 14

26 37 64 85 -19 28

-1 20 52 75 -36 15

-6 15 43 65 -35 8

-14 -12 13 16 -18 -4

33 26 32 36 6 25

Dataset = Noisy-dSprites

Beta

VA

E S

core

Fact

orV

AE S

core

MIG

DC

I D

isenta

ngle

ment

Modula

rity

SA

P

LR10

LR100

LR1000

LR10000

GBT10

GBT100

GBT1000

GBT10000

Efficiency (LR)

Efficiency (GBT)

61 65 60 55 48 60

83 79 77 75 46 81

86 79 80 80 41 84

89 83 82 82 43 87

67 68 68 65 41 64

85 78 85 93 37 80

83 78 85 94 35 79

81 75 82 92 33 78

-69 -63 -65 -67 -39 -67

-71 -65 -72 -82 -28 -69

Dataset = Scream-dSprites

Beta

VA

E S

core

Fact

orV

AE S

core

MIG

DC

I D

isenta

ngle

ment

Modula

rity

SA

P

LR10

LR100

LR1000

LR10000

GBT10

GBT100

GBT1000

GBT10000

Efficiency (LR)

Efficiency (GBT)

-19 -21 -12 -4 29 -12

1 -2 -41 -53 22 -45

24 23 -18 -29 0 -30

60 53 14 3 -33 -9

-46 -46 -26 -26 53 -21

63 39 83 86 -43 62

69 56 80 87 -56 53

74 63 77 82 -61 49

-49 -58 -59 -64 58 -38

-55 -77 -29 -34 64 0

Dataset = SmallNORB

Beta

VA

E S

core

Fact

orV

AE S

core

MIG

DC

I D

isenta

ngle

ment

Modula

rity

SA

P

LR10

LR100

LR1000

LR10000

GBT10

GBT100

GBT1000

GBT10000

Efficiency (LR)

Efficiency (GBT)

27 7 -27 -3 -23 -3

4 -10 -39 -25 -44 -12

12 -9 -44 -20 -49 -15

19 -3 -40 -12 -43 -13

15 10 -6 4 -8 -8

13 27 44 52 17 8

37 57 64 80 45 21

0 -13 -23 -32 -22 -1

-17 -7 -2 -18 -2 4

8 26 40 53 23 6

Dataset = Cars3D

Beta

VA

E S

core

Fact

orV

AE S

core

MIG

DC

I D

isenta

ngle

ment

Modula

rity

SA

P

LR10

LR100

LR1000

LR10000

GBT10

GBT100

GBT1000

GBT10000

Efficiency (LR)

Efficiency (GBT)

55 38 -14 11 19 14

73 56 -1 30 30 32

44 11 -47 -1 38 -14

42 7 -52 -6 33 -16

60 56 51 71 42 55

43 36 71 96 50 50

53 42 56 97 55 49

59 46 48 94 55 49

14 40 66 35 -11 49

23 21 77 80 38 45

Dataset = Shapes3D

Figure 19. Rank-correlation between the metrics and the performance on downstream task on different data sets. We observe somecorrelation between most disentanglement metrics and downstream performance. However, the correlation varies across data sets.

where the goal is to recover the true factors of variations from the learned representation using either multi-class logisticregression (LR) or gradient boosted trees (GBT). Our goal is to investigate the relationship between disentanglement and theaverage classification accuracy on these downstream tasks as well as whether better disentanglement leads to a decreasedsample complexity of learning.

To compute the classification accuracy for each trained model, we sample true factors of variations and observations fromour ground truth generative models. We then feed the observations into our trained model and take the mean of the Gaussianencoder as the representations. Finally, we predict each of the ground-truth factors based on the representations with aseparate learning algorithm. We consider both a 5-fold cross-validated multi-class logistic regression as well as gradientboosted trees of the Scikit-learn package. For each of these methods, we train on 10, 100, 1000 and 10 000 samples. Wecompute the average accuracy across all factors of variation using an additional set 10 000 randomly drawn samples.

Figure 19 shows the rank correlations between the disentanglement metrics and the downstream performance for allconsidered data sets. We observe that all metrics except Modularity seem to be correlated with increased downstreamperformance on the different variations of dSprites and to some degree on Shapes3D. However, it is not clear whether this isdue to the fact that disentangled representations perform better or whether some of these scores actually also (partially)capture the informativeness of the evaluated representation. Furthermore, the correlation is weaker or inexistent on other datasets (e.g., Cars3D). Finally, we report in Figure 24 the rank correlation between unsupervised scores computed after trainingon the mean and sampled representation and downstream performance. Depending on the data set, the rank correlationranges from from mildly negative, to mildly positive. In particular, we do not observe enough evidence supporting the claimthat decreased total correlation of the aggregate posterior proves beneficial for downstream task performance.

To assess the sample complexity argument we compute for each trained model a statistical efficiency score which we defineas the average accuracy based on 100 samples divided by the average accuracy based on 10 000 samples for either the logisticregression or the gradient boosted trees. The key idea is that if disentangled representations lead to sample efficiency, thenthey should also exhibit a higher statistical efficiency score. We remark that this score differs from the definition of samplecomplexity commonly used in statistical learning theory. The corresponding results are shown in Figures 20 and 21 wherewe plot the statistical efficiency versus different disentanglement metrics for different data sets and models and in Figure 19where we show rank correlations. Overall, we do not observe conclusive evidence that models with higher disentanglementscores also lead to higher statistical efficiency. We note that some AnnealedVAE models seem to exhibit a high statisticalefficiency on Scream-dSprites and to some degree on Noisy-dSprites. This can be explained by the fact that these models

Page 32: Challenging Common Assumptions in the Unsupervised ... · first theoretically show that the unsupervised learning of disentangled representations is fundamentally impossible without

Challenging Common Assumptions in the Unsupervised Learning of Disentangled Representations

0.4 0.6 0.8 1.00.4

0.6

0.8

1.0

Eff

icie

ncy

(LR

)

Metric = BetaVAE Score

Model β-VAE FactorVAE β-TCVAE DIP-VAE-I DIP-VAE-II AnnealedVAE

0.2 0.4 0.6 0.8 1.00.4

0.6

0.8

1.0 Metric = FactorVAE Score

0.0 0.2 0.40.4

0.6

0.8

1.0 Metric = MIG

0.0 0.2 0.4 0.60.4

0.6

0.8

1.0 Metric = DCI Disentanglement

0.6 0.7 0.8 0.9 1.00.4

0.6

0.8

1.0 Metric = Modularity

0.00 0.04 0.080.4

0.6

0.8

1.0

Data

set =

dSprite

s

Metric = SAP

0.60 0.75 0.90

0.60

0.75

0.90

Eff

icie

ncy

(LR

)

0.2 0.4 0.6 0.8 1.0

0.60

0.75

0.90

0.0 0.2 0.4

0.60

0.75

0.90

0.0 0.2 0.4 0.6

0.60

0.75

0.90

0.7 0.8 0.9 1.0

0.60

0.75

0.90

0.00 0.04 0.08

0.60

0.75

0.90 Data

set =

Colo

r-dSprite

s

0.25 0.50 0.75 1.000.4

0.6

0.8

1.0

Eff

icie

ncy

(LR

)

0.0 0.3 0.60.4

0.6

0.8

1.0

0.0 0.1 0.2 0.30.4

0.6

0.8

1.0

0.00 0.15 0.300.4

0.6

0.8

1.0

0.60 0.75 0.900.4

0.6

0.8

1.0

0.00 0.03 0.060.4

0.6

0.8

1.0

Data

set =

Noisy

-dSprite

s

0.25 0.50 0.75

0.75

0.90

Eff

icie

ncy

(LR

)

0.2 0.4 0.6 0.8

0.75

0.90

0.0 0.1 0.2 0.3

0.75

0.90

0.00 0.15 0.30

0.75

0.90

0.60 0.75 0.90

0.75

0.90

0.00 0.04 0.08

0.75

0.90

Data

set =

Scre

am

-dSprite

s

0.4 0.6 0.8 1.0

0.6

0.8

1.0

Eff

icie

ncy

(LR

)

0.2 0.4 0.6 0.8

0.6

0.8

1.0

0.00 0.15 0.30

0.6

0.8

1.0

0.0 0.1 0.2 0.3 0.4

0.6

0.8

1.0

0.75 0.90

0.6

0.8

1.0

0.00 0.08 0.16

0.6

0.8

1.0 Data

set =

Sm

allN

OR

B

0.97 0.98 0.99 1.000.2

0.4

0.6

0.8

Eff

icie

ncy

(LR

)

0.6 0.7 0.8 0.9 1.00.2

0.4

0.6

0.8

0.0 0.1 0.2 0.30.2

0.4

0.6

0.8

0.0 0.2 0.4 0.60.2

0.4

0.6

0.8

0.7 0.8 0.9 1.00.2

0.4

0.6

0.8

0.000 0.025 0.0500.2

0.4

0.6

0.8

Data

set =

Cars3

D

0.4 0.6 0.8 1.0Value

0.50

0.75

1.00

Eff

icie

ncy

(LR

)

0.3 0.6 0.9Value

0.50

0.75

1.00

0.0 0.4 0.8Value

0.50

0.75

1.00

0.0 0.4 0.8 1.2Value

0.50

0.75

1.00

0.75 0.90Value

0.50

0.75

1.00

0.0 0.1 0.2 0.3Value

0.50

0.75

1.00

Data

set =

Shapes3

D

Figure 20. Statistical efficiency (accuracy with 100 samples ÷ accuracy with 10 000 samples) based on a logistic regression versusdisentanglement metrics for different models and data sets. We do not observe that higher disentanglement scores lead to higher statisticalefficiency.

Page 33: Challenging Common Assumptions in the Unsupervised ... · first theoretically show that the unsupervised learning of disentangled representations is fundamentally impossible without

Challenging Common Assumptions in the Unsupervised Learning of Disentangled Representations

0.4 0.6 0.8 1.00.40

0.48

0.56

0.64

Eff

icie

ncy

(G

BT)

Metric = BetaVAE Score

Model β-VAE FactorVAE β-TCVAE DIP-VAE-I DIP-VAE-II AnnealedVAE

0.2 0.4 0.6 0.8 1.00.40

0.48

0.56

0.64

Metric = FactorVAE Score

0.0 0.2 0.40.40

0.48

0.56

0.64

Metric = MIG

0.0 0.2 0.4 0.60.40

0.48

0.56

0.64

Metric = DCI Disentanglement

0.6 0.7 0.8 0.9 1.00.40

0.48

0.56

0.64

Metric = Modularity

0.00 0.04 0.080.40

0.48

0.56

0.64 Data

set =

dSprite

s

Metric = SAP

0.60 0.75 0.900.4

0.5

0.6

0.7

Eff

icie

ncy

(G

BT)

0.2 0.4 0.6 0.8 1.00.4

0.5

0.6

0.7

0.0 0.2 0.40.4

0.5

0.6

0.7

0.0 0.2 0.4 0.60.4

0.5

0.6

0.7

0.7 0.8 0.9 1.00.4

0.5

0.6

0.7

0.00 0.04 0.080.4

0.5

0.6

0.7

Data

set =

Colo

r-dSprite

s

0.25 0.50 0.75 1.00

0.50

0.75

1.00

Eff

icie

ncy

(G

BT)

0.0 0.3 0.6

0.50

0.75

1.00

0.0 0.1 0.2 0.3

0.50

0.75

1.00

0.00 0.15 0.30

0.50

0.75

1.00

0.60 0.75 0.90

0.50

0.75

1.00

0.00 0.03 0.06

0.50

0.75

1.00 Data

set =

Noisy

-dSprite

s

0.25 0.50 0.750.4

0.6

0.8

1.0

Eff

icie

ncy

(G

BT)

0.2 0.4 0.6 0.80.4

0.6

0.8

1.0

0.0 0.1 0.2 0.30.4

0.6

0.8

1.0

0.00 0.15 0.300.4

0.6

0.8

1.0

0.60 0.75 0.900.4

0.6

0.8

1.0

0.00 0.04 0.080.4

0.6

0.8

1.0

Data

set =

Scre

am

-dSprite

s

0.4 0.6 0.8 1.0

0.56

0.64

0.72

0.80

Eff

icie

ncy

(G

BT)

0.2 0.4 0.6 0.8

0.56

0.64

0.72

0.80

0.00 0.15 0.30

0.56

0.64

0.72

0.80

0.0 0.1 0.2 0.3 0.4

0.56

0.64

0.72

0.80

0.75 0.90

0.56

0.64

0.72

0.80

0.00 0.08 0.16

0.56

0.64

0.72

0.80

Data

set =

Sm

allN

OR

B

0.97 0.98 0.99 1.00

0.30

0.45

0.60

Eff

icie

ncy

(G

BT)

0.6 0.7 0.8 0.9 1.0

0.30

0.45

0.60

0.0 0.1 0.2 0.3

0.30

0.45

0.60

0.0 0.2 0.4 0.6

0.30

0.45

0.60

0.7 0.8 0.9 1.0

0.30

0.45

0.60

0.000 0.025 0.050

0.30

0.45

0.60

Data

set =

Cars3

D

0.4 0.6 0.8 1.0Value

0.50

0.75

1.00

Eff

icie

ncy

(G

BT)

0.3 0.6 0.9Value

0.50

0.75

1.00

0.0 0.4 0.8Value

0.50

0.75

1.00

0.0 0.4 0.8 1.2Value

0.50

0.75

1.00

0.75 0.90Value

0.50

0.75

1.00

0.0 0.1 0.2 0.3Value

0.50

0.75

1.00

Data

set =

Shapes3

D

Figure 21. Statistical efficiency (accuracy with 100 samples ÷ accuracy with 10 000 samples) based on gradient boosted trees versusdisentanglement metrics for different models and data sets. We do not observe that higher disentanglement scores lead to higher statisticalefficiency (except for DCI Disentanglement and Mutual Information Gap on Shapes3D and to some extend in Cars3D).

Page 34: Challenging Common Assumptions in the Unsupervised ... · first theoretically show that the unsupervised learning of disentangled representations is fundamentally impossible without

Challenging Common Assumptions in the Unsupervised Learning of Disentangled Representations

0.0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

Valu

e

Dataset = Cars3D

Bottom 33% (DCI Disentanglement) Middle 33% (DCI Disentanglement) Top 33% (DCI Disentanglement)

0.0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8 Dataset = Color-dSprites

0.0

0.1

0.2

0.3

0.4

0.5

0.6 Dataset = Noisy-dSprites

GBT10 GBT100 GBT1000 GBT10000Metric

0.05

0.10

0.15

0.20

0.25

0.30

0.35

0.40 Dataset = Scream-dSprites

GBT10 GBT100 GBT1000 GBT10000Metric

0.0

0.2

0.4

0.6

0.8

1.0

1.2

Valu

e

Dataset = Shapes3D

GBT10 GBT100 GBT1000 GBT10000Metric

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8 Dataset = SmallNORB

GBT10 GBT100 GBT1000 GBT10000Metric

0.0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8 Dataset = dSprites

Figure 22. Downstream performance for three groups with increasing DCI Disentanglement scores.

have low downstream performance and that hence the accuracy with 100 samples is similar to the accuracy with 10 000samples. We further observe that DCI Disentanglement and MIG seem to be lead to a better statistical efficiency on thethe data set Shapes3D for gradient boosted trees. Figures 22 and 23 show the downstream performance for three groupswith increasing levels of disentanglement (measured in DCI Disentanglement and MIG respectively). We observe thatindeed models with higher disentanglement scores seem to exhibit better performance for gradient boosted trees with 100samples. However, considering all data sets, it appears that overall increased disentanglement is rather correlated with betterdownstream performance (on some data sets) and not statistical efficiency. We do not observe that higher disentanglementscores reliably lead to a higher sample efficiency.

Implications. While the empirical results in this section are negative, they should also be interpreted with care. After all,we have seen in previous sections that the models considered in this study fail to reliably produce disentangled representations.Hence, the results in this section might change if one were to consider a different set of models, for example semi-supervisedor fully supervised one. Furthermore, there are many more potential notions of usefulness such as interpretability andfairness that we have not considered in our experimental evaluation. Nevertheless, we argue that the lack of concreteexamples of useful disentangled representations necessitates that future work on disentanglement methods should make thispoint more explicit. While prior work (Steenbrugge et al., 2018; Laversanne-Finot et al., 2018; Nair et al., 2018; Higginset al., 2017b; 2018) successfully applied disentanglement methods such as β-VAE on a variety of downstream tasks, it is notclear to us that these approaches and trained models performed well because of disentanglement.

Page 35: Challenging Common Assumptions in the Unsupervised ... · first theoretically show that the unsupervised learning of disentangled representations is fundamentally impossible without

Challenging Common Assumptions in the Unsupervised Learning of Disentangled Representations

0.0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

Valu

e

Dataset = Cars3D

Bottom 33% (MIG) Middle 33% (MIG) Top 33% (MIG)

0.0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8 Dataset = Color-dSprites

0.0

0.1

0.2

0.3

0.4

0.5

0.6 Dataset = Noisy-dSprites

GBT10 GBT100 GBT1000 GBT10000Metric

0.05

0.10

0.15

0.20

0.25

0.30

0.35

0.40 Dataset = Scream-dSprites

GBT10 GBT100 GBT1000 GBT10000Metric

0.0

0.2

0.4

0.6

0.8

1.0

1.2

Valu

e

Dataset = Shapes3D

GBT10 GBT100 GBT1000 GBT10000Metric

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8 Dataset = SmallNORB

GBT10 GBT100 GBT1000 GBT10000Metric

0.0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8 Dataset = dSprites

Figure 23. Downstream performance for three groups with increasing MIG scores.

TC

(sa

mple

d)

TC

(m

ean)

avgM

I (s

am

ple

d)

avgM

I (m

ean)

Reco

nst

ruct

ion KL

LR10

LR100

LR1000

LR10000

GBT10

GBT100

GBT1000

GBT10000

-13 -55 65 -32 -80 15

-6 -40 67 -9 -66 28

-11 -47 85 -5 -85 33

-14 -50 85 -10 -88 32

-5 -15 31 -4 -37 3

-9 20 -3 34 4 -40

-10 -3 -12 2 3 -42

21 -28 28 -27 -30 29

Dataset = Cars3D

TC

(sa

mple

d)

TC

(m

ean)

avgM

I (s

am

ple

d)

avgM

I (m

ean)

Reco

nst

ruct

ion KL

LR10

LR100

LR1000

LR10000

GBT10

GBT100

GBT1000

GBT10000

-28 16 1 30 9 -9

-18 17 -31 7 22 -67

-31 -15 17 2 -21 -22

-28 -47 45 -16 -52 -0

17 19 -28 2 22 -59

20 41 -44 19 42 -83

18 46 -41 28 44 -82

15 44 -34 35 43 -74

Dataset = Color-dSprites

TC

(sa

mple

d)

TC

(m

ean)

avgM

I (s

am

ple

d)

avgM

I (m

ean)

Reco

nst

ruct

ion KL

LR10

LR100

LR1000

LR10000

GBT10

GBT100

GBT1000

GBT10000

-18 21 -17 19 20 -21

-14 1 -24 -9 8 -52

-36 -17 10 -7 -13 -13

-39 -54 48 -34 -54 11

10 17 -28 5 16 -53

14 34 -47 16 33 -80

15 35 -48 17 35 -83

14 30 -42 16 32 -80

Dataset = dSprites

TC

(sa

mple

d)

TC

(m

ean)

avgM

I (s

am

ple

d)

avgM

I (m

ean)

Reco

nst

ruct

ion KL

LR10

LR100

LR1000

LR10000

GBT10

GBT100

GBT1000

GBT10000

-7 14 3 23 21 -1

11 -2 30 10 9 40

8 -4 36 9 3 46

7 -19 49 -4 -15 57

11 31 -4 30 19 -2

21 40 -48 20 29 -43

5 17 -47 -2 6 -54

-8 1 -37 -13 -8 -52

Dataset = Noisy-dSprites

TC

(sa

mple

d)

TC

(m

ean)

avgM

I (s

am

ple

d)

avgM

I (m

ean)

Reco

nst

ruct

ion KL

LR10

LR100

LR1000

LR10000

GBT10

GBT100

GBT1000

GBT10000

-22 -36 27 -30 -38 -23

-19 -26 2 -33 -18 -42

-16 -29 -8 -44 -16 -43

-11 -28 -9 -43 -15 -45

-18 -36 20 -31 -30 -27

-22 -40 4 -27 -14 -41

-26 -43 9 -24 -15 -40

-26 -45 10 -24 -15 -39

Dataset = Scream-dSprites

TC

(sa

mple

d)

TC

(m

ean)

avgM

I (s

am

ple

d)

avgM

I (m

ean)

Reco

nst

ruct

ion KL

LR10

LR100

LR1000

LR10000

GBT10

GBT100

GBT1000

GBT10000

-2 -45 43 -12 -48 5

10 -35 47 -3 -46 -1

7 -66 73 3 -79 2

8 -66 71 -1 -79 3

-11 3 18 42 4 -27

-13 16 -2 51 27 -40

-8 3 13 46 12 -34

-7 -4 21 43 3 -32

Dataset = Shapes3D

TC

(sa

mple

d)

TC

(m

ean)

avgM

I (s

am

ple

d)

avgM

I (m

ean)

Reco

nst

ruct

ion KL

LR10

LR100

LR1000

LR10000

GBT10

GBT100

GBT1000

GBT10000

11 31 -2 27 17 8

3 -3 24 11 7 26

11 -23 36 -12 -22 37

12 -55 51 -53 -60 46

4 60 -21 54 46 -4

-15 -40 -1 -57 -66 -11

-3 -51 9 -72 -78 1

1 -60 17 -79 -84 9

Dataset = SmallNORB

Figure 24. Rank correlation between unsupervised scores and downstream performance.

Page 36: Challenging Common Assumptions in the Unsupervised ... · first theoretically show that the unsupervised learning of disentangled representations is fundamentally impossible without

Challenging Common Assumptions in the Unsupervised Learning of Disentangled Representations

J. Additional Figures

0.0 0.2 0.4 0.6 0.8 1.00.00

0.01

0.02

0.03

0.04

0.05

TC

(sa

mple

d)

Dataset = dSprites

Model VAE β-VAE FactorVAE β-TCVAE DIP-VAE-I DIP-VAE-II AnnealedVAE

0.0 0.2 0.4 0.6 0.8 1.00.00

0.02

0.04

0.06

0.08

0.10

0.12Dataset = Color-dSprites

0.0 0.2 0.4 0.6 0.8 1.00.00

0.02

0.04

0.06

0.08

0.10

0.12Dataset = Noisy-dSprites

0.0 0.2 0.4 0.6 0.8 1.0Regularization strength

0.00

0.01

0.02

0.03

0.04

0.05

0.06Dataset = Scream-dSprites

0.0 0.2 0.4 0.6 0.8 1.0Regularization strength

0.0

0.5

1.0

1.5

2.0

2.5

TC

(sa

mple

d)

Dataset = SmallNORB

0.0 0.2 0.4 0.6 0.8 1.0Regularization strength

0.00

0.05

0.10

0.15

0.20

0.25

0.30Dataset = Cars3D

0.0 0.2 0.4 0.6 0.8 1.0Regularization strength

0.00

0.05

0.10

0.15

0.20

0.25Dataset = Shapes3D

Figure 25. Total correlation of sampled representation plotted against regularization strength for different data sets and approaches(including AnnealedVAE).

0.0 0.2 0.4 0.6 0.8 1.00.0

0.5

1.0

1.5

2.0

2.5

3.0

TC

(m

ean)

Dataset = dSprites

Model VAE β-VAE FactorVAE β-TCVAE DIP-VAE-I DIP-VAE-II AnnealedVAE

0.0 0.2 0.4 0.6 0.8 1.00.0

0.5

1.0

1.5

2.0

2.5

3.0Dataset = Color-dSprites

0.0 0.2 0.4 0.6 0.8 1.00

1

2

3

4

5Dataset = Noisy-dSprites

0.0 0.2 0.4 0.6 0.8 1.0Regularization strength

0

1

2

3

4

5

6

7Dataset = Scream-dSprites

0.0 0.2 0.4 0.6 0.8 1.0Regularization strength

0

5

10

15

20

TC

(m

ean)

Dataset = SmallNORB

0.0 0.2 0.4 0.6 0.8 1.0Regularization strength

0

1

2

3

4

5

6Dataset = Cars3D

0.0 0.2 0.4 0.6 0.8 1.0Regularization strength

0.0

0.5

1.0

1.5

2.0

2.5Dataset = Shapes3D

Figure 26. Total correlation of mean representation plotted against regularization strength for different data sets and approaches (includingAnnealedVAE).

Page 37: Challenging Common Assumptions in the Unsupervised ... · first theoretically show that the unsupervised learning of disentangled representations is fundamentally impossible without

Challenging Common Assumptions in the Unsupervised Learning of Disentangled Representations

0.0 0.2 0.4 0.6 0.8 1.00.015

0.020

0.025

0.030

0.035

0.040

avgM

I (s

am

ple

d)

Dataset = dSprites

Model VAE β-VAE FactorVAE β-TCVAE DIP-VAE-I DIP-VAE-II

0.0 0.2 0.4 0.6 0.8 1.00.015

0.020

0.025

0.030

0.035

0.040Dataset = Color-dSprites

0.0 0.2 0.4 0.6 0.8 1.00.015

0.020

0.025

0.030

0.035

0.040Dataset = Noisy-dSprites

0.0 0.2 0.4 0.6 0.8 1.0Regularization strength

0.016

0.018

0.020

0.022

0.024

0.026

0.028

0.030

0.032Dataset = Scream-dSprites

0.0 0.2 0.4 0.6 0.8 1.0Regularization strength

0.015

0.020

0.025

0.030

0.035

avgM

I (s

am

ple

d)

Dataset = SmallNORB

0.0 0.2 0.4 0.6 0.8 1.0Regularization strength

0.015

0.020

0.025

0.030

0.035

0.040

0.045

0.050Dataset = Cars3D

0.0 0.2 0.4 0.6 0.8 1.0Regularization strength

0.02

0.04

0.06

0.08

0.10

0.12Dataset = Shapes3D

Figure 27. The average mutual information of the dimensions of the sampled representation generally decrease except for DIP-VAE-I.

0.0 0.2 0.4 0.6 0.8 1.00.02

0.03

0.04

0.05

0.06

0.07

0.08

0.09

0.10

0.11

avgM

I (m

ean)

Dataset = dSprites

Model VAE β-VAE FactorVAE β-TCVAE DIP-VAE-I DIP-VAE-II

0.0 0.2 0.4 0.6 0.8 1.00.02

0.03

0.04

0.05

0.06

0.07Dataset = Color-dSprites

0.0 0.2 0.4 0.6 0.8 1.00.03

0.04

0.05

0.06

0.07

0.08

0.09

0.10

0.11Dataset = Noisy-dSprites

0.0 0.2 0.4 0.6 0.8 1.0Regularization strength

0.020

0.025

0.030

0.035

0.040

0.045

0.050

0.055

0.060Dataset = Scream-dSprites

0.0 0.2 0.4 0.6 0.8 1.0Regularization strength

0.00

0.05

0.10

0.15

0.20

0.25

0.30

0.35

avgM

I (m

ean)

Dataset = SmallNORB

0.0 0.2 0.4 0.6 0.8 1.0Regularization strength

0.02

0.03

0.04

0.05

0.06

0.07

0.08Dataset = Cars3D

0.0 0.2 0.4 0.6 0.8 1.0Regularization strength

0.04

0.06

0.08

0.10

0.12

0.14

0.16Dataset = Shapes3D

Figure 28. The average mutual information of the dimensions of the mean representation generally increase.