From abstract items to latent spaces to observed data and back: Compositional Variational Auto-Encoder
Abstract
Conditional Generative Models are now acknowledged an essential tool in Machine Learning. This paper focuses on their control. While many approaches aim at disentangling the data through the coordinate-wise control of their latent representations, another direction is explored in this paper. The proposed CompVAE handles data with a natural multi-ensemblist structure (i.e. that can naturally be decomposed into elements). Derived from Bayesian variational principles, CompVAE learns a latent representation leveraging both observational and symbolic information. A first contribution of the approach is that this latent representation supports a compositional generative model, amenable to multi-ensemblist operations (addition or subtraction of elements in the composition). This compositional ability is enabled by the invariance and generality of the whole framework w.r.t. respectively, the order and number of the elements. The second contribution of the paper is a proof of concept on synthetic 1D and 2D problems, demonstrating the efficiency of the proposed approach.
Keywords: Generative model, semi-structured representation, neural networks
1 Introduction
Representation learning is at the core of machine learning, and even more so since the inception of deep learning [2]. As shown by e.g., [3, 12], the latent representations built to handle high-dimensional data can effectively support desirable functionalities. One such functionality is the ability to directly control the observed data through the so-called representation disentanglement, especially in the context of computer vision and image processing [25, 20] (more in section 2).
This paper extends the notion of representation disentanglement from a latent coordinate-wise perspective to a semi-structured setting. Specifically, we tackle the ensemblist setting where a datapoint can naturally be interpreted as the combination of multiple parts. The contribution of the paper is a generative model built on the Variational Auto-Encoder principles [17, 27], controlling the data generation from a description of its parts and supporting ensemblist operations such as the addition or removal of any number of parts.
The applicative motivation for the presented approach, referred to as Compositional Variational AutoEncoder (CompVAE), is the following. In the domain of Energy Management, a key issue is to simulate the consumption behavior of an ensemble of consumers, where each household consumption is viewed as an independent random variable following a distribution law defined from the household characteristics, and the household consumptions are possibly correlated through external factors such as the weather, or a football match on TV (attracting members of some but not all households). Our long term goal is to infer a simulator, taking as input the household profiles and their amounts: it should be able to simulate their overall energy consumption and account for their correlations. The data-driven inference of such a programmable simulator is a quite desirable alternative to the current approaches, based on Monte-Carlo processes and requiring either to explicitly model the correlations of the elementary random variables, or to proceed by rejection.
Formally, given the description of datapoints and their parts, the goal of CompVAE is to learn the distribution laws of the parts (here, the households) and to sample the overall distribution defined from a varying number of parts (the set of households), while accounting for the fact that the parts are not independent, and the sought overall distribution depends on shared external factors: the whole is not the sum of its parts.
The paper is organized as follows. Section 2 briefly reviews related work in the domain of generative models and latent space construction, replacing our contribution in context. Section 3 gives an overview of CompVAE, extending the VAE framework to multi-ensemblist settings. Section 4 presents the experimental setting retained to establish a proof of concept of the approach on two synthetic problems, and section 5 reports on the results. Finally section 6 discusses some perspectives for further work and applications to larger problems.
2 Related Work
Generative models, including VAEs [17, 27] and GANs [9], rely on an embedding from the so-called latent space onto the dataspace . In the following, data space and observed space are used interchangeably. It has long been observed that continuous or discrete operations in the latent space could be used to produce interesting patterns in the data space. For instance, the linear interpolation between two latent points and can be used to generate a morphing between their images [26], or the flip of a boolean coordinate of can be used to add or remove an elementary pattern (the presence of glasses or mustache) in the associated image [7].
The general question then is to control the flow of information from the latent to the observed space and to make it actionable. Several approaches, either based on information theory or on supervised learning have been proposed to do so. Losses inspired from the Information Bottleneck [31, 29, 1] and enforcing the independence of the latent and the observed variables, conditionally to the relevant content of information, have been proposed: enforcing the decorrelation of the latent coordinates in -VAE [12]; aligning the covariances of latent and observed data in [19]; decomposing the latent information into pure content and pure noise in InfoGAN [3]. Independently, explicit losses have been used to yield conditional distributions in conditional GANs [23], or to enforce the scope of a latent coordinate in [18, 32], (e.g. modeling the light orientation or the camera angle).
The structure of the observed space can be mimicked in the latent space, to afford expressive yet trainable model spaces; in Ladder-VAE [30], a sequence of dependent latent variables are encoded and reversely decoded to produce complex observed objects. Auxiliary losses are added in [22] in the spirit of semi-supervised learning. In [16], the overall generative model involves a classifier, trained both in a supervised way with labeled examples and in an unsupervised way in conjunction with a generative model.
An important case study is that of sequential structures: [5] considers fixed-length sequences and loosely mimics an HMM process, where latent variable controls the observed variable and the next latent . In [13], a linear relation among latent variables and is enforced; in [6], a recurrent neural net is used to produce the latent variable encoding the current situation. In a more general context, [34] provides a generic method for designing an appropriate inference network that can be associated with a given Bayesian network representing a generative model to train.
The injection of explicit information at the latent level can be used to support ”information surgery” via loss-driven information parsimony. For instance in the domain of signal generation [4], the neutrality of the latent representation w.r.t. the locutor identity is enforced by directly providing the identity at the latent level: as does not need to encode the locutor information, the information parsimony pressure ensures independence wrt the locutor. Likewise, fair generative processes can be enforced by directly providing the sensitive information at the latent level [35]. In [21], an adversarial mechanism based on Maximum Mean Discrepancy [10] is used to enforce the neutrality of the latent. In [24], the minimization of the mutual information is used in lieu of an adversary.
Discussion.
All above approaches (with the except of sequential settings [5, 13], see below) handle the generation of a datapoint as a whole naturally involving diverse facets; but not composed of inter-related parts. Our goal is instead to tackle the proper parts-and-whole structure of a datapoint, where the whole is not necessarily the simple sum of its parts and the parts of the whole are interdependent. In sequential settings [5, 13], the dependency of the elements in the sequence are handled through parametric restrictions (respectively considering fixed sequence-size or linear temporal dependency) to enforce the proper match of the observed and latent spaces. A key contribution of the proposed CompVAE is to tackle the parts-to-whole structure with no such restrictions, and specifically accommodating a varying number of parts possibly different between the training and the generation phases.
3 Overview of CompVAE
This section describes the CompVAE model, building upon the VAE principles [17] with the following difference: CompVAE aims at building a programmable generative model , taking as input the ensemble of the parts of a whole observed datapoint. A key question concerns the latent structure most appropriate to reflect the ensemblist nature of the observed data. The proposed structure (section 3.1) involves a latent variable associated to each part of the whole. The aggregation of the part is achieved through an order-invariant operation, and the interactions among the parts are modeled at an upper layer of the latent representation.
In encoding mode, the structure is trained from the pairs formed by a whole, and an abstract description of its parts; the latent variables are extracted along an iterative non-recurrent process, oblivious of the order and number of the parts (section 3.2) and defining the encoder model .
In generative mode, the generative model is supplied with a set of parts, and accordingly generates a consistent whole, where variational effects operate jointly at the part and at the whole levels.
Notations.
A datapoint is associated with an ensemble of parts noted . Each belongs to a finite set of categories . Elements and parts are used interchangeably in the following. In our illustrating example, a consumption curve involves a number of households; the -th household is associated with its consumer profile , with ranging in a finite set of profiles. Each profile in thus occurs 0, 1 or several times. The generative model relies on a learned distribution , that is decomposed into latent variables: a latent variable named associated to each part , and a common latent variable .
3.1 CompVAE: Bayesian architecture
The architecture proposed for CompVAE is depicted as a graphical model on Fig.
1. As said, the -th part belongs to category and is associated with a latent variable (different parts with same category are associated with different latent variables). The ensemble of the s is aggregated into an intermediate latent variable .
A key requirement is for to be invariant w.r.t. the order of elements in . In the following is set to the sum of the , . Considering other order-invariant aggregations is left for further work.
The intermediate latent variable is used to condition the latent variable; both and condition the observed datapoint . This scheme corresponds to the following factorization of the generative model :
(1) |
In summary, the distribution of is conditioned on the ensemble as follows: The -th part of is associated with a latent variable modeling the generic distribution of the underlying category together with its specifics. Variable is deterministically computed to model the aggregation of the , and finally models the specifics of the aggregation.
Notably, each is linked to a single element, while is global, being conditioned from the global auxiliary . The rationale for introducing is to enable a more complex though still learnable distribution at the level compared with the alternative of conditioning only on . It is conjectured that an information-effective distribution would store in (respectively in ) the local information related to the -th part (resp. the global information describing the interdependencies between all parts, e.g. the fact that the households face the same weather, vacation schedules, and so on). Along this line, it is conjectured that the extra information stored in is limited compared to that stored in the s; we shall return to this point in section 4.1.
The property of invariance of the distribution w.r.t. the order of the is satisfied by design. A second desirable property regards the robustness of the distribution w.r.t. the varying number of parts in . More precisely, two requirements are defined. The former one, referred to as size-flexibility property, is that the number of parts of an is neither constant, nor bounded a priori. The latter one, referred to as size-generality property is the generative model to accommodate larger numbers of parts than those seen in the training set.
3.2 Posterior inference and loss
Letting denote the empirical data distribution, the learning criterion to optimize is the data likelihood according to the sought generative model : .
The (intractable) posterior inference of the model is approximated using the Evidence Lower Bound (ELBO) [14],
following the Variational AutoEncoder approach [17, 27]. Accordingly, we proceed by optimizing a lower bound of the log-likelihood of the data given , which is equivalent to minimizing an upper bound of the Kullback-Leibler divergence between the two distributions :
(2) |
The learning criterion is, with the inference distribution:
(3) |
The inference distribution is further factorized as , yielding the final training loss:
(4) |
The training of the generative and encoder model distributions is described in Alg. 1.
3.3 Discussion
In CompVAE, the sought distributions are structured as a Bayesian graph (see in Fig. 1), where each node is associated with a neural network and a probability distribution family, like for VAEs. This neural network takes as input the parent variables in the Bayesian graph, and outputs the parameters of a distribution in the chosen family, e.g., the mean and variance of a Gaussian distribution. The reparametrization trick [17] is used to back-propagate gradients through the sampling.
A concern regards the training of latent variables when considering Gaussian distributions. A potential source of instability in CompVAE comes from the fact that the Kullback-Leibler divergence between and (Eq. (4)) becomes very large when the variance of some variables in becomes very small111Single-latent variable VAEs do not face such problems as the prior distribution is fixed, it is not learned.. To limit this risk, some care is exercised in parameterizing the variances of the normal distributions in to making them lower-bounded.
3.3.1 Modeling of .
The latent distributions , and are modeled using diagonal normal distributions as usual. Regarding the model , in order to be able to faithfully reflect the generative model , it is necessary to introduce the correlation between the s in [34].
As the aggregation of the is handled by considering their sum, it is natural to handle their correlations through a multivariate normal distribution over the . The proposed parametrization of such a multivariate is as follows. Firstly, correlations operate in a coordinate-wise fashion, that is, and are only correlated if . The parametrization (detailed in appendix LABEL:app:multivariate-parameter) of the s ensures that: i) the variance of the sum of the can be controlled and made arbitrarily small in order to ensure an accurate VAE reconstruction; ii) the Kullback-Leibler divergence between and can be defined in closed form.
The learning of is done using a fully-connected graph neural network [28] leveraging graph interactions akin message-passing [8]. The graph has one node for each element , and every node is connected to all other nodes. The state of the -th node is initialized to , where is some learned function of noted, is a learned embedding of , and is a random noise used to ensure the differentiation of the s. The state of each node of the graph at the -th layer is then defined by its -th layer state and the aggregation of the state of all other nodes:
(5) |
where and are learned neural networks: is meant to embed the current state of each node for an aggregate summation, and is meant to ”fine-tune” the -th node conditionally to all other nodes, such that they altogether account for .
4 Experimental Setting
This section presents the goals of experiments and describes the experimental setting used to empirically validate CompVAE.
4.1 Goals of experiments
As said, CompVAE is meant to achieve a programmable generative model. From a set of latent values , either derived from in a generative context, or recovered from some data , it should be able to generate values matching any chosen subset of the . This property is what we name the ”ensemblist disentanglement” capacity, and the first goal of these experiments is to investigate whether CompVAE does have this capacity.
A second goal of these experiments is to examine whether the desired properties (section 3.1) hold. The order-invariant property is enforced by design. The size-flexibility property will be assessed by inspecting the sensitivity of the extraction and generative processes to the variability of the number of parts. The size-generality property will be assessed by inspecting the quality of the generative model when the number of parts increases significantly beyond the size range used during training.
A last goal is to understand how CompVAE manages to store the information of the model in respectively the s and . The conjecture done (section 3.1) was that the latent s would take in charge the information of the parts, while the latent would model the interactions among the parts. The use of synthetic problems where the quantity of information required to encode the parts can be quantitatively assessed will permit to test this conjecture. A related question is whether the generative model is able to capture the fact that the whole is not the sum of its parts. This question is investigated using non-linear perturbations, possibly operating at the whole and at the parts levels, and comparing the whole perturbed obtained from the s, and the aggregation of the perturbed s generated from the parts. The existence of a difference, if any, will establish the value of the CompVAE generative model compared to a simple Monte-Carlo simulator, independently sampling parts and thereafter aggregating them.
4.2 1D and 2D Proofs of concept
Two synthetic problems have been considered to empirically answer the above questions.222These problems are publicly available at https://github.com/vberger/compvae .
In the 1D synthetic problem,
the set of categories is a finite set of frequencies . A given ”part” (here, curve) is a sine wave defined by its frequency in and its intrinsic features, that is, its amplitude and phase . The whole associated to is a finite sequence of size , deterministically defined from the non-linear combination of the curves:
with the number of sine waves in , a parameter controlling the non-linearity of the aggregation of the curves in , and a global parameter controlling the sampling frequency. For each part (sine wave), is sampled from , and is sampled from .
The part-to-whole aggregation is illustrated on Fig. 2, plotting the non-linear transformation of the sum of 4 sine waves, compared to the sum of non-linear transformations of the same sine waves. The sensitivity to is illustrated in supplementary material (Appendix LABEL:app:data-generation Fig. LABEL:fig:lambda-impact). is set to 3 in the experiments.
This 1D synthetic problem features several aspects relevant to the empirical assessment of CompVAE. Firstly, the impact of adding or removing one part can be visually assessed as it changes the whole curve: the general magnitude of the whole curve is roughly proportional to its number of parts. Secondly, each part involves, besides its category , some intrinsic variations of its amplitude and phase. Lastly, the whole is not the sum of its parts (Fig. 2).
The generative model is defined as a Gaussian distribution , the vector parameters and of which are produced by the neural network (architecture details in supplementary material, section LABEL:app:sines-nn-structure).
In the 2D synthetic problem,
each category in is composed of one out of five colors ( ) associated with a location in . Each thus is a colored site, and its internal variability is its intensity. The whole associated to a set of s is an image, where each pixel is colored depending on its distance to the sites and their intensity (Fig. 3).

Likewise, the observation model is a Gaussian distribution , the parameters and of which are produced by the neural network. The observation variance is shared for all three channel values (red, green, blue) of any given pixel. Architecture details are given in supplementary material (section LABEL:app:dots-nn-structure).
The 2D problem shares with the 1D problem the fact that each part is defined from its category (resp. a frequency, or a color and location) on the one hand, and its specifics on the other hand (resp, its amplitude and frequency, or its intensity); additionally, the whole is made of a set of parts in interaction. However, the 2D problem is significantly more complex than the 1D, as will be discussed in section 5.2.
4.3 Experimental setting
CompVAE is trained as a mainstream VAE, except for an additional factor of difficulty: the varying number of latent variables (reflecting the varying number of parts) results in a potentially large number of latent variables. This large size and the model noise in the early training phase can adversely affect the training procedure, and lead it to diverge. The training divergence is prevented using a batch size set to 256. The neural training hyperparameters are dynamically tuned using the Adam optimizer [15] with , and , which empirically provide a good compromise between training speed, network stability and good convergence. On the top of Adam, the annealing of the learning rate is achieved, dividing its value by 2 every 20,000 iterations, until it reaches .
For both problems, the data is generated on the fly during the training, 333The data generator is given in supplementary material, section LABEL:app:data-generation. preventing the risk of overfitting. The overall number of iterations (batches) is up to444Experimentally, networks most often converge much earlier. 500,000. The computational time on a GPU GTX1080 is 1 day for the 1D problem, and 2 days for the 2D problem.
Empirically, the training is facilitated by gradually increasing the number of parts in the datapoints. Specifically, the number of parts is uniformly sampled in at each iteration, with at the initialization and incremented by 1 every 3,000 iterations, up to 16 parts in the 1D problem and 8 in the 2D problem.
5 CompVAE: Empirical Validation
This section reports on the proposed proofs of concept of the CompVAE approach.
5.1 1D Proof of Concept
Fig. 4 displays in log-scale the losses of the s and latent variables along time, together with the reconstruction loss and the overall ELBO loss summing the other three (Eq. (4)). The division of labor between the s and the is seen as the quantity of information stored by the s increases to reach a plateau at circa 100 bits, while the quantity of information stored by steadily decreases to around 10 bits. As conjectured (section 3.1), carries little information.
Note that the reconstruction loss remains high, with a high ELBO even at convergence time, although the generated curves ”look good”. This fact is explained from the high entropy of the data: on the top of the specifics of each part (its amplitude and phase), is described as a -length sequence: the temporal discretization of the signal increases the variance of and thus causes a high entropy, which is itself a lower bound for the ELBO. Note that a large fraction of this entropy is accurately captured by CompVAE through the variance of the generative model .

The ability of ”ensemblist disentanglement” is visually demonstrated on Fig. 6: considering a set of , the individual parts are generated (Fig. 6, left) and gradually integrated to form a whole (Fig. 6, right) in a coherent manner.

The size-generality property is satisfactorily assessed as the model could be effectively used with a number of parts ranging up to 30 (as opposed to 16 during the training) without requiring any re-training or other modification of the model (results omitted for brevity).
5.2 2D Proof of Concept
As shown in Fig. 7, the 2D problem is more complex. On the one hand, a 2D part only has a local impact on (affecting a subset of pixels) while a 1D part has a global impact on the whole sequence. On the other hand, the number of parts has a global impact on the range of in the 1D problem, whereas each pixel value ranges in the same interval in the 2D problem. Finally and most importantly, is of dimension 200 in the 1D problem, compared to dimension () in the 2D problem. For these reasons, the latent variables here need to store more information, and the separation between the (converging toward circa 200-300 bits of information) and (circa 40-60 bits) is less clear.

Likewise, reconstruction loss remains high, although the generated images ”look good”, due to the fact that the loss precisely captures the discrepancies in the pixel values that the eye does not perceive.
Finally, the ability of ”ensemblist disentanglement” is inspected by incrementally generating the whole from a set of colored sites (Fig. 8). The top row displays the colors of from left to right. On the second row, the -th square shows an image composed from by the ground truth generator, and rows 3 to 6 show images generated by the model from the same . While the generated generally reflects the associated set of parts, some advents of black and white glitches are also observed (for instance on the third column, rows 3 and 5). These glitches are blamed on the saturation of the network (as black and white respectively are represented as and in RGB), since non linear combinations of colors are used for a good visual rendering555Color blending in the data generation is done taking into account gamma-correction..

6 Discussion and Perspectives
The main contribution of the paper is the generative framework CompVAE, to our best knowledge the first generative framework able to support the generation of data based on a multi-ensemble . Built on the top of the celebrated VAE, CompVAE learns to optimize the conditional distribution in a theoretically sound way, through introducing latent variables (one for each part ), enforcing their order-invariant aggregation and learning another latent variable to model the interaction of the parts. Two proofs of concepts for the approach, respectively concerning a 1D and a 2D problem, have been established with respectively very satisfactory and satisfactory results.
This work opens several perspectives for further research. A first direction in the domain of computer vision consists of combining CompVAE with more advanced image generation models such as PixelCNN [33] in a way similar to PixelVAE [11], in order to generate realistic images involving a predefined set of elements along a consistent layout.
A second perspective is to make one step further toward the training of fully programmable generative models. The idea is to incorporate explicit biases on the top of the distribution learned from unbiased data, to be able to sample the desired sub-spaces of the data space. In the motivating application domain of electric consumption for instance, one would like to sample the global consumption curves associated with high consumption peaks, that is, to bias the generation process toward the top quantiles of the overall distribution.
Acknowledgments
This work was funded by the ADEME #1782C0034 project NEXT (https://www.ademe.fr/next).
The authors would like to thank Balthazar Donon and Corentin Tallec for the many useful and inspiring discussions.
References
- [1] Alessandro Achille and Stefano Soatto. Emergence of invariance and disentanglement in deep representations. The Journal of Machine Learning Research, 19(1):1947–1980, 2018.
- [2] Yoshua Bengio, Aaron Courville, and Pascal Vincent. Representation learning: A review and new perspectives. IEEE Transactions on Pattern Analysis and Machine Intelligence, 35(8):1798–1828, 2013.
- [3] Xi Chen, Yan Duan, Rein Houthooft, John Schulman, Ilya Sutskever, and Pieter Abbeel. Infogan: Interpretable representation learning by information maximizing generative adversarial nets. In Advances in neural information processing systems, pages 2172–2180, 2016.
- [4] Jan Chorowski, Ron J. Weiss, Samy Bengio, and Aäron van den Oord. Unsupervised speech representation learning using WaveNet autoencoders. url: http://arxiv.org/abs/1901.08810, 2019.
- [5] Junyoung Chung, Kyle Kastner, Laurent Dinh, Kratarth Goel, Aaron C Courville, and Yoshua Bengio. A recurrent latent variable model for sequential data. In C. Cortes, N. D. Lawrence, D. D. Lee, M. Sugiyama, and R. Garnett, editors, Advances in Neural Information Processing Systems 28, pages 2980–2988. 2015.
- [6] John D. Co-Reyes, YuXuan Liu, Abhishek Gupta, Benjamin Eysenbach, Pieter Abbeel, and Sergey Levine. Self-consistent trajectory autoencoder: Hierarchical reinforcement learning with trajectory embeddings. International Conference on Machine Learning, 2018.
- [7] Vincent Dumoulin, Ishmael Belghazi, Ben Poole, Olivier Mastropietro, Alex Lamb, Martin Arjovsky, and Aaron Courville. Adversarially learned inference. International Conference on Learning Representations, 2017.
- [8] Justin Gilmer, Samuel S. Schoenholz, Patrick F. Riley, Oriol Vinyals, and George E. Dahl. Neural message passing for quantum chemistry. In International Conference on Machine Learning, pages 1263–1272, 2017.
- [9] Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. Generative adversarial nets. In Advances in Neural Information Processing Systems 27, pages 2672–2680. Curran Associates, Inc., 2014.
- [10] Arthur Gretton, Karsten M. Borgwardt, Malte J. Rasch, Bernhard Schölkopf, and Alexander Smola. A kernel two-sample test. The Journal of Machine Learning Research, 13:723–773, 2012.
- [11] Ishaan Gulrajani, Kundan Kumar, Faruk Ahmed, Adrien Ali Taiga, Francesco Visin, David Vazquez, and Aaron Courville. PixelVAE: A latent variable model for natural images. International Conference on Learning Representations, 2017.
- [12] Irina Higgins, Loic Matthey, Arka Pal, Christopher Burgess, Xavier Glorot, Matthew Botvinick, Shakir Mohamed, and Alexander Lerchner. beta-VAE: Learning basic visual concepts with a constrained variational framework. International Conference on Learning Representations, 2017.
- [13] Wei-Ning Hsu, Yu Zhang, and James Glass. Unsupervised learning of disentangled and interpretable representations from sequential data. In Advances in Neural Information Processing Systems 30, pages 1878–1889. Curran Associates, Inc., 2017.
- [14] Michael I. Jordan, Zoubin Ghahramani, Tommi S. Jaakkola, and Lawrence K. Saul. An introduction to variational methods for graphical models. Machine Learning, 37(2):183–233, 1999.
- [15] Diederik P. Kingma and Jimmy Ba. Adam: A method for stochastic optimization. International Conference on Learning Representations, 2014.
- [16] Diederik P. Kingma, Danilo J. Rezende, Shakir Mohamed, and Max Welling. Semi-supervised learning with deep generative models. Neural Information Processing Systems, 2014.
- [17] Diederik P. Kingma and Max Welling. Auto-encoding variational bayes. International Conference on Learning Representations, 2013.
- [18] Tejas D Kulkarni, William F. Whitney, Pushmeet Kohli, and Josh Tenenbaum. Deep convolutional inverse graphics network. In Advances in Neural Information Processing Systems 28, pages 2539–2547. Curran Associates, Inc., 2015.
- [19] Abhishek Kumar, Prasanna Sattigeri, and Avinash Balakrishnan. Variational inference of disentangled latent concepts from unlabeled observations. International Conference on Learning Representations, 2017.
- [20] Li Liu, Wanli Ouyang, Xiaogang Wang, Paul Fieguth, Jie Chen, Xinwang Liu, and Matti Pietikäinen. Deep learning for generic object detection: A survey. url: http://arxiv.org/abs/1809.02165, 2018.
- [21] Christos Louizos, Kevin Swersky, Yujia Li, Max Welling, and Richard Zemel. The variational fair autoencoder. url: http://arxiv.org/abs/1511.00830, 2015.
- [22] Lars Maaløe, Casper Kaae Sønderby, Søren Kaae Sønderby, and Ole Winther. Auxiliary deep generative models. In International Conference on Machine Learning, pages 1445–1453, 2016.
- [23] Mehdi Mirza and Simon Osindero. Conditional generative adversarial nets. url: http://arxiv.org/abs/1411.1784, 2014.
- [24] Daniel Moyer, Shuyang Gao, Rob Brekelmans, Aram Galstyan, and Greg Ver Steeg. Invariant representations without adversarial training. In Advances in Neural Information Processing Systems, pages 9084–9093, 2018.
- [25] Dilip K Prasad. Survey of the problem of object detection in real images. International Journal of Image Processing (IJIP), 6(6):441, 2012.
- [26] Alec Radford, Luke Metz, and Soumith Chintala. Unsupervised representation learning with deep convolutional generative adversarial networks. International Conference on Learning Representations, 2015.
- [27] Danilo Jimenez Rezende, Shakir Mohamed, and Daan Wierstra. Stochastic backpropagation and approximate inference in deep generative models. In International Conference on Machine Learning, pages 1278–1286, 2014.
- [28] F. Scarselli, M. Gori, A. C. Tsoi, M. Hagenbuchner, and G. Monfardini. The graph neural network model. IEEE Transactions on Neural Networks, 20(1):61–80, 2009.
- [29] Ravid Shwartz-Ziv and Naftali Tishby. Opening the black box of deep neural networks via information. url: http://arxiv.org/abs/1703.00810, 2017.
- [30] Casper Kaae Sønderby, Tapani Raiko, Lars Maaløe, Søren Kaae Sønderby, and Ole Winther. Ladder variational autoencoders. Advances in Neural Information Processing Systems, pages 3738–3746, 2016.
- [31] Naftali Tishby, Fernando C. Pereira, and William Bialek. The information bottleneck method. CoRR, physics/0004057, 2000.
- [32] Luan Tran, Xi Yin, and Xiaoming Liu. Disentangled representation learning GAN for pose-invariant face recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 1415–1424, 2017.
- [33] Aaron van den Oord, Nal Kalchbrenner, Lasse Espeholt, koray kavukcuoglu koray, Oriol Vinyals, and Alex Graves. Conditional image generation with PixelCNN decoders. In Advances in Neural Information Processing Systems 29, pages 4790–4798. Curran Associates, Inc., 2016.
- [34] Stefan Webb, Adam Golinski, Rob Zinkov, Siddharth Narayanaswamy, Tom Rainforth, Yee Whye Teh, and Frank Wood. Faithful inversion of generative models for effective amortized inference. In Advances in Neural Information Processing Systems, pages 3070–3080, 2018.
- [35] Rich Zemel, Yu Wu, Kevin Swersky, Toni Pitassi, and Cynthia Dwork. Learning fair representations. In International Conference on Machine Learning, pages 325–333, 2013.