Layer Collaboration in the Forward-Forward Algorithm
Abstract
Backpropagation, which uses the chain rule, is the de-facto standard algorithm for optimizing neural networks nowadays. Recently, Hinton (2022) proposed the forward-forward algorithm, a promising alternative that optimizes neural nets layer-by-layer, without propagating gradients throughout the network. Although such an approach has several advantages over back-propagation and shows promising results, the fact that each layer is being trained independently limits the optimization process. Specifically, it prevents the network’s layers from collaborating to learn complex and rich features. In this work, we study layer collaboration in the forward-forward algorithm. We show that the current version of the forward-forward algorithm is suboptimal when considering information flow in the network, resulting in a lack of collaboration between layers of the network. We propose an improved version that supports layer collaboration to better utilize the network structure, while not requiring any additional assumptions or computations. We empirically demonstrate the efficacy of the proposed version when considering both information flow and objective metrics. Additionally, we provide a theoretical motivation for the proposed method, inspired by functional entropy theory.
1 Introduction
Deep neural networks (DNNs) have been the backbone of many state-of-the-art machine learning systems in recent years. Their ability to learn complex representations from large amounts of data has led to breakthroughs in many fields such as image and speech recognition, natural language processing, and game playing (Lavin and Gray, 2016; Bojarski et al., 2016; Luss et al., 2019; Gat et al., 2021; Morais et al., 2022; Rosenberg et al., 2021; Ophir et al., 2020; Zhou et al., 2021; Clauset and Post, 2019; Fedorov et al., 2020). One of the key reasons for the success of deep neural networks is the backpropagation algorithm, which is currently the most widely used training paradigm.
Backpropagation (Rumelhart et al., 1986) utilizes the chain-rule to compute gradients for minimization of a loss function, which measures the discrepancy between the predicted output and the desired result. The algorithm starts with a forward pass through the network to compute the predicted output and then calculates the error between the prediction and the desired output. The error is then propagated back through the network using the chain rule of calculus, computing the gradient of the loss function with respect to the network’s parameters, i.e., the weights and biases. These gradients are then used to update the parameters via gradient descent variants, reducing the loss function and subsequently the error. This process is repeated until the network reaches a satisfactory error on the training data. The chain rule presented challenges for a variety of model architectures. In very deep networks, gradients of the loss function with respect to the weights can become small, resulting in vanishing gradients. Moreover, the gradient can also become too large. Both vanishing and exploding gradients lead to poor performance.
Recently, Hinton (2022) introduced the forward-forward algorithm, which offers an alternative to the backpropagation training process by replacing the forward and backward passes with two forward passes. The data that is forwarded in the network is the sample itself (e.g., image, text) concatenated with a one-hot encoded vector that represents a label. The first forward pass is done with positive data (i.e., data concatenated with the true label) and the second forward pass is done with negative data (i.e., data concatenated with a false label). The network then classifies whether a sample is concatenated with its true label. An important aspect of the forward-forward algorithm: the aforementioned optimization process is performed separately for each layer.

In this work, we examine the dynamics of individual layers during the forward-forward algorithm based training process. Our findings indicate that, during training, the layers exhibit a lack of effective intercommunication, which results in a suboptimal level of collaboration. Based on this observation, we propose a simple modification to the forward-forward algorithm to enhance inter-layer communication during training. Our empirical results demonstrate that incorporating this modification leads to a significant improvement in performance.
To further investigate this phenomenon, we analyze forward-forward-based models through the lens of functional entropy. We demonstrate that the forward-forward method implicitly maximizes the entropy within the network. Throughout this work, we analyze the dynamics of the entropy when training a model using the forward-forward algorithm.
Our contributions:
We begin by analyzing the layer collaboration of forward-forward-based models. We find that those models are limited in their ability to transfer information during training and consequently exhibit a poor collaboration between layers. Then, we propose a modification to the forward-forward algorithm that enables better information flow through the network. We show the effectiveness of this modification in terms of performance and layer collaboration. Our work also presents a theoretical connection between the forward-forward algorithm and functional entropy. Using it, we are able to investigate the dynamics of the layer through the lens of entropies.
The remainder of the work is organized as follows: Section 2 provides details regarding the forward-forward algorithm and how it differs from the popular backpropagation method. Section 3 analyzes the collaborative dynamics of the layers of forward-forward-based models and proposes a modification. Section 4 connects forward-forward optimization and the functional entropy of the layers. In Section 5 we extensively study the benefits of our proposed modification using performance metrics and our proposed theoretical framework. Lastly, Section 6 covers related works in the areas of layer collaboration and neural-network learning methods.
2 Background
Discriminative learning seeks to construct a mapping between data-instance and corresponding label , given training data . Classical discriminative deep fully-connected neural networks have a recursive structure. Each layer processes its predecessor’s output. Thus, a k-layer fully connected network can be viewed as a chain of its sub-functions,
(1) |
where each layer is denoted by . In the following, we cover backpropagation and forward-forward algorithms. In this work, we also point out some of their differences, as well as pros and cons.
2.1 Backpropagation
Backpropagation-based learning optimizes the parameters of a network given a set of training data points and a pre-defined loss function . During tuning, one or multiple data points are forwarded through the network to arrive at a prediction. The loss considers the output of the network and uses it to quantify the performance of the network. Formally, the optimization seeks to address the program:
(2) |
This program is often addressed by gradient-based methods, such as stochastic gradient descent (SGD) and its variants. For example, the optimization step for the i-th layer is:
(3) |
Consequently, during backpropagation-based learning, information first flows from the input to the output in the forward pass. Conversely, in the backward pass, information flows from the output to the input (see Figure 1).
2.2 The Forward-Forward algorithm
The forward-forward algorithm provides an alternative method to learn deep neural network parameters. Different from backpropagation, the samples are represented as the raw input data point linked with a one-hot encoding vector that points to a label . Then, it splits the training data into ‘positive’ and ‘negative’ sets. The one-hot vector in the positive samples is pointing at the true label. For the negative samples, the label vector is pointing at a randomly chosen wrong label. Intuitively, it seeks to learn whether a linked sample is positive or negative.
The main idea behind the forward-forward algorithm: make each layer ‘excited’ about positive samples and, at the same time, make each layer less excited about negative samples. Formally, given a data point linked with a label, we seek to classify whether this set is positive or negative based on the goodness of the network. Let us define the goodness of the i-th layer as its squared sum of ReLU activities. The probability of a sample to be positive given by the i-th layer is computed via
(4) |
where is a hyperparameter that controls the scale of the gradient. Normalization is applied between layers such that later layers aren’t able to rely on their predecessor’s decision.
Finally, to perform learning, we optimize each layer sequentially to maximize goodness on positive samples and minimize negative ones. Namely, we train each layer once until convergence and then optimize its successor. Using SGD for layer optimization, we obtain a positive optimization step for the i-th layer via
(5) |
Similarly, the negative step is performed with the opposite sign. Throughout the paper, we also refer to this loss as ‘Layer loss.’ Different from backpropagation, this allows an individual training process for each layer, and, as a result, one can learn layers with non-differentiable functions in between the optimized layers, such as solvers (Indelman and Hazan, 2021), argmax operation (Lorberbom et al., 2019), etc. Notice, similarly to energy-based models (LeCun and Huang, 2005), the forward-forward algorithm assigns a score to every pair. However, instead of training the network using backpropagation, the network is trained layer-by-layer.
Different from the training step, the inference process requires assessing the goodness of the network for each possible label. Then, the predicted label is determined by their goodness. The chosen label is the one with the highest overall goodness.






3 Layer collaboration
Deep nets consist of multiple layers. Each layer processes the input data and passes it to the next layer, allowing the network to learn increasingly complex representations of the data. Thus, layer collaboration is crucial in deep nets because it allows the network to learn and extract features from the data in a hierarchical and efficient manner. For example, lower layers of the network might learn to identify ‘simple’ features such as edges and textures, while later layers might learn complex features such as object shapes and patterns. This hierarchical structure allows the network to build a robust representation of the data and make accurate predictions.
The forward-forward algorithm permits communication between layers only through the forward pass because each layer only takes into account the output of its predecessor. In this work, we find that the forward collaboration solely is not sufficient for developing a complex representation of the data. One can observe from Equation 5 that the forward-forward process detailed above does not enable the flow of information to earlier layers (i.e., layers closer to the data) while training (see Figure 1). For example, the first layer is not aware of the existence of later layers and hence can not consider information from them in its optimization process. This is contrary to the backpropagation algorithm. From Equation 3 we observe that information can flow from a layer to its predecessors through the gradients of the optimization step.
In the following, we examine the collaboration between layers when training models using the forward-forward algorithm. We then propose a modification to the forward-forward algorithm that enables better collaboration between layers.
3.1 Analysis
The forward-forward inference process takes into account the sum of the goodnesses of the model’s layers. This process can be viewed as a vote among the layers regarding the label that is most likely to fit the data sample . Such an independent voting process allows us to examine the performance of each layer individually by considering their goodness. Similarly, one can evaluate the performance of subgroups of layers.
Figures 2 (a-c) present for three datasets (MNIST, Fashion-MNIST, and CIFAR-10) the error of each layer individually and when they are combined to operate together. Those figures empirically verify that the forward-forward algorithm is unable to learn an intricate and collective representation. To see this, note that individually, each layer consistently performs better than its successor. In addition, together, the marginal contribution of each layer to the performance is negative (i.e., higher error) and the first layer alone achieves the lowest error. This is an undesired property for deep neural networks. This implies that, in its current form, the forward-forward algorithm performs better using shallow networks, specifically, one layer networks.
Next, we propose a modification to the forward-forward algorithm that allows layers to collaborate regardless of their position in the network.
3.2 Collaborative Forward-Forward
The observations in Section 3.1 encourage to seek a method that enables layers to better collaborate during training. As discussed before, breaking down the forward-forward training procedure reveals that early layers are not aware of their predecessors. In the following, we propose a simple, yet highly effective modification to the forward-forward algorithm. Concretely, we propose to consider the global goodness of the network for the optimization of each layer.
The threshold in the forward-forward algorithm governs the magnitude of the learning. To account for the global goodness of the network with respect to a given sample , we propose to add the goodness of all layers to a given layer optimization step. Formally, we propose that the modified probability for positive samples reads as follows,
(6) |
where, the constant is a sum of goodness values from different layers. For example,
(7) |
Importantly, is a copy of with fixed parameters (i.e., we do not take gradients with respect to ). The proposed modification results in the following modified update step, with respect to :
(8) |
Such an update step allows information flow between layers. Intuitively, this ensures that the i-th layer is aware of its later layers and its parameters are trained accordingly. It is important to note that our method maintains the original forward-forward assumptions and requirements.

We illustrate our method in Figure 3. The dashed lines represent , which governs the magnitude of the gradient. Algorithms 1 and 2 illustrate the difference between the layer and the unified loss. The layer loss requires to manually set the hyperparameter , which turns out to be sensitive and its value varies between tasks. Moreover, it is not clear how should be modified during training. Our method modifies the value of based on the progress of the training, while considering performance and collaboration between layers.
We also suggest to modify the training procedure. The classical forward-forward algorithm trains each layer until convergence. Differently, we suggest to alternate the updates of all the layers’ parameters. The combination of the unified loss and the modified training procedure yields a collaborative forward-forward method. We later show that this simple modification of the forward-forward algorithm is highly effective and improves results.
4 Entropic view of layer collaboration
The functional entropy can measure the amount of information that is shared across layers in the forward-forward algorithm, and consequently allows us to quantify and compare layer collaboration. In the following section we formulate the connection between the forward-forward algorithm and functional entropy. Using this connection, we show that during training, the functional entropy of the network implicitly improves. In Section 5, we propose to directly use the functional entropy during training and show that it is beneficial to the performance.
The functional entropy is defined over a non-negative random variable, which we denote by that is defined over the space with a probability measure . Here and throughout, we use to refer to a stochastic variable, which we will integrate over. The functional entropy of the non-negative function is
(9) |
The functional entropy is non-negative, namely, and equals zero only if is a constant.
Our settings are defined over the input space where and . In the following, we denote by the data-label tuple combined with the layer index . We use to refer to its domain. The functional entropy can be intuitively considered as a scaled KL-divergence between the distribution of the input of a layer and the output of a layer . Formally, consider the prior and posterior distributions, , i.e.,
(10) |
One can verify that is a probability density function since and . The functional entropy can now be described as the scaled KL-divergence between the prior and posterior distributions above:
(11) |
Thus, intuitively, the functional entropy is the amount of learning that is done by a function , measured by the KL-divergence from the prior distribution of the input to the posterior distribution of the output .
In Figure 4 we validate that our method improves the overall functional entropy of the learned network. We attribute its improvement to our learning algorithm, which encourages functional entropy to flow between the layers of the network. To formulate this notion, we decompose the functional entropy by layers, and measure a notion of conditional functional entropy with respect to the layers themselves. The probability measure for accounts for the input of the -th layer when computing its output . The probability measure is which does not enforce statistical dependencies between the data generating distribution and the layer distribution , for which we take the uniform distribution. In this setting, the functional entropy decomposes into two components111This decomposition can be verified by adding and subtracting to .:
(12) |
The first component is the functional entropy of the layers’ distribution of the marginal function . This entropy accounts for the information across layers. The seconds component is an average of entropies that are conditioned on the i-th layer.




5 Results
In this section, we study various modifications to the forward-forward algorithm. Specifically, we begin by studying the collaboration of layers when training using classical and collaborative forward-forward. We study the collaboration from both a performance and an entropy point of view. Then, we present the benefits of our method when using activations other than ReLU. In addition, we find that the introduced functional entropy concept is highly effective in replacing the sigmoid in forward-forward optimization.
Experimental setup
Similar to Hinton (2022), we study all methods using the MNIST (Lecun et al., 1998) and CIFAR10 (Krizhevsky et al., 2009) datasets. We also study FashionMNIST (Xiao et al., 2017) as it is more challenging than MNIST. For all models, we use three layers with neurons in each layer. For the ‘Backpropagation Label Output’ model, we project last hidden layer to the label space using an extra layer, while for discriminator-based models, the label is concatenated with the input data. All models are optimized using the Adam (Kingma and Ba, 2014) optimizer. We apply a learning rate of and a batch size of . All models are trained on a single GPU (NVIDIA 2080 Ti) for epochs.
Method | Training | MNIST | Fashion-MNIST | CIFAR-10 |
Backpropagation Pairwise | Network | 2.0 | 10.7 | 45.4 |
Backpropagation Label | Network | 1.5 | 10.7 | 45.9 |
Forward-Forward | Layer | 3.3 | 13.2 | 54.2 |
Forward-Forward | Network | 3.4 | 13.6 | 53.7 |
Collaborative Forward-Forward | Layer | 2.2 | 11.6 | 53.2 |
Collaborative Forward-Forward | Network | 2.2 | 11.6 | 51.6 |
Collaborative Forward-Forward | Network | 2.1 | 12.0 | 52.7 |
5.1 Layer collaboration
This section covers collaboration between layers using various learning algorithms. We study information flow between layers when training a deep neural network with backpropagation, with forward-forward, and with our proposed modifications. In backpropagation-based models, early layers are updated according to their effect on later layers. Thus, they hierarchically learn features that are useful for the final output. In contrast, the forward-forward algorithm updates each layer separately without being aware of subsequent layers. This results in poor communication between layers and a limited information flow. Our results show that, in forward-forward-based models, the first layer on its own achieves the lowest prediction error and additional layers do not improve results further. This highlights the need for a modification that enables a better communication between layers. Our method allows layers to better take into account global information. We study this form of layer collaboration of the aforementioned methods from a performance and a functional information point of view.
For performance reference, we provide backpropagation-based models results. The first model presented, employs classical backpropagation. It consists of several layers and its output equals the size of the label space. The second model is trained to replicate the forward-forward methodology. Concretely, the input is the sample concatenated with a label . The model’s task is to discriminate between positive and negative samples, it is done by pairwise ranking of concatenated samples (Liu et al., 2009). In this model, in each optimization step, the last layer’s norm is maximized for positive samples and minimized for negative samples.
Performance.
Table 1 summarizes the performance of the evaluated methods. Under the forward-forward setup, the proposed approach performs the best, being superior to the classical forward-forward under all evaluated setups. We also perform ablation study for different choices of and training procedures. As expected, the backpropagation approach, in both classical and pairwise, reaches the best overall performance. Interestingly, optimizing the network using classical and pairwise objectives reaches comparable performance.
Entropy.
Next, we analyze the total entropy of the network for both classical and collaborative forward-forward, considering MNIST and Fashion-MNIST. Results are depicted in Figure 4. Results are reported for both positive and negative samples (Figure 4(a)), positive samples (Figure 4(b)), and negative samples (Figure 4(c)). Results suggest the proposed method better maximizes the entropy across all settings.
5.2 Entropy-based optimization
In Section 4 we observe that the forward-forward algorithm implicitly maximizes the entropy of the network. This raises the question of whether directly optimizing the entropy will boost performance. We formulate the learning process as an entropy maximization process. For the classical forward-forward optimization, this results in maximizing the following objective for the i-th layer:
(13) |
Note, is fixed, and the expectation is taken with respect to . It is straightforward to modify the above equation to the unified loss. It can be done similar to Equation 8, i.e., by adding the goodness of all other layers without considering their gradients.
This proposed optimization process, which does not include the sigmoid function, nor obtains on-par results to the classical and collaborative forward-forward. For the Fashion-MNIST dataset, considering layer optimization, it obtains an error of (vs. the original ) and for the collaborative version it obtains (vs. ).
6 Related work
Our work focuses on the collaboration between layers in forward-forward-based models. In this section, we review research directions that are related to the forward-forward method, layer collaboration in deep neural networks, and functional entropy.
Forward-Forward algorithm.
The forward-forward algorithm assigns a score to each pair using the goodness function. This process is similar to energy-based models (EBMs) which assign an energy to each similar pair. EBMs are a class of generative machine learning models that can to model complex probability distributions (LeCun and Huang, 2005; LeCun et al., 2006). They are based on the idea that the probability of a particular configuration of the network is determined by its energy. The energy is defined as a function of the network’s parameters, and the probability of a configuration given by the Boltzmann distribution, which is proportional to the exponential of the negative energy. Energy-based models have been used in a variety of applications, such as image generation, natural language processing, and reinforcement learning (Ngiam et al., 2011; Liu and Wang, 2017; Haarnoja et al., 2017; Sharma et al., 2021). One of the most popular types of energy-based models is the Boltzmann machine (Salakhutdinov and Hinton, 2009).
Another line of related work discusses the Restricted Boltzmann Machine (RBM) (Fischer and Igel, 2012), which, similarly to the forward-forward algorithm, trains the network layer-by-layer. A notable difference: RBMs are unsupervised learning models, while the forward-forward method is a fully supervised.
Layer collaboration.
Various works have studied features learned using neural networks. Bottom-up approaches observe that networks start with simple features and build up to more complex ones. For example, in image recognition tasks, the network may first learn to recognize edges, then shapes, and finally objects (Workman and Jacobs, 2015; Liu et al., 2021). In top-down approaches, the network starts with high level features and then breaks them down into lower level ones. In these approaches, high level features are used to guide the learning of lower level features. For example, in natural language processing tasks, the network may use high level semantic features to guide the learning of lower level syntactic features (Maulud et al., 2021; Adel et al., 2015). Both bottom-up and top-down approaches have been used in combination in several works, e.g., by (Mahdi et al., 2019; Anderson et al., 2018). These works have demonstrated that using high level features can help the network to learn more efficiently and improve the performance of the model. In general, the high-level features learned by the network are more abstract and semantically meaningful, while low-level features are more specific to the task at hand. This type of hierarchical representation learning is possible only when layers can collaborate with each other.
Functional entropy.
Functional entropy is gaining an increasing amount of attention from the machine learning community (Jahnel and Köppl, 2023; Gat et al., 2022a; Anceschi and Zhu, 2021; Gat et al., 2022b; Evans, 2021). Gat et al. (2022c) used it to study the importance of the gradient norm for a PAC-Bayes generalization bound. Different from our work, the authors are able to study gradient flow in the network as it is differentiable. Gat et al. (2020) maximize entropies of data modalities as a regularization of biased datasets. This is different from our work since they operate over differentiable models and data modalities (i.e., not parameters). In this work, we find the functional entropy to be useful as a guiding concept for collaboration between layers in forward-forward-based models.
7 Conclusions and limitations
This work studies layer collaboration of models that are trained using the forward-forward algorithm. We find that, in its current form, the forward-forward algorithm is limited in its collaborative capabilities, meaning information can not propagate backward throughout the network. We study this phenomenon and propose a method to overcome this challenge. We further propose an entropic-based view of the forward-forward algorithm, and, through extensive experiments, show the benefits of following the proposed method.
Nevertheless, the forward-forward algorithm is far from being comparable to the standard backpropagation. Many questions remain open as to how it can be applied to complex applications, e.g., image classification, text analysis, generative learning, and modern neural architectures, e.g., convnets, transformers, etc. We hope the community will keep pursing such research directions and challenge popular methods.
References
- Hinton [2022] G. Hinton. The forward-forward algorithm: Some preliminary investigations. ArXiv, 2212.13345, 2022.
- Lavin and Gray [2016] Andrew Lavin and Scott Gray. Fast algorithms for convolutional neural networks. In CVPR, 2016.
- Bojarski et al. [2016] Mariusz Bojarski, Davide Del Testa, Daniel Dworakowski, Bernhard Firner, Beat Flepp, Prasoon Goyal, Lawrence D Jackel, Mathew Monfort, Urs Muller, Jiakai Zhang, et al. End to end learning for self-driving cars. In arXiv preprint arXiv:1604.07316, 2016.
- Luss et al. [2019] Ronny Luss, Pin-Yu Chen, Amit Dhurandhar, Prasanna Sattigeri, Yunfeng Zhang, Karthikeyan Shanmugam, and Chun-Chen Tu. Generating contrastive explanations with monotonic attribute functions. In arXiv preprint arXiv:1905.12698, 2019.
- Gat et al. [2021] Itai Gat, Idan Schwartz, and Alex Schwing. Perceptual score: What data modalities does your model perceive? In NeurIPS, 2021.
- Morais et al. [2022] Edmilson Morais, Ron Hoory, Weizhong Zhu, Itai Gat, Matheus Damasceno, and Hagai Aronowitz. Speech emotion recognition using self-supervised features. In ICASSP, 2022.
- Rosenberg et al. [2021] Daniel Rosenberg, Itai Gat, Amir Feder, and Roi Reichart. Are VQA systems rad? measuring robustness to augmented data with focused interventions. 2021.
- Ophir et al. [2020] Yaakov Ophir, Refael Tikochinski, Christa SC Asterhan, Itay Sisso, and Roi Reichart. Deep neural networks detect suicide risk from textual facebook posts. Scientific reports, 2020.
- Zhou et al. [2021] S Kevin Zhou, Hayit Greenspan, Christos Davatzikos, James S Duncan, Bram Van Ginneken, Anant Madabhushi, Jerry L Prince, Daniel Rueckert, and Ronald M Summers. A review of deep learning in medical imaging: Imaging traits, technology trends, case studies with progress highlights, and future promises. Proceedings of the IEEE, 2021.
- Clauset and Post [2019] Aaron Clauset and Kollen Post. Machine learning improves hearing aids. Science, 2019.
- Fedorov et al. [2020] Igor Fedorov, Marko Stamenovic, Carl Jensen, Li-Chia Yang, Ari Mandell, Yiming Gan, Matthew Mattina, and Paul N Whatmough. Tinylstms: Efficient neural speech enhancement for hearing aids. arXiv preprint arXiv:2005.11138, 2020.
- Rumelhart et al. [1986] David E Rumelhart, Geoffrey E Hinton, and Ronald J Williams. Learning representations by back-propagating errors. nature, 1986.
- Indelman and Hazan [2021] Hedda Cohen Indelman and Tamir Hazan. Learning randomly perturbed structured predictors for direct loss minimization. In International Conference on Machine Learning. PMLR, 2021.
- Lorberbom et al. [2019] Guy Lorberbom, Andreea Gane, Tommi Jaakkola, and Tamir Hazan. Direct optimization through for discrete variational auto-encoder. Advances in neural information processing systems, 32, 2019.
- LeCun and Huang [2005] Yann LeCun and Fu Jie Huang. Loss functions for discriminative training of energy-based models. In International workshop on artificial intelligence and statistics, 2005.
- Lecun et al. [1998] Yann Lecun, Léon Bottou, Yoshua Bengio, and Patrick Haffner. Gradient-based learning applied to document recognition. In IEEE, 1998.
- Krizhevsky et al. [2009] Alex Krizhevsky, Vinod Nair, and Geoffrey Hinton. Cifar-10 (canadian institute for advanced research). 2009.
- Xiao et al. [2017] Han Xiao, Kashif Rasul, and Roland Vollgraf. Fashion-mnist: a novel image dataset for benchmarking machine learning algorithms. arXiv preprint arXiv:1708.07747, 2017.
- Kingma and Ba [2014] Diederik P Kingma and Jimmy Ba. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
- Liu et al. [2009] Tie-Yan Liu et al. Learning to rank for information retrieval. Foundations and Trends® in Information Retrieval, 3(3):225–331, 2009.
- LeCun et al. [2006] Yann LeCun, Sumit Chopra, Raia Hadsell, M Ranzato, and Fujie Huang. A tutorial on energy-based learning. Predicting structured data, 2006.
- Ngiam et al. [2011] Jiquan Ngiam, Zhenghao Chen, Pang W Koh, and Andrew Y Ng. Learning deep energy models. In ICML, 2011.
- Liu and Wang [2017] Qiang Liu and Dilin Wang. Learning deep energy models: Contrastive divergence vs. amortized mle. arXiv preprint arXiv:1707.00797, 2017.
- Haarnoja et al. [2017] Tuomas Haarnoja, Haoran Tang, Pieter Abbeel, and Sergey Levine. Reinforcement learning with deep energy-based policies. In ICML, 2017.
- Sharma et al. [2021] K Sharma, Bhopendra Singh, Edwin Herman, R Regine, S Suman Rajest, and Ved P Mishra. Maximum information measure policies in reinforcement learning with deep energy-based model. In ICCIKE, 2021.
- Salakhutdinov and Hinton [2009] Ruslan Salakhutdinov and Geoffrey Hinton. Deep boltzmann machines. In ICAIS, volume 5, 2009.
- Fischer and Igel [2012] Asja Fischer and Christian Igel. An introduction to restricted boltzmann machines. In CVPR, 2012.
- Workman and Jacobs [2015] Scott Workman and Nathan Jacobs. On the location dependence of convolutional neural network features. In CVPR, 2015.
- Liu et al. [2021] Caifeng Liu, Lin Feng, Guochao Liu, Huibing Wang, and Shenglan Liu. Bottom-up broadcast neural network for music genre classification. 2021.
- Maulud et al. [2021] Dastan Hussen Maulud, Subhi RM Zeebaree, Karwan Jacksi, Mohammed A Mohammed Sadeeq, and Karzan Hussein Sharif. State of art for semantic analysis of natural language processing. Qubahan Academic Journal, 2021.
- Adel et al. [2015] Heike Adel, Ngoc Thang Vu, Katrin Kirchhoff, Dominic Telaar, and Tanja Schultz. Syntactic and semantic features for code-switching factored language models. IEEE/ACM transactions on audio, speech, and language Processing, 2015.
- Mahdi et al. [2019] Ali Mahdi, Jun Qin, and Garth Crosby. Deepfeat: A bottom-up and top-down saliency model based on deep features of convolutional neural networks. IEEE Transactions on Cognitive and Developmental Systems, 2019.
- Anderson et al. [2018] Peter Anderson, Xiaodong He, Chris Buehler, Damien Teney, Mark Johnson, Stephen Gould, and Lei Zhang. Bottom-up and top-down attention for image captioning and visual question answering. In CVPR, 2018.
- Jahnel and Köppl [2023] Benedikt Jahnel and Jonas Köppl. Trajectorial dissipation of phi -entropies for interacting particle systems. arXiv preprint arXiv:2301.03922, 2023.
- Gat et al. [2022a] Itai Gat, Guy Lorberbom, Idan Schwartz, and Tamir Hazan. Latent space explanation by intervention. In AAAI, volume 36, 2022a.
- Anceschi and Zhu [2021] Francesca Anceschi and Yuzhe Zhu. On a spatially inhomogeneous nonlinear fokker-planck equation: Cauchy problem and diffusion asymptotics. arXiv preprint arXiv:2102.12795, 2021.
- Gat et al. [2022b] Itai Gat, Nitay Calderon, Roi Reichart, and Tamir Hazan. A functional information perspective on model interpretation. In ICML, 2022b.
- Evans [2021] Josephine Evans. Hypocoercivity in phi-entropy for the linear relaxation boltzmann equation on the torus. SIAM, 53, 2021.
- Gat et al. [2022c] Itai Gat, Yossi Adi, Alexander Schwing, and Tamir Hazan. On the importance of gradient norm in pac-bayesian bounds. In NeurIPS, volume 35, 2022c.
- Gat et al. [2020] Itai Gat, Idan Schwartz, Alexander Schwing, and Tamir Hazan. Removing bias in multi-modal classifiers: Regularization by maximizing functional entropies. NeurIPS, 33, 2020.