This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Deep Learning without Weight Symmetry

Li Ji-An
Neurosciences Graduate Program
University of California, San Diego
La Jolla, CA 92093
[email protected]
&Marcus K. Benna
Department of Neurobiology
University of California, San Diego
La Jolla, CA 92093
[email protected]
Abstract

Backpropagation (BP), a foundational algorithm for training artificial neural networks, predominates in contemporary deep learning. Although highly successful, it is often considered biologically implausible. A significant limitation arises from the need for precise symmetry between connections in the backward and forward pathways to backpropagate gradient signals accurately, which is not observed in biological brains. Researchers have proposed several algorithms to alleviate this symmetry constraint, such as feedback alignment and direct feedback alignment. However, their divergence from backpropagation dynamics presents challenges, particularly in deeper networks and convolutional layers. Here we introduce the Product Feedback Alignment (PFA) algorithm. Our findings demonstrate that PFA closely approximates BP and achieves comparable performance in deep convolutional networks while avoiding explicit weight symmetry. Our results offer a novel solution to the longstanding weight symmetry problem, leading to more biologically plausible learning in deep convolutional networks compared to earlier methods.

1 Introduction

Both artificial and biological neural networks must orchestrate complex synaptic weight updates in order to improve task performance. The correct organization of these weight updates becomes even more challenging in deeper multilayer neural networks, often referred to as the credit assignment problem. Over the past decades, the error backpropagation (BP) algorithm has revolutionized contemporary deep learning [1], serving as a fundamental algorithm for training artificial neural networks.

Despite its success, BP is frequently considered biologically implausible. Although recent proposals have (partially) addressed many biological implausibilities, such as nonlocal plasticity, multiple separate learning phases, and non-biological error representations [2, 3, 4, 5], a significant limitation known as the weight symmetry (or weight transport) problem persists [1, 6, 7, 8]. In BP, following the chain rule, the feedback weights WTW^{T} in the backward pass are precisely symmetric to the feedforward weights WW in the forward pass in order to accurately transmit error signals that match the gradients of the cost function. To implement BP in the brain, the error signals should be locally available for the feedforward weights, which implies that if wabw_{a\rightarrow b} is the feedforward weight from neuron aa to neuron bb, there should exist a symmetric feedback weight from neuron bb to neuron aa (such that wba=wabw_{b\rightarrow a}=w_{a\rightarrow b}). Why is this symmetric weight pattern not observed in the biological brain? Indeed, in local cortical circuits, two connected neurons aa and bb are either unidirectionally linked, with a probability of 69% (aba\rightarrow b or bab\rightarrow a), or bidirectional connected, with a probability of 31% (aba\rightarrow b and bab\rightarrow a) [9]. Even for bidirectionally connected neurons, wabw_{a\rightarrow b} and wbaw_{b\rightarrow a} are only modestly correlated (R0.36R\approx 0.36) [9]. These observations are in striking contrast to the perfect correlation (R=1R=1) for symmetric weight connectivity assumed by BP. This discrepancy thus calls for alternative explanations involving algorithms that biological mechanisms can implement.

In an effort to eliminate the symmetry assumption, it was demonstrated that random fixed feedback weights BB can transmit useful error signals to upstream layers, leading to a learning process whereby the feedforward weights WW approximately align with the feedback weights, i.e., WBTW\propto B^{T} (feedback alignment, FA) [10]. However, FA struggles to match BP’s performance in more advanced network architectures and in more challenging tasks, including deeper networks, convolutional layers, and large-scale image datasets (e.g., CIFAR10, ImageNet) [11]. A variant of FA, direct feedback alignment (DFA) [12], transmits error signals directly from the output layer to each upstream (hidden) layer, but also suffers from severe performance loss compared to BP [11]. Another FA variant was proposed by meta-learning the plasticity rule of feedforward weights to improve FA for online learning and low-data regimes [13], but it still significantly underperforms relative to BP.

Other proposals have explored various methods for updating feedback weights. In sign-concordant feedback (SF) algorithms, the sign of feedforward weights is transported (i.e., copied) to the feedback weights, both at initialization and during training [14]. This approach has shown considerable improvement over FA, approaching BP’s performance in simple tasks (though still with a significant performance gap in more complex tasks like ImageNet) [15, 16, 17]. Nonetheless, it remains unclear whether a biologically plausible plasticity rule can effectively transport the sign of the synaptic weight from the forward path to the corresponding synaptic weight in the backward path. In the weight mirror (WM) algorithm [18], the feedback weights are updated to track the feedforward weights, by injecting random noise into neurons during multiple learning phases (one phase for each layer). WM can reach a performance similar to BP; however, the biological feasibility of layer-specific learning phases and “bias blocking” (setting bias to zero) during the mirror mode remains to be established. The phaseless alignment learning (PAL) algorithm eliminates the need for the additional mirror mode, but again deviates from BP’s dynamics, showing a significant performance gap compared to BP [19]. The Kollen-Pollack (KP) algorithm [18], which uses a feedback weight update symmetric to the feedforward weight update and includes weight decay, has been shown to closely approximate BP, achieving similar performance. However, even starting from asymmetric initializations, both WM and KP ultimately lead to a scenario in which feedforward and feedback weights are symmetric, a connectivity pattern not observed in the brain. In short, KP and WM can achieve a BP-level task performance, but converge to a configuration with almost exact weight symmetry. FA, SF, and PAL alleviate the weight symmetry issue (although their weight configurations are still more aligned than and arguably incompatible with the biological observations), but significantly sacrifice task performance. None of these algorithms manages to achieve a BP-level performance while completely avoiding explicit weight symmetry.

In this study, we propose the Product Feedback Alignment (PFA) algorithm, which closely approximates BP. It completely avoids explicit weight symmetry by relying on alignment between forward and indirect backward pathways, using an additional population of neurons. Specifically, the feedforward weights WW align with the product of a pair of feedback weights RR and BB (such that W(RB)TW\propto(RB)^{T}). We show that PFA can achieve BP-level performance in deeper networks, convolutional layers, and more challenging datasets such as CIFAR10 and ImageNet. Further, PFA can outperform other algorithms with sparse feedback connections, an important biological constraint of the brain. Our results offer a novel solution to the longstanding weight symmetry problem, providing supportive evidence for the feasibility of implementing BP-like algorithms in the brain.

BP FA DFA SF KP WM PAL PFA
No need to transport weight sign
No need to transport weight magnitude
No separate feedback weight learning phase
No explicit weight symmetry after training
Accurate approximation to BP (path alignment)
BP-level task performance
Table 1: Detailed comparison of algorithms. BP: backpropagation. FA: feedback alignment. DFA: direct feedback alignment. SF: sign-concordant feedback. KP: Kollen-Pollack algorithm. WM: weight mirror. PAL: phaseless alignment learning. PFA: product feedback alignment. ✛: It is unclear how the feedback weights in SF can be learned in a biologically plausible way. ✽: these algorithms reduce, but do not fully eliminate explicit weight symmetry. ✦: These algorithms significantly outperform FA and DFA, but still underperform compared to BP in more challenging tasks (CIFAR10 for PAL and ImageNet for SF).

2 Product feedback alignment

Consider a fully-connected multilayer neural network with depth LL, mapping the input 𝒙0\bm{x}_{0} to the output 𝒙L\bm{x}_{L}. In the forward pass, the activation of layer l+1l+1 (with Nl+1N_{l+1} neurons), denoted as 𝒙l+1\bm{x}_{l+1}, is determined by

𝒙l+1=σ(Wl+1,l𝒙l+𝒃l+1),\bm{x}_{l+1}=\sigma(W_{l+1,l}\bm{x}_{l}+\bm{b}_{l+1})\ , (1)

where σ\sigma is the activation function, Wl+1,lW_{l+1,l} is the Nl+1×NlN_{l+1}\times N_{l} feedforward weight, and 𝒃l+1\bm{b}_{l+1} is the bias. For a training dataset consisting of data points (𝒙0,𝒚)(\bm{x}_{0},\bm{y}), the set of parameters Wl+1,l{W_{l+1,l}} and 𝒃l\bm{b}_{l} are trained to minimize the loss function \mathcal{L} that measures the difference between layer LL’s output 𝒙L\bm{x}_{L} and the target output 𝒚\bm{y} (e.g., the cross-entropy).

The loss at the output layer directly provides the teaching signal (error) 𝒆L\bm{e}_{L} that is locally available at 𝒙L\bm{x}_{L}, defined as the gradient 𝒆L=/𝒙L\bm{e}_{L}=-\partial\mathcal{L}/\partial\bm{x}_{L}. In the backward pass of BP (Fig. 1), the error (i.e., gradient) at layer 𝒙l\bm{x}_{l} is iteratively backpropagated as

𝒆lBP=σ(𝒙l)Wl+1,lT𝒆l+1BP,\bm{e}_{l}^{\rm BP}=\sigma^{\prime}(\bm{x}_{l})\odot W_{l+1,l}^{T}\bm{e}_{l+1}^{\rm BP}\ , (2)

where \odot is the Hadamard product. Subsequently, the feedforward weight Wl+1,lW_{l+1,l} is updated by

ΔWl+1,l=η𝒆l+1BP𝒙lT,\Delta W_{l+1,l}=\eta\bm{e}_{l+1}^{\rm BP}\bm{x}_{l}^{T}\ , (3)

where η\eta is the learning rate. If we assume that el+1ie_{l+1}^{i} (ii-th component of 𝒆l+1\bm{e}_{l+1}) is locally available at the neuron with activation xl+1ix_{l+1}^{i}, then the update of Wl+1,li,jW_{l+1,l}^{i,j} (forward synaptic weight between xl+1ix_{l+1}^{i} and xljx_{l}^{j}) relies only on locally available information (the product of el+1ie_{l+1}^{i} and xljx_{l}^{j}), which is biologically plausible. The aspect that lacks biological plausibility is the backpropagation of errors via Wl+1,lTW_{l+1,l}^{T}, since the same synaptic weight Wl+1,li,jW_{l+1,l}^{i,j} is used twice in both the forward and backward passes.

Because chemical synapses in the brain transmit information (carried by presynaptic spikes) only in one direction (from axon to dendrite), one must introduce an additional set of feedback weights Bl,l+1B_{l,l+1} to backpropagate the error signals from the output layer to upstream layers. FA employs fixed random feedback weights Bl,l+1FAB_{l,l+1}^{\rm FA} (Nl×Nl+1N_{l}\times N_{l+1}) that are independent of the feedforward weights, alleviating the weight symmetry problem. Then the error at the layer 𝒙l\bm{x}_{l} is computed as 𝒆lFA=σ(𝒙l)Bl,l+1FA𝒆l+1FA\bm{e}_{l}^{\rm FA}=\sigma^{\prime}(\bm{x}_{l})\odot B^{\rm FA}_{l,l+1}\bm{e}_{l+1}^{\rm FA}. DFA similarly calculates the error at the layer 𝒙l\bm{x}_{l} using 𝒆lDFA=σ(𝒙l)Bl,LDFA𝒆L\bm{e}_{l}^{\rm DFA}=\sigma^{\prime}(\bm{x}_{l})\odot B_{l,L}^{\rm DFA}\bm{e}_{L} via a fixed random feedback weight Bl,LDFAB_{l,L}^{\rm DFA} (Nl×NLN_{l}\times N_{L}). As shown in previous works, the feedforward weights Wl+1,lW_{l+1,l} in FA learn to gradually align with feedback weights Bl,l+1TB_{l,l+1}^{T}, leading to an approximate weight symmetry (to an extent that may still contradict biological observations). However, both FA and DFA failed to reach good performance for deeper networks, convolutional layers, and more challenging tasks. SF transports the sign of feedforward weights into backward weights, i.e., Bl,l+1SF=Sign(Wl+1,lT)B^{\rm SF}_{l,l+1}=\text{Sign}(W^{T}_{l+1,l}), and the error 𝒆l\bm{e}_{l} is computed as 𝒆lSF=σ(𝒙l)Bl,l+1SF𝒆l+1SF\bm{e}_{l}^{\rm SF}=\sigma^{\prime}(\bm{x}_{l})\odot B^{\rm SF}_{l,l+1}\bm{e}_{l+1}^{\rm SF}. SF significantly outperforms FA and DFA, but still underperforms compared to BP in deep convolutional layers for challenging tasks (e.g., ImageNet), and again exhibits approximate weight symmetry [17].

In PFA (Fig. 1), we introduce an intermediate population 𝒆¯l\bm{\bar{e}}_{l} (with N¯l\bar{N}_{l} neurons) for each layer ll, denoted as

𝒆¯l=Bl,l+1PFA𝒆l+1PFA,\bm{\bar{e}}_{l}=B_{l,l+1}^{\rm PFA}\bm{e}_{l+1}^{\rm PFA}\ , (4)

using a fixed random feedback weights Bl,l+1PFAB_{l,l+1}^{\rm PFA} (N¯l×Nl+1\bar{N}_{l}\times N_{l+1}). Subsequently, the PFA error for layer ll (with activations 𝒙l\bm{x}_{l}) is calculated as

𝒆lPFA=σ(𝒙l)Rl,l𝒆¯𝒍,\bm{e}_{l}^{\rm PFA}=\sigma^{\prime}(\bm{x}_{l})\odot R_{l,l}\bm{\bar{e}_{l}}\ , (5)

where Rl,lR_{l,l} represents a plastic feedback weight matrix (Nl×N¯lN_{l}\times\bar{N}_{l}). In addition to the feedforward weights Wl+1,lW_{l+1,l} updated by ΔWl+1,l=η𝒆l+1𝒙lT\Delta W_{l+1,l}=\eta\bm{e}_{l+1}\bm{x}_{l}^{T}, we also update Rl,lR_{l,l} as

ΔRl,l=η𝒙l𝒆¯lT.\Delta R_{l,l}=\eta\bm{x}_{l}\bm{\bar{e}}_{l}^{T}\ . (6)

Note that the update of the synaptic weight Rl,lj,iR_{l,l}^{j,i} only relies on locally available information xljx_{l}^{j} and e¯li\bar{e}_{l}^{i}. The backward path between layers l+1l+1 and ll consists of a pair of feedback weights Bl,l+1B_{l,l+1} (omitting superscripts from here on for simplicity) and Rl,lR_{l,l}. Weight decay is applied for both Wl+1,lW_{l+1,l} and Rl,lR_{l,l}, such that the influence of initializations gradually diminishes. We will provide an explanation for the PFA’s implicit alignment mechanism in Section 4. We note that when the weight decay and learning rate for Rl,lR_{l,l} are set to zero, PFA effectively implements FA. When Bl,l+1B_{l,l+1} is set to an identity matrix (with N¯l=Nl+1\bar{N}_{l}=N_{l+1}), PFA reduces to the standard KP algorithm.

Refer to caption
Figure 1: Comparison of learning algorithms for multilayer networks transmitting errors in the backward pass. BP: backpropagation. FA: feedback alignment. DFA: direct feedback alignment. SF: sign-concordant feedback. KP: Kollen-Pollack algorithm. WM: weight mirror. PAL: phaseless alignment learning. PFA: product feedback alignment.

3 PFA approximates BP in MNIST handwritten digit classification

We used the above learning algorithms to train a two-hidden-layer feedforward network (with layer sizes 784-512-512-10) with ReLU activation on the MNIST dataset (using the BioTorch framework 111https://github.com/jsalbert/biotorch (Apache-2.0 license) [17]). The expansion ratio (1/λ=N¯l/Nl+11/\lambda=\bar{N}_{l}/N_{l+1}) in PFA is set to 10. See Appendix A for training details. We found that PFA, PFA-o (a variant of the algorithm in which Bl,l+1B_{l,l+1} is chosen to be a semi-orthogonal matrix), and SF reach a test performance similar to BP, slightly outperforming FA and DFA (Fig. 2a).

We recorded several metrics throughout the training process. The first metric, backward-forward weight alignment [17], quantifies the angle between the feedforward weights and the corresponding feedback weights for FA, derived from the normalized inner products (Vec(Wl+1,lT),Vec(Bl,l+1))(\textbf{Vec}(W_{l+1,l}^{T}),\textbf{Vec}(B_{l,l+1})). We similarly define the backward-forward path alignment (Vec(Wl+1,lT),Vec(Rl,lBl,l+1))(\textbf{Vec}(W_{l+1,l}^{T}),\textbf{Vec}(R_{l,l}B_{l,l+1})) for PFA. Our PFA achieved an angle around 1818^{\circ} (Fig. 2b) for all layers in the network after the initial epochs (because we set a large initial weight decay for RR and WW). Due to the orthogonal initialization for BB (Fig. 2c), PFA-o gradually achieved an angle close to 00^{\circ}. These are substantially lower than the weight alignment angles for FA (between 6060^{\circ} and 9090^{\circ}; Fig. 2d) and SF (around 3030^{\circ}; Fig. 2e), indicating a superior alignment in PFA (but crucially without weight symmetry).

The second metric, backward-forward weight norm ratio, is designed to capture the risk of experiencing gradient exploding/vanishing problems observed in FA [16, 17], and is defined as Bl,l+12/Wl+1,lT2||B_{l,l+1}||_{2}/||W_{l+1,l}^{T}||_{2}. We similarly defined the backward-forward path norm ratio Rl,lBl,l+12/Wl+1,lT2||R_{l,l}B_{l,l+1}||_{2}/||W_{l+1,l}^{T}||_{2} for PFA. Our PFA and PFA-o attained a weight ratio near 11 for all layers in the network – the ideal ratio akin to BP – after the initial epochs, suggesting better stability in error propagation over FA (between 0.3 and 1.1) and SF (between 0.5 and 1.2) (Fig. 2b-e). Collectively, these findings imply that PFA achieves a close approximation to BP.

Refer to caption
Figure 2: Characterization of learning algorithms for two-hidden-layer feedforward networks trained to classify MNIST digit images. (a) Task performance. Shaded regions show standard deviations across 5 seeds. PFA and PFA-o curves are almost overlapping with the BP curve, suggesting a close approximation. (b-e) Backward-forward weight alignment for FA/SF, and path alignment for PFA/PFA-o (top). Backward-forward weight norm ratio for FA/SF, and path norm ratio for PFA/PFA-o (bottom).

4 Alignment mechanism between forward and backward paths in PFA

We now elucidate the alignment mechanism between the forward and backward pathways. During training, the feedforward weights Wl+1,lW_{l+1,l} align with the feedback weight matrix product (Rl,lBl,l+1)T(R_{l,l}B_{l,l+1})^{T} when the following two requirements are satisfied.

(1) Bl,l+1TBl,l+1B_{l,l+1}^{T}B_{l,l+1} approximates an identity matrix. Then the weight updates for Wl+1,lW_{l+1,l} and (Rl,lBl,l+1)T(R_{l,l}B_{l,l+1})^{T} are aligned:

Δ(Rl,lBl,l+1)T=Bl,l+1TΔRl,lT=ηBl,l+1T𝒆¯l𝒙lT=η(Bl,l+1TBl,l+1)𝒆l+1𝒙lTη𝒆l+1𝒙lT=ΔWl+1,l.\begin{split}\Delta(R_{l,l}B_{l,l+1})^{T}&=B_{l,l+1}^{T}\Delta R_{l,l}^{T}=\eta B_{l,l+1}^{T}\bm{\bar{e}}_{l}\bm{x}_{l}^{T}=\eta(B_{l,l+1}^{T}B_{l,l+1})\bm{e}_{l+1}\bm{x}_{l}^{T}\\ &\approx\eta\bm{e}_{l+1}\bm{x}_{l}^{T}=\Delta W_{l+1,l}\ .\end{split} (7)

(2) The influence of the initial values of Wl+1,lW_{l+1,l} and Rl,lR_{l,l}, which causes misalignment between Wl+1,lW_{l+1,l} and (Rl,lBl,l+1)T(R_{l,l}B_{l,l+1})^{T}, further diminishes over successive learning epochs (due to weight decay).

Different approaches can ensure that Bl,l+1TBl,l+1B_{l,l+1}^{T}B_{l,l+1} approximates an identity matrix. One approach is to initialize Bl,l+1B_{l,l+1} as a semi-orthogonal (N¯l×Nl+1\bar{N}_{l}\times N_{l+1}) matrix (with N¯lNl+1\bar{N}_{l}\geq N_{l+1}), as used in PFA-o.

Alternatively, elements in Bl,l+1B_{l,l+1} can be sampled independently from a distribution with a mean of 0 and a variance of 1/N¯l1/\bar{N}_{l}, as used in PFA. This initialization for PFA is easier to implement in the brain than PFA-o, as it is not obvious what biological plasticity rule can learn a precisely semi-orthogonal matrix. Following the Marchenko–Pastur law [20] and assuming the limit Nl+1N_{l+1}\rightarrow\infty and N¯l\bar{N}_{l}\rightarrow\infty with the ratio Nl+1/N¯l=λ<1N_{l+1}/\bar{N}_{l}=\lambda<1, the eigenvalue density μ(v)\mu(v) of Bl,l+1TBl,l+1B_{l,l+1}^{T}B_{l,l+1} satisfies

μ(v)=12π(λ+v)(vλ)λv𝟙v[λ,λ+]\mu(v)=\frac{1}{2\pi}\frac{\sqrt{(\lambda_{+}-v)(v-\lambda_{-})}}{\lambda v}\mathbbm{1}_{v\in[\lambda_{-},\lambda_{+}]} (8)

with λ±=(1±λ)2\lambda_{\pm}=(1\pm\sqrt{\lambda})^{2}. In the limit λ0\lambda\rightarrow 0, Bl,l+1TBl,l+1B_{l,l+1}^{T}B_{l,l+1} converges to the identity matrix with probability 1. See Fig. 3a for a numerical verification.

In short, although Wl+1,lW_{l+1,l} does not align with Rl,lR_{l,l} or Bl,l+1B_{l,l+1} individually, we still have Wl+1,lTRl,lBl,l+1W_{l+1,l}^{T}\approx R_{l,l}B_{l,l+1} (see Fig. 3b) and thus Wl+1,lT𝒆l+1Rl,lBl,l+1𝒆l+1W_{l+1,l}^{T}\bm{e}_{l+1}\approx R_{l,l}B_{l,l+1}\bm{e}_{l+1}, signifying a good approximation to BP (in the limit of N¯i\bar{N}_{i}\rightarrow\infty for PFA). Like FA and DFA, PFA obviates the need to transport the feedforward weight (or its sign, as in SF) at initialization or during training, but significantly outperforms FA and DFA. Like KP, PFA employs aligned weight updates for both forward and backward pathways. However, unlike KP and WM, PFA does not lead to explicit weight symmetry, i.e., forward and backward synapses between any pair of neurons do not share identical weights for PFA (in fact we can avoid bidirectionally connected pairs of neurons altogether). For a detailed comparison of algorithms, see Table 1.

Refer to caption
Figure 3: (a) Eigenvalues of BTBB^{T}B as a function of the expansion ratio (1/λ1/\lambda). The shaded region shows the standard deviation of the eigenvalues. (b) Backward-forward path alignment between WW and (RB)T(RB)^{T} as a function of the expansion ratio (1/λ1/\lambda), where we randomly sampled WW and set RT=BWR^{T}=BW (expected to hold after the effect of the weight initialization has fully decayed). This simplification is consistent with the path alignment after training observed in simulations.

5 PFA for convolutional layers

Similar to FA and DFA for deep convolutional networks [16], PFA extends seamlessly to convolutional layers. Consider a forward convolutional kernel Wl+1,lh×w×Nl+1×NlW_{l+1,l}\in\mathbb{R}^{h\times w\times N_{l+1}\times N_{l}}, where hh is the kernel height, ww is the kernel width, and NlN_{l} is the number of channels in the ll-th layer. The feature-map activation xl+1ix_{l+1}^{i} for the ii-th channel at layer l+1l+1 is calculated as

xl+1i=σ(j=1NlWl+1,li,jxlj+bl+1i),x_{l+1}^{i}=\sigma(\sum_{j=1}^{N_{l}}W_{l+1,l}^{i,j}*x_{l}^{j}+b_{l+1}^{i})\ , (9)

where Wl+1,li,jW_{l+1,l}^{i,j} is the h×wh\times w kernel and * denotes convolution. The BP error (i.e., gradient) in the backward pass for jj-th channel at the ll-th layer is iteratively calculated as

elBP,j=σ(xl)i=1Nl+1W~l+1,li,jel+1BP,i,e_{l}^{{\rm BP},j}=\sigma^{\prime}(x_{l})\odot\sum_{i=1}^{N_{l+1}}\tilde{W}_{l+1,l}^{i,j}*e_{l+1}^{{\rm BP},i}\ , (10)

where \odot is the Hadamard product, and W~\tilde{W} is the a rotation of WW by 180180^{\circ} (flipped kernel) [21]. The FA error for jj-th channel at layer ll is calculated as elFA,j=σ(xl)i=1Nl+1Bl,l+1FA,j,iel+1FA,ie_{l}^{{\rm FA},j}=\sigma^{\prime}(x_{l})\odot\sum_{i=1}^{N_{l+1}}B_{l,l+1}^{{\rm FA},j,i}*e_{l+1}^{{\rm FA},i}, where Bl,l+1FAh×w×Nl×Nl+1B_{l,l+1}^{\rm FA}\in\mathbb{R}^{h\times w\times N_{l}\times N_{l+1}} is the backward kernel. For PFA, the intermediate error for the kk-th intermediate channel at the ll-th layer is

e¯lk=i=1Nl+1Bl,l+1k,iel+1PFA,i,\bar{e}_{l}^{k}=\sum_{i=1}^{N_{l+1}}B_{l,l+1}^{k,i}*e_{l+1}^{{\rm PFA},i}\ , (11)

where Bl,l+11×1×N¯l×NlB_{l,l+1}\in\mathbb{R}^{1\times 1\times\bar{N}_{l}\times N_{l}}. The PFA error for the jj-th channel at layer ll is calculated as

elPFA,j=σ(xl)k=1N¯lR~l,lj,ke¯lk,e_{l}^{{\rm PFA},j}=\sigma^{\prime}(x_{l})\odot\sum_{k=1}^{\bar{N}_{l}}\tilde{R}_{l,l}^{j,k}*\bar{e}_{l}^{k}\ , (12)

where Rl,lh×w×Nl×N¯lR_{l,l}\in\mathbb{R}^{h\times w\times N_{l}\times\bar{N}_{l}}. The pair of tensors Bl,l+1B_{l,l+1} and Rl,lR_{l,l} accommodate different convolutional hyperparameters for Wl+1,lW_{l+1,l} (e.g., stride, padding, dilation, groups). The weight updates are calculated similarly as ΔWl+1,li,j=ηel+1ixlj\Delta W_{l+1,l}^{i,j}=\eta e_{l+1}^{i}*x_{l}^{j} (in BP and PFA) and ΔRl,lj,k=ηxlje¯lk\Delta R_{l,l}^{j,k}=\eta x_{l}^{j}*\bar{e}_{l}^{k} (in PFA).

6 PFA closely approximates BP in deep convolutional networks

We trained ResNet-20 on the CIFAR-10 dataset for these learning algorithms [22, 17]. See Appendix A for training details. Our results (Fig. 4a) show that networks trained with PFA and PFA-o attain a test accuracy comparable to that of BP and SF, significantly surpassing that of FA and DFA. PFA and PFA-o again achieve a close approximation to BP (Fig. 4b-e), indicated by the small angle of path alignment and the close-to-one backward-forward path norm ratio.

We additionally trained ResNet-18 on ImageNet for these algorithms (except FA and DFA, which were previously shown to perform poorly [18, 17]). See Appendix A for training details. Our results (Fig. 5a) again show that networks trained with PFA (68.46%) and PFA-o (69.30%) achieve a test accuracy comparable to that of BP (69.69%), in a situation where even SF (64.63%) significantly underperforms. PFA and PFA-o again achieve a close approximation to BP (Fig. 5b-d), as shown by their path alignment and path norm ratio.

Refer to caption
Figure 4: Characterization of learning algorithms for ResNet-20 on CIFAR-10. (a) Task performance. Shaded regions show standard deviations across 5 seeds. PFA and PFA-o curves are almost overlapping with the BP curve, suggesting a close approximation. (b-e) Backward-forward weight alignment for FA/SF, and path alignment for PFA/PFA-o (top). Backward-forward weight norm ratio for FA/SF, and path norm ratio for PFA/PFA-o (bottom).
Refer to caption
Figure 5: Characterization of learning algorithms for ResNet-18 on ImageNet. (a) Task performance. PFA and PFA-o curves are almost overlapping with the BP curve, suggesting a close approximation. (b-d) Backward-forward weight alignment for SF, and path alignment for PFA/PFA-o (top). Backward-forward weight norm ratio for SF, and path norm ratio for PFA/PFA-o (bottom).

7 PFA for sparsely connected populations

In the biological brain, connections between neurons are sparse: in local cortical circuits, a large proportion of pairs of neurons are not connected [9], and the connection density decreases quickly over distance (between two neurons) [23]. The sparse connectivity is even more prominent in long-range projections. Therefore, biologically plausible rules should be able to overcome challenges caused by sparsity (or even take advantage of it). We thus examined the effects of different sparsity levels in the backward path on task performance, by fixing a proportion of feedback weights to zero during training (i.e., BB in FA, DFA, SF, PFA, and also RR in PFA). We found that the performance of FA, DFA, and SF gradually degrades with sparser connectivity (Fig. 6). In contrast, the performance of PFA degrades more slowly, suggesting a potential advantage of our PFA algorithm.

Refer to caption
Figure 6: Effects of sparse connections in the feedback pathway, showing task performance on MNIST. A proportion of the feedback weights in FA, DFA, SF, and PFA (but not in BP) are fixed to zero during training. Shaded regions show standard deviations across 5 seeds.

Discussion

In conclusion, our PFA algorithm can approximate BP by fostering implicit alignment between forward and backward pathways, rather than relying on explicit weight symmetry. This algorithm provides a potential explanation for why we do not observe symmetric weights in the brain even though it presumably implements credit assignment across multilayer network architectures, serving as a demonstration of feasibility. We showed that PFA achieves a performance comparable to BP in both fully-connected feedforward networks and deeper convolutional networks, even on rather challenging datasets/tasks. In addition, our PFA outperforms other biology-inspired algorithms in a sparse-connectivity scenario.

Our PFA algorithm utilizes an additional neuronal population to transmit errors 𝒆¯l\bm{\bar{e}}_{l} to the population with activation 𝒙l\bm{x}_{l} within the same layer. This is broadly consistent with the brain’s diverse neuronal population structure, with predominantly local connections. We have demonstrated that introducing this population allows us to fully remove all bidirectional connections between neurons, leading to a weight alignment angle of 9090^{\circ} (i.e., zero weight correlation). To accurately match the weight correlation patterns in the brain, the observed values of weight alignment/correlation (an alignment angle of 7878^{\circ} and a correlation of R0.2R\approx 0.2, estimated in Appendix B) can be directly achieved by combining different groups of synapses separately learned by PFA (no correlation), FA (partial correlation), and KP (full correlation). Further, additional populations like the one introduced here provide extra flexibility and offer the possibility to deal with other biological constraints (such as sparse connections), a direction yet to be fully explored.

In our simulations, we fixed the feedback weights BB, and since in PFA (but not PFA-o) each synaptic weight in BB was independently sampled, a fairly large expansion ratio was necessary to ensure a close approximation to BP. Exploring biologically plausible plasticity rules for BB that could reduce the expansion ratio requirement, potentially leading to a BTBB^{T}B closer to the identity matrix, is one of our future research directions. We only considered an expansion ratio larger than one (λ<1\lambda<1) in this study because an expansion ratio smaller than one will reduce the rank of the transmitted errors, similar to the scenario of low-rank gradient approximation [24]. Further investigation is required to understand the effects of low-rank weight updates.

The plasticity rules for updating the feedforward weights WW, which are based on presynaptic activations and postsynaptic errors, are only biologically plausible when the error signal elie_{l}^{i} is locally available at the corresponding neuron with activation xlix_{l}^{i}. Several proposals have been put forward regarding how a single neuron can represent and transmit both forward activations and backward errors without interference. The dendritic cortical microcircuit framework posits that the errors are represented in the apical dendrites while the forward activations are represented in the basal dendrites [3]. An additional self-prediction pathway is introduced to facilitate error transmission. Another study, based on burst-dependent plasticity, proposes that the forward activations are represented by the event firing rates, while errors are encoded by burst probabilities [4], enabling the multiplexing of signals within the same neuron. Our PFA algorithm can be directly integrated with these proposals, providing solutions to the explicit weight symmetry problem in these frameworks.

Our study is subject to several limitations. First, while we solved the weight symmetry problem in feedforward networks, the convolutional kernels re-introduce weight sharing across spatial locations, potentially conflicting with biological observations again. Second, introducing the additional neuronal population increases the training time and memory cost compared to BP. Third, we only studied the effects of sparse feedback connections in MNIST, while more tasks/datasets are required to systematically examine them. Addressing these limitations is one of our future directions.

Acknowledgments and Disclosure of Funding

M.K.B was supported by NIH R01NS125298. M.K.B and L.J.-A. were supported by the Kavli Institute for Brain and Mind.

References

  • [1] David E Rumelhart, Geoffrey E Hinton, and Ronald J Williams. Learning representations by back-propagating errors. nature, 323(6088):533–536, 1986.
  • [2] James CR Whittington and Rafal Bogacz. An approximation of the error backpropagation algorithm in a predictive coding network with local hebbian synaptic plasticity. Neural computation, 29(5):1229–1262, 2017.
  • [3] João Sacramento, Rui Ponte Costa, Yoshua Bengio, and Walter Senn. Dendritic cortical microcircuits approximate the backpropagation algorithm. Advances in neural information processing systems, 31, 2018.
  • [4] Alexandre Payeur, Jordan Guerguiev, Friedemann Zenke, Blake A Richards, and Richard Naud. Burst-dependent synaptic plasticity can coordinate learning in hierarchical circuits. Nature neuroscience, 24(7):1010–1019, 2021.
  • [5] Nicholas Alonso and Emre Neftci. Tightening the biological constraints on gradient-based predictive coding. In International Conference on Neuromorphic Systems 2021, pages 1–9, 2021.
  • [6] Stork. Is backpropagation biologically plausible? In International 1989 Joint Conference on Neural Networks, pages 241–246. IEEE, 1989.
  • [7] Francis Crick. The recent excitement about neural networks. Nature, 337(6203):129–132, 1989.
  • [8] Timothy P Lillicrap, Adam Santoro, Luke Marris, Colin J Akerman, and Geoffrey Hinton. Backpropagation and the brain. Nature Reviews Neuroscience, 21(6):335–346, 2020.
  • [9] Sen Song, Per Jesper Sjöström, Markus Reigl, Sacha Nelson, and Dmitri B Chklovskii. Highly nonrandom features of synaptic connectivity in local cortical circuits. PLoS biology, 3(3):e68, 2005.
  • [10] Timothy P Lillicrap, Daniel Cownden, Douglas B Tweed, and Colin J Akerman. Random synaptic feedback weights support error backpropagation for deep learning. Nature communications, 7(1):13276, 2016.
  • [11] Sergey Bartunov, Adam Santoro, Blake Richards, Luke Marris, Geoffrey E Hinton, and Timothy Lillicrap. Assessing the scalability of biologically-motivated deep learning algorithms and architectures. Advances in neural information processing systems, 31, 2018.
  • [12] Arild Nøkland. Direct feedback alignment provides learning in deep neural networks. Advances in neural information processing systems, 29, 2016.
  • [13] Navid Shervani-Tabar and Robert Rosenbaum. Meta-learning biologically plausible plasticity rules with random feedback pathways. Nature Communications, 14(1):1805, 2023.
  • [14] Qianli Liao, Joel Leibo, and Tomaso Poggio. How important is weight symmetry in backpropagation? In Proceedings of the AAAI Conference on Artificial Intelligence, volume 30, 2016.
  • [15] Will Xiao, Honglin Chen, Qianli Liao, and Tomaso Poggio. Biologically-plausible learning algorithms can scale to large datasets. In International Conference on Learning Representations, 2018.
  • [16] Theodore H Moskovitz, Ashok Litwin-Kumar, and LF Abbott. Feedback alignment in deep convolutional networks. arXiv preprint arXiv:1812.06488, 2018.
  • [17] Albert Jiménez Sanfiz and Mohamed Akrout. Benchmarking the accuracy and robustness of feedback alignment algorithms. arXiv preprint arXiv:2108.13446, 2021.
  • [18] Mohamed Akrout, Collin Wilson, Peter Humphreys, Timothy Lillicrap, and Douglas B Tweed. Deep learning without weight transport. Advances in neural information processing systems, 32, 2019.
  • [19] Kevin Max, Laura Kriener, Garibaldi Pineda García, Thomas Nowotny, Walter Senn, and Mihai A Petrovici. Learning efficient backprojections across cortical hierarchies in real time. In International Conference on Artificial Neural Networks, pages 556–559. Springer, 2023.
  • [20] Zhidong Bai and Jack W Silverstein. Spectral analysis of large dimensional random matrices, volume 20. Springer, 2010.
  • [21] Ian Goodfellow, Yoshua Bengio, and Aaron Courville. Deep learning. MIT press, 2016.
  • [22] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770–778, 2016.
  • [23] Bernhard Hellwig. A quantitative analysis of the local connectivity between pyramidal neurons in layers 2/3 of the rat visual cortex. Biological cybernetics, 82(2):111–121, 2000.
  • [24] Thijs Vogels, Sai Praneeth Karimireddy, and Martin Jaggi. Powersgd: Practical low-rank gradient compression for distributed optimization. Advances in Neural Information Processing Systems, 32, 2019.
  • [25] Xavier Glorot and Yoshua Bengio. Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the thirteenth international conference on artificial intelligence and statistics, pages 249–256. JMLR Workshop and Conference Proceedings, 2010.

Appendix A Training details

The expansion ratio (1/λ=N¯l/Nl+11/\lambda=\bar{N}_{l}/N_{l+1}) in PFA is set to 10. All weights in the networks are initialized with the Glorot initialization with unit gain factor [25], except for the BB weights in PFA, which are determined using the Glorot initialization with gain factor (1+λ)/2\sqrt{(1+\lambda)/2}, ensuring that BTBB^{T}B approximates an identity matrix (or BB is directly set equal to a semi-orthogonal matrix for the “PFA-o” variant). For optimization, we employed stochastic gradient descent with a momentum of 0.9.

A.1 MNIST

The learning rate is 10210^{-2} for BP, SF, PFA, FA and DFA, multiplied by a factor of 0.5 at the 20th and 30th epoch. The weight decay coefficient is 10410^{-4} for WW in BP, FA, DFA and SF. It is initially 0.030.03 for WW and RR in PFA, but reduced to 10410^{-4} after the first epoch. The networks are trained for 40 epochs with a batch size of 64 and 5 seeds. Training each network takes 10\sim 10 minutes on a GeForce RTX 3090.

A.2 CIFAR10

The learning rate is 0.1 for BP, FA, SF and PFA, and 10310^{-3} for DFA, multiplied by a factor of 0.1 at the 100th, 150th and 200th epochs. The weight decay is 10410^{-4} for WW in BP, FA, DFA and SF, and 0.0050.005 for WW and RR in PFA (reduced to 10410^{-4} after the first epoch). We trained the networks for 200 epochs with a batch size of 128 and 5 seeds. Training each network takes 2\sim 2 hours on a GeForce RTX 3090.

A.3 ImageNet

The learning rate is 0.1 for BP, SF and PFA, multiplied by a factor of 0.1 at the 20th, 40th and 60th epochs. The weight decay is 10410^{-4} for WW in BP and SF, and 0.00050.0005 for WW and RR in PFA (reduced to 10410^{-4} after the first epoch). We trained the networks for 75 epochs with a batch size of 256. Training each network takes 1\sim 1 week on a GeForce RTX 3090.

Appendix B Weight correlation/alignment in the brain

We ran simulations based on experimental observations [9]. We first sampled feedforward weights from a normal distribution. For 69% of feedforward weights (unidirectionally connected neurons), we set the corresponding feedback weights to 0. For the remaining 31% (bidirectionally connected neurons), we randomly sampled feedback weights that correlate with feedforward weights (correlation coefficient R=0.36R=0.36). The weight correlation for all pairs of connected neurons is then R0.2R\approx 0.2, corresponding to a backward-forward weight alignment angle of 7878^{\circ}. This result is an approximate estimate and might change under different assumptions.