This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Spatial-aware decision-making with ring attractors in Reinforcement Learning systems

Marcos Negre Saura & Richard Allmendinger
Alliance Manchester Business School
The University of Manchester
Manchester, United Kingdom
[email protected]
[email protected]
\AND
Theodore Papamarkou
College of Mathematical Medicine
Zhejiang Normal University
Jinhua, China
[email protected]
&Wei Pan
Department of Computer Science
The University of Manchester
Manchester, United Kingdom
[email protected]
Abstract

This paper explores the integration of ring attractors, a mathematical model inspired by neural circuit dynamics, into the reinforcement learning (RL) action selection process. Ring attractors, as specialized brain-inspired structures that encode spatial information and uncertainty, offer a biologically plausible mechanism to improve learning speed and predictive performance. They do so by explicitly encoding the action space, facilitating the organization of neural activity, and enabling the distribution of spatial representations across the neural network in the context of deep RL. The application of ring attractors in the RL action selection process involves mapping actions to specific locations on the ring and decoding the selected action based on neural activity. We investigate the application of ring attractors by both building them as exogenous models and integrating them as part of a Deep Learning policy algorithm. Our results show a significant improvement in state-of-the-art models for the Atari 100k benchmark. Notably, our integrated approach improves the performance of state-of-the-art models by half, representing a 53% increase over selected baselines.

1 Introduction

This paper addresses the challenge of efficient action selection in reinforcement learning (RL), particularly in environments with spatial structures. Our primary contribution is the novel integration of ring attractors (Kim et al., 2017), a neural circuit model from neuroscience, into the RL framework. This approach improves spatial awareness in action selection and provides a mechanism for uncertainty-aware decision making in RL, leading to more accurate and efficient learning in complex environments. Ring attractors offer a unique framework to continuously and stably represent spatial information (Sun et al., 2020). In a ring attractor network, neurons interconnect circularly, forming a loop with tuned connections (Blair et al., 2014). This configuration allows for robust and localized activation patterns, maintaining accurate spatial representations even with noise or perturbations. Applying ring attractors to the selection of RL actions involves mapping actions to specific ring locations and decoding the selected action based on neural activity. This spatial embedding proves advantageous for continuous action spaces, particularly in tasks such as robotic control and navigation (Rivero-Ortega et al., 2023). Ring attractors improve decision making by exploiting spatial relations between actions, contributing to informed transitions between actions in sequential decision-making tasks in RL.

In what follows, we summarize our contributions. Briefly, our contributions include a novel approach to RL policies based on ring attractors, the inclusion of uncertainty-aware capabilities in our RL systems, and the development of Deep Learning (DL) modules for RL with ring attractors.

Integration of ring attractors into RL policies and spatial encoding for action selection. We propose a novel approach for incorporating ring attractors, a neural structure use for motor control and cognition, into RL as a lightweight, efficient and robust decision-making structure. The circular structure of ring attractors allows the model to represent spatial information and relations between actions. This spatial awareness significantly speeds up the learning rate of the RL agent. The relevant methodology and experiments can be found in Sections 3.1.2 and 4.1, respectively.

Uncertainty-aware RL. Ring attractors can encode uncertainty estimation to drive the action selection process. This paper utilizes Bayesian uncertainty estimation to influence the policy. The relevant methodology and experiments can be found in Sections 3.1.3 and 4.1, respectively.

DL module for ring attractors. We develop a reusable DL module based on recurrent neural networks that integrates ring attractors into DL-based RL agents. Additionally, this enables the adoption of our ring attractor approach across different RL models and tasks in various domains of application. The relevant methodology and experiments can be found in Sections 3.2 and 4.2, respectively.

2 Related Work

The integration of ring attractors into RL systems brings together neuroscience-inspired models and advanced machine learning techniques. Here, we review the literature on the key areas that form the foundation of our RL research: spatial awareness in RL, biologically inspired reinforcement learning approaches, and uncertainty quantification methods.

2.1 Spatial Awareness in Reinforcement Learning

Incorporating spatial awareness into RL systems has improved performance on tasks with inherent spatial structure. Regarding relational RL, Zambaldi et al. (2019) introduced an approach using attention mechanisms to reason about spatial relations between entities in an environment. This method demonstrated improved sample efficiency and generalization in tasks that require spatial reasoning. On the topic of navigation, Mirowski et al. (2017) developed a deep RL agent capable of navigating complex city environments using street-level imagery. Their approach incorporated auxiliary tasks, such as depth prediction and loop closure detection. Concerning explicit spatial representations, Gupta et al. (2017) proposed a cognitive mapping and planning approach for visual navigation, combining spatial memory with a differentiable neural planner. Similarly, Bapst et al. (2019) introduced a relational deep RL framework using graph neural networks to capture spatial relations between objects. Although these approaches demonstrate the importance of spatial awareness in RL, they often lack the biological plausibility found in neural circuits.

2.2 Biologically Inspired Machine Intelligence

Biologically inspired approaches to RL seek to leverage insights from neuroscience to improve the efficiency, adaptability, and interpretability of RL algorithms. These methods often draw upon neural circuit dynamics and cognitive processes observed in biological systems. The work presented in (Banino et al., 2018) demonstrated that incorporating grid-like representations, inspired by mammalian grid cells, into RL agents improved performance on navigation tasks. Their work showed that these biologically inspired representations emerged naturally in agents trained on navigation tasks and transfer well to new environments. Similarly, Cueva & Wei (2018) showed that recurrent neural networks trained on navigation tasks naturally developed grid-like representations, suggesting a deep connection between biological and artificial navigation systems. Wang et al. (2018) proposed a biologically inspired meta-RL algorithm that mimics the function of the prefrontal cortex and dopamine-based neuromodulation. Their approach demonstrated rapid learning and adaptation to new tasks, similar to the flexibility observed in biological learning systems.

2.3 Uncertainty Quantification

Regarding exploration strategies, Osband et al. (2016) introduced bootstrapped deep Q-networks (DQNs), addressing exploration by leveraging uncertainty in Q-value estimates by training multiple DQNs with shared parameters. Building on this theme, Burda et al. (2018) proposed random network distillation (RND), measuring uncertainty by comparing predictions between a target network and a randomly initialized network. For efficient uncertainty quantification, Durasov et al. (2020) and Bykovets et al. (2022) presented a novel ‘masksemble’ approach, applying masks across the input batch during the forward pass to generate diverse predictions. Addressing risk assessment in non-stationary environments, Jain et al. (2021) described a method to analyze sources of lack of knowledge by adding a second Bayesian model to predict algorithmic action risks, particularly relevant for multi-agent RL (MARL) systems. In the context of individual treatment effects, Lee et al. (2020) performed uncertainty quantification (UQ) using an exogenously prescribed algorithm, making the method agnostic to the underlying recommender algorithm.

Azizzadenesheli et al. (2018) developed a Bayesian approaches for RL in episodic high-dimensional Markov decision processes (MDPs). They introduced two novel algorithms: LINUCB and LINPSRL. These algorithms achieve significant improvements in sample efficiency and performance by incorporating uncertainty estimation into the learning process. The extension to Deep RL, called Bayesian deep Q-networks, BDQNs (Azizzadenesheli et al., 2018) , incorporates efficient Thompson sampling and Bayesian linear regression at the output layer to factor uncertainty estimation in the action-value estimates. On a similar line, Foerster et al. (2019) proposed a Bayesian action decoder. It is a learning algorithm based on approximate Bayesian updates to obtain a public belief that conditions the actions taken by other agents in the environment. This creates uncertainty-aware agents that are not biased by training data. It also generates a factorised, approximate belief state that provides the agents with efficient learning through informed actions.

In summary, the literature reveals a growing interest in incorporating spatial awareness, biological inspiration, and UQ into RL systems. However, there remains a gap in integrating these elements into a cohesive framework. Our work on ring attractors aims to bridge this gap by providing a biologically plausible model that inherently captures spatial relations and can be extended to handle uncertainty, potentially leading to more robust and efficient RL agents.

3 Methodology

In this section, we describe two main methods: an exogenous ring attractor model using spiking neural networks (SNNs) and a DL-based ring attractor integrated into the RL agent. Both leverage the ring attractors’ spatial encoding capabilities to enhance action selection and performance. We detail the ring attractor architecture, dynamics, and implementation, including uncertainty injection in the SNN model for robust decision making. Spiking neural networks (SNNs) are employed for their biological plausibility and efficient temporal information processing, which closely mimic natural neuronal behavior. The integrated approach offers end-to-end training for efficiency and scalability. Ring attractors in RL maintain stable spatial information representations, preserving action relations lost in traditional flattened action spaces. This circular spatial exploitation potentially yields smoother policy gradients and more efficient learning in spatial tasks, attributed to the ring attractors’ ability to maintain a stable representation of spatial information.

3.1 Exogenous Ring Attractor Model: Spiking Neural Network

During the first stage of the research, the focus is on developing a self-contained ring attractor as an SNN. This will be integrated into the output of the value-based policy model to perform action selection.

3.1.1 Ring Attractor Architecture

Ring attractors commonly consist of a configuration of excitatory and inhibitory neurons arranged in a circular pattern. We can model the dynamics of the ring using the Touretzky ring attractor network (Touretzky, 2005). In this model, each excitatory neuron establishes connections with all other excitatory neurons, and an inhibitory neuron is placed in the middle of the ring with equal weighted connections to all excitatory neurons. This creates a network that facilitates complex information processing.

Excitatory neurons’ input signal. Let xnisx^{i}_{n}\in\mathbb{R}^{s} denote the input signal from source ii to the excitatory neuron n=1,,Nn=1,\ldots,N. The total input to neuron nn is defined as the sum of all input signals II for that particular neuron: xn=i=1Ixnix_{n}=\sum_{i=1}^{I}x^{i}_{n}, where xnx_{n}\in\mathbb{R}. To model input signals xnix^{i}_{n} of varying strengths, these signals are commonly viewed as Gaussian functions xni:ssx^{i}_{n}:\mathbb{R}^{s}\rightarrow\mathbb{R}^{s}. These functions allow us to represent the input to each neuron as a sum of weighted Gaussian distributions. The key parameters of these Gaussian functions are: KiK_{i}, the magnitude variable for the input signal ii, which determines the overall strength of the signal; μi\mu_{i}, which defines the mean position of the the Gaussian curve in the ring for the input signal ii, representing the central focus of the signal; σi\sigma_{i}, the standard deviation of the Gaussian function, which determines the spread or reliability of the signal; and αn\alpha_{n}, which represents the preference for the orientation of the neuron nn in space. These parameters combine to the following:

xn(Ki)=i=1Ixni(αn)=i=1IKi2πσiexp(12(αnμi)2σi2),n=1,2,,Nx_{n}(K_{i})=\sum_{i=1}^{I}x^{i}_{n}(\alpha_{n})=\sum_{i=1}^{I}{\frac{K_{i}}{\sqrt{2\pi\sigma_{i}}}\exp\left(-\frac{1}{2}\frac{(\alpha_{n}-\mu_{i})^{2}}{\sigma_{i}^{2}}\right)},\quad n=1,2,\ldots,N (1)

Neuron activation function. We employ the rectified linear unit (ReLU) function f(x)=max(0,x+h)f(x)=\max(0,x+h), where h+h\in\mathbb{R}^{+} as the activation function for each neuron, where hh is a threshold that introduces the non-linear behaviour in the ring.

Excitatory neuron dynamics. The dynamics of excitatory neurons in the ring is described as:

dvndtΔvnΔt=vn+ΔtvnΔt=f(xn+en+in)vnτ.\frac{dv_{n}}{dt}\approx\frac{\Delta v_{n}}{\Delta t}=\frac{v_{n+\Delta t}-v_{n}}{\Delta t}=\frac{f(x_{n}+e_{n}+i_{n})-v_{n}}{\tau}. (2)

In Eq. 2 vnv_{n}\in\mathbb{R} represents the activation of the excitatory neuron nn, xnx_{n} is previously defined in Eq. 1 is the external input to the neuron nn of Eq. 1, ene_{n}\in\mathbb{R} represents the weighted influence of the other excitatory neurons activation, which is defined mathematically in Eq. 4. ini_{n}\in\mathbb{R} is the influence from the weighted inhibitory neuron activation to the target excitatory neuron nn, and τ=Δt\tau=\Delta t is the time integration constant. This equation captures the evolution of neuronal activation over time, considering both excitatory and inhibitory activations.

Inhibitory neuron dynamics. The activation of the inhibitory neuron, which regulates network dynamics, is described by:

dudtΔuΔt=u+ΔtuΔt=f(en+in)uτ\frac{du}{dt}\approx\frac{\Delta u}{\Delta t}=\frac{u_{+\Delta t}-u}{\Delta t}=\frac{f(e_{n}+i_{n})-u}{\tau} (3)

Here, uu\in\mathbb{R} represents the inhibitory neuron’s activation output, ene_{n} is the weighted sum of excitatory activations where in this case nn is the inhibitory neuron, and ini_{n}\in\mathbb{R} is the weighted self-inhibition activation term. This equation models how the inhibitory neuron integrates inputs from the excitatory population and its own state.

Synaptic weighted connections: The influence between neurons decreases with distance, as modeled by the weighted connections. This weighted connection applies to both excitatory and inhibitory neurons. For excitatory neurons: w(EmEn)=ed(m,n)2w^{(E_{m}\rightarrow E_{n})}=e^{-d^{2}_{(m,n)}}, where d(m,n)=|mn|d_{(m,n)}=|m-n| is the distance between neurons mm and nn. For the inhibitory neuron: w(IEn)=ed(m,n)2w^{(I\rightarrow E_{n})}=e^{-d^{2}_{(m,n)}}. Note that our model contains a single inhibitory neuron placed in the middle of the ring, with a distance of 1 unit to all excitatory neurons.The excitatory (ene_{n}) and inhibitory (ini_{n}) weighted connections are also known in the literature as neuron-proximal excitatory and inhibitory voltage or potential.These are then defined as follows:

en=m=1Nw(EmEn)vmin=w(ImEn)u\displaystyle e_{n}=\sum_{m=1}^{N}w^{(E_{m}\rightarrow E_{n})}v_{m}\ \ \ \ \ i_{n}=w^{(I_{m}\rightarrow E_{n})}u (4)

Full excitatory neuron dynamics. Combining all influences, the complete dynamics of excitatory neurons are described by:

dvndt=1τ(f(m=1Nwm,n(EmEn)vm+xn+w(IEn)u)vn)\frac{dv_{n}}{dt}=\frac{1}{\tau}\left(f\left(\sum_{m=1}^{N}w^{(E_{m}\rightarrow E_{n})}_{m,n}v_{m}+x_{n}+w^{(I\rightarrow E_{n})}u\right)-v_{n}\right) (5)

This equation updates the activation for all neurons based on recurrent excitation, external input from the summation of the input signals for neuron nn, xnx_{n}; and inhibitory influence.

Full inhibitory neuron dynamics. The complete dynamics of the inhibitory neuron are given by:

dudt=1τ(f(w(II)u+m=1Nwm(EmI)vm)u)\frac{du}{dt}=\frac{1}{\tau}\left(f\left(w^{(I\rightarrow I)}u+\sum_{m=1}^{N}w^{(E_{m}\rightarrow I)}_{m}v_{m}\right)-u\right) (6)

This equation models how the inhibitory neuron integrates self-inhibition and excitatory inputs from the entire network. These equations collectively describe the complex dynamics of the ring attractor network, capturing the interplay between excitatory and inhibitory neurons, external inputs, and synaptic connections.

3.1.2 Ring Attractor as Behavior Policy in Reinforcement Learning

To integrate the ring attractor model with RL, we need to establish a connection between the estimated value of state-action pairs and the input to the ring attractor network. This integration allows the ring attractor to serve as a behavior policy, guiding action selection based on the values learned. We begin by reformulating the input function for a target excitatory neuron nn. The key modification is setting the scale factor KiK_{i} to the Q-value Q(s,a)Q(s,a) of the state-action pair (s,a)(s,a), that is Ki=Q(s,a)K_{i}=Q(s,a).

This formulation ensures that actions with higher estimated values are given more weight in the ring attractor dynamics, naturally biasing the network towards more valuable actions. The orientation of the signal within the ring attractor is determined by the direction of movement in the action space. We represent this as μi=α(a)\mu_{i}=\alpha(a), where α(a)\alpha(a) is the angle corresponding to the action aa in the circular action space. We define our circular action space 𝒜\mathcal{A} as a subset of 2\mathbb{R}^{2}, where each action a𝒜a\in\mathcal{A} is represented by a point on the unit circle. The function α:𝒜[0,2π)\alpha:\mathcal{A}\rightarrow[0,2\pi) maps each action to its corresponding angle on this circle. To account for uncertainty in our value estimates, we incorporate the variance of the estimated value for each action into our model: σi=σa\sigma_{i}=\sigma_{a}.

This allows the network to represent not just the expected value of actions, but also our confidence in those estimates. Combining these elements, we arrive at the following equation for the action signal x(a)x(a):

xn(Q)=a=1AQ(s,a)2πσaexp(12(αnαa)2σa2)x_{n}(Q)=\sum_{a=1}^{A}{\frac{Q(s,a)}{\sqrt{2\pi\sigma_{a}}}\exp\left(-\frac{1}{2}\frac{(\alpha_{n}-\alpha_{a})^{2}}{\sigma_{a}^{2}}\right)} (7)

This equation represents the input to each neuron as a sum of Gaussian functions, where each function is centered on an action’s direction and scaled by its Q-value. The dynamics of the excitatory neurons in the ring attractor, now incorporating the Q-value inputs, are described by:

dvndt=1τ(max(0,(m=1m=Nwm,n(EE)vm+xn(Q)+wIEuu)))vn\begin{gathered}\frac{dv_{n}}{dt}=\frac{1}{\tau}\left(max\left(0,\left(\sum_{m=1}^{m=N}w^{(E\xrightarrow{}E)}_{m,n}v_{m}+x_{n}(Q)+w^{I\xrightarrow{}E_{u}}u\right)\right)\right)-v_{n}\end{gathered} (8)

This equation captures how the activation of each neuron evolves over time, influenced by the action-value functions xn(Q)x_{n}(Q), and both excitatory and inhibitory feedback. This equation captures how the activation of each neuron evolves over time, influenced by the action-value input Gaussian functions xn(Q)x_{n}(Q), the excitatory feedback en=m=1Nw(EmEn)vme_{n}=\sum_{m=1}^{N}w^{(E_{m}\rightarrow E_{n})}v_{m}, and the inhibitory feedback in(=w(IEn)ui_{n}(=w^{(I\rightarrow E_{n})}u.

To translate the ring attractor’s output into an action in the 2D space, we use the following equation:

action=argmaxn{𝐕}N(A)N(E)\text{action}=\operatorname*{arg\,max}_{n}\{\mathbf{V}\}\cdot\frac{N^{(A)}}{N^{(E)}} (9)

where n{1,,N(E)}n\in\{1,...,N^{(E)}\}, N(E)N^{(E)} is the number of excitatory neurons in the ring attractor, N(A)N^{(A)} is the number of discrete actions in the action space 𝒜\mathcal{A}, 𝐕=[v1,v2,,vN(E)]\mathbf{V}=[v_{1},v_{2},...,v_{N^{(E)}}].

This equation assumes that both the neurons in the ring attractor and the actions in the action space are uniformly distributed. This approach allows for nuanced action selection that takes into account both the spatial relations between actions and their estimated values. A visualization of the ring is presented in Fig. 1.

3.1.3 Uncertainty Quantification Model

In the field of DRL, for any state-action pair (s,a)(s,a), the Q-value Q(s,a)Q(s,a) can be expressed as a function approximation algorithm Φθ(s)\Phi_{\theta}(s) taking as input the current state ss. This function approximation algorithm (Φθ(s)\Phi_{\theta}(s)) can be expressed as the weight matrix of our function approximation algorithm transposed θT\theta^{T} times the feature vector extracted from the input state x(s)x(s): Q(s,a)=Φθ(s)=θTx(s)Q(s,a)=\Phi_{\theta}(s)=\theta^{T}x(s). As stated in Section 3.1.2, the variance of the Gaussian functions, input to the ring attractor, will be given by the variance of the estimate value for that particular action σi=σa\sigma_{i}=\sigma_{a}.

Refer to caption
Figure 1: Ring attractor Touretzky representation: Circular arrangement of excitatory neurons (N0-N7) with excitatory connections and central inhibitory neuron. Four input signals shown as colored gradients. Overall activation depicted by red outline. Includes connection weights and input signal parameters, illustrating ring attractor dynamics.

Among the diverse methods to compute the uncertainty of the action (σa\sigma_{a}) we have chosen to compute a posterior distribution with Bayesian linear regression (BLR). BLR acts as output layer for our neural network (NN) of choice. We choose a linear regression model because it does not compromise the efficiency of the NN, while at the same time it provides a distribution to compute the variance for the state-action pairs. The implementation is based on Azizzadenesheli et al. (2018), where a Bayesian value-based DQN model was instantiated to output an uncertainty-aware prediction for the state-action pairs. With this approach, the new QQ function is defined as:

Q(s,a)=Φθ(s)Twa,Q(s,a)=\Phi_{\theta}(s)^{T}w_{a}, (10)

where waw_{a} are the weights from the posterior distribution of the BLR model. Eq. 10 represents the parameters of the final Bayesian linear layer.

When provided with a state transition tuple (s,a,r,s)(s,a,r,s^{\prime}), where ss is the current state, aa is the action taken, rr is the reward received, and ss^{\prime} is the next state. This tuple represents a single step of interaction between the agent and the environment in the RL framework. The model learns to adjust the weights waw_{a} of the BLR and the function approximation algorithm, i.e. neural networks (NNs), (Φθ\Phi_{\theta}) to align the Q values with the optimal action a=argmax(γΦθ(s)Twa)a=argmax(\gamma\Phi_{\theta}(s)^{T}w_{a}), equation 11.

Q(s,a)=Φθ(s)Tway:=r+γΦθtarget(x)Twtargeta^Q(s,a)=\Phi_{\theta}(s)^{T}w_{a}\xrightarrow{}y:=r+\gamma\Phi_{\theta_{\text{target}}}(x^{\prime})^{T}w_{\text{target}}\hat{a} (11)

where γ\gamma is the discount factor, θtarget\theta_{target} refers to the parameters of the target function approximation algorithm used for learning, and a^\hat{a} is the predicted optimal action in the next state ss^{\prime}. The construction of the Gaussian BLR prior distribution and the weights ii sample wa,iw_{a,i} collected from the posterior distribution are performed through Thompson Sampling. This process allows us to incorporate uncertainty into our action-value estimates. For details on the construction of the Gaussian prior distribution and the specifics of the sampling process, we refer readers to Azizzadenesheli et al. (2018). Both the mean μa\mu_{a} and the variance σa\sigma_{a} from Eq. 12 are calculated from a finite number of samples II.

μa\displaystyle\mu_{a} =i=0i=Iwa,iTΦθ(st)I=i=0i=IQ(s,a)iI\displaystyle=\frac{\sum^{i=I}_{i=0}w_{a,i}^{T}\Phi_{\theta}(s_{t})}{I}=\frac{\sum^{i=I}_{i=0}{Q(s,a)_{i}}}{I} (12)
σa2\displaystyle\sigma^{2}_{a} =i=0i=I(wa,iTΦθ(st)μa)2I1\displaystyle=\frac{\sum^{i=I}_{i=0}{\left(w_{a,i}^{T}\Phi_{\theta}(s_{t})-\mu_{a}\right)}^{2}}{I-1}

3.2 Deep Learning Ring Attractor Model

To further enhance the ring attractor’s integration into RL frameworks and agents, we provide a Deep Learning (DL) implementation.This approach improves model learning and integration with DRL agents. Our implementation offers both algorithmic improvements, by benefiting from DL training process, and software integration improvements, easing the deployment processes. Recurrent Neural Networks (RNNs) offer a practical approach for integrating ring attractors within DRL agents. Recent studies by  Li et al. (2015) show that RNNs perform well in modeling sequential data and in capturing temporal dependencies for decision making. RNNs mirror ring attractors’ temporal dynamics, with their recurrent connections and flexible architecture emulating the interconnected nature of ring attractor neurons. This allows modeling of weighted connections for both forward and recurrent hidden states, as shown in the Appendix A.1.3. The premises for modeling the structure of the RNN are as follows.

Attractor state as recurrent connections. RNN recurrent connections model the attractor state, integrating information from previous time steps into the current network state, allowing retention of information over time.

Signal input as a forward pass. Forward connections from previous layers are arranged circularly, mimicking the ring’s spatial distribution. The attractor state encodes task context, influenced by current input and hidden state. A learnable time constant τ\tau, inherited from Eq. 2, controls input contributions and temporal evolution, enabling adaptive behavior and adaptable input contribution to the attractor state.

3.2.1 Deep Reinforcement Learning Agent Integration

To shape circular connectivity within a RNN, the weighted connections in the input signal V(s)V(s) and the hidden state or attractor state U(v)U(v) are computed as follows:

V(s)m,n\displaystyle V(s)_{m,n} =1τΦθ(s)Twm,nIH=1τΦθ(s)Ted(m,n)λ\displaystyle=\frac{1}{\tau}\Phi_{\theta}(s)^{T}w^{I\xrightarrow{}H}_{m,n}=\frac{1}{\tau}\Phi_{\theta}(s)^{T}e^{\frac{d(m,n)}{\lambda}} (13)
d(m,n)\displaystyle d(m,n) =min(abs(mnMN),Nabs(mnMN))\displaystyle=\min{(abs(m-n\frac{M}{N}),N-abs(m-n\frac{M}{N}))}
U(v)m,n\displaystyle U(v)_{m,n} =h(v)Twm,nHH=h(Φθ)Ted(m,n)λ\displaystyle=h(v)^{T}w^{H\xrightarrow{}H}_{m,n}=h(\Phi_{\theta})^{T}e^{\frac{d(m,n)}{\lambda}}
d(m,n)\displaystyle d(m,n) =min(abs(mn),Nabs(mm))\displaystyle=\min{(abs(m-n),N-abs(m-m))}

This circular structure mimics the arrangement of excitatory neurons in the ring attractor. Eq. 13 shows the input signal to the recurrent layer Vm,nV_{m,n} from neuron mm from the previous layer in the DL agent to neuron nn in the RNN. The hidden state, Um,nU_{m,n} mimics an attractor state, representing the recurrent connections in the RNN.. The weighted RNN connections input-to-hidden and hidden-to-hidden wm,nIHw^{I\xrightarrow{}H}_{m,n}, wm,nHHw^{H\xrightarrow{}H}_{m,n} depend on a parameter λ\lambda that drives the decay of the potential over distance and distance between neurons d(m,n)d(m,n) where NN is the total number of neurons for the RNN and MM is the count of neurons in the previous layer of the NN architecture. The function ϕθ(s):SM\phi_{\theta}(s):\mathbb{R}^{S}\rightarrow\mathbb{R}^{M} maps the input state ss of the DL agent to a representation of characteristics that will be the input of the recurrent layer. Likewise, θ\theta represents the parameters of this function (i.e., the weights and biases of the NN layers preceding the RNN layer, which extract relevant features from the input). The function h(v):NNh(v):\mathbb{R}^{N}\rightarrow\mathbb{R}^{N} is a parameterized by learnable weights transformation that maps the information from previous forward passes into the current hidden state. The learnable parameter τ\tau is the positive time constant responsible for the integration of signals in the ring. It defines the contribution of input states ϕθ(s)\phi_{\theta}(s) to the current hidden state, imitating the attractor state, applied to neural networks.

Finally, the action-value function Q(s,a)Q(s,a) is derived from the RNN layer’s output by applying the neurons activation function to the combined input V(s)V(s) and hidden state information U(s)U(s). The activation function of choice is a hyperbolic tangent tanh\tanh, this function is symmetric around zero, leading to faster convergence and stability. However, the output range of tanh\tanh (-1 to 1) is not fully compatible with value-based methods, where the DL agents needs to output action-value pairs in the range of the environment’s reward function. To address this issue and prevent saturation of the tanh\tanh activation function, we scale the action-value pairs by multiplying them with a learnable scalar β\beta, as Q(s,a)=βht=βtanh((V(s)+U(v))=βtanh((1τΦθ(st)TwIH+ht1(v)TwHH))Q(s,a)=\beta h_{t}=\beta\tanh((V(s)+U(v))=\beta\tanh((\frac{1}{\tau}\Phi_{\theta}(s_{t})^{T}w^{I\rightarrow H}+h_{t-1}(v)^{T}w^{H\rightarrow H})).

4 Experiments

This section presents the findings from our experiments that validate our proposed approach of integrating ring attractors into RL algorithms. To assess the effectiveness of our method, we conduct comparisons across multiple baseline models and action spaces. The evaluation encompasses two implementations: a traditional exogenous ring attractor and an innovative approach where the ring attractor is modeled directly into a DRL agents. In both implementations, action-value pairs Q(s,a)Q(s,a) are evenly distributed across the ring circumference. For the exogenous model, each action is associated with a specific angle on the ring, Section 3.1.2. In the DL implementation, each neuron in the RNN corresponds to one action-value, with the weights network modeling the circular topology of the action space Section 3.2.1.

Our results demonstrate the effectiveness of ring attractors in enhancing action selection and significantly speeding up the learning process of the Reinforcement Learning agent overall.

4.1 Exogenous Ring Attractor Model Performance Analysis

To evaluate our exogenous ring attractor model integrated with BDQN (Azizzadenesheli et al., 2018) we performed experiments in the OpenAI Super Mario Bros environment  (Kauten, 2018). This benchmark exhibits an spatially distributed complex decision-making scenario.

Refer to caption
Figure 2: Learning speed comparison in the OpenAI Gym Super Mario Bros environment (Kauten, 2018). The plot shows cummulative reward over 1 million frames for three models: Standard BDQN; BDQNRA with ring attractor behavior policy from Section 3.1.2, setting action-value pair variance constant to σa=π6\sigma_{a}=\frac{\pi}{6}; and BDQNRA-UA with RA and Uncertainty Awareness (UA) implementing the uncertainty quantification model from  3.1.3 to feed into the variance of the action-value pairs. Displaying mean episodic returns over 10 averaged seeds.

Fig. 2 shows that both ring attractor models (BDQNRA and BDQNRA-UA) consistently outperform standard BDQN. The uncertainty-aware version (BDQNRA-UA) shows the best overall performance, highlighting the benefits of combining ring attractors spatial distribution of the action space with uncertainty-aware action selection. Empirical evaluations revealed that the Spiking Neural Network-based ring attractor models exhibited a mean computational overhead of 297.3% (SD = 14.2%) compared to the baseline, significantly impacting runtime efficiency. To address this performance bottleneck and integrating the ring attractor spatial understanding into the DRL, we developed a DL implementation of the ring attractor. This DL implementation is evaluated in the subsections below.

4.2 Deep Learning Ring Attractor Model Performance Analysis

This subsection details the effects of incorporating uncertainty quantification through Bayesian Linear Regression. We evaluate the quality of uncertainty estimates and their impact on exploration strategies and overall agent performance. Fig. 3, shows that the DDQNRA model consistently outperforms the standard DDQN across tasks for both navigation and game like decision-making scenarios. This suggests that the ring attractor’s ability to encode spatial relationships in different actions spaces contributes significantly to the agent’s learning efficiency. These results indicate that the integration of ring attractors into DRL architectures can lead to significant improvements in both learning speed and overall performance, especially in environments with strong spatial components.

Refer to caption
Refer to caption
Figure 3: Performance comparison: DDQNRA vs standard DDQN (van Hasselt et al., 2015) in two environments. Right: OpenAI highway (Leurent, 2018), showing learning speed in spatial navigation tasks. Left: OpenAI Super Mario Bros (Kauten, 2018), demonstrating adaptability to complex, game-like scenarios. Displaying mean episodic returns over 10 averaged seeds.

4.3 Performance on Atari 100k Benchmark

In this results section, we provide a comprehensive analysis of our model’s performance on the Atari 100k benchmark (Bellemare et al., 2012). We present detailed comparisons with state-of-the-art models, highlighting the improvements achieved by our approach. We analyze the factors contributing to the significant performance increase mentioned in the introduction, breaking down results by game, and discussing notable trends or patterns observed across different types of tasks.

Table 1: Performance comparison on Atari 100k Benchmark (Bellemare et al., 2012). Benchmark performed across all environments where actions can be layout in one or more 2D action space planes (ring attractors). This is represented by the ring configuration column. The results are recorded at the end of training and averaged over 10 random seeds, 3 samples per seed. We display game score and overall mean and median human-normalized scores for each algorithm.
Game Agent: Reported Implemented
Environment Ring Human CURL SPR EffZero EffZero EffZeroRA
Alien Double 7127.7 558.2 801.5 808.5 738.1 1098.8
Asterix Single 8503.3 734.5 977.8 25557.8 14839.3 31037.3
Bank Heist Double 753.1 131.6 380.9 351.0 362.8 460.5
BattleZone Double 37187.5 14870.0 16651.0 13871.2 11908.7 15672.0
Boxing Double 12.1 1.2 35.8 52.7 30.5 62.4
Chopper C. Double 7387.8 1058.5 974.8 1117.3 1162.4 1963.0
Crazy Climber Single 35829.4 12146.5 42923.6 83940.2 83883.0 100649.7
Freeway Double 29.6 26.7 24.4 21.8 22.7 31.3
Frostbite Double 4334.7 1181.3 1821.5 296.3 287.5 354.8
Gopher Double 2412.5 669.3 715.2 3260.3 2975.3 3804.0
Hero Double 30826.4 6279.3 7019.2 9315.9 9966.4 11976.1
Jamesbond Double 302.8 471.0 365.4 517.0 350.1 416.4
Kangaroo Double 3035.0 872.5 3276.4 724.1 689.2 1368.8
Krull Double 2665.5 4229.6 3688.9 5663.3 6128.3 9282.1
Kung Fu M. Double 22736.3 14307.8 13192.7 30944.8 27445.6 49697.7
Ms Pacman Single 6951.6 1465.5 1313.2 1281.2 1166.2 2028.0
Private Eye Double 69571.3 218.4 124.0 96.7 94.3 155.8
Road Runner Double 7845.0 5661.0 669.1 17751.3 19203.1 29389.3
Seaquest Double 42054.7 384.5 583.1 1100.2 1154.7 1532.8
Human–normalised Score
Mean 1.000 0.428 0.638 1.101 0.959 1.454
Median 1.000 0.242 0.434 0.420 0.403 0.531

Table 1 presents a comprehensive comparison of our ring attractor-based RL model integrated with Efficient Zero (Ye et al., 2021), evaluating performance across multiple Atari games with a limited training budget of 100,000 environment steps. The table includes results from baseline methods and recent top-performing algorithms SPR (Schwarzer et al., 2020) and CURL (Srinivas et al., 2020) for context. Our model demonstrates significant improvements over baseline methods, particularly in games with inherent spatial components, such as Asterix and Boxing, showing 110% and 105% improvement respectively over the previous state-of-the-art.

These results reaffirm that the spatial encoding provided by ring attractors is especially beneficial in environments where spatial relationships between actions are key. The consistent performance improvement different games indicates that our approach provides a general enhancement that benefits a wide range of RL tasks. Even in games where the improvement is less dramatic, we still see substantial increases in performance, suggesting that the benefits of the ring attractor extend beyond just spatially-oriented games.

To ensure a fair comparison under identical experimental conditions, we re-implemented and evaluated both the baseline EffZero and our proposed EffZeroRA model using the same computational resources and experimental setup as employed throughout this study.

Ablation studies were conducted to isolate the impact of key components in our ring attractor models, detailed in Appendix A.1.1 and A.1.2. For the exogenous model, we compared performance with correct and randomized action distributions in the ring. In the DL implementation, we removed the circular weight distribution to assess its importance.

5 Conclusion

This paper presents a novel approach to RL, integrating ring attractors into action selection. Our work demonstrates that these neuroscience-inspired ring attractors significantly enhance learning capabilities for value-based RL agents, leading to more stable and efficient action selection, particularly in spatially structured tasks.

5.1 Key Findings and Implications

The integration of ring attractors as a DL module proves particularly effective, allowing for end-to-end training and easy incorporation into existing RL architectures. This approach improves performance and offers potential insight into explicit spatial encoding of actions.

Our results demonstrate significant improvements in action selection and learning speed. We achieve state-of-the-art performance on the Atari 100k benchmark, with an average 53% performance increase across all games tested compared to the baseline and previous state-of-the-art models. Notable improvements were observed in games with strong spatial components, such as Asterix (110 % improvement) and Boxing (105% improvement). Additionally, we observed improvements in other environments tested outside the Atari benchmark, further supporting the effectiveness of our approach across various RL tasks and agents.

5.2 Future Work

We acknowledge that our approach, while promising, has limitations and areas for potential improvement. Future research should investigate the scalability of this method in high-dimensional action spaces and explore its efficacy in domains where spatial relationships are less straightforward.

We believe that the success of this approach opens up several future research paths. The current work can be extended to multi-agent scenarios and policy-based RL agents. In the field of uncertainty-aware decision making, leveraging the spatial structure provided by attractor networks presents a promising avenue to map uncertainty explicitly to the action space. Deploying the techniques presented here into specific domains could yield performance boosts, especially in safe RL, leveraging their stability properties to enforce constraints and ensure predictable behavior.

This approach not only improves performance but also offers potential insight into spatial encoding of actions and decision-making processes, bridging the gap between neuroscience-inspired models and practical RL agents.

References

  • Azizzadenesheli et al. (2018) Kamyar Azizzadenesheli, Emma Brunskill, and Animashree Anandkumar. Efficient exploration through bayesian deep q-networks. CoRR, abs/1802.04412, 2018.
  • Banino et al. (2018) Andrea Banino, Caswell Barry, Benigno Uria, Charles Blundell, Timothy P. Lillicrap, Piotr Mirowski, Alexander Pritzel, Martin J. Chadwick, Thomas Degris, Joseph Modayil, Greg Wayne, Hubert Soyer, Fabio Viola, Brian Zhang, Ross Goroshin, Neil C. Rabinowitz, Razvan Pascanu, Charlie Beattie, Stig Petersen, Amir Sadik, Stephen Gaffney, Helen King, Koray Kavukcuoglu, Demis Hassabis, Raia Hadsell, and Dharshan Kumaran. Vector-based navigation using grid-like representations in artificial agents. Nat., 557(7705):429–433, 2018.
  • Bapst et al. (2019) Victor Bapst, Alvaro Sanchez-Gonzalez, Carl Doersch, Kimberly L. Stachenfeld, Pushmeet Kohli, Peter W. Battaglia, and Jessica B. Hamrick. Structured agents for physical construction. CoRR, abs/1904.03177, 2019.
  • Bellemare et al. (2012) Marc G. Bellemare, Yavar Naddaf, Joel Veness, and Michael Bowling. The arcade learning environment: An evaluation platform for general agents. CoRR, abs/1207.4708, 2012.
  • Blair et al. (2014) Hugh Blair, Di Wu, and Jason Cong. Oscillatory neurocomputing with ring attractors: A network architecture for mapping locations in space onto patterns of neural synchrony. Philosophical transactions of the Royal Society of London. Series B, Biological sciences, 369:20120526, 02 2014.
  • Burda et al. (2018) Yuri Burda, Harrison Edwards, Amos J. Storkey, and Oleg Klimov. Exploration by random network distillation. CoRR, abs/1810.12894, 2018.
  • Bykovets et al. (2022) Eugene Bykovets, Yannick Metz, Mennatallah El-Assady, Daniel A. Keim, and Joachim M. Buhmann. How to enable uncertainty estimation in proximal policy optimization, 2022.
  • Cueva & Wei (2018) Christopher J. Cueva and Xue-Xin Wei. Emergence of grid-like representations by training recurrent neural networks to perform spatial localization, 2018.
  • Durasov et al. (2020) Nikita Durasov, Timur M. Bagautdinov, Pierre Baqué, and Pascal Fua. Masksembles for uncertainty estimation. CoRR, abs/2012.08334, 2020.
  • Foerster et al. (2019) Jakob N. Foerster, H. Francis Song, Edward Hughes, Neil Burch, Iain Dunning, Shimon Whiteson, Matthew M. Botvinick, and Michael Bowling. Bayesian action decoder for deep multi-agent reinforcement learning. In ICML 2019: Proceedings of the Thirty-Sixth International Conference on Machine Learning, June 2019.
  • Goodridge et al. (1998) JP Goodridge, PA Dudchenko, KA Worboys, EJ Golob, and Taube JS. Cue control and head direction cells. The Journal of neuroscience : the official journal of the Society for Neuroscience, 112:RC154, 08 1998.
  • Gupta et al. (2017) Saurabh Gupta, James Davidson, Sergey Levine, Rahul Sukthankar, and Jitendra Malik. Cognitive mapping and planning for visual navigation. CoRR, abs/1702.03920, 2017.
  • Jain et al. (2021) Moksh Jain, Salem Lahlou, Hadi Nekoei, Victor Butoi, Paul Bertin, Jarrid Rector-Brooks, Maksym Korablyov, and Yoshua Bengio. DEUP: direct epistemic uncertainty prediction. CoRR, abs/2102.08501, 2021.
  • Kauten (2018) Christian Kauten. Super Mario Bros for OpenAI Gym. GitHub, 2018.
  • Kim et al. (2017) Sung Soo Kim, Hervé Rouault, Shaul Druckmann, and Vivek Jayaraman. Ring attractor dynamics in the ¡i¿drosophila¡/i¿ central brain. Science, 356(6340):849–853, 2017. doi: 10.1126/science.aal4835.
  • Lee et al. (2020) Hyun-Suk Lee, Yao Zhang, William Zame, Cong Shen, Jang-Won Lee, and Mihaela van der Schaar. Robust recursive partitioning for heterogeneous treatment effects with uncertainty quantification, 2020.
  • Leurent (2018) Edouard Leurent. An environment for autonomous driving decision-making, 2018.
  • Li et al. (2015) Xiujun Li, Lihong Li, Jianfeng Gao, Xiaodong He, Jianshu Chen, Li Deng, and Ji He. Recurrent reinforcement learning: A hybrid approach, 2015.
  • McNaughton et al. (1996) BL McNaughton, CA Barnes, JL Gerrard, K Gothard, MW Jung, JJ Knierim, H Kudrimoti, Y Qin, WE Skaggs, M Suster, and KL Weaver. Deciphering the hippocampal polyglot: the hippocampus as a path integration system. J Exp Biol., 12, 1996.
  • ME (2008) Hasselmo ME. Grid cell mechanisms and function: contributions of entorhinal persistent spiking and phase resetting. Hippocampus, pp.  19021258, 12 2008.
  • Mirowski et al. (2017) Piotr Mirowski, Razvan Pascanu, Fabio Viola, Hubert Soyer, Andrew J. Ballard, Andrea Banino, Misha Denil, Ross Goroshin, Laurent Sifre, Koray Kavukcuoglu, Dharshan Kumaran, and Raia Hadsell. Learning to navigate in complex environments, 2017.
  • Osband et al. (2016) Ian Osband, Charles Blundell, Alexander Pritzel, and Benjamin Van Roy. Deep exploration via bootstrapped DQN. CoRR, abs/1602.04621, 2016.
  • Rivero-Ortega et al. (2023) Jesús D. Rivero-Ortega, Juan S. Mosquera-Maturana, Josh Pardo-Cabrera, Julián Hurtado-López, Juan D. Hernández, Victor Romero-Cano, and David F. Ramírez-Moreno. Ring attractor bio-inspired neural network for social robot navigation. Frontiers in Neurorobotics, 17, 2023. ISSN 1662-5218. doi: 10.3389/fnbot.2023.1304597.
  • Schwarzer et al. (2020) Max Schwarzer, Ankesh Anand, Rishab Goel, R. Devon Hjelm, Aaron C. Courville, and Philip Bachman. Data-efficient reinforcement learning with momentum predictive representations. CoRR, abs/2007.05929, 2020.
  • Srinivas et al. (2020) Aravind Srinivas, Michael Laskin, and Pieter Abbeel. CURL: contrastive unsupervised representations for reinforcement learning. CoRR, abs/2004.04136, 2020.
  • Sun et al. (2020) Xuelong Sun, Shigang Yue, and Michael Mangan. A decentralised neural model explaining optimal integration of navigational strategies in insects. eLife, 9:e54026, jun 2020. ISSN 2050-084X. doi: 10.7554/eLife.54026.
  • Taube (2007) Jeffrey Taube. The head direction signal: Origins and sensory-motor integration. Annual review of neuroscience, 30:181–207, 02 2007. doi: 10.1146/annurev.neuro.29.051605.112854.
  • Taube (1995) JS Taube. Head direction cells recorded in the anterior thalamic nuclei of freely moving rats. Journal of Neuroscience, 15(1):70–86, 1995. ISSN 0270-6474. doi: 10.1523/JNEUROSCI.15-01-00070.1995.
  • Touretzky (2005) David Touretzky. Attractor network models of head direction cells. The Journal of neuroscience : the official journal of the Society for Neuroscience, pp.  411–432, 07 2005. doi: 10.7551/mitpress/3447.003.0026.
  • van Hasselt et al. (2015) Hado van Hasselt, Arthur Guez, and David Silver. Deep reinforcement learning with double q-learning. CoRR, abs/1509.06461, 2015.
  • Wang et al. (2018) Jane X Wang, Zeb Kurth-Nelson, Dharshan Kumaran, Dhruva Tirumala, Hubert Soyer, Joel Z Leibo, Demis Hassabis, and Matthew Botvinick. Prefrontal cortex as a meta-reinforcement learning system. Nature neuroscience, 21(6):860–868, 2018. ISSN 1097-6256.
  • Ye et al. (2021) Weirui Ye, Shaohuai Liu, Thanard Kurutach, Pieter Abbeel, and Yang Gao. Mastering atari games with limited data. CoRR, abs/2111.00210, 2021.
  • Zambaldi et al. (2019) Vinícius Flores Zambaldi, David Raposo, Adam Santoro, Victor Bapst, Yujia Li, Igor Babuschkin, Karl Tuyls, David P. Reichert, Timothy P. Lillicrap, Edward Lockhart, Murray Shanahan, Victoria Langston, Razvan Pascanu, Matthew M. Botvinick, Oriol Vinyals, and Peter W. Battaglia. Deep reinforcement learning with relational inductive biases. 2019.
  • Zhang (1996) K Zhang. Representation of spatial orientation by the intrinsic dynamics of the head-direction cell ensemble: a theory. Journal of Neuroscience, 16(6):2112–2126, 1996. ISSN 0270-6474. doi: 10.1523/JNEUROSCI.16-06-02112.1996.
  • Zugaro et al. (2001) Michaël Zugaro, Alain Berthoz, and Sidney Wiener. Background, but not foreground, spatial cues are taken as references for head direction responses by rat anterodorsal thalamus neurons. The Journal of neuroscience : the official journal of the Society for Neuroscience, 21:RC154, 08 2001.

Appendix A Appendix

A.1 Background:Attractor Networks Theoretical Foundations

Ring attractor networks are a type of biological neural structure that have been proposed to underlie the representation of various cognitive functions, including spatial navigation, working memory, and decision making (Kim et al., 2017).

Biological intuition. In the early 1990s, the research carried out by  Zhang (1996) proposed that ring neural structures could underlie the representation of heading direction in rodents. Zhang (1996) argued that the neural activity in an attractor network might encode the direction of the animal’s head, with the network transitioning from one attractor state to another state as the animal turns.

Empirical evidence. There is growing evidence from neuroscience supporting the role of ring attractors in neural processing. For example, electrophysiological recordings from head direction cells (HDCs) of rodents have revealed a circular organisation of these neurons, with neighbouring HDCs encoding slightly different heading directions (Taube, 1995). Furthermore, studies have shown that HDC activity can be influenced by sensory inputs, such as visual signals and vestibular signals, and that these inputs can cause the network to update its representation of heading direction (Taube, 2007).

Sensor fusion in ring attractors. Ring attractor networks provide a theoretical foundation for understanding cognitive functions such as spatial navigation, working memory, and decision making. In the context of action selection in RL, sensor fusion plays a pivotal role in augmenting the information processing capabilities of these networks. By combining data from various sensory modalities, ring attractors create a more nuanced and robust representation of the environment, essential for adaptive behaviors (ME, 2008). Research has elucidated the relatioshipn between ring attractors and sensory inputs, with the circular organisation of HDCs in rodents complemented by the convergence of visual and vestibular inputs, highlighting the integrative nature of sensory information within the ring attractor framework (Zugaro et al., 2001).

Modulation by sensory inputs. Beyond the spatial domain, sensory input dynamically influences the activity of ring attractor networks. Studies have shown that visual cues and vestibular signals not only update the representation of heading direction but also contribute to the stability of attractor states, allowing robust spatial memory and navigation (Goodridge et al., 1998).

Sensor fusion for action selection. The concept of sensor fusion within the context of ring attractors extends beyond traditional sensory modalities, encompassing diverse sources, such as proprioceptive and contextual cues (McNaughton et al., 1996). Building on the foundation of ring attractor networks discussed earlier, the integration of sensor fusion in the context of action selection involves fusing the action values associated with each potential action within the ring attractor framework. In particular, sensory information, previously shown to modulate the activity of ring attractor networks, extends its influence to the representation of action values. The inclusion of sensory information reflects a higher cognitive process, where the adaptable nature of ring attractor networks plays a central role in orchestrating optimal decision making and action selection in complex environments.

A.1.1 Exogenous Ring Attractor Model Ablation Study

To isolate the impact of the ring attractor structure, we conducted an ablation study comparing our full BDQNRA model against versions with the action space overlay in an incorrect distribution in the ring, Fig. 4. This incorrect distribution involves randomly rearranging the placement of actions within the ring, disrupting the natural topology of the action space.

For instance, this could mean placing opposing or unrelated actions side by side in the ring, such as pairing ”move left” with ”move down” instead of its natural opposite ”move right”. More generally, this incorrect distribution breaks the inherent relationships between actions that are typically preserved in the ring structure.

Refer to caption
Figure 4: Ablation study comparing BDQN variants in OpenAI Gym Super Mario Bros (Kauten, 2018). The plot shows cummulative reward over 1 million frames for three models: Standard BDQN (Azizzadenesheli et al., 2018) ; BDQNRA-UA with RA and Uncertainty Awareness (UA) implementing both the ring attractor behavior policy from Section 3.1.2 and the uncertainty quantification model from  3.1.3; and BDQNRA-RM, applying the same concepts from BDQNRA-UA, but randomly distributing the action space across the ring in each experiment. Displaying mean episodic returns over 10 averaged seeds.

A.1.2 DL RNN Ring Attractor Model Ablation Study

This ablation study focused on isolating the impact of the ring-shaped connectivity in our RNN-based ring attractor model. The key aspect of our experiment was to remove the circular weight distribution in both the forward pass (input-to-hidden connections) and the recurrent connections (hidden-to-hidden), while maintaining all other aspects of the RNN architecture. This approach allows us to directly assess the contribution of the spatial ring structure to the model’s performance.

Refer to caption
Refer to caption
Figure 5: Ablation study results comparing the performance of the full RNN-based ring attractor model against a version with the circular weight distribution removed. The graph illustrates a significant performance drop for the Ms Pacman and Chopper Command environments in the Atari 100K benchmark (Bellemare et al., 2012). This emphasises the role of the circular topology in encoding spatial information and enhancing learning. Displaying mean episodic returns over 10 averaged seeds.

In our original model, the weights between neurons were determined by a distance-dependent function that created a circular topology. This function assigned stronger connections between neurons that were close together in the ring and weaker connections between distant neurons. For the ablation, we replaced this distance-dependent weight function with standard weight matrices for both the input-to-hidden and hidden-to-hidden connections. This modification effectively transforms our ring attractor RNN into a standard RNN, where the weights are not constrained by the circular topology. We retained other key elements of the model, such as the learnable time constant and the non-linear transformation, to isolate the effect of the ring structure specifically.

A.1.3 Recurrent Neural Network Modeling

Refer to caption
Figure 6: RNN modeling ring attractor synaptic connections: The left shows the forward pass (input-to-hidden) as the agent output layer. The right depicts the hidden-to-hidden recurrent connection between inference time steps. The weighted connections of the sample for the neuron n3n3 demonstrated.

As seen before, excitatory neurons are organized in a circular pattern, with connection weights between neurons determined by a distance-weighted function mimicking the synaptic connection of biological neurons, as shown in Fig.6 is the structured connectivity of the RNN, which mimics the circular topology of biological ring attractors.