This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

A Novel Neuron Model of Visual Processor

Jizhao Liu [email protected] Jing Lian J C Sprott Yide Ma School of Information Science and Engineering, Lanzhou University, No.222, TianShui Road(south), Lanzhou, 730000, Gansu, China School of Electronics and Information Engineering, Lanzhou Jiaotong University, No.88, West Anning Road, Lanzhou, 730070, Gansu, China Department of Physics, University of Wisconsin, Madison, WI 53706, USA.
Abstract

Simulating and imitating the neuronal network of humans or mammals is a popular topic that has been explored for many years in the fields of pattern recognition and computer vision. Inspired by neuronal conduction characteristics in the primary visual cortex of cats, pulse-coupled neural networks (PCNNs) can exhibit synchronous oscillation behavior, which can process digital images without training. However, according to the study of single cells in the cat primary visual cortex, when a neuron is stimulated by an external periodic signal, the interspike-interval (ISI) distributions represent a multimodal distribution. This phenomenon cannot be explained by all PCNN models. By analyzing the working mechanism of the PCNN, we present a novel neuron model of the primary visual cortex consisting of a continuous-coupled neural network (CCNN). Our model inherited the threshold exponential decay and synchronous pulse oscillation property of the original PCNN model, and it can exhibit chaotic behavior consistent with the testing results of cat primary visual cortex neurons. Therefore, our CCNN model is closer to real visual neural networks. For image segmentation tasks, the algorithm based on CCNN model has better performance than the state-of-art of visual cortex neural network model. The strength of our approach is that it helps neurophysiologists further understand how the primary visual cortex works and can be used to quantitatively predict the temporal-spatial behavior of real neural networks. CCNN may also inspire engineers to create brain-inspired deep learning networks for artificial intelligence purposes.

keywords:
primary visual cortex model , pulse-coupled neural network, continuous-coupled neural network, brain-like computation.
journal: e-Print

1 Introduction

Creating brain-like machines is a long-standing goal in computer science. Neuromorphic computing efforts date back to the 1980s, when transistors were used to mimic the functionality of biological neurons and synapses. In the early 2000s, such research efforts (essentially spike-driven computations) facilitated the emergence of large-scale neuromorphic chips. Today, spiking neural networks (SNNs) are being actively explored using various algorithms. An SNN uses the time (pulse) of the signal to process information. Recently, research on SNNs has attracted the attention of many researchers from both hardware and algorithm aspects. Kaushik et al. stated that ”the field of neuromorphic computing is a synergistic effort equally weighted across both hardware and algorithmic domains to enable spike-based artificial intelligence. ”[1]

According to Maass, neural networks can be categorized into three different generations[2]. The first generation is based on using McCulloch-Pitts neurons as computational units, which perform threshold operations and output Boolean results. Hopfield networks and Boltzmann machines are based on these neural network models. The second generation is based on computational units that output continuous values. By applying a sigmoid unit or a rectified linear unit (ReLU) as an ”activation function”, these networks are able to evaluate a continuous set of output values and address complex and deep tasks. The current deep learning networks, which have multiple hidden layers between their input and output, are all based on such second-generation neurons. The third generation of networks primarily use spiking neurons of the ‘integrate-and-fire’ type that exchange information via spikes. Scientists currently believe that mammals use spikes to process information.

A pulse-coupled neural network (PCNN) is a third-generation artificial neural network proposed in the 1990s[3]. Inspired by the primary visual cortex of mammals such as cats, a PCNN can exhibit synchronous oscillation behavior and process digital images without training. In the past few decades, a large number of studies have proven that PCNN models achieve better performances in image segmentation[4], edge detection[5], noise reduction[6], fusion[7], enhancement[8] and quantization[9].

There are two main paths for studying PCNNs. The first path involves studying the parameter setting problem[10], because a PCNN is a multiparameter iteration network, and the network’s performance depends on settings such as the iteration numbers, the decay time constants and linking weight coefficients. Broussard was the first to attempt to set parameters adaptively by using the gradient descent method[11]. Yin et al. utilized the edge gradient to determine the linking coefficient β\beta of the PCNN, but other parameters were constants[12]. Li et al. proposed a novel autowave PCNN to solve the shortest path problem, adaptively setting the dynamic threshold according to the current network state[13]. Gao et al. proposed a modified PCNN mode by establishing a new neural threshold and a varying linking coefficient value[14]. Lian et al. utilized a parameter-adaptive PCNN model to construct an automatic segmentation algorithm[15, 16].

The second path proposes derivative models based on PCNN[3]. For example, Kinser simplified the PCNN model to the ICM model in 1996 and demonstrated that ICM was useful for target recognition[17]. In 2009, Zhan et al. proposed an SCM model by modifying the internal activity and firing condition of a PCNN. They showed that an SCM has lower computational complexity and achieves very high accuracy rates for image retrieval tasks compared with other common methods[18]. Two years later, Chen et al. simplified the SCM model to the SPCNN model by replacing the traditional firing condition[19]. Deng, et al. found that the firing time is not integer-based and proposed a quasi-continuous PCNN model in 2012. In recent years, heterogeneous PCNN models have been proposed because researchers realized that the models should be more consistent with real biological nervous systems[20].

The PCNN and its derivative models have achieved great success on image processing tasks because they simulate the transmission characteristics of neurons in the cerebral cortex visual zone. Ironically, these models do not fully mimic all the characteristics of visual cortical neurons. Siegel used an external periodic signal to stimulate a cat’s receptive fields and then recorded the signal of the primary visual cortex of the cat[21]. This experiment revealed that the interspike-interval (ISI) distributions represent a multimodal distribution, which all PCNNs cannot exhibit.

In recent years, Bittner el al. found that synaptic plasticity relies on the relative timing of pre- and post-synaptic spikes[22]. These features map to the synaptic weights of SNNs, including PCNN[23]. Therefore, it is very important to design a neuron model that generates more accurate predictions of stimulus responses. This motivated us to further study neuron models of the primary visual cortex. Using our method, not only can we predict the responses of neurons, but it may also inspire more efficient SNNs for artificial intelligence purposes.

In this work, we present a novel neuron model of the primary visual cortex by analyzing the working mechanism of the PCNN, i.e., the continuous-coupled neural network (CCNN). To model the CCNN, we follow the minimal criteria for a sensory encoding model proposed by Yamins and DiCarlo, which are stimulus-computability, mappability, and predictivity[24]. Our model inherits the threshold exponential decay and synchronous pulse oscillation property from the original PCNN model, and its ISI distributions are closer to those of real neurons when stimulated by periodic signals. Thus, the CCNN model more closely approaches real visual neural networks and may benefit the next generation of deep learning networks.

The rest of this paper is organized as follows. In Section II, we survey the relationship between the PCNN model and the results of neurophysiological experiments and analyze the output spike train data of the PCNN. Then, we analyzed the deficiencies and limitations of designing such SNNs for artificial intelligence purposes. In Section III, we introduce the CCNN model and analyze its dynamic behavior. In Section IV, we investigate the spiking feature of CCNN, demonstrate why the CCNN model is closer to the results of neurophysiological experiments and analyze how the CCNN may further benefit artificial intelligence. In Section V, we propose and evaluate an image segmentation algorithm; the comparative results indicate that our method achieves better performance than the image segmentation algorithm without training. Finally, Section VI provides conclusions and suggests future work.

2 Pulse-Coupled Neural Network

2.1 The biological feature of neurons

1. Neurons do not create an action potential (AP) immediately; instead, they undergo a refractory period[25]. The absolute refractory period (ARP) refers to the period during which a neuron creates two consecutive APs. The ARP is caused by the inactivation of the Na+ channels that originally opened to depolarize the membrane. These channels then remain inactivated until the membrane hyperpolarizes. The channels then close, de-inactivate, and regain their ability to open in response to a stimulus. The relative refractory period immediately follows the ARP. As the voltage-gated potassium channels open to terminate the action potential by repolarizing the membrane, the potassium conductance of the membrane increases dramatically. The K+ ions moving out of the cell move the membrane potential closer to the equilibrium potential for potassium, which causes a brief hyperpolarization of the membrane—that is, the membrane potential becomes transiently more negative than its normal resting potential. Until the potassium conductance returns to the resting value, a larger stimulus will be required to reach the initiation threshold to initiate a second depolarization. This period is called the relative refractory period (RRP).

2. In the primary visual cortex, coherent oscillations exist, which means that neurons create action potentials not only due to external stimuli but also from stimuli by other neurons. This process is called stimulus-induced movement. The original experiments were performed by Eckhorn et al., who recorded single and multiple spikes as well as local field potentials simultaneously from several locations in the primary visual cortex (A17 and A18) of a cat. They reported that ”Coherent stimulus-evoked resonances were found at distant cortical positions when at least one of the primary coding properties was similar. Coherence was found 1) within a vertical cortex column, 2) between neighboring hypercolumns, and 3) between two different cortical areas”[26].

2.2 The genesis of PCNN

Based on the above biological features, a pulse-coupled neural network (PCNN) was proposed[27]. The basic model’s main parts are couple linking, feeding input, modulation product, dynamic threshold, and spiking output. The couple-linking part represents the spiking signal received from other neurons. The feeding input part is the signal from external sources through the receptive fields, which can be pulses, analog time-varying signals, constants, or any combination. The received input signals are modulated through the modulation product part. The dynamic threshold part is used to map the refractory period of neurons. Spiking the output part creates an action potential signal when the signal satisfies the burst conditions. The model is shown in Fig. 1.

Refer to caption
Figure 1: Pulse-coupled neuron model

The following equation further describes the PCNN model:

Fij[n]=eαfFij[n1]+VFMijklYkl[n1]+SijLij[n]=eαlLij[n1]+VLWijklYkl[n1]Uij[n]=Fij[n](1+βLij[n])Yij[n]={1,ifUij[n]>Eij[n]0,otherwiseEij[n]=eαeEij[n1]+VEYij[n],\begin{split}\begin{aligned} F_{ij}[n]&=e^{-\alpha_{f}}F_{ij}[n-1]+V_{F}M_{ijkl}Y_{kl}[n-1]+S_{ij}\\ L_{ij}[n]&=e^{-\alpha_{l}}L_{ij}[n-1]+V_{L}W_{ijkl}Y_{kl}[n-1]\\ U_{ij}[n]&=F_{ij}[n](1+\beta L_{ij}[n])\\ Y_{ij}[n]&=\left\{\begin{aligned} 1,&\qquad ifU_{ij}[n]>E_{ij}[n]\\ 0,&\qquad otherwise\\ \end{aligned}\right.\\ E_{ij}[n]&=e^{-\alpha_{e}}E_{ij}[n-1]+V_{E}Y_{ij}[n]\end{aligned}\end{split}, (1)

where the five main parts are couple linking Lij[n]L_{ij}[n], feeding input Fij[n]F_{ij}[n], modulation product Uij[n]U_{ij}[n], dynamic threshold Eij[n]E_{ij}[n] and spiking output Yij[n]Y_{ij}[n]. SijS_{ij} is the external feeding input stimulus received by the receptive fields. αf\alpha_{f}, αl\alpha_{l} and αe\alpha_{e} denote exponential decay factors that record previous input states. VFV_{F} and VLV_{L} are weighting factors modulating the action potentials of surrounding neurons. Additionally, MijklM_{ijkl} and WijklW_{ijkl} denote the feeding and linking synaptic weights, respectively, and β\beta denotes the linking strength, which directly determines Lij[n]L_{ij}[n] in the modulation product Uij[n]U_{ij}[n].

From the above equations, the exponential decay factors ingeniously imitate the refractory period because if a neuron creates an action potential or a spike, the dynamic threshold will increase to a large value, but it decays as time passes; therefore, it will be easier to burst after a the refractory period. The couple linking part imitates coherent oscillation behavior if the right values are chosen.

Since PCNN and its derivative models can simulate the transmission characteristics of neurons in the cerebral cortex visual zone, they have achieved great success on some unsupervised image processing tasks.

2.3 The deficiencies and the limitations of PCNN

Action potentials are propagated along the axons of a neuron to reach the nerve terminals, where they can trigger the release of chemical messengers that affect other neurons. Therefore, spikes are an important way in which neurons in the brain carry information[28]. The interspike interval (ISI) is the time between successive action potentials (also known as spikes) of a neuron. Thus, ISI distributions are believed to be an effective tool for revealing the mechanism of brain code information[29].

Based on the data which represent the responses of V1 populations of neurons to natural image sequences [30], the spike train of a single neuron of the primary visual cortex (V1 area) of macaques can be calculated, which is shown as Fig. 2.111The data used here were collected in the laboratory of Dario Ringach at UCLA and downloaded from the CRCNS website. https://crcns.org/data-sets/vc/pvc-1/conditions. Fig. 2 indicate that the ISI distributions of a single neuron have many different periods when stimulated by natural images.

Refer to caption
Figure 2: Interspike interval of real neurons under natural image external stimulus

More detailed experiments were performed by Siegel, who used a hand-held projector to stimulate the cells of receptive fields with a high-speed square wave signal and collected the response spike train of a single neuron of the primary visual cortex of a cat (area 17). He found many patterns in the return maps of ISI distributions, see Fig. 5 in ref.[21].

However, the PCNN can output only a fixed-period response signal, which leads the ISI distributions of the PCNN to manifest as a single bar. To illustrate further, we use sine functions (periodically driven signals) and logistic maps (chaotically driven signals) as external stimulus signals SijS_{ij} and calculate the ISI distribution of the response spike train of the PCNN, as shown in Fig.3. Therefore, PCNN cannot mimic the dynamic behavior of real neurons.

Refer to caption
(a)
Refer to caption
(b)
Figure 3: Interspike interval of PCNN neurons under different stimuli. (a) S=0.5(1+sin(t))S=0.5(1+sin(t)); (b) S=S=Logistic Map

On the other hand, the general idea for using SNNs in artificial intelligence is to use the spike intervals as a type of error that can be backpropagated by algorithms[31]. The PCNN only outputs a spiking train with a fixed period. Therefore, it would seem impossible to utilize PCNN to design SNNs for artificial intelligence, despite PCNN having achieved great success on image processing tasks.

Based on the above analysis, a PCNN cannot predict the temporal-spatial behavior of real neural networks, and it cannot benefit artificial intelligence as other SNNs can. This limitation prevented further research and application into PCNNs.

3 Continuous-Coupled Neural Network

These deficiencies and limitations of the PCNN model motivated us to further explore models of the primary visual cortex. We followed Yamins’ criteria to design the CCNN. As described in[24], these are:

Stimulus computability: The model should accept arbitrary stimuli within the general stimulus domain of interest;

Mappability: The components of the model should correspond to experimentally definable components of the neural system; and

Predictivity: The units of the model should provide detailed predictions of stimulus-by-stimulus responses for arbitrarily chosen neurons in each mapped area.

3.1 CCNN Models

In neurophysiology experiments, nerve action potentials in response to changing stimuli are monitored and recorded by an electrode. Then, bandpass filtering is utilized to separate the relatively high frequency components in the action potential from low frequency noise, which stems from measurement equipment and the environment. Finally, the action potential is coded to the spike train by the all-or-none law: if the voltage exceeds the threshold potential, the nerve fiber will generate a complete response; otherwise, it provides no response. A flowchart of the coding action potential is shown in Fig.4.

Refer to caption
Figure 4: All-or-none coding with action potentials

Therefore, it is reasonable to modify the spiking part of the PCNN (Fig. 1) to a continuous output part (Fig. 5), to create the CCNN. When studying the spiking features of CCNN, one can generate the spiking train data by using an appropriate step function as the filter. The CCNN model is shown in Fig. 5.

The basic model’s main parts are couple linking, feeding input, modulation product, dynamic threshold, and continuous output. The couple linking part receives output signals from other neighboring neurons. The feeding input part consists of signals from external sources through the receptive fields. The modulation product part calculates all the received input signals. The dynamic activity part is used to map the refractory period of neurons. The continuous output part generates the response signal of the CCNN.

Refer to caption
Figure 5: Continuous coupled neuron model

The following equation further describes the CCNN model:

Fij[n]=eαfFij[n1]+VFMijklYkl[n1]+SijLij[n]=eαlLij[n1]+VLWijklYkl[n1]Uij[n]=Fij[n](1+βLij[n])Yij[n]=f(Uij[n]Eij[n])Eij[n]=eαeEij[n1]+VEYij[n1].\begin{split}\begin{aligned} F_{ij}[n]&=e^{-\alpha_{f}}F_{ij}[n-1]+V_{F}M_{ijkl}Y_{kl}[n-1]+S_{ij}\\ L_{ij}[n]&=e^{-\alpha_{l}}L_{ij}[n-1]+V_{L}W_{ijkl}Y_{kl}[n-1]\\ U_{ij}[n]&=F_{ij}[n](1+\beta L_{ij}[n])\\ Y_{ij}[n]&=f(U_{ij}[n]-E_{ij}[n])\\ E_{ij}[n]&=e^{-\alpha_{e}}E_{ij}[n-1]+V_{E}Y_{ij}[n-1]\end{aligned}\end{split}. (2)

3.2 The nonlinear function of continues output

As it is well known, the dynamic characteristics of a dynamic system vary with the nonlinear function. In order to design a system which has similar dynamic characteristics to real neurons, we use Sigmoid, TanH, ReLU and Softplus as the nonlinear function to test the dynamic behavior and ISI distributions. The function graph of the nonlinear functions are shown in Fig.6. When the system is stimulated by DC signals, the results are shown in Tab.1. When the system is stimulated by periodic signals, the results are shown in Tab.2.

Refer to caption
(a)
Refer to caption
(b)
Refer to caption
(c)
Refer to caption
(d)
Figure 6: Function graph of the nonlinear functions: (a) Sigmoid; (b) TanH; (c) ReLU; (d) Softplus
Nonlinear Function Equation Dynamic Behavior
Sigmoid f(x)=11+exf(x)=\frac{1}{1+e^{-x}} Periodic Behavior
TanH f(x)=exexex+exf(x)=\frac{e^{x}-e^{-x}}{e^{x}+e^{-x}} Periodic Behavior
ReLU f(x)=max(0,x)f(x)=max(0,x) Periodic Behavior
Softplus f(x)=log(1+ex)f(x)=log(1+e^{x}) Chaotic Behavior
Table 1: Dynamic behavior under DC signals
Nonlinear Function Equation Chaotic Behavior
Sigmoid f(x)=11+exf(x)=\frac{1}{1+e^{-x}} Chaotic Behavior
TanH f(x)=exexex+exf(x)=\frac{e^{x}-e^{-x}}{e^{x}+e^{-x}} Periodic Behavior
ReLU f(x)=max(0,x)f(x)=max(0,x) Periodic Behavior
Softplus f(x)=log(1+ex)f(x)=log(1+e^{x}) Chaotic Behavior
Table 2: Dynamic behavior under periodic signals

According to Hodgkin and Siegel, real neurons will exhibit periodic behavior under DC drive signal, and they will exhibit chaotic behavior under periodic drive signal. Therefore, using Sigmoid as the nonlinear function of CCNN model is a reasonable choice. The CCNN model is described by:

Fij[n]=eαfFij[n1]+VFMijklYkl[n1]+SijLij[n]=eαlLij[n1]+VLWijklYkl[n1]Uij[n]=Fij[n](1+βLij[n])Yij[n]=1/(1+e(Uij[n]Eij[n]))Eij[n]=eαeEij[n1]+VEYij[n1].\begin{split}\begin{aligned} F_{ij}[n]&=e^{-\alpha_{f}}F_{ij}[n-1]+V_{F}M_{ijkl}Y_{kl}[n-1]+S_{ij}\\ L_{ij}[n]&=e^{-\alpha_{l}}L_{ij}[n-1]+V_{L}W_{ijkl}Y_{kl}[n-1]\\ U_{ij}[n]&=F_{ij}[n](1+\beta L_{ij}[n])\\ Y_{ij}[n]&=1/(1+e^{-(U_{ij}[n]-E_{ij}[n])})\\ E_{ij}[n]&=e^{-\alpha_{e}}E_{ij}[n-1]+V_{E}Y_{ij}[n-1]\end{aligned}\end{split}. (3)

Later, we will illustrate that the CCNN model can exhibit both periodic and chaotic behavior, causing the model to generate numerous spike train sequences with a threshold filter. This property empowers the CCNN model to achieve better performances on image processing applications and may further benefit the use of spiking neural networks for artificial intelligence purposes. The various spike train patterns are also more similar to those of real visual networks.

3.3 Dynamic behavior of CCNN: Noncoupling with neighborhood neurons

3.3.1 Constant stimulus

The single neuron model can be simplified from Eq. (2) by setting Mijkl=0M_{ijkl}=0, Wijkl=0W_{ijkl}=0 and β=0\beta=0. The equations of the single neuron model are given by:

F[n]=eαfF[n1]+SU[n]=F[n]Y[n]=1/(1+e(U[n]E[n]))E[n]=eαeE[n1]+VEY[n1].\begin{split}\begin{aligned} F[n]&=e^{-\alpha_{f}}F[n-1]+S\\ U[n]&=F[n]\\ Y[n]&=1/(1+e^{-(U[n]-E[n])})\\ E[n]&=e^{-\alpha_{e}}E[n-1]+V_{E}Y[n-1]\end{aligned}\end{split}. (4)

When S=0S=0, the nerve neuron has no input signal, and Eq.2 can be simplified to:

F[n]=0U[n]=F[n]=0Y[n]=1/(1+eE[n])E[n]=eαeE[n1]+VEY[n1].\begin{split}\begin{aligned} F[n]&=0\\ U[n]&=F[n]=0\\ Y[n]&=1/(1+e^{E[n]})\\ E[n]&=e^{-\alpha_{e}}E[n-1]+V_{E}Y[n-1]\end{aligned}\end{split}. (5)

Substituting the fourth equation into the fifth equation, we obtained the recursive formula of EnE_{n}, which is:

En=eαeE[n1]+VE/(1+eE[n1]).\begin{split}\begin{aligned} E_{n}=e^{-\alpha_{e}}E[n-1]+V_{E}/(1+e^{E[n-1]})\end{aligned}\end{split}. (6)

The fixed point of Eq. (6) can be calculated by:

E=eαeE+VE/(1+eE);\begin{split}\begin{aligned} E=e^{-\alpha_{e}}E+V_{E}/(1+e^{E})\end{aligned}\end{split}; (7)

therefore, the fixed points are as follows:

E(1+eE)=VE/(1eαe).\begin{split}\begin{aligned} E(1+e^{E})=V_{E}/(1-e^{-\alpha_{e}})\end{aligned}\end{split}. (8)

Since f(E)=E(1+eE)f(E)=E(1+e^{E}) is a monotonically increasing function, when EE\rightarrow-\infty, f(E)0f(E)\rightarrow 0. Thus, when VE/(1eαe)>0V_{E}/(1-e^{-\alpha_{e}})>0, the CCNN model has only one fixed point; otherwise, it has no fixed point. In fact, VEV_{E} is the reset voltage, which is always positive, and the CCNN has only one fixed point if and only if αe>0\alpha_{e}>0.

Next, we investigate whether this fixed point is stable. We have the following theorem.

Theorem: The necessary condition for the fixed points or period-1 orbit to be stable is that the absolute value of the first derivative of Eq. (7) must be less than 11.

Proof: For a nonlinear function x=f(x)x=f(x), assume that a slight disturbance occurs at the fixed point xx^{*}. The equation of the fixed point is:

x+εn+1=f(x+εn),\begin{split}\begin{aligned} x^{*}+\varepsilon_{n+1}=f(x^{*}+\varepsilon_{n})\end{aligned}\end{split}, (9)

where εn\varepsilon_{n} is the distance before the disturbance and εn+1\varepsilon_{n+1} is the distance after the disturbance. We expand the right-hand side of Eq. (9) and obtain the linear term of εn\varepsilon_{n} as follows:

x+εn+1=f(x)+f(x)x|x=xεn+.\begin{split}\begin{aligned} x^{*}+\varepsilon_{n+1}=f(x)+\frac{\partial f(x)}{\partial x}|_{x=x^{*}}\varepsilon_{n}+\cdots\end{aligned}\end{split}. (10)

Substituting x=f(x)x^{*}=f(x^{*}) into Eq. (10), we have:

εn+1εn=f(x)x|x=xεn.\begin{split}\begin{aligned} \dfrac{\varepsilon_{n+1}}{\varepsilon_{n}}=\frac{\partial f(x)}{\partial x}|_{x=x^{*}}\varepsilon_{n}\end{aligned}\end{split}. (11)

For a stable fixed point, |εn+1|<|εn||\varepsilon_{n+1}|<|\varepsilon_{n}|. Therefore, the stable condition of the fixed point is:

s=f(x)x|x=xεn1,\begin{split}\begin{aligned} s=\frac{\partial f(x)}{\partial x}|_{x=x^{*}}\varepsilon_{n}\leq 1\end{aligned}\end{split}, (12)

where s=1s=1 is the stable boundary. For f(x)=1f^{\prime}(x)=1, the dynamic system exhibits tangent bifurcation behavior, while for f(x)=1f^{\prime}(x)=-1, the dynamic system exhibits period doubling bifurcation behavior.

The derivative function of the fixed point is given by:

f(E)E=eαeVEeE(1+eE)2.\begin{split}\begin{aligned} \frac{\partial f(E)}{\partial E}=e^{-\alpha_{e}}-V_{E}*\frac{e^{E}}{(1+e^{E})^{2}}\end{aligned}\end{split}. (13)

When |f(E)E|1|\frac{\partial f(E)}{\partial E}|\leq 1, the fixed points are stable. To solve the critical value of Eq. (13), we can use the variable substitution method. Assuming y=eEy=e^{E}, Eq. (13) can be simplified to:

y(1+y)2=1±eαeVE.\begin{split}\begin{aligned} \frac{y}{(1+y)^{2}}=\frac{1\pm e^{-\alpha_{e}}}{V_{E}}\end{aligned}\end{split}. (14)

Since y=eE0y=e^{E}\geq 0, when f(y)=12y(1+y)2=0f^{\prime}(y)=\frac{1-2y}{(1+y)^{2}}=0, y=0.5y=0.5, E=ln0.5E=ln0.5, f(y)f(y) reached its extreme point, fmax(y)=y(1+y)2=0.22f_{max}(y)=\frac{y}{(1+y)^{2}}=0.22. Therefore, if 1±eαeVE0.22\frac{1\pm e^{-\alpha_{e}}}{V_{E}}\leq 0.22, ln(0.22VE1)αeln(0.22VE1)ln(-0.22V_{E}-1)\leq\mp\alpha_{e}\leq ln(0.22V_{E}-1), the fixed point is stable; otherwise, the fixed point is unstable.

When S=C,F[0]=0S=C,F[0]=0, F[n]=eαfF[n1]+C=C(1+eαe+e2αe++e(n1)αe)=C(1enαe)/(1eαe)=CSF[n]=e^{-\alpha_{f}}F[n-1]+C=C(1+e^{-\alpha_{e}}+e^{-2\alpha_{e}}+\cdots+e^{-(n-1)\alpha_{e}})=C(1-e^{-n\alpha_{e}})/(1-e^{\alpha_{e}})=C_{S}. Eq. (5) can be simplified to:

Y[n]=1/(1+eE[n]CS)E[n]=eαeE[n1]+VEY[n1].\begin{split}\begin{aligned} Y[n]&=1/(1+e^{E[n]-C_{S}})\\ E[n]&=e^{-\alpha_{e}}E[n-1]+V_{E}Y[n-1]\end{aligned}\end{split}. (15)

Substituting Y[n1]Y[n-1] into E[n]E[n], we obtain the recursive formula of EnE_{n}, which is:

En=eαeE[n1]+VE/(1+eE[n1]CS).\begin{split}\begin{aligned} E_{n}=e^{-\alpha_{e}}E[n-1]+V_{E}/(1+e^{E[n-1]-C_{S}})\end{aligned}\end{split}. (16)

Therefore, the fixed points are:

E(1+eECS)=VE/(1eαe).\begin{split}\begin{aligned} E(1+e^{E-C_{S}})=V_{E}/(1-e^{-\alpha_{e}})\end{aligned}\end{split}. (17)

The derivative function of the fixed point is given by:

f(E)E=eαeVEeECS(1+eECS)2.\begin{split}\begin{aligned} \frac{\partial f(E)}{\partial E}=e^{-\alpha_{e}}-V_{E}*\frac{e^{E-C_{S}}}{(1+e^{E-C_{S}})^{2}}\end{aligned}\end{split}. (18)

When E=ln0.5+CSE=ln0.5+C_{S}, eECS(1+eECS)2\frac{e^{E-C_{S}}}{(1+e^{E-C_{S}})^{2}} reaches its extreme point of 0.220.22. If 1±eαeVE0.22\frac{1\pm e^{-\alpha_{e}}}{V_{E}}\leq 0.22, ln(0.22VE1)αeln(0.22VE1)ln(-0.22V_{E}-1)\leq\mp\alpha_{e}\leq ln(0.22V_{E}-1), the fixed point is stable; otherwise, the fixed point is unstable.

Let En=h(n)E_{n}=h(n); then, the derivative function of Eq. (16) can be given by:

f(E)n=f(E)EEn=f(E)Eh(n)n=(eαeVEeECS(1+eECS)2)h(n)n.\begin{split}\begin{aligned} \frac{\partial f(E)}{\partial n}&=\frac{\partial f(E)}{\partial E}*\frac{\partial E}{\partial n}=\frac{\partial f(E)}{\partial E}*\frac{\partial h(n)}{\partial n}\\ &=(e^{-\alpha_{e}}-V_{E}*\frac{e^{E-C_{S}}}{(1+e^{E-C_{S}})^{2}})*\frac{\partial h(n)}{\partial n}\end{aligned}\end{split}. (19)

If fT(E)=Ef^{T}(E)=E, then TT is the period of EE. Since eECS>0e^{E-C_{S}}>0, g(CS)=eαeVEeECS(1+eECS)2g(C_{S})=e^{-\alpha_{e}}-V_{E}*\frac{e^{E-C_{S}}}{(1+e^{E-C_{S}})^{2}} is a monotonically increasing function. Therefore, for a specific period of fixed points of EE, a larger CSC_{S} will lead to a larger f(E)/n{\partial f(E)}/{\partial n}; therefore, the fixed period-1 orbit will return to its starting point faster. In other words, increasing CSC_{S} decreases the periods of EE and YY. This phenomenon can be verified through an experiment, as shown in Fig.7.

3.3.2 Periodic stimulus

When the CCNN is driven by a periodic stimulus, it exhibits very complex dynamics, as shown in Fig. 7(d).

Refer to caption
(a)
Refer to caption
(b)
Refer to caption
(c)
Refer to caption
(d)
Figure 7: CCNN neuron under different stimulus signals, the parameter is set to αf=0.1\alpha_{f}=0.1, αe=1\alpha_{e}=1, VE=50V_{E}=50. (a) S=0S=0; (b) S=1S=1; (c) S=3S=3; (d) S=1+sin(t)S=1+sin(t)

A phase space is a space in which all possible states of a system are represented, with each possible state corresponding to one unique point in the phase space. We use values of F(t)F(t) as x-coordinate, values of E(t)E(t) as y-coordinate and values of Y(t)Y(t) as z-coordinate to form phase space plots. The phase space plots of Fig. 7(d) are shown in Fig. 8. The orbit of the CCNN does not repeat, diverge or converge, which indicates that the CCNN model exhibits chaotic behavior.

Refer to caption
Figure 8: Phase-space plots of a single CCNN neuron, S=sin(t)S=sin(t)

To further study this behavior, we calculated the Lyapunov exponent spectrum[32, 33]. When the largest Lyapunov exponent (LLE) is greater than zero, the system exhibits chaotic behavior. When the LLE is equal to zero, the system exhibits periodic behavior. When the LLE is less than zero, the trajectory of the system will converge to a stable fixed point.

When αf=0.1\alpha_{f}=0.1, αe=1.0\alpha_{e}=1.0, and VE=50V_{E}=50, S=0.5(1+sin(t))S=0.5(1+sin(t)), the LLE converges to 0.090.09, which indicates that the neuron is in a chaotic state. Furthermore, we used the Lyapunov exponent spectrum to investigate the overall dynamic behavior of single neuron under different stimulus signals. For S=A(1+sin(ωt))S=A(1+sin(\omega t)), A[0,4]A\in[0,4], ω[0,20]\omega\in[0,20], we calculate the corresponding LLE of CCNN. The results are shown in Fig.9.

Refer to caption
(a) S=A(1+sin(ωt))S=A(1+sin(\omega t))
Refer to caption
(b) S=Asin(ωt)S=Asin(\omega t)
Figure 9: Lyapunov exponent spectrum of single neuron under different stimulus signals.

Clearly, the CCNN model can exhibit both periodic and chaotic behavior when stimulated by different driven signals. As shown in Fig. 9, when S=A(1+sin(ωt))S=A(1+sin(\omega t)), the CCNN model is sensitive to the amplitude of the driven signal. However, when S=A(sin(ωt))S=A(sin(\omega t)), the CCNN is sensitive to the frequency of the driven signal. This interesting feature further empowers CCNN to generate a variety of response signals that reflect different input signals.

3.4 Dynamic behavior of CCNN: Coupling with the neighborhood neurons

When a CCNN neuron is coupled with neighborhood neurons, its equation is given by:

F[n]=eαfF[n1]+S+ΣML[n]=eαlL[n1]+ΣWU[n]=F[n](1+βL[n])Y[n]=1/(1+e(U[n]E[n]))E[n]=eαeE[n1]+VEY[n1]\begin{split}\begin{aligned} F[n]&=e^{-\alpha_{f}}F[n-1]+S+\Sigma M\\ L[n]&=e^{-\alpha_{l}}L[n-1]+\Sigma W\\ U[n]&=F[n](1+\beta L[n])\\ Y[n]&=1/(1+e^{-(U[n]-E[n])})\\ E[n]&=e^{-\alpha_{e}}E[n-1]+V_{E}Y[n-1]\end{aligned}\end{split} (20)

where ΣM\Sigma M is the input feed from neighborhood neurons in the receptive field and ΣW\Sigma W is the coupling linking input of connected neurons.

When S=CS=C, F[0]=0F[0]=0, L[0]=0L[0]=0,

F[n]=eαfF[n1]+C+ΣM=(C+ΣM)(1+eαe+e2αe++e(n1)αe)=(C+ΣM)(1enαe)/(1eαe)=CSm.\begin{split}\begin{aligned} F[n]&=e^{-\alpha_{f}}F[n-1]+C+\Sigma M\\ &=(C+\Sigma M)(1+e^{-\alpha_{e}}+e^{-2\alpha_{e}}+\cdots+e^{-(n-1)\alpha_{e}})\\ &=(C+\Sigma M)(1-e^{-n\alpha_{e}})/(1-e^{\alpha_{e}})=C_{Sm}.\end{aligned}\end{split} (21)
L[n]=eαfF[n1]+ΣW=ΣW(1+eαe+e2αe++e(n1)αe)=ΣW(1enαe)/(1eαe)=CSw.\begin{split}\begin{aligned} L[n]&=e^{-\alpha_{f}}F[n-1]+\Sigma W\\ &=\Sigma W(1+e^{-\alpha_{e}}+e^{-2\alpha_{e}}+\cdots+e^{-(n-1)\alpha_{e}})\\ &=\Sigma W(1-e^{-n\alpha_{e}})/(1-e^{\alpha_{e}})=C_{Sw}\end{aligned}\end{split}. (22)

Therefore, Eq. (20) can be simplified to:

Y[n]=1/(1+eE[n]CSm(1+βCSw))E[n]=eαeE[n1]+VEY[n1].\begin{split}\begin{aligned} Y[n]&=1/(1+e^{E[n]-C_{Sm}(1+\beta C_{Sw})})\\ E[n]&=e^{-\alpha_{e}}E[n-1]+V_{E}Y[n-1]\end{aligned}\end{split}. (23)

Substituting Y[n1]Y[n-1] into E[n]E[n], we obtained the recursive formula of EnE_{n}, which is:

En=eαeE[n1]+VE/(1+eE[n1]CSm(1+βCSw)).\begin{split}\begin{aligned} E_{n}=e^{-\alpha_{e}}E[n-1]+V_{E}/(1+e^{E[n-1]-C_{Sm}(1+\beta C_{Sw})})\end{aligned}\end{split}. (24)

Therefore, the fixed points are:

E(1+eECSm(1+βCSw))=VE/(1eαe).\begin{split}\begin{aligned} E(1+e^{E-C_{Sm}(1+\beta C_{Sw})})=V_{E}/(1-e^{-\alpha_{e}})\end{aligned}\end{split}. (25)

The derivative function of the fixed point is given by:

f(E)E=eαeVEeECSm(1+βCSw)(1+eECSm(1+βCSw)2.\begin{split}\begin{aligned} \frac{\partial f(E)}{\partial E}=e^{-\alpha_{e}}-V_{E}*\frac{e^{E-C_{Sm}(1+\beta C_{Sw})}}{(1+e^{E-C_{Sm}(1+\beta C_{Sw}})^{2}}\end{aligned}\end{split}. (26)

When E=ln0.5+CSm(1+βCSw)E=ln0.5+C_{Sm}(1+\beta C_{Sw}), eECSm(1+βCSw)(1+eECSm(1+βCSw))2\frac{e^{E-C_{Sm}(1+\beta C_{Sw})}}{(1+e^{E-C_{Sm}(1+\beta C_{Sw})})^{2}} reaches its extreme point of 0.220.22. If 1±eαeVE0.22\frac{1\pm e^{-\alpha_{e}}}{V_{E}}\leq 0.22, ln(0.22VE1)αeln(0.22VE1)ln(-0.22V_{E}-1)\leq\mp\alpha_{e}\leq ln(0.22V_{E}-1), the fixed point is stable; otherwise, the fixed point is unstable.

Let En=h(n)E_{n}=h(n); then, the derivative function of Eq. (16) can be given as follows:

f(E)n=f(E)EEn=f(E)Eh(n)n=(eαeVEeECSm(1+βCSw)(1+eECSm(1+βCSw)2)h(n)n.\begin{split}\begin{aligned} \frac{\partial f(E)}{\partial n}&=\frac{\partial f(E)}{\partial E}*\frac{\partial E}{\partial n}=\frac{\partial f(E)}{\partial E}*\frac{\partial h(n)}{\partial n}\\ &=(e^{-\alpha_{e}}-V_{E}*\frac{e^{E-C_{Sm}(1+\beta C_{Sw})}}{(1+e^{E-C_{Sm}(1+\beta C_{Sw}})^{2}})*\frac{\partial h(n)}{\partial n}\end{aligned}\end{split}. (27)

If fT(E)=Ef^{T}(E)=E, then TT is the period of EE. Since eECSm(1+βCSw)>0e^{E-C_{Sm}(1+\beta C_{Sw})}>0, g(CS)=eαeVEeECSm(1+βCSw)(1+eECSm(1+βCSw))2g(C_{S})=e^{-\alpha_{e}}-V_{E}*\frac{e^{E-C_{Sm}(1+\beta C_{Sw})}}{(1+e^{E-C_{Sm}(1+\beta C_{Sw})})^{2}} is a monotonically increasing function. Since Y[n]=1/(1+e(U[n]E[n]))>0Y[n]=1/(1+e^{-(U[n]-E[n])})>0, ΣM\Sigma M and ΣW\Sigma W are always positive. Therefore, when a neuron is coupled with its neighborhood neurons, it will have a higher output signal frequency.

4 Spiking Feature of CCNN

To evaluate the spiking feature of the CCNN, we utilized a threshold filter. The flowchart is shown in Fig. 10.

Refer to caption
Figure 10: Flowchart for the CCNN sequence of generating spike train

In Fig. 10 the threshold filter is designed to be:

YS[n]={1,ifYS[n]>μmax(Y)0,otherwise,\begin{split}\begin{aligned} Y_{S}[n]&=\left\{\begin{aligned} 1,&\qquad if\qquad Y_{S}[n]>\mu max(Y)\\ 0,&\qquad otherwise\\ \end{aligned}\right.\\ \end{aligned}\end{split}, (28)

where μ\mu is a coefficient. When μ=0.8\mu=0.8, S=0.21(square(ωt,dc)+1)S=0.21*(square(\omega*t,dc)+1) with angular velocity ω=0.2π\omega=0.2\pi and duty cycle dc=50dc=50. The spike train and ISI distributions are shown in Fig. 11.

Refer to caption
(a)
Refer to caption
(b)
Figure 11: Spike train and ISI distributions under a stimulus of y=0.21(square(ωt,dc)+1)y=0.21*(square(\omega*t,dc)+1), where ω=0.2π\omega=0.2\pi and dc=50dc=50

Comparing Fig. 11(b) with Fig. 2 indicates that the CCNN model can generate a pattern similar to that of real neurons.

Siegel performed detailed experiments of real signal neurons of the primary visual cortex and reported that ”Stimuli were presented at 271.5 ms intervals. A burst of action potentials can be seen following each stimulus. By plotting the distribution of interspike intervals, it can be seen that there is a major periodicity at approximately 270 ms. There are also peaks at 175, 540, and 1080 ms. The latter two values are roughly integer multiples of the driving period” [21].

The CCNN model can reproduce this phenomenon perfectly. When the stimulus is y=0.4(square(ωt,dc)+1)y=0.4*(square(\omega*t,dc)+1), where ω=0.2π\omega=0.2\pi (driven period is T=10T=10) and dc=50dc=50, the ISI distributions and return are shown in Fig. 12.

Refer to caption
Refer to caption
Figure 12: ISI distributions and return map of the i-th spike - (i+1)-th spike under the stimulus of y=0.4(square(ωt,dc)+1)y=0.4*(square(\omega*t,dc)+1), where ω=0.2π\omega=0.2\pi (driven period is T=10T=10) and dc=50dc=50

Furthermore, Siegel showed that a single neuron of the primary visual cortex can exhibit different patterns under different stimulus frequencies (see Fig. 3 Fig. 5 in ref[21]). We used different frequencies of the driven signal to stimulate the CCNN model and also found that different patterns of spike trains exist.

When the stimulus signal is set to y=0.21(square(ωt,dc)+1)y=0.21*(square(\omega*t,dc)+1), ω=2π\omega=2\pi is the angular velocity and dc=50dc=50 is the duty cycle. The spike train and ISI distributions are shown in Fig.13 (a) and (b), indicating that neurons exhibit periodic behavior.

When the frequency of the stimulus signal is decreased to ω=1.2π\omega=1.2\pi, the neuron starts to exhibit complex behavior. The main frequency is equal to the stimulus signal’s frequency, but frequencies of four to forty-four times the stimulus signal frequency can be observed, and their intensity decreases with increasing frequency. Small windows with frequencies of 1 and 4 times the frequency of the stimulus signal exist; however, the neuron does not output double and triple frequency signals, as shown in Fig.13 (c) and (d).

When the frequency of the stimulus signal is decreased to ω=1π\omega=1\pi, the neuron exhibits a complex temporal pattern, as shown in Fig.13 (e) and (f).

Refer to caption
(a)
Refer to caption
(b)
Refer to caption
(c)
Refer to caption
(d)
Refer to caption
(e)
Refer to caption
(f)
Figure 13: Spike train of CCNN under periodic stimulus signal: (a) shows the ISI distributions under the stimulus of y=0.21(square(ωt,dc)+1)y=0.21*(square(\omega*t,dc)+1), where ω=2π\omega=2\pi and dc=50dc=50; (b) is the return map under the stimulus of y=0.21(square(ωt,dc)+1)y=0.21*(square(\omega*t,dc)+1), where ω=2π\omega=2\pi and dc=50dc=50; (c) shows the ISI distributions under the stimulus of y=0.21(square(ωt,dc)+1)y=0.21*(square(\omega*t,dc)+1), where ω=1.2π\omega=1.2\pi and dc=50dc=50; (d) is the return map under the stimulus of y=0.21(square(ωt,dc)+1)y=0.21*(square(\omega*t,dc)+1), where ω=1.2π\omega=1.2\pi and dc=50dc=50; (e) shows the ISI distributions under the stimulus of y=0.21(square(ωt,dc)+1)y=0.21*(square(\omega*t,dc)+1), where ω=π\omega=\pi and dc=50dc=50; (f) is the return map under the stimulus of y=0.21(square(ωt,dc)+1)y=0.21*(square(\omega*t,dc)+1), where ω=π\omega=\pi and dc=50dc=50;

The results of the above experiments demonstrate that the CCNN model is more similar to a real primary visual cortex neuron network.

The ultimate advantage of SNNs comes from their ability to fully exploit spatiotemporal event-based information. Using CCNN as a unit neuron allows richer spike trains to be output from different input signals. These spike trains can be used as a type of ”error” that can be backpropagated by algorithms. Therefore, it is possible to utilize CCNN to design deep learning networks for artificial intelligence purposes.

5 Applications in Image Segmentation

Since CCNN model is more similar to the visual cortex, it is natural to utilize this model to perform image processing tasks. In this work, we use applied CCNN model into a image segmentation scheme. The flowchart is shown in Fig. 14.

Refer to caption
Figure 14: Flowchart of image segmentation

Where the filter is designed as follows:

YS[n]={1,ifYS[n]>μmax(I)0,otherwise,\begin{split}\begin{aligned} Y_{S}[n]&=\left\{\begin{aligned} 1,&\qquad if\qquad Y_{S}[n]>\mu max(I)\\ 0,&\qquad otherwise\\ \end{aligned}\right.\\ \end{aligned}\end{split}, (29)

where μ\mu is a coefficient set to 0.330.33 in this application, and max(I)max(I) is the largest pixel value of the image.

The main image segmentation procedures are described as follows:

Step 1: Generate the neural network which has the same size of the input image.

Step 2: Calculate standard deviation and otsu threshold of the input image and set the exponential decay factors based on them. Set the initial state of all neurons. The initial state is arbitrary, in this application we set all initial state to be zero for simplicity,

Step 3: Set coupling link mode. In this application, we set the coupling link mode so that each neuron only connects to the nearest 8 neurons around it. The linking weight is set to be the reciprocal of the Euclidean distance. For neuron N(i,j)N(i,j) and N(k,l)N(k,l), the linking weight is Mijkl,Wijkl=1(ik)2+(jl)2M_{ijkl},W_{ijkl}=\frac{1}{(i-k)^{2}+(j-l)^{2}}.

Step 4: Using input image as the feeding input FF, calculate each neuron’s modulation product UU, dynamic threshold EE and output YY and YSY_{S} Step 5: Repeat step 4 until the results is converged. Step 6: Using the final YSY_{S} to generate the segmentation image.

The pseudocode of the algorithm is shown in Algorithm 1.

Algorithm 1 Algorithm for Image Segmentation
0:  Original image (OI)
0:  Segmentation result (SR)
  theta=std2(OI)theta=std2(OI)
  Shist=gh(OI)Shist=gh(OI)
  Smax=max(max(OI))Smax=max(max(OI))
  af=log(1/theta)af=log(1/theta)
  vl=1vl=1
  ve=exp(af)+1+6Bvlve=exp(-af)+1+6*B*vl
  M3=(1exp((3af)))/(1exp(af))+6Bvlexp(af)M3=(1-exp(-(3*af)))/(1-exp(-af))+6*B*vl*exp(-af)
  ae=log(ve/(ShistM3))ae=log(ve/(Shist*M3))
  W=[0.510.51010.510.5]W=\begin{bmatrix}0.5&1&0.5\\ 1&0&1\\ 0.5&1&0.5\end{bmatrix}
  W=MW=M
  Y=0Y=0
  L=0L=0
  U=0U=0
  E=0E=0
  repeat
     K=conv2(Y,M,same)K=conv2(Y,M,^{\prime}same^{\prime})
     F=OIF=OI
     L=vlKL=vl*K
     U=exp(af)U+F(1+BL)U=exp(-af)*U+F*(1+B*L)
     Y=1/(1+exp(EU))Y=1/(1+exp(E-U))
     E=exp(ae)E+veYE=exp(-ae)*E+ve*Y
     if Y>μmax(OI)Y>\mu*max(OI)then
        Y=1Y=1
     else
        Y=0Y=0
     end if
     SRYSR\leftarrow Y
  until  Segmentation result is converged
  return  Segmentation result (SR)

To evaluate the effectiveness of the proposed segmentation method, we use BSDS500 dataset, which is designed for evaluating natural edge detection that includes not only object contours but also background boundaries. The visual segmentation results are shown in Fig.15.

Refer to caption
(a)
Refer to caption
(b)
Refer to caption
(c)
Refer to caption
(d)
Refer to caption
(e)
Refer to caption
(f)
Figure 15: Natural image segmentation results: (a) original image of boat; (b) segmentation results of boat; (c) original image of plane; (d) segmentation results of plane; (e) original image of duck; (f) segmentation results of duck;

The results of Fig.15(a)–Fig.15(d) have shown that the proposed method can effectively extract the target object from the background. However, our method is an unsupervised image segmentation method and depends on the difference of pixel values of the target object and the background. Therefore, our method cannot distinguish the object which has different pixel values, as shown in Fig.15(e) and Fig.15(f).

Different from natural images, medical images use the different characteristics of normal and abnormal tissues in X-rays (CT images) or magnetic fields (MRI images) to identify the lesion area. In medical images, if the pixel value of one area is obviously different from other areas, then the area is likely to be the lesion area. Therefore, our method can effectively perform medical image segmentation tasks.

To evaluate the effectiveness of the proposed segmentation method, we conducted experiments on images from the Mammographic Image Analysis Society (MIAS) database. For this database, the filter parameters μ\mu is set to be 0.450.45. We selected different types of breast images; the details are shown in Tab. 3.

Reference Number Background Tissue Class of abnormality present Severity of abnormality
mdb007 Fatty-glandular Normal N/A
mdb030 Fatty-glandular Other, ill-defined masses Benign
mdb158 Fatty Architectural distortion Malignant
mdb120 Fatty-glandular Architectural distortion Malignant
mdb121 Fatty-glandular Architectural distortion Benign
mdb238 Fatty Calcification Malignant
Table 3: The breast image types

The visual segmentation results are shown in Fig.16.

Refer to caption
(a)
Refer to caption
(b)
Refer to caption
(c)
Refer to caption
(d)
Refer to caption
(e)
Refer to caption
(f)
Refer to caption
(g)
Refer to caption
(h)
Refer to caption
(i)
Refer to caption
(j)
Refer to caption
(k)
Refer to caption
(l)
Figure 16: Breast image segmentation results: (a) original image of mdb007; (b) segmentation results of mdb007; (c) original image of mdb030; (d) segmentation results of mdb030; (e) original image of mdb158; (f) segmentation results of mdb158; (g) original image of mdb120; (h) segmentation results of mdb120; (i) original image of mdb121; (j) segmentation results of mdb121; (k) original image of mdb238; (l) segmentation results of mdb238;

Based on the segmentation results, our method can accurately distinguish breast masses from background tissue, as shown in Fig.16(e) and Fig.16(f). The images in 16(g)– (j) indicate that breast images of benign and malignant architectural distortion types have different segmentation results. Therefore, doctors can more easily distinguish benign and malignant architectural distortion. Combined with some morphological postprocessing tools, the algorithm can also predict the diagnostic result. From Fig.16(k) and Fig.16(l), our algorithm is effectively able to identify calcification points.

To compare our method with other segmentation methods, we randomly chose 100 mammogram images from the Digital Database for Screening Mammography (DDSM) as test images. A visual comparison of the segmentation results is shown in Fig. 17.

Refer to caption
(a)
Refer to caption
(b)
Refer to caption
(c)
Refer to caption
(d)
Refer to caption
(e)
Refer to caption
(f)
Refer to caption
(g)
Refer to caption
(h)
Figure 17: Comparative breast image segmentation results: (a) original image from the DDSM database (well-defined/circumscribed masses); (b) segmentation results by a professional doctor; (c) segmentation results by CCNN; (d) segmentation results by OTSU; (e) segmentation results by SCM; (f) segmentation results by SPCNN; (g) segmentation results by Dense-Unet; (h) segmentation results by Sk-Unet;

Therefore, our method is able to accurately identify a single type of lesion without training, as shown in Fig.16(a)–Fig.16(f). However, it cannot identify lesions with well-defined/circumscribed masses and calcification, as shown in Fig.16(g)–Fig.16(h). In this situation, we believe that applying some morphological tools could obtain better results.

We adopted Otsu, SPCNN, SCM, Dense-Unet and Selective Kernel U-Net as comparative algorithms to evaluate the effectiveness of our method. The Otsu method[34] is an effective and adaptive thresholding scheme that has been widely applied for image segmentation. SCM and SPCNN are two state-of-art networks of visual cortex neural network model, which has superior performance in application of image segmentation[35, 19]. Dense-Unet and Selective Kernel U-Net are deep learning method for biomedical image segmentation, which can provide good segmentation performance in medical image[36, 37].

We compared the algorithm segmentation results with the manual segmentation results of professional doctors and calculated the area overlap (OV) of the above algorithms. A visual comparison of the segmentation results is shown in Fig. 4. To quantitatively compare the above algorithms, we utilized five prevalent evaluation metrics, i.e., average pixel intensities (AVE), variance (VAR), area overlap (OV), sensitivity (SEN) and runtime (T). The results are shown in Tab.4.

Algorithms AVE VAR OV(%) SEN Time(s)
Otsu 0.4072 0.0103 0.0618 1 0.0894
SPCNN 0.4073 0.0091 0.0666 1 0.0768
SCM 0.5487 0.0140 0.2920 0.9711 0.0686
CCNN 0.9794 0.0058 0.8119 0.6819 0.1540
Dense-Unet 0.9990 0.0075 0.6849 0.9602 1.9470
SK-Unet 0.9987 0.0184 0.7384 0.9225 7.8167
Table 4: comparative results

In this comparative experiment, the deep learning network did not achieve better results, which may be caused by the small sample size. Compared to all method, although our method requires a longer runtime than other unsupervised image segmentation method, its ability to accurately segment lesions is much better than that of the other methods. Therefore, the proposed segmentation method performs competitively compared to other prevalent algorithms.

6 Conclusion and future work

In this work, we proposed a novel model of the primary visual cortex. Compared with the previous best visual cortex neuron model, our CCNN model generates more types of spike trains under different stimuli and exhibits frequency doubling characteristics similar to those of real neurons. We compared the results of spike train data from CNN with those from real primary visual cortex neurons, and the results verified that our model is more similar to the real neurons.

Because the results of existing medical imaging diagnoses rely on the subjective judgments of doctors and because the CCNN model is more similar to the visual cortex, it is natural to utilize our model to process medical images. In this work, we applied the CCNN model to a mammogram image segmentation task. The comparative results demonstrated that our method is superior at accurate lesion segmentation.

The CCNN outputs different spike train patterns, which can be used as a type of ”error” that can be backpropagated by algorithms, allowing CCNN to be used in deep learning networks.

Thus, the advantage of our approach is that it further helps neurophysiologists understand how the primary visual cortex works and quantitatively predicts the temporal-spatial behavior of real neural networks. It also helps empower engineers to build more applications using visual neural networks and may inspire the next generation of deep learning networks for artificial intelligence purposes.

In the future, some interesting work remains to be explored.

1. As is well known, visual neurons are sensitive to direction; however, it is unknown whether the CCNN model exhibits this behavior. If so, the mechanism of the neuron network should be explored.

2. Brain waves are rhythmic oscillation patterns that can be registered as macroscopic oscillations utilizing EEG sensors on the scalp. Different rhythms are related to different brain activities: delta rhythmic are related to sleep states; theta might be entrance to further understanding learning and memory; alpha is usually related to attention, lucid thinking and integration; beta is present during the state of alert and problem solving; and gamma rhythms modulate perception and consciousness. It will be an interesting work to study whether our model can exhibit similar behavior as EEG signals. According to Baravalle et al, realized and imagined activities in the brain is different[38]. If our model can generate similar output sequence of EEG signal, perhaps we can distinguish between realized and imagined activities in the brain by studying the different sequences generated by the model. If so, the resulting works will help neurophysiologists further understand how the brain works.

3. To use the CCNN for artificial intelligence purposes, designing the “error” that can be backpropagated by algorithms is an important task.

4. The hardware circuits corresponding to the CCNN should be explored.

5. The image segmentation algorithm proposed in the section V can only segment a single target object. The image segmentation algorithm of multi-target objects based on this model will be a meaningful work.

Acknowledgments

The data of Fig. 2 were collected in the Laboratory of Dario Ringach at UCLA and downloaded from the CRCNS web site. https://crcns.org/data-sets/vc/pvc-1/conditions.

This study is supported by the Fundamental Research Funds for the Central Universities (No. 19lgpy230).

Conflict of Interest: The authors declare that they have no conflicts of interest.

References

  • [1] K. Roy, A. Jaiswal, P. Panda, Towards spike-based machine intelligence with neuromorphic computing, Nature 575 (7784) (2019) 607–617.
  • [2] W. Maass, Networks of spiking neurons: the third generation of neural network models, Neural networks 10 (9) (1997) 1659–1671.
  • [3] Z. Yang, J. Lian, Y. Guo, S. Li, D. Wang, W. Sun, Y. Ma, An overview of pcnn model’s development and its application in image processing, Archives of Computational Methods in Engineering 26 (2) (2019) 491–505.
  • [4] J. Lian, Z. Yang, W. Sun, Y. Guo, L. Zheng, J. Li, B. Shi, Y. Ma, An image segmentation method of a modified spcnn based on human visual system in medical images, Neurocomputing 333 (2019) 292–306.
  • [5] W. Xie, Y. Li, Y. Ma, Pcnn-based level set method of automatic mammographic image segmentation, Optik-International Journal for Light and Electron Optics 127 (4) (2016) 1644–1650.
  • [6] C. Shen, D. Wang, S. Tang, H. Cao, J. Liu, Hybrid image noise reduction algorithm based on genetic ant colony and pcnn, The Visual Computer 33 (11) (2017) 1373–1384.
  • [7] J. Wang, Q. Li, Z. Jia, N. Kasabov, J. Yang, A novel multi-focus image fusion method using pcnn in nonsubsampled contourlet transform domain, Optik 126 (20) (2015) 2508–2511.
  • [8] J. Lian, Z. Yang, J. Liu, W. Sun, L. Zheng, X. Du, Z. Yi, B. Shi, Y. Ma, An overview of image segmentation based on pulse-coupled neural network, Archives of Computational Methods in Engineering (2019) 1–17.
  • [9] Z. Yang, J. Lian, S. Li, Y. Guo, Y. Ma, A study of sine–cosine oscillation heterogeneous pcnn for image quantization, Soft Computing (2019) 1–12.
  • [10] X. Deng, C. Yan, Y. Ma, Pcnn mechanism and its parameter settings, IEEE transactions on neural networks and learning systems (2019).
  • [11] R. P. Broussard, Physiologically-based vision modeling applications and gradient descent-based parameter adaptation of pulse coupled neural networks, Tech. rep., AIR FORCE INST OF TECH WRIGHT-PATTERSON AFB OH (1997).
  • [12] M. Yin, P. Duan, W. Liu, X. Liang, A novel infrared and visible image fusion algorithm based on shift-invariant dual-tree complex shearlet transform and sparse representation, Neurocomputing 226 (2017) 182–191.
  • [13] X. Li, Y. Ma, X. Feng, Self-adaptive autowave pulse-coupled neural network for shortest-path problem, Neurocomputing 115 (2013) 63–71.
  • [14] C. Gao, D. Zhou, Y. Guo, Automatic iterative algorithm for image segmentation using a modified pulse-coupled neural network, Neurocomputing 119 (2013) 332–338.
  • [15] J. Lian, Y. Ma, Y. Ma, B. Shi, J. Liu, Z. Yang, Y. Guo, Automatic gallbladder and gallstone regions segmentation in ultrasound image, International journal of computer assisted radiology and surgery 12 (4) (2017) 553–568.
  • [16] J. Lian, B. Shi, M. Li, Z. Nan, Y. Ma, An automatic segmentation method of a parameter-adaptive pcnn for medical images, International journal of computer assisted radiology and surgery 12 (9) (2017) 1511–1519.
  • [17] J. M. Kinser, Foveation by a pulse-coupled neural network, IEEE transactions on neural networks 10 (3) (1999) 621–625.
  • [18] K. Zhan, H. Zhang, Y. Ma, New spiking cortical model for invariant texture retrieval and image processing, IEEE Transactions on Neural Networks 20 (12) (2009) 1980–1986.
  • [19] Y. Chen, S.-K. Park, Y. Ma, R. Ala, A new automatic parameter setting method of a simplified pcnn for image segmentation, IEEE transactions on neural networks 22 (6) (2011) 880–892.
  • [20] Z. Yang, J. Lian, S. Li, Y. Guo, Y. Qi, Y. Ma, Heterogeneous spcnn and its application in image segmentation, Neurocomputing 285 (2018) 196–203.
  • [21] R. Siegel, Non-linear dynamical system theory and primary visual cortical processing, Physica D: Nonlinear Phenomena 42 (1-3) (1990) 385–395.
  • [22] K. C. Bittner, A. D. Milstein, C. Grienberger, S. Romani, J. C. Magee, Behavioral time scale synaptic plasticity underlies ca1 place fields, Science 357 (6355) (2017) 1033–1036.
  • [23] M. Osswald, S.-H. Ieng, R. Benosman, G. Indiveri, A spiking neural network model of 3d perception for event-based neuromorphic stereo vision systems, Scientific reports 7 (2017) 40703.
  • [24] D. L. Yamins, J. J. DiCarlo, Using goal-driven deep learning models to understand sensory cortex, Nature neuroscience 19 (3) (2016) 356.
  • [25] K. Novak, A. B. de Camargo, M. Neuwirth, K. Kothbauer, V. E. Amassian, V. Deletis, The refractory period of fast conducting corticospinal tract axons in man and its implications for intraoperative monitoring of motor evoked potentials, Clinical neurophysiology 115 (8) (2004) 1931–1941.
  • [26] R. Eckhorn, R. Bauer, W. Jordan, M. Brosch, W. Kruse, M. Munk, H. Reitboeck, Coherent oscillations: A mechanism of feature linking in the visual cortex?, Biological cybernetics 60 (2) (1988) 121–130.
  • [27] T. Lindblad, J. M. Kinser, T. Lindblad, J. Kinser, Image processing using pulse-coupled neural networks, Springer, 2005.
  • [28] A. Dettner, S. Münzberg, T. Tchumatchenko, Temporal pairwise spike correlations fully capture single-neuron information, Nature communications 7 (2016) 13805.
  • [29] J. Aljadeff, B. J. Lansdell, A. L. Fairhall, D. Kleinfeld, Analysis of neuronal spike trains, deconstructed, Neuron 91 (2) (2016) 221–259.
  • [30] D. L. Ringach, P. J. Mineault, E. Tring, N. D. Olivas, P. Garcia-Junco-Clemente, J. T. Trachtenberg, Spatial clustering of tuning in mouse primary visual cortex, Nature communications 7 (2016) 12270.
  • [31] S. M. Bohte, J. N. Kok, H. La Poutre, Error-backpropagation in temporally encoded networks of spiking neurons, Neurocomputing 48 (1-4) (2002) 17–37.
  • [32] J. Liu, J. Ma, J. Lian, P. Chang, Y. Ma, An approach for the generation of an nth-order chaotic system with hyperbolic sine, Entropy 20 (4) (2018) 230.
  • [33] J. Liu, S. Tang, J. Lian, Y. Ma, X. Zhang, A novel fourth order chaotic system and its algorithm for medical image encryption, Multidimensional Systems and Signal Processing 30 (4) (2019) 1637–1657.
  • [34] H. Du, X. Chen, J. Xi, An improved background segmentation algorithm for fringe projection profilometry based on otsu method, Optics Communications 453 (2019) 124206.
  • [35] K. Zhan, J. Shi, Q. Li, J. Teng, M. Wang, Image segmentation using fast linking scm, in: 2015 International Joint Conference on Neural Networks (IJCNN), IEEE, 2015, pp. 1–8.
  • [36] S. Guan, A. A. Khan, S. Sikdar, P. V. Chitnis, Fully dense unet for 2-d sparse photoacoustic tomography artifact removal, IEEE journal of biomedical and health informatics 24 (2) (2019) 568–576.
  • [37] M. Byra, P. Jarosik, A. Szubert, M. Galperin, H. Ojeda-Fournier, L. Olson, M. O’Boyle, C. Comstock, M. Andre, Breast mass segmentation in ultrasound with selective kernel u-net convolutional neural network, Biomedical Signal Processing and Control 61 (2020) 102027.
  • [38] O. A. Rosso, F. Montani, R. Baravalle, N. Guisande, M. Granado, Characterization of visuomotor/imaginary movements in eeg: An information theory and complex network approach, Frontiers in Physics 7 (2019) 115.