Modelling Human Visual Motion Processing with Trainable Motion Energy Sensing and a Self-attention Network
Abstract
Visual motion processing is essential for humans to perceive and interact with dynamic environments. Despite extensive research in cognitive neuroscience, image-computable models that can extract informative motion flow from natural scenes in a manner consistent with human visual processing have yet to be established. Meanwhile, recent advancements in computer vision (CV), propelled by deep learning, have led to significant progress in optical flow estimation, a task closely related to motion perception. Here we propose an image-computable model of human motion perception by bridging the gap between biological and CV models. Specifically, we introduce a novel two-stages approach that combines trainable motion energy sensing with a recurrent self-attention network for adaptive motion integration and segregation. This model architecture aims to capture the computations in V1-MT, the core structure for motion perception in the biological visual system, while providing the ability to derive informative motion flow for a wide range of stimuli, including complex natural scenes. In silico neurophysiology reveals that our model’s unit responses are similar to mammalian neural recordings regarding motion pooling and speed tuning. The proposed model can also replicate human responses to a range of stimuli examined in past psychophysical studies. The experimental results on the Sintel benchmark demonstrate that our model predicts human responses better than the ground truth, whereas the state-of-the-art CV models show the opposite. Our study provides a computational architecture consistent with human visual motion processing, although the physiological correspondence may not be exact.
1 Introduction
Visual motion perception is essential not only for humans and animals to perceive and interact with the world but also for artificial agents to process various dynamic visual tasks. As such, visual motion estimation has been extensively studied by both biological vision and computer vision research communities. The key issue lies in estimating optical flow, an array of instantaneous image motion vectors [1, 2].
Vision science has revealed that optical flow estimation in the human visual system (HVS) is primarily served by a pathway that includes the primary visual cortex (V1 [3, 4]) and middle temporal (MT [5, 6]) area. A significant portion of neurons in area V1 are sensitive to the direction of local motion, while those in area MT integrate and segregate local motion signals for global flow interpretation. The MT process also helps to overcome the aperture problem[7]. Several computational models of area MT have been proposed [8, 9], but mechanisms that can fully encapsulate a variety of motion integration abilities of MT neurons and/or human perception have yet to be developed [10]. Furthermore, previous studies tested the models using only simple artificial stimuli. It remains challenging to create a human-like optical flow extraction mechanism that can derive informative motion flow for a wide range of stimuli including complex natural videos.
On the other hand, optical flow estimation has recently made remarkable progress in the field of computer vision. FlowNet[11] pioneered the use of fully convolutional neural networks for dense optical flow estimation, and various approaches based on deep neural networks (DNNs) have emerged subsequently[12, 13, 14, 15]. Owing to the powerful representation ability of DNNs, these models outperform humans in estimating the ground-truth optical flow of natural scenes [16]. Consequently, one might expect DNN-based optical flow estimation algorithms to become promising models of human visual motion processing, similar to how ImageNet-trained DNN models provide good computational models of human object recognition[17]. However, since computer vision optical flow models are designed solely to find local image correspondences between pairs of frames on the image coordinates [11], they cannot explain systematic or adaptive deviations of human perceived motion from local ground truth (GT) [16]. Moreover, existing DNN models often exhibit instability when dealing with non-textured stimuli commonly used in vision science [18].
In this study, we leveraged the flexibility of DNNs to construct an image-computational model of human motion processing. While recent studies successfully used DNNs to elaborate the understanding of the neural mechanism of visual motion processing[19, 20, 21, 22, 23], we aimed to make a model that can explain a broader range of physiological and psychophysical phenomena, including those whose neural mechanisms are not yet clear. From an engineering standpoint, we aimed to make a human-aligned optic flow algorithm while maintaining a flow estimation capability comparable to the state-of-the-art (SOTA) CV models. Our model extracts dense informative motion flows for a wide range of inputs in a way consistent with physiologically measured neural responses and psychophysically measured human motion perception. It consisted of two stages. The first stage, which mimicked the function of V1, comprised of neurons with multi-scale spatiotemporal filters to extract local motion energy. Unlike previous models of V1 [24, 8, 25], the filter tunings were learnable to fit natural optic flow computation. The second stage, which mimicked the function of MT recurrently integrated local motion signals, and solved the aperture problem. We constructed an undirected fully connected graph from the map of local motion energy, and used the attention mechanism [26, 27] for adaptive global motion integration and segregation.
We evaluated the performance of our model from several aspects. In in-silico neurophysiology, we found that our model’s neurons exhibited direction and speed tunings similar to those observed in mammalian physiological recordings in V1 and MT [28, 29]. In simulations of psychophysical findings, our model showed good generalization from simple artificial stimuli (e.g., drifting Gabor) to complex naturalistic scenes. Our model produced human-like responses for several conventional motion stimuli and illusions, including global motion pooling and the barber-pole illusion. Furthermore, the mode’s response to natural scenes was closer to that of humans in comparison to other computer vision models. Our two-stages model provides a computational architecture consistent with human visual motion processing, although the physiological correspondence may not be exact. This achievement is not only valuable in terms of neuroscientific understanding of human visual computation but also for the development of human-aligned machine vision that stably recognizes the world as humans do.
2 Molding of two-stages motion perception system

2.1 First-stage: Local Motion Energy Computation
Spatiotemporal separable Gabor filter: Since we aimed at an image-computable model, the input is a sequence of grayscale images for all spatial positions within the image domain and for all times . The goal of the first-stage neuron is to capture local motion energy at a specific spatiotemporal frequency, which is associated with the function of a direction-selective neuron in the V1 cortex. The responses of neurons can be modeled as 3D Gabor filters [30, 31]. Gabor filters are known to be optimal in the sense that they achieve maximal resolution in both the spatiotemporal and associated frequency domains[32]. To save computational complexity, we decomposed 3D spatiotemporal filters into filters separable in space and time. The spatial component is described by 2D Gabor filters , and the temporal component is described by a 1D sinusoidal function with exponential decay. Specifically, given and within the receptive field, the impulse responses of spatial and temporal complex filters are defined as:
(1) |
The parameters in red are designed to be trained to fit the dataset, where and denote the spatiotemporal frequency tunings of the filter with the preferred speed , and determines the preferred moving orientation; and control the shape of the Gabor filter, and controls the degree of attenuation of the temporal impulse response. All parameters are subject to certain numerical constraints, such as being limited to [0,) to avoid redundancy; and are limited to less than 0.25 pixels per frame to avoid spectrum aliasing, etc. The response of a simple direction-selective cell to a video stimuli can be computed via separate convolutions:
(2) |
where can be learned as spontaneous firing rates. Further, local motion energy is captured by a phase-insensitive complex cell in the V1 cortex, which computes the squared summation of the response from a pair of simple V1 cells with approximately orthogonal spatiotemporal receptive fields[24]. We denote the pair of orthogonal (even and odd) simple V1 cells as:
(3) |
where and extract the real and imaginary parts of a complex number; denotes convolution operations. Then, the response of a complex cell is obtained from a combination of the quadrature pair of the simple cells using the motion energy formulation:
(4) |
Multi-scale Wavelet Processing: The convolution kernel of our spatial filter has a fixed size of , which imposes a physical limitation on the receptive field of each unit. To enhance the flexibility of the receptive field size, we employed a multi-scale processing strategy, as shown in Fig. 1 (A). Specifically, we constructed an image pyramid consisting of eight images scaled linearly from to . In total, 256 independent complex cells are deployed across different scales, with the lower scale having a larger receptive field and a preference for a faster motion speed. This approach is computationally efficient for capturing large-scale displacements in engineering applications[33, 34], and enables the representation of different groups of cells sensitive to short- and long-distance motions [35]. The N complex cells capture motion energy on multiple scales. We applied energy normalization to each cell to ensure consistent energy levels:
(5) |
where is the semi-saturation constant of the normalization, and determines the maximum attainable response. We interpret the response, denoted by , as the model equivalent of a post-stimulus time histogram (PSTH), which is a measure of the neuron’s firing rate. Physiologically, such responses could be computed via inhibitory feedback mechanisms[36, 37]. Considering the spatial arrangement of images, it is expected that motion energy responses should exist at each spatial location generated by the same complex cell groups. Bilinear interpolation is used to resize the multi-scale motion energy into the same spatial size = . In the context of DNNs, this is mainly to balance the trade-off between the spatial resolution and computational overhead, and the final output of the first stage is a 256-channel feature map that captures the underlying local motion energy, which partially characterizes the cellular patterns of the V1 cortex in a computational manner[24].

2.2 Second stage: Global Motion Integration and Segregation
The receptive field size of first-stage neurons limits their ability to capture only local motion. Advanced spatial integration is an essential requirement for a motion perception system to solve the aperture problem[7]. Spatial integration could involve several mechanisms, such as object recognition and segmentation[38], depth estimation[39], contour structure inference[40], etc., which rely on substantial prior information that may go beyond the capabilities of classical modeling methods. DNNs, with their massive number of parameters and flexibility, offer a suitable solution. However, the spatial integration of local motion requires more flexible connectivity relations than what can be achieved by general convolutions, which are limited by their local receptive field[41]. To overcome this limitation, we propose a computational model based on the attention mechanism and recurrent processing to model the function of motion integration.
Constructing a Graph Structure: First, to establish more flexible connections between cells, we discard the Euclidean space structure within the image and construct topological spaces with an undirected weighted graph , where is the set of nodes; represents the adjacent matrix based on the graph. Each spatial location is treated as a node, and the feature of each node is derived from the whole set of local motion energies . The connection between any pair of nodes is computed using a specific distance metric, and strong connection relationships are formed between nodes whose local motion energy patterns are similar.
Specifically, given a reshaped feature map , the features at each of its locations are first locally embedded into the graph space, as , where is a group of trainable parameters. The distance between any pair of nodes is calculated by the cosine similarity, which is similar to the self-attention mechanism in current transformer structures[26, 42, 27]. We employ the adjacency matrix to represent the connectivity of the whole topological space, and is a symmetric semi-positive definite matrix defined as:
(6) |
We perform exponential scaling of the connections between graphs using the matrix , given by , where is a learnable scalar restricted to the range (0,10) to avoid overflow. The smaller , the smoother the connections across nodes and vice versa. Finally, a symmetric normalization operation is utilized to balance the energy, given by , where is the degree matrix with . As such, an energy-normalized undirected graph structure is constructed, as illustrated in the top side of Fig. 2 (B). Intuitively, this adjacency matrix represents the neuron’s affinity or connectivity within the space, with strong and global connections built across neurons with related motion responses.
Recurrent Integration Processing: Recurrent neural networks (RNNs) are often used to simulate neurons in the brain, as they are flexible in modelling temporal dependencies and feedback loops, which are fundamental aspects of neural processing in the brain[43]. We use a recurrent network, rather than multiple feedforward blocks, to simulate the process of local motion signals being gradually integrated into MT and eventually converging to a stable state.
As shown in Fig. 2 (A), the local motion energy from the first stage is divided into two recurrent streamlines. One is the motion energy , which is continuously updated in the loop, while the other is embedded in the attention space to generate the graph adjacency matrix to control motion integration, denoted as . In each iteration, the adjacency matrix is first constructed using . Subsequent motion integration is achieved through a simple matrix multiplication, which is computationally similar to the information propagation mechanism in transformers[26, 42] and can also be considered a simplified version of graph convolution[44]. The integrated motion information is passed through two independent Conv-GRU blocks to update the motion energy and feature , respectively. The Conv-GRU represents a gated recurrent unit[45] implemented in a convolutional manner, and we adopt a spatio-temporal separable approach following RAFT[14]. The motion integration process approximates the ideal final convergence state of the motion energy through recurrent iteration.
Decoding the Motion Flow: We adopt the same strategy of decoding the 2D optical flow in each iteration. Initially, the integrated motion is projected back to the energy space with positive value using a square operation, followed by an energy normalization operation: The resulting response is interpreted as a post-stimulus time histogram and the motion energy is constrained to the same energy space as the local motion energy from stage I, as we further design an identical flow decoder to project the energy of each spatial location into the optical flow. The structure of the flow decoder is a nonlinear mapping process consisting of multiple convolution blocks with residual connections, which are referred to in several current advanced optical flow estimation models[33, 46]. Additionally, a convex upsampling strategy[14] is employed to restore the optical flow’s original resolution. The entire architecture of stage II is illustrated in Fig. 2 (A, B).
2.3 Training Strategy
Current methods for estimating optical flow using deep neural networks (DNNs) can be categorized as unsupervised/self-supervised and supervised learning approaches. While unsupervised learning methods are intuitively similar to creatures’ interaction with the world, most current methods based on differentiable image warping[47, 48, 49] still try to approximate physical motion ground truth. Therefore, we adopt a supervised learning approach in this work, which is more straightforward as recent research suggests that human perception of motion is reasonably similar to physical GT[16]. However, our primary goal lies in evaluating how well the model approximates human motion perception rather than its accuracy in predicting GT. To train and evaluate the model, we construct a dataset containing various natural and artificial motion scenes. Specifically, we incorporate the Sintel benchmark[50], the DAVIS[51] dataset with pseudo-labels generated by FlowFormer[52], as well as self-created multi-frame datasets with non-texture motions and drifting grating motion. Including simple motion stimuli and drifting gratings allows the model to generalize under different non-texture conditions while providing a potential slow-world Bayesian prior[9]. The model is first pre-trained with simple motion and subsequently fine-tuned on complex natural scenes to facilitate convergence[12]. More specific training details can be found in the supplementary materials.

3 Analyses
Fig. 3 shows the distribution of trained parameters in the first stage: (A) presents the distribution of velocity and orientation of the units; (B) displays the spatiotemporal frequency tuning of the units; (C) demonstrates that the complex cells in the first stage are capable of handling single-frequency component motions such as drifting gratings. The design of the trainable motion energy unit allows for optimal fitting to the flow statistics of the dataset. Although the distribution of the trained parameters appears to lack specific characteristics other than uniformity, it does reflect the effect of training. To validate the effectiveness of training the motion energy units, we conducted experiments with a fixed tuning parameter design using a uniform distribution sampled at equal intervals in terms of spatiotemporal frequency and orientation. The results showed that stage I without the fitting function significantly degraded the model’s ability to estimate motion. (See Our-fixed in Table 1 for details).
The following three parts are conducted for analysis: 1) In silico neurophysiological test of the activation pattern of the units; 2) Psychophysical stimulus tests comparing human perception and model response; 3) Natural scene test of the generalizability of the model in complicated scenarios.
3.1 In silico Neurophysiological Study
Directional Tuning: Some V1 and MT neurons respond selectively to a specific range of motion directions. To investigate the directional tuning of the units in our model, we utilized in silico neurophysiology to measure the activation patterns of 256 units in response to drifting Gabor and plaid stimuli. A plaid consists of two superimposed drifting Gabors, as illustrated in Fig. 4 (C). Analysis of the data revealed three distinct groups of units based on their partial correlation to Gabor and plaid stimuli: 1) Component cells, which always respond to the direction of a Gabor component; 2) Pattern cells, which respond to the integrated (coherent) motion direction of plaid; and 3) Unclassified cells, which do not show a clear preference for either component or pattern motions, as illustrated in Fig. 4 (C).

The distribution of these cell types is not uniform, with component cells being more commonly found in the primary visual cortex (V1), while pattern cells are more often observed in the MT and MST regions[29]. In agreement with this, as illustrated in Fig. 4 (A) (B), our model demonstrates that in the first stage, most units tend to be component cells, whereas the number of pattern and unclassified cells increases in the second stage. In addition, we employed the maximizing activation method [56] to reversely render their cellular preferences, showing that the second-stage unclassified units respond to more complicated motion patterns consisting of both central and background motions, as presented at the bottom of Fig. 4 (B). This suggests that the classical classification of motion neurons into component and pattern cells might be insufficient to characterize the motion integration properties of these neurons.
Spectral Receptive Field and Speed Tuning: We test the spectral (spatial-frequency-vs-temporal-frequency) receptive field of the model’s units using a combination of drifting Gabors with different spatiotemporal frequencies. Two-dimensional oriented Gaussian contours are used to depict the receptive field of the 256 cells, fitted by minimizing the least square error. The results are shown in Fig 4 (E) . Visually, the distribution of the receptive field tilt angle spreads from horizontal/vertical directions to oblique directions (Fig. 4, E). This indicates that the units in the stage II have significant speed tuning compared to the stage I. Speed tuning is a characteristic of higher-order visual motion neurons[28] and is often found in the MT area[57]. This tendency can be seen from the distribution of the partial correlation between the actual receptive field and its speed prediction/independent prediction[55], as demonstrated in Fig. 4 (F). Our two-stages process shows a degree of consistency with the change in mammalian neural distribution from the V1 to MT area.
3.2 Psychophysical Analysis

Fourier motion: In the "missing fundamental illusion"[60], as shown in Fig. 5 (C), when the first spatial harmonic is removed from a square-wave grating with a quarter-cycle shift, the perceived motion direction appears reversed. Our model, whose first stage estimates motion from the Fourier components[61], can predict this reversal. In contrast, computer vision models designed to infer optical flow based on structural correspondence do not exhibit this bias.
Motion Integration: The direction of 1D motion stimuli, such as drifting Gabors, is ambiguous due to the aperture problem. When presented alone, they are perceived to move in orthogonal directions of stripes. When superimposed with other 1D motion components in different orientations, 2D directions consistent with both components are perceived. Pattern motion neurons in MT may contribute to this phenomenon. The second stage of our model can explain this as well, as shown in Fig. 5. Psychophysics also demonstrates motion integration across space. Fig. 5 (A) shows a psychophysical stimulus consisting of drifting Gabors [58]. These Gabors have a variety of local directions and speeds, yet all of them are consistent with one global 2D motion (downward in this case). When viewed as a whole, humans do perceive coherent downward motion. Stage II of our model predicts both the perceived direction and speed of the global Gabor motion.
Fig. 5 (B) compares spatial motion integration between 1D Gabor motion (left) and 2D plaid motion (right). Humans are able to perceive global downward motion only in the former case: The heat maps depict how units with high activity establish long-distance connections to resolve the aperture problem when subjected to Gabor (ambiguous motion) stimuli. In contrast, the plaid stimuli (determined motion) suppress these long-distance connections. In the latter case, local integration of motion signals takes priority over global integration. Once the local ambiguity is resolved, the global integration process is suppressed. Our model can predict such adaptive motion pooling in human visual processing [58]. Furthermore, the barber pole illusion demonstrates how locally ambiguous 1D motion is affected by the shape of the moving area [62]. Specifically, as the height-width ratio of the visual area varies, human perception of direction shifts from oblique to vertical [59] (Fig. 5, D). Our model can predict the shift in perceived direction in stage II, showing its ability to integrate motion signals with boundary orientations. For more video demonstrations, such as the reverse phi illusion[63], please see the supplementary material.
Comparison of Complex Natural Scenes: The proposed model can effectively handle natural scenes, as demonstrated in Fig. 6 (A). Natural stimuli contain diverse spatiotemporal frequency components, leading to complex activation patterns in Stage I. From the decoded flow field, the motion of stage I is limited to localized areas. For example, local motion cannot be found in untextured road areas. This situation necessitates long-range interactions with the surrounding spatial context, which our recurrent integration process in stage II effectively accomplishes. It is evident from the affinity heat map (right side of subfigure A) that object and background areas are clearly segregated in stage II. This suggests that the integration mechanism based on attention has the potential to combine motion integration and object segmentation into a single framework. These two processes are considered highly relevant in the human visual system [64].
We used psychophysically measured optical flow from the Sintel dataset[16] as a benchmark for naturalistic scene flow perceived by humans. Our model was compared to several optical flow estimation methods used in computer vision, including classical algorithms like Farneback, as well as SOTA DNN-based models. We evaluated officially released DNN models that utilize a wide range of inference structures, such as multi-scale inference[11, 12], spatial recurrent models[14], graph reasoning[65], and vision transformers[15].
As shown in Table 1, we computed both the Pearson correlation coefficients and vector endpoint errors (EPEs) between the model prediction and human response, or ground truth (GT). Additionally, we examined the partial correlation between humans and models while controlling the impact of GT:
(7) |
where is the Pearson correlation. This measure is critical in validating the models’ ability to capture the characteristics of the human response, as any model could appear to have a high correlation with the human response by just approximating the GT due to a high correlation between the human response and GT[16].
Quantitatively, the proposed model outperformed all compared models in terms of partial correlation. The RAFT-val in the table is the RAFT framework trained on our dataset as a validation, and indicates that our mixed training set also improves the explanatory power of the human response. Fig. 6 (C) shows that our model significantly improves the partial correlation across iterations in Stage II, indicating that the proposed recurrent motion integration architecture can generate more human-like deviations from GT, which is a trend not present in a similar recurrent network, RAFT. In Fig. 6 (D), one can directly see that our model prediction is more similar to human-perceived flow than the GT flow.

Method | v.s. Human | v.s. GT | |||||||||
Farneback[66] | 0.27 | 0.23 | 0.11 | 0.41 | 0.91 | 0.34 | 2.02 | 0.34 | 0.33 | 0.92 | 1.96 |
FlowNet2.0[11] | 0.39 | 0.26 | 0.34 | 0.92 | 0.90 | 0.96 | 0.94 | 0.95 | 0.94 | 0.98 | 0.47 |
RAFT[14] | 0.20 | 0.22 | 0.14 | 0.92 | 0.90 | 0.96 | 0.93 | 0.98 | 0.99 | 0.99 | 0.25 |
RAFT-val | 0.43 | 0.17 | 0.42 | 0.92 | 0.89 | 0.96 | 1.01 | 0.92 | 0.89 | 0.98 | 0.69 |
AGFlow[65] | 0.30 | 0.16 | 0.20 | 0.93 | 0.90 | 0.96 | 0.92 | 0.98 | 0.98 | 0.98 | 0.27 |
GMFlow[15] | 0.34 | 0.32 | 0.17 | 0.91 | 0.84 | 0.96 | 1.03 | 0.93 | 0.90 | 0.97 | 0.73 |
FlowFormer[52] | 0.36 | 0.14 | 0.32 | 0.93 | 0.91 | 0.95 | 0.90 | 0.98 | 0.97 | 0.98 | 0.42 |
FFV1MT[67] | 0.31 | 0.16 | 0.31 | 0.83 | 0.64 | 0.92 | 1.48 | 0.59 | 0.84 | 0.94 | 1.29 |
3DCNN | 0.27 | 0.29 | 0.42 | 0.83 | 0.86 | 0.95 | 1.31 | 0.83 | 0.86 | 0.96 | 1.14 |
DorsalNet[23] | 0.17 | 0.19 | -0.10 | 0.20 | -0.08 | 0.86 | 2.35 | 0.20 | -0.04 | 0.86 | 2.33 |
Ours-fixed | -0.02 | 0.12 | 0.16 | 0.31 | 0.23 | 0.78 | 2.24 | 0.35 | 0.18 | 0.80 | 2.29 |
Ours-Stage I | 0.34 | 0.23 | 0.35 | 0.71 | 0.71 | 0.92 | 1.52 | 0.67 | 0.67 | 0.92 | 1.49 |
Ours-Stage II | 0.57 | 0.43 | 0.47 | 0.91 | 0.88 | 0.95 | 0.98 | 0.86 | 0.87 | 0.95 | 1.04 |
Table 1 also shows the results of three biologically-inspired models. FFV1MT[67] is a model capable of computing dense optical flow using direct decoding of the Simoncelli & Heeger V1-MT mechanism. The other two models are modified versions of MotionNet[20] and DorsalNet[23], recently proposed DNN-based models for the explanation of neural responses to visual motion stimuli. Since the original models are designed to recognize global motion only, we tested general multi-layer 3D CNN (a core component of MotionNet and DorsalNet) with residual connections, and a pre-trained DorsalNet with frozen parameters, with a linear flow decoder to compute dense flow, trained on natural dense optical flow datasets. Our model outperforms these biologically-inspired models in predicting human responses. The low performances of FFV1MT model and the modified DorsalNet also suggest that accurately estimating dense optical flow is a challenging task, requiring specific design considerations to address complex and long-range spatial interactions, large jumps, and boundary effects, the complexities of which are not adequately captured by simple mechanisms.
4 Discussion and Conclusion
DNNs have achieved impressive performance in various vision tasks, and their ability to explain the HVS is an active area of research. Recent studies have employed DNNs to model and understand the neural mechanism of visual motion. For instance, Rideaux et al.[19, 20] and Nakamura and Gomi [21] used multilayer feedforward networks, while Storrs et al. [22] used a predictive coding network (PredNet), to model biological visual motion processing, and found similarities to neurophysiological data. De Jong et al.[68] found that the spatiotemporal frequency tuning properties of some units in FlowNet resemble those found in mammalian neurons. DorsalNet[23] uses first-person perspective video stimuli to train a 3D ResNet model to predict self-motion parameters, which helps model recapitulate the neural representation of dorsal visual stream.
With a similar goal in mind, we simplified the biological motion process pipeline and proposed a two-stage architecture that models the complete pathway from images, through neural representations, to the perceptual response. Through end-to-end training using a wide range of datasets, our model generalizes well from simple stimuli to complex natural scenes and partially captures important characteristics of motion-processing neurons, including a change in spatiotemporal tuning from V1 to MT areas. To model the motion integration function, we introduced a novel recurrent process based on the attention mechanism. This process successfully explains a wide range of physiological findings (e.g., a change in the population of component and pattern cells from V1 to MT) and psychophysical findings (e.g., global motion pooling). It also improves the partial correlation with human psychophysical response. The success of the attention mechanism in motion integration could be attributed to its similarity to the human visual grouping mechanism[69] or similar feature grouping/binding that may be accomplished using a top-down attentional selection mechanism[70].
While we show that the attention-based recurrent network is a promising computational model of human visual motion grouping and segmentation, how its complex architecture (where all responses can be influenced by all other responses) is actually implemented in the human brain remains an open question. Another limitation of our current model is that it does not process several important abilities of human motion perception, including non-Fourier (second-order) motion detection [61], and motion integration sensitive to surface layout[71]. Our model does not take into account the benefits of biological processing, such as energy efficiency.
In conclusion, combining classical motion energy and advanced deep learning technology is a promising approach to bridge the gap between human and DNN motion perception systems. Our proposed architecture and recurrent process offer insights into the underlying mechanisms of motion perception and open up new avenues for future research.
Acknowledgement
This work was supported in part by JST, the establishment of university fellowships towards the creation of science technology innovation, Grant Number JPMJFS2123; in part by the JSPS Grants-in-Aid for Scientific Research (KAKENHI), Grant Numbers JP20H00603 and JP20H05957.
References
- [1] J. J. Gibson, “The perception of the visual world.,” 1950.
- [2] L. M. Vaina, S. A. Beardsley, and S. K. Rushton, Optic flow and beyond, vol. 324. Springer Science & Business Media, 2004.
- [3] D. H. Hubel and T. N. Wiesel, “Receptive fields of single neurones in the cat’s striate cortex,” The Journal of physiology, vol. 148, no. 3, p. 574, 1959.
- [4] D. H. Hubel and T. N. Wiesel, “Receptive fields and functional architecture in two nonstriate visual areas (18 and 19) of the cat,” Journal of neurophysiology, vol. 28, no. 2, pp. 229–289, 1965.
- [5] C. D. Salzman, C. M. Murasugi, K. H. Britten, and W. T. Newsome, “Microstimulation in visual area mt: effects on direction discrimination performance,” Journal of Neuroscience, vol. 12, no. 6, pp. 2331–2355, 1992.
- [6] S. Celebrini and W. T. Newsome, “Microstimulation of extrastriate area mst influences performance on a direction discrimination task,” Journal of Neurophysiology, vol. 73, no. 2, pp. 437–448, 1995.
- [7] C. C. Pack and R. T. Born, “Temporal dynamics of a neural solution to the aperture problem in visual area mt of macaque brain,” Nature, vol. 409, no. 6823, pp. 1040–1042, 2001.
- [8] E. P. Simoncelli and D. J. Heeger, “A model of neuronal responses in visual area mt,” Vision research, vol. 38, no. 5, pp. 743–761, 1998.
- [9] Y. Weiss, E. P. Simoncelli, and E. H. Adelson, “Motion illusions as optimal percepts,” Nature neuroscience, vol. 5, no. 6, pp. 598–604, 2002.
- [10] C. Pack and R. Born, “Cortical mechanisms for the integration of visual motion,” 2008.
- [11] A. Dosovitskiy, P. Fischer, E. Ilg, P. Hausser, C. Hazirbas, V. Golkov, P. Van Der Smagt, D. Cremers, and T. Brox, “Flownet: Learning optical flow with convolutional networks,” in Proceedings of the IEEE international conference on computer vision, pp. 2758–2766, 2015.
- [12] E. Ilg, N. Mayer, T. Saikia, M. Keuper, A. Dosovitskiy, and T. Brox, “Flownet 2.0: Evolution of optical flow estimation with deep networks,” in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 2462–2470, 2017.
- [13] D. Sun, X. Yang, M.-Y. Liu, and J. Kautz, “Pwc-net: Cnns for optical flow using pyramid, warping, and cost volume,” in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 8934–8943, 2018.
- [14] Z. Teed and J. Deng, “Raft: Recurrent all-pairs field transforms for optical flow,” in European conference on computer vision, pp. 402–419, Springer, 2020.
- [15] H. Xu, J. Zhang, J. Cai, H. Rezatofighi, and D. Tao, “Gmflow: Learning optical flow via global matching,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 8121–8130, 2022.
- [16] Y.-H. Yang, T. Fukiage, Z. Sun, and S. Nishida, “Psychophysical measurement of perceived motion flow of naturalistic scenes,” iScience in press, 2023.
- [17] D. L. Yamins and J. J. DiCarlo, “Using goal-driven deep learning models to understand sensory cortex,” Nature neuroscience, vol. 19, no. 3, pp. 356–365, 2016.
- [18] Z. Sun, Y.-J. Chen, Y.-H. Yang, and S. Nishida, “Comparative analysis of visual motion perception: Computer vision models versus human vision,” in Conference on Cognitive Computational Neuroscience, (Oxford, UK), August 24-27 2023.
- [19] R. Rideaux and A. E. Welchman, “But still it moves: static image statistics underlie how we see motion,” Journal of Neuroscience, vol. 40, no. 12, pp. 2538–2552, 2020.
- [20] R. Rideaux and A. E. Welchman, “Exploring and explaining properties of motion processing in biological brains using a neural network,” Journal of Vision, vol. 21, no. 2, pp. 11–11, 2021.
- [21] D. Nakamura and H. Gomi, “Decoding self-motion from visual image sequence predicts distinctive features of reflexive motor responses to visual motion,” Neural Networks, vol. 162, pp. 516–530, 2023.
- [22] K. Storrs, O. Kampman, R. Rideaux, G. Maiello, and R. Fleming, “Properties of v1 and mt motion tuning emerge from unsupervised predictive learning,” Journal of Vision, vol. 22, no. 14, pp. 4415–4415, 2022.
- [23] P. Mineault, S. Bakhtiari, B. Richards, and C. Pack, “Your head is there to move you around: Goal-driven models of the primate dorsal pathway,” Advances in Neural Information Processing Systems, vol. 34, pp. 28757–28771, 2021.
- [24] E. H. Adelson and J. R. Bergen, “Spatiotemporal energy models for the perception of motion,” Josa a, vol. 2, no. 2, pp. 284–299, 1985.
- [25] J. A. Perrone, “A visual motion sensor based on the properties of v1 and mt neurons,” Vision research, vol. 44, no. 15, pp. 1733–1755, 2004.
- [26] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, Ł. Kaiser, and I. Polosukhin, “Attention is all you need,” Advances in neural information processing systems, vol. 30, 2017.
- [27] A. Dosovitskiy, L. Beyer, A. Kolesnikov, D. Weissenborn, X. Zhai, T. Unterthiner, M. Dehghani, M. Minderer, G. Heigold, S. Gelly, et al., “An image is worth 16x16 words: Transformers for image recognition at scale,” arXiv preprint arXiv:2010.11929, 2020.
- [28] J. A. Perrone and A. Thiele, “Speed skills: measuring the visual speed analyzing properties of primate mt neurons,” Nature neuroscience, vol. 4, no. 5, pp. 526–532, 2001.
- [29] J. Movshon, E. Adelson, M. Gizzi, and W. T. Newsome, “The analysis of moving visual patterns,” in Frontiers in cognitive neuroscience, MIT Press, 1992.
- [30] J. P. Jones, A. Stepnoski, and L. A. Palmer, “The two-dimensional spectral structure of simple receptive fields in cat striate cortex,” Journal of Neurophysiology, vol. 58, no. 6, pp. 1212–1232, 1987.
- [31] J. P. Jones and L. A. Palmer, “An evaluation of the two-dimensional gabor filter model of simple receptive fields in cat striate cortex,” Journal of neurophysiology, vol. 58, no. 6, pp. 1233–1258, 1987.
- [32] R. N. Bracewell and R. N. Bracewell, The Fourier transform and its applications, vol. 31999. McGraw-Hill New York, 1986.
- [33] D. Sun, X. Yang, M.-Y. Liu, and J. Kautz, “Pwc-net: Cnns for optical flow using pyramid, warping, and cost volume,” in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 8934–8943, 2018.
- [34] J.-Y. Bouguet et al., “Pyramidal implementation of the affine lucas kanade feature tracker description of the algorithm,” Intel corporation, vol. 5, no. 1-10, p. 4, 2001.
- [35] E. Castet and J. Zanker, “Long-range interactions in the spatial integration of motion signals.,” Spatial Vision, vol. 12, no. 3, pp. 287–307, 1999.
- [36] D. J. Heeger, “Modeling simple-cell direction selectivity with normalized, half-squared, linear operators,” Journal of neurophysiology, vol. 70, no. 5, pp. 1885–1898, 1993.
- [37] M. Carandini and D. J. Heeger, “Summation and division by neurons in primate visual cortex,” Science, vol. 264, no. 5163, pp. 1333–1336, 1994.
- [38] S. Gilaie-Dotan, “Visual motion serves but is not under the purview of the dorsal pathway,” Neuropsychologia, vol. 89, pp. 378–392, 2016.
- [39] A. Noest and A. Van Den Berg, “The role of early mechanisms in motion transparency and coherence.,” Spatial Vision, vol. 7, no. 2, pp. 125–147, 1993.
- [40] J. Lorenceau and M. Shiffrar, “The influence of terminators on motion integration across space,” Vision research, vol. 32, no. 2, pp. 263–273, 1992.
- [41] Y. Wei, H. Xiao, H. Shi, Z. Jie, J. Feng, and T. S. Huang, “Revisiting dilated convolution: A simple approach for weakly-and semi-supervised semantic segmentation,” in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 7268–7277, 2018.
- [42] X. Wang, R. Girshick, A. Gupta, and K. He, “Non-local neural networks,” in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 7794–7803, 2018.
- [43] T. Serre, “Deep learning: the good, the bad, and the ugly,” Annual review of vision science, vol. 5, pp. 399–426, 2019.
- [44] T. N. Kipf and M. Welling, “Semi-supervised classification with graph convolutional networks,” arXiv preprint arXiv:1609.02907, 2016.
- [45] J. Chung, C. Gulcehre, K. Cho, and Y. Bengio, “Empirical evaluation of gated recurrent neural networks on sequence modeling,” arXiv preprint arXiv:1412.3555, 2014.
- [46] L. Liu, J. Zhang, R. He, Y. Liu, Y. Wang, Y. Tai, D. Luo, C. Wang, J. Li, and F. Huang, “Learning by analogy: Reliable supervision from transformations for unsupervised optical flow estimation,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 6489–6498, 2020.
- [47] R. Jonschkowski, A. Stone, J. T. Barron, A. Gordon, K. Konolige, and A. Angelova, “What matters in unsupervised optical flow,” in Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part II 16, pp. 557–572, Springer, 2020.
- [48] A. Stone, D. Maurer, A. Ayvaci, A. Angelova, and R. Jonschkowski, “Smurf: Self-teaching multi-frame unsupervised raft with full-image warping,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 3887–3896, 2021.
- [49] S. Meister, J. Hur, and S. Roth, “Unflow: Unsupervised learning of optical flow with a bidirectional census loss,” in Proceedings of the AAAI conference on artificial intelligence, vol. 32, 2018.
- [50] D. J. Butler, J. Wulff, G. B. Stanley, and M. J. Black, “A naturalistic open source movie for optical flow evaluation,” in European Conf. on Computer Vision (ECCV) (A. Fitzgibbon et al. (Eds.), ed.), Part IV, LNCS 7577, pp. 611–625, Springer-Verlag, Oct. 2012.
- [51] F. Perazzi, J. Pont-Tuset, B. McWilliams, L. Van Gool, M. Gross, and A. Sorkine-Hornung, “A benchmark dataset and evaluation methodology for video object segmentation,” in Computer Vision and Pattern Recognition, 2016.
- [52] Z. Huang, X. Shi, C. Zhang, Q. Wang, K. C. Cheung, H. Qin, J. Dai, and H. Li, “Flowformer: A transformer architecture for optical flow,” in Computer Vision–ECCV 2022: 17th European Conference, Tel Aviv, Israel, October 23–27, 2022, Proceedings, Part XVII, pp. 668–685, Springer, 2022.
- [53] N. J. Priebe, S. G. Lisberger, and J. A. Movshon, “Tuning for spatiotemporal frequency and speed in directionally selective neurons of macaque striate cortex,” Journal of Neuroscience, vol. 26, no. 11, pp. 2941–2950, 2006.
- [54] E. LeDue, M. Zou, and N. Crowder, “Spatiotemporal tuning in mouse primary visual cortex,” Neuroscience letters, vol. 528, no. 2, pp. 165–169, 2012.
- [55] N. J. Priebe, C. R. Cassanello, and S. G. Lisberger, “The neural representation of speed in macaque area mt/v5,” Journal of Neuroscience, vol. 23, no. 13, pp. 5650–5661, 2003.
- [56] J. Yosinski, J. Clune, A. Nguyen, T. Fuchs, and H. Lipson, “Understanding neural networks through deep visualization,” arXiv preprint arXiv:1506.06579, 2015.
- [57] N. J. Priebe, C. R. Cassanello, and S. G. Lisberger, “The neural representation of speed in macaque area mt/v5,” Journal of Neuroscience, vol. 23, no. 13, pp. 5650–5661, 2003.
- [58] K. Amano, M. Edwards, D. R. Badcock, and S. Nishida, “Adaptive pooling of visual motion signals by the human visual system revealed with a novel multi-element stimulus,” Journal of vision, vol. 9, no. 3, pp. 4–4, 2009.
- [59] N. Fisher and J. M. Zanker, “The directional tuning of the barber-pole illusion,” Perception, vol. 30, no. 11, pp. 1321–1336, 2001.
- [60] R. O. Brown and S. He, “Visual motion of missing-fundamental patterns: motion energy versus feature correspondence,” Vision Research, vol. 40, no. 16, pp. 2135–2147, 2000.
- [61] C. Chubb and G. Sperling, “Drift-balanced random stimuli: a general basis for studying non-fourier motion perception,” JOSA A, vol. 5, no. 11, pp. 1986–2007, 1988.
- [62] P. Sun, C. Chubb, and G. Sperling, “Two mechanisms that determine the barber-pole illusion,” Vision research, vol. 111, pp. 43–54, 2015.
- [63] S. M. Anstis and B. J. Rogers, “Illusory continuous motion from oscillating positive-negative patterns: Implications for motion perception,” Perception, vol. 15, no. 5, pp. 627–640, 1986.
- [64] T. Handa and A. Mikami, “Neuronal correlates of motion-defined shape perception in primate dorsal and ventral streams,” European journal of Neuroscience, vol. 48, no. 10, pp. 3171–3185, 2018.
- [65] A. Luo, F. Yang, K. Luo, X. Li, H. Fan, and S. Liu, “Learning optical flow with adaptive graph reasoning,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 36, pp. 1890–1898, 2022.
- [66] G. Farnebäck, “Two-frame motion estimation based on polynomial expansion,” in Image Analysis: 13th Scandinavian Conference, SCIA 2003 Halmstad, Sweden, June 29–July 2, 2003 Proceedings 13, pp. 363–370, Springer, 2003.
- [67] F. Solari, M. Chessa, N. K. Medathati, and P. Kornprobst, “What can we expect from a v1-mt feedforward architecture for optical flow estimation?,” Signal Processing: Image Communication, vol. 39, pp. 342–354, 2015.
- [68] D. B. de Jong, F. Paredes-Vallés, and G. C. de Croon, “How do neural networks estimate optical flow? a neuropsychology-inspired study,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 44, no. 11, pp. 8290–8305, 2021.
- [69] P. Mehrani and J. K. Tsotsos, “Self-attention in vision transformers performs perceptual grouping, not attention,” arXiv preprint arXiv:2303.01542, 2023.
- [70] J. K. Tsotsos, Y. Liu, J. C. Martinez-Trujillo, M. Pomplun, E. Simine, and K. Zhou, “Attending to visual motion,” Computer Vision and Image Understanding, vol. 100, no. 1-2, pp. 3–40, 2005.
- [71] J. McDermott, Y. Weiss, and E. H. Adelson, “Beyond junctions: nonlocal form constraints on motion interpretation,” Perception, vol. 30, no. 8, pp. 905–923, 2001.