End-to-end Learning of Object Motion Estimation from Retinal Events
for Event-based Object Tracking
Abstract
Event cameras, which are asynchronous bio-inspired vision sensors, have shown great potential in computer vision and artificial intelligence. However, the application of event cameras to object-level motion estimation or tracking is still in its infancy. The main idea behind this work is to propose a novel deep neural network to learn and regress a parametric object-level motion/transform model for event-based object tracking. To achieve this goal, we propose a synchronous Time-Surface with Linear Time Decay (TSLTD) representation, which effectively encodes the spatio-temporal information of asynchronous retinal events into TSLTD frames with clear motion patterns. We feed the sequence of TSLTD frames to a novel Retinal Motion Regression Network (RMRNet) to perform an end-to-end 5-DoF object motion regression. Our method is compared with state-of-the-art object tracking methods, that are based on conventional cameras or event cameras. The experimental results show the superiority of our method in handling various challenging environments such as fast motion and low illumination conditions.
Introduction
Biological eyes are one of the most efficient and sophisticated neural systems. As a vital part of eyes, the retina can precisely and efficiently capture motion information (?), especially for motions caused by moving objects (?), in natural scenes. Compared with the retina, most state-of-the-art motion estimation approaches are still limited by challenging conditions such as motion blur and high dynamic range (HDR) illuminations. There have been several attempts (e.g., (?)) trying to imitate the retina by using artificial neural networks (ANNs). However, it is difficult for ANNs to imitate the asynchronous nature of the retina. In contrast, event cameras (e.g., DAVIS (?)) are asynchronous visual sensors with very high dynamic range and temporal resolution (120 dB, 1 ms). These bio-inspired sensors help event-based methods to perform better in many computer vision and artificial intelligence tasks. In particular, event cameras can filter out non-motion information from the visual input, under stable illumination conditions or infrequent light variations: thus saving a lot of computation power, and giving clear clues about where object movement occurred.

Several 3D 6-DoF ego-motion estimation methods for event-based input, such as (?; ?; ?), have been proposed during recent years. They have shown the superiority of event cameras on the motion estimation task. However, there are only a few studies devoted to analyzing object-level motion, and most of these studies are designed for some special scenarios (e.g., (?) is for the pedestrian tracking scenario). Moreover, none of these methods are based on the regression methodology, which gives an explicit motion model for retinal events. As shown in Fig. 1, the retinal events, collected between the previous snapshot and the current snapshot, show clear visual patterns about object motions. With this intuition, we present a 5-DoF object-level motion estimation method based on the event camera inputs.
In this study, we propose a novel deep neural network (called Retinal Motion Regression Network, abbreviated as RMRNet) to regress the corresponding 5-DoF object-level motion for moving objects. Here, the 5-DoF object-level motion is a 2D transform between the object bounding box in the previous frame and the estimated object bounding box in the current frame. The proposed RMRNet has a lightweight network structure, and it is end-to-end trainable. In order to leverage the proposed network, we encode asynchronous retinal events into a sequence of synchronous Time-Surface with Linear Time Decay (TSLTD) frames, as shown in the second row of Fig. 1. The TSLTD representation, based on the Time-Surface representation (?), contains clear spatio-temporal motion patterns corresponding to the original retinal events, which is convenient for extracting motion information using RMRNet. Overall, this study makes the following contributions:
-
•
We present the TSLTD representation that is amenable for preserving the spatio-temporal information of retinal events, and for training motion regression networks.
-
•
We introduce a 5-DoF object-level motion model to explicitly regress object motions for visual tracking.
-
•
We propose a Retinal Motion Regression Network (RMRNet) that allows one to end-to-end estimate the 5-DoF object-level motion, directly from TSLTD frames.
We evaluate our method on an event camera dataset and an extreme event dataset. The results demonstrate the superiority of our method when it is compared with several state-of-the-art object tracking methods.

Related Work
Event cameras have achieved great success on a variety of computer vision and artificial intelligence tasks (?). Event-based motion estimation is a popular topic in these tasks. For 2D motion estimation, event-based optical flow can be calculated by using a sliding window variational optimization algorithm (?), or the best point trajectories of the event data (?), by using a self-supervised neural network (?), or a time-slice block-matching method(?). For 3D motion estimation, the event-based 6-DoF motion can be predicted by using a line constraint (?), using interleaved probabilistic filters (?), or using photometric depth maps (?). For event-based visual-inertial data, the 3D motion information can also be estimated by using an extended Kalman filter on image features (?), using a keyframe-based nonlinear optimization algorithm (?) or using a continuous-time representation (?). As described in these works, event cameras have shown a unique and superior ability for the motion estimation task. From their results, we can see that event cameras usually outperform conventional cameras, especially when coping with some harsh conditions such as fast motion and HDR scenes.
Despite the fact that event-based object tracking methods can benefit a lot from the high spatio-temporal resolution and high HDR features of event cameras (when compared with conventional tracking methods e.g., (?; ?; ?; ?; ?; ?)), there are only a few works done in this area. These works can be roughly divided into two categories. The works in the first category need a clustering process to group events into clusters. The works in the second category do not need the clustering process.
In the first category, (?) proposes a method for tracking multiple pedestrians with occlusions. They use a Gaussian mixture model for clustering. Similarly, (?) also rely on a clustering algorithm to track objects with occlusions using a stereo system of two event cameras. In (?), they propose a variant of particle filter to track the cluster centers grouped by a Hough transform algorithm.
In the second category, (?) proposes an object detection and tracking method, which is built on top of the motion compensation concept, and they use the Kalman filter tracker for tracking. (?) also proposes a detection and tracking method for long-term tracking using a local sliding window approach. (?) presents a real-time method for multi-target tracking based on both mean-shift clustering and Kalman tracking.
For the above-mentioned studies, we have found that all of them involve handcrafted strategies. However, in our study, we prefer to learn an explicit 2D object motion model from original retinal events, with minimal human intervention, in an end-to-end manner.
Proposed Method
Our method can directly predict frame-wise 5-DoF in-plane object-level motion (i.e., a 2D transform for bounding boxes) from retinal events. The full pipeline of our method is shown in Fig. 2. The retinal events are initially created by an event camera (we use event channel data from a DAVIS sensor (?), as the input in this work), and the events are converted into a series of synchronous Time-Surface with Linear Time Decay (TSLTD) frames. Then the TSLTD frames are fed to our new Retinal Motion Regression Network (RMRNet) to regress the corresponding 5-DoF in-plane motions. In the remainder of this section, we introduce the proposed method in detail.
Time-Surface with Linear Time Decay
The -th event of retinal events can be represented as a quadruple:
(1) |
where and are the horizontal and vertical coordinates of ; indicates the polarity (On or Off) of , and is the timestamp of . Retinal events can occur independently in an asynchronous manner, which makes it difficult for conventional computer vision algorithms to directly process the raw data. There are several attempts (?; ?; ?) that try to convert asynchronous retinal events to synchronous frames. For example, (?) adopts a hierarchical model, which is beneficial to the recognition task but containing less motion information. (?; ?) use specially designed histograms to format retinal events, which cut off the continuity of motion patterns in the temporal domain. In contrast, we prefer to create a clear and lightweight motion pattern for training our network with the help of event cameras, which allows our method to use a much smaller network structure, and maintain high precision estimation and real-time performance simultaneously.
In this work, we propose a synchronous representation of retinal events named Time-Surface with Linear Time Decay (TSLTD), as shown in Fig. 1. The TSLTD representation is based on the Time-Surface representation in (?), but with two major differences: (1) the Time-Surface representation, which is designed for object recognition, consists of clear object contours with little motion information. We replace the exponential time decay kernel in Time-Surface with a linear time decay kernel to efficiently create effective motion patterns; (2) our TSLTD representation does not need the two hyper-parameters (i.e., the size of neighborhood and the time constant of exponential kernel) in Time-Surface. Thus, TSLTD can be effectively generalized to various target objects for motion estimation.
In TSLTD, motion information, captured by an event camera, is encoded and represented as a set of TSLTD frames over every time window T. The time window T will be discussed later. Each of the TSLTD frames is initialized to a three-dimensional zero matrix . Here and are the height and width of the event camera resolution, the third dimension indicates the polarity of events (and the information from On or Off events will be stored separately). Then asynchronous retinal events within the current time window are used to update the matrix in an ascending order in terms of their timestamps.
More specifically, supposing that we are processing a retinal event set which is collected between the start timestamp and the end timestamp of the current time window, to yield a new TSLTD frame . is initialized to a 3D zero matrix . During the updating process, we process from to , where is the first event and is the last event. Each of the events in triggers an update, which assigns a value to at the coordinates corresponding to the triggering event. The value for the -th update caused by is calculated using the following equation:
(2) |
where is the timestamp of . So the assigned value for each update is proportional to the timestamp of the triggered event. is the linear time decay, and we use to normalize the decay. When a pixel of an object moves to a new coordinate, an event will occur at that coordinate and the TSLTD frame records a higher value of than the previous one at that coordinate. As a result, the time-motion variation information is naturally embedded in the TSLTD frames to form intensity gradients as shown in Fig. 2, which have shown a clear pattern of the direction and magnitude of the corresponding object motion. Therefore, the TSLTD format facilitates our network to extract the motion information through intensity gradient patterns that are embedded in the TSLTD frames.
There are two main problems that are related to the time window T value, used in generating TSLTD frames. The first problem is that if T is set to a large interval (e.g., more than 16 ), there are two consequences for fast moving objects. Since these objects move fast, they can return to the previous position or move far from the previous position, during a large time interval. As a result, new motion patterns overlap and contaminate the previous patterns, or become too large to recognize. On the contrary, if T is set to a small interval (e.g., less than 3 ), TSLTD can only capture a very small movement, which may not be distinguished from sensor noises, and thus it may cause an ambiguous motion pattern (especially for a low-resolution event camera). After testing T with different objects and motions, we experimentally set T to be (according to the sampling frequency of 150 Hz) for good generalization performance.
Network Structure
Nowadays, pre-trained deep models, such as VGGNet (?) and ResNet (?), are very popular among many computer vision tasks. But most of these pre-trained models were trained using three-channel RGB images, which is not the optimal option for two-channel TSLTD frames. In addition, since TSLTD frames have clear motion patterns, we do not need very deep and complex networks to extract very high-level features for general pattern recognition.
Here we design a lightweight network, named RMRNet, to learn object motions directly from TSLTD in an end-to-end manner. Between every two adjacent video frames, there are five TSLTD frames, which contain multiple objects. For individual object motion estimation, we crop object patches from the TSLTD frames. During the training stage, we crop the object patches from adjacent five TSLTD frames according to times of their axis-aligned bounding boxes, to preform a joint training. Here is a parameter that renders the cropped region slightly larger than the previous bounding box to capture the full pattern of object motion. During the test stage, we crop object patches frame-by-frame according to times of their axis-aligned bounding boxes. Finally, these object patches are resized to 64 64, and sent to the proposed RMRNet as the input.
As shown in Fig. 2, the first part of RMRNet contains four convolutional layers for feature extraction. The initial three layers share the same kernel size of 3 3 with stride of 2, which is similar to the VGG-Network (?). The kernel size of the final layer, which is used to reduce the feature dimensions, is 1 1 with stride of 1. The filter numbers of the four layers are 32, 64, 128 and 32, respectively. The four convolutional layers are followed by a batch normalization layer. A dropout layer is added in the end during the training stage. Finally, the output feature is flattened and sent to the next part. The second part of RMRNet is an LSTM module. This module contains three layers with 1568 channels in each layer. By adding the LSTM module, we can stack object patches in one regression process. Then the LSTM module can fuse the stacked motion features from the CNN part to regress an accumulated motion, which is the motion between the first object patch and the final object patch. The final part of RMRNet is a set of fully connected layers, which are used to predict 5-DoF motions. The first fully connected layer has 1568 channels. Then the following layers are divided into five branches for different components of the 5-DoF motion. Each branch contains two layers, which respectively have 512 and 128 channels. This network structure is chosen due to its desirable performance on balancing both precision and speed. The output of RMRNet is a 5-DoF transform ( to ), as described next.
Dataset | Sequence names | Feature | Challenges |
---|---|---|---|
ECD | shapes_translation | B&W shape objects mainly with translations | FM |
ECD | shapes_6dof | B&W shape objects with various 6-DoF motions | FM |
ECD | poster_6dof | Natural textures with cluttered background and various 6-DoF motions | FM+BC |
ECD | slider_depth | Various artifacts at different depths with only translations | BC |
EED | fast_drone | A fast moving drone under a very low illumination condition | FM+SO+HDR |
EED | light_variations | Same with the upper one with extra periodical abrupt flash lights | FM+SO+HDR |
EED | what_is_background | A thrown ball with a dense net as foreground | FM+OC |
EED | occlusions | A thrown ball with a short occlusion under a dark environment | FM+OC+HDR |
A 5-DoF transform between a frame and the next frame for an object can be defined as a subset of the 2D affine transform on an object bounding box in the -th frame:
(3) |
where is represented as a quintet, and are respectively the horizontal and vertical displacement factors, and are respectively the horizontal and vertical scale factors, and is the rotation factor. Note that the rotation factor and the scaling factors and in are “in-place operations”, which means that we will keep center alignment before and after these two operations. The resulting coordinate transform is as follows:
(4) |
Here the original coordinates of the bounding box of object in previous frame are transformed into the new coordinates in current frame through the transform . The operator indicates an in-place operation. Note that the five parameters to predicted by RMRNet are normalized to a range of using the Tanh activation function. Then we set five boundary parameters to for to to constrain the range of object movements. Finally, the transform is calculated as follows:
(5) |
In this paper, we respectively fix to to 72, 54, 30, 0.2 and 0.2, and fix to 1.2 according to the 240 180 resolution of the DAVIS camera for RMRNet. This parameter setting will allow a relatively large range for object movements. Thus, the setting is suitable for estimating the majority of object motions, which includes most of the fast movements.
Learning Approach
There are two key points that should be mentioned in relation to the training stage. The first one is that if we only use B&W object samples (i.e., black objects with a white background) as the training data, our network can only learn some relatively simple motion patterns for the corresponding object motions. Thus, we follow the standard five-fold cross-validation protocol and use the object pairs (refer to the next section) from the shapes_6dof, poster_6dof and light_variations sequences as the training and validation data (while the other five sequences are unseen during the training stage) to train and validate the proposed RMRNet. The second point is that it is difficult to learn an object motion model from a single TSLTD frame. There are five TSLTD frames for every object pair. A single TSLTD frame usually includes only a small movement and it has only a weak motion pattern. After extensive experiments, we find that stacking five TSLTD frames in one prediction has gained the optimal performance during the training stage.
The proposed network is trained using the ADAM solver with a mini-batch size of 16. We use a fixed learning rate of 0.0001, and our loss function is the MSE loss:
(6) |
where is the estimated 5-DoF motion, is the corresponding ground truth, is the number of training samples, and the subscript indicates the -th sample.
Experiments
Pre-processing
For our evaluation, we use a challenging mixed event dataset including a part of the popular Event Camera Dataset (ECD) (?) and the Extreme Event Dataset (EED) (?), which were recorded using a DAVIS event camera in real-world scenes. The details of the dataset can be found in Table 1. Note that the mixed dataset contains both the event data sequences and the corresponding video sequences for every sequence. Since the ECD dataset does not provide ground truth bounding boxes for object tracking, we labeled a rotated or an axis-aligned rectangle bounding box as the ground truth for each object in the dataset to evaluate all the competing methods.
We choose five state-of-the-art object tracking methods, including SiamFC (?), ECO (?), SiamRPN++ (?), ATOM (?) and E-MS (?), as our competitors. About these competitors: SiamFC and SiamRPN++ are fast and accurate methods, which are based on Siamese networks. ECO and ATOM are state-of-the-art methods that have achieved great performance on various datasets. E-MS (?) is recently proposed for event-based target tracking based on mean-shift clustering and Kalman filter. We extend E-MS to support bounding box-based tracking, by employing the minimum enclosing rectangle of those events that belong to an identical cluster center as its estimated bounding box.
To perform ablation studies, we also compare our RMRNet with an event-based variant of ECO (called as ECO-E) and a variant of RMRNet (called as RMRNet-TS). ECO-E, which uses our proposed TSLTD frames as its inputs, is an event-based variant of ECO (?). ECO-E is used to evaluate the performance of ECO on the event data sequences. RMRNet-TS uses the classical Time-Surface frames in Hots (?) instead of using the proposed TSLTD frames as its inputs. RMRNet-TS is used to evaluate the performance of the classical Time-Surface representation and our TSLTD representation. For SiamFC, ECO, SiamRPN++ and ATOM, we use the video sequences as their inputs. For ECO-E, E-MS, RMRNet-TS and the proposed RMRNet, we use the event data sequences (in quadruple format) as their inputs. For all the five competitors, we use their released codes, default parameters and best pre-trained models.
Method | shapes_translation | shapes_6dof | poster_6dof | slider_depth | ||||
---|---|---|---|---|---|---|---|---|
AOR | AR | AOR | AR | AOR | AR | AOR | AR | |
SiamFC(?) | 0.812 | 0.940 | 0.835 | 0.968 | 0.830 | 0.956 | 0.909 | 1.000 |
ECO(?) | 0.823 | 0.943 | 0.847 | 0.969 | 0.846 | 0.960 | 0.947 | 1.000 |
SiamRPN++(?) | 0.790 | 0.942 | 0.779 | 0.972 | 0.753 | 0.899 | 0.907 | 1.000 |
ATOM(?) | 0.815 | 0.945 | 0.803 | 0.974 | 0.835 | 0.961 | 0.897 | 1.000 |
ECO-E(?) | 0.821 | 0.941 | 0.834 | 0.960 | 0.783 | 0.878 | 0.771 | 0.993 |
E-MS(?) | 0.675 | 0.768 | 0.612 | 0.668 | 0.417 | 0.373 | 0.447 | 0.350 |
RMRNet-TS | 0.491 | 0.564 | 0.467 | 0.509 | 0.504 | 0.558 | 0.814 | 0.993 |
RMRNet | 0.836 | 0.951 | 0.866 | 0.980 | 0.859 | 0.962 | 0.915 | 1.000 |
Method | fast_drone | light_variations | what_is_background | occlusions | ||||
---|---|---|---|---|---|---|---|---|
AOR | AR | AOR | AR | AOR | AR | AOR | AR | |
SiamFC(?) | 0.766 | 1.000 | 0.772 | 0.947 | 0.712 | 0.833 | 0.148 | 0.000 |
ECO(?) | 0.830 | 1.000 | 0.782 | 0.934 | 0.675 | 0.750 | 0.209 | 0.333 |
SiamRPN++(?) | 0.717 | 0.941 | 0.497 | 0.500 | 0.653 | 0.833 | 0.096 | 0.167 |
ATOM(?) | 0.763 | 1.000 | 0.652 | 0.921 | 0.725 | 0.917 | 0.387 | 0.500 |
ECO-E(?) | 0.728 | 0.882 | 0.685 | 0.803 | 0.099 | 0.000 | 0.308 | 0.333 |
E-MS(?) | 0.313 | 0.307 | 0.325 | 0.321 | 0.362 | 0.360 | 0.356 | 0.353 |
RMRNet-TS | 0.199 | 0.118 | 0.096 | 0.066 | 0.108 | 0.000 | 0.000 | 0.000 |
RMRNet | 0.892 | 1.000 | 0.802 | 0.947 | 0.202 | 0.083 | 0.716 | 0.833 |
The output of our RMRNet is a 2D 5-DoF frame-wise bounding box transform model, estimated between two adjacent frames. In order to evaluate the quality of the estimated frame-wise bounding box transform model, we compare the proposed RMRNet with all the competitors on frame-wise object tracking, that is, our evaluation on these competing methods is based on object pairs, each of which includes two object regions on two adjacent frames corresponding to an identical object. During the evaluation, we treat each of the object pairs as a tracking instance in the corresponding frame. In addition, all the competitors only estimate axis-aligned bounding boxes, whereas the proposed RMRNet can estimate both rotated and axis-aligned bounding boxes (by using the 5-DoF motion model). Therefore, we evaluate all the methods on the axis-aligned ground truth bounding boxes for a fair precision comparison.
Evaluation Metrics
For evaluating the precision of all the methods, we calculate the Average Overlap Rate (AOR) as follow:
(7) |
where is the estimated bounding box in the -th round of the cross-validation for the -th object pair, and is the corresponding ground truth. Eq. (7) shows that the AOR measure is related to the Intersection over Union (IoU). is the repeat times of the cross-validation, and is the number of object pairs in the current sequence. We set to 5 for all the following experiments.
We also calculate the Average Robustness (AR) to measure the robustness of all the competing methods as follow:
(8) |
where indicates that whether the tracking in the -th round for the -th pair is successful or not (0 means failure and 1 means success). If the AOR value obtained by a method for one object pair is under 0.5, we will consider it as a tracking failure case.

Evaluation on the Event Dataset
We use the mixed event dataset to evaluate the eight competing methods. We choose the shapes_translation, shapes_6dof, poster_6dof and slider_depth sequences from the ECD dataset (?) as the representative sequences for comparison. The first three sequences have increasing motion speeds. The fourth sequence has a constant motion speed. The object textures of these sequences vary from simple B&W shapes to complicated artifacts. For these sequences, we are mainly concerned with the performance of all methods for various motions, especially for fast 6-DoF motion, and for different object shapes and textures.
For comparison, the quantitative results are reported in Table 2. We also provide some representative qualitative results obtained by SiamFC, ECO, SiamRPN++, ATOM, E-MS and our method in the top three rows of Fig. 3. From Table 2, we can see that our method achieves the best performance on the first three sequences and it achieves the second best performance on the fourth sequence. SiamFC, ECO, SiamRPN++ and ATOM also achieve competitive results. However, as we can see in Fig. 3, our method has achieved better performance in estimating fast motion. In comparison, SiamFC, ECO, SiamRPN++ and ATOM usually lose the tracked objects, due to the influence of motion blur. Comparing with the original ECO, ECO-E has achieved inferior performance, which shows that state-of-the-art object tracking methods, like ECO, are not suitable to be directly applied to event data. The classical Time-Surface frames are designed for object recognition and detection, which contain less motion information for motion estimation and object tracking. Thus, RMRNet-TS shows much inferior results in this evaluation. By leveraging the high temporal resolution feature of the event data, E-MS can also effectively handle most fast motions. However, E-MS is less effective to handle complicated object textures and cluttered backgrounds (e.g., for the poster_6dof and slider_depth sequences). In contrast, the proposed RMRNet outperforms E-MS by a large margin, which shows the superiority of our method in handling various object textures and cluttered backgrounds.
Moreover, we also choose the EED dataset to evaluate the eight competing methods. The EED dataset (?) contains four challenging sequences: fast_drone, light_variations, occlusions and what_is_background. The first three sequences respectively record a fast moving drone under low illumination environments. The fourth sequence records a moving ball with a net as foreground. Using this dataset, we want to evaluate the competing methods in low illumination conditions and in occlusion situations.
The quantitative results and some representative qualitative results are shown in Table 3 and Fig. 3, respectively. From the results, we can see that our method achieves the best performance on most sequences except for the what_is_background sequence, on which our method has obtained low AOR and AR. This is because that the foreground net in the what_is_background sequence covers the ball, which destroys the corresponding motion patterns. In contrast, our method has achieved the highest AOR and AR on the occlusions sequence. This is because that in this sequence, the occlusion time is short and only one object pair involves the occlusion. Our method fails at tracking that object pair but it successfully estimates the other object pairs. As the competitors, SiamFC, ECO, SiamRPN++ and ATOM show their low effectiveness with fast motion and low illumination conditions. Moreover, ECO-E and RMRNet-TS are respectively inferior to ECO and RMRNet. Although SiamFC and ECO achieve relatively good results for the first two sequences, the sequences include relatively clean backgrounds, which helps the two methods to achieve the performance. However, if we add a small amount of noise around the objects, the performance of the two methods will be significantly reduced. In contrast, our method can maintain its performance even with the severe sensor noises of an event camera, as shown in the seventh column of Fig. 3. Meanwhile, the performance of E-MS is highly affected by the sensor noises in the HDR environments. Thus, E-MS show unsatisfied results on the EED dataset.
Time Cost
Since the proposed RMRNet is a relatively shallow network, our method has an advantage of relatively high efficiency. The proposed RMRNet is implemented using PyTorch on a PC with an Intel i7 CPU and an NVIDIA GTX 1080 GPU. For the mixed dataset, our method achieves real-time performance and the average computational time for each object pair (including five TSLTD frames) is 38.57 ms.
Conclusion
In this paper, we demonstrate the great potential of event cameras for object tracking under severe conditions such as fast motion and low illumination scenes. By using the event camera, our method only extracts motion related information from the inputs. Then we present the TSLTD representation to encode input retinal events into a sequence of synchronous TSLTD frames. TSLTD can represent the spatio-temporal information with clear motion patterns. Finally, to leverage the motion clues contained in TSLTD, we propose the RMRNet to regress 5-DoF motion information in an end-to-end manner. Extensive experiments demonstrate the superiority of our method over several other state-of-the-art object tracking methods.
Acknowledgments
This work is supported by the National Natural Science Foundation of China under Grants U1605252 and 61872307. Haosheng Chen and David Suter acknowledge funding from ECU that enabled the first author to visit ECU and undertake a small portion of this work.
References
- [Bardow, Davison, and Leutenegger 2016] Bardow, P.; Davison, A. J.; and Leutenegger, S. 2016. Simultaneous optical flow and intensity estimation from an event camera. In CVPR, 884–892.
- [Barranco, Fermuller, and Ros 2018] Barranco, F.; Fermuller, C.; and Ros, E. 2018. Real-time clustering and multi-target tracking using event-based sensors. In IROS, 5764–5769.
- [Bertinetto et al. 2016] Bertinetto, L.; Valmadre, J.; Henriques, J. F.; Vedaldi, A.; and Torr, P. H. 2016. Fully-convolutional siamese networks for object tracking. In ECCV, 850–865.
- [Brandli et al. 2014] Brandli, C.; Berner, R.; Yang, M.; Liu, S.-C.; and Delbruck, T. 2014. A 240 180 130 dB 3 s latency global shutter spatiotemporal vision sensor. IEEE JSSC 49(10):2333–2341.
- [Camuñas-Mesa et al. 2017] Camuñas-Mesa, L. A.; Serrano-Gotarredona, T.; Ieng, S.-H.; Benosman, R.; and Linares-Barranco, B. 2017. Event-driven stereo visual tracking algorithm to solve object occlusion. IEEE TNNLS 29(9):4223–4237.
- [Danelljan et al. 2017] Danelljan, M.; Bhat, G.; Khan, F. S.; Felsberg, M.; et al. 2017. Eco: Efficient convolution operators for tracking. In CVPR, 6638–6646.
- [Danelljan et al. 2019] Danelljan, M.; Bhat, G.; Khan, F. S.; and Felsberg, M. 2019. Atom: Accurate tracking by overlap maximization. In CVPR, 4660–4669.
- [Fan and Xiang 2017] Fan, H., and Xiang, J. 2017. Robust visual tracking via local-global correlation filter. In AAAI, 4025–4031.
- [Gallego et al. 2017] Gallego, G.; Lund, J. E.; Mueggler, E.; Rebecq, H.; Delbruck, T.; and Scaramuzza, D. 2017. Event-based, 6-DOF camera tracking from photometric depth maps. IEEE TPAMI 40(10):2402–2412.
- [Gallego et al. 2019] Gallego, G.; Delbrück, T.; Orchard, G.; Bartolozzi, C.; Taba, B.; Censi, A.; Leutenegger, S.; Davison, A. J.; Conradt, J.; Daniilidis, K.; and Scaramuzza, D. 2019. Event-based vision: A survey. arXiv preprint arXiv:1904.08405.
- [Gallego, Rebecq, and Scaramuzza 2018] Gallego, G.; Rebecq, H.; and Scaramuzza, D. 2018. A unifying contrast maximization framework for event cameras, with applications to motion, depth, and optical flow estimation. In CVPR, 3867–3876.
- [Glover and Bartolozzi 2017] Glover, A., and Bartolozzi, C. 2017. Robust visual tracking with a freely-moving event camera. In IROS, 3769–3776.
- [He et al. 2016] He, K.; Zhang, X.; Ren, S.; and Sun, J. 2016. Deep residual learning for image recognition. In CVPR, 770–778.
- [Huang and Zhou 2019] Huang, J., and Zhou, W. 2019. Re2ema: Regularized and reinitialized exponential moving average for target model update in object tracking. In AAAI, 8457–8464.
- [Kim, Leutenegger, and Davison 2016] Kim, H.; Leutenegger, S.; and Davison, A. J. 2016. Real-time 3D reconstruction and 6-DoF tracking with an event camera. In ECCV, 349–364.
- [Lagorce et al. 2017] Lagorce, X.; Orchard, G.; Galluppi, F.; Shi, B. E.; and Benosman, R. B. 2017. Hots: A hierarchy of event-based time-surfaces for pattern recognition. IEEE TPAMI 39(7):1346–1359.
- [Lan et al. 2018] Lan, X.; Ye, M.; Zhang, S.; and Yuen, P. C. 2018. Robust collaborative discriminative learning for rgb-infrared tracking. In AAAI, 7008–7015.
- [Li et al. 2019a] Li, B.; Wu, W.; Wang, Q.; Zhang, F.; Xing, J.; and Yan, J. 2019a. Siamrpn++: Evolution of siamese visual tracking with very deep networks. In CVPR, 4282–4291.
- [Li et al. 2019b] Li, Y.; Zhu, J.; Hoi, S. C.; Song, W.; Wang, Z.; and Liu, H. 2019b. Robust estimation of similarity transformation for visual object tracking. In AAAI, 8666–8673.
- [Liu and Delbruck 2018] Liu, M., and Delbruck, T. 2018. Adaptive time-slice block-matching optical flow algorithm for dynamic vision sensors. In BMVC.
- [Maqueda et al. 2018] Maqueda, A. I.; Loquercio, A.; Gallego, G.; Garcıa, N.; and Scaramuzza, D. 2018. Event-based vision meets deep learning on steering prediction for self-driving cars. In CVPR, 5419–5427.
- [McIntosh et al. 2016] McIntosh, L.; Maheswaranathan, N.; Nayebi, A.; Ganguli, S.; and Baccus, S. 2016. Deep learning models of the retinal response to natural scenes. In NIPS, 1369–1377.
- [Mitrokhin et al. 2018] Mitrokhin, A.; Fermüller, C.; Parameshwara, C.; and Aloimonos, Y. 2018. Event-based moving object detection and tracking. In IROS, 1–9.
- [Mueggler et al. 2017] Mueggler, E.; Rebecq, H.; Gallego, G.; Delbruck, T.; and Scaramuzza, D. 2017. The event-camera dataset and simulator: Event-based data for pose estimation, visual odometry, and SLAM. IJRR 36(2):142–149.
- [Mueggler et al. 2018] Mueggler, E.; Gallego, G.; Rebecq, H.; and Scaramuzza, D. 2018. Continuous-time visual-inertial odometry for event cameras. IEEE TRO 34(6):1425–1440.
- [Mueggler, Huber, and Scaramuzza 2014] Mueggler, E.; Huber, B.; and Scaramuzza, D. 2014. Event-based, 6-DOF pose tracking for high-speed maneuvers. In IROS, 2761–2768.
- [Murphy-Baum and Awatramani 2018] Murphy-Baum, B. L., and Awatramani, G. B. 2018. An old neuron learns new tricks: redefining motion processing in the primate retina. Neuron 97(6):1205–1207.
- [Olveczky, Baccus, and Meister 2003] Olveczky, B. P.; Baccus, S. A.; and Meister, M. 2003. Segregation of object and background motion in the retina. Nature 423(6938):401.
- [Piatkowska et al. 2012] Piatkowska, E.; Belbachir, A. N.; Schraml, S.; and Gelautz, M. 2012. Spatiotemporal multiple persons tracking using dynamic vision sensor. In CVPRW, 35–40.
- [Qi et al. 2019] Qi, Y.; Zhang, S.; Zhang, W.; Su, L.; Huang, Q.; and Yang, M.-H. 2019. Learning attribute-specific representations for visual tracking. In AAAI, 8835–8842.
- [Ramesh et al. 2018] Ramesh, B.; Zhang, S.; Lee, Z. W.; Gao, Z.; Orchard, G.; and Xiang, C. 2018. Long-term object tracking with a moving event camera. In BMVC.
- [Rebecq, Horstschaefer, and Scaramuzza 2017] Rebecq, H.; Horstschaefer, T.; and Scaramuzza, D. 2017. Real-time visualinertial odometry for event cameras using keyframe-based nonlinear optimization. In BMVC.
- [Simonyan and Zisserman 2014] Simonyan, K., and Zisserman, A. 2014. Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556.
- [Sironi et al. 2018] Sironi, A.; Brambilla, M.; Bourdis, N.; Lagorce, X.; and Benosman, R. 2018. Hats: histograms of averaged time surfaces for robust event-based object classification. In CVPR, 1731–1740.
- [Yun et al. 2017] Yun, S.; Choi, J.; Yoo, Y.; Yun, K.; and Young Choi, J. 2017. Action-decision networks for visual tracking with deep reinforcement learning. In CVPR, 2711–2720.
- [Zhu et al. 2019] Zhu, A. Z.; Yuan, L.; Chaney, K.; and Daniilidis, K. 2019. Unsupervised event-based learning of optical flow, depth, and egomotion. In CVPR, 989–997.
- [Zihao Zhu, Atanasov, and Daniilidis 2017] Zihao Zhu, A.; Atanasov, N.; and Daniilidis, K. 2017. Event-based visual inertial odometry. In CVPR, 5391–5399.