This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Indoor Future Person Localization from an Egocentric Wearable Camera

Jianing Qiu1,2, Frank P.-W. Lo2, Xiao Gu1,2, Yingnan Sun2, Shuo Jiang3 and Benny Lo2 1Department of Computing, Imperial College London {jianing.qiu17, xiao.gu17}@imperial.ac.uk2The Hamlyn Centre, Imperial College London {po.lo15, y.sun16, benny.lo}@imperial.ac.uk3College of Electronics and Information Engineering, Tongji University [email protected]
Abstract

Accurate prediction of future person location and movement trajectory from an egocentric wearable camera can benefit a wide range of applications, such as assisting visually impaired people in navigation, and the development of mobility assistance for people with disability. In this work, a new egocentric dataset was constructed using a wearable camera, with 8,250 short clips of a targeted person either walking 1) toward, 2) away, or 3) across the camera wearer in indoor environments, or 4) staying still in the scene, and 13,817 person bounding boxes were manually labelled. Apart from the bounding boxes, the dataset also contains the estimated pose of the targeted person as well as the IMU signal of the wearable camera at each time point. An LSTM-based encoder-decoder framework was designed to predict the future location and movement trajectory of the targeted person in this egocentric setting. Extensive experiments have been conducted on the new dataset, and have shown that the proposed method is able to reliably and better predict future person location and trajectory in egocentric videos captured by the wearable camera compared to three baselines. Our dataset can be downloaded from here

I Introduction

With the recent advances in wearable technologies, wearable cameras are becoming popular and able to capture egocentric (first-person) videos with less motion blur and wider field of view. Wearable cameras have been applied to many human-centric applications, such as gaze prediction to understand human intention and attention [1, 2], human action recognition [3, 4, 5], and also learning environment affordances [6], in egocentric settings. Accurate prediction of future person location and movement trajectory in egocentric videos captured by a wearable camera can help visually impaired people safely navigate, and can also facilitate the development of accurate and personalized mobility assistance (such as autonomous wheelchairs or exoskeleton robots).

Apart from visual images, some wearable cameras are also able to record other sensing signals, such as audio, GPS and IMU, synchronously. Fusing these signals recorded by the wearable cameras could enable better navigation assistance for the user, and may also enable the transfer of assistance developed using wearable cameras to robots. For a social or service robot deployed in an indoor environment, for instance, in a large shopping mall, accurate prediction of the future locations and movement trajectories of people in the vicinity can help the robot better plan its navigating path, avoid potential collisions with people, and if needed, track and follow a targeted person more precisely.

Most works on future person location and trajectory prediction, however, have only been focused on scenes captured by birds-eye or oblique views [7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17], or from vehicle-mounted cameras [18, 19, 20, 21, 22, 23]. Little research has been carried out in egocentric scenarios captured by wearable cameras [24, 25, 26]. In the use of a wearable camera, the motion of the camera wearer plays an important role in predicting the future location and movement trajectory of a targeted person relative to the wearer’s viewpoint, which is very different from scenarios captured by fixed cameras. Although both wearable and vehicle-mounted cameras provide egocentric vision, vehicles are mainly on the road, whereas a camera wearer can be in many different indoor and outdoor places, and the wearer can change directions more flexibly and abruptly compared to a vehicle. Furthermore, the use of deduced information from vehicle-mounted and wearable cameras is different. One is for assisted or autonomous driving, and the other can be used for human mobility assistance. Egocentric vision from a wearable camera reflects how the camera wearer perceives their surroundings, and with synchronized IMU signals, the intent of the camera wearer can be better inferred. For example, whether the wearer is going to change his/her path or to stop to avoid collision. Therefore, the prediction of future location of a targeted person relative to the wearer’s viewpoint can be more accurate. Egocentric vision also provides a close up view of the people in the camera wearer’s vicinity, which offers additional cues for predicting their future locations, such as the targeted person’s pose [25, 23], and person-person interactions. In this work, we aim to predict the future location and trajectory of a targeted person in videos captured by a wearable camera in indoor environments.

The contributions of this work are twofold:

  • A novel egocentric video dataset was constructed for future person location and movement trajectory prediction. The dataset contains over 8k short clips. For each targeted person, a tight bounding box was manually labelled in each frame, resulting in nearly 14k bounding boxes in total. The estimated pose of each targeted person in each frame is also included along with the synchronized 6-axis IMU signals of the wearable camera. Although this dataset was constructed for future location and trajectory prediction, it can also be used to develop and evaluate a tracking system in the egocentric setting, and it also complements the existing FPL  [25] and Citywalks [27] datasets, whose annotations are based on automated detection methods. To the best of our knowledge, this is the first egocentric dataset that provides human annotated person bounding boxes, detected human body poses, and synchronized IMU signals. The dataset is applicable to multiple tasks and will be made available upon request.

  • An LSTM-based encoder-decoder framework was designed for the prediction of the location and movement trajectory of a targeted person, which takes the previous locations and poses of the targeted person, and IMU signals of the camera as the input, and outputs the predicted future location and movement trajectory of the targeted person.

II Related Work

Future person location and movement trajectory prediction have been an active research area, but predictions in most studies have only been carried out in videos captured from birds-eye or oblique views [7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17]. Social-LSTM was proposed in [11] to predict human trajectory, which uses social pooling layers to model the effect of nearby people’s behaviours on a person’s future trajectory. Social-GAN [14] further introduces adversarial learning in future trajectory prediction. In [15], a person’s future trajectory is jointly predicted with his/her future activities using rich visual semantic features, such as scene and objects. Predicting multiple possible future trajectories of people have also been studied in [17]. However, these approaches are not targeted for egocentric scenarios, in which the motion of the camera plays an important role, and the camera view is also largely different from the top-down or oblique views. In the following, we review two egocentric scenarios relevant to or involving future person localization.

Vehicle-mounted cameras. With the use of visual data captured by vehicle-mounted cameras, predicting pedestrian trajectory has been an active research area in autonomous driving [18, 19, 20, 21, 22]. Studies have also been carried out in predicting pedestrian’s intent of crossing or not crossing the street [28, 29], forecasting pedestrian locomotion [23], and detecting traffic accident with pedestrian future location prediction [30]. Although not targeted at pedestrians, [31] and [32] have studied future vehicle localization, as their works are targeted at vehicles, the designed framework and the used cues for future localization differ from ours (e.g., we are able to use human pose as a cue, similar to those in [25, 23]). All works above are part of an assisted or autonomous driving system, whereas our work focuses on using wearable cameras to provide future person localization.

Wearable cameras. Predicting the future location and trajectory of the camera wearer in egocentric stereo images have been studied in [24]. Our work is inspired by [25], which is the first work that predicts future locations of people (not the camera wearer) in egocentric videos from wearable cameras. However, our work differs from them in that 1) we aim to predict a bounding box as the future location of the targeted person instead of predicting a point as the location as in their work. A bounding box can better reflect how much physical space the person occupies, and therefore can enable better navigation. The predicted bounding box can also facilitate tracking if the method is integrated into a tracking system; 2) IMU signals extracted from the wearable camera are used, along with manually labelled human bounding boxes that can represent the scale of a targeted person; 3) we propose an LSTM-based encoder-decoder framework for the future location and movement trajectory prediction in egocentric videos whereas their predictions are based on a 1D convolution-deconvolution framework (conv-deconv). Built on the conv-deconv framework of [25], an adaptive online learning scheme is proposed in [26] to further improve the prediction accuracy. Another relevant work is [27], which proposes a GRU-CNN based framework (STED) and the Citywalks dataset. STED encodes past optical flow and bounding boxes of a pedestrian, and then decodes his/her future bounding boxes. Although we both use bounding boxes, we differ from them in the framework for prediction, and we use human pose and IMU data as extra cues for predicting future person location.

III Method

In this work, we aim to predict the locations of a targeted person in future frames of an egocentric video based on the observation of his/her locations, body poses, and the motion of the camera wearer in the past frames. Formally, denoting each frame as a time step tt, the location of the targeted person in each frame is defined as lt+1×4l_{t}\in\mathbb{R_{+}}^{1\times 4} (i.e., the coordinates of the top-left and bottom-right corners of the person bounding box), the body pose in each frame is defined as pt+K×2p_{t}\in\mathbb{R_{+}}^{K\times 2} where KK is the number of body keypoints, and the motion of the camera wearer in each frame is defined as it1×6i_{t}\in\mathbb{R}^{1\times 6} (i.e., 6-axis IMU signals from a 3-axis accelerometer and a 3-axis gyroscope). Given an observation period TobsvT_{obsv}, a prediction period TpredT_{pred}, and the time step of the current frame t=t0t=t_{0}, our goal is to predict a set of future locations Lpred=(lt0+1,,lt0+Tpred)L_{pred}=(l_{t_{0}+1},...,l_{t_{0}+T_{pred}}) of the targeted person based on his/her past locations Lobsv=(lt0Tobsv+1,,lt0)L_{obsv}=(l_{t_{0}-T_{obsv}+1},...,l_{t_{0}}) and past body poses Pobsv=(pt0Tobsv+1,,pt0)P_{obsv}=(p_{t_{0}-T_{obsv}+1},...,p_{t_{0}}), and the camera wearer’s past motion Iobsv=(it0Tobsv+1,,it0)I_{obsv}=(i_{t_{0}-T_{obsv}+1},...,i_{t_{0}}). To achieve this, a 2-layer LSTM-based encoder-decoder framework is designed, which is shown in Figure 1. LSTM [33] is able to model long-term time evolution, and therefore is suitable for the task of future location and trajectory prediction. The encoder takes xt=lt+it+pt,t(t0Tobsv+1,,t0)x_{t}=l_{t}+i_{t}+p_{t},t\in(t_{0}-T_{obsv}+1,...,t_{0}) as the input per time step of recurrent calculation. Note that ltl_{t}, iti_{t}, and ptp_{t} are first flattened and then concatenated, which makes xt1×(10+2×K)x_{t}\in\mathbb{R}^{1\times(10+2\times K)}. The initial hidden and cell states are both set to 0 (i.e., when xt=xt0Tobsv+1x_{t}=x_{t_{0}-T_{obsv}+1}).

The final hidden and cell states from both layers of the encoder, denoted as context vectors z1z^{1} and z2z^{2}, are then used as the initial hidden and cell states of the respective layers of the decoder. The decoder then decodes the location of the targeted person in each future frame (i.e., t(t0+1,,t0+Tpred)t\in(t_{0}+1,...,t_{0}+T_{pred})). yt=lty_{t}=l_{t} is the true location of the targeted person in the time step t(t0,,t0+Tpred1)t\in(t_{0},...,t_{0}+T_{pred}-1). The hidden state st2,t(t0,,t0+Tpred1)s_{t}^{2},t\in(t_{0},...,t_{0}+T_{pred}-1) from the top layer of the decoder is fed into a fully connected layer to predict the location y^t+1\hat{y}_{t+1} at next future time step. The center of the predicted person bounding box is used to form the movement trajectory of the targeted person in the future frames.

We denote the proposed framework as LIP-LSTM as it takes in the location, IMU, and pose as the input.

Refer to caption

Figure 1: Overview of the LSTM-based encoder-decoder framework for future person location and trajectory prediction. The encoder takes the past locations, body poses of a targeted person, and the past motion of the camera wearer as the input, and encodes them into context vectors z1z^{1} and z2z^{2}. The decoder receives the context vectors, and decodes the locations of the targeted person at future time steps. We use the bounding box of the targeted person to represent his/her location at each time step, and the person’s movement trajectory is formed using the centers of the bounding boxes. Note that in our experiments, the decoder does not receive the true location at t0t_{0}. Therefore, it starts decoding the future location from time step t0+2t_{0}+2.

IV Dataset

TABLE I: The Number of Samples of Different Walking Directions in the Training and Test Sets
Dataset Toward Away Across Still Total
Train 496 5440 320 388 6644
Test 253 1187 65 101 1606

Refer to caption

Figure 2: One targeted person is shown for each walking direction. A targeted person at each time step is manually labelled with a tight bounding box and his/her body pose is estimated by OpenPose. From top to bottom, the walking direction of the targeted person with respect to the camera wearer is: 1) toward, 2) away, 3) across, 4) still. (Best viewed in color).

IV-A Data Collection

A GoPro Hero 7 Black camera was mounted on the chest of the camera wearer to record egocentric video data in different indoor environments. Data were collected from 2 national museums and 1 large shopping mall while the camera wearer was walking around. All video data were recorded using the wide angle at 30fps with the resolution set to 1920×10801920\times 1080. 9 videos were recorded from the shopping mall, and 7 videos from the museums. The total length of the recorded data lasts for 1 h 56 m.

IV-B Data Pre-processing

The recorded egocentric videos were first downsampled to 10 fps and 455×256455\times 256 using ffmpeg111https://ffmpeg.org/. OpenPose [34] was then used to detect body poses in each frame after downsampling. We adopted the setting of detecting 25 body keypoints (i.e., K=25K=25). The 6-axis IMU (3-axis accelerometer and 3-axis gyroscope) signals associated with each downsampled frame were extracted from GoPro meta data using dashware222http://www.dashware.net/. We pre-defined 4 different types of people movements with regards to the camera wearer, among which 3 types are adopted from [25] (i.e., toward, away, and across), and one new type still is introduced (i.e., the targeted person is not moving in the scene). We then manually went through the OpenPose processed frames, identifying those 4 types and labelling tight bounding boxes of the targeted people with VIA [35]. We label each targeted person for at least 20 consecutive frames (2 seconds at 10 fps), in which this person’s body poses are successfully detected by OpenPose. For those longer than 20 frames, a sliding window was applied to generate multiple 2-second samples. In total, there are 8,250 samples obtained, with 13,817 unique bounding boxes. The dataset is then split into training and test sets (training and test samples are drawn from different videos). Table I shows the sample distribution across 4 types of walking direction in each set. Figure 2 shows some targeted person samples. We call this new dataset EILT (Egocentric Indoor future Location and Trajectory prediction dataset).

V Experiment

V-A Implementation Details

We implement our encoder-decoder framework using PyTorch. The hidden dimension of all LSTM layers is 384. Dropout [36] is set to 0.5 for the LSTM layers. The entire network is trained end-to-end with 2\ell_{2} loss, which measures the mean squared error between the predicted and true locations. Adam [37] is used as the optimizer. We set batch size to 64 and train the network for 100 epochs. Teacher forcing is used with a ratio of 0.5 during training. Following [25], we set both TobsvT_{obsv} and TpredT_{pred} to be 10, i.e., a setting of observing 1 second, and predicting for the next second.

V-B Baselines

The following baselines are used to compare the performance of the prediction of a person’s location and movement trajectory with our proposed method on the EILT dataset:

STATS: A statistics-based approach. Given a time step t(t0+1,,t0+10)t\in(t_{0}+1,...,t_{0}+10), we first compute the average location lt1avgl_{t-1}^{avg} of time steps from t09t_{0}-9 up to t1t-1. We then calculate the displacement of the location ltl_{t} with respect to lt1avgl_{t-1}^{avg}, which results in a displacement matrix m10×4m\in\mathbb{R}^{10\times 4}. For each type of walking direction (i.e., toward, away, across, or still), we average n=1Nmn\sum_{n=1}^{N}m_{n} where NN is the total number of samples of a walking direction jj to obtain an average displacement matrix mjavg10×4m_{j}^{avg}\in\mathbb{R}^{10\times 4} of that walking direction. All mjavgm_{j}^{avg} are calculated using the training set. During the test, we obtain the walking direction of the targeted person, calculate his/her average location in the past 10 frames, and use the corresponding average displacement matrix of his/her walking direction to iteratively predict his/her locations in the future 10 frames.

LR: We use linear regression to estimate the locations of the targeted person in the future frames. Concretely, linear regression is applied to a set of top-left corners, and a set of bottom-right corners of the person bounding boxes in the observation period, after which two linear functions are obtained, each for predicting two corners of the person bounding boxes in the prediction period. Assuming that the targeted person is moving at a constant speed and does not change his/her walking direction, we compute the average displacement α\alpha of the xx coordinate of each corner in the observation period. For example, the average displacement of the xx coordinate of the top-left corner is calculated as αtl=(lt0tlxlt09tlx)/9\alpha^{tl}=(l_{t_{0}}^{tl_{x}}-l_{t_{0}-9}^{tl_{x}})/9. The xx coordinate of each corner in the prediction period is then incrementally added with respective α\alpha, starting from the respective xx coordinate in the t0t_{0} frame, and the yy coordinate is estimated using its corresponding linear function.

L-LSTM: The same encoder-decoder architecture as the LIP-LSTM with the input to the encoder being only the past locations of the targeted person.

The predictions from t0+2t_{0}+2 to t0+10t_{0}+10 of all three baselines are used to compare with our method.

V-C Evaluation Metrics

We adopted the following 3 metrics (also used by [27]):

Mean IOU: The intersection over union (IOU) averaged across IOUs calculated between the true and predicted bounding boxes of all future time steps starting from t0+2t_{0}+2.

Mean Final IOU: The IOU averaged across those calculated for the final time step (i.e., the final future location at the end of the prediction period t0+10t_{0}+10).

Mean DE: The mean displacement error (DE) is the average 2\ell_{2} distance between the true point of a trajectory and the predicted point in each prediction time step [7, 11]. A point of a trajectory is the center of the person’s bounding box in our case which starts from t0+2t_{0}+2.

V-D Overall Results

Refer to caption

Figure 3: Qualitative results of different methods in predicting future person location and trajectory. Top 5 rows are examples that are successfully predicted by our method (LIP-LSTM). Bottom 2 rows are failure cases. From top to bottom, the walking direction of the targeted person with respect to the camera wearer is 1) toward, 2) away, 3) away, 4) across, 5) still, 6) across, 7) still. Apart from the final location and trajectory at the end of prediction period (t0+10t_{0}+10), we also show the location and trajectory at the middle of prediction period (t0+5t_{0}+5).

Table II shows the overall results and comparisons between the proposed LIP-LSTM and baseline methods. It can be seen that LIP-LSTM shows the best performance in both future location and trajectory predictions as it produces the highest IOUs and the lowest DE. The STATS method has the worst performance in both tasks. With the past locations being the only cue, the LSTM-based encoder-decoder framework is already able to have a 6.6% relative increase in the measured Mean Final IOU compared to that of LR (0.501 vs 0.470), and reduce Mean DE from 15.8 to 13.9, which is a 12.0% relative improvement. With the addition of IMU and pose information, the LIP-LSTM is able to further increase the Mean IOU (0.619 to 0.630) and Mean Final IOU (0.501 to 0.513), and reduce the Mean DE from 13.9 to 13.7.

TABLE II: Overall Results and Comparisons with Baselines
Method Mean IOU Mean Final IOU Mean DE
STATS 0.499 0.419 24.0
LR 0.605 0.470 15.8
L-LSTM 0.619 0.501 13.9
LIP-LSTM 0.630 0.513 13.7
TABLE III: Results on Each Walking Direction
Method Toward Away Across Still
Mean IOU Mean Final IOU Mean DE Mean IOU Mean Final IOU Mean DE Mean IOU Mean Final IOU Mean DE Mean IOU Mean Final IOU Mean DE
STATS 0.252 0.157 41.8 0.602 0.525 15.7 0.033 0.000 64.5 0.208 0.106 50.2
LR 0.439 0.223 25.7 0.665 0.551 12.5 0.477 0.310 19.0 0.403 0.245 29.0
L-LSTM 0.524 0.420 19.8 0.663 0.555 11.8 0.496 0.412 19.1 0.318 0.249 34.2
LIP-LSTM 0.534 0.389 18.5 0.667 0.551 11.4 0.536 0.384 15.9 0.448 0.370 23.8

Figure 3 shows some qualitative results. We visualize the predicted and true locations and trajectories at both the middle (t0+5t_{0}+5) and end (t0+10t_{0}+10) of the prediction period. In the first row, in which the targeted person is walking toward the camera wearer, the predicted trajectory of LIP-LSTM well matches the person’s true trajectory at both the middle and end time point (please refer to the red and green lines), and its predicted future person locations better align with the ground truth than other three baseline methods’. The predicted trajectories of all three baselines in this case deviate from the true trajectory, and the predicted locations are all behind the ground truth. In the case of the targeted person walking away from the camera wearer, we show 2 examples in the second and third rows. In the second row, the targeted person is walking at the approximately same speed as the camera wearer, and therefore, this person appears in roughly the same location in the future frames. The predicted locations of all methods are close to the ground truth at t0+5t_{0}+5. However, at t0+10t_{0}+10, only LIP-LSTM and LR are still able to correctly predict the person’s location. In the third row, the targeted person is walking at a different speed away from the camera wearer, and the camera wearer is also changing the walking path. In this case, only LIP-LSTM is able to reliably predict the targeted person’s location at t0+10t_{0}+10. We hypothesize that it is because LIP-LSTM receives additional IMU information to make correct prediction. In the fourth row, the targeted person is walking across the camera wearer. In this case, although LR, L-LSTM, and LIP-LSTM correctly predict the person is passing from left to right, LIP-LSTM produces a more accurate final location at t0+10t_{0}+10. In the fifth row, the targeted person is standing still in the clip and the camera wearer is turning left as well as moving forward. Both STATS and LR fail to predict the person’s future location and trajectory. While L-LSTM is able to predict the relative movements of the person in the future frames, it still fails to predict the person’s final location at t0+10t_{0}+10. In this case, only the predictions of LIP-LSTM are close to the ground truth at both t0+5t_{0}+5 and t0+10t_{0}+10. Two failure cases are shown in the sixth and seventh rows. In the sixth row, although LIP-LSTM is able to predict the targeted person is walking toward the left, it fails to catch the walking speed of the targeted person, and therefore the predicted location and trajectory lag behind the ground truth. In the seventh row, the targeted person is standing still in the scene. Although the camera wearer is moving toward the targeted person, we hypothesize that the LIP-LSTM overestimates the walking speed of the camera wearer, and therefore fails to predict the future location of the targeted person at both t0+5t_{0}+5 and t0+10t_{0}+10 in this case.

V-E Results on Each Walking Direction

As shown in the Table I, the samples of the targeted person walking away from the camera wearer account for 80% of the entire dataset, which may induce bias in network learning. Therefore, we further divide the dataset into 4 splits based on the walking direction. Each split has only one type of walking direction, and both L-LSTM and LIP-LSTM are retrained on each split. We use the same setting as in Section V-A to retrain them on the away split. For the rest three splits, we use a batch size of 32, and train the networks for 1,000 epochs. Table III shows the results on each split of different walking directions. In all scenarios, LIP-LSTM has the highest Mean IOU and lowest Mean DE. In the away split, although the measured Mean IOU and Mean Final IOU of LR, L-LSTM, and LIP-LSTM are close, in terms of Mean DE, LIP-LSTM has a 3.4% relative improvement with regard to L-LSTM, and 8.8% relative improvement with regard to LR. It is also worth noting that all methods perform the best on the away split compared to their performance on the other three splits. This may suggest that increasing the number of samples of the other three walking directions may lead to better performance on them, as the number of samples in the away split is roughly 10 times more than the other three. On both the toward and across splits, although L-LSTM achieves the best Mean Final IOU, LIP-LSTM is able to reduce the Mean DE on both splits compared to L-LSTM, 6.6% and 16.8% relative improvements on the toward and across splits, respectively. On the still split, LIP-LSTM outperforms the other three methods across all metrics, a 48.6% relative increase compared to L-LSTM in the measured Mean Final IOU.

Figure 4 shows some qualitative results on each walking direction split. In the example of the targeted person walking toward the camera wearer, the predicted future location and trajectory of LIP-LSTM well match the ground truth. In the example of the targeted person walking away, both L-LSTM and LIP-LSTM make better predictions than STATS and LR. In the example at the bottom-left of Figure 4, the targeted person is walking across the camera wearer. Although the predicted future location of LIP-LSTM does not completely align with the ground truth, its predicted movement trajectory well matches the true trajectory. In the example of the targeted person staying still, LIP-LSTM also makes better predictions than the other three methods.

Refer to caption

Figure 4: Qualitative results on each walking direction split. L-LSTM and LIP-LSTM are retrained on each split. We show one example for each split. The predicted future location (at t0+10t_{0}+10) and walking trajectory of a targeted person are rendered on the t0+10t_{0}+10 frame.

VI Conclusions

In this work, we address the task of predicting future person location and trajectory in egocentric videos captured by a wearable camera. An LSTM-based encoder-decoder framework is designed, which utilizes the cues of the past locations and body poses of the targeted person, and the past motion of the camera wearer recorded by the camera IMU. A new egocentric indoor dataset is constructed to evaluate the performance of our method. Both quantitative and qualitative results have shown that our method is able to reliably predict a targeted person’s future location and trajectory relative to the camera wearer’s viewpoint. The constructed new dataset can also be used to develop and evaluate the performance of a tracking system in the egocentric setting.

References

  • [1] Y. Li, M. Liu, and J. M. Rehg, “In the eye of beholder: Joint learning of gaze and actions in first person video,” in Proceedings of the European Conference on Computer Vision (ECCV), September 2018.
  • [2] Y. Huang, M. Cai, Z. Li, and Y. Sato, “Predicting gaze in egocentric video by learning task-dependent attention transition,” in Proceedings of the European Conference on Computer Vision (ECCV), 2018, pp. 754–769.
  • [3] D. Damen, H. Doughty, G. Maria Farinella, S. Fidler, A. Furnari, E. Kazakos, D. Moltisanti, J. Munro, T. Perrett, W. Price et al., “Scaling egocentric vision: The epic-kitchens dataset,” in Proceedings of the European Conference on Computer Vision (ECCV), 2018, pp. 720–736.
  • [4] E. Kazakos, A. Nagrani, A. Zisserman, and D. Damen, “Epic-fusion: Audio-visual temporal binding for egocentric action recognition,” in Proceedings of the IEEE International Conference on Computer Vision, 2019, pp. 5492–5501.
  • [5] S. Sudhakaran, S. Escalera, and O. Lanz, “Lsta: Long short-term attention for egocentric action recognition,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2019, pp. 9954–9963.
  • [6] T. Nagarajan, Y. Li, C. Feichtenhofer, and K. Grauman, “Ego-topo: Environment affordances from egocentric video,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 163–172.
  • [7] S. Pellegrini, A. Ess, K. Schindler, and L. Van Gool, “You’ll never walk alone: Modeling social behavior for multi-target tracking,” in 2009 IEEE 12th International Conference on Computer Vision.   IEEE, 2009, pp. 261–268.
  • [8] K. M. Kitani, B. D. Ziebart, J. A. Bagnell, and M. Hebert, “Activity forecasting,” in European Conference on Computer Vision.   Springer, 2012, pp. 201–214.
  • [9] S. Huang, X. Li, Z. Zhang, Z. He, F. Wu, W. Liu, J. Tang, and Y. Zhuang, “Deep learning driven visual path prediction from a single image,” IEEE Transactions on Image Processing, vol. 25, no. 12, pp. 5892–5904, 2016.
  • [10] S. Yi, H. Li, and X. Wang, “Pedestrian behavior understanding and prediction with deep neural networks,” in European Conference on Computer Vision.   Springer, 2016, pp. 263–279.
  • [11] A. Alahi, K. Goel, V. Ramanathan, A. Robicquet, L. Fei-Fei, and S. Savarese, “Social lstm: Human trajectory prediction in crowded spaces,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 961–971.
  • [12] A. Robicquet, A. Sadeghian, A. Alahi, and S. Savarese, “Learning social etiquette: Human trajectory understanding in crowded scenes,” in European conference on computer vision.   Springer, 2016, pp. 549–565.
  • [13] W.-C. Ma, D.-A. Huang, N. Lee, and K. M. Kitani, “Forecasting interactive dynamics of pedestrians with fictitious play,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2017, pp. 774–782.
  • [14] A. Gupta, J. Johnson, L. Fei-Fei, S. Savarese, and A. Alahi, “Social gan: Socially acceptable trajectories with generative adversarial networks,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018, pp. 2255–2264.
  • [15] J. Liang, L. Jiang, J. C. Niebles, A. G. Hauptmann, and L. Fei-Fei, “Peeking into the future: Predicting future person activities and locations in videos,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2019, pp. 5725–5734.
  • [16] C. Choi and B. Dariush, “Looking to relations for future trajectory forecast,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2019, pp. 921–930.
  • [17] J. Liang, L. Jiang, K. Murphy, T. Yu, and A. Hauptmann, “The garden of forking paths: Towards multi-future trajectory prediction,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 10 508–10 518.
  • [18] J. F. P. Kooij, N. Schneider, F. Flohr, and D. M. Gavrila, “Context-based pedestrian path prediction,” in European Conference on Computer Vision.   Springer, 2014, pp. 618–633.
  • [19] A. Bhattacharyya, M. Fritz, and B. Schiele, “Long-term on-board prediction of people in traffic scenes under uncertainty,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018, pp. 4194–4202.
  • [20] O. Styles, A. Ross, and V. Sanchez, “Forecasting pedestrian trajectory with machine-annotated training data,” in 2019 IEEE Intelligent Vehicles Symposium (IV).   IEEE, 2019, pp. 716–721.
  • [21] R. Chandra, U. Bhattacharya, A. Bera, and D. Manocha, “Traphic: Trajectory prediction in dense and heterogeneous traffic using weighted interactions,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2019, pp. 8483–8492.
  • [22] O. Makansi, O. Cicek, K. Buchicchio, and T. Brox, “Multimodal future localization and emergence prediction for objects in egocentric view with a reachability prior,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 4354–4363.
  • [23] K. Mangalam, E. Adeli, K.-H. Lee, A. Gaidon, and J. C. Niebles, “Disentangling human dynamics for pedestrian locomotion forecasting with noisy supervision,” in The IEEE Winter Conference on Applications of Computer Vision, 2020, pp. 2784–2793.
  • [24] H. Soo Park, J.-J. Hwang, Y. Niu, and J. Shi, “Egocentric future localization,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2016, pp. 4697–4705.
  • [25] T. Yagi, K. Mangalam, R. Yonetani, and Y. Sato, “Future person localization in first-person videos,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2018.
  • [26] M. Huynh and G. Alaghband, “Aol: Adaptive online learning for human trajectory prediction in dynamic video scenes,” arXiv preprint arXiv:2002.06666, 2020.
  • [27] O. Styles, V. Sanchez, and T. Guha, “Multiple object forecasting: Predicting future object locations in diverse environments,” in Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, 2020, pp. 690–699.
  • [28] P. Gujjar and R. Vaughan, “Classifying pedestrian actions in advance using predicted video of urban driving scenes,” in 2019 International Conference on Robotics and Automation (ICRA).   IEEE, 2019, pp. 2097–2103.
  • [29] B. Liu, E. Adeli, Z. Cao, K.-H. Lee, A. Shenoi, A. Gaidon, and J. C. Niebles, “Spatiotemporal relationship reasoning for pedestrian intent prediction,” IEEE Robotics and Automation Letters, vol. 5, no. 2, pp. 3485–3492, 2020.
  • [30] Y. Yao, M. Xu, Y. Wang, D. J. Crandall, and E. M. Atkins, “Unsupervised traffic accident detection in first-person videos,” in 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2019, pp. 273–280.
  • [31] Y. Yao, M. Xu, C. Choi, D. J. Crandall, E. M. Atkins, and B. Dariush, “Egocentric vision-based future vehicle localization for intelligent driving assistance systems,” in 2019 International Conference on Robotics and Automation (ICRA).   IEEE, 2019, pp. 9711–9717.
  • [32] S. Malla, I. Dwivedi, B. Dariush, and C. Choi, “Nemo: Future object localization using noisy ego priors,” arXiv preprint arXiv:1909.08150, 2019.
  • [33] S. Hochreiter and J. Schmidhuber, “Long short-term memory,” Neural computation, vol. 9, no. 8, pp. 1735–1780, 1997.
  • [34] Z. Cao, G. Hidalgo Martinez, T. Simon, S. Wei, and Y. A. Sheikh, “Openpose: Realtime multi-person 2d pose estimation using part affinity fields,” IEEE Transactions on Pattern Analysis and Machine Intelligence, 2019.
  • [35] A. Dutta and A. Zisserman, “The VIA annotation software for images, audio and video,” in Proceedings of the 27th ACM International Conference on Multimedia, ser. MM ’19.   New York, NY, USA: ACM, 2019. [Online]. Available: https://doi.org/10.1145/3343031.3350535
  • [36] N. Srivastava, G. Hinton, A. Krizhevsky, I. Sutskever, and R. Salakhutdinov, “Dropout: a simple way to prevent neural networks from overfitting,” The journal of machine learning research, vol. 15, no. 1, pp. 1929–1958, 2014.
  • [37] D. P. Kingma and J. Ba, “Adam: A method for stochastic optimization,” arXiv preprint arXiv:1412.6980, 2014.