Grouptron: Dynamic Multi-Scale Graph Convolutional Networks for Group-Aware Dense Crowd Trajectory Forecasting
Abstract
Accurate, long-term forecasting of pedestrian trajectories in highly dynamic and interactive scenes is a long-standing challenge. Recent advances in using data-driven approaches have achieved significant improvements in terms of prediction accuracy. However, the lack of group-aware analysis has limited the performance of forecasting models. This is especially nonnegligible in highly crowded scenes, where pedestrians are moving in groups and the interactions between groups are extremely complex and dynamic. In this paper, we present Grouptron, a multi-scale dynamic forecasting framework that leverages pedestrian group detection and utilizes individual-level, group-level and scene-level information for better understanding and representation of the scenes. Our approach employs spatio-temporal clustering algorithms to identify pedestrian groups, creates spatio-temporal graphs at the individual, group, and scene levels. It then uses graph neural networks to encode dynamics at different scales and aggregate the embeddings for trajectory prediction. We conducted extensive comparisons and ablation experiments to demonstrate the effectiveness of our approach. Our method achieves 9.3% decrease in final displacement error (FDE) compared with state-of-the-art methods on ETH/UCY benchmark datasets, and 16.1% decrease in FDE in more crowded scenes where extensive human group interactions are more frequently present.
I Introduction
Generating long-term and accurate predictions of human pedestrian trajectories is of enormous significance for developing reliable autonomous systems (e.g. autonomous vehicles, mobile robots), which are often required to safely navigate in crowded scenarios with many human pedestrians present, such as in crowded open traffic or in warehouses. Therefore, the capability of understanding and predicting the dense crowd behavior is instrumental in order to avoid collisions with the highly dynamic crowds and to behave in a socially-aware way.
Through daily interactions, humans are able to develop the remarkable capability of understanding crowded and interactive environments and inferring the potential future movements of all other traffic participants. Similarly, recent advances using machine learning techniques leverage the massive human interaction data are capable of achieving state-of-the-art performance on trajectory prediction [1, 2, 3, 4, 5, 6]. However, most state-of-the-art methods fail to consider the densely populated scenarios, which are extremely challenging due to the immensely dynamic and complex interactions. Moreover, in such highly crowded scenarios, groups of pedestrians are common. [7] estimates that 50% to 70% of pedestrians walk in groups, which exhibit dramatically distinct behaviors from individuals. However, existing methods lack the mechanism of representing such group dynamics, which limits their performance in highly crowded situations.
In this paper, we seek to resolve this challenge by leveraging group detection and dynamic spatio-temporal graph representations at different scales as visualized in Fig. 1. Concretely, we construct a dynamic multi-scale graph convolutional neural network that uses pedestrian group information to better learn and model the extremely complex dynamics at different scales and between different scales.

The contributions of this paper are summarized as follows:
-
•
We propose to leverage spatio-temporal clustering algorithms to detect pedestrians in groups.
-
•
We design a hierarchical prediction framework using spatio-temporal graph neural networks which encode the scene at three levels:
-
1.
The individual level encodes the historical trajectory of the predicted pedestrian.
-
2.
The group level encodes the dynamics and trajectory information within each pedestrian group.
-
3.
The scene level encodes the dynamics and interactions between pedestrian groups.
-
1.
-
•
The proposed multi-scale spatio-temporal architecture outperforms existing methods by 9.3% in terms of final displacement errors. In particular, for densely crowded scenarios, the performance improvement can be as significant as 16.1%.
II Related Work

II-A Human Trajectory Forecasting
One of the pioneering works of human trajectory forecasting is the Social Force model [8], which applies Newtonian forces to model human motion. Similar methods with strong priors have also been proposed [9]; yet, most of them rely on hand-crafted energy potential, such as relative distances and rules, to model human motion.
Recently, machine learning methods have been applied to the problem of human trajectory forecasting to obtain models with better performance. One line of work is to formulate this problem as a deterministic time-series regression problem and then solve it using, e.g., Gaussian Process Regression (GPR) [10], inverse reinforcement learning (IRL) [11], and recurrent neural networks (RNNs) [1, 12, 13].
However, the issue of these deterministic regressors is that human behavior is rarely deterministic or unimodal. Hence, generative approaches have become the state-of-the-art trajectory forecasting methods, due to recent advancements in deep generative models [14, 15] and their ability of generating distributions of potential future trajectories (instead of a single future trajectory). Most of these methods use a recurrent neural network architecture with a latent variable model, such as a conditional variational auto-encoder (CVAE) [2, 16, 17, 18], or a generative adversarial network (GAN) [19, 20, 21, 22, 23] to encode multi-modality. Compared to previous work, we not only consider multi-modality from the perspective of a single agent, but also from the group level; we take into account the phenomenon that people usually move in groups. We show that our group-aware prediction has better understanding of the scenes and achieves better forecasting performance.
II-B Graph Convolutional Networks
Of the methods mentioned above, RNN-based methods have achieved better performance. However, recurrent architectures are parameter inefficient and expensive in training [24]. Besides, to handle spatial context, RNN-based methods need additional structures. Most of them use graph models to encode neighboring pedestrians’ information since the topology of graphs is a natural way to represent interactions between pedestrians. Graph convolutional networks (GCN) introduced in [25] is more suitable for dealing with non-Euclidean data. The Social-BiGAT [20] introduces a graph attention network [26] to model social interactions. GraphSAGE [27] aggregates nodes and fuses adjacent nodes in different orders to extract node embeddings. To capture both the spatial and temporal information, Spatio-Temporal Graph Convolutional Networks (STGCN) extends the spatial GCN to spatio-temporal GCN for skeleton-based action recognition [28]. STGCN is adapted by Social-STGNN [29] for trajectory forecasting, where trajectories are modeled by graphs with edges representing social interactions and weighted by the distances between pedestrians. A development related to our paper is dynamic multi-scale GNN (DMGNN) [30], which proposes a multi-scale graph to model human body relations and extract features at multiple scales for motion prediction. There are two kinds of sub-graphs in the multi-scale graph: (i) single-scale graphs, which connect body components at the same scales, and (ii) cross-scale graphs, which form cross-scale connections among body components. Based on the multi-scale graphs, a multi-scale graph computational unit is proposed to extract and fuse features across multiple scales. Motivated by this work, we adopt the multi-scale graph strategy for dense crowd forecasting which includes scene-level graphs, group-level graphs, and individual-level graphs.
II-C Group-aware Prediction
People moving in groups (such as friends, family members, etc.) is a common phenomenon and people in each group tend to exhibit similar motion patterns. Motivated by this phenomenon, group-aware methods [31] consider the possibility of human agents being in groups or formations to have more correlated motions than independent ones. They therefore can also model reactions of agents to the moving groups. Human agents can be assigned to different groups by clustering trajectories with similar motion patterns based on methods such as -means clustering [32], support vector clustering [33], coherent filtering [34], and spectral clustering methods [35].
III Grouptron
III-A Problem Formulation
We aim to predict the trajectory distributions of a time-varying number of pedestrians . At time step , given the 2-D position of each pedestrian and their previous trajectories of time steps , our goal is to predict the distributions of their trajectories in the next future time steps, . We denote the distributions as . The performance of the predictions is evaluated by standard distance-based metrics on benchmark datasets: the average displacement error (ADE) and the final displacement error (FDE).
III-B Model Overview
Rooted in the CVAE architecture in Trajectron++ [2], we design a more expressive multi-scale scene encoding structure, which actively takes into consideration the group-level and scene-level information for better representation of crowded scenes where groups of pedestrians are present. Concretely, we leverage spatio-temporal graphs for each level to model information and interactions at the corresponding level. We refer to our model as Grouptron. Our model is illustrated in Fig. 2. In this subsection, we provide an overview of the architecture, and in Section III-C, we elaborate on the details of the Grouptron model.
At the individual level, we construct spatio-temporal graphs for individual pedestrians. The graph is centered at the node whose trajectory we want to predict. We call it the “current node”. Long short term memory (LSTM) networks [36] are used to encode this graph. We group the pedestrians with the agglomerative clustering algorithm based on Hausdorff distances [35]. STGCN is used to encode dynamics within the groups.
At the scene level, spatio-temporal graphs are created to model dynamics among pedestrian groups and are encoded using a different STGCN. Lastly, the information across different scales is combined. A decoder is then used to obtain trajectory predictions and the model can output the most possible trajectory or the predicted trajectory distributions.
III-C Multi-Scale Scene Encoder
III-C1 Individual-Level Encoder
The first level of encoding is for the individual pedestrians. We represent information at the individual level using a spatio-temporal graph for the current node. The nodes include the current node and all other nodes that are in the perception range of the current node and nodes whose perception range covers the current node. The node states are the trajectories of the nodes. The edges are directional and there is an edge if pedestrian is in the perception range of pedestrian . To encode the current node’s history trajectory, we use an LSTM network with hidden dimension 32. To encode the edges, we first perform an element-wise sum on the states of all neighboring nodes to form a single vector that represents all neighbors of the current node. This vector is then fed into the edge LSTM network which is an LSTM network with hidden dimension 8. In this way, we obtain two vectors: a vector encoding the trajectory history of the current node and a vector encoding the representation of all the neighbors of the current node.
III-C2 Pedestrian Group Clustering
To cluster nodes into groups based on trajectories, we propose to leverage the agglomerative clustering algorithm [35], which uses similarity scores based on Hausdorff distances between trajectories. The number of clusters (groups) to create for each scene is determined by:
(1) |
where C is the number of clusters and N is the total number of nodes to be clustered.
Furthermore, We only include nodes with an edge to or from the current node. This is because we only want to include nodes that can potentially influence the current node to avoid unhelpful information from nodes that are too far away from the current node.
III-C3 Group-Level Encoder
For each group, we create a spatio-temporal graph consisting of , where is the time step and is the group id. are all the nodes in the group . The node states are the trajectories of the represented pedestrians. are the set of edges between the nodes in the current group such that to allow maximum interaction modeling within pedestrian groups.
After forming the aforementioned graphs for each group, they are then passed to the group-level trajectory encoder to obtain the encoded vectors for nodes in each group. The group-level trajectory encoder is an STGCN proposed in [28] and used in [29]. We set the convolution filter size to 3 and use the same weight parameters for all the groups.
We then average the encoded vectors of all nodes in each group to obtain the representations for the corresponding groups. That is, , where is the encoded vector for group , is the encoded vector for node in the output from the group-level trajectory encoder, and is the number of nodes in group .
III-C4 Scene-Level Encoder
After obtaining the encoded vectors for each group, a scene-level spatio-temporal graph with nodes representing groups is created. That is, . , where G is the total number of groups and is the timestep. The state for each node is from the group-level trajectory encoder. are the set of edges between the groups in the scene. Each is set to to allow maximum message passing between group nodes.
We then select the encoded vector corresponding to the last timestep and the group id of the current node as the scene-level encoding: , where is the group id of the current node we are encoding for and T is the total number of time steps.
III-C5 Multi-Scale Encoder Output
The output of the multi-scale scene encoder is the concatenation of the following level encoded vectors: the output from the node history encoder, the output from individual-level edge encoder, and the output from the scene-Level encoder. That is, , where is the encoded vector for the current node’s history trajectory, is the vector representing individual-level neighbors, is the encoded vector from the scene-level encoder.
III-D Decoder Network
Together with the latent variable , is passed to the decoder that is a Gated Recurrent Unit (GRU) [37] with 128 dimensions. The output from the GRU decoder is then fed into dynamics integration modules as control actions to output the predicted trajectory distributions, or the single most-likley trajectory, depending on the task.
III-E Loss Functions
We adopt the following objective function for the overall CVAE model:
(2) | ||||
where is the mutual information between x and z under the distribution . We follow the process given in [38] to compute . We approximate with , and obtain the unconditioned latent distribution by summing out over the batch.
IV Experiments
Method | ETH | HOTEL | UNIV | ZARA1 | ZARA2 | AVG |
---|---|---|---|---|---|---|
Social-LSTM | 2.35/1.09 | 1.76/0.79 | 1.40/0.67 | 1.00/0.47 | 1.17/0.56 | 1.54/0.72 |
Social-GAN | 1.52/0.81 | 1.61/0.72 | 1.26/0.60 | 0.69/0.34 | 1.84/0.42 | 1.18/0.58 |
SoPhie | 1.43/0.70 | 1.67/0.76 | 1.24/0.54 | 0.63/0.34 | 0.78/0.38 | 1.15/0.54 |
Trajectron++ | 1.68/0.71 | 0.46/0.22 | 1.07/0.41 | 0.77/0.30 | 0.59/0.23 | 0.91/0.37 |
Grouptron | 1.56/0.70 | 0.46/0.21 | 0.97/0.38 | 0.76/0.30 | 0.56/0.22 | 0.86/0.36 |
Method | UNIV | UNIV-40 | UNIV-45 | UNIV-50 |
---|---|---|---|---|
Trajectron++ | 1.07/0.41 | 1.17/0.436 | 1.24/0.46 | 1.25/0.47 |
Grouptron | 0.97/0.38 | 1.00/0.39 | 1.04/0.40 | 1.07/0.42 |
IV-A Datasets
The model is trained on two publicly available datasets that are benchmarks in the field: The ETH [39], with subsets named ETH and HOTEL, and the UCY [40] datasets, with subsets named ZARA1, ZARA2, and UNIV. The trajectories are sampled at seconds intervals. The model observes 8 time steps, which corresponds to 3.2 seconds, and predicts the next 12 time steps, which corresponds to 4.8 seconds.
To further evaluate and demonstrate Grouptron’s performance in densely populated scenarios, we create UNIV-N test sets, where is the minimum number of people present simultaneously at each time step in the test sets. Each UNIV-N test set contains all time steps that have at least people in the scene simultaneously from the original UNIV test set. In this way, we created test sets UNIV-40, UNIV-45, and UNIV-50 and, at each time step, there are at least 40, 45, and 50 people in the scene at the same time. These test sets are far more challenging than the original UNIV test set because of the more complex and dynamic interactions at different scales. We train the models on the original UNIV training set and evaluate the models on the UNIV-40, UNIV-45, and UNIV-50 test sets.
IV-B Evaluation Metrics
IV-B1 Final Displacement Error
(3) |
which is the distance between the predicted final position and the ground truth final position with prediction horizon .
IV-B2 Average Displacement Error
(4) |
where is the total number of pedestrians, is the number of future timesteps we want to predict for, is the predicted trajectory for pedestrian at timestep . is the mean distance between the ground truth and predicted trajectories.
IV-C Experiment settings
The Grouptron model is implemented using PyTorch. The model is trained using an Intel I7 CPU and NVIDIA GTX 1080 Ti GPUs for 100 epochs. The batch size is 256. For the HOTEL, UNIV, ZARA1, and ZARA2 datasets, the output dimension for the group-level and scene level encoders are 16. For the ETH dataset, we set the output dimension of the group-level and scene-level encoders to be 8. This is because the ETH test set contains only 2 timesteps with at least 5 people in the scene, out of the total 1161 timesteps. In comparison, the training set contains 1910 timesteps, out of 4976 in total, with at least 5 people. Thus, to help the model learn generalizable representations in this case, we decrease the output dimension of the STGCNs to 8. The learning rate is set to initially and decayed exponentially every epoch with a decay rate of 0.9999. The model is trained using Adam gradient descent and gradients are clipped at .
IV-D Evaluation of the Group Clustering Algorithm

In Fig. 3, it is shown that the groups created by the agglomerative clustering method are very close to the natural definition of pedestrian groups. We can see that pedestrians 5 and 6 are travelling in a highly correlated fashion and pedestrians 1, 2, 3, and 4’s trajectories are highly similar as well. In both cases, the clustering algorithm is able to correctly cluster these pedestrians into their corresponding groups. This shows that by using agglomerative clustering based on Hausdorff distances, Grouptron is able to successfully generate naturally defined groups. To quantitatively evaluate the groups generated by the algorithm, we selected 10 random time steps from the ETH training dataset. We invited 10 human volunteers to label groups for the time steps and used the agglomerative clustering method described in Section III-C to generate group clusters, respectively. For both human-generated groups and algorithm-generated groups, the number of groups to be formed is computed using Equation 1. We then compute the average Sørensen–Dice coefficient between human-generated and algorithm-generated groups. That is, we use
(5) |
where is the number of timesteps, is the total number of human annotators, is the grouping created by the agglomerative clustering method for time step , is the grouping created by humans for time step , and measures how many of the groups by humans and the algorithm are exactly the same. The Average Dice coefficient between human annotators and the algorithm is 0.72. The higher the Dice coefficient, the more similar are the groups created by the agglomerative clustering method and human annotators. Thus, The average Dice Coefficient of 0.72 indicates the groups output by agglomerative clustering method are really similar to human-generated ones.
IV-E Quantitative Results
We compare Grouptron’s performance with state-of-the-art methods and common baselines in the field in terms of the FDE and ADE metrics, and the results are shown in Table I. Overall, Grouptron outperforms all state-of-the-art methods with considerable decrease in displacement errors. Since Grouptron is built on Trajectron++, we also compare the FDE and ADE values of Grouptron and Trajectron++. We find that Grouptron outperforms Trajectron++ on all 5 datasets by considerable margins. Particularly, on the ETH dataset, Grouptron achieves an FDE of 1.56m, which is 7.1% better than the FDE value of 1.68m by Trajectron++. Furthermore, Grouptron achieves an FDE of 0.97m on the UNIV dataset. This is 9.3% reduction in FDE error when compared with the FDE value of 1.07m by Trajectron++ on the same dataset.
Moreover, we compare Grouptron’s performance with Trajectron++’s in dense crowds with the UNIV-N datasets in Table II. Overall, Grouptron outperforms Trajectron++ on all the UNIV-N test sets by enormous margins. In particular, Grouptron achieves an FDE of 1.04m and ADE of 0.40m on the UNIV-45 test set, which contains all timesteps from the original UNIV test set that have at least 45 pedestrians in the scene at the same time. This is 16.1% in FDE improvement when compared with the FDE value of Trajectron++ and 13.0% in ADE improvement when compared with the ADE value of Trajectron++ on the same test set.
Furthermore, we notice that the state-of-the-art method, Trajectron++, performs substantially worse as the number of pedestrians in the scene increases. Specifically, Trajectron++’s FDE increases from 1.07m to 1.25m as the minimum number of pedestrians in the scene increases from 1 to 50. In contrast, Grouptron’s FDE remains relatively stable as the number of pedestrians increases. This shows that Grouptron performs much better and is more robust in densely populated scenarios.
IV-F Qualitative Analysis


Fig. 4 shows Grouptron’s most likely predictions for some examples of the UNIV dataset. Fig. 4a shows two pedestrian groups crossing paths. We can see that Grouptron’s predictions are consistent with the groups. Furthermore, it accurately predicts when and where the two groups’ trajectories intersect. Fig. 4b shows a case where pedestrians are forming groups and merging paths. Grouptron again successfully predicts the formation of this group. Fig. 4c and Fig. 4d show Grouptron’s performance in densely populated scenes with more than 40 pedestrians. Even in these extremely challenging scenarios for state-of-the-art methods, Grouptron still produces predictions of high quality and the predictions are consistent with pedestrian groups. Furthermore, we can see that even when pedestrian groups are crossing paths or influencing each other, Grouptron successfully predicts these highly dynamic and complex scenarios.
In Fig. 5, we compare Grouptron’s distributions of 20 most likely predictions with those of Trajectron++’s. Comparing Fig. 5a with 5b and 5c with 5d shows that Grouptron’s predictions for the pedestrians of interest reflect the interactions within pedestrian groups more accurately. Furthermore, Grouptron’s prediction distributions have much smaller ranges, indicating that it is much more confident with prediction outcomes.
V Conclusions
In this paper, we present Grouptron, a multi-scale graph neural network for pedestrian trajectory forecasting. The spatio-temporal graphs model complex and highly dynamic pedestrian interactions at three scales: the individual level, the group level, and the scene level. Grouptron detects pedestrian groups using agglomerative clustering based on Hausdorff distances, uses GCNs and LSTM-based networks to encode information at different scales, and combines the embeddings across scales. With these novel designs, Grouptron achieves high performance even in situations that are tough for state-of-the-art methods. Through extensive experiments, we show that Grouptron outperforms state-of-art methods on ETH and UCY datasets, which are standard benchmarks in the field. Furthermore, through qualitative comparisons and experiments run on densely populated UNIV-N test sets, we show that Grouptron outperforms state-of-art methods by even larger margins and is robust to the change in the number of pedestrians in the scene.
References
- [1] A. Alahi, K. Goel, V. Ramanathan, A. Robicquet, L. Fei-Fei, and S. Savarese, “Social lstm: Human trajectory prediction in crowded spaces,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 961–971.
- [2] T. Salzmann, B. Ivanovic, P. Chakravarty, and M. Pavone, “Trajectron++: Dynamically-feasible trajectory forecasting with heterogeneous data,” in Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XVIII 16. Springer, 2020, pp. 683–700.
- [3] J. Li, F. Yang, M. Tomizuka, and C. Choi, “Evolvegraph: Multi-agent trajectory prediction with dynamic relational reasoning,” Advances in neural information processing systems, vol. 33, pp. 19 783–19 794, 2020.
- [4] J. Li, F. Yang, H. Ma, S. Malla, M. Tomizuka, and C. Choi, “Rain: Reinforced hybrid attention inference network for motion forecasting,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, pp. 16 096–16 106.
- [5] C. Choi, J. H. Choi, J. Li, and S. Malla, “Shared cross-modal trajectory prediction for autonomous driving,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 244–253.
- [6] H. Ma, Y. Sun, J. Li, M. Tomizuka, and C. Choi, “Continual multi-agent interaction behavior prediction with conditional generative memory,” IEEE Robotics and Automation Letters, vol. 6, no. 4, pp. 8410–8417, 2021.
- [7] J. Sochman and D. C. Hogg, “Who knows who - inverting the social force model for finding groups,” 2011 IEEE International Conference on Computer Vision Workshops (ICCV Workshops), pp. 830–837, 2011.
- [8] D. Helbing and P. Molnar, “Social force model for pedestrian dynamics,” Physical review E, vol. 51, no. 5, p. 4282, 1995.
- [9] G. Antonini, M. Bierlaire, and M. Weber, “Discrete choice models of pedestrian walking behavior,” Transportation Research Part B: Methodological, vol. 40, no. 8, pp. 667–687, 2006.
- [10] J. M. Wang, D. J. Fleet, and A. Hertzmann, “Gaussian process dynamical models for human motion,” IEEE transactions on pattern analysis and machine intelligence, vol. 30, no. 2, pp. 283–298, 2007.
- [11] N. Lee and K. M. Kitani, “Predicting wide receiver trajectories in american football,” in 2016 IEEE Winter Conference on Applications of Computer Vision (WACV). IEEE, 2016, pp. 1–9.
- [12] A. Jain, A. R. Zamir, S. Savarese, and A. Saxena, “Structural-rnn: Deep learning on spatio-temporal graphs,” in Proceedings of the ieee conference on computer vision and pattern recognition, 2016, pp. 5308–5317.
- [13] A. Vemula, K. Muelling, and J. Oh, “Social attention: Modeling attention in human crowds,” in 2018 IEEE international Conference on Robotics and Automation (ICRA). IEEE, 2018, pp. 4601–4607.
- [14] K. Sohn, H. Lee, and X. Yan, “Learning structured output representation using deep conditional generative models,” Advances in neural information processing systems, vol. 28, pp. 3483–3491, 2015.
- [15] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio, “Generative adversarial nets,” Advances in neural information processing systems, vol. 27, 2014.
- [16] N. Lee, W. Choi, P. Vernaza, C. B. Choy, P. H. Torr, and M. Chandraker, “Desire: Distant future prediction in dynamic scenes with interacting agents,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2017, pp. 336–345.
- [17] H. Ma, J. Li, W. Zhan, and M. Tomizuka, “Wasserstein generative learning with kinematic constraints for probabilistic interactive driving behavior prediction,” in 2019 IEEE Intelligent Vehicles Symposium (IV). IEEE, 2019, pp. 2477–2483.
- [18] J. Li, H. Ma, Z. Zhang, J. Li, and M. Tomizuka, “Spatio-temporal graph dual-attention network for multi-agent prediction and tracking,” IEEE Transactions on Intelligent Transportation Systems, 2021.
- [19] A. Gupta, J. Johnson, L. Fei-Fei, S. Savarese, and A. Alahi, “Social gan: Socially acceptable trajectories with generative adversarial networks,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2018, pp. 2255–2264.
- [20] V. Kosaraju, A. Sadeghian, R. Martín-Martín, I. Reid, S. H. Rezatofighi, and S. Savarese, “Social-bigat: Multimodal trajectory forecasting using bicycle-gan and graph attention networks,” arXiv preprint arXiv:1907.03395, 2019.
- [21] A. Sadeghian, V. Kosaraju, A. Sadeghian, N. Hirose, H. Rezatofighi, and S. Savarese, “Sophie: An attentive gan for predicting paths compliant to social and physical constraints,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2019, pp. 1349–1358.
- [22] T. Zhao, Y. Xu, M. Monfort, W. Choi, C. Baker, Y. Zhao, Y. Wang, and Y. N. Wu, “Multi-agent tensor fusion for contextual trajectory prediction,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2019, pp. 12 126–12 134.
- [23] J. Li, H. Ma, and M. Tomizuka, “Conditional generative neural system for probabilistic trajectory prediction,” in 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE, 2019, pp. 6150–6156.
- [24] S. Bai, J. Z. Kolter, and V. Koltun, “An empirical evaluation of generic convolutional and recurrent networks for sequence modeling,” arXiv preprint arXiv:1803.01271, 2018.
- [25] T. N. Kipf and M. Welling, “Semi-supervised classification with graph convolutional networks,” arXiv preprint arXiv:1609.02907, 2016.
- [26] P. Veličković, G. Cucurull, A. Casanova, A. Romero, P. Lio, and Y. Bengio, “Graph attention networks,” arXiv preprint arXiv:1710.10903, 2017.
- [27] W. L. Hamilton, R. Ying, and J. Leskovec, “Inductive representation learning on large graphs,” in Proceedings of the 31st International Conference on Neural Information Processing Systems, 2017, pp. 1025–1035.
- [28] S. Yan, Y. Xiong, and D. Lin, “Spatial temporal graph convolutional networks for skeleton-based action recognition,” in Thirty-second AAAI conference on artificial intelligence, 2018.
- [29] A. Mohamed, K. Qian, M. Elhoseiny, and C. Claudel, “Social-stgcnn: A social spatio-temporal graph convolutional neural network for human trajectory prediction,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 14 424–14 432.
- [30] M. Li, S. Chen, Y. Zhao, Y. Zhang, Y. Wang, and Q. Tian, “Dynamic multiscale graph neural networks for 3d skeleton based human motion prediction,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 214–223.
- [31] A. Rudenko, L. Palmieri, M. Herman, K. M. Kitani, D. M. Gavrila, and K. O. Arras, “Human motion trajectory prediction: A survey,” The International Journal of Robotics Research, vol. 39, no. 8, pp. 895–935, 2020.
- [32] J. Zhong, W. Cai, L. Luo, and H. Yin, “Learning behavior patterns from video: A data-driven framework for agent-based crowd modeling,” in Proceedings of the 2015 International Conference on Autonomous Agents and Multiagent Systems, 2015, pp. 801–809.
- [33] I. A. Lawal, F. Poiesi, D. Anguita, and A. Cavallaro, “Support vector motion clustering,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 27, no. 11, pp. 2395–2408, 2016.
- [34] N. Bisagno, B. Zhang, and N. Conci, “Group lstm: Group trajectory prediction in crowded scenarios,” in Proceedings of the European Conference on Computer Vision (ECCV) Workshops, 2018, pp. 0–0.
- [35] S. Atev, G. Miller, and N. P. Papanikolopoulos, “Clustering of vehicle trajectories,” IEEE transactions on intelligent transportation systems, vol. 11, no. 3, pp. 647–657, 2010.
- [36] S. Hochreiter and J. Schmidhuber, “Long short-term memory,” Neural computation, vol. 9, no. 8, pp. 1735–1780, 1997.
- [37] J. Chung, C. Gulcehre, K. Cho, and Y. Bengio, “Empirical evaluation of gated recurrent neural networks on sequence modeling,” arXiv preprint arXiv:1412.3555, 2014.
- [38] S. Zhao, J. Song, and S. Ermon, “Infovae: Balancing learning and inference in variational autoencoders,” in Proceedings of the aaai conference on artificial intelligence, vol. 33, no. 01, 2019, pp. 5885–5892.
- [39] S. Pellegrini, A. Ess, K. Schindler, and L. Van Gool, “You’ll never walk alone: Modeling social behavior for multi-target tracking,” in 2009 IEEE 12th International Conference on Computer Vision. IEEE, 2009, pp. 261–268.
- [40] A. Lerner, Y. Chrysanthou, and D. Lischinski, “Crowds by example,” in Computer graphics forum, vol. 26, no. 3. Wiley Online Library, 2007, pp. 655–664.
- [41] B. Ivanovic and M. Pavone, “The trajectron: Probabilistic multi-agent trajectory modeling with dynamic spatiotemporal graphs,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2019, pp. 2375–2384.