Nothing Special   »   [go: up one dir, main page]

Next Article in Journal
Critic Learning-Based Safe Optimal Control for Nonlinear Systems with Asymmetric Input Constraints and Unmatched Disturbances
Next Article in Special Issue
Multivariate Modeling for Spatio-Temporal Radon Flux Predictions
Previous Article in Journal
Double-Layer Detection Model of Malicious PDF Documents Based on Entropy Method with Multiple Features
Previous Article in Special Issue
Flight Delay Regression Prediction Model Based on Att-Conv-LSTM
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Trajectory Feature-Boosting Network for Trajectory Prediction

School of Computer Science and Engineering, Southeast University, Nanjing 211189, China
*
Author to whom correspondence should be addressed.
Entropy 2023, 25(7), 1100; https://doi.org/10.3390/e25071100
Submission received: 5 June 2023 / Revised: 6 July 2023 / Accepted: 21 July 2023 / Published: 23 July 2023

Abstract

:
Trajectory prediction is an essential task in many applications, including autonomous driving, robotics, and surveillance systems. In this paper, we propose a novel trajectory prediction network, called TFBNet (trajectory feature-boosting network), that utilizes trajectory feature boosting to enhance prediction accuracy. TFBNet operates by mapping the original trajectory data to a high-dimensional space, analyzing the change rules of the trajectory in this space, and finally aggregating the trajectory goals to generate the final trajectory. Our approach presents a new perspective on trajectory prediction. We evaluate TFBNet on five real-world datasets and compare it to state-of-the-art methods. Our results demonstrate that TFBNet achieves significant improvements in the ADE (average displacement error) and FDE (final displacement error) indicators, with increases of 46% and 52%, respectively. These results validate the effectiveness of our proposed approach and its potential to improve the performance of trajectory prediction models in various applications.

1. Introduction

Trajectory prediction is a crucial technology with practical applications in various fields, including automatic driving [1], robotics [2], and surveillance systems [3]. In autonomous driving systems, accurate prediction of the trajectories of surrounding vehicles, pedestrians, and other traffic participants is crucial for safety and decision-making. In robot navigation and path planning, predicting the trajectories of objects in the surrounding environment helps robots avoid obstacles, prevent collisions, and optimize path selection. In security surveillance, trajectory prediction assists security systems in analyzing and identifying suspicious behavior or abnormal movements, used in areas such as video monitoring, intrusion detection, and behavior analysis. In traffic management, by predicting the motion trajectories of vehicles, pedestrians, and bicycles, traffic authorities can better plan roads, traffic signals, and traffic flow to improve efficiency and reduce congestion. Its ability to predict and prevent traffic accidents, improve traffic flow management, and enhance logistics and distribution efficiency makes it an essential tool in our daily lives.
In the context of traffic management, trajectory prediction can help us forecast traffic congestion and prevent accidents, improving road planning and traffic flow management. Additionally, in the navigation of autonomous vehicles and robots, trajectory prediction can enhance safety by predicting the trajectories of pedestrians and other vehicles, and by providing better path planning for making safe and efficient driving decisions. Moreover, in monitoring systems, trajectory prediction can be used to forecast the movement of people or vehicles in crowded scenes, improving the efficiency of tracking and monitoring systems. Finally, in logistics and distribution, trajectory prediction can optimize distribution efficiency, forecast the movement of distribution vehicles, and provide better support for distribution route planning.
However, trajectory prediction is a challenging research topic. The motion mode of an object is complex and uncertain and will be affected by various factors, such as obstacles, interactions with other objects, and the uncertain environment. These factors make the prediction of future motion nonlinear and complex, making it difficult to accurately forecast the trajectory of an object. Additionally, the future movement of an object also depends on its goals, plans, and intentions, which are difficult to model and predict. As a result, developing accurate and efficient trajectory prediction methods remains an ongoing research challenge.
Early work on trajectory prediction primarily relied on Markov process models [4,5] and social force models [6,7]. However, with the rapid development of deep learning in recent years, researchers have started using deep neural networks to tackle trajectory prediction problems. AMGB [8] handles trajectory prediction in dynamic environments from a new perspective of predicting vehicle movement direction and distance. Zhang [9] proposed a dual-branch spatial-temporal graph neural network to automatically model and group view regions. Hui [10] proposed a trajectory prediction model based on a deep encoder–Cdecoder and a deep neural network (DNN), which introduces an attention mechanism into the traditional encoder–Cdecoder framework. Shi [11] proposed an integrated deep-learning-based two-dimensional trajectory prediction model that can predict combined behaviors. Recurrent neural networks (RNN), including long short-term memory (LSTM) and gated recurrent unit (GRU), have been used to model trajectory prediction [12,13,14,15,16,17,18]. These models can handle continuous data and consider the object’s past movements while making predictions.
Social-LSTM [13] is an architecture that connects LSTMs of different sequences and proposes a social pooling layer to handle different agent interactions. However, in many scenarios, multiple objects interact with each other, making it challenging to model uncertainty, unpredictability, and the intentions of different agents. Simple recurrent neural networks may not handle these interactions effectively. In recent years, graph neural networks (GNN) have gained popularity and achieved good results in many fields. GNNs are designed to handle interactions between different nodes. As a result, researchers have applied GNNs to trajectory prediction [19,20,21,22,23,24,25]. The primary function of GNNs is to model the interactions between different nodes, which can help in dealing with the interaction between multiple objects, such as vehicles, pedestrians, and obstacles, in the case of autonomous vehicles. GNNs have shown great potential to improve the performance of trajectory prediction models.
Recent research has focused on generating multiple possible future trajectories because the future motion of objects is uncertain, and this approach may provide more realistic predictions and improve the safety of autonomous systems. For instance, a car may turn left or go straight at an intersection, leading to different future trajectories depending on the driver’s intended destination. Similarly, a person may produce multiple possible future trajectories depending on their intended destination. At present, multi-modal trajectory prediction is mainly based on generation methods, which can be divided into two categories: GAN generation methods [26,27,28] and conditional variational auto-encoder (CVAE) methods [16,21,29,30,31,32,33].
The proposed trajectory prediction network in this paper, called TFBNet, aims to improve the accuracy of trajectory prediction by using trajectory feature boosting. This approach maps the original trajectory data to a higher-dimensional space where the motion intention of the object can be mined. TFBNet consists of four main modules: (1) a trajectory feature-boosting module, which maps the trajectory data to a higher dimension and learns the distribution of future trajectories through a CVAE module; (2) a multi-recurrent goal mining module, which uses GRU to mine trajectory data in the high-dimensional space and form multiple prediction goals; (3) a goal feedback module, which links the prediction goals with the module and feeds back the prediction results to dynamically adjust the model and has the function of memory backtracking. It feeds back the historical trajectory prediction results to the subsequent prediction, which is convenient for the model to capture similar trajectories and assist in prediction; (4) a target aggregation module, which evaluates multiple predicted goals, selects effective goals, reduces high-dimensional trajectory data to low-dimensional, and generates the final predicted trajectory.
Understanding the intention of object motion is crucial to improving the accuracy of multi-modal trajectory prediction. Different intentions may produce completely different results, and effectively mining an object’s potential intention based on past trajectory observations is, therefore, essential. The key to improve the accuracy of multi-modal trajectory prediction is to determine the object’s intention accurately. In order to better capture the motion intent of objects, this paper maps the original trajectory data into a high-dimensional space and explores the motion intent of objects in that space.
The main contributions of this paper are as follows:
  • A novel approach for trajectory prediction is proposed, which maps the original trajectory data to a high-dimensional space to better mine the object’s motion intention. The experimental results demonstrate that this approach can improve the accuracy of trajectory prediction by capturing the change law of the object’s motion trajectory more effectively.
  • A dilated attention gating structure (DAGConv) is introduced and applied to both the CVAE module and the goal aggregation module. The results show that DAGConv can effectively extract useful information and significantly enhance the accuracy of trajectory prediction.
  • A goal feedback structure is designed, which not only provides real-time feedback to the model, but also evaluates the output results of the trajectory representation module.
  • A goal aggregation module is developed, which integrates the attention mechanism and the dilated attention gating structure. This module can evaluate multiple prediction goals, select effective ones adaptively, and generate the final predicted trajectory.
In summary, we propose a novel trajectory prediction model that exhibits improved predictive performance compared to other trajectory prediction models. Additionally, we provide a new approach for trajectory prediction by mapping the data into a high-dimensional space to better explore trajectory features.
The organizational structure of this paper is as follows: Section 1 provides an introduction; Section 2 introduces the relevant methods; Section 3 presents the model proposed in this paper to address the trajectory prediction problem and provides a detailed description of the model; Section 4 compares and analyzes the experimental results; Section 5 concludes the paper with a summary of the key findings.

2. Related Work

Trajectory prediction involves predicting the future motion trajectory of an object based on its past motion trajectory. Accurate prediction of an object’s future motion trajectory would greatly benefit applications such as autonomous driving and robotics. In trajectory prediction, better exploration and understanding of an object’s motion intent are key to improving the performance of trajectory prediction. In recent years, trajectory prediction has made a lot of progress. In order to enable agents to interact better, Agentformer (agent-aware transformers for socio-temporal multi-agent forecasting) [33] is proposed. Based on the Transformer architecture, Agentformer has designed an agent-aware attention structure, which allows the model to distinguish different agents, thus improving the ability to interact with information. Social-GAN (socially acceptable trajectories with generative adversarial networks) [26] improves on Social-LSTM, focusing on generating trajectories that conform to social norms. At the same time, the social LSTM model is pooled every step into only one pool in the known trajectory change stage, reducing the computational overhead and improving the prediction speed of the model. In order to better deal with the interaction between agents, a graph neural network is undoubtedly a good choice [19,20,21,22,23,24,25], Social-bigat (social bicycle-gan and graph attention networks) [19] combines GAT (graph attention network) and GAN (generative adversarial networks). This method also takes into account the relationship between noise and prediction trajectory, and proposes a latent encoder, which links the prediction trajectory with noise to improve the accuracy of the prediction trajectory. In contrast to the model with a cyclic recursive structure, Social-STGCNN (social spatio-temporal graph convolutional neural network) [20] uses a TCN (temporary revolutionary network) to replace the cyclic recursive structure. Compared with Social-bigat, Social-STGCNN improves the prediction accuracy with fewer parameters and better training speed. STAR (spatio-temporal graph transformer networks) [34] is different from the graph convolution network. STAR takes the agent dimension as the time dimension of the transformer and uses the transformer structure to realize the interaction between agents. RSBG (recursive social behavior graph) [22] proposes a social behavior graph and uses a recursive structure to mine interactive information. Although the graph neural network can extract the interaction well, it is more suitable for the prediction of fixed nodes, such as traffic flow prediction. The node of traffic flow prediction is the sensor fixed on the road, and its position is unchanged. However, for trajectory prediction, the scene is likely to change constantly. The number of agents in different scenes is generally different, and the number of agents at different times in the same scene may also be different. However, to construct a graph structure, the number of nodes of the graph needs to be determined in advance. Therefore, for trajectory prediction, the graph neural network cannot play its role well. The semantic information of the trajectory context also has an impact on prediction. CSCnet (contextual semantic consistency network) [35] considers the semantic consistency of the context of trajectory prediction in a congested space to improve the prediction accuracy. Ynet (a scene compliant trajectory forecasting network) [36] uses multimodal input to input a semantic map, RGB map and other information into the model. However, this requires large datasets. Some datasets do not have such data, and the adaptability is not so good. Use of a physical model is also one of the trajectory prediction methods. It is also one of the feasible methods to integrate a physical model into a deep learning model. NSP-SFM (neural social physics models–social force models) [37] combines a physical model with a deep learning model, uses a physical model with learnable parameters, and obtains good prediction results. For multi-modal trajectory prediction, use of a goal-driven method has become one of the research hotspots in recent years [29,30,31,38,39,40]. Tnt (target-driven trajectory prediction) [38] is guided by the endpoint of the trajectory and predicts the trajectory in three stages. First, the final goal is generated, and then the trajectory is generated based on these goals. Finally, these trajectories are filtered and scored to obtain the prediction results. PECNet (predicted endpoint conditioned network) [31] is also used to generate the final goal point first, and then uses the goal point to assist in prediction, while paying attention to the generation of trajectories that conform to social norms. Bitrap (bi-directional pedestrian trajectory prediction) [29] uses a bidirectional loop decoder based on the goal condition to estimate the endpoint of the trajectory and improve the long-term prediction ability of the model. SGNet (stepwise goal-driven networks) [30] improves trajectory prediction accuracy by generating multiple goals. MemoNet (retrospective-memory-based network) [40] proposes a memory backtracking structure, which predicts the movement intention of the agent by looking for similar scenes in the training data. Each prediction can find similar scenes in the memory, and, at the same time, the current scene is also stored in the memory. The prediction process of MemoNet is also divided into two steps: first, predict the endpoint, and then generate the trajectory according to the end point.
In contrast to previous models, the model proposed in this paper proposes a new idea, which involves mapping the original trajectory data to a higher dimension to space, thus improving the representation of the trajectory features and facilitating better mining of trajectory data. In order to better mine the trajectory features, this paper proposes a dilated attention gating structure and a multi-recurrent goal aggregation structure to improve the ability of the model to mine trajectory features. At the same time, this paper considers the impact of the goal on the trajectory prediction, and designs a goal feedback structure to assist in the prediction.

3. Proposed Model

3.1. Problem Formulation

The trajectory prediction in this paper is based on the observed trajectory of the agent in the past to predict the trajectory of the agent in the future. Specifically, X = X t obs + 1 ,   X t obs + 2 ,   ,   X t R B × o b s × C represents the trajectory data with the length of obs observed in the past, where B represents the batch size, and C represents the number of features entered, including the trajectory features (coordinates, speed, acceleration, etc.). Y = y t + 1 ,   y t + 2 ,   ,   y t + p r e d R B × p r e d × p o s represents the prediction goal, where pred represents the predicted time length and pos represents the predicted coordinates. The prediction content in this paper is multimodal trajectory prediction, so the prediction result Y pred = y ˜ t + 1 ,   y ˜ t + 2 ,   ,   y ˜ t + pred R B × pred × K × pos , where K represents the number of samples output.

3.2. The Architecture of TFBNet

Figure 1 depicts the overall architecture of TFBNet. The main idea behind TFBNet is to map the original trajectory data to a high-dimensional space, extract the trajectory features in that space, generate multiple goal trajectories, and finally use the goal aggregation module to combine them and reduce the high-dimensional data to a lower-dimensional space. TFBNet comprises four key components: a trajectory feature-boosting module, a multi-recurrent goal mining module, a goal aggregation module, and a goal feedback module. The trajectory feature-boosting module improves the representation of the original trajectory data, while the multi-recurrent goal mining module is responsible for extracting the trajectory features. The goal aggregation module is used to reduce the high-dimensional data to a low-dimensional space and to generate the final prediction results. Additionally, this paper introduces a dilated attention gating convolution (DAGConv), which is a crucial component of the CVAE and the goal aggregation module. The experimental results obtained demonstrate that DAGConv significantly improves the prediction accuracy of the model.

3.3. Trajectory Feature Boosting

3.3.1. Trajectory Representation Boosting

To map the low-dimensional trajectory data to the high-dimensional space, this paper processes the data of each time step separately. In particular, multiple GRU structures are utilized to process the trajectory data of each time step and these structures are linked together according to the time relationship. This stacking method helps to mine the time dependence of the trajectory data and provides a more comprehensive representation of the trajectory. By processing each time step independently, the model is able to capture the information of not only the current time step, but also of the other time steps.
Furthermore, each GRU passes through an independent full connection layer, allowing for more information to be extracted from each time step. Finally, the information of all time steps is aggregated to obtain a more comprehensive representation of the trajectory data. By employing this method, the model is able to capture the complex and dynamic nature of the trajectory data and provide more accurate trajectory predictions. The formula is as follows:
r t = W r G R U Dropout Concat x t , g r + b r
r o u t = Concat r t , r t + 1 , , r t + o b s
where, x t represents the original input at time t, g r represents the variable output from the goal feedback mechanism to the trajectory representing the lifting module, r o u t represents the output of the representation boosting module, W r , and b r represent the training parameters of the model. Dropout refers to randomly dropping out some neuron connections, while Concat represents concatenation.

3.3.2. Dilated Attention Gating

The structure of DAGConv (dilated attention gating convolution) is shown in Figure 2, which is mainly composed of DAConv (dilated attention convolution), MLP (multilayer perceptron), and a gating mechanism. DAConv and MLP separately mine the information and then each pass the gating, and, finally, the two results are added to obtain the final output. The following introduces the structure of DAConv:
As shown in Figure 2, DAConv uses multi-resolution expansion convolution to mine trajectory information at different levels. For the input data, the expansion convolution of different size convolution kernels is used to process the data separately, and then the results are organically combined through the attention mechanism. The attention mechanism can be introduced to adaptively select multi-resolution convolution results. It should be noted that the size of the convolution result is different, so this paper uses the boundary padding method to make the output result size the same. The expansion convolution increases the receptive field of the convolution kernel while keeping the number of parameters unchanged. Compared with RNN, it can greatly improve the training speed of the model. The attention formula is:
α = W a · σ W l ( input ) + b l + x ˜ + b a
γ i = exp α i j = 1 n exp α j
O u t D = [ f ( γ 0 x ˜ 0 , γ 1 x ˜ 1 , , γ n u m x ˜ n u m ) ]
Among them, σ is the activation function tanh, input represents the initial input of DAGConv, and x ˜ indicates the DAConv output. W a , W l , b l , b A are trainable parameters of the model. O u t D represents the output of the attention mechanism and num represents the number of different convolutions. In order to better select the trajectory goals and generate more accurate prediction results, this paper introduces a channel attention mechanism ECA layer [41] in DAConv. The ECA layer avoids dimensionality reduction and local cross-channel interaction, which improves the performance of the model while reducing the complexity of the model. The weight of channel C i is calculated as follows:
β i = σ j = 1 k w i j C i j ,   C i j Ψ i k
among Ψ i k represents the set of k adjacent channels of C i , W k R k × d , d represents the number of channels, W k is set to:
w 1 , 1 w 1 , 1 0 0 0 0 w 2 , 2 w 2 , k + 1 0 0 0 0 0 w d , d k + 1 w d , d
The above k is a super parameter, which can be set manually or adaptively:
k = f ( d ) = log 2 γ + b γ o d d
Among them, γ is 2, and b is 1.

3.3.3. CVAE

The CVAE module in this paper consists of three parts: a front network p θ Z X t , an identification network q φ Z X t , Y t , and a generating network p ω c out X t , Z . θ , φ , ω is the model parameter, c o u t is the output of the CVAE module. The mean value of the prior network generation is μ z p , the variance is σ z p distribution N μ z p , σ z p , the mean value of the recognition network generation is μ z q , the variance is σ z q distribution N μ z q , σ z q . Since identifying the network requires real future trajectory data Y t , it can only be used in the training stage. When training, the model is distributed N μ z q , σ z q ; when testing, the model is distributed N μ z p , σ z p . In particular, the DAGConv structure is introduced into the CVAE generation network. The experimental results show that DAGConv improves the performance of CVAE greatly.

3.4. Multi-Recurrent Goal Mining

The objective of this module is to thoroughly extract trajectory features and generate multiple trajectory sequences for different goals. Since an agent’s trajectory at a given time is influenced by its past trajectory and destination, the trajectory can be seen as a continuous sequence. Therefore, this paper introduces a multi-recurrent structure, which includes an attention mechanism, a forward GRU, and a backward GRU. Unlike the bidirectional GRU, we introduce the input of the goal feedback module. Additionally, our forward GRU and backward GRU are concatenated, and the input of the backward GRU is related to the output of the forward GRU. The forward GRU uses the attention mechanism to process the output of the goal feedback module and the pre-input. The output of the forward GRU is then fed into the backward GRU in reverse order of time. The backward GRU uses the output of the forward GRU to initialize its hidden layer state, and its final output is obtained by combining the backward GRU output with the output of the forward GRU. By incorporating this multi-recurrent structure, the model can more effectively capture the temporal dependencies in the trajectory data and generate multiple trajectory sequences for different goals.
F out = G R U f Concat F C h f , Attention g m , h f
Attention g m = g m Softmax W m σ g m + b m
B out = G R U b Concat F C h b , h b
M out = F C Concat F out , B out
where, G R U f represents the forward GRU, G R U b represents the backward GRU, F o u t represents the output of the forward GRU, B o u t represents the output to the GRU, g m represents the output of the goal feedback module to this module, h f represents the hidden layer state of the forward GRU, h b represents the hidden layer state of the backward GRU, and M o u t indicates the output of this module.

3.5. Goals Aggregation

The module described above maps trajectory data to a high-dimensional space and generates multiple potential trajectories for the model to follow. Its function is to collect several candidate trajectories, map them back to a low-dimensional space, and produce the final prediction results. To accomplish this, the module combines the attention mechanism with DAGConv.
First, an adaptive method is employed to select the most suitable trajectories from among the many potential ones. Next, DAGConv is used to gather and merge these selected trajectories, eliminating any invalid ones and retaining only those that accurately represent the model’s intended goals. Finally, the resulting set of goal trajectories is used to generate the model’s predictions. The specific formula is:
M a t t = M out Softmax W ρ M out + b ρ
Y p r e d = DAGConv M a t t

3.6. Goal Feedback

The goal feedback module plays a crucial role in improving the trajectory prediction accuracy. By feeding back the generated trajectory goals to the model, the auxiliary model can better learn the trajectory sequence. Essentially, the goal feedback module functions as a memory backtracking module that feeds back the historical trajectory prediction results to aid in subsequent predictions. This feedback mechanism allows the model to capture similar trajectories and facilitates prediction.
As previously mentioned, the goal feedback module produces two outputs: one for the trajectory representation boosting module and another for the multi-recurrent goal mining module. However, the prediction goals generated by the model may contain errors that could accumulate over time and negatively impact prediction accuracy. To address this issue, this paper proposes an attention mechanism to filter out potentially problematic goals before feeding them back to the model. The attention mechanism works by evaluating the similarity between the predicted trajectory goals and the actual ground truth trajectories. Goals that deviate too far from the actual trajectory are deemed inaccurate and filtered out. This approach helps ensure that the feedback mechanism reinforces the model’s ability to predict accurate trajectories.
In summary, the goal feedback module is a vital component of the trajectory prediction model, facilitating improved learning and more accurate predictions. By incorporating an attention mechanism, the module can further enhance accuracy by filtering out inaccurate predictions and reinforcing accurate ones.

4. Experimental Results

4.1. Datasets

The dataset utilized in this paper is the publicly available ETH and UCY pedestrian trajectory dataset. These datasets are recorded from a third-person perspective and contain five subsets: ETH, HOTEL, UNIV, ZARA1, and ZARA2, featuring four unique scenes and a total of 1536 pedestrians. The dataset presents numerous challenging behaviors, such as couples walking together, group crossing, and formation and dispersion of groups. The data is collected at a frequency of 2.5 Hz, meaning that one frame of data is captured every 0.4 s.
To evaluate the model’s prediction performance, the paper adopts the leave-one-out method for data segmentation. This method involves using four of the datasets as training sets, while the remaining one is reserved for testing. For instance, to evaluate the model’s prediction performance for the HOTEL dataset, the ETH, UNIV, ZARA1, and ZARA2 subsets are utilized for training, while HOTEL is set aside for testing. This method ensures that the model is evaluated using all available data and that its predictive accuracy is robust across different datasets.

4.2. Experimental Settings

The experiments in this paper were all completed under the Linux server. The server configuration was CPU: Intel Xeon Gold 6226R × 2, GPU: NVIDIA TITAN RTX × 3.
This paper proposes that the hidden layer state of the GRU of the model is set to 512, the batch size is set to 128, and the training iteration is 50 times. The data of the past 8 time steps is used to predict the data of the next 12 time steps, and the number of samples is K = 20.

4.2.1. Evaluation Metrics

In this paper, the average distance error (ADE) and the final distance error (FDE) are used as the evaluation indicators of the model. The smaller the value of these two evaluation indicators, the better the prediction performance of the model.
A D E = t T pred y t ^ y t 2 T pred
F D E = y ^ t y t 2 , t = T pred
where, T p r e d indicates the predicted goal time length, y t ^ represents the predicted value of the track at the time t, and y t is the true value of the track at time t.

4.2.2. Baselines

This paper compares the proposed model with the following baselines:
  • Social-LSTM [13]: This model introduces a “social” pooling layer that allows LSTMs of spatially adjacent sequences to share their hidden states with each other.
  • SGAN [26]: This model combines sequence prediction and generating adversarial networks to predict trajectories.
  • Sophie [27]: This model utilizes two information sources, namely all path history and scene context information in the scene, and combines physical and social information using a social attention mechanism and physical attention.
  • Social-bigat [19]: This model is based on a graphical attention network, encoding reliable feature representations of social interactions between humans in the scene, and combining them with generative adversarial networks to generate multiple future trajectories.
  • RSBG [22]: A group-based social interaction model, which uses a graph convolution neural network to disseminate social interaction information in such a graph by recursively extracting social representation.
  • MATF GAN [42]: This model encodes the past trajectories and scene contexts of multiple agents into multi-agent tensors, and then applies convolutional fusion to capture multi-agent interactions while preserving the spatial structure and scene context of the agents.
  • PSA-GRU [17]: The model adopts a human social dual-attention network based on gated recursive units, fully utilizing important location nodes of personal historical trajectories and social information between pedestrians.
  • Social-STGCNN [20]: This model replaces aggregation methods by modeling interactions as graphs.
  • CGNS [43]: This model combines the advantages of conditional potential space learning and variable dispersity minimization, and uses an attention mechanism to utilize static context and interactive information.
  • PIF [44]: Adopting an end-to-end multitasking learning model that utilizes rich visual features about human behavior information and its interaction with the surrounding environment.
  • NMMP [28]: This model uses a separate branch to simulate the behavior of a single agent, an interaction branch to simulate the interaction between agents, and different wrappers to handle different input formats and features.
  • FvTraj [45]: This model is based on a multi-head attention mechanism and uses a social perception attention module to simulate social interaction between pedestrians, as well as a view perception attention module to capture the relationship between historical motion states and visual features.
  • DSCMP [46]: This model simulates dynamic interaction between agents by learning the spatial and temporal consistency of agents, as well as understanding the layout of contextual scenes. At the same time, a differentiable queue mechanism is designed, which can clarify the correlation between memory and learning long trajectories.
  • STGAT [23]: This model is based on a sequence-to-sequence architecture to predict the future trajectory of pedestrians. In addition to the spatial interaction captured by the graph attention mechanism at each time step, additional LSTM is also used to encode the temporal correlation of the interaction.
  • TPNet [47]: This model is divided into two stages to predict trajectories: first creating some suggested target trajectories, and then classifying and refining these trajectories to obtain the final predicted trajectory.

4.3. Experiment Results and Analysis

Table 1 presents the comparative experimental results of the proposed TFBNet and the baseline model for predicting the next 12 time steps on the ETH-UCY dataset. As shown in the table, the proposed model outperforms the baseline in the five datasets across both evaluation metrics. Compared to the baseline with the second-best prediction performance, the ADE index increased by 25% to 57%, the FDE index increased by 34% to 62%, and, on average, the ADE and FDE indices increased by 46% and 52%, respectively. These results demonstrate that the proposed model has a strong ability to predict the endpoint of trajectories, with the FDE index improvement being more pronounced than that of the ADE index.
Social-LSTM performed the worst among the 15 baseline models, while SGAN, which optimizes the social pool based on Social-LSTM, significantly improved the prediction performance. None of the baseline models outperformed the proposed model in all indicators on the five datasets. This indicates that the adaptability of the baseline models is not strong, underscoring the strength of the proposed model’s prediction performance.
It is worth noting that TPNet showed the best prediction performance on the HOTEL, UNIV, and ZARA2 datasets among the baseline models, but none of its average indicators were superior. This was due to its poor performance on the ETH dataset, which brought down its average prediction performance. Additionally, the average ADE index of the baseline models was mostly around 0.4 without breaking through, whereas the proposed model achieved a lower ADE index of 0.22, demonstrating a breakthrough in prediction performance.
Most existing models directly mine data from the original trajectory data. In contrast, our model first elevates the dimensionality of the trajectory data by mapping the original trajectory data into a high-dimensional space for exploration. This leads to better trajectory prediction performance compared to baseline models. Additionally, we designed the novel DAGConv module to extract trajectory features at different levels, which is applied in the goal aggregation and CVAE modules, enhancing the predictive performance. Furthermore, unlike previous models, we also considered the similarity between different object motion trajectories and designed the goal feedback module to capture similar trajectory features. This allows for improved trajectory prediction when similar trajectories exist.
Table 2 presents the comparative experimental results of the proposed TFBNet model and the baseline model on the ETH-UCY dataset to predict the next eight time steps. The results indicate that the proposed model outperforms the baseline in both evaluation indicators on all five datasets, with an average ADE increase of 44% and an FDE increase of 60% compared to the second-best model. The improvement in performance is slightly more significant than that of the next 12 steps, further highlighting the high prediction performance and strong adaptability of the proposed model.
Figure 3, Figure 4, Figure 5, Figure 6 and Figure 7 illustrate the prediction process of TFBNet, where K = 20. The blue line represents the observed track in the past, the green line represents the actual future track, and the red line represents the future track predicted by the model. The predicted trajectory of TFBNet shows relatively accurate final destinations and relatively divergent predicted future trajectories, which reflect the characteristics of multi-modal prediction well. This again demonstrates the good prediction performance of TFBNet.
To account for the uncertainty in an object’s future trajectory, this paper employs multiple trajectory prediction to simulate the diversity of potential outcomes. Following previous research, K is set to 20 samples, which equates to predicting the future 20 trajectories of the object. Notably, the choice of K can significantly affect the prediction results. To demonstrate this effect on the proposed model’s predictive performance, we conducted further experiments using different K values and prediction lengths ( T p r e d = 8 and T p r e d = 12 ) and performed a comparative analysis. The results are presented in Table 3 and Table 4. Interestingly, as K increases, the two metrics’ values of the experimental results for all five datasets decrease. This suggests that the proposed model’s predictive performance improves as K increases.
It is worth noting that the current research employs a verification method that compares the best prediction result of multiple trajectories with the actual trajectories. In theory, the more predicted trajectories, the better the prediction performance.

4.4. Ablation Experiments

To validate the effectiveness of each module in TFBNet, we conducted an ablation experiment on the ETH-UCY dataset, the results of which are presented in Table 5. The best result is in bold for each dataset. Specifically, we evaluated the performance of TFBNet with individual modules removed: TFBNet-DAGConv with the DAGConv structure removed, TFBNet-MR with the multi-recurrent goal mining module removed, TFBNet-GA with the goal aggregation module removed, and TFBNet-GF with the goal feedback module removed.
The experimental results demonstrate that each module proposed in this paper can improve the model’s predictive performance, with the DAGConv structure having the greatest impact on model performance. This suggests that the DAGConv structure is highly effective in improving the model’s predictive performance. The TFBNet-GA and TFBNet-MR models exhibit similar predictive performance, indicating that the goal aggregation and multi-recurrent goal mining modules have a similar degree of impact on the model. Notably, while the impact of each module on the average displacement error (ADE) is relatively low, their impact on the final displacement error (FDE) is relatively high. This suggests that the models proposed in this paper are especially useful in improving the model’s predictive performance for the end of the trajectory.

5. Discussion

A large number of experimental results conducted on five real pedestrian trajectory datasets show that the model proposed in this paper outperforms the baseline model in all indicators. This paper also conducted ablation experiments to verify the effectiveness of each module of the proposed model. At the same time, this paper also conducted analysis experiments on the key parameter K, and the experimental results also indicated that our proposed model has good robustness. Therefore, this paper proposes a model that can effectively predict pedestrian trajectories.

6. Conclusions

In this paper, we proposed a novel trajectory prediction network called the trajectory feature-boosting network (TFBNet). By mapping original trajectory data to high-dimensional space and learning the change rule of the trajectory, we provide a new approach for predicting time-series data, such as trajectories. To achieve this, we designed several modules, including a trajectory feature-boosting module, a conditional variational autoencoder (CVAE) module, a multi-recurrent goal mining module, a goal aggregation module, and a goal feedback module. We also introduced a dilated attention gating convolution (DAGConv) structure, which we applied in the CVAE module and goal aggregation module. Our experimental results show that DAGConv can improve the model’s data mining ability and significantly enhance its prediction performance. Our experiments, conducted on five real datasets, demonstrate that the proposed model’s prediction performance is significantly better than that of state-of-the-art methods.
In conclusion, our proposed TFBNet provides a promising approach for predicting time-series data, such as trajectories. Our findings also suggest that the individual modules we designed, including DAGConv, can contribute to enhancing the predictive performance of the model. Overall, our proposed model outperforms state-of-the-art methods, highlighting the potential of our approach in real-world applications.

Author Contributions

Conceptualization, Q.N. and W.P.; formal analysis, W.P., Y.Z. and R.Y.; funding acquisition, Q.N.; methodology, W.P.; project administration, Q.N., W.P., Y.Z. and R.Y.; software, W.P.; supervision, Q.N.; validation, W.P.; visualization, W.P.; writing—original draft, W.P.; writing—review and editing, Q.N., W.P., Y.Z. and R.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (12273003).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The dataset used in this paper is a publicly available real trajectory dataset, and the dataset link is as follows: ETH dataset: https://icu.ee.ethz.ch/research/datsets.html (accessed on 5 January 2022). UCY dataset: https://graphics.cs.ucy.ac.cy/research/downloads/crowd-data (accessed on 5 January 2022).

Acknowledgments

The authors would like to acknowledge the reviewers for their valuable feedback on the draft to improve the flow and quality of the work.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Yurtsever, E.; Lambert, J.; Carballo, A.; Takeda, K. A survey of autonomous driving: Common practices and emerging technologies. IEEE Access 2020, 8, 58443–58469. [Google Scholar] [CrossRef]
  2. Garcia, E.; Jimenez, M.A.; De Santos, P.G.; Armada, M. The evolution of robotics research. IEEE Robot. Autom. Mag. 2007, 14, 90–103. [Google Scholar] [CrossRef]
  3. Valera, M.; Velastin, S.A. Intelligent distributed surveillance systems: A review. IEE Proc.-Vis. Image Signal Process. 2005, 152, 192–204. [Google Scholar] [CrossRef] [Green Version]
  4. Wang, J.M.; Fleet, D.J.; Hertzmann, A. Gaussian process dynamical models for human motion. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 30, 283–298. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Kitani, K.M.; Ziebart, B.D.; Bagnell, J.A.; Hebert, M. Activity forecasting. In Computer Vision—ECCV 2012: 12th European Conference on Computer Vision, Florence, Italy, 7–13 October 2012; Springer: Berlin/Heidelberg, Germany, 2012; pp. 201–214. [Google Scholar]
  6. Helbing, D.; Molnar, P. Social force model for pedestrian dynamics. Phys. Rev. E 1995, 51, 4282. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Mehran, R.; Oyama, A.; Shah, M. Abnormal crowd behavior detection using social force model. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 935–942. [Google Scholar]
  8. Li, R.; Qin, Y.; Wang, J.; Wang, H. AMGB: Trajectory prediction using attention-based mechanism GCN-BiLSTM in IOV. Pattern Recognit. Lett. 2023, 169, 17–27. [Google Scholar] [CrossRef]
  9. Zhang, X.; Angeloudis, P.; Demiris, Y. Dual-branch Spatio-Temporal Graph Neural Networks for Pedestrian Trajectory Prediction. Pattern Recognit. 2023, 142, 109633. [Google Scholar] [CrossRef]
  10. Hui, F.; Wei, C.; ShangGuan, W.; Ando, R.; Fang, S. Deep encoder–Cdecoder-NN: A deep learning-based autonomous vehicle trajectory prediction and correction model. Phys. A Stat. Mech. Its Appl. 2022, 593, 126869. [Google Scholar] [CrossRef]
  11. Shi, K.; Wu, Y.; Shi, H.; Zhou, Y.; Ran, B. An integrated car-following and lane changing vehicle trajectory prediction algorithm based on a deep neural network. Phys. A Stat. Mech. Its Appl. 2022, 599, 127303. [Google Scholar] [CrossRef]
  12. Morton, J.; Wheeler, T.A.; Kochenderfer, M.J. Analysis of recurrent neural networks for probabilistic modeling of driver behavior. IEEE Trans. Intell. Transp. Syst. 2016, 18, 1289–1298. [Google Scholar] [CrossRef]
  13. Alahi, A.; Goel, K.; Ramanathan, V.; Robicquet, A.; Fei-Fei, L.; Savarese, S. Social lstm: Human trajectory prediction in crowded spaces. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 961–971. [Google Scholar]
  14. Vemula, A.; Muelling, K.; Oh, J. Social attention: Modeling attention in human crowds. In Proceedings of the 2018 IEEE International Conference on Robotics and Automation (ICRA), Brisbane, QLD, Australia, 21–25 May 2018; pp. 4601–4607. [Google Scholar]
  15. Fernando, T.; Denman, S.; Sridharan, S.; Fookes, C. Soft+ hardwired attention: An lstm framework for human trajectory prediction and abnormal event detection. Neural Netw. 2018, 108, 466–478. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Lee, N.; Choi, W.; Vernaza, P.; Choy, C.B.; Torr, P.H.; Chandraker, M. Desire: Distant future prediction in dynamic scenes with interacting agents. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 336–345. [Google Scholar]
  17. Yan, X.; Yang, J.; Song, L.; Liu, Y. PSA-GRU: Modeling Person-Social Twin-Attention Based on GRU for Pedestrian Trajectory Prediction. In Proceedings of the 2021 40th Chinese Control Conference (CCC), Shanghai, China, 26–28 July 2021; pp. 8151–8157. [Google Scholar]
  18. Bilro, P.; Barata, C.; Marques, J.S. Pedestrian Trajectory Prediction Using LSTM and Sparse Motion Fields. In Proceedings of the 2022 26th International Conference on Pattern Recognition (ICPR), Montreal, QC, Canada, 21–25 August 2022; pp. 3529–3535. [Google Scholar]
  19. Kosaraju, V.; Sadeghian, A.; Martín-Martín, R.; Reid, I.; Rezatofighi, H.; Savarese, S. Social-bigat: Multimodal trajectory forecasting using bicycle-gan and graph attention networks. Adv. Neural Inf. Process. Syst. 2019, 32, 137–146. [Google Scholar]
  20. Mohamed, A.; Qian, K.; Elhoseiny, M.; Claudel, C. Social-STGCNN: A Social Spatio-Temporal Graph Convolutional Neural Network for Human Trajectory Prediction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  21. Ivanovic, B.; Pavone, M. The trajectron: Probabilistic multi-agent trajectory modeling with dynamic spatiotemporal graphs. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 2375–2384. [Google Scholar]
  22. Sun, J.; Jiang, Q.; Lu, C. Recursive social behavior graph for trajectory prediction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 660–669. [Google Scholar]
  23. Huang, Y.; Bi, H.; Li, Z.; Mao, T.; Wang, Z. Stgat: Modeling spatial-temporal interactions for human trajectory prediction. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 6272–6281. [Google Scholar]
  24. Li, J.; Ma, H.; Zhang, Z.; Tomizuka, M. Social-wagdat: Interaction-aware trajectory prediction via wasserstein graph double-attention network. arXiv 2020, arXiv:2002.06241. [Google Scholar]
  25. Shi, L.; Wang, L.; Long, C.; Zhou, S.; Zhou, M.; Niu, Z.; Hua, G. SGCN: Sparse graph convolution network for pedestrian trajectory prediction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 8994–9003. [Google Scholar]
  26. Gupta, A.; Johnson, J.; Fei-Fei, L.; Savarese, S.; Alahi, A. Social gan: Socially acceptable trajectories with generative adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 2255–2264. [Google Scholar]
  27. Sadeghian, A.; Kosaraju, V.; Sadeghian, A.; Hirose, N.; Rezatofighi, H.; Savarese, S. Sophie: An attentive gan for predicting paths compliant to social and physical constraints. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 1349–1358. [Google Scholar]
  28. Hu, Y.; Chen, S.; Zhang, Y.; Gu, X. Collaborative motion prediction via neural motion message passing. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 6319–6328. [Google Scholar]
  29. Yao, Y.; Atkins, E.; Johnson-Roberson, M.; Vasudevan, R.; Du, X. Bitrap: Bi-directional pedestrian trajectory prediction with multi-modal goal estimation. IEEE Robot. Autom. Lett. 2021, 6, 1463–1470. [Google Scholar] [CrossRef]
  30. Wang, C.; Wang, Y.; Xu, M.; Crandall, D.J. Stepwise goal-driven networks for trajectory prediction. IEEE Robot. Autom. Lett. 2022, 7, 2716–2723. [Google Scholar] [CrossRef]
  31. Mangalam, K.; Girase, H.; Agarwal, S.; Lee, K.H.; Adeli, E.; Malik, J.; Gaidon, A. It is not the journey but the destination: Endpoint conditioned trajectory prediction. In Computer Vision—ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Springer: Cham, Switzerland, 2020; pp. 759–776. [Google Scholar]
  32. Salzmann, T.; Ivanovic, B.; Chakravarty, P.; Pavone, M. Trajectron++: Dynamically-feasible trajectory forecasting with heterogeneous data. In Computer Vision—ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Springer: Cham, Switzerland, 2020; pp. 683–700. [Google Scholar]
  33. Yuan, Y.; Weng, X.; Ou, Y.; Kitani, K.M. Agentformer: Agent-aware transformers for socio-temporal multi-agent forecasting. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Nashville, TN, USA, 20–25 June 2021; pp. 9813–9823. [Google Scholar]
  34. Yu, C.; Ma, X.; Ren, J.; Zhao, H.; Yi, S. Spatio-temporal graph transformer networks for pedestrian trajectory prediction. In Computer Vision—ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Springer: Cham, Switzerland, 2020; pp. 507–523. [Google Scholar]
  35. Xia, B.; Wong, C.; Peng, Q.; Yuan, W.; You, X. CSCNet: Contextual semantic consistency network for trajectory prediction in crowded spaces. Pattern Recognit. 2022, 126, 108552. [Google Scholar] [CrossRef]
  36. Mangalam, K.; An, Y.; Girase, H.; Malik, J. From goals, waypoints & paths to long term human trajectory forecasting. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Nashville, TN, USA, 20–25 June 2021; pp. 15233–15242. [Google Scholar]
  37. Yue, J.; Manocha, D.; Wang, H. Human trajectory prediction via neural social physics. In Proceedings of the European Conference on Computer Vision; Springer: Cham, Switzerland, 2022; pp. 376–394. [Google Scholar]
  38. Zhao, H.; Gao, J.; Lan, T.; Sun, C.; Sapp, B.; Varadarajan, B.; Shen, Y.; Shen, Y.; Chai, Y.; Schmid, C.; et al. Tnt: Target-driven trajectory prediction. In Proceedings of the Conference on Robot Learning, PMLR, London, UK, 8–11 November 2021; pp. 895–904. [Google Scholar]
  39. Rhinehart, N.; McAllister, R.; Kitani, K.; Levine, S. Precog: Prediction conditioned on goals in visual multi-agent settings. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 2821–2830. [Google Scholar]
  40. Xu, C.; Mao, W.; Zhang, W.; Chen, S. Remember Intentions: Retrospective-Memory-based Trajectory Prediction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 6488–6497. [Google Scholar]
  41. Wang, Q.; Wu, B.; Zhu, P.; Li, P.; Zuo, W.; Hu, Q. ECA-Net: Efficient Channel Attention for Deep Convolutional Neural Networks. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 11531–11539. [Google Scholar]
  42. Zhao, T.; Xu, Y.; Monfort, M.; Choi, W.; Baker, C.; Zhao, Y.; Wang, Y.; Wu, Y.N. Multi-agent tensor fusion for contextual trajectory prediction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 12126–12134. [Google Scholar]
  43. Li, J.; Ma, H.; Tomizuka, M. Conditional generative neural system for probabilistic trajectory prediction. In Proceedings of the 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Macau, China, 3–8 November 2019; pp. 6150–6156. [Google Scholar]
  44. Liang, J.; Jiang, L.; Niebles, J.C.; Hauptmann, A.G.; Fei-Fei, L. Peeking into the future: Predicting future person activities and locations in videos. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 5725–5734. [Google Scholar]
  45. Bi, H.; Zhang, R.; Mao, T.; Deng, Z.; Wang, Z. How can i see my future? fvtraj: Using first-person view for pedestrian trajectory prediction. In Proceedings of the Computer Vision—ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020, Proceedings, Part VII 16; Springer: Cham, Switzerland, 2020; pp. 576–593. [Google Scholar]
  46. Tao, C.; Jiang, Q.; Duan, L.; Luo, P. Dynamic and static context-aware lstm for multi-agent motion prediction. In Proceedings of the European Conference on Computer Vision; Springer: Cham, Switzerland, 2020; pp. 547–563. [Google Scholar]
  47. Fang, L.; Jiang, Q.; Shi, J.; Zhou, B. Tpnet: Trajectory proposal network for motion prediction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 6797–6806. [Google Scholar]
Figure 1. Model architecture of the proposed TFBNet.
Figure 1. Model architecture of the proposed TFBNet.
Entropy 25 01100 g001
Figure 2. The architecture of DAGConv.
Figure 2. The architecture of DAGConv.
Entropy 25 01100 g002
Figure 3. TFBNet’s prediction results on ETH.
Figure 3. TFBNet’s prediction results on ETH.
Entropy 25 01100 g003
Figure 4. TFBNet’s prediction results on HOTEL.
Figure 4. TFBNet’s prediction results on HOTEL.
Entropy 25 01100 g004
Figure 5. TFBNet’s prediction results on UNIV.
Figure 5. TFBNet’s prediction results on UNIV.
Entropy 25 01100 g005
Figure 6. TFBNet’s prediction results on ZARA1.
Figure 6. TFBNet’s prediction results on ZARA1.
Entropy 25 01100 g006
Figure 7. TFBNet’s prediction results on ZARA2.
Figure 7. TFBNet’s prediction results on ZARA2.
Entropy 25 01100 g007
Table 1. Performance comparison of different models for trajectory prediction. K = 20 samples, T p r e d = 12. The best result is in bold, the second-best result is underlined, and “IMP” represents the relative improvement of the model performance proposed in this paper to the second-best result. ↓ means lower is better. The same for other tables.
Table 1. Performance comparison of different models for trajectory prediction. K = 20 samples, T p r e d = 12. The best result is in bold, the second-best result is underlined, and “IMP” represents the relative improvement of the model performance proposed in this paper to the second-best result. ↓ means lower is better. The same for other tables.
MethodADE/FDE ( 4.8 s ) ( m ) , Best of 20
ETHHOTELUNIVZARA1ZARA2Average
Social-LSTM1.09/2.350.79/1.760.67/1.400.47/1.000.56/1.170.72/1.54
SGAN0.81/1.520.72/1.610.60/1.260.34/0.690.42/0.840.58/1.18
Sophie0.70/1.430.76/1.670.54/1.240.30/0.630.38/0.780.54/1.15
Social-bigat0.69/1/290.49/1.010.55/1.320.30/0.620.36/0.750.48/1.00
RSBG0.80/1.530.33/0.640.59/1.250.40/0.860.30/0.650.48/0.99
MATF GAN1.01/1.750.43/0.800.44 0.910.26/0.450.26/0.570.48/0.90
PSA-GRU0.79/1.630.52/1.070.53/1.130.41/0.770.34/0.740.52/1.07
Social-STGCNN0.64/1.110.49/0.850.44/0.790.34/0.530.30/0.480.44/0.75
CGNS0.62/1.400.70/0.930.48/1.220.32/0.590.35/0.710.49/0.97
PIF0.73/1.650.30/0.590.60/1.270.38/0.810.31/0.680.46/1.00
NMMP0.61/1.080.33/0.630.52/1.110.32/0.660.43/0.850.41/0.82
FvTraj0.56/1.140.28/0.550.52/1.120.37/0.780.32/0.680.41/0.85
DSCMP0.66/1.210.27/0.460.50/1.070.33/0.680.28/0.600.41/0.80
STGAT0.65/1.120.35/0.660.52/1.100.34/0.690.29/0.600.43/0.83
TPNet0.84/1.730.24/0.460.42/0.940.33/0.750.26/0.600.42/0.90
TFBNet ( o u r s ) 0.42/0.580.15/0.250.29/0.520.15/0.250.11/0.180.22/0.36
IMP ( % ) 25%/46%37%/45%30%/34%42%/44%57%/62%46%/52%
Table 2. Performance comparison of different models for trajectory prediction. T p r e d = 8.
Table 2. Performance comparison of different models for trajectory prediction. T p r e d = 8.
MethodADE/FDE ( 3.2 s ) ( m ) , Best of 20
ETHHOTELUNIVZARA1ZARA2Average
Social-LSTM0.73/1.480.49/1.010.41/0.840.27/0.560.33/0.700.45/0.91
SGAN0.61/1.220.48/0.950.36/0.750.21/0.420.27/0.540.39/0.78
PSA-GRU0.58/1.170.44/0.870.33/0.690.25/0.400.22/0.460.36/0.72
STGAT0.56/1.100.27/0.500.32/0.660.21/0.420.20/0.400.31/0.62
TPNet0.54/1.120.19/0.370.24/0.530.19/0.410.16/0.360.27/0.56
TFBNet ( o u r s ) 0.29/0.380.10/0.150.19/0.320.09/0.150.07/0.110.15/0.22
IMP46%/65%47%/59%20%/39%52%/62%56%/69%44%/60%
Table 3. Experimental results of different K values, T p r e d = 8 .
Table 3. Experimental results of different K values, T p r e d = 8 .
MetricDatasetK = 1K = 5K = 10K = 15K = 20K = 160
ADEETH0.580.380.330.320.290.20
HOTEL0.250.170.150.120.100.07
UNIV0.420.270.240.210.180.13
ZARA10.240.160.120.120.090.07
ZARA20.160.110.090.080.070.05
FDEDatasetK = 1K = 5K = 10K = 15K = 20K = 160
ETH1.010.630.500.490.380.23
HOTEL0.450.300.250.180.150.07
UNIV0.790.490.430.380.320.18
ZARA10.480.290.210.200.150.09
ZARA20.330.210.150.130.110.07
Table 4. Experimental results of different K values, T p r e d = 12 .
Table 4. Experimental results of different K values, T p r e d = 12 .
MetricDatasetK = 1K = 5K = 10K = 15K = 20K = 160
ADEETH0.850.580.490.470.420.31
HOTEL0.390.270.230.180.150.11
UNIV0.640.420.370.330.290.21
ZARA10.390.250.20.190.150.12
ZARA20.270.180.140.130.110.09
FDEDatasetK = 1K = 5K = 10K = 15K = 20K = 160
ETH1.621.050.810.750.580.38
HOTEL0.870.570.490.330.250.14
UNIV1.250.80.710.620.520.29
ZARA10.820.470.360.330.250.14
ZARA20.580.340.260.230.180.11
Table 5. Ablation experiments results.
Table 5. Ablation experiments results.
MethodETHHOTELUNIVZARA1ZARA2Average
TFBNet-DAGConv0.49/0.800.19/0.360.31/0.570.18/0.310.13/0.240.26/0.46
TFBNet-GA0.43/0.640.19/0.360.34/0.600.16/0.270.13/0.230.25/0.42
TFBNet-MR0.46/0.710.19/0.370.31/0.570.18/0.310.11/0.180.25/0.43
TFBNet-GF0.43/0.670.18/0.360.31/0.560.16/0.260.11/0.190.24/0.41
TFBNet0.42/0.580.15/0.250.29/0.520.15/0.250.11/0.180.22/0.36
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ni, Q.; Peng, W.; Zhu, Y.; Ye, R. A Novel Trajectory Feature-Boosting Network for Trajectory Prediction. Entropy 2023, 25, 1100. https://doi.org/10.3390/e25071100

AMA Style

Ni Q, Peng W, Zhu Y, Ye R. A Novel Trajectory Feature-Boosting Network for Trajectory Prediction. Entropy. 2023; 25(7):1100. https://doi.org/10.3390/e25071100

Chicago/Turabian Style

Ni, Qingjian, Wenqiang Peng, Yuntian Zhu, and Ruotian Ye. 2023. "A Novel Trajectory Feature-Boosting Network for Trajectory Prediction" Entropy 25, no. 7: 1100. https://doi.org/10.3390/e25071100

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop