Research on Time Series Forecasting Models Based on Hybrid Attention Mechanism and Graph Neural Networks

YiRui Cheng, Guo Li, Xu Zhou, ShuHui Ye

Abstract


In wireless data transmission, packet loss and missing data caused by environmental interference and network congestion significantly impact the stability of time series. To address these challenges, this study proposes a time series forecasting model named FGDLNet. FGDLNet is based on the Transformer architecture and integrates Graph Neural Networks (GNN) to enhance the performance of long sequence predictions, particularly in handling complex time series patterns. The model simplifies its structure and reduces computational complexity by removing the Decoder module from the traditional Transformer and replacing it with a linear layer for direct connection and prediction. To enhance the feature extraction capability of time series data, FGDLNet incorporates a multi-scale feature extraction module that extracts features at different temporal scales using multiple convolution kernels in parallel. Specifically, the model employs a single-channel processing approach to reduce interference between features and improve prediction accuracy. The introduced GNN module enables feature propagation and enhancement within the single channel, better capturing short-term fluctuations and long-term trends. In terms of the attention mechanism, this study designs a hybrid attention mechanism that combines global linear attention and local window attention. The global linear attention optimizes the computation to improve the efficiency of capturing global contextual information, while the local window attention strengthens the model’s ability to handle short-term dependencies. To evaluate the effectiveness of the model, we selected a dataset recorded during the flight of a specific aircraft, which includes longitude, latitude, and inertial navigation parameters, and conducted long-term trend forecasting. In the experiments, we used MAE (Mean Absolute Error), MSE (Mean Squared Error), and training time to assess the model’s performance. The experimental results show that FGDLNet outperforms traditional models such as Autoformer, Transformer, Informer, Reformer, DLinear, and ITransformer in long-term forecasting tasks. Specifically, FGDLNet achieves the following MAE values: 0.1400, 0.0595, 0.0092, 0.0324, 0.0493, and 0.122, which are significantly lower than those of the other models. In terms of MSE, FGDLNet also demonstrates lower errors: 0.0231, 0.0584, 0.0987, 0.0825, 0.0798, and 0.1925. Additionally, FGDLNet’s training time per epoch is 1156.47 seconds, which is about 7% faster than the Transformer model (1243.14 seconds).


Full Text:

PDF

References


Vaswani A. Attention is all you need(2023). Advances in Neural Information Processing Systems, NIPS'17: Proceedings of the 31st International Conference on Neural Information Processing Systems Pages 6000 - 6010.

https://doi.org/10.48550/arXiv.1706.03762

Zhou, H., Zhang, S., Peng, J., Zhang, S., Li, J., Xiong, H., and Zhang, W. (2021). Informer: Beyond efficient transformer for long sequence time-series forecasting. In Proceedings of the AAAI conference on artificial intelligence Vol. 35, No. 12, pp. 11106-11115.

https://doi.org/10.1609/aaai.v35i12.17325

Kitaev, Nikita ,Łukasz Kaiser, and Anselm Levskaya. (2020)."Reformer: The efficient transformer." arXiv preprint arXiv:2001.04451 .

https://doi.org/10.48550/arXiv.2001.04451

Wu, H., Xu, J., Wang, J., & Long, M. (2021). Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting. Advances in neural information processing systems, Vol.34, 22419-22430.

https://doi.org/10.48550/arXiv.2106.13008

Zeng A, Chen M, Zhang L, et al(2023). Are Transformers Effective for Time Series Forecasting? Proceedings of the AAAI Conference on Artificial Intelligence. Vol.37(9): 11121-11128.

Https://doi.org/10.1609/aaai.v37i9.26317

Liu Y, Hu T, Zhang H, et al(2023). iTransformer: Inverted Transformers are Effective for Time Series Forecasting. arXiv preprint arXiv:2310.06625, 2023.

https://doi.org/10.48550/arXiv.2310.06625

Xu, Zhijian, Ailing Zeng, and Qiang Xu (2023). "FITS: Modeling time series with $10 k $ parameters." .Montavon G, Orr G B, Müller K R. Neural Networks: Tricks of the Trade. arxiv preprint arxiv:2307.03756

https://doi.org/10.48550/arXiv.2307.03756

Cao, D., Jia, F., Arik, S. O., Pfister, T., Zheng, Y., Ye, W., & Liu, Y. (2023). Tempo: Prompt-based generative pre-trained transformer for time series forecasting. arxiv preprint arxiv:2310.04948.

https://doi.org/10.48550/arXiv.2310.04948

Yitian Zhang, Liheng Ma, Soumyasundar Pal, Yingxue Zhang, Mark Coates(2023). Multi-resolution Time-Series Transformer for Long-term Forecasting. arxiv preprint arXiv:2311.04147v2

https://doi.org/10.48550/arXiv.2311.04147

Arjun Ashok, Étienne Marcotte , Valentina Zantedeschi, Nicolas Chapados, Alexandre Drouin (2023). TACTiS-2: Better, Faster, Simpler Attentional Copulas for Multivariate Time Series.arxiv preprint arxiv:2310.01327.

https://doi.org/10.48550/arXiv.2310.01327

Yang, Z., Liu, L., Li, N., & Tian, J. (2022). Time series forecasting of motor bearing vibration based on informer. Sensors, 22(15), 5858.

https://doi.org/10.3390/s22155858

Gong, M., Zhao, Y., Sun, J., Han, C., Sun, G., & Yan, B. (2022). Load forecasting of district heating system based on Informer. Energy, 253, 124179.

https://doi.org/10.1016/j.energy.2022.124179

Pan, K., Lu, J., Li, J., & Xu, Z. (2023). A Hybrid Autoformer Network for Air Pollution Forecasting Based on External Factor Optimization. Atmosphere, 14(5), 869.

https://doi.org/10.3390/atmos14050869

Feng, H., & Zhang, X. (2023). A novel encoder-decoder model based on Autoformer for air quality index prediction. Plos one, 18(4), e0284293.

https://doi.org/10.1371/journal.pone.0284293

Huang, Y., & Wu, Y. (2023). Short-Term Photovoltaic Power Forecasting Based on a Novel Autoformer Model. Symmetry, 15(1), 238.

https://doi.org/10.3390/sym15010238

González Hernández, A. L. (2024). Time Series Forecasting Using Transformer-based Models: Exploring Transformer, Informer, and Autoformer. Becerra González, Juan Antonio .

https://hdl.handle.net/11441/163116

Frissou N, Kimour M T, Selmane S. Optimized Support Vector Regression for Predicting Leishmaniasis Incidences[J]. Informatica (Slovenia), 2021, 45(7).

https://doi.org/10.31449/inf.v45i7.3665

Liu Y, Pan B. Profit estimation model and financial risk prediction combining multi-scale convolutional feature extractor and BGRU model[J]. Informatica (Slovenia), 2024, 48(11).

https://doi.org/10.31449/inf.v48i11.5941

Wang T, Yu J, Cao Y, et al. Fault Prediction of CNC Machine Tools Based on Gutenberg-Richter Law and Fuzzy Neural Networks[J]. Informatica (Slovenia), 2024, 48(18).

https://doi.org/10.31449/inf.v48i18.6292

Flores A, Tito-Chura H. Multi-Step Forecasting of Guillain Barré Cases using Deep Learning[J]. Informatica (Slovenia), 2024, 48(20).

https://doi.org/10.31449/inf.v48i20.6358




DOI: https://doi.org/10.31449/inf.v49i21.7580

Creative Commons License
This work is licensed under a Creative Commons Attribution 3.0 License.