Research on Time Series Forecasting Models Based on Hybrid Attention Mechanism and Graph Neural Networks
Abstract
In wireless data transmission, packet loss and missing data caused by environmental interference and network congestion significantly impact the stability of time series. To address these challenges, this study proposes a time series forecasting model named FGDLNet. FGDLNet is based on the Transformer architecture and integrates Graph Neural Networks (GNN) to enhance the performance of long sequence predictions, particularly in handling complex time series patterns. The model simplifies its structure and reduces computational complexity by removing the Decoder module from the traditional Transformer and replacing it with a linear layer for direct connection and prediction. To enhance the feature extraction capability of time series data, FGDLNet incorporates a multi-scale feature extraction module that extracts features at different temporal scales using multiple convolution kernels in parallel. Specifically, the model employs a single-channel processing approach to reduce interference between features and improve prediction accuracy. The introduced GNN module enables feature propagation and enhancement within the single channel, better capturing short-term fluctuations and long-term trends. In terms of the attention mechanism, this study designs a hybrid attention mechanism that combines global linear attention and local window attention. The global linear attention optimizes the computation to improve the efficiency of capturing global contextual information, while the local window attention strengthens the model’s ability to handle short-term dependencies. To evaluate the effectiveness of the model, we selected a dataset recorded during the flight of a specific aircraft, which includes longitude, latitude, and inertial navigation parameters, and conducted long-term trend forecasting. In the experiments, we used MAE (Mean Absolute Error), MSE (Mean Squared Error), and training time to assess the model’s performance. The experimental results show that FGDLNet outperforms traditional models such as Autoformer, Transformer, Informer, Reformer, DLinear, and ITransformer in long-term forecasting tasks. Specifically, FGDLNet achieves the following MAE values: 0.1400, 0.0595, 0.0092, 0.0324, 0.0493, and 0.122, which are significantly lower than those of the other models. In terms of MSE, FGDLNet also demonstrates lower errors: 0.0231, 0.0584, 0.0987, 0.0825, 0.0798, and 0.1925. Additionally, FGDLNet’s training time per epoch is 1156.47 seconds, which is about 7% faster than the Transformer model (1243.14 seconds).
Full Text:
PDFReferences
Vaswani A. Attention is all you need(2023). Advances in Neural Information Processing Systems, NIPS'17: Proceedings of the 31st International Conference on Neural Information Processing Systems Pages 6000 - 6010.
https://doi.org/10.48550/arXiv.1706.03762
Zhou, H., Zhang, S., Peng, J., Zhang, S., Li, J., Xiong, H., and Zhang, W. (2021). Informer: Beyond efficient transformer for long sequence time-series forecasting. In Proceedings of the AAAI conference on artificial intelligence Vol. 35, No. 12, pp. 11106-11115.
https://doi.org/10.1609/aaai.v35i12.17325
Kitaev, Nikita ,Łukasz Kaiser, and Anselm Levskaya. (2020)."Reformer: The efficient transformer." arXiv preprint arXiv:2001.04451 .
https://doi.org/10.48550/arXiv.2001.04451
Wu, H., Xu, J., Wang, J., & Long, M. (2021). Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting. Advances in neural information processing systems, Vol.34, 22419-22430.
https://doi.org/10.48550/arXiv.2106.13008
Zeng A, Chen M, Zhang L, et al(2023). Are Transformers Effective for Time Series Forecasting? Proceedings of the AAAI Conference on Artificial Intelligence. Vol.37(9): 11121-11128.
Https://doi.org/10.1609/aaai.v37i9.26317
Liu Y, Hu T, Zhang H, et al(2023). iTransformer: Inverted Transformers are Effective for Time Series Forecasting. arXiv preprint arXiv:2310.06625, 2023.
https://doi.org/10.48550/arXiv.2310.06625
Xu, Zhijian, Ailing Zeng, and Qiang Xu (2023). "FITS: Modeling time series with $10 k $ parameters." .Montavon G, Orr G B, Müller K R. Neural Networks: Tricks of the Trade. arxiv preprint arxiv:2307.03756
https://doi.org/10.48550/arXiv.2307.03756
Cao, D., Jia, F., Arik, S. O., Pfister, T., Zheng, Y., Ye, W., & Liu, Y. (2023). Tempo: Prompt-based generative pre-trained transformer for time series forecasting. arxiv preprint arxiv:2310.04948.
https://doi.org/10.48550/arXiv.2310.04948
Yitian Zhang, Liheng Ma, Soumyasundar Pal, Yingxue Zhang, Mark Coates(2023). Multi-resolution Time-Series Transformer for Long-term Forecasting. arxiv preprint arXiv:2311.04147v2
https://doi.org/10.48550/arXiv.2311.04147
Arjun Ashok, Étienne Marcotte , Valentina Zantedeschi, Nicolas Chapados, Alexandre Drouin (2023). TACTiS-2: Better, Faster, Simpler Attentional Copulas for Multivariate Time Series.arxiv preprint arxiv:2310.01327.
https://doi.org/10.48550/arXiv.2310.01327
Yang, Z., Liu, L., Li, N., & Tian, J. (2022). Time series forecasting of motor bearing vibration based on informer. Sensors, 22(15), 5858.
https://doi.org/10.3390/s22155858
Gong, M., Zhao, Y., Sun, J., Han, C., Sun, G., & Yan, B. (2022). Load forecasting of district heating system based on Informer. Energy, 253, 124179.
https://doi.org/10.1016/j.energy.2022.124179
Pan, K., Lu, J., Li, J., & Xu, Z. (2023). A Hybrid Autoformer Network for Air Pollution Forecasting Based on External Factor Optimization. Atmosphere, 14(5), 869.
https://doi.org/10.3390/atmos14050869
Feng, H., & Zhang, X. (2023). A novel encoder-decoder model based on Autoformer for air quality index prediction. Plos one, 18(4), e0284293.
https://doi.org/10.1371/journal.pone.0284293
Huang, Y., & Wu, Y. (2023). Short-Term Photovoltaic Power Forecasting Based on a Novel Autoformer Model. Symmetry, 15(1), 238.
https://doi.org/10.3390/sym15010238
González Hernández, A. L. (2024). Time Series Forecasting Using Transformer-based Models: Exploring Transformer, Informer, and Autoformer. Becerra González, Juan Antonio .
https://hdl.handle.net/11441/163116
Frissou N, Kimour M T, Selmane S. Optimized Support Vector Regression for Predicting Leishmaniasis Incidences[J]. Informatica (Slovenia), 2021, 45(7).
https://doi.org/10.31449/inf.v45i7.3665
Liu Y, Pan B. Profit estimation model and financial risk prediction combining multi-scale convolutional feature extractor and BGRU model[J]. Informatica (Slovenia), 2024, 48(11).
https://doi.org/10.31449/inf.v48i11.5941
Wang T, Yu J, Cao Y, et al. Fault Prediction of CNC Machine Tools Based on Gutenberg-Richter Law and Fuzzy Neural Networks[J]. Informatica (Slovenia), 2024, 48(18).
https://doi.org/10.31449/inf.v48i18.6292
Flores A, Tito-Chura H. Multi-Step Forecasting of Guillain Barré Cases using Deep Learning[J]. Informatica (Slovenia), 2024, 48(20).
https://doi.org/10.31449/inf.v48i20.6358
DOI: https://doi.org/10.31449/inf.v49i21.7580

This work is licensed under a Creative Commons Attribution 3.0 License.