TY - GEN
T1 - Convolutional Spatio-Temporal Prediction Network with Trainable Positional Encoding
AU - Xu, Yuan
AU - Zhang, Yi Zhou
AU - Sun, Da Zhi
AU - Li, Kai
AU - Zhu, Qun Xiong
AU - Ke, Wei
AU - Zhang, Yang
AU - Zhang, Ming Qing
N1 - Publisher Copyright:
© 2025 IEEE.
PY - 2025
Y1 - 2025
N2 - Existing spatio-temporal prediction networks that rely on recurrent neural networks face significant parallelization challenges, leading to high computational costs and prolonged training durations. In contrast, recurrent-free methods, particularly convolutional neural networks (CNNs), have attracted significant attention because of their lightweight models and straightforward architectures. Traditional CNNs, however, are limited by their inherent weight-sharing property, which imposes a strong inductive bias and restricts their ability to capture spatial heterogeneity. To overcome these challenges, we propose a novel trainable positional encoding convolutional neural network (TPE-CNN). TPE-CNN incorporates a positional encoding module to dynamically learn location-specific spatio-temporal correlations and temporal evolution patterns, addressing the inability of CNNs to handle absolute positional dependencies effectively. Additionally, we integrate large-kernel convolutions and inception modules, significantly enhancing global information awareness. This design enables the incorporation of global attention mechanisms while maintaining a low parameter count, achieving an optimal balance between performance and computational efficiency. To validate the effectiveness of TPE-CNN, we performed extensive experiments on three real-world datasets. Experimental results demonstrate that TPE-CNN delivers superior prediction accuracy while maintaining low computational complexity.
AB - Existing spatio-temporal prediction networks that rely on recurrent neural networks face significant parallelization challenges, leading to high computational costs and prolonged training durations. In contrast, recurrent-free methods, particularly convolutional neural networks (CNNs), have attracted significant attention because of their lightweight models and straightforward architectures. Traditional CNNs, however, are limited by their inherent weight-sharing property, which imposes a strong inductive bias and restricts their ability to capture spatial heterogeneity. To overcome these challenges, we propose a novel trainable positional encoding convolutional neural network (TPE-CNN). TPE-CNN incorporates a positional encoding module to dynamically learn location-specific spatio-temporal correlations and temporal evolution patterns, addressing the inability of CNNs to handle absolute positional dependencies effectively. Additionally, we integrate large-kernel convolutions and inception modules, significantly enhancing global information awareness. This design enables the incorporation of global attention mechanisms while maintaining a low parameter count, achieving an optimal balance between performance and computational efficiency. To validate the effectiveness of TPE-CNN, we performed extensive experiments on three real-world datasets. Experimental results demonstrate that TPE-CNN delivers superior prediction accuracy while maintaining low computational complexity.
KW - Convolutional Neural Networks
KW - Spatio-temporal Prediction
KW - Trainable Positional Encoding
UR - https://www.scopus.com/pages/publications/105013958196
U2 - 10.1109/CCDC65474.2025.11090881
DO - 10.1109/CCDC65474.2025.11090881
M3 - Conference contribution
AN - SCOPUS:105013958196
T3 - Proceedings of the 37th Chinese Control and Decision Conference, CCDC 2025
SP - 406
EP - 411
BT - Proceedings of the 37th Chinese Control and Decision Conference, CCDC 2025
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 37th Chinese Control and Decision Conference, CCDC 2025
Y2 - 16 May 2025 through 19 May 2025
ER -