Advances in Mathematical Physics (Jan 2024)

Constructing Attention-LSTM-VAE Power Load Model Based on Multiple Features

  • Chaoyue Ma,
  • Ying Wang,
  • Feng Li,
  • Huiyan Zhang,
  • Yong Zhang,
  • Haiyan Zhang

DOI
https://doi.org/10.1155/2024/1041791
Journal volume & issue
Vol. 2024

Abstract

Read online

With the complexity of modern power system and the susceptibility to external weather influences, it brings challenges to build an accurate load model. This paper proposes a variational autoencoder (VAE) long short-term memory (LSTM) load model based on the attention mechanism (Attention). First, the Prophet data decomposition method is used to decompose long sequences of load data at multiple time scales. Second, the correlation-based feature selection with maximum information coefficient (CFS-MIC) method is employed to select weather features based on their relevance, a subset of features with high correlation and low redundancy is chosen as model inputs. Finally, the Attention-LSTM-VAE model is constructed to capture the temporal variations laws of load. The dataset includes 2 years of load values and weather data collected in Caojiaping, Hunan Province, China. The experimental results show that the Attention-LSTM-VAE model has the lowest mean absolute error of 0.0374 and the highest R-squared value of 0.9714, verifying the accuracy of the model. Therefore, the performance of the Attention-LSTM-VAE model is better than the general deep learning load models, which has important reference for the research of power load models. Comparisons with other deep learning methods, the experimental results show that the Attention-LSTM-VAE model has the lowest mean absolute error of 0.0374 and the highest R-squared value of 0.9714. The Attention-LSTM-VAE has better robustness, stability, and accuracy in load modeling, which has an important reference for the research of power load models.