Improving the Bi-LSTM model with XGBoost and attention mechanism: A combined approach for short-term power load prediction.

Autor: Dai, Yeming, Zhou, Qiong, Leng, Mingming, Yang, Xinyu, Wang, Yanxin
Předmět:
Zdroj: Applied Soft Computing; Nov2022, Vol. 130, pN.PAG-N.PAG, 1p
Abstrakt: Short term power load forecasting plays an important role in the management and development of power systems with a focus on the reduction in power wastes and economic losses. In this paper, we construct a novel, short-term power load forecasting method by improving the bidirectional long short-term memory (Bi-LSTM) model with Extreme Gradient Boosting (XGBoost) and Attention mechanism. Our model differs from existing methods in the following three aspects. First, we use the weighted grey relational projection algorithm to distinguish the holidays and non-holidays in the data preprocessing. Secondly, we add the Attention mechanism to the Bi-LSTM model to improve the validity and accuracy of prediction. Thirdly, XGBoost is a newly-developed, well-performing prediction model, which is used together with the Attention mechanism to optimize the Bi-LSTM model. Therefore, we develop a novel, combined power load prediction model "Attention-Bi-LSTM + XGBoost" with the weight determination theory-error reciprocal method. Using two power market datasets, we evaluate our prediction method by comparing it with two benchmark models and four other models. With our prediction method, the MAPE, MAE, and RMSE for the Singapore's power market are 0.387, 43.206, and 54.357, respectively; and those for the Norway's power market are 0.682, 96.278, and 125.343, respectively. The test results are smaller than the results for six other models. This indicates that our prediction method outperforms the LSTM, Bi-LSTM, Attention-RNN, Attention-LSTM, Attention-Bi-LSTM, and XGBoost in effectiveness, accuracy, and practicability. [ABSTRACT FROM AUTHOR]
Databáze: Supplemental Index