Next Article in Journal
Trend Lines and Japanese Candlesticks Applied to the Forecasting of Wind Speed Data Series
Next Article in Special Issue
Irradiance Nowcasting by Means of Deep-Learning Analysis of Infrared Images
Previous Article in Journal
Hybrid Surrogate Model for Timely Prediction of Flash Flood Inundation Maps Caused by Rapid River Overflow
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Short Term Electric Power Load Forecasting Using Principal Component Analysis and Recurrent Neural Networks

by
Venkataramana Veeramsetty
1,*,
Dongari Rakesh Chandra
2,
Francesco Grimaccia
3 and
Marco Mussetta
3
1
Center for Artificial Intelligence and Deep Learning, Department of Electrical and Electronics Engineering, SR University, Warangal 506371, India
2
Department of Electrical and Electronics Engineering, Kakatiya Institute of Technology and Science (KITS), Warangal 506015, India
3
Department of Energy, Politecnico di Milano, Via Lambruschini 4, 20156 Milan, Italy
*
Author to whom correspondence should be addressed.
Submission received: 13 December 2021 / Revised: 18 January 2022 / Accepted: 18 January 2022 / Published: 24 January 2022
(This article belongs to the Special Issue Feature Papers of Forecasting 2022)

Abstract

:
Electrical load forecasting study is required in electric power systems for different applications with respect to the specific time horizon, such as optimal operations, grid stability, Demand Side Management (DSM) and long-term strategic planning. In this context, machine learning and data analytics models represent a valuable tool to cope with the intrinsic complexity and especially design future demand-side advanced services. The main novelty in this paper is that the combination of a Recurrent Neural Network (RNN) and Principal Component Analysis (PCA) techniques is proposed to improve the forecasting capability of the hourly load on an electric power substation. A historical dataset of measured loads related to a 33/11 kV MV substation is considered in India as a case study, in order to properly validate the designed method. Based on the presented numerical results, the proposed approach proved itself to accurately predict loads with a reduced dimensionality of input data, thus minimizing the overall computational effort.

1. Introduction

Nowadays, the energy system is facing a radical revolution towards a green transition, with increasing penetration of renewable energy sources (RES), migration to distributed systems, with new actors like prosumers, and storage integration, both utility scale and domestic, which represent a key technology to decouple energy production and consumption [1].
In this regard, distributed sensor architectures, digital technology, data analytics and computational tools would represent crucial enabling technologies for monitoring, forecasting and maintenance purposes, to better manage the balance between power demand and supply, and to improve embedding of distributed RES; additionally, for the particular case of stand-alone hybrid systems, energy forecasting will particularly help anticipating customers’ behavior, sizing the electrical infrastructure and improving overall system reliability [2]. Therefore, forecasting capability brings helpful insights for security of energy supply, supporting power companies in providing their end-users with advanced demand-side services, and safe and stable systems.
Utility companies have several advantages with accurate load forecasting, such as reduced operation and maintenance costs, optimized management of demand supply, increased system reliability, effective long-term strategic planning for future investments [3,4]. Electrical load forecasting can be generally divided into four main categories based on forecasting time, such as very short-term, short-term [5], medium-term and long-term load forecasting [6]. Load forecasting with different applications with respect to the specific time horizon, such as optimal operations [7], grid stability [8], Demand Side Management (DSM) [9] or long-term strategic planning [10].
On the other hand, with respect to short-term load forecasting, energy trading is another important task for utilities to successfully increase revenues on the day-ahead energy market models. Power wholesale markets have around the world many different mechanisms and day-ahead or infra-day sessions, e.g., in India two categories exist based on trading time, such as Hourly Ahead Market (HAM) and Day Ahead Market (DAM). In HAM, one hour before the time of energy use, energy trading will be open. Similarly, for DAM, one day before the time of energy use, energy can be traded [11].
A methodology was developed for short-Term load forecasting by combining Light Gradient Boosting Machine (LGBM), eXtreme Gradient Boosting machine (XGB) and Multi-Layer Perceptron (MLP) models in [12]. In this hybrid model both XGB-LGBM combining is used for meta-data generation. A multi-temporal-spatial-scale temporal convolutional network was used in [13] to predict active power load. The multi-temporal-spatial-scale technique is used to minimize noise in load data. A hybrid clustering-based deep learning methodology was developed in [14] for short-term load forecasting. Clustering technique was used to make different clusters of distribution transformers based on load profile. Markov-chain mixture distribution model is developed in [15] to predict the load of residential customers by 30 min ahead. A study was done for load forecasting using various machine learning models like SVM, Random Forest and LSTM [16] both individually and with a fusion prediction approach. Short-Term load forecasting was done using convolutional neural networks (CNN) and sequence models like LSTM and GRU in [17]. CNN was used for feature extraction and sequence models are used for load forecasting. A CNN and Deep Residual Network based machine learning model was developed in [18] for short-Term load forecasting. Various regression models along with correlation concept for dimensionality reduction were used for load forecasting in [19]. LSTM and factor analysis based deep learning model was developed in [20] for load foresting within a smart cities environment. Artificial neural network based machine learning models were developed both for photovoltaic power forecasting [21], and load forecasting on MV distribution networks [22]. Most of the papers on probabilistic renewable generation forecasting literature over the last ten years or so have focused on different variants of statistical and machine learning approaches: in [23] a comparison of non-parametric approaches to this probabilistic forecasting problem has been performed. All these methodologies in literature contributed significantly to face short-term electric power load forecasting problems. In order to improve the forecasting accuracy and also to build a light weight model for active power load forecasting applied to a 33/11 kV substation, a new approach was developed in this paper by using recurrent neural networks for load forecasting and Principal Component Analysis for dimensional reduction.
The novelty of the proposed approach consists in a hybrid approach combining the heterogeneous input structure with PCA: in particular, the new approach considers the temporal impact of the previous three hours data and three days at the same hour data, and the previous three weeks at the same hour data, thus enabling the model to predict load with good accuracy by properly capturing temporal resolution diversity (e.g., the weekends load pattern); additionally, PCA is able to extract the most essential features from the given nine input information, thus compacting the input layer and reducing computational load, maintaining the same overall accuracy. The combination of RNN and PCA is used for the first time in short-term load forecasting problem. RNN models were trained using self adaptive Adam optimizer as shown in [24]. Complete literature summary on short-term load forecasting domain with various machine learning approaches is presented in Table 1. All these methodologies provides valuable contribution towards short-term load forecasting but have some limitations like model complexity, accuracy and weekly impact not considered. In this paper, accuracy in load prediction is improved by tuning the RNN model parameters, model complexity reduced by using principle component analysis and weekly impact considered by using features like P ( h 168 ) , P ( h 336 ) and P ( h 504 ) .

2. Methodology

2.1. Dimensionality Reduction Using Principal Component Analysis (PCA)

Principal Component Analysis (PCA) uses the extraction features approach to compress the original dataset to a lower subspace feature, with the aim of maintaining most of the relevant information. Detailed procedure for most relevant feature extraction using PCA is drawn from [25].

2.2. Recurrent Neural Network (RNN)

The Recurrent Neural Network (RNN) is a network where the activation status of each hidden neuron for the previous input is used to calculate the activation status of the hidden neuron for the current input [26]. The main and most important feature of RNN is the Hidden state, which recalls some information about previous samples. This work produced four distinct models of RNN i.e., RNN-HAM-Model1 (RHM-1), RNN-HAM-Model2 (RHM-2), RNN-DAM-Model1 (RDM-1) and RNN-DAM-Model2 (RDM-2) to forecast power for effective energy trading in Hourly Ahead Market(HAM) and Day Ahead Market (DAM).
In this study, RHM-1 is designed to predict the load based on the last three hours of load, load at the same time for the last three days and loading at the same time but for the last three weeks. The architecture for the proposed RNN model is shown in Figure 1.
The PCA algorithm is applied to the input features of the load dataset to find the principal components. It has been observed from the CEVR that six principal components cover almost 90% of the load dataset variance. Thus, the 9 input features i.e., P ( h 1 ) , P ( h 2 ) , P ( h 3 ) , P ( h 24 ) , P ( h 48 ) , P ( h 72 ) , P ( h 168 ) , P ( h 336 ) and P ( h 504 ) in each dataset sample are replaced by the corresponding six principal components. These six principal components were used to train the RHM-2. RHM-2 was therefore designed with six input neurons and one output neuron. The architecture of the proposed RNN model is the same as shown in Figure 1, where the number of inputs ( N i ) is reduced to 6 by the PCA.
RDM-1 is designed to predict the load based on load at the time of forecasting for the last three days and load at forecast time but for the last three weeks. The architecture for the proposed RNN model is the same as shown in Figure 1, where only six input features are considered, i.e., P ( h 24 ) , P ( h 48 ) , P ( h 72 ) , P ( h 168 ) , P ( h 336 ) and P ( h 504 ) , thus N i = 6 .
The PCA algorithm is applied to the input features of the load dataset to find the principal components. Load dataset consists in total of 6 input features i.e., P ( h 24 ) , P ( h 48 ) , P ( h 72 ) , P ( h 168 ) , P ( h 336 ) and P ( h 504 ) , and one output feature P ( h ) . It has been observed from CEVR that four principal components cover almost 90% of the load dataset variance. Thus, six input features, i.e., P ( h 24 ) , P ( h 48 ) , P ( h 72 ) , P ( h 168 ) , P ( h 336 ) and P ( h 504 ) for each dataset sample, are converted into four principal components. These four principal components were used to train the RDM-2. RDM-2 was therefore designed with four input neurons and one output neuron. The architecture of the proposed RNN model is the same as shown in Figure 1, where the N i is finally reduced to 4 by the PCA. Table 2 resumes all this information about the analyzed RNN models, with respect to the considered architecture.
Trained RNN model can predict P ( h ) based on input (X) features using Equations (1) and (2). Performance of the all these RNN models have been observed in terms of error metrics like Mean Square Error (MSE), Mean Absolute Error (MAE) and Root Mean Square Error (RMSE) [27] as shown in Equations (3)–(5), respectively.
a < t > = t a n h ( W a x X + W a a a < t 1 > + b a )
P ( h ) = s i g m o i d ( W y a a < t > + b y )
M S E = 1 n i = 1 n ( Y i t r u e Y i p r e d ) 2
M A E = i = 1 n | Y i t r u e Y i p r e d | n
R M S E = 1 n i = 1 n ( Y i t r u e Y i p r e d ) 2
The complete work done in this paper is presented in Figure 2.

3. Result Analysis

Data was captured from [28] to train and test the models. This load data consists of a total of 2184 samples (91 days × 24 h) and these data are rearranged into a 1680, i.e., ( 2184 ( 3 ( W ) × 7 ( D ) × 24 ( h ) ) ) × 10 matrix. The first nine columns represent nine input features, whereas the 10th column represents target output (load). Statistical features of the load dataset that have been used to train the RNN model is presented in Table 3. The frequency distribution of output load data values is represented in terms of histogram plot as shown in Figure 3,

3.1. Load Forecasting for HAM-(RHM-1)

The train and test datasets for “RHM-1” comprise a total of 1680 observations. Out of 1680 load data samples, 1512 were chosen for training and 168 for validation. For several hidden neurons the performance of the model in terms of performance metrics was observed during both training and testing, as shown in Table 4. Table 4 indicates that models performance in terms of training and test accuracy grows to 13 hidden neurons. If the number of occluded neurons exceeds 13 it is excessively fit and leads in further test errors. At this point, the “RHM-1” is deemed an optimum model, with 13 hidden neurons.
In addition, the number of hidden layers in the RNN model was raised to boost the model’s performance (RHM-1). The performance of the model with different levels was measured using the performance metrics as illustrated in Table 5. Each hidden layer consists of 13 neurons. It was seen from Table 5 that the model performs well with only one hidden layer. The test error values rise for the same loss of training if the number of hidden layers is more than one. This indicates that if the number of hidden layers is greater than one, then the model gets overfit. Furthermore, as the number of hidden layers rises, the number of training parameters increases the needed memory and processing time.
The suggested model, i.e., RNN-HAM-Mode11, has been trained 10 fold in the same data set and is judged to be the ideal load prediction model in real time when the best values for training and validation errors were given. The performance of the suggested ‘RHM-1’ model is observed in stochastic environment and shown in Table 6. For all error matrices that reflect the sturdy behavior of the “RHM-1” architecture, the standard deviation is noted to be virtually null.

3.2. Load Forecasting for HAM-(RHM-2)

The PCA algorithm is applied to the input features of the load dataset to find the principal components. The total variance in the dataset covered by each principal component and the cumulative variance covered are shown in Figure 4. It shows that six principal components cover almost 90% of the variance in the load dataset. Outcome of PCA that feeda as input to RNN for first 10 datasamples are presented in Table 7.
The suggested “RNN-HAM-Model2” has been trained and tested with different number of hidden neurons to detect the optimal “RNN-HAM-Model2”. The model is observed in terms of performance measures in the form of Table 8 throughout both the training and testing. The performance of the model has been growing up to 11 hidden neurons with regard to training and test accuracy, the Table 8 was found. So the optimum model is at this point in “RHM-2” with 11 hidden neurons.
The number of layers covered by this model (RHM-2), which was meant to predict loading one hour sooner, has also been increased. Each layer is comprised of 11 neurons and performance measurement metrics as shown in Table 9 of a model with different layers were observed. In Table 9, good test performance was found with only one hidden layer. If the number of layers concealed is more than one, then the numbers for the test error rise, then it is overfit if the number of layers hidden is higher than the one.
The model, i.e., RNN-HAM-Model 2, is trained ten times in an identical data set and is regarded to be the ideal model for real-time load prediction for training and validation errors. Table 10 presents the performance of the suggested model, that is, ‘RHM-2’ inside stochastic environments, and it is shown that for all error matrices, which indicate strong performance of the RHM-2 architecture, a standard deviation of practically zero is present.
In Table 11, An original model, i.e., the RHM-1, is compared with the compressed model, i.e., RHM-2. The RHM-2 is tiny with 210 parameters in relation to the 313 RHM-1. Due to the little dimensional compression of the model, RHM-2 losses compared to RHM-1 are somewhat greater. Although the workout parameters of “RHM-2” were compressed in 32.91%, losses of MSE, RMSE and MAE correspondingly rose by 4.5%, 1.7% and 5%.

3.3. Load Forecasting for DAM-(RDM-1)

The suggested model is conditioned and assessed using different numbers of hidden neurons in order to identify the best RDM-1. In terms of the performance metrics provided in Table 12, the model performance during training and testing is noted. The outputs of a model have been seen in Table 12 as regards training and test accuracy increases up to 13 hidden neurons. RDM-1 is deemed an optimum model at this moment with 13 hidden neurons.
In order to enhance performance (RDM-1), the numbers of hidden layers in the RNN models are increased. There are 13 neurons per hidden layer, the performance of which is demonstrated in Table 13 is illustrated by performance metrics for the model with different layers. The model with only one hidden layer has been noticeable in Table 13 for a positive test performance.
The model recommended, i.e., RDM-1 was trained on the same dataset 10 times and is regarded the best way to forecast loads in real time in terms of training and validation errors. Statistical analyses of the training behaviour, shown in Table 14, indicate that the standard deviation in the RDM-1 Architecture is practically zero for all error matrices described as robust behaviour.

3.4. Load Forecasting for DAM-(RDM-2)

The PCA algorithm is applied to the input features of the load dataset to find the principal components. Load dataset consists of a total of 6 input features, i.e., P ( h 24 ) , P ( h 48 ) , P ( h 72 ) , P ( h 168 ) , P ( h 336 ) and P ( h 504 ) , and one output P ( h ) . The total variance in the dataset covered by each principal component and the cumulative variance covered are shown in Figure 5. Figure 5 shows that four principal components cover almost 90% of the variance in the load dataset. Thus, six input features, i.e., P ( h 24 ) , P ( h 48 ) , P ( h 72 ) , P ( h 168 ) , P ( h 336 ) and P ( h 504 ) for each dataset sample, are translated into four principal components. These four principal components were used to train the RDM-2. RDM-2 was therefore equipped with four input neurons and one output neuron.
In order to find the optimal “RDM-2” in terms of the number of hidden neurons, the proposed “RDM-2” is equipped and evaluated with different numbers of hidden neurons. The performance of the model during both training and testing is observed in terms of performance metrics as shown in Table 15. From Table 15, it has been observed that the performance of the model has increased to 7 hidden neurons in terms of training and test accuracy. At this point, therefore, RDM-2 with 7 hidden neurons is considered to be an optimal model.
In addition, there have been higher numbers of hidden layers to improve the model’s efficiency (RDM-2) for load prediction. Each hidden layer has 7 neurons and performance metrics as given in Table 15 demonstrate the output of the model with different layers. From Table 16, good test performance with just one hidden layer has been noticed. If the number is more than one, the values for the test error are increased, the model becomes over-fit if the number of hidden layers is higher than one.
The recommended model, i.e., RDM-2, is trained ten times on the same data set and is deemed an ideal model for forecasting the load in real time when it has given the best values in relation to training and validation errors. In Table 17 the statistical analysis of the suggested model workouts reveals that the standard deviation is practically Nil for all the error matrices defining resilient behaviour in the RDM-2 architecture.
In Table 18, the comparison is shown to the original model, namely the RDM-1, and the compressed model. In comparison with the RDM-1 with 274 parameters the size of RDM-2 is modest with 92 parameters. The model RDM-2 exhibited somewhat higher test losses than the model RDM-1, due to the distortion of the model with the lower dimensionality. Although the training size of “RDM-2” has been reduced by 66.42%, losses, i.e, MSE, RMSE and MAE correspondingly have risen by 2.5%, 0.7% and 1.9%.

3.5. Comparative Result Analysis

The performance of the proposed RNN model was verified by comparing with ANN models [22,29,30], Regression models [19] and LSTM model [20] as presented in Table 19. It can be observed that the RNN model was able to predict the load with good accuracy. The performance of the model was compared statistically with models proposed in [22,29,30] and statistical metrics presented in Table 20, showing that the proposed RNN model is statistically robust with zero standard deviation.
The comparison with real load on 30 November 2018 of the loads forecast is shown in Figure 6 utilising several suggested RNN models for hourly and day ahead markets. The expected load of RHM-1 and RHM-2 is closer to real load than RDM-1 and RDM-2, since the former model forecast loads an hour earlier and one day in advance.
In Table 21, the total training time for various RNN systems with varying batch sizes is reported. As clearly shown, if we refer to batch size 32 (last row) the number of back-propagation is significantly reduced with respect to batch size 1, thus resulting in a lower computational effort as wanted by the authors’ initial design.
In order to show the advantages of using a non-linear approach, the performance of the proposed RNN model was verified by comparing with commonly used linear models like Auto Regression (AR) [31], Moving Average (MA) [32], Auto-regressive Moving Average (ARMA) [33], Auto-regressive Integrated Moving Average (ARIMA) [34] and Simple Exponential Smoothing (SES) [35], as presented in Table 22. It can be observed that the RNN model was able to perform better than traditional linear methods in terms of both RMSE and MAE values of predicted load. Although some concerns have been reported in literature with respect to using MAE as an accuracy indicator [36], we preferred to show both RMSE and MAE error metrics for the sake of comparison with results in previously cited references.

4. Conclusions

An accurate short-term projection of the electric load allows utilities to efficiently sell their electricity and manage the system on more steady, trustworthy expected information.
In order to ensure that utilities can efficiently trade in energy, the authors proposed different RNN models, notably RHM-1 and RDM-1 for predicting the load accurately. Lightweight models, i.e., RHM-2 and RDM-2, present reduced input features by means of PCA. These light weight models predicted the load with nearly the almost near accuracy as the original ones but reducing the complexity of the model a lot comparing to original models.
In this paper, real time load data were obtained from a 33/11 kV substation near the Kakatiya University in Warangal (India) for training and testing different RNN models in a practical case study. In order to identify outliers and also to observe the skewedness of data, suitable preprocessing techniques were employed.
The suggested RNN models were verified in terms of error measures by correlating them to those reported in the literature. Randomness in forecast using suggested RNN models is noticed and compared to current models.
Future works could additional take into account external factors and habits, e.g., climate, weather conditions and particular human behavioral patterns.

Author Contributions

V.V. constructed the research theories and methods, developed the basic idea of the study, performed the computer simulation and analyses; V.V., F.G. and M.M. conducted the preliminary research; D.R.C., F.G. and M.M. worked on proof read of this article; V.V., F.G. and M.M. worked on document preparation; V.V. served as the head researcher in charge of the overall content of this study as well as modifications made. All authors have read and agreed to the published version of the manuscript.

Funding

This research activity was partly supported by the EU Horizon 2020 project PLATOON (Grant agreement ID: 872592) https://platoon-project.eu/ (accessed on 17 January 2022).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets generated during and/or analysed during the current study are available in the Mendeley Data repository, https://data.mendeley.com/datasets/ycfwwyyx7d/2 (accessed on 17 January 2022).

Acknowledgments

We thank S R Engineering College Warangal, Telangana State, India and POLITECNICO DI MILANO, Italy for supporting us during this work. We also thank engineers in 33/11 kV substation near Kakatiya University in Warangal for providing the historical load data.

Conflicts of Interest

The authors declare that they have no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
P ( h ) Load at h th hour
P ( h 1 ) Load at one hour before from the time of prediction
P ( h 2 ) Load at two hours before from the time of prediction
P ( h 3 ) Load at three hours before from time of prediction
P ( h 24 ) Load at one day before from the time of prediction
P ( h 48 ) Load at two days before from the time of prediction
P ( h 72 ) Load at three days before from time of prediction
P ( h 168 ) Load at one week before from the time of prediction
P ( h 336 ) Load at two weeks before from the time of prediction
P ( h 504 ) Load at three weeks before from time of prediction
MSEMean Square Error
MAEMean Absolute Error
RMSERoot Mean Square Error
a < t > Hidden neuron current activation state
a < t 1 > Hidden neuron previous activation state
b a Bias parameter for hidden layer
b y Bias parameter for output layer
W a x Weight matrix between input and hidden layer
W y a Weight matrix between output and hidden layer
DAMDay ahead market
HAMHourly ahead market
RHM-1Recurrent Neural Network Model for Hourly Ahead Market
RHM-2Light weight recurrent neural network Model for Hourly Ahead Market
RDM-1Recurrent Neural Network Model for day ahead market
RDM-2Light weight recurrent neural network Model for day ahead market
Y i t r u e Actual load from i th sample
Y i p r e d Predicted load with i th sample

References

  1. Ahmad, T.; Chen, H. A review on machine learning forecasting growth trends and their real-time applications in different energy systems. Sustain. Cities Soc. 2020, 54, 102010. [Google Scholar] [CrossRef]
  2. Akhavan-Hejazi, H.; Mohsenian-Rad, H. Power systems big data analytics: An assessment of paradigm shift barriers and prospects. Energy Rep. 2018, 4, 91–100. [Google Scholar] [CrossRef]
  3. Almeshaiei, E.; Soltan, H. A methodology for electric power load forecasting. Alex. Eng. J. 2011, 50, 137–144. [Google Scholar] [CrossRef] [Green Version]
  4. Khodayar, M.E.; Wu, H. Demand forecasting in the Smart Grid paradigm: Features and challenges. Electr. J. 2015, 28, 51–62. [Google Scholar] [CrossRef]
  5. Mansoor, M.; Grimaccia, F.; Leva, S.; Mussetta, M. Comparison of echo state network and feed-forward neural networks in electrical load forecasting for demand response programs. Math. Comput. Simul. 2021, 184, 282–293. [Google Scholar] [CrossRef]
  6. Su, P.; Tian, X.; Wang, Y.; Deng, S.; Zhao, J.; An, Q.; Wang, Y. Recent trends in load forecasting technology for the operation optimization of distributed energy system. Energies 2017, 10, 1303. [Google Scholar] [CrossRef] [Green Version]
  7. Zheng, X.; Ran, X.; Cai, M. Short-term load forecasting of power system based on neural network intelligent algorithm. IEEE Access 2020. [Google Scholar] [CrossRef]
  8. Vasudevan, S. One-Step-Ahead Load Forecasting for Smart Grid Applications. Ph.D. Thesis, The Ohio State University, Columbus, OH, USA, 2011. [Google Scholar]
  9. Neusser, L.; Canha, L.N. Real-time load forecasting for demand side management with only a few days of history available. In Proceedings of the 4th International Conference on Power Engineering, Energy and Electrical Drives, Istanbul, Turkey, 13–17 May 2013; pp. 911–914. [Google Scholar]
  10. Singh, A.K.; Khatoon, S.; Muazzam, M.; Chaturvedi, D.K. Load forecasting techniques and methodologies: A review. In Proceedings of the 2012 2nd International Conference on Power, Control and Embedded Systems, Allahabad, India, 17–19 December 2012; pp. 1–10.
  11. Ahmad, F.; Alam, M.S. Assessment of power exchange based electricity market in India. Energy Strategy Rev. 2019, 23, 163–177. [Google Scholar] [CrossRef]
  12. Massaoudi, M.; Refaat, S.S.; Chihi, I.; Trabelsi, M.; Oueslati, F.S.; Abu-Rub, H. A novel stacked generalization ensemble-based hybrid LGBM-XGB-MLP model for Short-Term Load Forecasting. Energy 2021, 214, 118874. [Google Scholar] [CrossRef]
  13. Yin, L.; Xie, J. Multi-temporal-spatial-scale temporal convolution network for short-term load forecasting of power systems. Appl. Energy 2021, 283, 116328. [Google Scholar] [CrossRef]
  14. Syed, D.; Abu-Rub, H.; Ghrayeb, A.; Refaat, S.S.; Houchati, M.; Bouhali, O.; Bañales, S. Deep learning-based short-term load forecasting approach in smart grid with clustering and consumption pattern recognition. IEEE Access 2021, 9, 54992–55008. [Google Scholar] [CrossRef]
  15. Munkhammar, J.; van der Meer, D.; Widén, J. Very short term load forecasting of residential electricity consumption using the Markov-chain mixture distribution (MCM) model. Appl. Energy 2021, 282, 116180. [Google Scholar] [CrossRef]
  16. Guo, W.; Che, L.; Shahidehpour, M.; Wan, X. Machine-Learning based methods in short-term load forecasting. Electr. J. 2021, 34, 106884. [Google Scholar] [CrossRef]
  17. Eskandari, H.; Imani, M.; Moghaddam, M.P. Convolutional and recurrent neural network based model for short-term load forecasting. Electr. Power Syst. Res. 2021, 195, 107173. [Google Scholar] [CrossRef]
  18. Sheng, Z.; Wang, H.; Chen, G.; Zhou, B.; Sun, J. Convolutional residual network to short-term load forecasting. Appl. Intell. 2021, 51, 2485–2499. [Google Scholar] [CrossRef]
  19. Veeramsetty, V.; Mohnot, A.; Singal, G.; Salkuti, S.R. Short Term Active Power Load Prediction on A 33/11 kV Substation Using Regression Models. Energies 2021, 14, 2981. [Google Scholar] [CrossRef]
  20. Veeramsetty, V.; Chandra, D.R.; Salkuti, S.R. Short-term electric power load forecasting using factor analysis and long short-term memory for smart cities. Int. J. Circuit Theory Appl. 2021, 49, 1678–1703. [Google Scholar] [CrossRef]
  21. Grimaccia, F.; Mussetta, M.; Zich, R. Neuro-fuzzy predictive model for PV energy production based on weather forecast. In Proceedings of the 2011 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE 2011), Taipei, Taiwan, 27–30 June 2011; pp. 2454–2457. [Google Scholar] [CrossRef]
  22. Veeramsetty, V.; Deshmukh, R. Electric power load forecasting on a 33/11 kV substation using artificial neural networks. SN Appl. Sci. 2020, 2, 855. [Google Scholar] [CrossRef] [Green Version]
  23. Hong, T.; Pinson, P.; Fan, S.; Zareipour, H.; Troccoli, A.; Hyndman, R.J. Probabilistic energy forecasting: Global Energy Forecasting Competition 2014 and beyond. Int. J. Forecast. 2016, 32, 896–913. [Google Scholar] [CrossRef] [Green Version]
  24. Veeramsetty, V.; Reddy, K.R.; Santhosh, M.; Mohnot, A.; Singal, G. Short-term electric power load forecasting using random forest and gated recurrent unit. Electr. Eng. 2021, 1–23. [Google Scholar] [CrossRef]
  25. Abdi, H.; Williams, L.J. Principal component analysis. Wiley Interdiscip. Rev. Comput. Stat. 2010, 2, 433–459. [Google Scholar] [CrossRef]
  26. Mandic, D.P.; Chambers, J. Recurrent Neural Networks for Prediction: Learning Algorithms, Architectures and Stability; John Wiley & Sons, Inc.: Hoboken, NJ, USA, 2001. [Google Scholar]
  27. Karri, C.; Durgam, R.; Raghuram, K. Electricity Price Forecasting in Deregulated Power Markets using Wavelet-ANFIS-KHA. In Proceedings of the 2018 International Conference on Computing, Power and Communication Technologies (GUCON), Greater Noida, India, 28–29 September 2018; pp. 982–987. [Google Scholar]
  28. Veeramsetty, V. Active Power Load Dataset. 2020. Available online: https://data.mendeley.com/datasets/ycfwwyyx7d/2 (accessed on 13 December 2021).
  29. Shaloudegi, K.; Madinehi, N.; Hosseinian, S.; Abyaneh, H.A. A novel policy for locational marginal price calculation in distribution systems based on loss reduction allocation using game theory. IEEE Trans. Power Syst. 2012, 27, 811–820. [Google Scholar] [CrossRef]
  30. Veeramsetty, V.; Chintham, V.; Vinod Kumar, D. Proportional nucleolus game theory–based locational marginal price computation for loss and emission reduction in a radial distribution system. Int. Trans. Electr. Energy Syst. 2018, 28, e2573. [Google Scholar] [CrossRef]
  31. Hannan, E.J.; Kavalieris, L. Regression, autoregression models. J. Time Ser. Anal. 1986, 7, 27–49. [Google Scholar] [CrossRef]
  32. Johnston, F.; Boyland, J.; Meadows, M.; Shale, E. Some properties of a simple moving average when applied to forecasting a time series. J. Oper. Res. Soc. 1999, 50, 1267–1271. [Google Scholar] [CrossRef]
  33. Chen, J.F.; Wang, W.M.; Huang, C.M. Analysis of an adaptive time-series autoregressive moving-average (ARMA) model for short-term load forecasting. Electr. Power Syst. Res. 1995, 34, 187–196. [Google Scholar] [CrossRef]
  34. Contreras, J.; Espinola, R.; Nogales, F.J.; Conejo, A.J. ARIMA models to predict next-day electricity prices. IEEE Trans. Power Syst. 2003, 18, 1014–1020. [Google Scholar] [CrossRef]
  35. Haben, S.; Giasemidis, G.; Ziel, F.; Arora, S. Short term load forecasting and the effect of temperature at the low voltage level. Int. J. Forecast. 2019, 35, 1469–1484. [Google Scholar] [CrossRef] [Green Version]
  36. Gneiting, T. Making and Evaluating Point Forecasts. J. Am. Stat. Assoc. 2011, 106, 746–762. [Google Scholar] [CrossRef] [Green Version]
Figure 1. The considered architecture.
Figure 1. The considered architecture.
Forecasting 04 00008 g001
Figure 2. Proposed work flow diagram.
Figure 2. Proposed work flow diagram.
Forecasting 04 00008 g002
Figure 3. Histogram plot for output feature P ( h ) .
Figure 3. Histogram plot for output feature P ( h ) .
Forecasting 04 00008 g003
Figure 4. Variance in the load dataset (for HAM) covered by principal components.
Figure 4. Variance in the load dataset (for HAM) covered by principal components.
Forecasting 04 00008 g004
Figure 5. Variance in the load dataset (for DAM) covered by principal components.
Figure 5. Variance in the load dataset (for DAM) covered by principal components.
Forecasting 04 00008 g005
Figure 6. Actual load vs. predicted load.
Figure 6. Actual load vs. predicted load.
Forecasting 04 00008 g006
Table 1. Literature summary.
Table 1. Literature summary.
ReferenceYearContributionDisadvantage
 [12]2021novel stacking ensemble-based algorithmModel complexity
[13]2021multi-temporal-spatial-scale techniqueMissing Weekly impact
[14]2021k-Medoid based algorithmModel complexity
[15]2021Markov-chain mixture distribution modelAccuracy
[16]2021Fusion forecasting approachAccuracy
[17]2021Bi-directional GRU and LSTMModel complexity
[18]2021Deep Residual Network with convolution layerModel complexity
[19]2021Regression ModelsAccuracy
[20]2021LSTM and Factor AnalysisAccuracy
[22]2020ANNAccuracy
Table 2. Summary of the RNN models.
Table 2. Summary of the RNN models.
ParametersRHM-1RHM-2RDM-1RDM-2
Input neurons ( N i )9664
Output Neurons ( N o )1111
Hidden Neurons ( N h )1311137
Hidden Layers1111
Hidden Layer activationTanhTanhTanhTanh
Output Layer activationSigmoidSigmoidSigmoidSigmoid
Weights & bias31321027492
Table 3. Load data statistics.
Table 3. Load data statistics.
Statistical ParametersOutput P ( h )
Count1680.00
Mean5904.52
Std.1077.75
Min3377.92
25%5138.90
50%5795.62
75%6618.66
Max8841.67
Number of training samples1512
Number of testing samples168
Table 4. Impact of hidden neurons on the performance of the model “RHM-1”.
Table 4. Impact of hidden neurons on the performance of the model “RHM-1”.
NodesTrainingTestingTrainable
MSERMSEMAEParam
210.01040.1240.093673
180.01030.1200.088523
150.01020.1150.081391
130.01010.1150.08313
110.01020.1170.083243
100.01040.1170.083211
Table 5. Impact of hidden layers on the performance of the model “RHM-1”.
Table 5. Impact of hidden layers on the performance of the model “RHM-1”.
No. of HiddenTrainingTestingTrainable
LayersNodesMSERMSEMAEParameters
1130.010.1150.08313
2130.010.1240.094664
3130.010.1310.11015
Table 6. Statistical performance of the model (RHM-1).
Table 6. Statistical performance of the model (RHM-1).
Statistical ParametersTrainingTesting
MSERMSEMAE
count101010
mean0.01030.11680.0831
std0.0001250.0017510.003725
min0.01010.1150.079
25%0.01030.115250.08
50%0.01030.11650.082
75%0.0103750.117750.08575
max0.01050.120.09
Table 7. Pricipal components for first 10 load data samples used for HAM.
Table 7. Pricipal components for first 10 load data samples used for HAM.
PC-1PC-2PC-3PC-4PC-5PC-6
1.15773−0.03658−0.159480.080498−0.03406−0.09375
1.2067160.022011−0.068650.102240.020807−0.02302
1.3179270.087764−0.154420.2866190.1735710.137578
1.4740230.247519−0.140540.3277710.0176630.282082
1.5851020.250539−0.109910.3010180.049170.098137
1.5289440.003412−0.098010.148669−0.026320.071544
1.675344−0.22242−0.346020.148969−0.084340.103249
1.571563−0.28011−0.518460.037141−0.231630.15251
1.335613−0.03608−0.487650.030066−0.050950.214678
1.0350980.156347−0.32840.417524−0.173990.139219
Table 8. Impact of hidden neurons on the performance of the model “RHM-2”.
Table 8. Impact of hidden neurons on the performance of the model “RHM-2”.
Hidden NodesTrainingTestingTrainable Parameters
MSERMSEMAE
90.01110.1220.089154
100.01140.1190.086181
110.01100.1170.084210
120.01110.1210.088241
130.01100.1210.089274
Table 9. Impact of hidden layers on the performance of the model “RHM-2”.
Table 9. Impact of hidden layers on the performance of the model “RHM-2”.
No. of HiddenTrainingTestingTrainable Parameters
LayersNodesMSERMSEMAE
1110.01100.1170.084210
2110.01120.1190.086463
3110.01130.120.088716
4110.01130.20.087969
Table 10. Observations of performance of the RHM-2 in stochastic environment.
Table 10. Observations of performance of the RHM-2 in stochastic environment.
Statistical ParametersTrainingTesting
MSERMSEMAE
Count101010
mean0.01120.11940.0861
std0.00010.00140.0018
min0.01100.11700.0840
25%0.01120.11900.0850
50%0.01120.11900.0860
75%0.01130.12080.0868
max0.01150.12100.0890
Table 11. Comparison between RNN models for HAM.
Table 11. Comparison between RNN models for HAM.
ModelTrainable
Parameters
Testing
RMSEMAE
RHM-13130.1150.080
RHM-22100.1170.084
% of absolute change32.911.75
Table 12. Impact of hidden neurons on the performance of the model “RDM-1”.
Table 12. Impact of hidden neurons on the performance of the model “RDM-1”.
Hidden
Nodes
TrainingTestingTrainable
Parameters
MSERMSEMAE
180.01550.15100.1140469
150.01550.15000.1100346
130.01550.14200.1030274
120.01550.14600.1090241
110.01550.14800.1100210
Table 13. Impact of hidden layers on the performance of the model “RDM-1”.
Table 13. Impact of hidden layers on the performance of the model “RDM-1”.
No. of HiddenTrainingTestingTrainable
Parameters
LayersNodesMSERMSEMAE
1130.01550.1420.103274
2130.01540.1480.108625
3130.01560.1480.109976
Table 14. Statistical training performance of model RDM-1.
Table 14. Statistical training performance of model RDM-1.
Statistical
Parameter
TrainingTesting
MSERMSEMAE
Count101010
mean0.01550.14750.1089
std0.00010.00400.0041
min0.01540.14200.1030
25%0.01540.14400.1065
50%0.01550.14750.1090
75%0.01560.14980.1100
max0.01570.15400.1160
Table 15. Impact of hidden neurons on the performance of the model “RDM-2”.
Table 15. Impact of hidden neurons on the performance of the model “RDM-2”.
Hidden NodesTrainingTestingTrainable Param
MSERMSEMAE
50.01650.1450.10756
60.01640.1440.10773
70.01650.1430.10692
90.01670.1450.107136
110.01640.1460.109188
Table 16. Impact of hidden layers on the performance of the model “RDM-2”.
Table 16. Impact of hidden layers on the performance of the model “RDM-2”.
No. of HiddenTrainingTestingTrainable Parameters
LayersNodesMSERMSEMAE
170.01650.1430.10692
270.01650.1440.108197
370.01660.1500.114302
Table 17. Statistical analysis of RDM-2 architecture.
Table 17. Statistical analysis of RDM-2 architecture.
Statistical ParametersTrainingTesting
MSERMSEMAE
count101010
mean0.01650.14650.1092
std0.00020.00210.0029
min0.01630.14300.1050
25%0.01640.14480.1065
50%0.01650.14700.1095
75%0.01660.14800.1115
max0.01680.14900.1130
Table 18. Comparison between RNN models for Day Ahead Markets (DAM).
Table 18. Comparison between RNN models for Day Ahead Markets (DAM).
ModelTrainable ParametersTesting
RMSEMAE
RDM-12740.1420.103
RDM-292 50.1430.105
% of absolute change66.420.71.9
Table 19. Validation of models in testing environment.
Table 19. Validation of models in testing environment.
ModelMSERMSE
TrainingTestingTrainingTesting
ANN Model [29]0.291.590.541.26
ANN Model [30]0.230.440.480.66
ANN Model [22]0.20.320.450.57
SLR Model [19]0.09730.01630.3120.128
PR Model [19]0.01710.01580.1310.126
MLR Model [19]0.07230.01190.2690.109
LSTM-HAM-Model1 [20]0.01090.0130.1040.114
LSTM-HAM-Model2 [20]0.01250.01460.1120.121
LSTM-DAM-Model1 [20]0.01560.020.1250.141
LSTM-DAM-Model2 [20]0.01660.020.1290.1414
RHM-10.01010.01320.10.115
RHM-20.0110.01380.1050.117
RDM-10.01540.020.1240.141
RDM-20.01630.02050.1280.143
Table 20. Validation in probabilistic environment.
Table 20. Validation in probabilistic environment.
Parameter[29][30][22]RHM-1RHM-2RDM-1RDM-2
Mean0.29750.25000.22500.01350.01430.02150.0215
SD0.02000.01000.01000.00020.00030.00100.0006
Min0.28000.24000.20000.01320.01380.02000.0205
25%0.28000.24750.21750.01330.01410.02080.0209
50%0.29500.25000.22000.01350.01430.02170.0216
75%0.30500.25250.23500.01360.01450.02210.0218
Max0.33000.26000.25000.01390.01470.02320.0223
Table 21. Training computation time (s).
Table 21. Training computation time (s).
Batch SizeRHM-1RHM-2RDM-1RDM-2No. of Back Propagations
162411678201182151,200
813114810615318,900
16748250859500
32242433474800
Table 22. Validation of models in testing environment by comparing with classical models.
Table 22. Validation of models in testing environment by comparing with classical models.
ModelTesting
RMSEMAE
AR [31]0.1830.151
MA [32]0.1940.168
ARMA [33]0.1710.141
ARIMA [34]0.1750.135
SES [35]0.1490.124
RHM-10.1150.079
RHM-20.1170.084
RDM-10.1420.103
RDM-20.1430.105
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Veeramsetty, V.; Chandra, D.R.; Grimaccia, F.; Mussetta, M. Short Term Electric Power Load Forecasting Using Principal Component Analysis and Recurrent Neural Networks. Forecasting 2022, 4, 149-164. https://0-doi-org.brum.beds.ac.uk/10.3390/forecast4010008

AMA Style

Veeramsetty V, Chandra DR, Grimaccia F, Mussetta M. Short Term Electric Power Load Forecasting Using Principal Component Analysis and Recurrent Neural Networks. Forecasting. 2022; 4(1):149-164. https://0-doi-org.brum.beds.ac.uk/10.3390/forecast4010008

Chicago/Turabian Style

Veeramsetty, Venkataramana, Dongari Rakesh Chandra, Francesco Grimaccia, and Marco Mussetta. 2022. "Short Term Electric Power Load Forecasting Using Principal Component Analysis and Recurrent Neural Networks" Forecasting 4, no. 1: 149-164. https://0-doi-org.brum.beds.ac.uk/10.3390/forecast4010008

Article Metrics

Back to TopTop