Next Article in Journal
Towards Secure Fitness Framework Based on IoT-Enabled Blockchain Network Integrated with Machine Learning Algorithms
Next Article in Special Issue
Additive Ensemble Neural Network with Constrained Weighted Quantile Loss for Probabilistic Electric-Load Forecasting
Previous Article in Journal
Design and Optimization of ECG Modeling for Generating Different Cardiac Dysrhythmias
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Attention-Based Multilayer GRU Model for Multistep-Ahead Short-Term Load Forecasting †

School of Electrical Engineering, Korea University, 145 Anam-ro, Seongbuk-gu, Seoul 02841, Korea
*
Author to whom correspondence should be addressed.
This paper is an extended version of our paper published In Proceedings of the 6th International Conference on Next Generation Computing 2020 (ICNGC 2020), Busan, Korea, 17–19 December 2020.
Sensors 2021, 21(5), 1639; https://doi.org/10.3390/s21051639
Submission received: 25 January 2021 / Revised: 19 February 2021 / Accepted: 23 February 2021 / Published: 26 February 2021
(This article belongs to the Special Issue Smart and Predictive Strategies in Data Fusion)

Abstract

:
Recently, multistep-ahead prediction has attracted much attention in electric load forecasting because it can deal with sudden changes in power consumption caused by various events such as fire and heat wave for a day from the present time. On the other hand, recurrent neural networks (RNNs), including long short-term memory and gated recurrent unit (GRU) networks, can reflect the previous point well to predict the current point. Due to this property, they have been widely used for multistep-ahead prediction. The GRU model is simple and easy to implement; however, its prediction performance is limited because it considers all input variables equally. In this paper, we propose a short-term load forecasting model using an attention based GRU to focus more on the crucial variables and demonstrate that this can achieve significant performance improvements, especially when the input sequence of RNN is long. Through extensive experiments, we show that the proposed model outperforms other recent multistep-ahead prediction models in the building-level power consumption forecasting.

1. Introduction

Smart grid technologies have attracted much attention because of their potential to cope with climate change and energy crises [1]. A smart grid optimizes energy efficiency through bidirectional interaction between suppliers and consumers and employs renewable energy (i.e., solar and wind power), combined cooling, heating, and power, and energy storage systems (ESSs) [1,2,3]. As a result, greenhouse gases can also be reduced during electricity generation. Short-term load forecasting (STLF) should be performed to determine the power supply required for the power system to establish an operational plan of the smart grid [3,4]. The STLF plays a crucial role in deciding when to run the power generation system from the next hour to one week [4]. Diverse STLF models include daily peak load forecasting, total daily load forecasting, hourly electrical load forecasting, and very STLF (e.g., 10-min and 15-min interval electrical load forecasting). Particularly, hourly electrical load forecasting and very STLF are popularly used for performing demand-side management or operating ESSs from the next hour up to one day [5].
STLF is a challenging task because electrical energy consumption exhibits complicated patterns and fluctuations due to diverse unexpected external factors [6]. Thus, effectively learning the relationship between electrical load and external factors when constructing a forecasting model is essential. Many studies have been done to construct STLF models using artificial intelligence (AI) techniques. For instance, support vector regression (SVR) [7], random forest (RF) [8], and extreme gradient boosting (XGB) [9] have demonstrated excellent prediction performance by considering the nonlinear relationship between input and output variables. However, these STLF models focus on the day-ahead point forecasting; they cannot easily and properly cope with various unexpected events that could affect electrical energy consumption for a day from the present time. Moreover, if the time difference between the training and testing sets is extensive, their prediction performance could deteriorate because they do not adequately reflect recent electrical load trends [10].
Recently, a gated recurrent unit (GRU) was used for STLF to solve these problems because it is simple and easy to implement and can carry out multistep-ahead forecasting [11]. However, the GRU has the disadvantage that its forecasting accuracy may deteriorate for a long input sequence because it concentrates on all variables equally. In this paper, we propose a GRU-based multistep-ahead STLF model and augment it using an attention mechanism to improve the forecasting performance by focusing more on crucial variables. To do this, we first collected hourly power consumption data of three office buildings. Then, we carried out preprocessing and exploratory data analysis to verify the relationship between various factors and the power consumption of buildings for model training. Lastly, we constructed an attention-augmented GRU network for multistep-ahead (24 points) forecasting for a day from the current time.
The main contributions of this paper are as follows:
  • We conducted multistep-ahead forecasting for the hourly power consumption of buildings to adequately cope with sudden changes in power consumption caused by various unexpected events, such as peak and blackout, instead of day-ahead point STLF.
  • We constructed an attention-based multilayered GRU model to achieve faster and more stable multistep-ahead STLF than other deep learning (DL) architectures.
  • We verified the superiority of the proposed model through extensive comparisons with several state-of-the-art forecasting models using the power consumption data of three office buildings.
The remainder of this paper is organized as follows. In Section 2, we review several related studies. Section 3 describes the input variable configuration. Section 4 presents the steps for constructing our attention based GRU model. Section 5 shows the experimental results to demonstrate the superiority of the model. Finally, we conclude the study and present the directions for future research in Section 6.

2. Related Studies

Many STLF models have been proposed to accurately predict energy consumption based on diverse AI techniques. In this section, we briefly introduce some recent STLF approaches based on machine learning (ML) and DL-based methods and summarize them in Table 1.
Lahouar and Slama [12] proposed a day-ahead hourly electrical load forecasting model based on the RF. They configured input variables by considering two cases: (1) exogenous and relative to the predicted day and (2) endogenous and relative to the previous days. They used the RF model’s variable importance to explain the relationship between the input variables and electrical load. They demonstrated the superiority of the proposed model by comparing it with the persistence analysis, artificial neural network, and SVR. Moon et al. [13] developed an RF-based total daily load forecasting model for university campuses. They first used a moving average method to consider the electrical load patterns on the days of the week. They built an RF model using several input variables, such as the timestamp, temperature, academic year, historical load, and prediction value from the moving average method. They demonstrated that their proposed model outperformed other popular ML methods, such as the decision tree, MLR, gradient boosting machine, SVR, and artificial neural network. Park et al. [14] proposed a two-stage hourly electrical load forecasting model. In the first stage, they constructed two forecasting models using the XGB and RF methods. Then, they combined their prediction values using a sliding window-based multiple linear regression (MLR) model in the second stage. They demonstrated that their proposed model outperforms several single ML methods.
Ryu et al. [15] proposed two deep neural network-based STLF models for applying a demand-side empirical load database. They compared these STLF models with the shallow neural network, double-seasonal holt-winters, and autoregressive integrated moving average (ARIMA) and demonstrated that their STLF models exhibited excellent performance. Izonin et al. [16] used non-iterative approaches based on a successive geometric transformation model (SGTM) to predict the net hourly electrical energy output of a combined cycle power plant. The SGTM neural-like structure showed a better prediction performance than MLR, SVR, and general regression neural networks. Motepe et al. [17] developed an improved load forecasting process using hybrid AI and deep learning methods. They determined the optimal hyperparameters through several experiments to enhance the performance of long short-term memory (LSTM) networks. They demonstrated that their proposed model, the LSTM network, had lower prediction errors than an adaptive neuro-fuzzy inference system and optimally pruned extreme learning machine.
Kuan et al. [18] constructed a multilayered self-normalizing GRU model for STLF. They demonstrated that the multilayered self-normalizing technology could improve the prediction performance of the GRU and LSTM models through several experiments. Chitalia et al. [19] presented a short-term load forecasting framework that is robust regardless of building types or locations. They collected five commercial buildings of five different building types located at five different locations. They explored nine different deep learning models and determined the best model for load forecasting in each cluster by considering unsupervised k-mean clustering. Sehovac and Grolinger [20] proposed a recurrent neural network (RNN) with attention for load forecasting. Their RNN had the ability to model time dependencies. They used sequence to sequence (S2S) approach to strengthen the ability by using encoder and decoder. Moreover, they added an attention mechanism to ease the connection between encoder and decoder to further improve the performance. They proved the superiority of their S2S approach and attention mechanism through comparison with a non S2S model and vanilla RNN.

3. Data Preprocessing

3.1. Data Collection

This section describes the data preprocessing and input variable configuration for the STLF model. For model construction and testing, we used one publicly available dataset and one confidential dataset. The former represents electrical energy consumption data collected from two office buildings in Richland, Washington, USA, from 2 January 2009 to 31 December 2011 [21]. The datasets consist of the timestamp information, temperature in Fahrenheit, and hourly electrical energy consumption. For the three missing values (e.g., 2009/04/05/02:00, 2010/04/04/02:00, and 2011/04/03/02:00) and some anomalous data in the datasets, we handled them properly using a linear interpolation method. The latter represents hourly electrical energy consumption data collected from the headquarters of a midsized company in Seoul, South Korea from 1 January 2015 to 31 December 2017 [10]. We also collected the holiday information in Washington and South Korea at ‘Time and Date [22]’ that can confirm national public holiday information in many countries.
Table 2 lists the simple building information with some statistical analyses on the collected electrical energy consumption data. Holiday includes Saturday, Sunday, and national holidays in the table. The minimum and count in the table represent the lowest electrical energy consumption and the number of data points, respectively. In addition, the first quartile, median, and third quartile represent the values of the lower 25%, 50%, and 75% points, respectively, in the energy consumption. Figure 1 shows the box plots for the energy consumption of three buildings by weekday, holiday, and total.

3.2. Feature Extraction

We considered the temperature, timestamp information, and historical load data to build the model and performed feature extraction for the given datasets to train the model as represented in Table 3.
In particular, as electrical energy consumption tends to vary with date and time, we considered all variables that represent date and time such as month, day, hour, day of the week, and holiday. Month, day, and hour have a sequence format. When data with a sequence format are used directly in the AI techniques, it is difficult to reflect their periodicity [23]. For instance, 11 p.m. and midnight are temporally contiguous; however, the data difference in sequence format is 23.
Hence, we represented the month, day, hour, and day of the week into two-dimensional data using Equations (1) to (8) to reflect their periodicity [23,24]. In Equations (3) and (4), LDM represents the last day of the month. For instance, if the month of the prediction is March, its LDM becomes 31. In addition, we used a variable indicating whether it is a holiday to consider different electrical energy consumption patterns on weekdays and holidays:
Month x = sin ( ( 360 ° / 12 ) × Month )
Month y = cos ( ( 360 ° / 12 ) × Month )
Day x = sin ( ( 360 ° / LDM ) × Day )
Day y = cos ( ( 360 ° / LDM ) × Day )
Hour x = sin ( ( 360 ° / 24 ) × Hour )
Hour y = cos ( ( 360 ° / 24 ) × Hour )
Day   of   the   week x = sin ( ( 360 ° / 7 ) × Day   of   the   week )
Day   of   the   week y = cos ( ( 360 ° / 7 ) × Day   of   the   week )
The temperature, which is closely related to power consumption, is also used for the model training [13]. The datasets of Building 1 and Building 2 included outdoor temperature in Fahrenheit, while the dataset of Building 3 only included power consumption data according to the timestamp information. Hence, we collected the temperature near the Building 3 provided by the Korea Meteorological Administration (KMA) [25] and converted them from Celsius to Fahrenheit.
We also considered the historical electricity load data comprising the electricity load simultaneously from one day to one week ago as the input variables to reflect the recent electricity consumption pattern [26,27]. In addition, we added a new historical electricity load value as an input variable to reflect different electricity load patterns of holidays and weekdays [13]. When the prediction point was a weekday, we used the weekday electricity consumption average of the past week. Likewise, when the prediction point is a holiday, we used the average of the past week’s holiday electricity consumption.

3.3. Correlation Analysis

We configured various input variables, such as the timestamp, temperature, and historical load to construct the forecasting model. We calculated the Pearson correlation coefficients with p-values for each dataset to confirm their relevance to the actual electrical energy consumption. Table 4 illustrates the results.
In the table, the average load considering the weekdays/holidays, historical load one day before, and historical load one week before exhibit a strong correlation with the actual electrical energy consumption. We also verified that the electric load pattern differs depending on the hour, day of the week, and holiday because Houry, Day of the weeky, and Holiday present a negative correlation with the actual electrical energy consumption.
The remaining variables are positively related to the building electrical energy consumption. Although they exhibited a nonlinear relationship, we can adequately handle this by applying a deep learning method, such as the LSTM and GRU networks. Overall, most input variables exhibit a meaningful correlation with the building electrical energy consumption because the p-values are less than 0.01.

4. Forecasting Model Construction

4.1. Gated Recurrent Unit Model

An artificial neural network (ANN), also known as a multilayer perceptron (MLP), is a popular AI technique implemented based on human biological neurons that can process large amounts of data in parallel and learn efficiently [24,28]. The ANN is a static input-output mapping model that only considers the input and output and does not consider time. It usually represents input, weights, output, etc., in vector form and the output is calculated internally from the input and weight vectors. The decision boundary is orthogonal to the weight vector [29]. In addition, the ANN presents various decision boundaries because an activation function determines the perceptron’s response to its input [23].
In contrast, a recurrent neural network (RNN), although it is a kind of ANN, is dynamic input-output mapping model that considers the input of all time. RNNs are well-suited to time-series data because they can process a time-series step-by-step, maintaining an internal state from time step to time step [30]. The LSTM is the most successful and widely used RNN [31]. The LSTM preserves the differential values of old inputs during backpropagation to solve the long-term dependency problem of the RNN. As a variant of the RNN architecture, GRU has the advantage of simplifying the LSTM structure by reducing the computation to update the hidden state while solving the long-term dependency problem and maintaining the performance of the LSTM [11,32].
Figure 2 illustrates a GRU cell architecture. In the figure, the GRU cell has input and forget gates. A gate controller, z, controls both the input and forget gates. When z is 1, the forget gate is closed, and the input gate is open. When z is 0, the forget gate is open, and the input gate is closed. At each step, the previous ( t 1 ) memory is saved, and the input of the time step is cleared. The GRU cell is controlled according to Equations (9)–(12):
r t = σ ( W r h t 1 + U r x t )
z t = σ ( W z h t 1 + U z x t )
c t = tan h ( W c ( h t 1 r ) + U c x t )
h c = ( z c ) + ( ( 1 z ) h t 1 )
We used a GRU network to predict the building electrical energy consumption at 24 points in time (from one hour later to one day later) and considered several hyperparameters to construct our GRU model. We set the number of hidden layers in the GRU model to two. The input layer of the GRU model consists of 18 nodes, and the hidden layer consists of 13 nodes per layer by applying two-thirds of the input layer, plus the size of the output layer [10,23].
We used the scaled exponential linear unit (SELU) as an activation function [33,34], defined by Equation (13), where α is a stochastic variable sampled from a uniform distribution at training time and is fixed to 1.67326, which is the expectation value of the distribution at testing time. Moreover, λ is an extra parameter for determining the slope and is set to 1.0507. This is because SELU can effectively train DL models due to its superior self-normalization quality and no vanishing gradient problem [23,34].
In addition, we used the Huber loss [35] and adaptive moment estimation (Adam) [34] as a loss function and optimization algorithm, respectively. The Huber loss is calculated using Equation (14) with δ = 1. We set the learning rate and learning epoch to 0.001 and 500, respectively.
f ( α , λ , x ) = { λ ( α e x α )   for   x < 0 λ x   for   x 0
L δ ( y , f ( x ) ) = { 1 2 ( y f ( x ) ) 2   for   | y f ( x ) | 0 δ | y f ( x ) | 1 2     δ 2   otherwise

4.2. Attention Mechanism

A more extended input sequence in the GRU network results in a worse prediction accuracy of the output sequence because it focuses on all input variables equally, even though they could have different correlations to the forecasting. An attention mechanism can be used to alleviate this problem by focusing on more relevant input variables.
An attention mechanism [36] consists of an encoder that generates an attention vector from the input and a decoder that generates a hidden state by taking the encoder output as input. The hidden states are divided by a view, and the encoder assigns an attention score to the hidden state of each step using the hidden state of the previous view decoder. An attention vector is created by performing a soft-max operation on the created attention score. In this way, whenever the decoder predicts the output value, the encoder concentrates on the input variables similar to the predicted value.
For instance, Kwon et al. [37] developed an attention mechanism based RNN model on electronic medical records. Hence, we constructed an attention mechanism to focus on input variables with high correlation to improve the model accuracy. We set the size of the attention window to 96. Figure 3 presents the overall attention based GRU model architecture.

5. Experimental Results

5.1. Experimental Design

In the experiments, we used an Intel (R) Core (TM) i7-8700 CPU (Santa Clara, CA, USA), Samsung 32G DDR4 memory (Suwon, Korea), NVIDIA Geforce GTX 1080ti (Santa Clara, CA, USA) and the operating system is Windows 10 version. For multistep-ahead hourly electricity load forecasting, we performed the experiments in Python 3.7.6, and the RNN-based models were constructed using TensorFlow 1.13.1 [38]. Table 5 presents the entire collection, training set, and testing set period for each building.
In AI-based STLF model training, normalization is usually conducted to prevent large-scale features from having too much influence by adjusting all the input variables within a close range. For this, we applied the min-max normalization for all input variables using Equation (15). By doing so, we transformed all input variables of the training set and then applied the X norm values obtained from the training set on the testing set. Hence, the minimum and maximum values of an input variable could be 0 and 1, respectively.
X norm = X X min X max X min
We used the mean absolute percentage error (MAPE) and coefficient of variation of the root mean squared error (CVRMSE) to compare the forecasting performance, which can be calculated using Equations (16) and (17), respectively. Both metrics represent the accuracy as a percentage error. Hence, they are more intuitive and easier to understand than other well-known metrics, such as the root mean squared error and mean squared error [23,26]. A lower value for these metrics indicates better prediction performance.
MAPE = 100 n t = 1 n | A t F t A t |
CVRMSE = 100 A ¯ t = 1 n ( A t F t ) 2 n
Here, A ¯ is an average of the actual values, and A t and F t are the actual and predicted values at time t, respectively.

5.2. Experimental Results and Discussion

We compared it with other state-of-the-art models, such as multivariate RF (MRF), DNN, LSTM, ATT-LSTM, GRU, and ensemble models to evaluate the validity of the proposed model. We considered two ensemble models: Park’s stacking ensemble model [14] and Moon’s stacking ensemble model, called COSMOS [10]. We considered all the input variables using each input variable for the prediction time point to construct the MRF and DNN models. Therefore, we used 432 input variables, that is, 18 (number of input variables) × 24 (number of prediction time points), for the multistep-ahead STLF [39]. The two stacking ensemble models consist of two stages, and the second-stage model used the prediction results of the first stage and demonstrated better forecasting performance than many single ML models and existing forecasting models. We also compared the prediction performance of the proposed model with that of Kuan’s GRU model [18]. We implemented an RF-based STLF model using MRF [40] in R packages and the two stacking ensemble models using xgboost 1.3.0 [41] and scikit-learn [42] in the Python environment. The selected hyperparameters for each model are listed in Table 6.
Table 7, Table 8 and Table 9 and Table 10, Table 11 and Table 12 present the MAPE and CVRMSE comparison of the three buildings by month and date type, respectively. The values in bold font indicate the best performance for the month (or date type). When comparing MAPE (CVRMSE) by month, the MAPE (CVRMSE) values from April to September, which corresponds to the summer, showed the lowest values. When comparing the forecasting performance of weekdays and holidays, the forecasting performance for weekdays was generally better than that for the weekend.
As listed in Table 2, the scale of power consumption of Building 3 is larger than that of Buildings 1 and 2, so the MAPE and CVRMSE values of Building 3 are generally smaller than Buildings 1 and 2. Table 7, Table 8, Table 9, Table 10, Table 11 and Table 12 indicate that the proposed model (ATT-GRU) showed the lowest MAPE among most STLF models in the comparisons by month and date type. In addition, from the results of ATT-LSTM and ATT-GRU, the attention mechanism could improve the prediction performance of the general LSTM and GRU models by 10% or more.
Although the GRU model performed better than LSTM in most experiments, the GRU model took less time to construct because it was a lighter model than the LSTM model. The average elapsed times for LSTM and GRU-based models are 10,937, 11,581, 7074, and 7263 s for LSTM, ATT-LSTM, GRU, and ATT-GRU, respectively. Also, their training time is depicted in Table 13. In terms of multistep-ahead STLF, the prediction accuracy of the ATT-GRU model did not reveal any significant decrease, even for the rear points. Moreover, our model exhibits the lowest MAPE and CVRMSE values point by point. The point-by-point forecasting results are depicted in the Appendix A. This is because the attention mechanism calculates the weights of the points and focuses on the points with large weights.
Figure 4, Figure 5 and Figure 6 represent scatter plots for the power consumption and temperature of the three buildings. According to the plots, the average power consumption of Buildings 1 and 2 tended to increase similarly as the temperature approached 0 or 100 degrees Fahrenheit and Building 3 showed an increase in power consumption as the temperature approached 100 degrees Fahrenheit. On the contrary, Buildings 1 and 2 showed frequent power consumption peaks when the temperature got close to 0 degrees Fahrenheit. The power forecast accuracy for Buildings 1 and 2 in November and December deteriorated due to unexpected peaks at low temperatures. Our proposed multi-step ahead forecasting method can be used to predict such unexpected peaks and handle them.

6. Conclusions

In this paper, we proposed a reliable hourly electrical energy consumption forecasting scheme using the attention mechanism and GRU network. The GRU network performs multistep-ahead forecasting, and the attention mechanism enables the network to concentrate more on the input variables with a higher correlation to energy consumption. We collected publicly available datasets from the United States Department of Energy and constructed 18 variables related to electricity consumption and an STLF model using the attention based GRU network to verify the proposed scheme performance. Through extensive experiments, we compared the proposed model with other popular state-of-the-art STLF models, including the stacking ensemble model, DNN, COSMOS, LSTM, ATT-LSTM, and GRU. We showed that our model outperformed the other models in terms of MAPE and CVRMSE. In addition, we demonstrated that the prediction performance of LSTM and GRU can be improved by more than 10% by using attention mechanism.
However, our model still has a black box. That is, it cannot explain the evidence of the output like any other deep learning models. To overcome this, we plan to extract feature importance using layer-wise relevance back-propagation and implement explainable AI based on it as a future work. Moreover, we will investigate how further to improve the prediction performance of multistep-ahead electricity load forecasting.

Author Contributions

Conceptualization, S.J. and J.M.; methodology, S.J., J.M. and S.P.; software, S.J., J.M. and S.P.; validation, S.J. and J.M.; formal analysis, E.H.; data curation, S.J., J.M. and S.P.; writing—original draft preparation, S.J.; writing—review and editing, E.H.; visualization, S.J., J.M. and S.P.; supervision, E.H.; project administration, E.H.; funding acquisition, E.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported in part by the Energy Cloud R&D Program (grant number: 2019M3F2A1073184) through the National Research Foundation of Korea (NRF) funded by the Ministry of Science and ICT and in part by the Korea Electric Power Corporation (grant number: R18XA05).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. This data can be found here: https://openei.org/datasets/dataset/consumption-outdoor-air-temperature-11-commercial-buildings (accessed on 25 February 2021).

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Table A1, Table A2, Table A3, Table A4, Table A5 and Table A6 present the MAPE and CVRMSE values of the forecasting methods for 24 forecasting points.
Table A1. Mean absolute percentage error comparison for Building 1.
Table A1. Mean absolute percentage error comparison for Building 1.
PointsMRFDNNPark’s [14]COSMOS [10]Kuan’s [18]LSTMATT-LSTMGRUATT-GRU
117.4414.7023.2113.5615.6714.8913.5813.5413.41
217.4315.4423.3913.6714.8914.2713.5113.5813.44
317.4815.8723.6513.7715.3914.3413.8914.3913.49
417.5916.3723.9213.8515.5514.3313.7014.4513.51
517.5716.3724.1913.9115.8214.4013.7414.4213.52
617.7017.0324.4913.9615.9814.4913.8114.4013.61
717.9017.0324.6714.0115.9014.5813.7814.3413.59
818.0317.0324.8914.0415.8414.6613.7314.2513.62
918.0316.8525.0914.0715.8314.7613.6714.1913.64
1018.2717.2425.2614.1015.8814.7613.6614.1613.64
1118.3517.1925.4314.1115.9314.9413.6714.3013.64
1218.5817.1925.5414.1315.9514.9613.6814.1113.64
1318.6717.4025.5014.1415.9614.9913.6914.1013.65
1418.4717.2925.4314.1615.9615.0313.7014.1213.65
1518.4017.1725.3014.1615.9615.0613.7314.1413.67
1618.4017.0825.1014.1715.9915.0913.7514.1513.67
1718.4217.4124.9614.1816.0115.1013.7914.1813.68
1818.4917.0124.7614.1916.0215.1113.8314.2013.68
1918.4617.1624.4814.2016.0215.1113.8514.2013.69
2018.4217.2224.3014.2216.0315.1013.8914.2013.71
2118.3517.2124.1414.2416.0215.0913.9114.2213.72
2218.7517.2824.0214.2616.0315.0913.9314.2313.73
2319.1116.9823.9214.2816.0415.0713.9814.2313.76
2419.8417.4523.8814.3116.0615.0614.0114.2413.78
Avg.18.2616.8724.5614.0715.8614.8513.7714.1813.63
Table A2. Mean absolute percentage error comparison for Building 2.
Table A2. Mean absolute percentage error comparison for Building 2.
PointsMRFDNNPark’s [14]COSMOS [10]Kuan’s [18]LSTMATT-LSTMGRUATT-GRU
120.6215.9629.2311.8218.8912.1011.9111.5311.56
221.5116.6729.4011.9119.4412.1011.9412.8811.73
322.1616.7929.6711.9821.3912.7311.9612.9311.74
422.8317.1529.9412.0422.9913.0611.9713.6211.75
523.3718.4230.1512.0823.7213.2511.9713.5511.77
623.8718.6930.4012.1123.9413.3011.9813.8311.80
724.1418.2930.6512.1424.0313.2611.9813.7411.82
824.6718.7730.8712.1624.1213.2211.9913.7611.82
925.1919.7831.0912.1724.2213.1911.9913.6811.82
1025.3519.6031.3412.1824.2813.1712.0013.6211.83
1125.5320.6831.5712.1924.3313.1812.0013.5911.83
1225.7120.4331.7912.2024.3413.1812.0013.5711.83
1325.6020.8031.9312.2024.3213.2012.0113.5611.85
1424.8420.6431.9812.2024.2613.2012.0113.4911.84
1524.4320.1831.9912.2024.2213.1912.0113.4211.84
1624.1819.5931.9412.2024.1713.1812.0213.3711.85
1724.0119.7631.8512.2124.1513.1612.0213.3411.85
1823.5819.1531.7312.2124.1613.1612.0213.3511.86
1923.5719.2331.5812.2224.1613.1512.0213.3611.86
2023.2719.6231.4212.2324.1713.1412.0313.3411.87
2123.1419.7231.2912.2524.1813.1112.0313.3411.88
2223.1819.6731.1512.2724.2113.1712.0313.3311.88
2323.7520.1030.9912.2924.2313.2112.0313.3111.88
2424.6320.0530.8212.3124.2613.2612.0313.3011.88
Avg.23.8819.1631.0312.1623.5913.0812.0013.3711.81
Table A3. Mean absolute percentage error comparison for Building 3.
Table A3. Mean absolute percentage error comparison for Building 3.
PointsMRFDNNPark’s [14]COSMOS [10]Kuan’s [18]LSTMATT-LSTMGRUATT-GRU
16.327.768.806.316.426.696.276.685.91
26.377.258.826.377.026.696.276.676.01
36.447.608.876.417.136.676.296.676.03
46.517.848.906.457.196.796.286.836.05
56.587.778.936.487.216.736.296.816.07
66.657.868.986.517.236.766.286.866.07
76.677.939.046.537.266.746.286.846.08
86.688.179.096.557.256.746.286.816.09
96.698.129.146.567.276.756.296.786.09
106.778.319.216.577.296.776.296.786.09
116.808.469.266.577.316.806.296.786.09
126.808.559.306.587.336.846.296.786.10
136.778.459.316.587.346.866.296.796.10
146.788.599.306.587.346.876.296.796.10
156.788.729.286.587.356.896.306.806.10
166.818.779.276.597.356.916.306.816.11
176.798.499.236.607.366.936.306.816.11
186.838.639.196.617.366.956.306.816.13
196.928.639.166.627.376.956.306.816.12
207.008.689.126.637.386.976.306.826.12
217.108.359.076.647.406.976.306.846.12
227.248.729.026.667.406.976.306.856.12
237.508.698.986.677.406.986.306.856.12
247.848.828.946.697.397.006.306.856.12
Avg.6.828.309.096.567.266.856.296.806.09
Table A4. Coefficient of variation of the root mean squared error comparison for Building 1.
Table A4. Coefficient of variation of the root mean squared error comparison for Building 1.
PointsMRFDNNPark’s [14]COSMOS [10]Kuan’s [18]LSTMATT-LSTMGRUATT-GRU
127.4022.3437.8522.8024.9725.5622.9323.5622.44
227.5923.7638.5623.0123.6925.5322.4923.6122.46
327.9024.5339.4123.1523.3025.8522.3223.6922.49
428.1024.9640.3823.2523.2825.6922.3623.7322.50
528.1325.4241.6923.3323.5225.8622.6223.8622.60
628.2225.7043.2523.4023.7326.1622.7823.9022.62
728.2725.7544.1723.4523.6726.3922.9624.0022.63
828.3625.5345.9623.5023.6126.5623.0624.0622.80
928.3025.5947.5123.5423.5926.7323.1024.1523.06
1028.5226.0947.9123.5623.6226.8523.1724.2523.16
1128.5625.8348.2823.5923.6626.9523.2224.3823.16
1228.8526.2248.2223.6123.6827.0323.3224.4123.31
1328.8626.1248.0323.6223.6827.1223.3124.4123.22
1428.6626.0948.0023.6323.7127.2123.3424.4523.31
1528.5625.8447.7423.6323.7527.2623.4024.5023.35
1628.5625.7046.7723.6323.7727.2923.4324.5423.36
1728.4226.0645.7423.6323.8027.3023.4724.5623.39
1828.4825.6244.5423.6323.8227.3023.5124.5723.38
1928.4525.6142.8523.6323.8027.3023.5124.5823.39
2028.3325.7542.0323.6423.7827.2923.5324.5523.52
2128.4125.7241.3823.6523.7627.3623.5424.5623.50
2228.9025.8440.7923.6723.7627.4123.5424.5623.55
2329.5625.7440.1123.6823.7427.5523.5624.5323.63
2430.4626.1339.4823.6923.7627.6223.5324.4923.65
Avg.28.4925.5043.7823.5023.7326.8023.1724.2522.44
Table A5. Coefficient of variation of the root mean squared error comparison for Building 2.
Table A5. Coefficient of variation of the root mean squared error comparison for Building 2.
PointsMRFDNNPark’s [14]COSMOS [10]Kuan’s [18]LSTMATT-LSTMGRUATT-GRU
124.0319.2834.5419.0921.2122.0319.2120.8819.11
223.8319.9935.0219.2420.0421.8219.2621.6919.21
323.8320.5735.5419.3321.4422.2219.2921.4319.27
423.8321.1636.1219.4122.8122.4219.3021.8819.26
523.8821.6036.6319.4723.4822.5219.3621.8519.33
623.8622.0037.2619.5123.6122.5819.4122.1419.39
724.0021.9637.9519.5523.6922.6019.4422.1719.42
824.0822.1638.6819.5823.7522.5819.5122.1819.45
924.0622.6239.1719.6023.8622.5919.5222.1519.49
1024.0422.7039.8419.6223.9922.6219.5322.1019.47
1123.9123.2040.4319.6424.1022.6319.5322.1319.48
1224.1322.9740.8419.6524.1322.6519.5422.1319.46
1324.1323.3541.0919.6624.1322.6919.5522.0719.53
1423.9123.5841.0819.6624.0922.7219.5722.0519.49
1523.8523.2740.9119.6624.0722.7519.5722.0019.51
1623.6922.7940.7019.6624.0322.7719.6022.0019.53
1723.6123.1240.3419.6624.0022.7919.6222.0419.60
1823.6422.6539.9319.6524.0022.8019.6122.0619.59
1923.6922.7239.4319.6624.0022.8019.6222.0819.59
2023.7522.9138.9019.6723.9922.7919.6322.0819.56
2123.8722.9638.3919.6923.9922.7819.6522.0919.59
2224.0422.9137.8219.7123.9922.7219.6622.0919.57
2324.7023.1437.2619.7524.0022.6919.6522.0819.61
2425.5123.1036.8019.7724.0022.6519.6622.1119.58
Avg.23.9922.3638.5319.5823.5222.5919.5121.9819.46
Table A6. Coefficient of variation of the root mean squared error comparison for Building 3.
Table A6. Coefficient of variation of the root mean squared error comparison for Building 3.
PointsMRFDNNPark’s [14]COSMOS [10]Kuan’s [18]LSTMATT-LSTMGRUATT-GRU
111.4011.0216.6210.2410.9611.5610.1611.2910.09
211.4611.1916.6610.3711.2111.8610.3711.1010.08
311.6011.8616.6810.4711.6211.6910.3211.2410.25
411.6611.8416.6610.5511.8512.0210.3511.4610.46
511.7512.1216.6410.6111.8711.9910.5611.2610.33
611.8212.3816.6510.6612.2512.1610.5111.4210.52
711.7612.3616.7510.7112.4212.2210.5511.5210.54
811.7412.3916.8710.7412.5212.2910.7211.6010.57
911.6712.6317.0210.7612.6112.3610.7111.6310.63
1011.8312.8817.1810.7712.7012.4010.6711.6710.61
1111.8712.6817.3210.7712.7612.4310.7511.6810.74
1211.8512.8817.4610.7812.8012.4810.7111.6710.77
1311.6412.9717.5010.7812.8612.5110.6211.6610.67
1411.5812.6617.4610.7812.8712.5310.6511.6710.59
1511.5712.8017.4210.7912.8912.5610.7711.6910.58
1611.5712.8117.3210.7912.9012.5710.7511.7010.56
1711.4312.8617.2210.8012.8912.5810.7711.6910.74
1811.4612.8617.2010.8112.8812.6010.8111.6710.51
1911.6912.9417.1810.8212.8912.6210.7811.6510.66
2011.9512.8417.1310.8412.9112.6410.6611.6510.70
2112.1812.3916.9910.8612.9312.6410.8211.6610.80
2212.3012.8216.7910.8812.9412.6410.7111.6610.88
2312.9012.6316.5610.9112.9612.6410.8411.6510.67
2413.7313.0816.3810.9412.9712.6510.7611.6510.84
Avg.11.8512.5016.9910.7312.5212.3610.6411.5610.57

References

  1. Desai, S.; Alhadad, R.; Mahmood, A.; Chilamkurti, N.; Rho, S. Multi-State Energy Classifier to Evaluate the Performance of the NILM Algorithm. Sensors 2019, 19, 5236. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Cheng, L.; Yu, T. A new generation of AI: A review and perspective on machine learning technologies applied to smart energy and electric power systems. Int. J. Energy Res. 2019, 43, 1928–1973. [Google Scholar] [CrossRef]
  3. Rathor, S.K.; Saxena, D. Energy management system for smart grid: An overview and key issues. Int. J. Energy Res. 2020, 44, 4067–4109. [Google Scholar] [CrossRef]
  4. Khan, A.R.; Mahmood, A.; Safdar, A.; Khan, Z.A.; Khan, N.A. Load forecasting, dynamic pricing and DSM in smart grid: A review. Renew. Sustain. Energy Rev. 2016, 54, 1311–1322. [Google Scholar] [CrossRef]
  5. Palensky, P.; Dietrich, D. Demand side management: Demand response, intelligent energy systems, and smart loads. IEEE Trans. Ind. Inform. 2011, 7, 381–388. [Google Scholar] [CrossRef] [Green Version]
  6. Fallah, S.N.; Deo, R.C.; Shojafar, M.; Conti, M.; Shamshirband, S. Computational Intelligence Approaches for Energy Load Forecasting in Smart Energy Management Grids: State of the Art, Future Challenges, and Research Directions. Energies 2018, 11, 596. [Google Scholar] [CrossRef] [Green Version]
  7. Zhang, X.; Wang, J.; Zhang, K. Short-term electric load forecasting based on singular spectrum analysis and support vector machine optimized by Cuckoo search algorithm. Electr. Power Syst. Res. 2017, 146, 270–285. [Google Scholar] [CrossRef]
  8. Son, M.; Moon, J.; Jung, S.; Hwang, E. A short-term load forecasting scheme based on auto-encoder and random forest. In Proceedings of the 3rd International Conference on Applied Physics, System Science and Computers (APSAC), Dubrovnik, Croatia, 26–28 September 2018; pp. 138–144. [Google Scholar] [CrossRef]
  9. Abbasi, R.A.; Javaid, N.; Ghuman, M.N.J.; Khan, Z.A.; Rehman, S.U. Short Term Load Forecasting Using XGBoost. In Workshops of the International Conference on Advanced Information Networking and Applications; Springer: Berlin/Heidelberg, Germany, 2019; pp. 1120–1131. [Google Scholar] [CrossRef]
  10. Moon, J.; Jung, S.; Rew, J.; Rho, S.; Hwang, E. Combination of short-term load forecasting models based on a stacking ensemble approach. Energy Build. 2020, 216, 109921. [Google Scholar] [CrossRef]
  11. Sajjad, M.; Khan, S.U.; Khan, N.; Haq, I.U.; Ullah, A.; Lee, M.Y.; Baik, S.W. Towards Efficient Building Designing: Heating and Cooling Load Prediction via Multi-Output Model. Sensors 2020, 20, 6419. [Google Scholar] [CrossRef] [PubMed]
  12. Lahouar, A.; Slama, J.B.H. Day-ahead load forecast using random forest and expert input selection. Energy Convers. Manag. 2015, 103, 1040–1051. [Google Scholar] [CrossRef]
  13. Moon, J.; Kim, K.-H.; Kim, Y.; Hwang, E. A Short-Term Electric Load Forecasting Scheme Using 2-Stage Predictive Analytics. In Proceedings of the IEEE International Conference on Big Data and Smart Computing (BigComp), Shanghai, China, 15–17 January 2018; pp. 219–226. [Google Scholar] [CrossRef]
  14. Park, S.; Moon, J.; Hwang, E. 2-Stage Electric Load Forecasting Scheme for Day-Ahead CCHP Scheduling. In Proceedings of the IEEE International Conference on Power Electronics and Drive System (PEDS), Toulouse, France, 9–12 July 2019. [Google Scholar] [CrossRef]
  15. Ryu, S.; Noh, J.; Kim, H. Deep Neural Network Based Demand Side Short Term Load Forecasting. Energies 2017, 10, 3. [Google Scholar] [CrossRef]
  16. Izonin, I.; Tkachenko, R.; Kryvinska, N.; Tkachenko, P.; Gregušml, M. Multiple Linear Regression Based on Coefficients Identification Using Non-Iterative SGTM Neural-Like Structure. In Proceedings of the International Work-Conference on Artificial Neural Networks, Gran Canaria, Spain, 12–14 June 2019; pp. 467–479. [Google Scholar] [CrossRef]
  17. Motepe, S.; Hasan, A.N.; Stopforth, R. Improving Load Forecasting Process for a Power Distribution Network Using Hybrid AI and Deep Learning Algorithms. IEEE Access 2019, 7, 82584–82598. [Google Scholar] [CrossRef]
  18. Kuan, L.; Yan, Z.; Xin, W.; Yan, C.; Xiangkun, P.; Wenxue, S.; Zhe, J.; Yong, Z.; Nan, X.; Xin, Z. Short-term electricity load forecasting method based on multilayered self-normalizing GRU network. In Proceedings of the 2017 IEEE Conference on Energy Internet and Energy System Integration, Beijing, China, 26–28 November 2018; pp. 1–5. [Google Scholar] [CrossRef]
  19. Chitalia, G.; Pipattanasomporn, M.; Garg, V.; Rahman, S. Robust short-term electrical load forecasting framework for commercial buildings using deep recurrent neural networks. Appl. Energy 2020, 278, 115410. [Google Scholar] [CrossRef]
  20. Sehovac, L.; Grolinger, K. Deep Learning for Load Forecasting Sequence to Sequence Recurrent Neural Networks with Attention. IEEE Access 2020, 8, 36411–36426. [Google Scholar] [CrossRef]
  21. Long-Term Energy Consumption & Outdoor Air Temperature for 11 Commercial Buildings—Datasets—OpenEI Datasets. Available online: https://openei.org/datasets/dataset/consumption-outdoor-air-temperature-11-commercial-buildings (accessed on 21 July 2020).
  22. Holidays and Observances around the World. Available online: https://www.timeanddate.com/holidays/ (accessed on 21 July 2020).
  23. Moon, J.; Park, S.; Rho, S.; Hwang, E. A comparative analysis of artificial neural network architectures for building energy consumption forecasting. Int. J. Distrib. Sens. Netw. 2019, 15. [Google Scholar] [CrossRef] [Green Version]
  24. Jung, S.; Moon, J.; Park, S.; Rho, S.; Baik, S.W.; Hwang, E. Bagging Ensemble of Multilayer Perceptrons for Missing Electricity Consumption Data Imputation. Sensors 2020, 20, 1772. [Google Scholar] [CrossRef] [Green Version]
  25. Dong-Nae Forecast (Digital Forecast). Available online: https://www.weather.go.kr/eng/weather/forecast/timeseries.jsp (accessed on 21 July 2020).
  26. Park, S.; Moon, J.; Jung, S.; Rho, S.; Baik, S.W.; Hwang, E. A Two-Stage Industrial Load Forecasting Scheme for Day-Ahead Combined Cooling, Heating and Power Scheduling. Energies 2020, 13, 443. [Google Scholar] [CrossRef] [Green Version]
  27. Moon, J.; Jung, S.; Park, S.; Hwang, E. Conditional Tabular GAN-Based Two-Stage Data Generation Scheme for Short-Term Load Forecasting. IEEE Access 2020, 8, 205327–205339. [Google Scholar] [CrossRef]
  28. Poniszewska-Maranda, A.; Kaczmarek, D.; Kryvinska, N.; Xhafa, F. Studying usability of AI in the IoT systems/paradigm through embedding NN techniques into mobile smart service system. Computing 2018, 101, 1661–1685. [Google Scholar] [CrossRef]
  29. Teslyuk, V.; Kazarian, A.; Kryvinska, N.; Tsmots, I. Optimal Artificial Neural Network Type Selection Method for Usage in Smart House Systems. Sensors 2021, 21, 47. [Google Scholar] [CrossRef]
  30. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  31. Guo, T.; Xu, Z.; Yao, X.; Chen, H.; Aberer, K.; Funaya, K. Robust online time series prediction with recurrent neural networks. In Proceedings of the 2016 IEEE International Conference on Data Science and Advanced Analytics (DSAA), Montreal, QC, Canada, 17–19 October 2016; pp. 816–825. [Google Scholar] [CrossRef] [Green Version]
  32. Cho, K.; van Merriënboer, B.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, 25–29 October 2014; Association for Computational Linguistics: Doha, Qatar, 2014; pp. 1724–1734. [Google Scholar]
  33. Klambauer, G.; Unterthiner, T.; Mayr, A.; Hochreiter, S. Self-Normalizing Neural Networks. In Proceedings of the Advances in Neural Information Processing Systems 30 (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  34. Jung, S.M.; Park, S.; Jung, S.W.; Hwang, E. Monthly Electric Load Forecasting Using Transfer Learning for Smart Cities. Sustainability 2020, 12, 6364. [Google Scholar] [CrossRef]
  35. Balasundaram, S.; Meena, Y. Robust Support Vector Regression in Primal with Asymmetric Huber Loss. Neural Process. Lett. 2019, 49, 1399–1431. [Google Scholar] [CrossRef]
  36. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention Is All You Need. In Advances in Neural Information Processing Systems 30; Guyon, I., Luxburg, U.V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., Garnett, R., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2017; pp. 5998–6008. [Google Scholar]
  37. Kwon, B.C.; Choi, M.J.; Kim, J.T.; Choi, E.; Kim, Y.B.; Kwon, S.; Sun, J.; Choo, J. Retainvis: Visual analytics with interpretable and interactive recurrent neural networks on electronic medical records. IEEE Trans. Vis. Comput. Graph. 2018, 25, 299–309. [Google Scholar] [CrossRef] [Green Version]
  38. Abadi, M.; Barham, P.; Chen, J.; Chen, Z.; Davis, A.; Dean, J.; Devin, M.; Ghemawat, S.; Irving, G.; Isard, M.; et al. TensorFlow: A System for Large-Scale Machine Learning. In Proceedings of the 12th USENIX Symposium on Operating Systems Design and Implementation, Savannah, GA, USA, 2–4 November 2016; pp. 265–283. [Google Scholar]
  39. Moon, J.; Kim, J.; Kang, P.; Hwang, E. Solving the Cold-Start Problem in Short-Term Load Forecasting Using Tree-Based Methods. Energies 2020, 13, 886. [Google Scholar] [CrossRef] [Green Version]
  40. Rahman, R.; Otridge, J.; Pal, R. IntegratedMRF: Random forest-based framework for integrating prediction from different data types. Bioinformatics 2017, 33, 1407–1410. [Google Scholar] [CrossRef]
  41. Chen, T.; Guestrin, C. XGBoost: A scalable tree boosting system. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 785–794. [Google Scholar] [CrossRef] [Green Version]
  42. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
Figure 1. Box plots for the electricity consumptions of three buildings (B1, B2, and B3 represent Building 1, Building 2, and Building 3, respectively).
Figure 1. Box plots for the electricity consumptions of three buildings (B1, B2, and B3 represent Building 1, Building 2, and Building 3, respectively).
Sensors 21 01639 g001
Figure 2. Gated recurrent unit (GRU) cell architecture.
Figure 2. Gated recurrent unit (GRU) cell architecture.
Sensors 21 01639 g002
Figure 3. Attention based GRU model architecture.
Figure 3. Attention based GRU model architecture.
Sensors 21 01639 g003
Figure 4. Scatter plot of electricity consumption and temperature for Building 1.
Figure 4. Scatter plot of electricity consumption and temperature for Building 1.
Sensors 21 01639 g004
Figure 5. Scatter plot of electricity consumption and temperature for Building 2.
Figure 5. Scatter plot of electricity consumption and temperature for Building 2.
Sensors 21 01639 g005
Figure 6. Scatter plot of electricity consumption and temperature for Building 3.
Figure 6. Scatter plot of electricity consumption and temperature for Building 3.
Sensors 21 01639 g006
Table 1. Summary of related works (RF, random forest; TSCV, time-series cross-validation; XGB, extreme gradient boosting; MLR, multiple linear regression; KEPCO, Korea Electric Power Corporation; DNN, deep neural network, DBN, deep belief network; SBTM, successive geometric transformations model; OP-ELM, optimally pruned extreme learning machines; LSTM, long short-term memory; GRU, gated recurrent unit; BiLSTM, bidirectional long short-term memory; CNN, convolution neural network; S2S, sequence to sequence; RNN, recurrent neural network).
Table 1. Summary of related works (RF, random forest; TSCV, time-series cross-validation; XGB, extreme gradient boosting; MLR, multiple linear regression; KEPCO, Korea Electric Power Corporation; DNN, deep neural network, DBN, deep belief network; SBTM, successive geometric transformations model; OP-ELM, optimally pruned extreme learning machines; LSTM, long short-term memory; GRU, gated recurrent unit; BiLSTM, bidirectional long short-term memory; CNN, convolution neural network; S2S, sequence to sequence; RNN, recurrent neural network).
ReferenceDatasetsAI MethodsGranularityCharacteristics
Lahouar and Slama [12]Tunisian Company of Electricity and GasRF with TSCVHourlyModel interpretation via variable importance
Moon et al. [13]Private university in Seoul, KoreaRF with TSCVDailyReflecting recent power consumption trends
Park et al. [14]Commercial and industrial buildings in KoreaStage 1: RF, XGB
Stage 2: MLR with sliding window
HourlyPerformance reinforcement
Ryu et al. [15]Demand side load data provided by KEPCODNN and DBNHourlyMultistep-ahead load forecasting
Izonin et al. [16]Combined cycle power plant datasetSGTMHourlyLess time consuming
Motepe et al. [17]South African power utility systemOP-ELM and LSTM30-minPerformance reinforcement
Kuan et al. [18]Electricity load from Junan, Shanhong provinceGRU15-minMultistep-ahead load forecasting
Chitalia et al. [19]Commercial buildings of five different types, in Bangkok, Hyderabad, Virginia, New York, and Massachusetts.LSTM/BiLSTM, LSTM/BiLSTM with attention, CNN+LSTM, CNN+BiLSTM, and convLSTM/convBiLSTM15-minRobust regardless of building types or locations
Sehovac and Grolinger [20]Commercial buildingS2S RNN with attention5-minPerformance reinforcement
Table 2. Energy consumption statistics of three buildings.
Table 2. Energy consumption statistics of three buildings.
StatisticalBuilding 1Building 2Building 3
WeekdaysHolidaysTotalWeekdaysHolidaysTotalWeekdaysHolidaysTotal
Minimum8.307.607.6010.109.409.4080.4580.4780.45
1st quartile17.4012.9014.7020.8017.0019.10124.11109.53117.22
Median33.9017.7023.5041.2020.9032.20199.72131.14157.82
Mean34.4320.0629.9840.1025.3735.53231.46138.17201.24
3rd quartile47.5023.3042.5053.4031.1248.80293.38157.44255.41
Maximum141.1077.30141.10135.0092.30135.00702.53336.77702.53
Standard deviation17.999.4517.1818.9612.1918.45128.1235.99115.86
Count18,120813626,25618,120813626,25617,784852026,304
LocationRichland, WashingtonRichland, WashingtonSeoul, South Korea
Public accessYesYesNo
Table 3. List of input variables.
Table 3. List of input variables.
No.Input VariableVariable Type
1MonthxContinuous [−1:1]
2MonthyContinuous [−1:1]
3DayxContinuous [−1:1]
4DayyContinuous [−1:1]
5HourxContinuous [−1:1]
6HouryContinuous [−1:1]
7Day of the weekxContinuous [−1:1]
8Day of the weekyContinuous [−1:1]
9HolidayBinary [1: holiday, 0: weekday]
10TemperatureContinuous
11Historical loadD−7Continuous
12Historical loadD−6Continuous
13Historical loadD−5Continuous
14Historical loadD−4Continuous
15Historical loadD−3Continuous
16Historical loadD−2Continuous
17Historical loadD−1Continuous
18Average loadContinuous
Table 4. Pearson correlation coefficients (PCCs) and p-values for each dataset.
Table 4. Pearson correlation coefficients (PCCs) and p-values for each dataset.
Input VariableBuilding 1Building 2Building 3
PCCp-ValuePCCp-ValuePCCp-Value
Monthx0.042***0.118***−0.156***
Monthy0.117***0.218***−0.177***
Dayx0.015**0.018*−0.009***
Dayy0.002***0.008***0.001***
Hourx0.134***0.139***−0.153***
Houry−0.541***−0.471***−0.640***
Day of the weekx0.333 0.322 0.110***
Day of the weeky−0.048***−0.053***−0.255***
Holiday−0.387***−0.369***−0.377***
Temperature0.014***−0.129***0.319***
Historical loadD−70.841*0.819***0.901**
Historical loadD−60.657***0.642***0.726***
Historical loadD−50.441***0.432***0.554**
Historical loadD−40.409***0.403***0.560***
Historical loadD−30.424***0.421***0.564***
Historical loadD−20.483***0.481***0.573***
Historical loadD−10.741***0.740***0.763***
Average load0.903***0.885***0.959***
(p-values: 0 ‘***’ 0.001 ‘**’ 0.01 ‘*’ 0.05 ‘.’ 0.1 ‘ ’ 1).
Table 5. Dataset period configuration.
Table 5. Dataset period configuration.
Dataset PeriodBuildings 1 and 2Building 3
Entire collection period2009/01/09–2011/12/312015/01/08–2017/12/31
Training set2009/01/09–2010/12/312015/01/08–2016/12/31
Testing set2011/01/01–2011/12/312017/01/01–2017/12/31
Table 6. Selected hyperparameters for each model.
Table 6. Selected hyperparameters for each model.
ModelsPackage or ModuleSelected Hyperparameters
MRFMultivariateRandomForestNo. of trees: 128 [39]
No. of features: 144 [39]
DNNTensorFlowNo. of hidden nodes: 289 [23]
No. of hidden layers: 5 [23]
Activation function: SELU [23]
Optimizer: Adam
Learning rate: 0.001
No. of epochs: 150
Batch size: 24
Park’s [14]
Stage 1: RF, XGB
Stage 2: MLR
scikit-learn
xgboost
RF
No. of estimators: 256
XGB
No. of estimators: 256
Max depth: 4
MLR
Sliding window size: 24
COSMOS [10]
Stage 1: DNNs
Stage 2: PCR
scikit-learnDNN
No. of hidden nodes: 13
No. of hidden layers: 2, 3, 4, 5
Activation function: ReLU
Optimizer: Adam
Learning rate: 0.001
No. of epochs: 150
Batch size: 24
PCR
Principal components: 1
Sliding window size: 168
Kuan’s [18]TensorFlowGRU
No. of cells: 100 [18]
No. of hidden layers: 5 [18]
Activation function: SELU [18]
Time step: 12 [18]
Batch size: 15 [18]
LSTM
ATT-LSTM
GRU,
ATT-GRU (Ours)
TensorFlowLSTM, GRU
No. of hidden nodes: 13
No. of hidden layers: 2
Activation function: SELU
Optimizer: Adam
Learning rate: 0.001
No. of epochs: 150
Batch size: 24
Table 7. MAPE (CVRMSE) comparison by months for Building 1.
Table 7. MAPE (CVRMSE) comparison by months for Building 1.
MonthMRFDNNPark’s [14]COSMOS [10]Kuan’s [18]LSTMATT-LSTMGRUATT-GRU
Jan.21.79 (27.40)20.36 (27.48)22.27 (30.62)13.89 (18.83)17.34 (26.21)14.10 (18.31)13.44 (18.06)13.86 (18.18)13.16 (17.77)
Feb.21.32 (29.31)21.60 (28.55)18.49 (27.67)16.46 (20.55)17.38 (27.93)16.68 (20.82)15.67 (20.44)16.45 (20.55)14.51 (19.63)
Mar.22.25 (32.89)22.16 (33.61)23.24 (32.71)14.85 (23.00)18.12 (32.12)15.24 (22.67)14.91 (20.61)14.88 (22.57)14.66 (21.33)
Apr.14.24 (19.97)13.26 (21.43)20.82 (37.57)8.71 (14.62)11.11 (18.92)10.72 (16.90)10.77 (15.81)10.77 (15.99)9.12 (13.37)
May7.97 (13.88)14.67 (22.25)16.06 (41.02)6.44 (11.71)12.13 (20.70)6.82 (12.97)7.12 (12.13)6.47 (13.51)6.39 (12.08)
Jun.12.54 (13.33)18.19 (24.70)20.99 (39.29)12.56 (12.08)15.72 (24.11)12.66 (14.94)14.14 (18.67)12.69 (15.48)12.18 (14.11)
Jul.14.22 (16.42)16.70 (27.32)19.21 (35.39)8.38 (11.37)16.10 (28.63)9.04 (16.11)7.76 (12.51)8.44 (15.07)7.67 (12.02)
Aug.14.22 (15.51)19.40 (28.45)13.39 (22.53)7.52 (9.47)17.30 (28.49)9.73 (11.11)8.24 (10.91)9.44 (11.74)7.01 (8.74)
Sep.14.20 (15.40)18.85 (26.70)28.65 (54.11)9.66 (13.12)17.18 (28.55)10.32 (12.44)11.41 (13.69)9.78 (12.54)8.96 (11.04)
Oct.18.84 (22.80)15.14 (23.40)17.94 (28.88)12.13 (16.54)17.20 (23.46)12.40 (14.14)13.67 (17.76)12.12 (13.82)11.26 (13.51)
Nov.31.33 (45.73)27.22 (42.51)44.11 (61.05)24.49 (40.11)32.57 (39.30)24.89 (35.82)20.88 (37.53)24.47 (35.22)20.76 (26.78)
Dec.45.11 (47.95)41.43 (43.57)41.43 (45.24)36.64 (38.13)43.11 (42.18)37.28 (39.63)31.76 (41.74)36.71 (37.88)29.12 (31.31)
Table 8. MAPE (CVRMSE) comparison by months for Building 2.
Table 8. MAPE (CVRMSE) comparison by months for Building 2.
MonthMRFDNNPark’s [14]COSMOS [10]Kuan’s [18]LSTMATT-LSTMGRUATT-GRU
Jan.15.81 (18.50)14.56 (20.43)13.95 (23.85)9.10 (11.80)13.35 (19.99)8.81 (12.64)7.64 (10.44)9.16 (12.12)7.16 (9.18)
Feb.14.68 (18.83)17.12 (21.76)13.05 (20.24)10.37 (13.99)14.25 (18.86)12.73 (15.03)10.66 (13.61)10.37 (14.06)9.58 (12.27)
Mar.12.58 (13.81)14.60 (17.78)13.32 (15.77)9.02 (11.30)14.56 (18.57)10.86 (9.19)9.81 (12.46)9.16 (11.98)9.09 (11.85)
Apr.15.82 (16.17)16.86 (21.08)20.23 (27.87)10.60 (13.39)17.90 (21.28)12.93 (13.98)10.32 (13.38)10.70 (12.40)10.88 (12.56)
May17.60 (16.80)17.62 (25.27)24.98 (34.07)7.92 (13.20)20.04 (24.90)8.70 (12.22)7.88 (11.99)9.08 (12.29)7.58 (11.44)
Jun.25.31 (18.22)22.32 (26.79)33.24 (37.19)10.31 (11.19)22.16 (30.44)10.17 (13.99)10.12 (13.45)10.39 (13.91)9.56 (12.60)
Jul.29.20 (21.19)27.35 (30.80)39.94 (46.91)7.63 (9.57)28.40 (37.39)8.82 (8.28)7.37 (9.67)8.69 (10.24)6.14 (8.15)
Aug.28.13 (19.27)29.73 (30.13)30.88 (30.37)7.74 (8.63)24.06 (34.45)10.42 (11.48)7.61 (10.10)11.31 (13.25)6.88 (9.55)
Sep.34.78 (23.66)27.71 (30.75)43.38 (55.64)9.56 (11.22)24.40 (34.78)9.90 (10.52)7.96 (10.43)8.67 (10.63)6.97 (9.56)
Oct.34.34 (28.69)19.98 (27.05)31.51 (34.68)10.18 (13.62)21.81 (30.59)10.23 (9.79)8.48 (8.20)9.44 (10.27)8.67 (8.37)
Nov.36.29 (37.48)33.78 (37.00)65.14 (62.40)21.93 (31.58)31.25 (48.13)19.78 (31.98)19.41 (28.97)20.75 (30.00)18.98 (26.88)
Dec.45.07 (39.50)42.15 (39.32)38.69 (39.17)32.50 (34.73)32.29 (38.27)32.17 (37.40)28.80 (36.18)33.89 (38.71)26.77 (33.10)
Table 9. MAPE (CVRMSE) comparison by months for Building 3.
Table 9. MAPE (CVRMSE) comparison by months for Building 3.
MonthMRFDNNPark’s [14]COSMOS [10]Kuan’s [18]LSTMATT-LSTMGRUATT-GRU
Jan.9.02 (12.02)14.95 (19.29)9.62 (11.71)9.38 (11.49)10.75 (12.37)9.19 (12.12)8.71 (11.16)9.33 (12.41)8.05 (10.81)
Feb.7.60 (8.94)13.83 (16.95)14.74 (33.66)7.65 (9.18)9.28 (15.61)10.02 (13.94)6.94 (8.32)9.03 (11.69)6.04 (7.66)
Mar.5.87 (8.89)11.28 (15.17)6.54 (8.93)4.53 (6.58)7.52 (10.48)5.76 (8.44)4.86 (7.16)7.61 (9.70)4.95 (7.70)
Apr.4.28 (5.83)8.93 (12.83)5.51 (7.62)2.99 (4.03)7.00 (8.93)4.07 (6.55)3.17 (5.49)7.59 (9.29)2.38 (3.50)
May6.06 (12.31)11.49 (16.55)8.26 (14.06)6.30 (11.95)7.84 (11.36)6.85 (8.91)5.23 (8.71)7.26 (9.72)5.11 (8.45)
Jun.7.17 (13.10)14.76 (23.38)6.75 (13.84)5.74 (10.89)10.05 (17.63)5.23 (7.19)5.09 (8.84)9.77 (11.90)4.48 (6.92)
Jul.10.18 (15.36)15.46 (26.38)9.33 (14.34)5.89 (8.63)12.26 (22.42)5.49 (9.54)5.63 (8.92)6.21 (8.14)5.33 (8.06)
Aug.11.54 (17.38)16.03 (26.11)9.51 (15.32)6.00 (8.95)10.09 (19.16)7.14 (9.16)6.22 (10.21)7.62 (8.87)5.57 (8.48)
Sep.6.26 (12.97)13.33 (19.38)9.24 (17.82)5.85 (12.00)9.26 (15.24)4.99 (6.19)4.38 (5.95)5.10 (6.00)3.57 (5.12)
Oct.8.08 (13.18)13.17 (18.16)7.64 (11.87)7.97 (14.05)8.04 (10.92)7.43 (11.92)7.51 (10.17)7.53 (10.55)7.20 (9.72)
Nov.7.50 (11.15)13.55 (18.28)8.76 (13.24)8.87 (12.67)9.32 (13.10)9.10 (14.75)8.52 (12.72)9.50 (13.99)8.10 (11.20)
Dec.10.37 (16.75)16.39 (23.36)11.81 (17.74)9.81 (17.04)12.88 (19.98)11.20 (17.59)9.96 (16.77)11.02 (17.36)9.79 (15.96)
Table 10. MAPE (CVRMSE) comparison by date types for Building 1. WD and HD denote weekdays and holidays, respectively.
Table 10. MAPE (CVRMSE) comparison by date types for Building 1. WD and HD denote weekdays and holidays, respectively.
TypeMRFDNNPark’s [14]COSMOS [10]Kuan’s [18]LSTMATT-LSTMGRUATT-GRU
Mon.18.75 (33.39)21.78 (33.81)21.76 (31.80)13.77 (25.03)15.47 (29.76)14.24 (27.35)13.47 (24.88)14.15 (26.12)12.86 (23.87)
Tue.17.57 (25.52)18.86 (33.85)32.52 (58.81)13.08 (21.53)14.22 (29.76)13.47 (21.61)12.31 (20.50)13.41 (21.41)12.06 (20.09)
Wed.18.06 (27.77)21.84 (33.14)18.63 (33.08)12.79 (19.81)17.79 (36.35)12.56 (19.70)12.23 (19.12)12.47 (20.23)11.23 (18.40)
Thu.18.77 (28.41)22.08 (30.76)16.27 (26.51)12.96 (19.89)19.17 (38.47)12.29 (19.36)10.52 (16.45)11.75 (18.31)9.93 (14.95)
Fri.21.20 (29.09)17.18 (27.93)17.18 (25.76)12.66 (23.02)17.05 (34.21)12.97 (19.08)12.66 (18.79)12.61 (19.51)11.43 (18.76)
Sat.25.21 (39.61)21.22 (31.65)33.97 (45.92)19.31 (33.02)13.90 (28.41)15.03 (28.16)13.02 (19.00)14.43 (26.45)12.41 (19.82)
Sun.19.30 (29.84)21.62 (32.29)26.92 (40.74)15.57 (29.39)14.88 (27.88)16.49 (29.22)15.48 (28.33)16.07 (29.54)14.66 (27.42)
WD18.18 (28.46)20.16 (31.98)20.22 (37.28)12.46 (21.32)15.61 (31.17)13.43 (19.10)12.37 (19.27)13.06 (19.16)11.86 (18.56)
HD23.58 (36.96)21.41 (32.04)32.20 (45.39)18.45 (33.35)14.39 (28.14)16.01 (28.57)14.26 (24.30)15.20 (27.96)13.93 (23.69)
Table 11. MAPE (CVRMSE) comparison by date types for Building 2. WD and HD denote weekdays and holidays, respectively.
Table 11. MAPE (CVRMSE) comparison by date types for Building 2. WD and HD denote weekdays and holidays, respectively.
TypeMRFDNNPark’s [14]COSMOS [10]Kuan’s [18]LSTMATT-LSTMGRUATT-GRU
Mon.23.20 (24.79)24.56 (30.73)26.13 (28.65)12.56 (19.32)17.96 (27.26)13.59 (20.18)12.64 (19.15)13.48 (20.29)12.44 (18.28)
Tue.22.42 (20.78)20.60 (25.60)33.89 (48.70)11.57 (17.54)12.57 (21.73)12.09 (18.19)11.53 (17.59)12.41 (18.86)11.12 (16.70)
Wed.21.02 (20.81)24.28 (30.05)24.85 (32.74)9.66 (14.90)12.05 (20.06)10.24 (15.24)9.24 (14.52)10.25 (15.28)9.01 (13.56)
Thu.21.71 (21.49)26.38 (31.33)24.63 (25.98)11.04 (16.95)16.75 (25.80)11.07 (16.50)10.06 (15.89)11.33 (16.42)9.67 (13.32)
Fri.25.72 (24.44)20.61 (25.22)26.50 (28.33)11.40 (21.24)18.92 (29.06)12.84 (20.86)10.96 (16.15)11.85 (19.37)10.11 (16.01)
Sat.36.37 (44.15)23.71 (28.44)43.78 (52.58)16.51 (28.31)21.34 (35.08)17.17 (26.07)15.60 (25.91)17.09 (26.47)14.58 (25.27)
Sun.30.82 (34.81)24.76 (31.16)36.05 (42.14)13.45 (26.88)22.66 (33.83)13.91 (23.52)12.91 (21.09)13.55 (23.84)12.69 (20.56)
WD22.12 (21.96)22.42 (28.97)25.62 (33.14)10.53 (17.13)18.85 (26.94)12.01 (19.08)10.45 (17.24)11.80 (19.14)10.15 (16.51)
HD34.39 (40.14)24.55 (30.86)42.61 (51.59)16.31 (30.69)21.80 (33.77)16.84 (32.83)15.98 (30.24)16.55 (31.92)13.98 (27.80)
Table 12. MAPE (CVRMSE) comparison by date types for Building 3. WD and HD denote weekdays and holidays, respectively.
Table 12. MAPE (CVRMSE) comparison by date types for Building 3. WD and HD denote weekdays and holidays, respectively.
TypeMRFDNNPark’s [14]COSMOS [10]Kuan’s [18]LSTMATT-LSTMGRUATT-GRU
Mon.7.89 (13.79)14.35 (23.50)8.24 (13.39)6.95 (11.24)9.65 (14.17)6.88 (11.61)6.47 (11.31)6.74 (11.64)6.20 (10.90)
Tue.7.60 (8.94)13.13 (21.35)9.79 (16.18)7.08 (11.41)8.05 (13.17)7.69 (12.51)7.42 (12.11)7.61 (12.40)6.79 (11.74)
Wed.5.87 (8.89)13.08 (20.18)8.85 (17.34)6.64 (10.23)6.32 (10.52)6.93 (10.25)6.61 (9.83)6.86 (10.13)6.05 (8.69)
Thu.4.28 (5.83)13.87 (23.24)7.19 (12.32)5.25 (8.91)7.47 (13.16)6.16 (9.97)4.70 (8.10)5.82 (9.13)4.40 (8.16)
Fri.6.06 (12.31)14.17 (23.64)10.43 (21.75)7.78 (12.44)8.12 (15.97)7.80 (12.84)7.01 (12.33)7.65 (12.67)6.54 (11.87)
Sat.7.17 (13.10)12.74 (18.25)9.27 (13.94)6.81 (10.23)8.15 (16.48)6.45 (10.90)6.10 (9.39)6.37 (10.21)5.93 (8.87)
Sun.10.18 (15.36)13.72 (22.04)8.81 (12.14)6.34 (8.88)8.08 (16.30)6.23 (8.58)5.99 (8.07)6.08 (8.93)5.17 (7.71)
WD6.89 (12.16)13.27 (22.67)8.48 (16.24)6.02 (10.30)8.52 (16.99)6.60 (10.99)5.43 (10.25)6.33 (10.67)5.52 (9.84)
HD9.79 (19.05)13.40 (21.09)9.88 (14.72)8.07 (12.68)8.51 (16.46)8.47 (12.89)7.78 (12.01)8.29 (12.70)7.95 (11.01)
Table 13. Training time of each model.
Table 13. Training time of each model.
Training TimeMRFDNNPark’s [14]COSMOS [10]Kuan’s [18]LSTMATT-LSTMGRUATT-GRU
Building 1361503274134510,74510,95211,58169747108
Building 23724982811426 10,76911,76112,16973627658
Building 3364500282138810,69310,09710,99268857023
Average367500279138610,73610,93711,58170747263
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Jung, S.; Moon, J.; Park, S.; Hwang, E. An Attention-Based Multilayer GRU Model for Multistep-Ahead Short-Term Load Forecasting. Sensors 2021, 21, 1639. https://doi.org/10.3390/s21051639

AMA Style

Jung S, Moon J, Park S, Hwang E. An Attention-Based Multilayer GRU Model for Multistep-Ahead Short-Term Load Forecasting. Sensors. 2021; 21(5):1639. https://doi.org/10.3390/s21051639

Chicago/Turabian Style

Jung, Seungmin, Jihoon Moon, Sungwoo Park, and Eenjun Hwang. 2021. "An Attention-Based Multilayer GRU Model for Multistep-Ahead Short-Term Load Forecasting" Sensors 21, no. 5: 1639. https://doi.org/10.3390/s21051639

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop