Next Article in Journal
Comparing Novel MMIC and Hybrid Circuit High Efficiency GaAs Schottky Diode mm-Wave Frequency Doublers
Previous Article in Journal
Attention Collaborative Autoencoder for Explicit Recommender Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Ultra-Short-Term Forecasting of Photo-Voltaic Power via RBF Neural Network

1
School of Electronic and Information, Shanghai Dianji University, Shanghai 201306, China
2
NARI Technology Co., Ltd., Nanjing 211106, China
*
Author to whom correspondence should be addressed.
Electronics 2020, 9(10), 1717; https://doi.org/10.3390/electronics9101717
Submission received: 14 September 2020 / Revised: 10 October 2020 / Accepted: 13 October 2020 / Published: 18 October 2020
(This article belongs to the Section Power Electronics)

Abstract

:
With the fast expansion of renewable energy systems during recent years, the stability and quality of smart grids using solar energy have been challenged because of the intermittency and fluctuations. Hence, forecasting photo-voltaic (PV) power generation is essential in facilitating planning and managing electricity generation and distribution. In this paper, the ultra-short-term forecasting method for solar PV power generation is investigated. Subsequently, we proposed a radial basis function (RBF)-based neural network. Additionally, to improve the network generalization ability and reduce the training time, the numbers of hidden layer neurons are limited. The input of neural network is selected as the one with higher Spearman correlation among the predicted power features. The data are normalized and the expansion parameter of RBF neurons are adjusted continuously in order to reduce the calculation errors and improve the forecasting accuracy. Numerous simulations are carried out to evaluate the performance of the proposed forecasting method. The mean absolute percentage error (MAPE) of the testing set is within 10%, which show that the power values of the following 15 min. can be predicted accurately. The simulation results verify that our method shows better performance than other existing works.

1. Introduction

With the growing scarcity of fossil energy, the use of sustainable energy for power generation has increasingly attracted the attention of global power industry. The fact that solar energy is the largest sustainable energy source has put photo-voltaic (PV) power generation in the spotlight. However, because the output power depends on variable meteorological factors, the PV generation system reveals a high level of uncertainty and periodicity. The integration of large scales PV energy into the power grid will challenge the power balance of the grid, leading to a series of disturbance. Therefore, the development of large-scale grid-connected PV power plants is limited [1]. The accurate prediction of PV power generation is an effective way to provide guidance for grid operation scheduling and regulation of the PV system itself, and to reduce the instability of PV grid-connected operation. These are strong motivations for ultra-short-term forecasting of the output of PV power plants.
In terms of the specific forecasting horizon, the forecasting can be divided into: ultra-short-term forecasting (a few minutes to 1 h ahead), short-term forecasting (1 h to several hours ahead), medium-term forecasting (several hours to one week ahead), and long-term forecasting (one week to one year or more) [2]. According to different forecasting logics, PV power generation forecasting methods are often classified into two basic types: direct forecasting methods and indirect forecasting methods.
The direct forecasting methods are linear modeling methods, which are also known as classical physical modeling methods. These methods combine the geographical location and numerical weather forecast of the PV generator in order to obtain solar vertical illuminance data, and then the principle of physical power generation of the PV system is used to develop linear mathematical models of PV power generation; finally, the output power of PV systems is directly predicted, without historical power generation data [3,4,5,6].
The indirect forecasting methods are non-linear forecasting methods, which mainly contain statistical methods [7] and machine learning methods, such as nonlinear multivariable regression [8], support vector machine [9], k-means clustering [10], artificial neural network [11,12], etc. The physical principle of PV power generation is not considered in these methods, while historical output power data and meteorological environmental factors are analyzed generally. Subsequently, these methods can obtain the output power of the system based on the PV power generation forecasting model.
From the articles published in recent years, the machine learning methods were increasingly applied to build PV power forecasting models by relevant scholars [13,14]. For example, a prediction model inspired of similar-day detection engine was proposed for day-ahead small-scale solar power output forecasting in [15]. In [16], a high-precision deep neural network model was proposed, in which the PV output in the next 24 hours was predicted. The results verified that the PVPNet algorithm for short-term forecasting achieved good forecasting performance combined with strong generalization ability; however, the quantity of input features was relatively too large. In [17], a grey-box model that is based on the hourly forecast of the global irradiance and quantile random forest model consisting of multiple classification models were discussed. The results showed that the improvement of more complicated approach was consistent over months in a year. In [18], a PV model was constructed upon the quantile regression forests, which was trained with forecasts of several meteorological variables that were produced by a numerical weather prediction model, and spatial and temporal indexes was estimated from the forecasted variables. A median and confidence interval were calculated in order to predict the hourly power of the PV plant. In [19], a statistical spatio-temporal method was proposed to forecast PV plant power output in six hours. A procedure inspired from the clear sky index for solar radiation was employed to address the non-stationarity issue of the production series. In [20], the aerosol index, which was able to detect the presence of UV-absorbing aerosols such as dust and soot, was added to the input data of a one-day-ahead PV output forecasting model based on BP neural network. When compared with a model that only choses traditional input features, the precision of the model using the aerosol index was indicated to be highly improved in both sunny and cloudy weather conditions.
In addition, the combination of artificial neural networks and various other methods were applied to the study of PV output forecasting [21,22,23,24,25,26]. For example, the statistical features of historical PV power time series were extracted by the stationary wavelet transform (SWT) that was proposed in [27], and a final deep neural network (DNN) was designed to make use of the reconstructed values estimated by four LSTM networks to predict the final forecasted value of the next 30-min. PV power. The grey theory was used in data processing in order to improve the forecasting accuracy and address the over-fitting problem presented in [28]. A deep belief network (DBN) model was combined with a feed-forward network (FFN) for day-ahead PV power output forecasting, which was superior than DBN alone in forecasting accuracy. Bayesian optimization in [29] was applied to select the optimal combined features, and a deep learning model was constructed based on a LSTM block and an embedding block with the connection of a merge layer in order to predict the day-ahead PV power output. A support vector regression (SVR) method was proposed in one-hour-ahead PV power prediction [30]. Big data and minimum redundancy maximum relevance (mRMR) technology were applied to input features dimension reduction, while GA was used to optimize the learning parameters inside SVR machine; thus, the computational speed and the forecasting precision were both increased. The wavelet packet decomposition (WPD) in [31] was utilized to decompose the original PV power series into four sub-series, each of which was fed into a LSTM networks. Subsequently, the final one-hour-ahead PV power forecasting results were obtained via a linear weighting method, reaching an average RMSE in four seasons of 0.2357 kW. In [32], a Weight-Varying Combination Forecast Mode (WVCFM) model was developed in order to generate deterministic point predictions for following 5 min, achieving an mean absolute percentage error (MAPE) under 7.1 % . In [33], the radiation classification coordinate (RCC) method was applied to process data set, and the RCC-LSTM model reached MAPE of 2.74∼7.25% in predicting the PV power generation of a next day. A novel hybrid improved Kmeans-GRA-Elman (HKGE) model was proposed in [34] for one-day-ahead PV power prediction with one-hour interval. The average daily root mean square error (RMSE) and R 2 of the HKGE model are 4.3210 kW and 0.9953, which outweighed other eight compared models.
However, these existing indirect forecasting methods are generally employed in a large forecasting time span, ranging from several hours to one day, and the model structure or the data processing method was complex, while the forecasting accuracy is limited. Meanwhile, the problem of how to accurately predict the PV power in ultra-short-term, reduce the complexity of the network structure, and improve the calculation speed and forecasting accuracy, has not been well solved.
In this paper, the ultra-short-term forecasting method that is based on radial basis function (RBF) neural network for solar PV power generation is investigated. RBF neural network is characterized by simple structure, good generalization ability and fast learning convergence rate. To promote these advantages, the superior limit number of hidden layer neurons is set. Moreover, the input of neural network is reduced according to their Spearman correlation with the power value to be predicted. To reduce the calculation errors, the data are normalized. To improve the forecasting accuracy in generating next 15 min, forecasts, the expansion parameter of RBF neurons is adjusted continuously, which affects the sensitivity of neurons to the input data. From the comparison with other three existed forecasting model, it is verified that the proposed model achieves excellent performance in both forecasting accuracy and robustness.
The paper is organized, as follows. In Section 2, the mathematical principle and hyperparameters of RBF neural network are presented. In Section 3, the obtained sample data are prepossessed. In Section 4, the forecasting model is proposed, appropriate parameters are determined via repeatedly tests, and the experimental results and comparisons with three existed forecasting methods are analyzed. Finally, Section 5 gives the conclusions.

2. Mathematical Model of RBF Neural Network

2.1. Structure of RBF Neural Network

The RBF neural network is a forward neural network. It is a three-layer feed-forward neural network that is based on high-dimensional spatial interpolation technology. RBF is designed as the activation function of the neuron, making the RBF neural network able to approximate any continuous function with arbitrary precision. Thanks to the strong nonlinear approximation combined with the simple network structure, it is often used for function approximation and analysis.
When compared with the more well-known standard back-feed propagation neural network (BPNN) that requires multiple hidden layers, the RBF neural network only employs one hidden layer. To achieve the same training accuracy, as a local approximation network, the later often spends less training time than the BPNN. Moreover, when the number of training samples increase, the performance of RBF network learning is better [35]. Therefore, in this paper, the RBF neural network is used to approximate the ultra-short-term output of the PV power generation system.
The standard form of the RBF neural network is a fixed three-layer structure of “input layer-hidden layer-output layer”. The structure of the RBF neural network in the form of “N-M-L” is shown in Figure 1 [36].
Here, x i i = 1 , 2 , , N denotes the data fed into i-th input node, N denotes the number of input nodes. c j j = 1 , 2 , , M denotes the center vector of the j-th radial basis neuron and M denotes the number of radial basis neurons in the hidden layer. y k k = 1 , 2 , , L denotes the k-th output and L denotes the number of output nodes.
The first layer of RBF neural network is the input layer. The number of nodes in this layer is the same as the dimension of input vectors. The input layer only plays the role of transmitting signals, so that the input vectors can be directly fed into the hidden space.
The second layer is the hidden layer, which is the most critical layer of the RBF neural network. The activation function that is used in hidden neurons is the RBF, as the previous section introduced. The number of RBF neurons in this layer depends on the complexity of the problem to be solved and the amount of data provided. In fact, RBF neurons only respond to a relatively small area of the input space, resulting that the larger number of input vectors and its ranging, the more RBF neurons required in order to achieve the same training accuracy. The goal of this layer is to perform a non-linear transformation of the input vector from the input space to the high-dimensional feature space (usually M > N ). Subsequently, in this high-dimensional feature space, a linear model is employed in order to train samples. Thus, through spatial transformation, a non-linear problem is transformed into a linear problem, thereby simplifying calculations. The process of training the hidden layer can be described as adjusting the parameters of the activation function, including the center vectors and the thresholds. A nonlinear optimization strategy is applied in this process, so the learning speed is comparatively slower than other layers. The output a j j = 1 , 2 , , M of j-th neuron in hidden layer can be calculated as
a j = Φ x c j · b j 1 = e x c j · b j 1 2 ,
where the input vector x R N denotes the position or coordinate of the input and R N denotes N-dimensional real number space. c j R N denotes the center vector or the position of j-th neuron in hidden layer, also called the weight vector of the neuron.
x c j denotes the dot product of the input vector x and the center vector c j , which is also known as the Euclidean distance, and it is calculated, as follows:
x c j = x c j T x c j 1 2 .
b j 1 denotes the threshold or sensitivity of j-th node in hidden layer with the superscript presenting the hidden layer, and this parameter regulates the response range of neurons to the input intuitively.
Φ denotes the radial basis activation function of the RBF neuron, which is a radially symmetric and non-negative function. The Gaussian function is often chosen as the radial basis transfer function in RBF neural networks because of the excellent properties of smoothness, radial symmetry, and simple form. The Gaussian function is demonstrated in Figure 2, with n denoting the input of the function and a denoting the output, and the relation between them is expressed as
a = Φ n = e n 2 .
The third layer is the output layer, and the number of nodes in this layer depends on the dimension of the output vectors. The output of each hidden layer neurons is linearly weighted, and then the final output of RBF network is generated in this layer. The linear optimization strategy of the LS method is applied in training the linear weights in output layer, leading the learning speed relatively fast. The output y k k = 1 , 2 , , L of k-th node in the output layer can be expressed as
y k = j = 1 M w j , k a j + b k 2 ,
where w j , k denotes the weight of j-th RBF neuron to k-th output node, and b k 2 denotes the threshold of k-th node in output layer with the superscript presenting the output layer [37].

2.2. Learning Algorithm of RBF Neural Network

The precise and approximate RBF neural network design functions are provided in the MATLAB platform. The precise type makes the number of hidden layer neurons directly equal to the number of training input samples, and its calculation speed is consequently fast, while the approximate type sets the upper limit of the number of hidden layer neurons, which is more effective when the number of training samples is large [38]. Obviously, the approximate type is more suitable for the case of more than 10,000 sets of data in the training set of this paper.
The function in MATLAB to design an approximate radial basis network is n e t = n e w r b ( P , T , G O A L , S P R E A D , M , D F ) , where the parameters P, T, G O A L , S P R E A D , M, and D F denote the input matrix, output matrix of the training set, the mean squared error goal, the maximum number of hidden neurons, and the number of neurons to add between displays, respectively.
With the target error and maximum number of hidden layer nodes set as G O A L an M separately, the training process of the approximate RBF neural network with training set data is as following.
Step 1: calculate the loss function of the simulating network with all the training samples. According to the LS method, the loss function can be expressed as
E = 1 N i = 1 N T ^ i T i 2 ,
where N denotes the number of training samples in training set, T ^ i denotes the forecasting output of the model, and T i denotes the true output.
Step 2: find the sample x that contributes the most to the network error.
Step 3: add a radial basis neuron to the hidden layer, and set its center vector c to the transpose of the input vector of the training sample, which is
c = x T ,
and set its threshold as b, which is determined by another parameter S P R E A D in MATLAB as following.
b = 1 l n 2 S P R E A D ,
which means that the neuron responds an output higher than 0.5 to any input vector with an Euclidean distance to its center vector within the value of S P R E A D .
Step 4: readjust the weights of each neuron in the hidden layer to each neuron in the linear output layer according to LS method, in order to minimize the network error after adding RBF neurons.
Step 5: re-update the MES of the network. Repeat the previous steps 2 to 4 until the specified error performance index G O A L is met, or the upper limit of the number of hidden layer neurons M is breached.
As the steps finished, the function n e w r b returns a new radial basis network n e t .

3. Data Analysis and Process

The data https://www.dcjingsai.com/v2/cmptDetail.html?id=309 used in this article are the meteorological data and power generation data of one PV power station in China from 1 January 2017 to 15 November 2018 https://github.com/AceBBang/PV-forecasting-data. Meteorological data and power generation data are measured at 15-min. intervals. In order to simplify the experiment, the data at night and invalid data are excluded. The invalid data mainly include some days with no change in solar irradiance, PV power output, or other types of data. Finally, more than 20,000 sets of data are obtained, including wind speed, wind direction, ambient temperature, pressure, humidity, solar irradiance, and power generation at 48 quarters during the day time.

3.1. Data Analysis

Solar panels for PV power generation rely on radiant energy in sunlight and the “PV effect” of the panels to generate electricity. Their working conditions are subject to two major factors. One is the internal electrical properties of the PV power plant, including the efficiency of photoelectric conversion devices, aging degree of PV modules, ash accumulation of PV array panels, differences in installation of panels, performance differences of transmission lines and inverters, losses in power conversion and transmission, etc. The other type is relatively external environmental factors, not only meteorological environmental factors, such as solar irradiance, weather type, season, temperature, wind speed, and cloud cover, but also geographical environment factors, including latitude and longitude of the PV system, topographical conditions, and nearby buildings.
For a given PV power plant, all output power data are provided by the same set of PV power generation components. Therefore, the time series of historical power generation has taken the impact of system information and internal factors of the PV array into account. Moreover, during the service life of the system, most of the qualified PV inverter modules operate in a relatively stable state for a long time. Although the power conversion efficiency and photoelectric conversion efficiency can change with time, the amount of change over a long period is so slight that it can be considered to be constant in short-term forecasts. Therefore, it is not necessary to consider the factors that are implicit in the historical output power data, including the power conversion efficiency of the inverter, the PV conversion efficiency of the PV array, the total area of the PV array, and the geographical environment of the power station.
Distinct from the geographic environmental factors and internal electrical factors of PV power plants, meteorological environmental factors, such as solar irradiance, ambient temperature, and humidity, can experience active changes. Even within a few minutes, non-negligible fluctuations can occur, and it is verified that they introduce a large influence into the performance of power generation [39,40]. Accordingly, meteorological factors should be taken as input features of the forecasting model.
Because of the hysteresis of photoelectric effect, there is inertia behind changes in its influencing factors, such as solar irradiance and ambient temperature. The time series of PV power station power generation shows a high autocorrelation, which is known as the sequence correlation. This means that the output power value at the next time step is closely connected with the outputs of the several historical time steps. Therefore, in addition to the meteorological environment factors, another important input feature is the historical power generation data of the PV system.
However, the amount of meteorological data and historical data are too large and, if a large amount of input features are fed into the network, the speed and accuracy of network calculation would be greatly reduced, and the computational burden of the network would be increased. Therefore, it is necessary to select some features with high correlation with the output power of the time to be predicted. Moreover, the correlation coefficients between these feathers and the data to be predicted are required.
In statistics, the Pearson correlation coefficient is commonly used to measure the degree of correlation between two sets of data, but it requires that the data of the two variables are normally distributed as a whole, and only the linear correlation and direction of the two variables are given. When compared with this, the analysis of Spearman correlation coefficient pays more attention to whether the changing trends of the two variables are the same, and it is less sensitive to anomalous data and more applicable. Thus, the Spearman correlation coefficient is more suitable for the non-linearity of PV power generation forecasting studied in this paper. The Spearman correlation coefficient between set X and set Y can be expressed as following,
ρ = i = 1 N x i x ¯ y i y ¯ i = 1 N x i x ¯ 2 i = 1 N y i y ¯ 2 ,
where x i and y i is the rank of elements X i in X and Y i in Y , respectively. x ¯ and y ¯ is the average of the rank of all elements in X and Y , respectively. N is the number of elements contained in X and Y . The ranking method can be described as: every element is arranged in descending order, with the corresponding rank of the largest value being 1 and the smallest N. If there are equal values of some elements, then the rank of these elements takes the average of the positions of these equal elements.
The function ρ = c o r r ( X , Y , t y p e , S p e r m a n ) in MATLAB calculates the Spearman correlation coefficient between X and Y , returning the number ρ .
To choose the highly correlated features, Figure 3 and Figure 4 shows the correlation coefficient between the power value 15 min later, the moment to be predicted, and meteorological data and historical output power data, respectively, with the specific results of the correlation analysis on these variables labeled. The label P ( t i ) ( i = 0 , 1 , 2 , 3 , 4 , 5 , 6 ) presents the output power before i quarter or quarters.
It can be seen from Figure 3 that only the correlation of irradiance reaches above 0.8 , while the correlation coefficients of other meteorological parameters are all below 0.5. Meanwhile, Figure 4 illustrates that the correlation generally decreases with the increase of time interval. Therefore, in order to ensure the input features quality and streamline the features, irradiation I ( t 0 ) , generated power P ( t 0 ) , P ( t 1 ) , P ( t 2 ) , P ( t 3 ) , and P ( t 4 ) with ρ > 0.7 are selected as the input features for the forecasting model.

3.2. Data Processing

When considering that RBF neural network learning belongs to supervised learning, this paper divides the sample data into three data sets according to a certain ratio: a collection of 60 % data constitutes the training set, which is used to train the forecasting model under different hyperparameters, including the number of hidden layer neurons, sensitivity coefficient, target error, etc. 20 % is divided into the validation set, which is used in orderto verify the forecasting accuracy of the trained neural network, so as to select the most suitable hyperparameters. Finally, the rest 20 % is the testing set, which is used to finally check the generalization performance of the forecasting model.
The solar irradiance and PV power generation been chosen as features with different orders of magnitude and dimensions, using data directly may lead to low network convergence speed, oversaturation of hidden layer neurons, numerical problems (i.e., due to the calculation accuracy of MATLAB, in the case of a large difference in the order of magnitude, the two quantities that are actually obviously equal will be misjudged as unequal), and other problems. Therefore, before the network training, the data should be normalized. When considering that the average relative error percentage is needed later, in order to avoid the result that the relative error is calculated as infinite when the original output is 0, all the data in the training set are normalized to the interval [ 0.1 , 0.9 ] .
The formula for normalizing set X to set Y is as following
y = y m a x y m i n x x m i n x m a x x m i n + y m i n ,
where x is the element to be normalized in the original sequence and y is the corresponding element after normalization, y m i n and y m a x are normalized interval boundary values, which are 0.1 and 0.9 here, x m i n and x m a x are the minimum and maximum values in the to-be-processed set X , respectively.
The function [ Y , P S ] = m a p m i n m a x ( X , y m i n , y m a x ) provided in MATLAB returns the normalized set Y and the process settings, which are x m i n and x m a x .
Subsequently, with the function Y = m a p m i n m a x ( a p p l y , X , P S ) , testing set data combined with validation set data can be normalized according to (9) applying the same x m i n and x m a x as in training set.
All of the forecasted value will be reverse normalized before the final forecasted results are obtained through the function X = m a p m i n m a x ( r e v e r s e , Y , P S ) .

3.3. Performance Metrics

The PV power forecasting model is used in order to predict the amount of power generation, and the pursuit of constant research, modeling, and improvement is to continuously reduce errors. Therefore, scientific evaluation of errors is particularly important. There are five common types of performance evaluation index tables for the power forecasting model used in [41].
The mean bias error (MBE)
MBE = 1 N i = 1 N T ^ i T i ,
where the N is the number of testing instances, T ^ i is the forecasting output of the model, and T i is the true output.
The mean absolute error (MAE)
MAE = 1 N i = 1 N T ^ i T i .
The mean absolute percentage error (MAPE)
MAPE = 1 N i = 1 N T ^ i T i T i .
The mean square error (MSE)
MSE = 1 N i = 1 N T ^ i T i 2 .
The root mean square error (RMSE)
RMSE = 1 N i = 1 N T ^ i T i 2 .

4. Experiments and Results Analysis

This section reports the process of model establishment and the experiment and comparison results for performance evaluation in order to validate the effectiveness of the proposed model.

4.1. Model Establishment

The six-dimensional data with high correlation with power P ( t + 1 ) are selected as the network input features to predict the PV output. The input features consist of the solar irradiance at current moment I ( t 0 ) and the PV output at current as well as historical moment, including the PV output from P ( t 0 ) to P ( t 4 ) , while the network output is P ( t + 1 ) , the output of the PV power generation system at the next moment. Figure 5 shows the "6-input-1-output" RBF neural network structure to be used in this paper.
After the training data, validation data and the test data divided in the previous section, the proposed PV power forecasting model is established according to the process that is shown in Figure 6.
The PV power forecasting model is generated with the training function of approximate RBF neural networks on MATLAB, which means fewer hidden layer neurons than the number of input data sets are required for training.
Subsequently, the training set and the verification set are simulated respectively through the simulating function to optimize the model parameters. In this process, the iterative trial method [42] is applied in order to optimize the maximum of hidden layer neurons and the expansion parameter S P R E A D . With the optimized model, the samples of testing set, which are not applied in the adjusting process, are used to test the performance of the forecasting model.
After several experiments, it is found that the greater the number of hidden layer neurons, the higher the learning accuracy, but the slower the training speed,, as illustrated in Figure 7. For clear demonstration, the error denotes the ratio of each error type with different hidden neurons to the corresponding error of the model with 20 hidden neurons with a fixed S P R E A D on validation set in Figure 7. When there are 120 neurons in the hidden layer, a better learning performance can be achieved, and the calculation speed of the training is not too slow with the training time slightly higher than 300 s. In addition, as the hidden neurons increased, the MSE and RMSE begin to increase again.
The expansion parameter S P R E A D in MATLAB functionally replaces b, and it adjusts the sensitivity of the hidden layer neurons to the input vector. The S P R E A D value should be large enough for the RBF neurons to respond to overlapping areas of the input space, but a too large S P R E A D may lead to all neurons responding to various input in the same degree. Hence, it is necessary to call the RBF network design function multiple times with different S P R E A D values and try to find the best value. Based on this, the RBF network design function is called in a loop structure with Continuously changing S P R E A D s. When comparing the errors in different S P R E A D cases, the expansion parameter with effective performance is selected. The learning effect can be demonstrated by small error in the training set, and the generalization ability can be proved by small error in the testing set.
The predetermined requirements that are mentioned in the flowchart mainly include that the RMSE of both the training set and the validation set drop below 10%, and the gap is not too large. Finally, under the premise of ensuring that the learning effect and generalization ability meet the requirements, the expansion parameter is set to be 0.5700 . Additionally, the structure of RBF neural network is “6-120-1” structure, which means 6 nodes in the input layer, 120 neurons in the hidden layer, and 1 output node.

4.2. Results Analysis

Figure 8 shows the mean square error (MSE) curve of the training set in the process of gradually adding hidden layer neurons to the forecasting model.
In Figure 8, the horizontal axis denotes the number of epochs, and in each epoch the entire training set data is passed forward the neural network once. One hidden layer neuron is addicted after each epoch during the RBF neural network training process. The vertical axis is a logarithmic axis, denoting the MSE value of the network, which highlights the magnitude of the error. The curve shows that, with the continuous increase of radial basis neurons, the magnitude of the network error decreases rapidly at first and then gently, which illustrates the convergence of the algorithm clearly.
Next, the training set, validation set, and testing set are separately simulated, and various errors of the three data sets are calculated. Table 1 shows the specific values of the five types of error evaluation indicators for these three data sets.
MBE is the average of the bias errors of the output by the network, which is also known as the average error, and it reflects the overall average deviation of the predicted value from the true value. The MBE of the verification set is greater than 0, and the test set is less than 0, which indicates that the predicted value of the PV power ultra-short-term forecasting model is generally higher for the validation set and lower for the testing set, respectively.
MAE is the average value of the absolute error of the predicted output. Analyzing it can avoid the situation where the positive and negative errors cancel out; thereby, the total absolute error reduced to a certain extent. Therefore, MAE can better reflect the magnitude of the absolute error between the predicted value and the actual value than MBE. The MAE of all three data sets is greater than the absolute value of MBE, which means that the error values of the forecasting results of the model are offset partly.
MAPE refers to the average value of the absolute value of the relative error between the predicted value and the true value, and then multiplied by 100 % . This indicator can well reflect the degree of confidence in the forecasting. The MAPEs of the training set, validation set, and testing set all drop less than 10 % , meeting the main requirements for the forecasting model in the previous section.
MSE is the default evaluation index for measuring network errors when MATLAB creates a neural network primarily owing to its simple calculation, presenting the sum of the absolute errors. The MSE of each data set is between 0.3 % and 0.4 % , which means that the learning effect of the forecasting model is better.
The RMSE is calculated from the square root of the mean square error. Its value is particularly sensitive to the maximum and minimum errors in a set of predicted values, so the dispersion of the forecasting can be well indicated. The RMSEs are all around 6 % , indicating that the deviation between the predicted value and the true value is very slight, and the forecasting model reaches a high degree of confidence.
The generalization performance of the model refers to robustness, which is, for untrained data, the forecasting model can also give forecasting values close to the true value. Therefore, a large testing set error is equivalent to the poor robustness of the model. From the simulation results, the low error of the testing set (the MAPE is below 10 % and the RMSE is slightly above 6 % ) fully reflects the high generalization ability of the model, and it is indicated that the forecasting task for the unlearned samples is finished accurately.
From the training set, the number of neurons in the hidden layer of the model is 120, which is much lower than the number of training samples of 14 , 301 , indicating that the model efficiently utilizes the RBF neurons. When compared to the number of training samples, the average utilization rate of each neuron is increased by 119 times. In addition, the large number of training set samples will inevitably impact on the error. However, even so, the training set error also meets the requirements, that is, the MAPE is below 10 % , which also shows that the model has excellent robustness and strong generalization ability.
To show the forecasting performance more intuitively and concisely, Figure 9 illustrates a five-day forecasting curve randomly selected from the testing set. The forecasting value is rather close to the actual value in general. The moment with a large absolute value of error is often the moment when the output power of the PV system is fluctuating. In addition, the more severe the fluctuation, the more unpredictable the forecast; the smoother the output power, the smaller the error.
Fve error types of proposed model are compared with other three contemporary prediction models, including BPNN, general regression neural network (GRNN), and linear regression (LR), in order to fully validate the performance of the proposed PV forecasting model based on RBF neural network.
Figure 10 presents the comparison of the five type of errors on testing set among the proposed approach and the other three methods. With the same input data and training set, the errors of four methods differ. For comparison, the scales of MBEs and MSEs are adjusted in Figure 10.
Although the MBE of the proposed model is greater than other models, which merely indicates the total bias on testing set is larger than other models, the MAE, MAPE, MSE, and RMSE of proposed forecasting model are superior to those of the other models. Because the comparison is conducted on a testing set, the satisfied performance of proposed model on both forecasting accuracy and robustness is revealed similarly.

4.3. Applicability Analysis of the Model in Four Generalized Weather Types

The forecasting models from four general weather types are discussed in [29,43,44], while the PV power ultra-short-term forecasting model that is proposed in this paper does not distinguish different weather types. To quantify how well the forecasting performance of the forecasting model under different weather conditions is, this section analyzes the applicability of the forecasting performance of the model for different weather types.
From the test set that is not involved in model training and verification, samples in four generalized weather types are collected. Figure 11 shows the comparison of the forecasting curves.
The comparison of the two error types under the four generalized weather types is shown in Figure 12, and the specific error values are shown in Table 2.
From Figure 12 and Table 2, it can be seen that the forecasting error in the sunny type is the smallest, while, in the remaining three, weather conditions are larger.
On the sunny day, the weather conditions are relatively stable, and the change in solar irradiance over time is relatively smooth and less fluctuating. The autocorrelation of the PV power time series is the strongest. For the RBF network, the output and input data in this case are highly correlated; therefore, the proposed forecasting model can reflect the tendency of the PV power generation accurately.
In cloudy, overcast, and rainy/snowy weather, the cloud covering the PV power plant often changes sharply, thus the fluctuation of solar irradiance in a short period of time is relatively considerable. The autocorrelation of the PV power time series is weak, so the forecasting effect is poor.
Figure 11 shows the fluctuations of power generation during these four days. Obviously, the amount of power generation changes most gently on a sunny day and most dramatically on a cloudy day, which corresponds to the difference in error values over four days, as mentioned above. As the degree of fluctuations of weather types increases, the forecasting accuracy of the ultra-short-term forecasting model gradually decreases. Although the error of the forecasting model is relatively large for some days with fluctuating power generation, the forecasting performance of the model is, in general, still excellent.
To further evaluate the performance of the proposed model, the comparisons of MAPEs and RMSEs of each model in each weather types are provided in Figure 13 and Figure 14, separately. The figures show that the proposed model remarkably outperforms other three ultra-short-term PV power prediction models in most of four seasons, although, in sunny day, the RMSE of BPNN is slightly lower, and in rainy or snowy day, the LR method performs marginally better with respect to MAPE and RMSE.

5. Conclusions

The increasingly penetration of PV power has challenged the planning and operation of the existing power system. An ultra short-term PV power forecasting model based on RBF neural network has been proposed in this paper in order to reduce the adverse effects of the intermittency and randomicity of PV power generation. By comparing the Spearman correlation coefficients of each feature with the power value to be predicted, the solar irradiance and historical output power have been selected as the input of the proposed network, in order to generate the prediction of the PV power after 15 min. The appropriate number of hidden layer neurons and expansion constant have been applied, and the structure and parameters of the optimal network have been fixed. Due to the limitation of hidden layer neuron numbers, the simple structure and short training time of model has been achieved. The performance of the proposed model has been verified to be accurate by simulation with the error MAPE of testing set below 10 % . The comparisons of proposed model with BPNN, GRNN and LR in testing set and four general weather types have validated the superior performance of the proposed PV power forecasting model.

Author Contributions

Formal analysis, W.M. and Z.C.; Funding acquisition, Z.C. and Q.Z.; Methodology, W.M.; Project administration, Q.Z.; Visualization, W.M. and Z.C.; Writing—original draft, W.M.; Writing—review and editing, Z.C. and Q.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded in part by the National Natural Science Foundation of China (Grant No. 61801112), the Natural Science Foundation of Jiangsu Province (Grant No. BK20180357), the Open Program of State Key Laboratory of Millimeter Waves at Southeast University (Grant No. K202029).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Tao, C.; Shanxu, D.; Changsong, C. Forecasting Power Output for Grid-Connected Photovoltaic Power System Without Using Solar Radiation Measurement. In Proceedings of the 2nd International Symposium on Power Electronics for Distributed Generation Systems, Hefei, China, 16–18 June 2010; pp. 773–777. [Google Scholar]
  2. Verma, A.; Tripathi, M.; Upadhyay, K.; Kim, H.j. A Review Article on Green Energy Forecasting. Asia Pac. J. Multimed. Serv. Converg. Art Humanit. Sociol. 2016, 6, 637–647. [Google Scholar] [CrossRef]
  3. Arbizu-Barrena, C.; Ruiz-Arias, J.A.; Rodriguez-Benitez, F.J.; Pozo-Vazquez, D.; Tovar-Pescador, J. Short-Term Solar Radiation Forecasting by Advecting and Diffusing MSG Cloud Index. Sol. Energy 2017, 155, 1092–1103. [Google Scholar] [CrossRef]
  4. Hu, K.; Cao, S.; Wang, L.; Li, W.; Lv, M. A New Ultra-Short-Term Photovoltaic Power Prediction Model Based on Ground-Based Cloud Images. J. Clean. Prod. 2018, 200, 731–745. [Google Scholar] [CrossRef]
  5. Wang, F.; Zhen, Z.; Liu, C.; Mi, Z.; Hodge, B.M.; Shafie-khah, M.; Catalao, J.P. Image Phase Shift Invariance Based Cloud Motion Displacement Vector Calculation Method for Ultra-Short-Term Solar PV Power Forecasting. Energy Convers. Manag. 2018, 157, 123–135. [Google Scholar] [CrossRef]
  6. Wang, F.; Li, K.; Wang, X.; Jiang, L.; Ren, J.; Mi, Z.; Shafie-Khah, M.; Catalao, J.P.S. A Distributed PV System Capacity Estimation Approach Based on Support Vector Machine With Customer Net Load Curve Features. Energies 2018, 11, 1750. [Google Scholar] [CrossRef] [Green Version]
  7. Yang, M.; Huang, X. Ultra-Short-Term Prediction of Photovoltaic Power Based on Periodic Extraction of PV Energy and LSH Algorithm. IEEE Access 2018, 6, 51200–51205. [Google Scholar] [CrossRef]
  8. Kim, G.G.; Choi, J.H.; Park, S.Y.; Bhang, B.G.; Nam, W.J.; Cha, H.L.; Park, N.; Ahn, H. Prediction Model for PV Performance With Correlation Analysis of Environmental Variables. IEEE J. Photovoltaics 2019, 9, 832–841. [Google Scholar] [CrossRef]
  9. Shi, J.; Lee, W.; Liu, Y.; Yang, Y.; Wang, P. Forecasting Power Output of Photovoltaic Systems Based on Weather Classification and Support Vector Machines. IEEE Trans. Ind. Appl. 2012, 48, 1064–1069. [Google Scholar] [CrossRef]
  10. McCandless, T.; Haupt, S.; Young, G. A Regime-Dependent Artificial Neural Network Technique for Short-Range Solar Irradiance Forecasting. Renew. Energy 2016, 89, 351–359. [Google Scholar] [CrossRef] [Green Version]
  11. De Giorgi, M.G.; Congedo, P.M.; Malvoni, M. Photovoltaic Power Forecasting Using Statistical Methods: Impact of Weather Data. IET Sci. Meas. Technol. 2014, 8, 90–97. [Google Scholar] [CrossRef]
  12. Khan, I.; Zhu, H.; Khan, D.; Panjwani, M.K. Photovoltaic Power Prediction by Cascade Forward Artificial Neural Network. In Proceedings of the 2017 International Conference on Information and Communication Technologies (ICICT), Karachi, Pakistan, 30–31 December 2017; pp. 145–149. [Google Scholar]
  13. Carriere, T.; Vernay, C.; Pitaval, S.; Kariniotakis, G. A Novel Approach for Seamless Probabilistic Photovoltaic Power Forecasting Covering Multiple Time Frames. IEEE Trans. Smart Grid 2020, 11, 2281–2292. [Google Scholar] [CrossRef] [Green Version]
  14. Ran, C.; Xia, X.; Yang, M.; Zhang, Z.; Li, Y.; Zeng, X.; Huang, H.; Teng, X.; Cai, Y.; Cao, B. BP Network PV Power Forecast Based on Daily Type and Fusion Theory. Zhongnan Daxue Xuebao (Ziran Kexue Ban)/J. Cent. South Univ. (Sci. Technol.) 2018, 49, 2232–2239. [Google Scholar]
  15. Zhang, Y.; Beaudin, M.; Taheri, R.; Zareipour, H.; Wood, D. Day-Ahead Power Output Forecasting for Small-Scale Solar Photovoltaic Electricity Generators. IEEE Trans. Smart Grid 2015, 6, 2253–2262. [Google Scholar] [CrossRef]
  16. Huang, C.; Kuo, P. Multiple-Input Deep Convolutional Neural Network Model for Short-Term Photovoltaic Power Forecasting. IEEE Access 2019, 7, 74822–74834. [Google Scholar] [CrossRef]
  17. Gigoni, L.; Betti, A.; Crisostomi, E.; Franco, A.; Tucci, M.; Bizzarri, F.; Mucci, D. Day-Ahead Hourly Forecasting of Power Generation From Photovoltaic Plants. IEEE Trans. Sustain. Energy 2018, 9, 831–842. [Google Scholar] [CrossRef] [Green Version]
  18. Almeida, M.P.; Perpinan, O.; Narvarte, L. PV Power Forecast Using a Nonparametric PV Model. Sol. Energy 2015, 115, 354–368. [Google Scholar] [CrossRef] [Green Version]
  19. Agoua, X.G.; Girard, R.; Kariniotakis, G. Short-Term Spatio-Temporal Forecasting of Photovoltaic Power Production. IEEE Trans. Sustain. Energy 2018, 9, 538–546. [Google Scholar] [CrossRef] [Green Version]
  20. Kou, J.; Liu, J.; Li, Q.; Fang, W.; Chen, Z.; Liu, L.; Guan, T. Photovoltaic Power Forecasting Based on Artificial Neural Network and Meteorological Data. In Proceedings of the 2013 IEEE International Conference of IEEE Region 10 (TENCON 2013), Xi’an, China, 22–25 October 2013; pp. 1–4. [Google Scholar]
  21. Wang, X.; Huang, K.; Zheng, Y.; Li, L.; Shao, F.; Jia, L.; Xu, Q. Combined PV Power Forecast Based on Firefly Algorithm-Generalized Regression Neural Network. Dianwang Jishu/Power Syst. Technol. 2017, 41, 455–461. [Google Scholar]
  22. Capizzi, G.; Napoli, C.; Bonanno, F. Innovative Second-Generation Wavelets Construction With Recurrent Neural Networks for Solar Radiation Forecasting. IEEE Trans. Neural Netw. Learn. Syst. 2012, 23, 1805–1815. [Google Scholar] [CrossRef] [Green Version]
  23. Almonacid, F.; Rus, C.; Perez-Higueras, P.; Hontoria, L. Calculation of the Energy Provided by a PV Generator. Comparative Study: Conventional Methods vs. Artificial Neural Networks. Energy 2011, 36, 375–384. [Google Scholar] [CrossRef]
  24. Lee, W.; Kim, K.; Park, J.; Kim, J.; Kim, Y. Forecasting Solar Power Using Long-Short Term Memory and Convolutional Neural Networks. IEEE Access 2018, 6, 73068–73080. [Google Scholar] [CrossRef]
  25. Sanjari, M.J.; Gooi, H.B. Probabilistic Forecast of PV Power Generation Based on Higher Order Markov Chain. IEEE Trans. Power Syst. 2017, 32, 2942–2952. [Google Scholar] [CrossRef]
  26. Han, Y.; Wang, N.; Ma, M.; Zhou, H.; Dai, S.; Zhu, H. A PV Power Interval Forecasting Based on Seasonal Model and Nonparametric Estimation Algorithm. Sol. Energy 2019, 184, 515–526. [Google Scholar] [CrossRef]
  27. Ospina, J.; Newaz, A.; Faruque, M.O. Forecasting of PV Plant Output Using Hybrid Wavelet-Based LSTM-DNN Structure Model. IET Renew. Power Gener. 2019, 13, 1087–1095. [Google Scholar] [CrossRef]
  28. Chang, G.W.; Lu, H. Integrating Gray Data Preprocessor and Deep Belief Network for Day-Ahead PV Power Output Forecast. IEEE Trans. Sustain. Energy 2020, 11, 185–194. [Google Scholar] [CrossRef]
  29. Yang, T.; Li, B.; Xun, Q. LSTM-Attention-Embedding Model-Based Day-Ahead Prediction of Photovoltaic Power Output Using Bayesian Optimization. IEEE Access 2019, 7, 171471–171484. [Google Scholar] [CrossRef]
  30. Liu, J.; Sun, H.; Chang, P.; Jiao, Z.; Wei, P.; Ke, X.; Sun, X.; Cheng, L. Research of Photovoltaic Power Forecasting Based on Big Data and mRMR Feature Reduction. In Proceedings of the 2018 IEEE Power Energy Society General Meeting (PESGM), Portland, OR, USA, 5–10 August 2018; pp. 1–5. [Google Scholar]
  31. Li, P.; Zhou, K.; Lu, X.; Yang, S. A Hybrid Deep Learning Model for Short-Term PV Power Forecasting. Appl. Energy 2019, 259, 114216. [Google Scholar] [CrossRef]
  32. Liu, L.; Zhao, Y.; Chang, D.; Xie, J.; Ma, Z.; Sun, Q.; Yin, H.; Wennersten, R. Prediction of short-term PV power output and uncertainty analysis. Appl. Energy 2018, 228, 700–711. [Google Scholar] [CrossRef]
  33. Chen, B.; Lin, P.; Lai, Y.; Cheng, S.; Chen, Z.; Wu, L. Very-Short-Term Power Prediction for PV Power Plants Using a Simple and Effective RCC-LSTM Model Based on Short Term Multivariate Historical Datasets. Electronics 2020, 9, 289. [Google Scholar] [CrossRef] [Green Version]
  34. Lin, P.; Peng, Z.; Lai, Y.; Cheng, S.; Chen, Z.; Wu, L. Short-term power prediction for photovoltaic power plants using a hybrid improved Kmeans-GRA-Elman model based on multivariate meteorological factors and historical power datasets. Energy Convers. Manag. 2018, 177, 704–717. [Google Scholar] [CrossRef]
  35. Hui-Qiang, Z.; Zeng-Jun, Y.; Liang, T. A Comparative Study on BP Network and RBF Network in Function Approximation. Bull. Ence Technol. 2005. [Google Scholar]
  36. Liu, J. RBF Neural Network Control for Mechanical Systems Design, Analysis and MATLAB Simulation; Tsinghua University Press: Beijing, China; Springer: Berlin, Germany, 2013. [Google Scholar]
  37. Zhang, D. MATLAB Neural Network Programming; Chemical Industry Press: Beijing, China, 2011. [Google Scholar]
  38. Gu, Y. MATLAB R2016a: 27 Cases of Neural Network Design And Application; Publishing House of Electronics Industry: Beijing, China, 2018. [Google Scholar]
  39. Xiao, W.; Dai, J.; Wu, H.; Nazario, G.; Cheng, F. Effect of Meteorological Factors on Photovoltaic Power Forecast Based on the Neural Network. RSC Adv. 2017, 7, 55846–55850. [Google Scholar] [CrossRef] [Green Version]
  40. Salam, A.; Oral, B.; Gorgulu, S. Measurements of Meteorological Parameter Effects on Photovoltaic Energy Production. Int. J. Circuits Syst. Signal Process. 2015, 9, 240–246. [Google Scholar]
  41. Bo, J.; Lunnong, T.; Zheng, Q.; Yan, P.; Jingyi, W. An overview of research progress of short-term photovoltaic forecasts. Electr. Meas. Instrum. 2017, 12, 1–6. [Google Scholar]
  42. Kermanshahi, B. Recurrent Neural Network for Forecasting Next 10 Years Loads of Nine Japanese Utilities. Neurocomputing 1998, 23, 125–133. [Google Scholar] [CrossRef]
  43. Brecl, K.; Topic, M. Photovoltaics (PV) System Energy Forecast on the Basis of the Local Weather Forecast: Problems, Uncertainties and Solutions. Energies 2018, 11, 1143. [Google Scholar] [CrossRef] [Green Version]
  44. Tan, J.; Deng, C. Ultra-Short-Term Photovoltaic Generation Forecasting Model Based on Weather Clustering and Markov Chain. In Proceedings of the 2017 IEEE 44th Photovoltaic Specialist Conference (PVSC), Washington, DC, USA, 25–30 June 2017; pp. 1158–1162. [Google Scholar]
Figure 1. Structure of radial basis function (RBF) neural network in N-M-L form.
Figure 1. Structure of radial basis function (RBF) neural network in N-M-L form.
Electronics 09 01717 g001
Figure 2. Gaussian function of RBF neuron.
Figure 2. Gaussian function of RBF neuron.
Electronics 09 01717 g002
Figure 3. Correlation analysis of photo-voltaic (PV) power and meteorological parameters.
Figure 3. Correlation analysis of photo-voltaic (PV) power and meteorological parameters.
Electronics 09 01717 g003
Figure 4. Autocorrelation analysis of PV power.
Figure 4. Autocorrelation analysis of PV power.
Electronics 09 01717 g004
Figure 5. RBF neural network for PV forecasting model structure.
Figure 5. RBF neural network for PV forecasting model structure.
Electronics 09 01717 g005
Figure 6. The flowchart of PV power forecasting.
Figure 6. The flowchart of PV power forecasting.
Electronics 09 01717 g006
Figure 7. Training time and comparative performance of models with different number of hidden neurons.
Figure 7. Training time and comparative performance of models with different number of hidden neurons.
Electronics 09 01717 g007
Figure 8. Mean square error (MSE) curve during RBF neural network training.
Figure 8. Mean square error (MSE) curve during RBF neural network training.
Electronics 09 01717 g008
Figure 9. A random five-day forecasting curve.
Figure 9. A random five-day forecasting curve.
Electronics 09 01717 g009
Figure 10. Comparative performance of the proposed method.
Figure 10. Comparative performance of the proposed method.
Electronics 09 01717 g010
Figure 11. The comparison of forecast results under four generalized weather types.
Figure 11. The comparison of forecast results under four generalized weather types.
Electronics 09 01717 g011
Figure 12. Performance of forecast results under four generalized weather types.
Figure 12. Performance of forecast results under four generalized weather types.
Electronics 09 01717 g012
Figure 13. The mean absolute percentage errors (MAPEs) of four PV forecasting models.
Figure 13. The mean absolute percentage errors (MAPEs) of four PV forecasting models.
Electronics 09 01717 g013
Figure 14. The RMSEs of four PV forecasting models.
Figure 14. The RMSEs of four PV forecasting models.
Electronics 09 01717 g014
Table 1. Performance of forecasting model.
Table 1. Performance of forecasting model.
Types of ErrorTraining SetValid SetTesting Set
MBE(%) 0.0000 0.0434 0.2108
MAE(%) 2.1053 2.4795 1.6432
MAPE(%) 9.3716 9.8142 9.6427
MSE(%) 0.3385 0.3741 0.3742
RMSE(%) 5.8184 6.1168 6.1174
Table 2. Forecasting performance under four generalized weather types.
Table 2. Forecasting performance under four generalized weather types.
Weather PatternSunnyOvercastCloudyRainy or Snowy
MAPE(%) 5.2184 9.0483 12.6302 9.5702
RMSE(%) 2.7498 6.6906 7.3988 5.1635
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Ma, W.; Chen, Z.; Zhu, Q. Ultra-Short-Term Forecasting of Photo-Voltaic Power via RBF Neural Network. Electronics 2020, 9, 1717. https://doi.org/10.3390/electronics9101717

AMA Style

Ma W, Chen Z, Zhu Q. Ultra-Short-Term Forecasting of Photo-Voltaic Power via RBF Neural Network. Electronics. 2020; 9(10):1717. https://doi.org/10.3390/electronics9101717

Chicago/Turabian Style

Ma, Wanxing, Zhimin Chen, and Qing Zhu. 2020. "Ultra-Short-Term Forecasting of Photo-Voltaic Power via RBF Neural Network" Electronics 9, no. 10: 1717. https://doi.org/10.3390/electronics9101717

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop