Next Article in Journal
Alternative Design of Double-Suction Centrifugal Pump to Reduce the Effects of Silt Erosion
Next Article in Special Issue
Comparison of LSSVR, M5RT, NF-GP, and NF-SC Models for Predictions of Hourly Wind Speed and Wind Power Based on Cross-Validation
Previous Article in Journal
A Modified One-Cycle-Control Method for Modular Multilevel Converters
Previous Article in Special Issue
Using Smart Persistence and Random Forests to Predict Photovoltaic Energy Production
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Probabilistic Wind Power Forecasting Approach via Instance-Based Transfer Learning Embedded Gradient Boosting Decision Trees

School of Electronic Information and Electrical Engineering, Shanghai Jiao Tong University, Shanghai 200240, China
*
Author to whom correspondence should be addressed.
Submission received: 3 December 2018 / Revised: 25 December 2018 / Accepted: 1 January 2019 / Published: 3 January 2019
(This article belongs to the Special Issue Solar and Wind Energy Forecasting)

Abstract

:
With the high wind penetration in the power system, accurate and reliable probabilistic wind power forecasting has become even more significant for the reliability of the power system. In this paper, an instance-based transfer learning method combined with gradient boosting decision trees (GBDT) is proposed to develop a wind power quantile regression model. Based on the spatial cross-correlation characteristic of wind power generations in different zones, the proposed model utilizes wind power generations in correlated zones as the source problems of instance-based transfer learning. By incorporating the training data of source problems into the training process, the proposed model successfully reduces the prediction error of wind power generation in the target zone. To prevent negative transfer, this paper proposes a method that properly assigns weights to data from different source problems in the training process, whereby the weights of related source problems are increased, while those of unrelated ones are reduced. Case studies are developed based on the dataset from the Global Energy Forecasting Competition 2014 (GEFCom2014). The results confirm that the proposed model successfully improves the prediction accuracy compared to GBDT-based benchmark models, especially when the target problem has a small training set while resourceful source problems are available.

1. Introduction

Wind energy has grown to an extent whereby its impact on the power system has become relevant in many regions. According to a report by the World Wind Energy Association (WWEA), the total capacity of wind turbines installed worldwide reached 539 GW at the end of 2017 [1]. In recent years, an increasing number of countries, including Germany, Ireland, Portugal, Spain, Sweden, and Uruguay, have reached a double-digit wind power share. Considering the higher variability of wind power generation, the increased wind power penetration level will introduce considerable uncertainties to power systems, resulting in higher requirements on power system transmission capacity, reserve capacity and system flexibility [2]. Probabilistic wind power forecasting is proposed to model the uncertainty of wind power by estimating the probability distribution of wind power generation. Based on the probability distribution, many decision-making applications, including unit commitment [3,4,5,6,7,8], wind power trading [9,10,11,12], reserve procurement [13,14], demand response [15,16], probabilistic power flow [17,18], and economic dispatch [19,20,21], can be developed.
To model the wind power distribution, some researchers developed a point wind power forecasting model, and subsequently, parametric models [22,23,24,25,26,27,28] and non-parametric models [29,30,31,32] were built to model the error distribution. In recent years, much attention has been paid to quantile regression. In the wind power forecasting track of the Global Energy Forecasting Competition 2014 (GEFCom2014) [32], three of the five most-effective models incorporated quantile regression to solve the wind power distribution. By utilizing the pinball loss function, quantile regression estimates the wind power quantiles directly, which has been shown to be an effective probabilistic wind power forecasting method. For example, a quantile regression forest model and a stacked random forest-gradient boosting decision trees (GBDT) model were built in [33]. These two models form a voted ensemble for forecasting the probability distribution of wind power. In [34], a quantile linear regression model was built, whereby the non-linear feature transformations of the input variable are taken as the model input. The winner of the GEFCom2014 wind power track developed a gradient boosted machine (GBM) approach for multiple quantile regression, therein fitting each quantile and zone independently [35]. Notably, the model developed in [35] utilizes the information on correlated wind farms in the other zones. The case study shows that when the information of the wind farms in the other zones is used as the input, a 2.5% decrease in the pinball loss can be achieved. Similarly, rather than solely using one numerical weather predictions (NWP) sample point measured at the location of the power plant, the work in [36] explored information from the NWP data of both the local zone and the nearby zones, which constitutes a spatial grid of NWP. By constructing new variables from the raw NWP data of the NWP grid, [36] significantly improved the forecast skill of state of the art forecasting systems by 16.09% and 12.85% for solar and wind power, respectively. The success of [35,36] suggested that introducing information of the wind farms in other zones is a viable technique to reduce the prediction error.
To utilize the information of wind power generations in other zones, this paper proposes a wind power quantile forecasting method based on instance-based transfer learning. Transfer learning focuses on solving a specific problem (target problem) using knowledge gained from different, but related problems (source problems) [37]. With this knowledge, transfer learning can improve the results achieved on a target problem [38]. Many successful applications can be found in visual adaptation [39] and text classification [40]. In instance-based transfer learning, the data of the related problems are used as training examples for the target problem [41,42]. For wind power quantile regression, the wind power generation of the wind farm in different zones shows spatial cross-correlation [43]. Therefore, it is reasonable to apply instance-based transfer learning techniques to reduce forecasting error on the target zone by introducing the data of related wind farms to the training set.
To review existing works that use instance-based transfer learning in forecasting, the work in [44] focused on adding data from the source problems (auxiliary training data) to the training process together with the data from the target problem (base training data). As the source problem set may mix with unrelated problems and the resulting negative knowledge transfers can degrade the prediction accuracy [45], thus [44] introduced a source problem selection technique based on the covariance coefficient of the load vectors. However, the selected auxiliary training data from different source problems are directly added to the training set and treated equally, neglecting the fact that the relatedness to the target problem is not the same for each source problem.
In this paper, an IBT-GBDT (instance-based transfer learning embedded gradient boosting decision trees) model is proposed. The gradient boosting decision trees algorithm is very effective in wind power generation probabilistic forecasting, which is chosen as the core forecasting method. Following the instance-based transfer learning technique, the base training data from the target problem and auxiliary training data from the source problems together constitute the training set, but with different weights assigned. To derive the weights, this paper analyzes two types of errors, i.e., random errors and systematic errors. Then, the formula for the weights, which considers the distribution of the errors, is given. However, the errors are unknown before the model is trained; thus, in practice, based on the theoretical analysis, the weights of the auxiliary training sets are solved by iteration, whereas the weight of the base training set is a hyperparameter chosen by cross-validation. As a result of the advantages of transfer learning, the combination successfully utilizes information about other zones. With the enlarged training set, the model is well trained using the GBDT algorithm, which results in the improvement in prediction accuracy. The IBT-GBDT model is tested on a public dataset from the wind track of GEFCom2014. This dataset consists of 10 zones. Measured by quantile forecasting score (QS), the results show that the IBT-GBDT method can increase the forecasting accuracy for the target zone, especially when this target problem has many closely-correlated problems and a small training set.
To the best of our knowledge, this is the first time that instance-based transfer learning has been used for wind power quantile regression. The contributions of this paper are three-fold:
  • Instance-based transfer learning is utilized to increase the forecasting accuracy of probabilistic wind power forecasting. Different weights are assigned to different auxiliary training sets according to their relatedness to the target problem to reflect the real relatedness between source problems and the target problem. Based on the maximum likelihood method, the theoretical formula for the weights is derived.
  • A unique method for solving the weights is proposed in this paper. The weight for a target zone is a hyperparameter chosen by cross-validation, and the weights for the source problems are solved by iteration.
  • Several GBDT-based benchmark models are developed in this paper to illustrate the effect of the instanced-based transfer learning method. Compared to those benchmark models, the IBT-GBDT model earned the highest prediction accuracy.
The remainder of this paper is organized as follows. In Section 2, the mathematical formulation of the proposed IBT-GBDT model and the corresponding training method are described. Case studies are conducted to validate the proposed approach in Section 3. Finally, Section 4 gives a summary of this paper.

2. The Instance-Based Transfer Learning Embedded Gradient Boosting Decision Trees

At the beginning of this section, the methodology of GBDT and its application in quantile regression are introduced. Then, the architecture of the IBT-GBDT model is described. Finally, the weight formula and the corresponding iterative weight-solving algorithm are derived.

2.1. Gradient Boosting Decision Trees for Quantile Regression

2.1.1. Pinball Loss Function and Weighted Pinball Loss

The pinball loss function is an error measure for quantile regression. Given a target percentile τ , f τ ( · ) denotes the predictive function for the wind power quantile. The corresponding pinball loss, denoted by l τ ( y , f τ ( X ) ) , is defined as:
l τ ( y , f τ ( X ) ) = τ ( y f τ ( X ) ) , y > f ø ( X ) ( 1 τ ) ( f τ ( X ) y ) , y f ø ( X ) .
where X denotes the model inputs such as the weather, wind speed, and wind direction. y denotes the actual wind power generation.
For a sample set denoted by S , by summing the pinball losses across all samples from S , the pinball loss of the set S is obtained. As some data points contribute more than others, large weights are assigned to the samples with higher contributions. In this case, the weighted pinball loss for the sample set S is defined as the weighted sum of all the pinball losses of samples from S . Let ( X i , y i ) denote the ith sample of S and w i denote the corresponding weight (the determination of w i depends on ( X i , y i ) , which is introduced in detail in Section 2.3). The weighted pinball loss is defined as:
L τ ( y , f τ ( X ) ) = i = 1 N w i l τ ( y i , f τ ( X i ) ) = τ y i > f τ ( X i ) w i ( y i f τ ( X i ) ) + ( 1 τ ) y i f τ ( X i ) w i ( f τ ( X i ) y i ) .

2.1.2. Gradient Boosting Decision Trees with Weighted Pinball Loss

Gradient boosting is a machine learning technique for regression and classification problems, which produces a prediction model in the form of an ensemble of base learners. Typically, gradient boosting is used with decision trees of a fixed size as base learners, e.g., Classification And Regression Trees (CART). By utilizing the weighted pinball loss, the use of gradient boosting decision trees (GBDT) becomes an effective technique for quantile regression.
GBDT builds the model in a stage-wise manner. In every stage, the target is to fit the residual error. f ( t ) denotes the fitted function after stage t. The number of boosting stages is denoted by T. Before the tth stage, the residual error for point i is:
ε i ( t 1 ) = y i f ( t 1 ) ( X i ) .
v i ( t ) denotes the fitted value on the ith sample in stage t.
f ( t ) ( X i ) = f ( t 1 ) ( X i ) + v i ( t )
After the residual error is fitted, the updated weighted pinball loss for the sample set S becomes:
L τ ( t ) = τ v i ( t ) < ε i ( t 1 ) w i ( ε i ( t 1 ) v i ( t ) ) + ( 1 τ ) v i ( t ) ε i ( t 1 ) w i ( v i ( t ) ε i ( t 1 ) ) .
The gradient of L τ ( t ) on v i ( t ) is:
g i ( t ) = L τ ( t ) v i ( t ) = w i τ , v i ( t ) < ε i ( t 1 ) w i ( 1 τ ) , v i ( t ) ε i ( t 1 ) .
Recall the definition of gradient boosting. The base learner should fit the negative gradient of the loss function ( g i ( t ) ). When choosing CART as the base learner, the updated weighted pinball loss is:
L τ ( t ) = τ k = 1 K ( t ) ρ k ( t ) < ε i ( t 1 ) w i ( ε i ( t 1 ) ρ k ( t ) ) + ( 1 τ ) k = 1 K ( t ) ρ k ( t ) ε i ( t 1 ) w i ( ρ k ( t ) ε i ( t 1 ) ) ,
where ρ k ( t ) denotes the value of the kth tree leaf in stage t and K ( t ) denotes the number of tree leaves in stage t. To minimize the weighted pinball loss function L τ ( t ) , we take the partial derivative of L τ ( t ) with respect to the value of the kth node ( ρ k ( t ) ).
L τ ( t ) ρ k ( t ) = τ ρ k ( t ) < ε i ( t 1 ) w i + ( 1 τ ) ρ k ( t ) ε i ( t 1 ) w i
ρ k ( t ) is chosen according to the following equation.
ρ k ( t ) = arg min ρ k ( t ) τ ρ k ( t ) < ε i ( t ) w i ( 1 τ ) ρ k ( t ) ε i ( t ) w i

2.1.3. Hyperparameters of Gradient Boosting Decision Trees

In the application of the GBDT algorithm, hyperparameter tuning is crucially important. In this paper, we consider several key hyperparameters, which can be classified into two groups: (1) regression tree-related hyperparameters; (2) gradient boosting-related hyperparameters.
The regression tree-related hyperparameters are (1) maximum depth D m a x and (2) minimum number of samples in leaf node N l e a f . Maximum depth controls the number of nodes used in the individual tree. The common value of maximum depth is between five and nine, which gives a suitable model complexity for the decision tree as a base learner of gradient boosting. The minimum number of samples in the leaf node is always between 20 and 80 [36].
The gradient boosting-related hyperparameters are (1) the number of boosting iterations T, (2) learning rate λ , and (3) bag fractions b a g . The bag fraction stands for the fraction of samples to be used to train the individual decision tree and is a real number satisfying b a g [ 0 , 1 ] . The number of boosting stages controls the number of trees used in the model. With sufficient boosting stages, it is a common pattern that the smaller the learning rate is, the better the generalization achieved [46]. However, the quantile GBDT model has to repeat the training process 99 times at each percentile ( τ = 0.01 , 0.02 , , 0.99 ), which is time consuming. Considering the computation time, the value of λ is chosen as 0.05. For this specific λ , the optimal value of T is between 400 and 800 [36].
The optimal values for these super parameters are determined by cross-validation, which will be discussed later.

2.2. The Architecture of IBT-GBDT

Consider M source problems { P ( m ) } , where m is the source problem index. When training the forecasting model for the target zone, the sample data of the target problem (base training data) constitute the training set. In an ideal scenario, wind power generations of the wind farm in different zones are strongly correlated. For the wind power quantile regression in the target zone, other than the base training data, the auxiliary training data can be added directly to the training set. With the enlarged training set, the model can be better trained, and the forecasting error will thus be reduced. In this case, the samples from the auxiliary training set and base training set satisfy the following equations.
y i ( m ) = f ( X i ( m ) ) + ε i ( m ) y i ( target ) = f ( X i ( target ) ) + ε i ( target ) ,
where f denotes the real mapping function between the input variable X i ( target ) and the wind power y i ( target ) for the target problem; X i ( m ) and y i ( m ) represent the ith input variable and output wind power from the mth source problem, respectively; and ε i ( target ) and ε i ( m ) denote the random error for samples from P ( target ) and P ( m ) , respectively.
However, in practice, the cross-correlation between the target problem and source forecasting problems is not ideal. In the worst case, the source problem set may be mixed with totally irrelevant problems. Therefore, the systematic error δ i ( m ) is introduced to denote the imperfect relatedness of P ( m ) to P ( target ) . With the systematic error considered, Formula (10) becomes:
y i ( m ) = f ( m ) ( X i ( m ) ) + δ i ( m ) + ε i ( m ) y i ( target ) = f ( X i ( target ) ) + ε i ( target ) .
where f ( m ) denotes the real mapping function between the input variable X i ( m ) and the wind power y i ( m ) for the mth source problem. f ( m ) is different from f.
In this case, the base training data and auxiliary training data should not be treated equally. In this paper, different weights are assigned to the target problem and source problems. The weight of the base training set is defined as w ( target ) , and w ( m ) denotes the weight of the mth auxiliary training set. w ( m ) should reflect the relatedness of different source problems P ( m ) to the target problem. A larger w ( m ) for P ( m ) implies a stronger relatedness to the target problem. The formula for weights is derived in Section 2.3. In this formula, the weights are calculated based on the performance of the GBDT model. Based on the formula for weights, an iterative algorithm is developed to solve the weight, where the weights are adjusted according to the weight formula in every iteration. Then, the updated weight is sent back to the GBDT model. The iteration stops when the weights converge.
When the target percentile τ changes, this paper assumes that the relatedness between the target problem and source problems does not change. Thus, the weights are the same for different quantiles and only need to be solved once. Therefore, the IBT-GBDT described in this paper is divided into two training steps. In Step 1, the weight is solved, and the quantile regression model with τ = 0.5 is trained. In Step 2, the weight from Step 1 is assigned to different datasets accordingly, and then, the quantile regression with τ = 0.01 0.49 and 0.51 0.99 is solved. In both two steps, the inner GBDT model is trained using the algorithm described in Section 2.1.2. The structure of the proposed IBT-GBDT model is depicted in Figure 1.

2.3. Derivation of the Weight Formula

The analysis in [27] shows that the error distribution of wind power generation forecasting is a heavy-tailed distribution. Thus, when solving the quantile regression model with τ = 0.5 in Step 1, we assume that the random variants δ ( m ) + ε ( m ) and ε i ( target ) are independent and follow the Laplace distribution. The expressions are:
ε ( target ) Laplace ( 0 , b ε ( target ) ) δ ( m ) + ε ( m ) Laplace ( 0 , b δ + ε ( m ) ) .
where b ε ( target ) and b δ + ε ( m ) represent the scale parameter of ε ( target ) and δ ( m ) + ε ( m ) , respectively.
For the hypothesis set of the prediction functions f θ ( θ is a parameter), the likelihood of the prediction function f θ being the correct prediction function is:
L i k e l i h o o d ( θ ) = log ( P ( Y | X ; θ ) ) = N ( target ) log 1 2 b ε ( target ) i = 1 N ( target ) 1 b ε ( target ) y i ( target ) f θ ( X i ( target ) ) + m = 1 M N ( m ) log 1 2 b δ + ε ( m ) m = 1 M i = 1 N ( m ) 1 b δ + ε ( m ) y i ( m ) f θ ( X i ( m ) ) .
Using the maximum likelihood method, the most likely value for the parameter θ ^ is:
θ ^ = arg max ( L i k e l i h o o d ( θ ) ) = arg min m = 1 M i = 1 N ( m ) 1 b δ + ε ( m ) y i ( m ) f θ ( X i ( m ) ) + i = 1 N ( target ) 1 b ε ( target ) y i ( target ) f θ ( X i ( target ) ) .
The target optimization function for solving the most likely prediction function f θ ^ is:
l o s s = m = 1 M 1 b δ + ε ( m ) i = 1 N ( m ) y i ( m ) f θ ^ ( X i ( m ) ) + 1 b ε ( target ) i = 1 N ( target ) y i ( target ) f θ ^ ( X i ( target ) ) .
According to (15), the weight formulas for w ( target ) and w ( m ) are:
w ( target ) = 1 / b ε ( target )
w ( m ) = 1 / b δ + ε ( m ) .
For weights, only their relative size matters. Thus, this paper normalizes the weights of the source problems w ( m ) to the range 0–1.
w ( m ) = min m ( b δ + ε ( m ) ) / b δ + ε ( m )
The corresponding w ( target ) becomes:
w ( target ) = min m ( b δ + ε ( m ) ) / b ε ( target ) .

2.4. Iterative Weight Assignment Algorithm

According to (18) and (19), the calculation of weights depends on the error distributions, and the error is calculated after the model is trained; however, the model is trained after the weights are set. To solve for the circular dependency, an iterative algorithm is presented in this paper, where w ( m ) is solved via iteration, while w ( target ) is a hyperparameter chosen in advance via cross-validation.
Because the relatedness between all the source problems P ( m ) and the target problem P ( target ) is initially unknown, all the source problems P ( m ) are assigned to the same weight (according to (18), the initial value for all w ( m ) is one). With w ( target ) and w ( m ) determined, the inner quantile regression model is trained using the gradient boosting decision trees described in Section 2.1.2. After the inner layer is trained, b δ + ε ( m ) (the scale parameter of the Laplace error distribution) is calculated. The weights w ( m ) , based on (18), are updated according to the corresponding b δ + ε ( m ) . With the updated w ( m ) , the next iteration begins. Finally, the outer iteration stops when every w ( m ) converges (Equation (18) holds simultaneously). The pseudocode of the weight assignment algorithm is described in Algorithm 1.
Algorithm 1 Weight assignment via iteration and validation (pseudocode).
1:
Choose w ( target )
2:
Initialize every w ( m ) to 1
3:
repeat
4:
  Complete the training of the inner quantile GBDT layer with τ = 0.5 according to the updated weight w ( m )
5:
  Calculate b δ + ε ( m )
6:
  Calculate w ( m ) according to (18)
7:
until Every w ( m ) converges
8:
Complete the training of the inner quantile GBDT layer with τ = 0.01 0.49 and 0.51 0.99 based on the converged weight w ( m )

3. Application and Results

3.1. Data Specification

To examine the IBT-GBDT model proposed in this paper, the model is tested on a public dataset from the probabilistic wind power forecasting track of GEFCom2014 (the dataset of GEFCom2014 is available in the Supplementary Material of [32]). The aim of this track was to forecast the normalized wind power generation in 10 zones, corresponding to 10 wind farms in Australia. The competition consisted of 15 tasks. The first three tasks were trial tasks, and the last 12 tasks were the evaluation tasks. To mimic real-world forecasting processes, the tasks were designed in a rolling forecasting manner. In the first trail task, months of data were provided as training data. With one month of incremental data released, the aim of each task was to forecast the wind power generation of the next month. The provided input included a 10-m wind speed vector and a 100-m wind speed vector. The desired output was the probabilistic distribution of wind power generation described by multiple quantiles. For each task, the quantile regression model was trained independently. The data period of the corresponding task is listed in Table 1.

3.2. Benchmark Models

In GEFCom2014, GBDT was shown to be a very effective algorithm for probabilistic forecasting of wind power, as GBDT was used in the top two effective models. In this paper, three GBDT-based models were developed and chosen as the benchmark models. The first one is the DL-GBDT [35], which was the winner of the wind power forecasting track of GEFCom2014. The authors of this paper reproduced the DL-GBDT model with exactly the same structure, input features, and hyperparameters. The last two benchmark models were the basic GBDT models that were developed to exhibit the effectiveness of the weight assignment algorithm. One was the GBDT model trained only by the base training data. The other was the GBDT model trained by both the base training data and auxiliary training data.

3.3. Feature Selection

The input data provided were hourly wind forecasts of the zonal and meridional wind components at 10 and 100 m, for ten separate (but correlated) wind zones. Based on the model input, the wind speed (WS), wind direction (WD), and wind energy (WE) are defined as follows, where u and v are the wind components provided.
WS = u 2 + v 2 WD = 180 / π × arctan ( u , v ) WE = 0.5 × w s 3
In DL-GBDT, features are derived from the provided model input by feature engineering, and different combinations of input features are tested by cross-validation. As a result, the input features of DL-GBDT model are very effective for the quantile model fit. Therefore, the proposed IBT-GBDT model chose the same input features as the DL-GBDT model, except the feature “hour” was added. The corresponding input features for the forecasting model of different zones are listed in Table 2.

3.4. Error Measure

The sharpness and reliability are two key measures for probabilistic forecasting. According to [32], the wind power quantile forecasting score (QS) is defined. QS is a comprehensive evaluation of sharpness and reliability, which is the average of pinball loss over all target percentiles. For the convenience of comparison, the target percentiles were chosen to be τ = 0.01 , 0.02 , 0.99 , the same as in GEFCom2014.
QS = 1 99 τ = 0.01 0.99 L τ ( y , f ( x ) )

3.5. Hyperparameter Tuning through Cross-Validation

For IBT-GBDT and the two basic GBDT models, the hyperparameters were determined through cross-validation. The hyperparameters included decision tree-related hyperparameters and gradient boosting-related hyperparameters. For IBT-GBDT, there was another hyperparameter w target . The data of the first three trial tasks were divided into five distinct parts, which constituted the five-fold cross-validation sets. A brief grid search was conducted to determine the optimal value of those hyperparameters. The result suggested that two sets of parameters should be used, one for percentiles in the middle range 0.16–0.84 and the for other percentiles. Table 3 presents the hyperparameter settings for the proposed IBT-GBDT model and the GBDT-based benchmark models (note: in the following tables, GBDT1 represents the GBDT model trained by the base training data and GBDT2 the GBDT model trained by both the base training data and the auxiliary training data.)
Table 4 presents the optimal value of w target in the proposed IBT-GBDT model of different zones.

3.6. Illustration of Training Process

To illustrate the training process of the IBT-GBDT model, wind power generation quantile forecasting for Zone 7 in the fourth evaluation task was chosen as the target problem. The auxiliary training sets were the data of wind power generation in zones other than Zone 7.
The cross-validation results showed that the best value of the hyperparameter w target was 50. Thus, 50 was chosen as the weight of the dataset from Zone 7 (target zone). The weights of the others zones were initialized to one. Then, the quantile regression model with τ = 0.5 was trained according to Step 1 of Algorithm 1. The converging process of the weights is presented in Table 5.
Table 5 shows that the weights converged after almost seven iterations. According to Table 5, the weight of Zone 8 always had the value of one. This means Zone 8 had the highest relatedness to Zone 7 (the same result was also given in [33]).
For the trained quantile regression model with τ = 0.5 , the pinball losses on the base training set and auxiliary training sets are calculated and shown in Table 6. Compared to the training loss after the first iteration, the training loss for zones except Zone 7 and Zone 8 increased after the second iteration. The reason is that the weights of these training sets dropped after the first iteration; thus, the inner GBDT algorithm was less fit to these training sets in the second iteration.
After the weight is determined, Step 2 of Algorithm 1 is executed. The wind power quantiles were forecast for the test set of Zone 7. Line plots of different prediction intervals, measurements, and forecast medium during the first 48 h are drawn in Figure 2.

3.7. The Relationship between Forecasting Error and w ( target )

As stated above, w ( target ) is a hyperparameter. To illustrate the effect of w ( target ) , for Zone 7, the relationship between QS and the hyperparameter w ( target ) is drawn in Figure 3.
As w ( target ) increases from 2–500, Figure 3 shows that the QS is first decreasing, but then increases. The reason is as follows. If w ( target ) is too small, the model is more like a common model of all the zones, where negative transfer occurs. When w ( target ) is too large, the model is more like a model trained only on data of the target zone, and no positive transfer is involved. When w ( target ) is given a suitable value, the positive transfer is strengthened, whereas the negative transfer is prevented; thus, the QS obtained the best result.

3.8. Analysis on Model Reliability

For any probabilistic forecasting method, reliability is seen as a primary requirement [47]. For the TL-GBDT model and benchmark models, the average proportion deviations between nominal and empirical proportions [47], which is the measure of reliability, are depicted in Figure 4.
According to Figure 4, the effectiveness of these models is analyzed. For all these GBDT-based methods, the quantiles were slightly overestimated for proportions lower than 0.5 and slightly underestimated for proportions above that value, which indicates that the corresponding predictive distributions were slightly too narrow. Among them, the worst-performing model was the GBDT model trained with both base training data and auxiliary training data. Negative transfer should be the main reason causing this. The second worst-performing model was the GBDT model trained with base training data. Comparing with the IBT-GBDT model, those two basic GBDT models did not perform well when the target percentile τ was between 80 and 95. For those quantiles, the distribution of training data was sparser, and the complicated model would easily overfit the training data; whereas for IBT-GBDT, with more training data, the over-fitting was suppressed.
To further evaluate the reliability of probabilistic forecasting, average coverage error (ACE) is introduced, which is defined as:
ACE τ = 1 n i = 1 n C i × 100 % ( 1 τ ) × 100 % ,
where C i is the indicative function of whether y i are covered by the prediction interval.
For any probabilistic forecasting model, the reliability may not be the same in different forecast horizons. Thus, the proposed model and benchmark models were tested on different forecast horizons. Measured by the ACE (the chosen prediction interval nominal confidences (PINCs) is 90%), a comparison of the reliability across different forecast horizons is shown in Table 7.
According to the results, the criteria ACE maintained within certain ranges in all forecasting horizons, which indicates that the proposed IBT-GBDT model achieved a rather high reliability in short probabilistic forecasting.

3.9. Comparison of Forecasting Error

In this paper, the forecasting errors were measured by the quantile score (QS), which was defined in (21). By definition, the QS of a model is the average of the pinball loss over all target percentiles. However, for different probabilistic wind power forecasting methods, their performance on different target percentiles is uneven. Thus, their pinball losses on several selected percentiles are listed and compared in Table 8.
When the target percentile τ is away from the central percentile, Table 8 shows that the improvement of the IBT-GBDT model over the DL-GBDT model is larger. The reason is discussed in the following. Around the edge of the distribution of the wind power generation, the training data were relatively scarce and had higher volatility. Furthermore, according to the definition of pinball loss, these higher volatility data had higher weights when the target percentile τ was away from the central percentile. Those training samples with high volatility caused the quantile regression model to fluctuate more, which led to lower predictive accuracy. In the proposed IBT-GBDT model, the volatility and variation of training data were reduced when more training data were available; whereas for the DL-GBDT model, the shortage of training data and the resulting high volatility still existed. This explains why the proposed IBT-GBDT model was more effective when the target percentile τ was away from the central percentile.
In addition, according to the bias variance theory, a complicated model will easily overfit a small dataset, but a complicated model is better than a simple model when more data are available. For the tree-based model, max depth D m a x reflects the complexity of the model. In Table 3, the validation result shows that the optimal value of D m a x in the IBT-GBDT model is larger than the optimal value of D m a x of the other benchmark models. This means, with more training data available in IBT-GBDT, the training dataset was large enough for the IBT-GBDT model with higher model complexity getting the upper hand, which led to the improvement in prediction accuracy.
To further test the applicability of the proposed IBT-GBDT model, this model was then applied to the zones other than Zone 7. For different zones, the QS of IBT-GBDT model was calculated and compared with the QS of the benchmark models in Table 9.
According to Table 9, the QS improvement over DL-GBDT achieved by IBT-GBDT ranged from 0.54%–2.40% for different zones, and the average improvement was 1.46%. Furthermore, for different tasks, the QS of the IBT-GBDT model was calculated and compared with the generalized additive tree model (GAT) [33], the linear regression model (LR) [34], and three benchmark models. The QS of these models (The QS of the DL-GBDT model, the GAT model, and the LR model came from the provisional leader board of the GEFCom2014. The provisional leader board can be download from the Supplementary Material of [32]) for different tasks are recorded in Table 10.

3.10. The Forecasting Error under a Small Base Training Set

One advantage of transfer learning is that it performs well when the training set is small. Thus, it is reasonable to infer that IBT-GBDT would be more effective than benchmark models when the training set becomes smaller. To illustrate this effect, the forecasting errors of the models trained with a small base training set (5%, 10%, 20%, and 50% of the samples of the original base training data) were calculated for both the IBT-GBDT model and the three benchmark models. Their QS are compared in Table 11.
According to Table 11, the maximum improvement over DL-GBDT is 4.74%. The forecasting error of IBT-GBDT was always smaller than other GBDT-based benchmark models, especially when the training set of the target problem was small. For different sample percentages of base training data, the weights of different zones are recorded in Table 12. It shows that those weights are almost the same across different sample percentages of the base training set. This result is in agreement with the above analysis, which is that the weights of the source problems just depend on their relatedness to the target problem.

3.11. The Relatedness between Different Zones

Based on the converged weight, the relatedness between different zones is analyzed. One characteristic of the proposed IBT-GBDT model is that the converged weight varies with different tasks. Thus, the converged weights were averaged over all tasks and formed a correlation matrix (the value of elements on diagonal axis were set to one). In Figure 5, a heat map is drawn to present the correlation. According to Figure 5, Zone 7 and Zone 8 are closely cross-correlated. This result agrees with the analysis of the correlation between zones conducted by [35].

3.12. The Comparison of Computational Time

The training process of those GBDT-based models is very time consuming. For the proposed IBT-GBDT model, which added extra structure onto the GBDT model, it is necessary to compare the computational time. In this paper, the GBDT algorithm was built on the R GBM package [48]. The training times (the tests were performed on an Intel(R) Core(TM) i7-4790 CPU, with 8 GB RAM, running a 64-bit Windows 10 operating system) of the IBT-GBDT model and the three benchmark models are compared in Table 13.
According to Table 13, the DL-GBDT model and GBDT model have similar training times. However, that of the IBT-GBDT model was three-times larger than that of the basic GBDT model, mainly because there were more training data involved in the training process.

4. Conclusions

To increase the performance of probabilistic wind power forecasting, instance-based transfer learning is described in this paper and combined with GBDT to form the IBT-GBDT model. To the best of our knowledge, this is the first time that instance-based transfer learning has been used for wind power quantile regression.
The IBT-GBDT model was tested on a public dataset from the wind track of GEFCom2014. Measured by quantile forecasting score, the IBT-GBDT model outperformed GBDT-based benchmark models across different zones. This result shows that the IBT-GBDT method can increase the forecasting accuracy for the target zone. Furthermore, the IBT-GBDT model became more effective when the training set of the target problem became smaller.

Author Contributions

Conceptualization, L.C. and Z.J.; methodology, L.C. and J.M.; software, L.C.; validation, J.G.; writing, review and editing, J.M. and J.G.

Funding

This research was funded by the National Key Research and Development Program of China, Grant Number 2016YFB0900100; Key Project of Shanghai Science and Technology Committee, Grant Number 18DZ1100303.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wind Power Capacity in 2017. Available online: https://wwindea.org/blog/2018/02/12/2017-statistics/ (accessed on 24 December 2018).
  2. Albadi, M.H.; El-Saadany, E.F. Overview of wind power intermittency impacts on power systems. Electr. Power Syst. Res. 2010, 80, 627–632. [Google Scholar] [CrossRef]
  3. Zhao, C.; Wang, J.H.; Watson, P.; Guan, Y. Multi-Stage Robust Unit Commitment Considering Wind and Demand Response Uncertainties. IEEE Trans. Power Syst. 2013, 28, 2708–2717. [Google Scholar] [CrossRef]
  4. Wang, Q.; Guan, Y.; Wang, J. A Chance-Constrained Two-Stage Stochastic Program for Unit Commitment with Uncertain Wind Power Output. IEEE Trans. Power Syst. 2012, 27, 206–215. [Google Scholar] [CrossRef]
  5. Wang, Y.; Xia, Q.; Kang, C. Unit Commitment with Volatile Node Injections by Using Interval Optimization. IEEE Trans. Power Syst. 2011, 26, 1705–1713. [Google Scholar] [CrossRef]
  6. Zhao, C.; Guan, Y. Data-Driven Stochastic Unit Commitment for Integrating Wind Generation. IEEE Trans. Power Syst. 2016, 31, 2587–2596. [Google Scholar] [CrossRef]
  7. Ackooij, W.V.; Finardi, E.C.; Ramalho, G.M. An Exact Solution Method for the Hydrothermal Unit Commitment Under Wind Power Uncertainty with Joint Probability Constraints. IEEE Trans. Power Syst. 2018, 33, 6487–6500. [Google Scholar] [CrossRef]
  8. Wen, Y.; Li, W.; Huang, G.; Liu, X. Frequency Dynamics Constrained Unit Commitment with Battery Energy Storage. IEEE Trans. Power Syst. 2016, 31, 5115–5125. [Google Scholar] [CrossRef]
  9. Morales, J.M.; Conejo, A.J.; Pérez-Ruiz, J. Short-Term Trading for a Wind Power Producer. IEEE Trans. Power Syst. 2010, 25, 554–564. [Google Scholar] [CrossRef]
  10. Xu, Q.; Zhang, N.; Kang, C.; Xia, Q.; He, D.; Liu, C.; Huang, Y.; Cheng, L.; Bai, J. A Game Theoretical Pricing Mechanism for Multi-Area Spinning Reserve Trading Considering Wind Power Uncertainty. IEEE Trans. Power Syst. 2016, 31, 1084–1095. [Google Scholar] [CrossRef]
  11. Zugno, M.; Jónsson, T.; Pinson, P. Trading wind energy on the basis of probabilistic forecasts both of wind generation and of market quantities. Wind Energy 2013, 16, 909–926. [Google Scholar] [CrossRef]
  12. Pinson, P.; Chevallier, C.; Kariniotakis, G.N. Trading Wind Generation From Short-Term Probabilistic Forecasts of Wind Power. IEEE Trans. Power Syst. 2007, 22, 1148–1156. [Google Scholar] [CrossRef] [Green Version]
  13. Matos, M.A.; Bessa, R.J. Setting the Operating Reserve Using Probabilistic Wind Power Forecasts. IEEE Trans. Power Syst. 2011, 26, 594–603. [Google Scholar] [CrossRef]
  14. Paterakis, N.G.; Erdinc, O.; Bakirtzis, A.G.; Catalão, J.P. Load-Following Reserves Procurement Considering Flexible Demand-Side Resources Under High Wind Power Penetration. IEEE Trans. Power Syst. 2015, 30, 1337–1350. [Google Scholar] [CrossRef]
  15. Sahin, C.; Shahidehpour, M.; Erkmen, I. Allocation of Hourly Reserve Versus Demand Response for Security-Constrained Scheduling of Stochastic Wind Energy. IEEE Trans. Sustain. Energy 2013, 4, 219–228. [Google Scholar] [CrossRef]
  16. Fang, X.; Hu, Q.; Li, F.; Wang, B.; Li, Y. Coupon-Based Demand Response Considering Wind Power Uncertainty: A Strategic Bidding Model for Load Serving Entities. IEEE Trans. Power Syst. 2016, 31, 1025–1037. [Google Scholar] [CrossRef]
  17. Jabr, R.A. Adjustable Robust OPF with Renewable Energy Sources. IEEE Trans. Power Syst. 2013, 28, 4742–4751. [Google Scholar] [CrossRef]
  18. Li, Y.; Li, W.; Yan, W.; Yu, J.; Zhao, X. Probabilistic Optimal Power Flow Considering Correlations of Wind Speeds Following Different Distributions. IEEE Trans. Power Syst. 2014, 29, 1847–1854. [Google Scholar] [CrossRef]
  19. Han, L.; Zhang, R.; Wang, X.; Dong, Y. Multi-Time Scale Rolling Economic Dispatch for Wind/Storage Power System Based on Forecast Error Feature Extraction. Energies 2018, 11, 2124. [Google Scholar] [CrossRef]
  20. Lorca, A.; Sun, X.A. Adaptive Robust Optimization with Dynamic Uncertainty Sets for Multi-Period Economic Dispatch Under Significant Wind. IEEE Trans. Power Syst. 2015, 30, 1702–1713. [Google Scholar] [CrossRef]
  21. Alham, M.H.; Elshahed, M.; Ibrahim, D.K.; El Zahab, E.E.D.A. A dynamic economic emission dispatch considering wind power uncertainty incorporating energy storage system and demand side management. Renew. Energy 2016, 96, 800–811. [Google Scholar] [CrossRef]
  22. Bracale, A.; Falco, P.D. An Advanced Bayesian Method for Short-Term Probabilistic Forecasting of the Generation of Wind Power. Energies 2015, 8, 10293–10314. [Google Scholar] [CrossRef] [Green Version]
  23. Zhang, Z.S.; Sun, Y.Z.; Gao, D.W.; Lin, J.; Cheng, L. A Versatile Probability Distribution Model for Wind Power Forecast Errors and Its Application in Economic Dispatch. IEEE Trans. Power Syst. 2013, 28, 3114–3125. [Google Scholar] [CrossRef]
  24. Wang, Z.; Shen, C.; Liu, F. A Conditional Model of Wind Power Forecast Errors and Its Application in Scenario Generation. Appl. Energy 2017, 212, 771–785. [Google Scholar] [CrossRef]
  25. Tewari, S.; Geyer, C.J.; Mohan, N. A Statistical Model for Wind Power Forecast Error and its Application to the Estimation of Penalties in Liberalized Markets. IEEE Trans. Power Syst. 2011, 26, 2031–2039. [Google Scholar] [CrossRef]
  26. Pinson, P. Very-short-term probabilistic forecasting of wind power with generalized logit—Normal distributions. J. R. Stat. Soc. 2012, 61, 555–576. [Google Scholar] [CrossRef]
  27. Bruninx, K.; Delarue, E.; Dhaeseleer, W. Statistical Description of the Error on Wind Power Forecasts via a Lévy Alpha-Stable Distribution. EUI RSCAS Working Paper 2013/50. 2013, pp. 1–8. Available online: http://cadmus.eui.eu/handle/1814/27520 (accessed on 24 December 2018).
  28. Bludszuweit, H.; Dominguez-Navarro, J.A.; Llombart, A. Statistical Analysis of Wind Power Forecast Error. IEEE Trans. Power Syst. 2008, 23, 983–991. [Google Scholar] [CrossRef]
  29. Taylor, J.; Jeon, J. Forecasting wind power quantiles using conditional kernel estimation. Renew. Energy 2015, 80, 370–379. [Google Scholar] [CrossRef] [Green Version]
  30. Qin, Z.; Li, W.; Xiong, X. Estimating wind speed probability distribution using kernel density method. Electr. Power Syst. Res. 2011, 81, 2139–2146. [Google Scholar] [CrossRef]
  31. Jeon, J.; Taylor, J.W. Using Conditional Kernel Density Estimation for Wind Power Density Forecasting. J. Am. Stat. Assoc. 2012, 107, 66–79. [Google Scholar] [CrossRef] [Green Version]
  32. Hong, T.; Pinson, P.; Fan, S.; Zareipour, H.; Troccoli, A.; Hyndman, R.J. Probabilistic energy forecasting: Global Energy Forecasting Competition 2014 and beyond. Int. J. Forecast. 2016, 32, 896–913. [Google Scholar] [CrossRef] [Green Version]
  33. Nagy, G.I.; Barta, G.; Kazi, S.; Borbély, G.; Simon, G. GEFCom2014: Probabilistic solar and wind power forecasting using a generalized additive tree ensemble approach. Int. J. Forecast. 2016, 32, 1087–1093. [Google Scholar] [CrossRef]
  34. Juban, R.; Ohlsson, H.; Maasoumy, M.; Poirier, L.; Kolter, J.Z. A multiple quantile regression approach to the wind, solar, and price tracks of GEFCom2014. Int. J. Forecast. 2016, 32, 1094–1102. [Google Scholar] [CrossRef]
  35. Landry, M.; Erlinger, T.P.; Patschke, D.; Varrichio, C. Probabilistic gradient boosting machines for GEFCom2014 wind forecasting. Int. J. Forecast. 2016, 32, 1061–1066. [Google Scholar] [CrossRef]
  36. Andrade, J.R.; Bessa, R.J. Improving Renewable Energy Forecasting with a Grid of Numerical Weather Predictions. IEEE Trans. Sustain. Energy 2017, 8, 1571–1580. [Google Scholar] [CrossRef]
  37. West, J.; Ventura, D.; Warnick, S. Inductive Transfer. In Spring Research Presentation: A Theoretical Foundation for Inductive Transfer; College of Physical and Mathematical Sciences, Brigham Young University: Provo, UT, USA, 2007. [Google Scholar]
  38. Pan, S.J.; Yang, Q. A Survey on Transfer Learning. IEEE Trans. Knowl. Data Eng. 2010, 22, 1345–1359. [Google Scholar] [CrossRef] [Green Version]
  39. Zhang, L.; Zuo, W.; Zhang, D. LSDT: Latent Sparse Domain Transfer learning for visual adaptation. IEEE Trans. Image Process. 2016, 25, 1177–1191. [Google Scholar] [CrossRef] [PubMed]
  40. Do, C.B.; Ng, A.Y. Transfer Learning for Text Classification. Adv. Neural Inf. Process. Syst. 2006, 299–306. [Google Scholar]
  41. Kamishima, T.; Hamasaki, M.; Akaho, S. TrBagg: A Simple Transfer Learning Method and its Application to Personalization in Collaborative Tagging. In Proceedings of the 2009 Ninth IEEE International Conference on Data Mining, Miami, FL, USA, 6 December 2009. [Google Scholar]
  42. Dai, W.; Yang, Q.; Xue, G.R.; Yu, Y. Boosting for transfer learning. In Proceedings of the 24th International Conference on Machine Learning, Corvalis, OR, USA, 20 June 2007. [Google Scholar]
  43. Zhao, Y.; Ye, L.; Pinson, P.; Tang, Y.; Lu, P. Correlation-Constrained and Sparsity-Controlled Vector Autoregressive Model for Spatio-Temporal Wind Power Forecasting. IEEE Trans. Power Syst. 2018, 33, 5029–5040. [Google Scholar] [CrossRef]
  44. Zhang, Y.; Luo, G. Short term power load prediction with knowledge transfer. Inf. Syst. 2015, 53, 161–169. [Google Scholar] [CrossRef]
  45. Rosenstein, M.T.; Marx, Z.; Kaelbling, L.; Dietterich, T.G. To transfer or not to transfer. In NIPS 2005 Workshop on Inductive Transfer: 10 Years Later; 2005; Available online: http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.94.7909 (accessed on 24 December 2018).
  46. Natekin, A.; Knoll, A. Gradient boosting machines, a tutorial. Front. Neurorobot. 2013, 7, 21. [Google Scholar] [CrossRef]
  47. Pinson, P.; Nielsen, H.A.; MØLler, J.K.; Madsen, H.; Kariniotakis, G.N. Non-parametric probabilistic forecasts of wind power: Required properties and evaluation. Wind Energy 2010, 10, 497–516. [Google Scholar] [CrossRef]
  48. Ridgeway, G. Generalized Boosted Regression Models. Available online: https://cran.r-project.org/web/packages/gbm/index.html (accessed on 24 December 2018).
Figure 1. The structure of the proposed IBT-GBDT model.
Figure 1. The structure of the proposed IBT-GBDT model.
Energies 12 00159 g001
Figure 2. Example forecast from the IBT-GBDT model.
Figure 2. Example forecast from the IBT-GBDT model.
Energies 12 00159 g002
Figure 3. The relationship between w ( target ) and quantile forecasting score (QS).
Figure 3. The relationship between w ( target ) and quantile forecasting score (QS).
Energies 12 00159 g003
Figure 4. Comparison of reliability.
Figure 4. Comparison of reliability.
Energies 12 00159 g004
Figure 5. The correlation between different zones.
Figure 5. The correlation between different zones.
Energies 12 00159 g005
Table 1. Data period of the scoring task.
Table 1. Data period of the scoring task.
TaskTraining SetTesting Set
Task 1January 2012–December 2012January 2013
Task 2January 2012–January 2013February 2013
Task 3January 2012–February 2013March 2013
Task 4January 2012–March 2013April 2013
Task 5January 2012–April 2013May 2013
Task 6January 2012–May 2013June 2013
Task 7January 2012–June 2013July 2013
Task 8January 2012–July 2013August 2013
Task 9January 2012–August 2013September 2013
Task 10January 2012–September 2013October 2013
Task 11January 2012–October 2013November 2013
Task 12January 2012–November 2013December 2013
Table 2. Input features for forecasting model of each zone.
Table 2. Input features for forecasting model of each zone.
FeatureZones
Current 10-m energyAll zones
Current energyAll zones
Current 10-m directionAll zones
Current directionAll zones
Hour of dayAll zones
−1 period energyAll zones, but Zone 1
−2 period energyAll zones, but Zone 1
−3 period energyZones 2, 3, 5, 6, 7
+1 period energyAll zones
+3 period energyZones 2, 3, 5, 6, 7
+4 period energyZone 1
−1 period directionZones 1, 6
+1 period directionZones 1, 6
Average energy, Periods −3–0All zones, but Zones 1, 2, 7
Average energy, Periods 0–+3All zones
Standard deviation of direction, Periods −3–0Zones 1, 2, 3, 10
Direction difference: 100 m and 10 mAll zones
Energy ratio: 100 m and 10 mAll zones
Percentage speed change: current and −1 periodAll zones
Note: all values are for the 100-m wind speed, unless otherwise specified.
Table 3. Hyperparameter settings.
Table 3. Hyperparameter settings.
τ = 0.16 0.84 τ = 0.01 0.15 and 0.85 0.99
ModelGBDT1GBDT2DL-GBDTIBT-GBDTGBDT1GBDT2DL-GBDTIBT-GBDT
D m a x 77785556
N l e a f 3030303030303030
T500500500500400400400500
λ 0.050.050.050.050.050.050.050.05
b a g 0.750.750.750.750.750.750.750.75
Table 4. w target settings for the IBT-GBDT model.
Table 4. w target settings for the IBT-GBDT model.
Zone Index12345678910
w target 50201001002050502020100
Table 5. Weights of the training sets from different zones after each iteration.
Table 5. Weights of the training sets from different zones after each iteration.
ZoneIteration Step
Initial12345678
Zone 110.5460.5380.5390.5350.5350.5360.5350.535
Zone 210.8350.8190.8160.8110.8170.8160.8170.817
Zone 310.3800.3640.3630.3610.3640.3630.3630.363
Zone 410.4640.4470.4490.4450.4460.4480.4480.448
Zone 510.3310.3140.3150.3120.3150.3160.3150.315
Zone 610.2940.2810.2810.2780.2820.2820.2820.282
Zone 7505050505050505050
Zone 8111111111
Zone 910.6810.6700.6710.6670.6710.6700.6710.671
Zone 1010.2080.2010.2010.1990.2010.2010.2010.201
Table 6. Pinball loss after each iteration step.
Table 6. Pinball loss after each iteration step.
ZoneIteration Step
1234567
Zone 10.06820.06860.06860.06860.06870.06860.0686
Zone 20.05420.05460.05470.05470.05470.05490.0548
Zone 30.07960.08120.08130.08130.08130.08120.0814
Zone 40.07340.07460.07450.07470.07480.07450.0747
Zone 50.08500.08710.08700.08720.08700.08670.0871
Zone 60.08830.09020.09030.09040.09030.08990.0903
Zone 70.03940.03910.03900.03890.03900.03890.0390
Zone 80.05090.05080.05080.05070.05090.05080.0509
Zone 90.05130.05160.05160.05170.05170.05160.0517
Zone 100.10150.10300.10330.10330.10330.10310.1033
Table 7. Comparison of the criterion average coverage error (ACE) (%).
Table 7. Comparison of the criterion average coverage error (ACE) (%).
ModelGBDT1GBDT2DL-GBDTIBT-GBDT
First Week−2.59−3.10−1.87−2.21
Second Week−2.64−4.60−2.77−2.84
Third Week1.031.510.920.63
Fourth Week0.470.820.380.27
Table 8. The pinball loss of the IBT-GBDT model and benchmark models.
Table 8. The pinball loss of the IBT-GBDT model and benchmark models.
τ GBDT1GBDT2DL-GBDTIBT-GBDTImprovement on GBDT1Improvement on GBDT2Improvement on DL-GBDT
0.10.01500.02160.01390.013212.25%38.84%4.94%
0.20.02500.03290.02370.02307.83%29.92%3.05%
0.30.03230.03670.03150.03104.10%15.59%1.56%
0.40.03670.04030.03620.03592.18%10.87%0.83%
0.50.03960.04240.03920.03901.35%7.98%0.42%
0.60.03790.04220.03750.03721.68%11.73%0.71%
0.70.03600.04070.03520.03473.67%14.65%1.44%
0.80.02910.03770.02780.02707.33%28.40%2.89%
0.90.01860.02580.01740.016610.63%35.63%4.18%
Table 9. The QS improvement of the IBT-GBDT model over the benchmark models.
Table 9. The QS improvement of the IBT-GBDT model over the benchmark models.
ZoneGBDT1GBDT2DL-GBDTIBT-GBDTImprovement on GBDT1Improvement on GBDT2Improvement on DL-GBDT
Zone 10.02570.03890.02510.02483.62%36.22%1.39%
Zone 20.02360.02940.02290.02264.27%23.07%1.05%
Zone 30.03020.03860.02940.02933.02%24.25%0.54%
Zone 40.03420.04390.03320.03274.24%25.39%1.50%
Zone 50.03380.04090.03300.03273.40%20.15%0.94%
Zone 60.03560.04430.03490.03462.89%21.90%0.86%
Zone 70.02710.03240.02620.02565.57%21.05%2.22%
Zone 80.02990.03290.02880.02815.90%14.65%2.40%
Zone 90.03110.03420.03030.02974.55%13.09%1.94%
Zone 100.04350.05380.04280.04213.22%21.76%1.75%
Table 10. Comparison of QS.
Table 10. Comparison of QS.
TaskGBDT1GBDT2DL-GBDTIBT-GBDTGATLR
Task 10.03650.04430.03490.03560.03730.0394
Task 20.04100.04980.03890.03990.04120.0431
Task 30.03990.04850.03810.03890.04090.0392
Task 40.03150.03850.03010.03070.03150.0320
Task 50.03980.04860.03850.03890.03960.0390
Task 60.03250.03970.03140.03190.03330.0330
Task 70.03740.04540.03580.03640.03800.0380
Task 80.03790.04640.03660.03710.03900.0375
Task 90.03990.04840.03820.03880.03940.0407
Task 100.04280.05190.04110.04170.04180.0424
Task 110.04000.04890.03860.03920.03940.0401
Task 120.04030.04880.03870.03920.03860.0398
All0.03830.04660.03680.03740.03830.0387
Table 11. QS of different models across different amounts of training data.
Table 11. QS of different models across different amounts of training data.
ModelGBDT1GBDT2DL-GBDTIBT-GBDTImprovement on GBDT1Improvement on GBDT2Improvement on DL-GBDT
5%0.04290.04740.04190.03997.05%15.97%4.74%
10%0.04130.04710.04040.03905.61%17.16%3.28%
20%0.03980.04680.03890.03814.44%18.63%2.05%
50%0.03920.04660.03830.03763.97%19.32%1.74%
100%0.03830.04660.03740.03683.80%21.01%1.41%
Table 12. Weights after convergence (different sample percentages of base training data).
Table 12. Weights after convergence (different sample percentages of base training data).
Zone5%10%20%50%100%
Zone 10.5500.6310.5120.5040.535
Zone 20.7890.7840.7720.8420.817
Zone 30.3600.3190.3260.3510.363
Zone 40.4740.4350.3860.4190.448
Zone 50.3260.3180.2900.2690.315
Zone 60.3020.2620.2270.2490.282
Zone 75050505050
Zone 811111
Zone 90.7480.6610.6250.5960.671
Zone 100.2190.2000.1710.1830.201
Table 13. The comparison of the training times.
Table 13. The comparison of the training times.
GBDT1GBDT2DL-GBDTIBT-GBDT
Training time567 s1864 s620 s2014 s

Share and Cite

MDPI and ACS Style

Cai, L.; Gu, J.; Ma, J.; Jin, Z. Probabilistic Wind Power Forecasting Approach via Instance-Based Transfer Learning Embedded Gradient Boosting Decision Trees. Energies 2019, 12, 159. https://0-doi-org.brum.beds.ac.uk/10.3390/en12010159

AMA Style

Cai L, Gu J, Ma J, Jin Z. Probabilistic Wind Power Forecasting Approach via Instance-Based Transfer Learning Embedded Gradient Boosting Decision Trees. Energies. 2019; 12(1):159. https://0-doi-org.brum.beds.ac.uk/10.3390/en12010159

Chicago/Turabian Style

Cai, Long, Jie Gu, Jinghuan Ma, and Zhijian Jin. 2019. "Probabilistic Wind Power Forecasting Approach via Instance-Based Transfer Learning Embedded Gradient Boosting Decision Trees" Energies 12, no. 1: 159. https://0-doi-org.brum.beds.ac.uk/10.3390/en12010159

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop