Next Article in Journal
Artificial Floating Island with Vetiver for Treatment of Arsenic-Contaminated Water: A Real Scale Study in High-Andean Reservoir
Previous Article in Journal
The Effect of Small Density Differences at River Confluences
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Short-Term Soil Moisture Forecasting via Gaussian Process Regression with Sample Selection

1
Shunde Graduate School of University of Science and Technology Beijing, Shunde 528300, China
2
State Key Laboratory of Power Transmission Equipment & System Security and New Technology, Chongqing University, Chongqing 400044, China
3
School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, China
4
Key Laboratory of Fluid and Power Machinery (Xihua University), Ministry of Education, Chengdu 610039, China
5
Beijing Key Laboratory of Knowledge Engineering for Materials Science, Beijing 100083, China
6
School of Economics and Management, University of Science and Technology Beijing, Beijing 100083, China
*
Author to whom correspondence should be addressed.
Submission received: 8 August 2020 / Revised: 19 October 2020 / Accepted: 31 October 2020 / Published: 3 November 2020
(This article belongs to the Section Hydraulics and Hydrodynamics)

Abstract

:
Soil moisture is a critical limiting factor for crop growth. Accurate soil moisture prediction helps to schedule irrigation and improve the crop production. A soil moisture prediction method based on Gaussian Process Regression (GPR) is proposed in this paper. In order to reduce the computation time of the GPR model, the Radially Uniform (RU) design algorithm was incorporated into the sample selection during the training procedure. Thus, representative training samples are identified and less training time is required. To validate the proposed prediction model, the soil moisture data collected in Beijing, China, was fully utilized. The experimental results demonstrate that the forecasting performance of the GPR model with the RU design algorithm is generally better than that of the generic GPR model in terms of less forecasting errors for both deterministic and probabilistic forecasting, while less computing time is needed for the model training.

1. Introduction

The structure and function of the natural hydrological system rests with the soil moisture in the hydrological cycle [1,2,3,4]. The distribution of mass and energy flux over land and atmosphere is decided by soil moisture. Thus, soil moisture is a vital indicator for evaluating water and energy balance [5].
Drought is one of the main natural disasters for agriculture all over the world. Different from other natural disasters, it has the characteristics of frequent occurrence, wide coverage and long duration. As a consequence, agriculture production is significantly influenced by drought. Soil moisture is an important variable for drought assessment and forecasting [6,7,8], as well as flood and landslide simulation and prediction [9,10,11,12]. By analyzing and predicting changes in soil moisture, it is possible to predict the potential drought and schedule irrigation [13]. Besides predicting droughts, soil moisture forecasting is also of great importance to flood management. Soil moisture in the functional landscape describes initial conditions of the watershed well and may provide valuable information for flood forecasting and early warning systems [14].
In addition to preventing drought and flood, soil moisture and agricultural constants are also closely linked, since soil moisture indicates the water content in soil for plant growth. Therefore, soil moisture is a critical limiting factor for crop growth. The accurate collection of soil moisture information is the base and guarantee for water-saving irrigation, optimal regulation of farmland, as well as the effective implementation of water-saving technologies [15]. Soil moisture also plays a significant role in the agricultural system. Continuous monitoring and prediction of soil moisture is an ideal strategy to develop the sustainability and productivity of the agricultural system and make positive plans and decision-making measures for agriculture.
Compared with traditional soil moisture forecasting models, such as autoregressive moving average model (ARMA) [13], empirical formula of the model [16], the water balance model [17], the dynamic model of soil water [18], remote sensing model [19,20], and the neural network model [21,22] hybrid model which incorporates the superior features of several algorithms could overcome the deficiency of any single method and reduce the prediction error. Thus, hybrid prediction methods were widely discussed in the literature. Ramendra [23] presented a hybrid machine learning technique in which extreme learning machine (ELM) models were explored to predict the monthly soil moisture. Huang et al. [24] proposed a hybrid genetic algorithm-backpropagation algorithm, which has a good practical value in soil moisture prediction. The weights and threshold values of the backpropagation network were optimized with the Genetic Algorithm (GA), which is able to obtain the global optimal solution.
The general regression algorithms, reported in the above-mentioned literature, fit the relationship between the independent variable x and the dependent variable y in the training set and predicts the new corresponding dependent variable y* according to the new independent variable x*. Differently, Gaussian Process Regression (GPR) algorithm can provide the distribution of predicted values, including the predictive mean value and predictive variance. As a result, the confidence interval of the predictive value can be derived and the relationship between the predicted and actual values is more clearly described. Recently, the GPR algorithm has received great attention from both academia and industry. It has been widely used in forecasting and inferring industrial variables. Liu [25] proposed auto-switch probabilistic soft sensors for industrial multi-stage processes with transitions according to the GPR model. In addition, the GPR model was applied in different industrial domains, such as fouling formation [26] and soft sensors [27]. These applications have proved that the GPR model can achieve good performance in real systems. Due to the abovementioned reasons, the GPR model is applied to predict the soil moisture in this paper.
Besides point forecasting of soil moisture, the prediction intervals (PIs) of soil moisture are built to quantify the uncertainty in connection with point forecasts in this paper. The forecasting variance of GPR is employed to develop the PIs for the probabilistic interval forecasting of soil moisture. Since matrix inversion is included in the GPR model, the large training dataset may induce the significance computation time in model training and inferring. To address this issue, the Radially Uniform (RU) design algorithm is employed to select training samples and thus reduce the computational time. Meanwhile, it improves the accuracy of soil moisture point prediction, which improves the efficiency and accuracy of the forecasting model simultaneously. Soil moisture data were collected from Beijing, China. The date range is from 28 February 2012 to 8 November 2016. In order to analyze the effects of the season factor on soil moisture prediction, comparative experiments were conducted with test datasets on different seasons. The main contributions of this paper are summarized as follows:
(1) A novel GPR-based soil moisture forecasting model is proposed. In the proposed model, both the deterministic and probabilistic forecasts are generated for different levels of decisions.
(2) The GPR computation is accelerated with the RU design algorithm and the most representative training samples are selected accordingly.
(3) The effectiveness and the efficiency of the proposed model are thoroughly evaluated based on real collected soil moisture data.
The remaining part of this paper is arranged as follows. In Section 2, the proposed forecasting model is introduced. Section 3 shows the experimental results of the proposed model, including the error analysis of point prediction and PI assessment. Finally, the conclusions are depicted in Section 4.

2. Prediction Model

To improve the forecasting performance of the GPR model, the training dataset is processed and only the most representative subset is selected. The RU algorithm is employed to sample the most representative subset from the full dataset.

2.1. Gaussian Process Regression

Gaussian process (GP) is a data-driven model according to the Bayesian theory and statistics. It is applied to addressing complex regression issues such as high dimensional, small sample size and highly non-linear while it has mighty generalization performance. Different from neural networks and support vector machines, GP has numerous advantages, including easy implementation, hyper-parametric adaptive acquisition, non-parametric inference flexibility, and output probabilistic significance.
GP is a random variable set, and the linear combination of discretionary random variables in the GP follows the normal distribution. Moreover, each finite-dimensional distribution is a joint normal distribution [28]. The GP is totally derived from its mean and covariance functions and has many similar properties to the Gaussian distribution [29] Given N groups of dataset, D = {( X i , y i ), i = 1,2,…,N}, where X i is the input and y i is the noisy output. For the simplest case, the noise is assumed to be independent, normal and additive, so that the relation between the latent function f(X) and the observed noisy target y is
y = f ( X ) + ε , where   ε N ( 0 , σ n o i s e 2 )
where ε is the noise, σ n o i s e 2 is the variance of the noise.
Given a random process with mean function μ(X) and covariance function k ( X , X ) in real value, they satisfy Equations (2) and (3).
μ ( X ) = E [ f ( X ) ]
k ( X , X ) = E [ ( f ( X ) μ ( X ) ) ( f ( X ) μ ( X ) ) ]
The random process f(X) is a Gaussian process, which can be described as:
f ( X ) G P ( μ ( X ) , k ( X , X ) )
In general, μ(x) is set to 0 and formula (4) can be rewritten as:
f ( X ) G P ( 0 , k ( X , X ) )
where X is the learning sample whose measure in the GP is the finite-dimensional distribution of the GP. As defined by the GP, the finite-dimensional distribution is a joint normal distribution as:
[ f ( X 1 ) , f ( X 2 ) , , f ( X n ) ] T N ( μ , K )
The predictive distribution of the function values f* is computed at the test location X*. The probability distribution of f * is described in Equation (7).
f * N 0 , k ( X * , X *
If taking the effects of noise into account, the probability distribution of the observed noisy target y is shown in Equation (8):
y N ( 0 , k ( X , X ) + σ n o i s e 2 I )
The joint probability distribution is depicted in Equation (9),
y f * N 0 , k ( X , X ) + σ n o i s e 2 I k ( X , X * ) k ( X * , X ) k ( X * , X * )
Using the conditional distribution properties of the Gaussian distribution, equations are obtained in Equation (10).
p ( f * | X , y , X * ) = N ( μ ^ , Σ ^ ) p ( y * | X , y , X * ) = N ( μ ^ , Σ ^ + σ n o i s e 2 I )
where μ ^ = k ( X * , X ) ( k ( X , X ) + σ n o i s e 2 I ) 1 y and Σ ^ = k ( X * , X * ) k ( X * , X ) ( k ( X , X ) + σ n o i s e 2 I ) 1 k ( X , X * ) .
The above formulas give the prediction form of GPR, which is also the finite-dimensional distribution of the test sample for the test space posterior. The prediction of the response corresponding to X* assumes to be:
y * = μ ^
Covariance function K is a covariance matrix, whose entries are given by the covariance function K i , j = k ( X i , X j ) . The covariance matrix is required to be a semi-positive definite matrix, and the kernel functions are semi-positive definite matrices so that the covariance matrix can be obtained by the kernel function. The Radial Basis Function (RBF) kernel [30] is employed in this paper.
k ( x i , x j ) = exp 1 2 d ( x i / l , x j / l ) 2
Since the large dataset may increase the kernel size and further significantly increase the computation time, controlling the size of the dataset can help to execute the GPR algorithm in a reasonable time. In this study, RU design was employed to select the most representative subset from large datasets so that the new training dataset is obtained and the kernel size is better controlled. In the meantime, as the newly selected subset is the characteristic concentration of the previous complete dataset, the prediction accuracy is hardly affected. Therefore, RU design can improve the prediction performance of the GPR model for large dataset problems.

2.2. Radially Uniform Design

2.2.1. Description of RU Design

Suppose that we have T = { x 1 , Y 1 ) , , ( x N , Y N ) } , where N is the number of a big dataset, and each x i is a normalized vector of d-dimensional inputs. If N is very big, the training time of the GPR model built with the full training dataset will be too long. Thus, a representative subset (the chosen data) D = ( x 1 , Y 1 ) , , ( x n , Y n ) is generated, and it can greatly improve the model training speed without inducing the loss of the accuracy. If D is randomly chosen from T, it is possible that the chosen data are not representative, so the prediction accuracy is influenced. It shows that the chosen data D ought to be representative of T and there must be a design point nearby each point in the full training data T.
RU algorithm is proposed in [31] and pursues to maintain the density of x 1 , , x N in the radial direction while minimizing the difference between the distributions of the response values in full training data and those of the chosen data.
The chosen data are featured by minimizing a standard φ if the following conditions are met: the radial distribution of the response values in the chosen data F ( r ) = i = 1 n I ( x i m 2 r ) / n is adequately close to the radial distribution of the response values in full training data G ( r ) = i = 1 N I ( x i m 2 r ) / N , where m is the median of the training data. This standard φ evaluates the difference between the distributions of response points in the chosen data and the full training data.

2.2.2. Procedure of RU Design

(1) Sort x 1 , , x N in ascending order according to the distances from the median x i m 2 , i = 1,…,N. x i is the point with i-th smallest distance to the median, i.e., x 1 is the point which is closest to the median.
(2) Set ( k 1 , , k n ) = ( n N , 2 n N , N n N ) . The notation i n N demonstrates the i n N rounded up. Then, take the N sorted training data into n teams, where x i is allocated to the team k i .
(3) Select the design points at random from each team to constitute the initial chosen data D 0 = { x 1 , , x n } , where x j is from team j. Next, set up t = 1.
(4) Put up D t = D t 1 . For each j = 1, …, n, run step 5).
(5) If x b e s t from group j satisfies D j = { x 1 , , x j 1 , x b e s t , x j + 1 , x n } which has the best standard value φ ( D j ) , take place of D t with D j .
(6) Stop the execution if D t = D t 1 . Else, put up t = t + 1 and back to step (4).
In RU design, x 1 , , x N are always standardized that can assure x i [ 0 , 1 ] d . If the chosen data and full training data meet the condition (13), it demonstrates that the radial distribution of the chosen data F(r) is an approach to the radial distribution of the full training data G(r).
sup r 0 | F ( r ) G ( r ) | 1 2

2.2.3. Choice of the Optimization Standard

Given G different responses: Y(1), …, Y(G), define the 1 n + 1 , 2 n + 1 , , n n + 1 quantiles of Y(g) in the full training data as Y ( 1 ) ( g ) , , Y ( n ) ( g ) . Y i ( g ) indicates the value of Y(g) corresponding to the i-th design point x i , and Y ( 1 ) ( g ) , , Y ( n ) ( g ) indicates the sorted values of Y 1 ( g ) , , Y n ( g ) . The proposed standard value is computed in Equation (14).
φ ( D ) = g = 1 G i = 1 n Y ( i ) ( g ) Y ( i ) ( g ) 2 n i = 1 n Y ( i ) ( g ) 2 n
φ ( D ) evaluates the representativeness of the response points chosen by the RU design algorithm. In this paper, x 1 , , x N denote N soil moisture time series data for 8 consecutive days. Y(1), Y(2), and Y(3) indicate the daily average soil moisture on day 9, day 10, and day 11, respectively.

2.3. Prediction Interval Formulation

Suppose a collection of observed points ( x i , t i ) i = 1 N , the forecasting target can be presented as
t i = g ( x i ) + ε ( x i )
where t i is the i-th forecasting target corresponding to x i , the vector of the input, ε ( x i ) indicates the noise whose mean is zero and g ( x i ) is the true regression mean. The prediction of the GPR model g ^ ( x i ) is an assessment of the true regression value g ( x i ) .
If the evaluated error and noise are independent in statistics, the variance of the total prediction errors σ ^ t 2 ( x i ) is able to be gained from the variance of model uncertainty σ ^ g 2 ( x i ) and the variance of noise σ ^ ε 2 ( x i ) , shown in Equation (19),
σ ^ t 2 ( x i ) = σ ^ g 2 ( x i ) + σ ^ ε 2 ( x i )
Given a time series ( x i , t i ) , L t ( α ) ( x i ) , U t ( α ) ( x i ) is the 100(1−α)% confidence level prediction interval (PI) of t i , where U t ( α ) ( x i ) and L t ( α ) ( x i ) are the upper and lower bounds of PI respectively, denoted as:
L t ( α ) ( x i ) = g ^ ( x i ) z 1 α / 2 σ ^ t 2 ( x i )
U t ( α ) ( x i ) = g ^ ( x i ) + z 1 α / 2 σ ^ t 2 ( x i )
where z 1 α / 2 is the critical value of the standard Gaussian distribution, depending on the prospective confidence level 100(1−α)%, and α is the significant level. The future target t i is expected to lie in the constructed PI with the nominal probability as,
P t i L t ( α ) ( x i ) , U t ( α ) ( x i ) = 100 ( 1 α ) %

2.4. Performance Assessment

2.4.1. Point Prediction Evaluation

To better evaluate the accuracy of point predictive results, three different assessment metrics, root-mean-squared-error (RMSE), mean absolute error (MAE) and mean absolute percentage error (MAPE) are utilized to compute the prediction errors of the proposed model. The formulations are listed in Equations (15)–(17), respectively.
R M S E = 1 M i = 1 M ( y i y ^ i ) 2
M A E = 1 M i = 1 M y i y ^ i
M A P E = 1 M i = 1 M ( | y ^ i y i y ¯ i | ) × 100 %
where y i is the actual value of the soil moisture, y ^ i is the predictive value of the soil moisture, y ¯ i is the mean value of the actual value of test dataset, and M is the number of the test dataset.

2.4.2. Interval Prediction Evaluation

Reliability

The reliability is a significant attribute for assessing the accuracy of the approximated PIs. According to the previous introduction of PI, the feature target t i is supposed to be included in the constructed interval with the nominal probability 100(1−α)%, known as the PI nominal confidence (PINC). Given N test data points, PI coverage probability (PICP) indicates the pragmatic coverage probability of the corresponding PIs, defined in Equations (23) and (24).
P I C P = 1 N i = 1 N I i ( α )
I i ( α ) = 1 , t i L t ( α ) ( x i ) , U t ( α ) ( x i ) 0 , t i L t ( α ) ( x i ) , U t ( α ) ( x i )
The closer the obtained PICP is to the corresponding PINC, the higher the reliability of PIs. Therefore, the average coverage error (ACE) is able to evaluate the reliability of PIs.
A C E = P I C P P I N C
With the reliability increasing, the ACE is getting near to zero.

Sharpness

Increasing the width of PIs can easily improve the reliability of PIs, but it makes no sense for practical applications. The reliability index, PICP, is also interrelated to the sharpness, which is applied to evaluating the quality of PIs in this study.
The width of PI [ L t ( α ) ( x i ) , U t ( α ) ( x i ) ] , w t ( α ) ( x i ) , is calculated according to Equation (26).
w t ( α ) ( x i ) = U t ( α ) ( x i ) L t ( α ) ( x i )
The interval score can assess the overall skill of soil moisture PIs and reflect the sharpness perspective in Equation (27).
S t ( α ) ( x i ) = 2 α w t ( α ) ( x i ) 4 L t ( α ) ( x i ) t i ,   i f   t i < L t ( α ) ( x i ) 2 α w t ( α ) ( x i ) ,   i f   t i L t ( α ) ( x i ) , U t ( α ) ( x i ) 2 α w t ( α ) ( x i ) 4 t i U t ( α ) ( x i ) ,   i f   t i > U t ( α ) ( x i )
The overall score value Ω t ( α ) can be computed as,
Ω t ( α ) = 1 N i = 1 N S t ( α ) ( x i )
The interval score grants the narrow PI and gives penalties when the target is not in the PI. Thus, it takes both reliability and sharpness into consideration and it can play a significant role in evaluating the overall skill of PIs. When the reliabilities of PIs are similar, the larger score is equivalent to the higher sharpness and overall skill [32].

3. Experiment Results and Analysis

3.1. Data Description

In this paper, the soil moisture data were collected during the period from 28 February 2012 to 8 November 2016, in Beijing, China. The collected original daily average soil moisture data were utilized to form the soil moisture time series as training and test datasets for analysis and prediction. The collected soil moisture data are plotted in Figure 1.
During the experiment, the 8 continuous historical soil moisture data points and its corresponding response points (the 9th, 10th, 11th data points) form a data item. In other words, the continuous soil moisture of 8 days is used to forecast the soil moisture in the following 3 days. The full dataset consists of 1272 data items. The first 1000 data items make up the complete training set, and the rest of the 272 data items form the test set.
Y(1), Y(2), and Y(3) indicate the daily average soil moisture of the 9th, 10th, 11th days, respectively. The full training dataset makes it very time-consuming to develop the GPR model, and thus the RU algorithm is applied to sample the representative dataset from the large full dataset. The sample sizes of 100 and 500 are considered for the sample selection to examine the performance of the proposed method. Finally, the GPR models built with the full training data and the chosen data separately are tested on the same test dataset.

3.2. Point Prediction Analysis

The prediction performance of the proposed model built with the full training data and the chosen training data in predicting Y(1), Y(2), and Y(3) are evaluated. Meanwhile, the published forecasting model based on ARMA algorithm [13] is considered as a benchmark. Four metrics, RMSE, MAE, MAPE, and time, are utilized to assess the forecasting performance, and experimental results are shown in Table 1, Table 2 and Table 3. It can be seen from Table 1, Table 2 and Table 3 that the GPR model with the training sample sizes of 100 obtained the best performance over three models in terms of the lowest RMSE, MAE, and MAPE values. The GPR model trained using the 500 selected training samples yielded worse prediction results for Y(1) and Y(2) compared to the GPR model based on the full training dataset. However, the error difference is minor between these two models, and the forecasting performance of the GPR model trained using the 500 selected training samples is better for Y(3) prediction.
The point prediction results from Table 1, Table 2 and Table 3 show that using the chosen data can improve the prediction accuracy and much less computation time is required. Thanks to the RU design algorithm, the most representative subset is selected from the original training dataset. The better setting of the selected sample size can assist in improving the prediction performance of the GPR model and reducing the computation time simultaneously. Therefore, the proposed method is applicable for GPR point prediction using the large dataset. Compared with the ARMA-based model, the GPR model with the training sample sizes of 100 yields better forecasting performance in terms of lower RMSE, MAE, and MAPE values as well as shorter computing time.
To further illustrate the performance of the proposed method, the results of the GPR models built with different training samples for Y(1), Y(2), and Y(3) prediction are plotted in Figure 2, Figure 3 and Figure 4. In point prediction, the 100 training samples were selected as the chosen data.
Compared with the prediction results shown in Figure 3 and Figure 4, the prediction results in Figure 2 are closer to the actual soil moisture. It indicates that the prediction accuracy decreases with the increasing of the forecasting horizon.
From Figure 2, Figure 3 and Figure 4, the red dotted line denotes the actual value curve, the blue line demonstrates the forecasting value curve of the GPR model trained in the chosen data whose size is 100, and the green intermittent line indicates the forecasting value curve of the GPR model trained in the full training dataset. It can be seen that the blue curve is closer to the red curve, showing that the prediction accuracy of the GPR model trained using the chosen training samples is higher than the GPR model based on the full training dataset.

3.3. Interval Prediction Result Analysis

As shown in Table 4, Table 5 and Table 6, the GPR model built with the training dataset size of 500 generally performs the best for Y(1), Y(2), and Y(3) prediction. However, the GPR model based on the full training dataset generates better PICP and ACE values for confidence level 90% in Y(2) and Y(3) prediction. The GPR model trained with the 100 samples obtained the worst interval prediction performance over the three models. A reasonable reason is that the GPR model cannot learn the true distribution well with fewer training data samples. To demonstrate the interval predication results, the PIs of the proposed method are provided in Figure 5, Figure 6, Figure 7 and Figure 8. In interval prediction, the 500 training samples are selected as the chosen data.

3.4. Discussion

In terms of point prediction, it can be seen from Table 1, Table 2 and Table 3 that the GPR model with the training sample sizes of 100 obtained the best performance based on the two datasets. Thus, the extracted subset training dataset includes the most representative training samples and the well-performed forecasting model is developed. Meanwhile, because only 100 data samples are utilized, two benefits can be observed: (1) since noise data is less likely to be sample, the forecasting accuracy can be improved; (2) the computational time is significantly reduced due to the small size of the kernel matrix. The GPR model trained using the 500 selected training samples yielded worse prediction results for Y(1) and Y(2) compared to the GPR model based on the full training dataset. However, regarding the Y(3) forecasting, the GPR model trained using the 500 selected training samples performed better than the latter model. A possible reason is that some noise points are included in the 500 samples, and thus the model performance degrades. Compared with the widely applied ARMA algorithm [16], the proposed model has achieved better performance. The execution time of the GRP model based on the 100 training samples is much less than that of the ARMA model. Therefore, the proposed model is more suitable for real-time forecasting applications.
For interval prediction, as shown in Table 4, Table 5 and Table 6, the GPR model built with the training dataset size of 500 generally performs the best for Y(1), Y(2), and Y(3) prediction. However, the GPR model based on the full training dataset generates better PICP and ACE values for confidence level 90% in Y(2) and Y(3) prediction. The GPR model trained with the 100 samples obtained the worst interval prediction performance over the three models. This phenomenon may occur because the GPR model cannot learn the true distribution well when the training data sample is small, and 500 sample points are enough for the model to learn the sample distribution, and the entire data set is not required. The interval prediction verifies the effectiveness of the RU design algorithm in selecting representative samples.

3.5. Seasonal Factor Analysis

Taking the impact of the season factor on the forecasting performance into account, 30-day data of June (Summer) and September (Fall) was used as the test dataset respectively to compare the prediction performance based on different training sample size. The test data of June and September are noted as test June and test Sep.
From Table 7 and Table 8, it is observed that the RMSE, MAE, MAPE values of the GPR models for test June are all higher than those of models for full test data. However, the performance of GPR models for test Sep is better than that of GPR models for full test data. The interval prediction results of the two selected months are drawn in Figure 8, Figure 9, Figure 10 and Figure 11.
It is demonstrated in Figure 11, Figure 12, Figure 13 and Figure 14 that the seasonal factor has an impact on the prediction performance of the GPR models. In June, the weather changes dramatically with much rainfall. Therefore, compared with September, the prediction in June is more difficult. It is more challenging to forecast soil moisture in summer.

4. Conclusions

In this paper, an RU design algorithm was applied to sample representative data points to develop the GPR model in predicting the soil moisture. The following conclusions are drawn from this study.
(1) Through the RU design algorithm, the most representative subset (the chosen data) was sampled from the complete training dataset. Based on the chosen data with proper sample size setting, the performance of the GPR model was improved and much less computing time was required.
(2) The RMSE, MAE, and MAPE values of the GPR model built with the chosen data were smaller compared with the GPR model based on the full training data in terms of the point prediction.
(3) Regarding the interval prediction, PIs of the GPR model built with the chosen data were better considering the reliability, sharpness and overall skill.
In future work, the integration of the RU design algorithm and other learning algorithms, such as deep neural networks [33] and boosted regression trees [34], will be investigated for the soil moisture prediction.

Author Contributions

Methodology, L.W. and C.H.; validation, M.L., Y.Z. and X.L.; writing, M.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Key R&D Program of China under Grant 2018YFC0810600, in part by Scientific and Technological Innovation Foundation of Shunde Graduate School, USTB under Grants BK19BF006 and BK20BF010, in part by the Interdisciplinary Research Project for Young Teachers of USTB (Fundamental Research Funds for the Central Universities) under Grant FRF-IDRY-19-017, in part by the Fundamental Research Funds for the Central Universities under Grants 06500078 and 06500103, in part by the Beijing Natural Science Foundation under Grant 9204028, in part by the Beijing Talents Plan under Grant BJSQ2020008, in part by the Visiting Scholarship of State Key Laboratory of Power Transmission Equipment & System Security and New Technology (Chongqing University), and in part by the Open Research Subject of Key Laboratory of Fluid and Power Machinery (Xihua University), Ministry of Education, under Grant szjj2019-011.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Seneviratne, S.I.; Corti, T.; Davin, E.L.; Hirschi, M.; Jaeger, E.B.; Lehner, I.; Orlowsky, B.; Teuling, A.J. Investigating soil moisture–climate interactions in a changing climate: A review. Earth Sci. Rev. 2010, 99, 125–161. [Google Scholar] [CrossRef]
  2. Garg, A.; Gadi, V.K.; Feng, Y.C.; Lin, P.; Qinhua, W.; Ganesan, S.; Mei, G. Dynamics of soil water content using field monitoring and AI: A case study of a vegetated soil in an urban environment in China. Sustain. Comput. Inform. Syst. 2019, 100301. [Google Scholar] [CrossRef]
  3. Chatterjee, S.; Dey, N.; Sen, S. Soil moisture quantity prediction using optimized neural supported model for sustainable agricultural applications. Sustain. Comput. Inform. Syst. 2018. [Google Scholar] [CrossRef]
  4. Corradini, C. Soil moisture in the development of hydrological processes and its determination at different spatial scales. J. Hydrol. 2014, 516, 1–5. [Google Scholar] [CrossRef]
  5. Brocca, L.; Ciabatta, L.; Massari, C.; Camici, S.; Tarpanelli, A. Soil moisture for hydrological applications: Open questions and new opportunities. Water 2017, 9, 140. [Google Scholar] [CrossRef]
  6. Sehgal, V.; Venkataramana, S. Watershed-scale retrospective drought analysis and seasonal forecasting using multi-layer, high-resolution simulated soil moisture for Southeastern US. Weather Clim. Extrem. 2019, 23, 100191. [Google Scholar] [CrossRef]
  7. Ford, T.W.; Quiring, S.M. Comparison of contemporary in situ, model, and satellite remote sensing soil moisture with a focus on drought monitoring. Water Resour. Res. 2019, 55, 1565–1582. [Google Scholar] [CrossRef]
  8. Pendergrass, A.G.; Meehl, G.A.; Pulwarty, R.; Hobbins, M.; Hoell, A.; AghaKouchak, A.; Bonfils, C.J.; Gallant, A.J.; Hoerling, M.; Hoffmann, D.; et al. Flash droughts present a new challenge for subseasonal-to-seasonal prediction. Nat. Clim. Chang. 2020, 10, 191–199. [Google Scholar] [CrossRef]
  9. Brocca, L.; Melone, F.; Moramarco, T.; Wagner, W.; Naeimi, V.; Bartalis, Z.; Hasenauer, S. Improving runoff prediction through the assimilation of the ASCAT soil moisture product. Hydrol. Earth Syst. Sci. 2010, 14, 1881–1893. [Google Scholar] [CrossRef] [Green Version]
  10. Koster, R.D.; Mahanama, S.P.; Livneh, B.; Lettenmaier, D.P.; Reichle, R.H. Skill in streamflow forecasts derived from large-scale estimates of soil moisture and snow. Nat. Geosci. 2010, 3, 613. [Google Scholar] [CrossRef]
  11. Brocca, L.; Ponziani, F.; Moramarco, T.; Melone, F.; Berni, N.; Wagner, W. Improving landslide forecasting using ASCAT-derived soil moisture data: A case study of the Torgiovannetto landslide in central Italy. Remote Sens. 2012, 4, 1232–1244. [Google Scholar] [CrossRef] [Green Version]
  12. Bittelli, M.; Valentino, R.; Salvatorelli, F.; Pisa, P.R. Monitoring soil-water and displacement conditions leading to landslide occurrence in partially saturated clays. Geomorphology 2012, 173, 161–173. [Google Scholar] [CrossRef]
  13. Yang, S.; Yiming, W.; Zhengqin, G. Research on soil moisture forecast based on ARIMA model. Agric. Res. Arid Areas 2006, 24, 114–118. [Google Scholar]
  14. Chifflard, P.; Kranl, J.; Zur Strassen, G.; Zepp, H. The significance of soil moisture in forecasting characteristics of flood events. A statistical analysis in two nested catchments. J. Hydrol. Hydromech. 2018, 66, 1–11. [Google Scholar] [CrossRef] [Green Version]
  15. Graham, S.L.; Laubach, J.; Hunt, J.E.; Eger, A.; Carrick, S.; Whitehead, D. Predicting soil water balance for irrigated and non-irrigated lucerne on stony, alluvial soils. Agric. Water Manag. 2019, 226, 105790. [Google Scholar] [CrossRef]
  16. Chen, H. Research on Soil Moistrue Prediction Models of Inadequate Irrigation Rice Field in South China. Ph.D. Thesis, Yangzhou University, Yangzhou, China, 2004. [Google Scholar]
  17. Mahmood, R.; Hubbard, K.G. An analysis of simulated long-term soil moisture data for three land uses under contrasting hydroclimatic conditions in the Northern Great Plains. J. Hydrometeorol. 2004, 5, 160–179. [Google Scholar] [CrossRef] [Green Version]
  18. Liang-chen, Z. Study on Estimation of Soil-water Content by Using Soil-Water Dynamics Model. Water Sav. Irrig. 2007, 3, 10–13. [Google Scholar]
  19. Ines, A.V.; Das, N.N.; Hansen, J.W.; Njoku, E.G. Assimilation of remotely sensed soil moisture and vegetation with a crop simulation model for maize yield prediction. Remote Sens. Environ. 2013, 138, 149–164. [Google Scholar] [CrossRef] [Green Version]
  20. Zhang, X.; Tang, Q.; Liu, X.; Leng, G.; Li, Z. Soil moisture drought monitoring and forecasting using satellite and climate model data over southwestern China. J. Hydrometeorol. 2017, 18, 5–23. [Google Scholar] [CrossRef]
  21. Xiao-gang, M.A. Forecast of Soil Moisture Content during Critical Period of Spring Sowing Based on Precipitation in Last Autumn. Chin. J. Agrometeorol. 2008, 29, 55–57. [Google Scholar]
  22. Sukhwinder, S.; Kaur, S.; Kumar, P. Forecasting soil moisture based on evaluation of time series analysis. In Advances in Power and Control Engineering; Springer: Singapore, 2020; pp. 145–156. [Google Scholar]
  23. Prasad, R.; Deo, R.C.; Li, Y.; Maraseni, T. Soil moisture forecasting by a hybrid machine learning technique: ELM integrated with ensemble empirical mode decomposition. Geoderma 2018, 330, 136–161. [Google Scholar] [CrossRef]
  24. Huang, C.; Li, L.; Ren, S.; Zhou, Z. Research of soil moisture content forecast model based on genetic algorithm BP neural network. In International Conference on Computer and Computing Technologies in Agriculture; Springer: Berlin/Heidelberg, Germany, 2010. [Google Scholar]
  25. Yi, L.; Tao Chen, T.; Chen, J. Auto-switch Gaussian process regression-based probabilistic soft sensors for industrial multigrade processes with transitions. Ind. Eng. Chem. Res. 2015, 54, 5037–5047. [Google Scholar]
  26. Liu, Y.; Chou, C.P.; Chen, J.; Lai, J.Y. Active learning assisted strategy of constructing hybrid models in repetitive operations of membrane filtration processes: Using case of mixture of bentonite clay and sodium alginate. J. Membr. Sci. 2016, 515, 245–257. [Google Scholar] [CrossRef]
  27. Yi, L.; Wu, Q.-Y.; Chen, J. Active selection of informative data for sequential quality enhancement of soft sensor models with latent variables. Ind. Eng. Chem. Res. 2017, 56, 4804–4817. [Google Scholar]
  28. Huang, C.; Zhao, Z.; Wang, L.; Zhang, Z.; Luo, X. The Point and Interval Forecasting of Solar Irradiance with an Active Gaussian Process. IET Renew. Power Gener. 2020, 14, 1020–1030. [Google Scholar] [CrossRef]
  29. Long, W.; Zhang, Z. Automatic detection of wind turbine blade surface cracks based on UAV-taken images. IEEE Trans. Ind. Electron. 2017, 64, 7293–7303. [Google Scholar]
  30. Rasmussen, C.E. Gaussian processes in machine learning. In Summer School on Machine Learning; Springer: Berlin/Heidelberg, Germany, 2003. [Google Scholar]
  31. Matthias, T.; Zhang, Z. Wind turbine modeling with data-driven methods and radially uniform designs. IEEE Trans. Ind. Inform. 2016, 12, 1261–1269. [Google Scholar]
  32. Huang, C.; Zhao, Z.; Wang, L.; Zhang, Z.; Luo, X. A hybrid approach for probabilistic forecasting of electricity price. IEEE Trans. Smart Grid 2013, 5, 463–470. [Google Scholar]
  33. Wang, L.; Zhang, Z.; Long, H.; Xu, J.; Liu, R. Wind Turbine Gearbox Failure Identification with Deep Neural Networks. IEEE Trans. Ind. Inform. 2017, 13, 1360–1368. [Google Scholar] [CrossRef]
  34. Huang, C.; Wang, L.; Lai, L.L. Data-Driven Short-Term Solar Irradiance Forecasting Based on Information of Neighboring Sites. IEEE Trans. Ind. Electron. 2019, 66, 9918–9927. [Google Scholar] [CrossRef]
Figure 1. Full data of daily average soil moisture.
Figure 1. Full data of daily average soil moisture.
Water 12 03085 g001
Figure 2. Points forecasting results in Y(1) prediction.
Figure 2. Points forecasting results in Y(1) prediction.
Water 12 03085 g002
Figure 3. Points forecasting results in Y(2) prediction.
Figure 3. Points forecasting results in Y(2) prediction.
Water 12 03085 g003
Figure 4. Points forecasting results in Y(3) prediction.
Figure 4. Points forecasting results in Y(3) prediction.
Water 12 03085 g004
Figure 5. GPR model built with full training data in Y(1) prediction.
Figure 5. GPR model built with full training data in Y(1) prediction.
Water 12 03085 g005
Figure 6. GPR model built with full training data in Y(2) prediction.
Figure 6. GPR model built with full training data in Y(2) prediction.
Water 12 03085 g006
Figure 7. GPR model built with full training data in Y(3) prediction.
Figure 7. GPR model built with full training data in Y(3) prediction.
Water 12 03085 g007
Figure 8. GPR model built with the chosen data in Y(1) prediction.
Figure 8. GPR model built with the chosen data in Y(1) prediction.
Water 12 03085 g008
Figure 9. GPR model built with the chosen data in Y(2) prediction.
Figure 9. GPR model built with the chosen data in Y(2) prediction.
Water 12 03085 g009
Figure 10. GPR model built with the chosen data in Y(3) prediction.
Figure 10. GPR model built with the chosen data in Y(3) prediction.
Water 12 03085 g010
Figure 11. Y(1) prediction with full training data and testing in test June.
Figure 11. Y(1) prediction with full training data and testing in test June.
Water 12 03085 g011
Figure 12. Y(1) prediction with the chosen training data and testing in test June.
Figure 12. Y(1) prediction with the chosen training data and testing in test June.
Water 12 03085 g012
Figure 13. Y(1) prediction with full training data and testing in test Sep.
Figure 13. Y(1) prediction with full training data and testing in test Sep.
Water 12 03085 g013
Figure 14. Y(1) prediction with the chosen training data and testing in test Sep.
Figure 14. Y(1) prediction with the chosen training data and testing in test Sep.
Water 12 03085 g014
Table 1. Comparison of the proposed models built with two datasets in Y(1) prediction.
Table 1. Comparison of the proposed models built with two datasets in Y(1) prediction.
ModelRMSEMAEMAPETime(s)
GPR + full dataset0.9080.4594.20910.204
GPR + 100 data samples0.8780.4464.0920.339
GPR + 500 data samples0.9180.4784.3802.594
ARMA0.9200.5905.6261.058
Note: RMSE (root-mean-squared-error); MAE (mean absolute error); MAPE (mean absolute percentage error). Statistics in bold indicate best results.
Table 2. Comparison of the proposed models built with two datasets in Y(2) prediction.
Table 2. Comparison of the proposed models built with two datasets in Y(2) prediction.
Training Dataset SizeRMSEMAEMAPETime(s)
GPR + full dataset1.2670.7927.25514.795
GPR + 100 data samples1.2330.7757.1010.393
GPR + 500 data samples1.2800.8167.4803.833
ARMA1.1680.8247.8161.091
Statistics in bold indicate best results.
Table 3. Comparison of the proposed models built with two datasets in Y(3) prediction.
Table 3. Comparison of the proposed models built with two datasets in Y(3) prediction.
Training Dataset SizeRMSEMAEMAPETime(s)
GPR + full dataset1.5081.0289.41312.579
GPR + 100 data samples1.4570.9708.8860.324
GPR + 500 data samples1.5050.9979.1313.148
ARMA1.6361.2719.8801.119
Statistics in bold indicate best results.
Table 4. Evaluation results of PIs in Y(1) prediction.
Table 4. Evaluation results of PIs in Y(1) prediction.
PINC (%)Training Dataset SizePICP (%)ACE (%)Score
90Full93.1303.130−0.945
10083.969−6.03−0.889
9550093.1293.129−0.970
Full93.893−1.107−0.670
10087.786−7.213−0.702
50093.893−1.106−0.670
99Full95.802−3.198−0.340
10093.511−5.488−0.494
50096.564−2.435−0.321
Statistics in bold indicate best results.
Table 5. Evaluation results of PIs in Y(2) prediction.
Table 5. Evaluation results of PIs in Y(2) prediction.
PINC (%)Training Dataset SizePICP (%)ACE (%)Score
90Full92.3662.366−1.296
10085.496−4.503−1.262
9550093.1293.129−1.311
Full94.275−0.725−0.874
10089.312−5.687−0.921
50095.0380.038−0.870
99Full95.420−3.580−0.381
10093.511−5.488−0.519
50095.801−3.198−0.351
Statistics in bold indicate best results.
Table 6. Evaluation results of PIs in Y(3) prediction.
Table 6. Evaluation results of PIs in Y(3) prediction.
PINC (%)Training Dataset SizePICP (%)ACE (%)Score
90Full90.8390.839−1.518
10080.152−9.847−1.551
9550091.9841.984−1.507
Full92.748−2.252−0.975
10086.641−8.358−1.114
50092.748−2.251−0.960
99Full94.656−4.344−0.317
10091.221−7.778−0.603
50095.419−3.580−0.304
Statistics in bold indicate best results.
Table 7. Y(1) Prediction of the proposed models built with full training dataset.
Table 7. Y(1) Prediction of the proposed models built with full training dataset.
Test DatasetRMSEMAEMAPETime(s)
full test0.9080.4594.20910.204
test June1.2530.7198.94010.686
test Sep0.6700.4254.0549.529
Statistics in bold indicate best results.
Table 8. Y(1) Prediction of the proposed models built with the chosen data whose size is 100.
Table 8. Y(1) Prediction of the proposed models built with the chosen data whose size is 100.
Test DatasetRMSEMAEMAPETime(s)
full test0.8780.4464.0920.339
test June1.2060.81610.1560.187
test Sep0.5840.4143.9470.207
Statistics in bold indicate best results.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Liu, M.; Huang, C.; Wang, L.; Zhang, Y.; Luo, X. Short-Term Soil Moisture Forecasting via Gaussian Process Regression with Sample Selection. Water 2020, 12, 3085. https://0-doi-org.brum.beds.ac.uk/10.3390/w12113085

AMA Style

Liu M, Huang C, Wang L, Zhang Y, Luo X. Short-Term Soil Moisture Forecasting via Gaussian Process Regression with Sample Selection. Water. 2020; 12(11):3085. https://0-doi-org.brum.beds.ac.uk/10.3390/w12113085

Chicago/Turabian Style

Liu, Mingshuai, Chao Huang, Long Wang, Yu Zhang, and Xiong Luo. 2020. "Short-Term Soil Moisture Forecasting via Gaussian Process Regression with Sample Selection" Water 12, no. 11: 3085. https://0-doi-org.brum.beds.ac.uk/10.3390/w12113085

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop