Figures
Abstract
The shortage of available water resources and climate change are major factors affecting agricultural irrigation. In order to improve the irrigation water use efficiency, it is necessary to predict the water requirements for crops in advance. Reference evapotranspiration (ETo) is a hypothetical standard reference crop evapotranspiration, many types of artificial intelligence models have been applied to predict ETo; However, there are still few in the literature regarding the application of hybrid models for deep learning model parameters optimization. This paper proposes two hybrid models based on particle swarm optimization (PSO) and long-short-term memory (LSTM) neural network, used to predict ETo at the four climate stations, Shaanxi province, China. These two hybrid models were trained using 40 years of historical data, and the PSO was used to optimize the hyperparameters in the LSTM network. We applied the optimized model to predict the daily ETo in 2019 under different datasets, the result showed that the optimized model has good prediction accuracy. The optimized hybrid models can help farmers and irrigation planners to make plan earlier and precisely, and can provide valuable information to improve tasks such as irrigation planning.
Citation: Jia W, Zhang Y, Wei Z, Zheng Z, Xie P (2023) Daily reference evapotranspiration prediction for irrigation scheduling decisions based on the hybrid PSO-LSTM model. PLoS ONE 18(4): e0281478. https://doi.org/10.1371/journal.pone.0281478
Editor: Andrew Lewis, Griffith University, AUSTRALIA
Received: April 11, 2022; Accepted: January 24, 2023; Published: April 18, 2023
Copyright: © 2023 Jia et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: All relevant data are within the paper and its Supporting Information files.
Funding: This work was supported in part by the National Key Research and Development Project of the 13th five-year plan fertilizer-water source-equipment adaptation technology and control equipment (No.2017YFD0201504)(http://www.most.gov.cn/index.html), by the Key R&D Program of Shaanxi Province (2023-YBNY-202), by Ningbo Science and Technology Plan Project (2021S022), by the Zhejiang Province Basic Public welfare Research Program (LGN20F030001), and by the Key Industrial Innovation Chain Projects of Shaaxi Province (2023-ZDLNY-67).
Competing interests: The authors have declared that no competing interests exist.
Introduction
Reference evapotranspiration is a hypothetical standard reference crop evapotranspiration, which plays a broad and important role in irrigation decision-making, hydrological prediction and scheduling, crops growth simulation and climate disaster monitoring [1]. The ETo can be determined using lysimeter, which provide accurate measurements typically using in the development and validation of other methods. Given the cost and complexity of lysimeters, its use is typically restricted to research. Thus, the use of mathematical models based on meteorological data (temperature, relative humidity, solar radiation and wind speed) recorded by weather stations is a more suitable approach for practical applications [2]. The Penman-Monteith FAO-56 method is used as the standard method to estimate the ETo, and this method has served as a criterion for comparing the forecast values of other models [3].
Over the last few years, with the rapid development of computers’ computing ability and artificial intelligence theory, computing ETo using weather data has been considered a regression task that can be solved by some classical machine learning model for estimating ETo [4–6], different types of artificial intelligence methods have been applied to estimate and predict ETo, including ANN, gene expression programming (GEP), support vector machine (SVM), random forest (RF), extreme gradient boosting (XGBoost), adaptive neuro-fuzzy inference system (ANFIS), and multi-layer perceptron neural network [7–11]. The deep learning techniques such as the Temporal Convolution Network, the Convolutional Neural Networks (CNN), the Long Short-Term Memory model (LSTM) and hybrid the Convolutional Neural Networks and the Long Short-Term Memory model (CNN-LSTM) also have been recently employed to predict ETo, and these models show outstanding performances [12–16]. Although these models have been used to estimate daily or monthly ETo, estimation studies of model hyper parameters are not as common in the literature as ETo forecasting studies. The decision making in irrigation scheduling depends on forecasts of 1 to 10 days ahead, which is critically important to determine crop water requirements and real time irrigation scheduling. The shorter time from daily to 7 days can be useful in planning the use of irrigation systems as well as in optimizing system power consumption. Longer forecast horizons also help in the water management of irrigation channels and reservoirs for irrigation use [17, 18].
In recent year, the LSTM and CNN are probably the most popular, efficient and widely used deep techniques for time series forecasting. LSTMs were introduced to overcome the problems of vanishing gradients of RNNs with capability of storing important information containing long sequences, which were used in a number of applications, including speech recognition, stock price prediction, image text recognition, traffic flow forecasting, agriculture rainfall forecast, and grammar learning [19, 20]. However, the application of LSTM in the field of hydrology has not been widely reported in the literature, but can be used in the estimation of hydrologic variables since several climatic variables used in hydrology exhibit time series behavior [21]. In the LSTM neural network, since the model training is an automatic process of adjusting weights and thresholds, the values of the hyperparameters will directly affect the occurrence of convergence, learning time, and local minima. However, in the existing study, the LSTM models provided only slight performance gains. Furthermore, the models generally have more hyperparameters to be adjusted, requiring more intelligent algorithm to optimize them. Hybrid techniques, such as ensemble modeling, usually offer better results than simple techniques since the combination of the models tends to capture the best of each one. Hybrid models are also developed to combine the advantages of different methods and form a new forecasting strategy. Therefore, many of them are considered to be more effective than pure classical methods or artificial intelligence models [22]. Particle swarm optimization (PSO) is a global optimization algorithm with simple rules and fast convergence. It has been widely used in neural network training and structural optimization design [23]. However, most parameters in the LSTM model need to be set manually, which is inefficient and unreasonable. The potential of a hybrid algorithm based on LSTM and PSO is still unexplored in the literature for ETo forecasting. Considering the optimization of the parameters of the shortcomings of the LSTM model, the most recently advanced hybrid technique was selected to improve the performance of the LSTM forecasting model. Specifically, the PSO was used to optimize the hyperparameters in the LSTM network, and proposes two hybrid model used to predict the daily ETo in four climate stations, Shaanxi province, China. Compared to existing models, the optimized model has better feasibility and prediction accuracy.
The innovation of this paper lies mainly in the following: two hybrid deep neural network model were proposed for daily ETo forecasting. The number of hidden neurons, dropout and look-back in LSTM are optimized by PSO. The optimized models were tested with four different datasets, and the proposed models achieved good results in forecast accuracy under different datasets.
This paper is organized as follows, in material and methods, an overview of study area, reference evapotranspiration and LSTM prediction model were given, and two hybrid models based on PSO and LSTM were introduced. In results and discussion, the result and analysis of optimized forecasting model were given, and we compare the proposed model under four different datasets. In conclusions, we conclude this paper with future work.
Material and methods
Study site
The Guanzhong Basin is a typical and important grain producing area in China, is located in the central part of Shaanxi province in China, and covers an area of 20 440 km2 (33°39′ - 35°50′ N, 107°30′ -110°37′ E), it has a warm temperate semi-humid monsoon climate, with the average annual temperature of 13.7° C, and the mean annual precipitation recorded from 2000 to 2010 was approximately 615 mm. The Wugong (WG, 34°19′ N, 108°14′E, 471.0m), Fengxiang (FX, 34°31′ N, 107°23′E, 781.1m), Xianyang (XY, 34°24′ N, 108°43′E, 472.8m) and Pucheng (PC, 34°53′ N, 109°38′E, 387.2m) weather stations are located in the Guanzhong Basin [24, 25], as the study sites for ETo forecasting.
Reference evapotranspiration
The general FAO Penman-Monteith approach (FAO56) [26] has been widely used to calculate reference evapotranspiration, because most users appreciate its simplicity and consider that it has acceptable accuracy. The FAO56 is expressed as Eq (1):
(1)
Herein, ETo represents the reference evapotranspiration (mm day−1); Rn represents net radiation at the crop surface (MJ m−2 day−1); G represents soil heat flux density (MJ m−2day−1); T represents mean daily air temperature at 2 m height (°C); u2 represents wind speed at 2 m height (m s−1); es represents saturation vapor pressure (kPa); ea represents actual vapor pressure (kPa); (es − ea) represents saturation vapor pressure deficit (kPa); Δ represents slope vapor pressure curve (kPa °C−1); γ represents psychrometric constant (kPa ° C−1).
The China National Meteorological Data Center (http://data.cma.cn, accessed on 30 December 2019) provides accurate and free weather data, the weather data types include daily mean temperature (Tmean,°C), daily maximum temperature (Tmax,°C), daily minimum temperature (Tmin,°C), average relative humidity (RH, %), daily average wind speed (U2, m/s) and sunshine hours (Sh, h). From January 1, 1980 to December 31, 2019, a total of 14600 weather data sample were obtained for each weather station (accessed on December 31, 2019). Table 1 presents the data statistics for each weather parameters. Since the elevations and latitude of the four stations are similar, the maximum, average and minimum values of the six weather parameters at the four stations are close to each other. It can be inferred that a predictive model optimized using weather data from one of the four stations can be used to predict parameters for the other sites.
Using weather data and the FAO56 approach, a total of 14600 daily ETo data were obtained for each weather station (Fig 1). It can be seen from the figure the value ranges for ETo are from 0.3 mm to 10 mm, and the peak period for each year’s data is from July to September. Due to cyclical changes in weather parameters, the data of ETo fluctuate greatly in different days, ETo changes with periodic changes in weather parameters. Table 1 presents the data statistics for the weather data and ETo.
Long short-term memory network model
The LSTM are widely applied in time series forecasting, which consists of input layer, LSTMs layer and output layer. The main components of the LSTM network consist of a sequence input layer that is used to input a sequence (time series data), and a sequence output layer that is used to learn long-term reliance among the time steps of a sequence data [27].
An LSTM-NN cell consists of three gates: input gate, output gate and forget gate. The framework of LSTM cell is shown in (Fig 2), the input gate gives new input to the cell. The output gate specifies the output of the cell, and the forget gate is responsible for specifying the prior values that need to be retained for future reference. The work principle of LSTM neural network at time step t are as follows [27, 28]:
(2)
(3)
(4)
(5)
(6)
(7)
(8)
(9)
The σ refers to the sigmoid function and controls the information passing state. When the σ is 0, the nothing can pass. When the σ is 1, everything can pass. Wf, Wi, Wc and Wo refer to the input weight. The corresponding bf, bi, bc and bo refer to the biasing. The t and t-1 refer to the current and previous time status. The x and h refer to the input and output, and C refers to the cell status.
Particle swarm optimization algorithm
As one of the evolutionary calculation technologies, the PSO algorithm searches an optimal solution of each particle in each iteration and records it as the current individual extremum (particle best, pbest), compares all current individual in search space, the best is denoted as the global extremum (global best, gbest) of the entire particle swarm.
Each particle has its speed and position in the next each subsequent iteration, all the particles in the particle swarm adjust their speed and position by pbest and gbest [29]. The particle updates its speed and position according to the following Formulas (10) and (11) [30]:
(10)
(11)
where k denotes the number of iterations. c1 and c2 are acceleration constant, which are used to adjust the maximum learning step. w is the inertia factor, which is used to adjust the search range of solution space, and r1 and r2 are uniform random numbers within the range [0, 1] to increase the randomness of the search.
Hybrid model based on LSTM and PSO
Similar to other neural networks, the LSTM is very sensitive to hyperparameters, including the size of the time window, the batches, hidden layer neurons, dropout, activate function.
In this study, the framework of two PSO-LSTM model is constructed, as shown in (Fig 3). the first model consisting of a LSTM layer, a flatten layer and a dense layer, and two important hyperparameters, the number of the first hidden layer neurons (N1) and the look back (time windows, T1) were used to be optimized. The second model consisting of two LSTM layer, a flatten layer and a dense layer, and four important hyperparameters, the number of the first hidden layer neurons (N2), the second hidden layer neurons (N3), the dropout (D) and the look back (T2) were used to be optimized. Those hyperparameters are regarded as the particles in PSO. The mean absolute error (MAE) of the model’s predicted and actual values are taken as the fitness function.
The implementation process of the hybrid model was shown in (Fig 4). The detailed optimized process of the hybrid model is as follows [29, 31]:
Step 1: Import the dataset and normalization, and dividing the dataset into training set, validation set and test set.
Step 2: PSO is used to optimize the hyperparameters.
- Parameter initialization. The particle dimension, population size, iterations, learning factors c1 and c2, inertia weight w, velocity and position are determined.
- Initialize the velocity and position of particles, and randomly generate population particles.
- The mean absolute error is taken as the fitness value of the model, and the pbest and gbest are calculated with the fitness value of the particle.
- After each iteration, the position and velocity are updated, and the fitness is also calculated, at the same time, the pbest and gbest are also updated.
- Determine whether the termination condition is met. If satisfied, the optimization results were obtained. Otherwise, go back to step 3).
Step 3: The optimized LSTM model were obtained, and training and evaluate the optimized model with other weather stations data.
Data pre-processing
The stationarity of daily ETo is important for LSTM model forecasting, long-term observational ETo records are needed to detect the stationarity. The detection method for the sequence data stationarity includes the intuitive judgment of sequence data image, the method of the Dickey-Fuller test [32], the method of the autocorrelation coefficient diagram and the partial correlation coefficient diagram. The partial autocorrelation coefficient (PAC) of the sequence data represents the correlation between any two different time steps for the same time series. The PAC of daily ETo was shown in (Fig 5), and show the lags from 0 to 50 days. It can be seen from the figure that the PAC of daily ETo has a rapid decay to near zero with increases of the lag numbers, the confidence limits for the autocorrelations were ±0.02 at the 95% confidence intervals. Although the PAC of daily ETo changes when the lag is greater than 10, the fluctuation range is little, so the sequence data of daily ETo was a stationary random parameter.
Experimental and parameter setting
Daily ETo sequence data from the study site were recorded from 1980 to 2019, over the 40-year period in question, there have been marked changes in global ambient temperature and other weather data. In this study, the ETo was calculated by mixing six weather parameters, it can be seen from Table 1 that the standard deviation of the ETo is much smaller than the standard deviation of five weather parameters at each station. The research interval in this paper is 1 day, and the change of meteorological parameters in one day is random, this paper cannot accurately describe the change of meteorological parameters, so this paper ignores the influence of meteorological parameter changes on the prediction model. The data were split into training set (1980–2007), test set (2008–2018) and prediction set (2019), the data number of training set, test set and prediction set were 10227, 4021, and 365, respectively. Since the ETo are all positive, the sequence data of ETo was normalized using the function of MinMaxScaler in Pandas (https://pandas.pydata.org/) [33], and transform the data into the range [0, 1]. After training of models, the data were back transformed to the original scale.
Based on preliminary training, under 20 particles, the training time of the single hidden layer LSTM model is not less than 4h, For the four stations, the time cost of building the prediction models separately is huge and difficult. Given this, the prediction model was trained using data from the WG station. According to relevant literature [34], the setting of the hyperparameters for the hybrid model adopts the grid search method and the swarm intelligence method. In this study, the first model was used for preliminary experiments with hyperparameter setting, it was determined that the N1 is from 10 to 60, the T1 is from 1 to 50, the activation function = ReLU, optimizer = Adam, dropout = 0.1, batch size = 64 and the number of train steps is 500. At the same time, the PSO algorithm parameters are set as: w = 0.5, c1 = c2 = 0.5, and the number of particles = 20.
After the first model experiment, in order to obtain better performing model, we increased the range of hidden layers and hyperparameters of the model, and the second model was constructed. It was determined that the N2 is from 1 to 200, the N3 is from 1 to 200, the T2 is from 1 to 200, dropout = 0.1,0.2, or 0.3, the activation function = ReLU, optimizer = Adam, batch size = 64, and the number of train steps is 100. At the same time, to ensure a fair comparison, the PSO algorithm parameters are set as: w = 0.5, c1 = c2 = 0.5, and the number of particles = 20.
This paper uses the Keras framework (https://keras.io/) [35], which is an API designed for human beings, not machines, and regards the popular Deep Learning framework TensorFlow as the backstage supporter, to build the prediction model. The CPU in the experimental is Inter(R) Core (TM) i5 8500 @3.00GHz, and the RAM is 8GB, the version of Python is 3.6, and the software platform is the PyCharm 2021.3.
In order to evaluate the performance of the optimized model at other three stations, the optimized model was evaluated using the data from other three stations. At the same time, in order to verify the accuracy of the optimized model under different period, four types of periods are divided, including type A (training set 1980–2007,10220, test set 2008–2018, 4015), type B (training set 1980–2017,13869, test set, 1981–2018,13869), type C (training set 2003–2014, 4380, test set, 2015–2018,1460) and type D (training set 2011–2017, 2190, test set, 2017–2018, 730). The amount of data for each type of training set and test set is shown in the (Fig 6). The optimized model is trained again using the above training set, and the following evaluation criteria are used to evaluate hybrid model.
Model evaluation criteria
Four evaluation measures were selected to indicate the performance of the different models [36].
Mean Absolute Error (MAE) is:
(12)
The Mean Squared Error (MSE) is:
(13)
The root mean square error (RMSE) is:
(14)
In the above formula, yi represents the predicted value. xi represents the true value. is the average value. N represent the number of prediction value. MAE is the mean absolute error. It can reflect the actual situation of the predicted value error. The MSE is the expected value of the square of the difference between the parameter estimate and the parameter true value, it can evaluate the degree of the data change, and the smaller value of the MSE, the better accuracy of the prediction model. RMSE is the square root of MSE. R2 can eliminate the influence of dimension on evaluation measure.
Results and discussion
The result of the first optimization model
For the first optimized model, the PSO-LSTM optimization iterated 20 times. (Fig 7) present the training time, maximum fitness, average fitness and minimum fitness in each iteration. From the figures, we can find that the training time cost increases rapidly with the number of iterations. Although the batch size = 64, T1 and N1 of hybrid model is [1,50] and [10, 60], respectively, the utilization of CPU gradually decreases during the training process. The training time for the first and last iterations is 4.27h and 17.30h, respectively, and the total training time of 20 times iterations is 221.25h (9.21 days). The fitness gradually decreases as the number of iterations increases, and the minimum fitness is much smaller than the maximum fitness of each iteration. After 20 times iterations, the maximum, average and minimum fitness of optimal particle are 0.637, 0.608 and 0.597, respectively. After 20 times iterations, the hyperparameter combination corresponding to the minimum fitness is [5, 24], it means that the look back = 5, the number of first layer neurons = 24 is the optimal combination for the first model.
Prediction and evaluation of the first model.
In order to evaluate the performance of model, the optimized model was trained again using the four different data set (type A, type B, type C and type D) for each station. The number of training steps during the optimization process is 500, in order to analyze the effect of different training steps on the model accuracy, the number of training steps were set as 500, 1000, 1500, and 2000, respectively. Table 2 shows the training results.
For the four stations and four datasets, it can be seen from the table that the most RMSE and MAE of training set gets smaller with the number of training steps increases, on the contrary, the RMSE and MAE of test set get bigger with the number of training steps increases. Therefore, when the number of training steps is 500, the RMSE and MAE of test set are the smallest among the four training steps. The model trained when the number of training steps is 500 can be better used to predict the ETo.
The first optimized model was used to predict ETo data in 2019, and the forecast results are shown in Table 3. As we can see from the table the MSE, RMSE, MAE and R2 between the actual value and predicted value of four different data set are similar for each site, which shows that the optimized model is suitable for these four different datasets. Based on the MAE, the data types with the highest prediction accuracy for the four stations are Type B, Type C, Type C, and Type B, respectively.
The prediction results with the highest prediction accuracy are shown in (Fig 8). As we can see from the figure that the change trend of between the predicted value and the actual value is similar, and the predicted value of the optimized model is relatively close to the true value at each station. The model predicted ETo value is less than the actual value from March to September, and the model predicted and actual ETo value are very similar from the October to the February, and the four stations have the same result.
Statistics of the error distribution between the predicted value of the model and the actual value in 2019 are shown in (Fig 9). From the figure, the errors of model predictions for the four stations are all within [–4,4] mm, and the range of the most of the prediction residual data was -1.0 to 1.0 mm. For four stations, the number of predicted values higher than the true value was 183, 164, 148, and 199, respectively, where the minimum error was -3.474 mm, -3.429 mm, -3.369 mm, and -2.239 mm, respectively. and the maximum error was 3.279 mm, 3.470 mm, 3.313 mm, and 4.155 mm, respectively. The main reason for this result is that due to the random fluctuation of meteorological data, it is difficult for the hybrid model to form a high-precision simulation of stochastic fluctuations.
In order to find out the correlation between prediction and true value, the fitting between the prediction and true values for validation set was computed. The fitting between the predicted and actual values for four stations in 2019 is shown in (Fig 10). For the four stations, the correlation coefficient of the optimized model for test set is 0.816, 0.824, 0.819, and 0.860, respectively. The experiments above also show that the first optimized model displays good prediction accuracy. It remains within a stable acceptable error range, which ensures that the daily ETo data predicted by the model can be used in actual guidance.
The result of the second optimization model.
For the second optimized model, the PSO-LSTM optimization iterated 9 times. (Fig 11) present the training time, maximum, average and minimum fitness in each iteration. From the figures, we can find that the training time increases rapidly with the number of iterations. Although the batch size = 64, the T2, N2, and N3 of hybrid model is [1,200], [1,200], and [1, 200], respectively, and D = 0.1,0.2, and 0.3, the utilization of CPU gradually decreases during the training process. The training time for the first and last iterations is 18.20h and 147.84h, respectively, and the total training time of 9 times iterations is 806.03h (33.58 days). The fitness gradually decreases as the number of iterations increases, and the minimum fitness is much smaller than the maximum fitness of each iteration. After 9 times iterations, the maximum, average and minimum are 0.619, 0.606 and 0.600, respectively. After 9 times iterations, the hyperparameter combination of the optimized model corresponding to the minimum fitness is [22, 175, 39, 0.2], it means that the look back = 22, the number of first layer neurons = 175, the number of second layer neurons = 39, and dropout = 0.2 is the optimal combination for the model.
Prediction and evaluation of the second model.
In order to evaluate the performance of the second optimized model, the model was also trained again using the four different data set of each station. The number of training steps during the optimization process is 100, in order to study the effect of different training steps on the model accuracy, the number of training steps were set as 50, 100, 150 and 200, respectively. Table 4 shows the training results.
Similar to the result of the first model, for each station and each dataset, it can be seen from the table that the RMSE and MAE of most training set gets smaller with the number of training steps increases, on the contrary, the RMSE and MAE of most test set get bigger with the number of training steps increases. Therefore, when the number of training steps is 50, the RMSE and MAE of test set are the smallest among the four training steps. Therefore, the model trained when the number of training steps is 50 can be better used to predict the ETo for each station.
The optimized model was also used to predict ETo data in 2019, and the forecast results are shown in Table 5. As we can see from the table the MSE, RMSE, MAE and R2 between the actual value and predicted value under four different data set are similar, which shows that the optimized model is suitable for these four different datasets. Based on the MAE, the data types with the highest prediction accuracy for the four stations are Type C, Type C, Type D, and Type B, respectively.
Similar to the result of the first model, the prediction results with the highest prediction accuracy are shown in (Fig 12). As we can see from the figure that the change trend of between the predicted value and the actual value is similar, and the prediction value of the optimized model is relatively close to the true value at each station. The predicted ETo value of the second model is also less than the actual value from March to September, and the predicted ETo value and actual value are also very close to the value from the October to the February. All four stations have the same result.
Statistics of the error distribution between the predicted value of the model and the actual value in 2019 are shown in (Fig 13). From the figure, the errors of model predictions for the four stations are also all within [–4,4] mm, and the range of the most of the prediction residual data was -1.0 to 1.0 mm. For four stations, the number of predicted values higher than the true value was 179, 189, 191, and 187, respectively. where the minimum error was -3.323mm, -3.192mm, -3.091mm, and -2.712mm, respectively. and the maximum error was 3.4149 mm, 3.671mm, 3.357mm, and 3.748mm, respectively. The main reason for this result also is that due to the random fluctuation of meteorological data, it is difficult for the hybrid model to form a high-precision simulation of stochastic fluctuations.
The fitting between the prediction value and actual values in 2019 is shown in (Fig 14). For the four stations, the correlation coefficient of the optimized model for test set is 0.818, 0.823, 0.822, and 0.857, respectively. The experiments above also show that the first optimized model displays good prediction accuracy. It remains within a stable acceptable error range, which ensures that the daily ETo data predicted by the model can be used in actual guidance.
Comparison of two models.
It can be seen from the two PSO-LSTM predicted results, the January, August, November and December were the months with the higher proportion of absolute errors, mainly because the value of daily ETo in these months is small, and smaller than the value of daily ETo in other months. The February, March, April and May are the months with the lower proportion of absolute error. Although the errors are relatively large in June, July, October and September, the overall errors are controlled within the range of from –1.0 to 1.0 mm. At the same time, June, July, October and September are not critical periods for the growth of crops, so the relatively low prediction accuracy has no impact on crop irrigation decision.
Compared to the first model, the second model has a larger range of hyperparameters and a more complex model topology, even though the second model has a much smaller number of training steps (100) than the first model (500), the training time of the second model is much greater than the training time cost of the first model. In fact, in the iterative process, the changes of fitness values of the two hybrid models are not significant. There are many factors that affect the performance of the hybrid model, including the characteristics of data, the hyperparameters in model and training methods. During the optimization process, the two models obtained the prediction 400 and 180 results under different hyperparameters combinations, respectively. Due to the long iteration time, the hybrid model unable to training more iterations in a short time. For the four datasets, each station obtained 8 optimized prediction models, at the same time, the prediction accuracy of the 8 models is similar. For the four sites, this study can meet the forecasting needs under different dataset. However, the optimized hybrid models provided only slight performance gains in the study.
it is impossible to consider all factors in the study. The optimized models have not exhibited high accuracies, Furthermore, deep learning models generally have more hyperparameters to be optimized with the meteorological data features.
Conclusions
In this paper, we investigate the use of the LSTM neural network to predict daily ETo. Two different topology LSTM models were constructed and optimized using the PSO algorithm hyperparameters in the LSTM neural network. The accuracy of the two hybrid models were evaluated using four different datasets in the WG, FX, XY, and PC stations, Shaanxi province, China. A single hidden layer LSTM model with 24 nodes was selected, and the value of look back was selected as 5. And the first and second hidden layer LSTM with 179 and 39, respectively, and the value of look back and dropout were selected as 22 and 0.2, respectively.
The optimized hybrid models were also predicted under different dataset, it can be found the optimized hybrid model has better accuracy under four different data set. Furthermore, the hybrid models developed in this study do not depend on external data, requiring only data measured at a local weather station, the successful validation of these models will allow agricultural producers to appropriately schedule the irrigation of crops according to ETo forecasts. In these situations, they can provide valuable information to improve tasks such as irrigation planning.
Supporting information
S1 File. The weather data and ETo of the four stations.
https://doi.org/10.1371/journal.pone.0281478.s001
(XLSX)
S2 File. The training time and fitness of two models for each iteration.
https://doi.org/10.1371/journal.pone.0281478.s002
(XLSX)
S3 File. The result of training set and test set for the first model.
https://doi.org/10.1371/journal.pone.0281478.s003
(XLSX)
S4 File. The result of prediction set for the first model.
https://doi.org/10.1371/journal.pone.0281478.s004
(XLSX)
S5 File. The result of training set and test set for the second model.
https://doi.org/10.1371/journal.pone.0281478.s005
(XLSX)
S6 File. The result of prediction set for the second model.
https://doi.org/10.1371/journal.pone.0281478.s006
(XLSX)
S7 File. The results of grid search method for two models.
https://doi.org/10.1371/journal.pone.0281478.s007
(RAR)
References
- 1. Gong X, Qiu R, Sun J, Ge J, Li Y, Wang S. Evapotranspiration and crop coefficient of tomato grown in a solar greenhouse under full and deficit irrigation. Agricultural Water Management. 2020;235.
- 2. Ferreira LB, da Cunha FF, de Oliveira RA, Fernandes Filho EI. Estimation of reference evapotranspiration in Brazil with limited meteorological data using ANN and SVM–A new approach. Journal of Hydrology. 2019;572:556–70.
- 3. Incrocci L, Thompson RB, Fernandez-Fernandez MD, De Pascale S, Pardossi A, Stanghellini C, et al. Irrigation management of European greenhouse vegetable crops. Agricultural Water Management. 2020;242.
- 4. Chia MY, Huang YF, Koo CH, Fung KF. Recent Advances in Evapotranspiration Estimation Using Artificial Intelligence Approaches with a Focus on Hybridization Techniques—A Review. Agronomy. 2020;10(1).
- 5. Gocić M, Arab Amiri M. Reference Evapotranspiration Prediction Using Neural Networks and Optimum Time Lags. Water Resources Management. 2021;35(6):1913–26.
- 6. Hebbalaguppae Krishnashetty P, Balasangameshwara J, Sreeman S, Desai S, Bengaluru Kantharaju A. Cognitive computing models for estimation of reference evapotranspiration: A review. Cognitive Systems Research. 2021;70:109–16.
- 7. Dias SHB, Filgueiras R, Fernandes Filho EI, Arcanjo GS, Silva GHD, Mantovani EC, et al. Reference evapotranspiration of Brazil modeled with machine learning techniques and remote sensing. PLoS One. 2021;16(2):e0245834. Epub 2021/02/10. pmid:33561147; PubMed Central PMCID: PMC7872264.
- 8. Ehteram M, Singh VP, Ferdowsi A, Mousavi SF, Farzin S, Karami H, et al. An improved model based on the support vector machine and cuckoo algorithm for simulating reference evapotranspiration. PLoS One. 2019;14(5):e0217499. Epub 2019/06/01. pmid:31150443; PubMed Central PMCID: PMC6544354.
- 9. Wu L, Fan J. Comparison of neuron-based, kernel-based, tree-based and curve-based machine learning models for predicting daily reference evapotranspiration. PLoS One. 2019;14(5):e0217520. Epub 2019/06/01. pmid:31150448; PubMed Central PMCID: PMC6544265.
- 10. Wu T, Zhang W, Jiao X, Guo W, Alhaj Hamoud Y. Evaluation of stacking and blending ensemble learning methods for estimating daily reference evapotranspiration. Computers and Electronics in Agriculture. 2021;184.
- 11. Wu T, Zhang W, Jiao X, Guo W, Hamoud YA. Comparison of five Boosting-based models for estimating daily reference evapotranspiration with limited meteorological variables. PLoS One. 2020;15(6):e0235324. Epub 2020/07/01. pmid:32598399; PubMed Central PMCID: PMC7347040.
- 12. Chen Z, Zhu Z, Jiang H, Sun S. Estimating daily reference evapotranspiration based on limited meteorological data using deep learning and classical machine learning methods. Journal of Hydrology. 2020;591.
- 13. Ferreira LB, da Cunha FF. New approach to estimate daily reference evapotranspiration based on hourly temperature and relative humidity using machine learning and deep learning. Agricultural Water Management. 2020;234.
- 14. Granata F, Di Nunno F. Forecasting evapotranspiration in different climates using ensembles of recurrent neural networks. Agricultural Water Management. 2021;255.
- 15. Jayasinghe WJMLP Deo RC, Ghahramani A, Ghimire S, Raj N. Deep Multi-Stage Reference Evapotranspiration Forecasting Model: Multivariate Empirical Mode Decomposition Integrated With the Boruta-Random Forest Algorithm. IEEE Access. 2021;9:166695–708.
- 16. Long X, Wang J, Gong S, Li G, Ju H. Reference evapotranspiration estimation using long short‐term memory network and wavelet‐coupled long short‐term memory network. Irrigation and Drainage. 2022.
- 17. de Oliveira e Lucas P, Alves MA, de Lima e Silva PC, Guimarães FG. Reference evapotranspiration time series forecasting with ensemble of convolutional neural networks. Computers and Electronics in Agriculture. 2020;177.
- 18. Yin J, Deng Z, Ines AVM, Wu J, Rasu E. Forecast of short-term daily reference evapotranspiration under limited meteorological variables using a hybrid bi-directional long short-term memory model (Bi-LSTM). Agricultural Water Management. 2020;242.
- 19. Afzaal H, Farooque AA, Abbas F, Acharya B, Esau T. Computation of Evapotranspiration with Artificial Intelligence for Precision Water Resource Management. Applied Sciences. 2020;10(5).
- 20. Sharma V, Nicholson C, Bergantino A, Irmak S, Peck D. Temporal Trend Analysis of Meteorological Variables and Reference Evapotranspiration in the Inter-mountain Region of Wyoming. Water. 2020;12(8).
- 21. Jung D-H, Kim HS, Jhin C, Kim H-J, Park SH. Time-serial analysis of deep neural network models for prediction of climatic conditions inside a greenhouse. Computers and Electronics in Agriculture. 2020;173.
- 22. Bian J, Wang L, Scherer R, Wozniak M, Zhang P, Wei W. Abnormal Detection of Electricity Consumption of User Based on Particle Swarm Optimization and Long Short Term Memory With the Attention Mechanism. IEEE Access. 2021;9:47252–65.
- 23. Wang Y, Lv L, Kong W, Qi J, Zhang J. An improved long short-term memory neural network for stock forecast. MATEC Web of Conferences. 2018;232.
- 24. Wang X, Zhang X, Yang M, Gou X, Liu B, Hao Y, et al. Multi-Site Evaluation of Accumulated Temperature and Rainfall for Maize Yield and Disease in Loess Plateau. Agriculture. 2021;11(4).
- 25. Zhao S, Tie X, Cao J, Li N, Li G, Zhang Q, et al. Seasonal variation and four-year trend of black carbon in the Mid-west China: The analysis of the ambient measurement and WRF-Chem modeling. Atmospheric Environment. 2015;123:430–9.
- 26. Allen R, Pereira L, Raes D, Smith M, Allen RG, Pereira LS, et al. Crop Evapotranspiration: Guidelines for Computing Crop Water Requirements, FAO Irrigation and Drainage Paper 56. FAO. 1998;56.
- 27. Wang S, Li P, Ji H, Zhan Y, Li H. Prediction of air particulate matter in Beijing, China, based on the improved particle swarm optimization algorithm and long short-term memory neural network. Journal of Intelligent & Fuzzy Systems. 2021;41(1):1869–85.
- 28. Cao Z, Han X, Lyons W, O’Rourke F. Energy management optimisation using a combined Long Short-Term Memory recurrent neural network–Particle Swarm Optimisation model. Journal of Cleaner Production. 2021;326.
- 29. He Q-Q, Wu C, Si Y-W. LSTM with particle Swam optimization for sales forecasting. Electronic Commerce Research and Applications. 2022;51.
- 30. Feng R, Fan G, Lin J, Yao B, Guo Q. Enhanced Long Short-Term Memory Model for Runoff Prediction. Journal of Hydrologic Engineering. 2021;26(2).
- 31. Gundu V, Simon SP. PSO–LSTM for short term forecast of heterogeneous time series electricity price signals. Journal of Ambient Intelligence and Humanized Computing. 2020;12(2):2375–85.
- 32. Zhang Q, Pan W, Li C, Wang X. The conditional distance autocovariance function. Canadian Journal of Statistics. 2021;49(4):1093–114.
- 33. Christ M, Braun N, Neuffer J, Kempa-Liehr AW. Time Series FeatuRe Extraction on basis of Scalable Hypothesis tests (tsfresh–A Python package). Neurocomputing. 2018;307:72–7.
- 34. Kara A. A data-driven approach based on deep neural networks for lithium-ion battery prognostics. Neural Computing and Applications. 2021;33(20):13525–38.
- 35. Harjoseputro Yulius. A Classification Javanese Letters Model using a Convolutional Neural Network with KERAS Framework[J]. International Journal of Advanced Computer Science and Applications (IJACSA),2020,11(10).
- 36. Ferreira LB, da Cunha FF, Zanetti SS. Selecting models for the estimation of reference evapotranspiration for irrigation scheduling purposes. PLoS One. 2021;16(1):e0245270. Epub 2021/01/12. pmid:33428674; PubMed Central PMCID: PMC7799801.